Let's reverse engineer this demo. You need 3 things: (1) robust hardware and motor designs that treat simulation as first-class citizen; (2) a human motion capture ("mocap") dataset, such as those for film and gaming characters; (3) massively parallel RL training in GPU-accelerated simulation. Last October, our team trained a 1.5M parameter foundation model called HOVER for such agile motor control. It follows this recipe, roughly speaking: (1) Simulation used to be an after-thought. Now, it has to be part of the hardware design process. If your robot doesn't simulate well, you can kiss RL goodbye. Hardware-simulation co-design is a very interesting emergent topic that only becomes meaningful with today's compute capability. (2) Human mocap dataset to produce natural-looking walking and running gaits. That's one huge advantage of using humanoid robot - you get to imitate from tons of human motions that were originally captured for movies or AAA games. At least 3 ways to use the data: - For initialization: pre-train the neural net to imitate human, and then finetune it into the robot form factor with physics turned on; - For reward function: penalize any deviations from the target pose; - For representation learning: treat the human poses as a "motion prior" to constrain the space of robot behaviors. (3) Shove the above into Isaac sim, add a lot of randomization, pump it through PPO, throw in a bunch of GPUs, and then watch Netflix till loss converges. If you have an urge to comment this is CGI, let me save you a few keystrokes — many academic labs now own the G1 robot in the flesh. Read about our team's HOVER work: https://lnkd.in/gfKW9K5U
Advanced Robotics Applications In Engineering
Explore top LinkedIn content from expert professionals.
-
-
🚀 Meet RAVEN: The Flying Robot That Walks, Jumps, and Soars 🦅 Drones are clumsy. They need open space, stable launch points, and struggle with rough terrain. Birds, on the other hand, dominate both air and land. That’s exactly what researchers at EPFL’s Laboratory of Intelligent Systems have captured in RAVEN—a robotic bird that walks, hops, jumps, and flies. 🔥 Inspired by ravens and crows, RAVEN’s multifunctional legs allow it to take off without a runway, land on rough surfaces, and even traverse obstacles that ground-based robots can’t handle. Traditional flying robots had to choose: either walk or fly—RAVEN does both. ✨ Why this matters: 🔹 Built for agility – It can jump-start its flight, making takeoff more energy-efficient. ⚡ 🔹 Nature’s blueprint, optimized – Lightweight avian-inspired legs mimic tendons and muscles. 🦵 🔹 Real-world impact – Imagine drones that can land in disaster zones, navigate tight spaces, or deliver aid without human intervention. 🎯 The future of robotics isn’t about copying nature—it’s about surpassing it. RAVEN isn’t just a flying robot. It’s a glimpse of what’s next: machines that move seamlessly across worlds, just like nature intended. 🌍✨ 🤔 What other real-world challenges do you think robots like RAVEN could help solve? Drop your thoughts below! ⬇️ #AI #Robotics #FlyingRobots #Drones #Innovation #FutureTech #Biomimicry #Aerospace #TechForGood
-
Presenting FEELTHEFORCE (FTF): a robot learning system that models human tactile behavior to learn force-sensitive manipulation. Using a tactile glove to measure contact forces and a vision-based model to estimate hand pose, they train a closed-loop policy that continuously predicts the forces needed for manipulation. This policy is re-targeted to a Franka Panda robot with tactile gripper sensors using shared visual and action representa- tions. At execution, a PD controller modulates gripper closure to track predicted forces -enabling precise, force-aware control. This approach grounds robust low- level force control in scalable human supervision, achieving a 77% success rate across 5 force-sensitive manipulation tasks. #research: https://lnkd.in/dXxX7Enw #github: https://lnkd.in/dQVuYTDJ #authors: Ademi Adeniji, Zhuoran (Jolia) Chen, Vincent Liu, Venkatesh Pattabiraman, Raunaq Bhirangi, Pieter Abbeel, Lerrel Pinto, Siddhant Haldar New York University, University of California, Berkeley, NYU Shanghai Controlling fine-grained forces during manipulation remains a core challenge in robotics. While robot policies learned from robot-collected data or simulation show promise, they struggle to generalize across the diverse range of real-world interactions. Learning directly from humans offers a scalable solution, enabling demonstrators to perform skills in their natural embodiment and in everyday environments. However, visual demonstrations alone lack the information needed to infer precise contact forces.
-
Micro drones are no longer niche tools — they are becoming a core pillar of surveillance, security, and tactical intelligence across defense, public safety, and critical infrastructure. Have you seen this one? What’s remarkable is not just the capability — it’s the speed of evolution. 📈 The Numbers Behind the Momentum • The global micro-drone market is growing at 16–19% CAGR, with forecasts projecting: • From ~$10B in 2024 to over $24B by 2029 • Small UAV market expected to exceed $11B by 2030 • Defense and surveillance account for one of the largest and fastest-growing segments due to: • Border security expansion • Urban surveillance demand • ISR (Intelligence, Surveillance, Reconnaissance) modernization 🧠 What Changed the Game? Modern micro drones now combine: • AI-powered navigation & object recognition • Real-time video transmission • Autonomous flight and obstacle avoidance • Swarm coordination capabilities • Ultra-miniaturized thermal + optical sensors Some nano-drones weigh under 20 grams, fly for 20–25 minutes, and transmit encrypted HD video over 1.5–2 km, all while operating with extremely low acoustic signatures. This level of capability was military-exclusive just a few years ago. Today, it’s rapidly becoming standard Micro surveillance drones are now actively used for: • Tactical reconnaissance in conflict zones • Law enforcement situational awareness • Crowd monitoring & perimeter security • Disaster response in collapsed or dangerous environments • Critical infrastructure inspection (energy, transport, telecom) At the tactical level, they allow frontline units to “see first” before entering hostile or uncertain environments — reducing risk and improving decision speed. 🤖 The Rise of Swarm Intelligence One of the most disruptive developments is coordinated micro-drone swarms: • Multiple drones operating as a single intelligent system • Real-time terrain mapping • Autonomous target identification • Dynamic mission adaptation This shifts surveillance from isolated viewpoints to distributed intelligence networks in the air. ⚠️ The Strategic Challenge With power comes responsibility. Micro drone surveillance forces critical conversations around: • Privacy and civil liberties • Airspace governance • Ethical deployment • Counter-drone defense systems • Digital sovereignty At the same time, governments and enterprises are investing heavily in anti-drone and RF-neutralization technologies, signaling that the drone vs counter-drone race has already begun. #Drones #SurveillanceTechnology #DefenseTech #AI #AutonomousSystems #SecurityInnovation #FutureOfSurveillance
-
Robotic assembly is proving to be increasingly useful in various applications. A recent demo from Kyber Labs showcases a robot assembling a spring-loaded pin endstop, inspired by a real aerospace component. The full sequence runs end-to-end, including: - Picking parts - Inserting the pin - Threading standard M6 (and larger) nuts - Performing in-hand adjustments along the way While each of these steps may seem straightforward for a human, the challenge lies in executing them reliably, thousands of times, without relying on fixtures tailored to a single geometry. What is particularly noteworthy in this demonstration is not the speed or precision, but the generality of the system. This robotic setup can manage insertion, fastening, and manipulation without being confined to a single task. This flexibility allows for easier integration into existing production setups, enabling operation only when necessary and the ability to adapt to nearby variants without extensive retooling.
-
A few years ago, I learned the hard way that jumping straight into hardware, sensors, motors, and wiring can lead to costly mistakes and late-night headaches. That’s when I discovered the true importance of #simulation in robotics and engineering. During the early phase of my final-year thesis, I spent weeks recreating our school cafeteria with Iman Tokosi in Blender, exporting it as an SDF model and loading it into Gazebo using #ROS2. Suddenly, I could drive a virtual robot through aisles and around tables without the fear of damaging anything real. It was challenging and eye-opening, and it saved me countless hours and resources. Then came the moment that changed everything: integrating #SLAM so the robot could build its own map while moving, and setting up #Nav2 to let it plan and follow paths autonomously. Watching it navigate the environment with precision and independence was a powerful confirmation that the system worked. Now, imagine a world where every structure, product, and system is simulated down to the smallest detail. The result? Reduced costs, faster development, increased reliability, enhanced safety, and stronger adherence to standards. Some may still view simulation as “just for show,” but I’ve experienced firsthand that it’s the foundation of true innovation. Are you leveraging simulation in your next robotics or engineering project? Let’s connect and exchange ideas!
-
+6
-
Very promising! A new open-source platform for research on Human-AI teaming from Duke University uses real-time human physiological and behavioral data such as eye gaze, EEG, ECG, across a wide range of test situations to identify how to improve Human-AI collaboration. Selected insights from the CREW project paper (link in comments): 💡 Comprehensive Design for Collaborative Research. CREW is built to unify multidisciplinary research across machine learning, neuroscience, and cognitive science by offering extensible environments, multimodal feedback, and seamless human-agent interactions. Its modular design allows researchers to quickly modify tasks, integrate diverse AI algorithms, and analyze human behavior through physiological data. 🔄 Real-Time Interaction for Dynamic Decision-Making. CREW’s real-time feedback channels enables researchers to study dynamic decision-making and adaptive AI responses. Unlike traditional offline feedback systems, CREW supports continuous and instantaneous human guidance, crucial for simulating real-world scenarios, and making it easier to study how AI can best align with human intentions in rapidly changing environments. 📊 Benchmarking Across Tasks and Populations. CREW enables large-scale benchmarking of human-guided reinforcement learning (RL) algorithms. By conducting 50 parallel experiments across multiple tasks, researchers could test the scalability of state-of-the-art frameworks like Deep TAMER. This ability to scale the study of the interaction of human cognitive traits with AI training outcomes is a first. 🌟 Cognitive Traits Driving AI Success. The study highlighted key human cognitive traits—spatial reasoning, reflexes, and predictive abilities—as critical factors in enhancing AI performance. Overall, individuals with superior cognitive test scores consistently trained better-performing agents, underscoring the value of understanding and leveraging human strengths in collaborative AI development. Given that Humans + AI should be at the heart of progress, this platform promises to be a massive enabler of better Human-AI collaboration. In particular, it can help in designing human-AI interfaces that apply specific human cognitive capabilities to improve AI learning and adaptability. Love it!
-
Imagine smarter robots for your business. New research from Google puts advanced Gemini AI directly into robots, which can now understand complex instructions, perform intricate physical tasks with dexterity (like assembly) and adapt to new objects or situations in real time. The paper introduces "Gemini Robotics," a family of AI models based on Google's Gemini 2.0, designed specifically for robotics. They present Vision-Language-Action (VLA) models capable of direct robot control, performing complex, dexterous manipulation tasks smoothly and reactively. The models demonstrate generalization to unseen objects and environments and can follow open-vocabulary instructions. It also introduces "Gemini Robotics-ER" for enhanced embodied reasoning (spatial/temporal understanding, detection, prediction), bridging the gap between large multimodal models and physical robot interaction. Here's why this matters: At scale, this will unlock more flexible, intelligent automation for the future of manufacturing, logistics, warehousing, and more, potentially boosting efficiency and enabling tasks previously too complex for robots as we've imagined in the past. Very, very promising! (Link in the comments.)
-
From Image to 3D: Why Lyra Could Unstick Brownfield Workflows Most brownfield digital-twin projects stall on day one: scanning is costly, photogrammetry is fussy, and CAD clean-up eats budgets. NVIDIA Research’s Lyra points to a different on-ramp. The workflow: > Ingest a single image (or video). > Self-distill a pre-trained, camera-controlled video diffusion model into a 3D Gaussian Splat (3DGS) scene—no real multi-view training data required. > Feed-forward generate 3D (and even 4D, i.e., dynamic) scenes for real-time rendering. > Export to simulation (e.g., Isaac Sim) to test perception, locomotion, and task policies before you touch the real site. Why this matters for industry: > Instead of waiting weeks for LiDAR + mesh clean-up, teams can bootstrap a plausible 3D context from minimal visuals—good enough to start layout studies, Gemba walk-throughs, and retrofit planning. > Physical AI Data flywheel: Robots and agents need diverse, realistic worlds. Lyra’s pipeline can go text→image→3D and text→video→4D, then drop those 3DGS worlds into Isaac Sim for policy training and regression tests—exactly the kind of closed loop we want for factory autonomy and site inspection. > Faster iteration, lower embodied carbon: Early feasibility and clash checks happen in synthetic scenes, reducing site visits, rework, and wasted materials—small steps that add up on SDG-aligned projects. This is not about throwing away scanning or BIM but rather giving teams a fast start so design, safety, and robotics folks can begin validating assumptions in few hours vs weeks. If you are in the business of industrial twins and Physical AI, this is for you. Read more about Lyra : https://lnkd.in/gpTjtxEx Our Model and code: https://lnkd.in/gtVJFhFE #NVIDIAResearch #Lyra #3DGaussianSplatting #DigitalTwin #PhysicalAI #BrownfieldEngineering #IndustrialDigitalization #RoboticsTraining #Simulation #AI4Industry #OpenUSD #Omniverse #IndustrialMetaverse #SustainableAI University of Toronto Vector Institute Simon Fraser University NVIDIA
-
𝗗𝗶𝗱 𝘆𝗼𝘂 𝗽𝗹𝗮𝗰𝗲 𝗮 "𝗖𝗮𝗻𝗮𝗿𝘆" 𝗶𝗻 𝘆𝗼𝘂𝗿 𝗣𝗿𝗼𝗰𝗲𝘀𝘀 𝗔𝘂𝘁𝗼𝗺𝗮𝘁𝗶𝗼𝗻? The sort of early warning detection system which monitors your automated processes and sings when irregularities occur? Why a 🐤 𝗰𝗮𝗻𝗮𝗿𝘆 you ask? Around 1911, miners started to take canary birds into the coal mines to detect the accumulation of toxic gases. These birds, would even sense the smallest traces and emissions, starting to erratically chirp and with that giving miners early warnings to immediately evacuate the mine. Just as the canaries once did in the mines, 𝗮 𝗱𝗶𝗴𝗶𝘁𝗮𝗹 "𝗰𝗮𝗻𝗮𝗿𝘆" can play a vital role in monitoring the health of your automated workflows signalling potential issues before they escalate and perhaps, cause scaled harm. But how do you implement a digital canary into your workflows in your process automation? 𝗜𝗻𝗰𝗼𝗿𝗽𝗼𝗿𝗮𝘁𝗲 𝗶𝘁 𝗳𝗿𝗼𝗺 𝘀𝘁𝗮𝗿𝘁: into your design by using code, reconciliation reports, and validation rules to establish effective in-process control checks and monitoring mechanisms and visual dashboards to analyse red flags. Find here 5 examples how to get early alerts in your process automation, even if your automation bots don't know how to sing: ▪️𝗣𝗿𝗼𝗰𝗲𝘀𝘀𝗶𝗻𝗴 𝗖𝗵𝗲𝗰𝗸𝘀: Implement automated checks at various stages of the process to ensure accuracy and completeness and volume variations. ▪️𝗜𝗻𝘁𝗲𝗴𝗿𝗮𝘁𝗶𝗼𝗻 𝗘𝗿𝗿𝗼𝗿 𝗩𝗮𝗹𝗶𝗱𝗮𝘁𝗶𝗼𝗻: Monitor integration and break points like API's for errors or failures to maintain seamless data flow across systems. ▪️𝗗𝗮𝘁𝗮 𝗜𝗻𝘁𝗲𝗴𝗿𝗶𝘁𝘆 𝗦𝗰𝗮𝗻𝘀: Validate for duplicate records or inconsistencies to maintain data integrity and remove manual overrides or corrections. ▪️𝗨𝘀𝗮𝗯𝗶𝗹𝗶𝘁𝘆 𝗙𝗲𝗲𝗱𝗯𝗮𝗰𝗸𝘀: Analyse insights from user feedbacks to check on usability issues, frequent issues and detect sentiment drops with NLP / AI. ▪️𝗖𝗼𝗺𝗽𝗹𝗶𝗮𝗻𝗰𝗲 𝗖𝗼𝗰𝗸𝗽𝗶𝘁: Create a centralised dashboard to monitor compliance metrics to detect red flags and and detect deviations from policies. By integrating digital canaries into your process automation strategy, you are not only enhance your ability to detect and respond to issues rapidly but also promote a culture of self-monitoring and continuous improvement. So, did you already place a digital "canary" into your process design and automations? If not, maybe it's time to reconsider adding this early warning system to your automation approach ensuring the health and resilience of your tasks, data & process performance. What early warning systems have worked for you best? #processautomation #intelligentautomation #rpa #processexcellence