RIEM News LogoRIEM News

Articles tagged with "reinforcement-learning"

  • NVIDIA tech helps humanoid robot beat human operators at opening doors

    NVIDIA researchers have developed “DoorMan,” a robotic learning system enabling a humanoid robot—the $16,000 Unitree G1—to open doors more efficiently than human operators. Utilizing only built-in RGB cameras and trained entirely through simulation-based reinforcement learning in NVIDIA’s Isaac Lab, the system allows the robot to open various real-world doors faster and with higher success rates than humans remotely controlling it. In tests, DoorMan completed door-opening tasks up to 31% faster than expert teleoperators and achieved an 83% success rate, outperforming both expert (80%) and non-expert (60%) human operators. This advancement represents significant progress in “loco-manipulation,” where robots must simultaneously walk, perceive, coordinate limbs, and manipulate objects. The DoorMan system employs a novel pixel-to-action training approach, relying solely on raw RGB input without specialized sensors like depth cameras or motion-capture markers. To overcome common reinforcement learning challenges, the researchers introduced a “staged-reset” mechanism

    roboticshumanoid-robotreinforcement-learningNVIDIA-Isaac-Labrobotic-manipulationAI-roboticsdoor-opening-robot
  • Teaching robot policies without new demonstrations: interview with Jiahui Zhang and Jesse Zhang - Robohub

    The article discusses the ReWiND framework introduced by Jiahui Zhang, Jesse Zhang, and colleagues in their CoRL 2025 paper, which enables robots to learn manipulation policies for novel, language-specified tasks without requiring new task-specific demonstrations. ReWiND operates in three stages: first, it learns a dense reward function from a small set of demonstrations in the deployment environment by predicting per-frame progress toward task completion. A novel video rewind augmentation technique is used to synthetically generate sequences simulating both progress and failure, improving the reward model’s accuracy and generalization. Second, the framework pre-trains a policy offline using these dense rewards relabeled on the demonstration data. Finally, the pre-trained policy is fine-tuned online in the deployment environment on unseen tasks, using the frozen reward function to provide feedback without additional demonstrations. The researchers evaluated ReWiND in both simulated (MetaWorld) and real-world (Koch) environments, focusing on the reward model’s generalization and the policy

    roboticsrobot-learningreinforcement-learninglanguage-guided-roboticsrobot-manipulationreward-function-learningpolicy-adaptation
  • Humanoid says its first bipedal robot can start walking just 48 hours after assembly - The Robot Report

    Humanoid, a London-based robotics company founded in 2024, has introduced its first humanoid bipedal robot, the HMND 01 Alpha Bipedal, developed from design to working prototype in just five months. Remarkably, the robot achieved stable walking within 48 hours after final assembly. Standing 179 cm tall with 29 degrees of freedom and a bimanual payload capacity of 15 kg, Alpha Bipedal is equipped with advanced sensors including six RGB cameras, depth sensors, a microphone array, and haptic and force sensors. Powered by NVIDIA Jetson Orin AGX and Intel i9 processors, it offers three hours of swappable battery life. The robot can perform a variety of movements such as walking on straight and curved paths, turning, sidestepping, squatting, hopping, running, and precise manipulation, and it can interact with people and coordinate with other robots. Humanoid utilized ultra-precise 3D modeling and NVIDIA’s Isaac Sim

    roboticshumanoid-robotbipedal-robotAIautomationNVIDIA-Jetsonreinforcement-learning
  • NVIDIA debuts first open reasoning AI for self-driving vehicles

    NVIDIA has introduced a suite of open-source AI models and tools aimed at advancing autonomous vehicles, robotics, and speech processing. Central to this launch is Alpamayo-R1 (AR1), the world’s first open reasoning vision-language-action (VLA) model designed for self-driving cars. AR1 integrates chain-of-thought reasoning with path planning to navigate complex driving scenarios by evaluating possible trajectories and contextual data, enabling human-like decision-making in challenging environments such as crowded intersections or lane closures. Built on NVIDIA’s Cosmos Reason platform, AR1 is available for customization by researchers for non-commercial use and has demonstrated improved reasoning capabilities through reinforcement learning post-training. Beyond AR1, NVIDIA’s Cosmos platform offers additional tools for physical AI development, including LidarGen for generating lidar data, Omniverse NuRec Fixer for neural reconstruction cleanup, Cosmos Policy for robot behavior creation, and ProtoMotions3 for training humanoid robots in simulated settings. These resources are already being utilized by ecosystem partners and academic

    robotautonomous-vehiclesAI-modelsNVIDIA-Cosmosreinforcement-learninglidar-simulationhumanoid-robots
  • US firm teaching humanoid robot brains to do laundry, make coffee, light candles

    Physical Intelligence (PI), a Silicon Valley robotics startup, is advancing the development of humanoid robots capable of learning and reliably performing complex physical tasks such as folding laundry, making coffee, and lighting candles. The company recently raised $400 million from investors including OpenAI and Jeff Bezos, valuing it above $2 billion. PI’s innovation centers on a new training method called Recap (Reinforcement Learning with Experience and Corrections via Advantage-conditioned Policies), which enables robots to learn more like humans—through instruction, correction, and autonomous practice—addressing a key challenge in robotics where small errors during task execution often compound and cause failure. Recap enhances robot learning by incorporating corrective human interventions when errors occur and by allowing the robot to evaluate its own actions using reinforcement learning. This approach uses a value function to assign credit or blame to specific moves, enabling the system to learn from imperfect experiences rather than discarding them. PI’s vision-language-action model, π*0.6, trained with Rec

    roboticshumanoid-robotsAI-trainingreinforcement-learningrobotic-manipulationphysical-intelligenceautomation
  • Video: Humanoid robot goes 'mountain-grade' while picking up litter

    Flexion Robotics has released a demonstration video showcasing its humanoid robot equipped with a new autonomy framework that enables it to navigate challenging outdoor terrain and perform litter cleanup tasks independently, without prior training. The robot identifies scattered objects, picks them up, and deposits them into a trash bin, highlighting advances in real-world robotic autonomy. The company’s technology integrates reinforcement learning and sim-to-real transfer to train low-level motor skills in simulation, which are then executed reliably on physical robots. This approach addresses the scalability challenge of collecting real-world data for every possible scenario by combining core learned skills with high-level decision-making powered by large language and vision-language models. The system is structured as a three-layer modular hierarchy: at the top, a language or vision-language model handles task planning, common-sense reasoning, and breaking down goals into actionable steps; the middle layer generates safe, short-range motions based on perception and instructions; and the base layer uses reinforcement learning controllers to execute these motions robustly across different environments and robot

    roboticshumanoid-robotreinforcement-learningrobot-autonomysim-to-real-transfermachine-learningrobotic-manipulation
  • Flexion to use Series A to build sim-to-real, AI systems powering humanoids - The Robot Report

    Flexion Robotics AG, a Zurich-based startup founded in 2024, has raised $50 million in Series A funding to develop a reinforcement learning and sim-to-real AI platform designed to power humanoid robots across various morphologies and tasks. The company leverages generative AI and large language models (LLMs) to move beyond brittle, task-specific robotic systems by enabling robots to perform reasoning, writing, and creative tasks autonomously. Flexion’s autonomy stack consists of three layers: a command layer using language models for common-sense reasoning and task decomposition; a motion layer featuring a vision-language-action model trained primarily on synthetic data and fine-tuned for real-world scenarios; and a control layer utilizing transformer-based, low-latency whole-body control with a modular skill library for rapid behavior composition. Flexion’s AI architecture, exemplified by its Reflect v0 system, integrates LLM and vision-language model agents for task scheduling and tool selection, a general motion generator for collision-aware trajectory planning,

    roboticshumanoid-robotsartificial-intelligencereinforcement-learningsim-to-reallarge-language-modelsautomation
  • Humanoid robot surprises scientists by mastering key basketball skills

    Researchers at the Hong Kong University of Science and Technology (HKUST) have developed a humanoid robot, a Unitree G1 nicknamed “Little Potato,” capable of performing advanced basketball skills such as dribbling, passing, and executing a clean three-step layup. The robot’s fluid and stable movements result from SkillMimic, an AI framework that learns from human demonstrations captured via video and motion-tracking suits. SkillMimic refines these movements through extensive virtual training, enabling the robot to switch smoothly between multiple skills—even transitions not explicitly shown in the training data. During demonstrations, the robot displayed impressive balance and adaptability, quickly recovering after a blocked layup attempt. SkillMimic-V2 addresses key challenges in reinforcement learning from imperfect human motion data by employing three innovations: the Stitched Trajectory Graph (STG) connects similar body positions across different skills to create new transition paths; the State Transition Field (STF) trains the robot to recover from errors by starting

    robothumanoid-robotAIreinforcement-learningroboticsSkillMimicrobot-basketball-skills
  • US: Robot dog balances rolling load on back with tactile sensing tech

    Researchers at Carnegie Mellon University have developed LocoTouch, a novel tactile sensing system that enables quadrupedal robots to carry loose, unsecured cylindrical and irregularly shaped objects on their backs without the items rolling off during movement. Unlike traditional robots that rely on rigid containers or mounted boxes to secure cargo, LocoTouch uses a high-density tactile sensor array made from piezoresistive film and conductive electrodes spread flat across the robot’s back. This sensor continuously detects shifts in the load’s position and orientation, allowing the robot to adjust its posture and gait in real time to maintain balance, similar to how humans instinctively stabilize objects while walking. The system was trained through reinforcement learning using over 4,000 digital twins in simulation, enabling the robot to experience a wide variety of object movements and disturbances. The learned balancing skills transferred directly to a physical Unitree Go1 quadruped robot, which successfully carried various objects over 60 meters, navigating obstacles and remaining stable even when bumped. This represents the

    robottactile-sensingquadrupedal-robotsreinforcement-learningrobotic-balancepiezoresistive-sensorsrobotic-assistants
  • Video: New brain helps humanoid robot to clean spaces without prior training

    Swiss company Flexion Robotics has developed a novel “brain” for humanoid robots that enables them to autonomously perform diverse tasks, such as cleaning spaces, without prior training or human intervention. Their general-purpose autonomy stack integrates multiple layers: a command layer using language models for common-sense reasoning and task decomposition; a motion layer combining vision, language, and action models trained on synthetic data and fine-tuned for real-world scenarios; and a control layer employing transformer-based whole-body control with a modular skill library. This architecture allows robots to quickly adapt to and interact with their environment intelligently, overcoming the limitations of brittle, task-specific programming. Flexion emphasizes that while humanoid robots have impressive mechanics, the key missing element for practical deployment in uncontrolled settings has been intelligence. By leveraging the same computational and training infrastructure that enabled large language models (LLMs), Flexion aims to create adaptable, autonomous systems capable of scalable, useful work. This development is particularly timely given global demographic shifts and labor shortages, with

    roboticshumanoid-robotsautonomous-systemsreinforcement-learningartificial-intelligencemotion-controlrobotics-innovation
  • Disney trains robots to fall, roll, and land safely without damage

    Disney researchers, collaborating with university engineers, have developed a reinforcement learning-based system that enables bipedal robots to fall safely by controlling their landing poses to protect sensitive components. Traditional robots often suffer damage from uncontrolled falls due to stiff joints or flailing limbs, leading to costly repairs. Instead of resisting gravity, the new approach teaches robots to absorb impacts by rolling or shifting limbs during a fall to land in stable, damage-minimizing positions, prioritizing damage prevention over strict balance control. The training involved thousands of simulated falls with randomized velocities and directions, allowing the robot to learn a variety of safe landing strategies. A scoring system rewarded moves that reduced impact forces and protected vulnerable parts like the head and battery pack, while penalizing erratic motions. The researchers generated 24,000 stable poses, including artist-designed ones within realistic joint limits, to expand the robot’s repertoire of safe landings. After two days of training on powerful GPUs, the learned policy was transferred to a real 16-kil

    roboticsreinforcement-learningrobot-safetybipedal-robotsrobot-fall-protectionDisney-researchrobot-simulation
  • World's first AI firefighting system extinguishes oil fires on ships

    The Korea Institute of Machinery and Materials (KIMM) has developed the world’s first AI-driven autonomous firefighting system specifically designed to detect and extinguish oil fires aboard naval vessels, even under challenging sea conditions. Unlike traditional systems that flood entire compartments with extinguishing agents, KIMM’s technology uses AI-based fire verification and reinforcement learning to accurately identify real fires and target suppression precisely at the source. This approach minimizes unnecessary damage from false alarms. The system integrates sensors, fire monitors, and a control unit capable of estimating fire location with over 98% accuracy, and can discharge foam up to 24 meters. It has been successfully tested in simulated ship compartments and real-world conditions aboard the ROKS Ilchulbong amphibious assault ship, demonstrating stable operation in waves up to one meter high. Developed by Senior Researcher Hyuk Lee and his team, the system adapts to ship movement using a reinforcement learning algorithm that adjusts nozzle aiming based on six degrees of freedom acceleration data. It

    AIautonomous-systemsfirefighting-technologyroboticssensorsreinforcement-learningmaritime-safety
  • World's first AI firefighting system extinguishes oil fires on ships

    The Korea Institute of Machinery and Materials (KIMM) has developed the world’s first AI-driven autonomous fire suppression system specifically designed to detect and extinguish oil fires aboard naval vessels, even under challenging sea conditions. Utilizing reinforcement learning, the system compensates for ship motion by continuously adjusting its nozzle aiming angle based on acceleration data, enabling it to accurately target fire sources up to 24 meters away. Unlike traditional systems that flood entire compartments, KIMM’s technology precisely directs foam only at confirmed fire locations, reducing unnecessary damage from false alarms. The system integrates sensors, fire monitors, and an AI-based control unit that verifies fire authenticity with over 98% accuracy and adapts to sea states of level 3 or higher. Extensive testing was conducted in a full-scale simulation facility replicating ship compartments and various fire scenarios, including open-area and shielded oil fires typical on aircraft carriers. Subsequent real-ship trials aboard the ROKS Ilchulbong amphibious assault ship demonstrated the system

    robotAIautonomous-systemsfire-suppressionreinforcement-learningmaritime-safetyKorea-Institute-of-Machinery-and-Materials
  • AgiBot deploys its Real-World Reinforcement Learning system - The Robot Report

    AgiBot has achieved a significant milestone by deploying its Real-World Reinforcement Learning (RW-RL) system in a manufacturing pilot with Longcheer Technology. This marks the first time AgiBot’s RW-RL has been applied on an active production line, bridging advanced AI innovation with large-scale precision manufacturing. The system enables robots to quickly learn and adapt to complex assembly tasks directly on the factory floor, reducing training times from weeks to minutes and allowing for flexible reconfiguration with minimal hardware changes. This approach addresses long-standing challenges in precision manufacturing, such as the rigidity of traditional automation systems, lengthy deployment cycles, and maintenance complexity. The RW-RL system offers several key advantages: rapid deployment, high adaptability to variations in part positioning and tolerances, and flexible reconfiguration without the need for custom fixtures or tooling. AgiBot emphasizes the system’s industrial-grade stability, maintaining a 100% task completion rate over extended operations, and its generality across different workspace layouts and production lines,

    robotreinforcement-learningmanufacturing-automationAI-roboticsindustrial-robotsflexible-manufacturingintelligent-automation
  • US team's sound-guided drones can fly where cameras fail to see

    Researchers at Worcester Polytechnic Institute (WPI) are developing tiny aerial robots that navigate using sound rather than traditional cameras or light sensors, enabling operation in environments with smoke, dust, or darkness where vision-based systems fail. Inspired by the echolocation abilities of bats and birds, the project combines metamaterials to reduce propeller noise, alternative propulsion methods like flapping wings, and bio-inspired designs to improve ultrasonic signal capture and emission. These drones will be compact—under 100 millimeters and 100 grams—and aim to be energy-efficient, affordable, and capable of autonomous navigation in challenging conditions. Funded by a $704,908 National Science Foundation grant over three years starting in September 2025, the project integrates physics-informed deep learning and hierarchical reinforcement learning to process ultrasonic signals and enable obstacle avoidance and goal-directed movement. Sensor fusion combining echolocation with inertial and other data enhances situational awareness and reliability. The research seeks to create deployable drone swarms for search, rescue, and hazardous environment

    roboticsdronesmetamaterialsbio-inspired-navigationultrasonic-sensingreinforcement-learningaerial-robots
  • China's humanoid robot takes over presentation, car salesperson gig

    China’s automaker Chery, in collaboration with AiMOGA Robotics, unveiled Mornine, a humanoid robot designed to integrate automotive technology with embodied intelligence. At the AiMOGA Global Business Conference in Wuhu, China, Mornine delivered a 30-minute multilingual presentation on robotics and automotive innovations, acted as an autonomous car sales assistant by greeting visitors, explaining car features, and even opening a car door—making it the world’s first humanoid robot to do so autonomously. Mornine’s capabilities stem from advanced technologies including full-body motion control, reinforcement learning, and a multilingual AI model called MoNet, enabling it to perceive, plan, and interact naturally with humans using vision-language understanding and semantic reasoning. Powered by AiMOGA’s L3 Assistance Level framework, Mornine features high-torque joints and dexterous hands with 17 degrees of freedom, allowing smooth and precise movements. The robot’s AI adapts its gestures and tone based on visitor reactions,

    robothumanoid-robotAIautonomous-systemsautomotive-technologyreinforcement-learninghuman-robot-interaction
  • Coco Robotics taps UCLA professor to lead new physical AI research lab

    Coco Robotics, a startup specializing in last-mile delivery robots, has established a new physical AI research lab led by UCLA professor Zhou, who has also joined the company as chief AI scientist. The move aims to leverage the extensive data—spanning millions of miles collected over five years in complex urban environments—to advance autonomous operation of their delivery bots and reduce delivery costs. Coco Robotics co-founder and CEO Zach Rash emphasized that the company now has sufficient data scale to accelerate research in physical AI, particularly in robot navigation and reinforcement learning, areas where Zhou is a leading expert. The new research lab operates independently from Coco Robotics’ partnership with OpenAI, which provides access to language models, while the lab focuses on utilizing the company’s proprietary robot-collected data. Coco Robotics plans to use the insights gained exclusively to enhance its own automation capabilities and improve the efficiency of its local robot models, rather than selling the data. Additionally, the company intends to share relevant research findings with the cities where it operates to help address

    roboticsartificial-intelligenceautonomous-deliveryphysical-AIrobot-navigationreinforcement-learninglast-mile-delivery
  • China’s wearable suit trains humanoid robots with high accuracy

    Researchers at China’s National University of Defense Technology, in collaboration with Midea Group, have developed HumanoidExo, a wearable suit system designed to train humanoid robots with high accuracy by capturing real-time human motion. Unlike traditional training methods that rely on videos and simulations—often causing robots to lose balance—HumanoidExo uses motion sensors and a LiDAR scanner to track seven arm joints and body movements, providing robots with precise, real-world data. The system’s AI component, HumanoidExo-VLA, combines a Vision-Language-Action model to interpret human tasks and a reinforcement learning controller to maintain robot balance during learning. Testing on the Unitree G1 humanoid robot demonstrated significant improvements: after training with data from five teleoperated and 195 exoskeleton-recorded sessions, the robot’s success rate on a pick-and-place task rose from 5% to nearly 80%, approaching the performance level of 200 human demonstrations. The robot also learned to walk effectively

    robothumanoid-robotswearable-suitmotion-captureAI-trainingreinforcement-learningexoskeleton
  • Humanoid robot walks naturally down passageway into a 'friends' den

    PND Robotics recently showcased its humanoid robot, Adam, which demonstrates a natural, human-like gait and directional sense as it walks down a hallway into a room filled with other robots performing various tasks. Using a proprietary reinforcement learning algorithm, Adam was trained through simulation-to-real-world methods to mimic human walking and movement with impressive fluidity and accuracy. In the video, Adam is greeted by another robot, Adam-U, highlighting PND Robotics’ focus on creating affordable, modular humanoid robots that combine biomimetic design with real-world adaptability and continuous self-learning. Adam and Adam-U made notable public debuts at events in Shanghai and Zhejiang Province, where they impressed audiences by performing human-like actions such as passing objects and navigating environments naturally. PND Robotics aims to democratize personal robotics by developing machines that integrate physical interaction, perception, and learning in a lifelike manner. Compared to other advanced humanoid robots like Cassie, Digit, and HRP-5P, Adam stands out for its human

    robothumanoid-robotreinforcement-learningbiomimetic-designmodular-robotspersonal-roboticsrobot-locomotion
  • China’s humanoid robot Bumblebee now walks with human-like gait

    Shanghai Kepler Robotics has unveiled a significant advancement in its humanoid robot K2 “Bumblebee,” showcasing China’s first “hybrid-architecture disturbance-resistant” gait that enables the robot to walk with a natural, human-like straight-knee motion. This breakthrough is achieved through a novel hybrid actuation system combining planetary roller screw linear actuators and rotary actuators in a series-parallel configuration. The linear actuators act as the robot’s primary “leg muscles,” providing walking force, while rotary actuators manage fine adjustments and terrain adaptation. This design offers high energy efficiency (81.3%), precise positioning, and strong load-bearing capacity, allowing Bumblebee to maintain balance over uneven surfaces such as bricks and grass, and to carry payloads up to 30 kilograms (66 pounds). To bridge the gap between simulation and real-world performance, Kepler addressed mechanical and control challenges by integrating reinforcement learning, imitation learning, and torque control, enabling dynamic gait switching and robust stability despite sensor noise, actuator

    robothumanoid-robothybrid-actuatorlocomotion-technologyreinforcement-learningindustrial-roboticsrobotic-gait
  • Humanoid robot HITTER plays table tennis with human-like speed

    UC Berkeley has developed a humanoid robot named HITTER that can play table tennis with human-like speed and agility. Demonstrated in a video, HITTER successfully engaged in rallies exceeding 100 shots against human opponents, using its left hand for balance and executing precise, fluid movements. The robot’s performance relies on a dual-system design: a high-level planner that tracks and predicts the ball’s trajectory using external cameras, and a low-level controller that converts these calculations into coordinated arm and leg motions. Trained on human motion data, HITTER can move naturally, reacting to balls traveling up to 5 m/s in under a second. The development team combined model-based planning with reinforcement learning to overcome the challenges of split-second decision-making and unpredictable shots inherent in table tennis. This hybrid approach enabled HITTER to fine-tune its movements through trial and error, resulting in lifelike swings and footwork. Tested on a general-purpose humanoid platform (likely the Unitree G1), HITTER demonstrated its

    roboticshumanoid-robotreinforcement-learningAI-planninghuman-robot-interactiontable-tennis-robotrobot-motion-control
  • Inside Singapore's physical AI revolution

    The article summarizes Episode 210 of The Robot Report Podcast, which centers on Singapore’s emerging leadership in physical AI and robotics. Key guests from the Singapore Economic Development Board (EDB), Certis Group, and the Home Team Science & Technology Agency discuss Singapore’s strategic initiatives to grow its robotics sector. The country leverages its strong manufacturing base, government incentives, and a collaborative ecosystem involving industry and academia to foster innovation and talent development. Emphasis is placed on the importance of integration, reliability, and scalability for successful deployment of robotics and AI technologies. The episode also covers notable robotics news, including Boston Dynamics’ Spot robot performing a public triple backflip, showcasing advancements in reinforcement learning for robot agility and recovery. Despite the impressive feat, Spot’s performance in America’s Got Talent did not advance to the quarterfinals. Additionally, Intuitive Surgical announced a permanent layoff of 331 employees (about 2% of its workforce) at its Sunnyvale headquarters. Lastly, John Deere expanded its agricultural

    roboticsartificial-intelligencephysical-AISingaporeBoston-Dynamicsreinforcement-learningautomation
  • RoboBallet makes robotic arms dance in sync on factory floors

    RoboBallet is a new AI system developed by a team from UCL, Google DeepMind, and Intrinsic that choreographs the movements of multiple robotic arms on factory floors, significantly improving efficiency and scalability in manufacturing. Traditional robotic coordination requires extensive manual programming to avoid collisions and complete tasks, a process that is time-consuming and prone to errors. RoboBallet overcomes these challenges by using reinforcement learning combined with graph neural networks, enabling it to plan coordinated movements for up to eight robotic arms performing 40 tasks in seconds, even in previously unseen layouts. This approach treats obstacles and tasks as points in a network, allowing rapid and adaptable planning that outperforms existing methods by generating plans hundreds of times faster than real-time. The system’s scalability is a major breakthrough, as it learns general coordination rules rather than memorizing specific scenarios, making it capable of handling complex, dynamic environments where factory layouts or robot configurations change frequently. RoboBallet’s ability to instantly generate high-quality plans could prevent costly

    roboticsindustrial-automationAIrobotic-armsmanufacturing-technologyreinforcement-learningfactory-efficiency
  • #IJCAI2025 distinguished paper: Combining MORL with restraining bolts to learn normative behaviour - Robohub

    The article discusses advancements presented at IJCAI 2025 concerning the integration of Multi-Objective Reinforcement Learning (MORL) with restraining bolts to enable AI agents to learn normative behavior. Autonomous agents, powered by reinforcement learning (RL), are increasingly deployed in real-world applications such as self-driving cars and smart urban planning. While RL agents excel at optimizing behavior to maximize rewards, unconstrained optimization can lead to actions that, although efficient, may be unsafe or socially inappropriate. To address safety, formal methods like linear temporal logic (LTL) have been used to impose constraints ensuring agents act within defined safety parameters. However, safety constraints alone are insufficient when AI systems interact closely with humans, as normative behavior involves compliance with social, legal, and ethical norms that go beyond mere safety. Norms are expressed through deontic concepts—obligations, permissions, and prohibitions—that describe ideal or acceptable behavior rather than factual truths. This introduces complexity in reasoning, especially with contrary-to-duty

    robotartificial-intelligencereinforcement-learningautonomous-agentssafe-AImachine-learningnormative-behavior
  • Google DeepMind, Intrinsic build AI for multi-robot planning

    The article discusses a new AI-driven approach to programming and coordinating multiple industrial robots in shared workspaces, developed through a collaboration between Google DeepMind Robotics, Intrinsic, and University College London. Traditional methods for robot motion planning rely heavily on manual programming, teach pendants, and trial-and-error, which are time-consuming and become increasingly complex when managing multiple robots to avoid collisions. The researchers introduced "RoboBallet," an AI model that leverages reinforcement learning and graph neural networks (GNNs) to generate collision-free motion plans efficiently. This model represents robots, tasks, and obstacles as nodes in a graph and learns generalized planning strategies by training on millions of synthetic scenarios, enabling it to produce near-optimal trajectories rapidly without manual intervention. Intrinsic, a company spun out of Alphabet’s X in 2021, aims to simplify industrial robot programming and scaling. Their RoboBallet system requires only CAD files and high-level task descriptions to generate motion plans, eliminating the need for detailed coding or fine

    roboticsartificial-intelligencemulti-robot-planningreinforcement-learninggraph-neural-networksindustrial-robotsautomation
  • Humanoid robots lack data to keep pace with explosive rise of AI

    The recent International Humanoid Olympiad held in Olympia, Greece, showcased humanoid robots competing in sports like boxing and soccer, highlighting their growing capabilities. Despite these advances, humanoid robots remain significantly behind AI software in learning from data, with experts estimating they are roughly "100,000 years" behind due to limited data availability. Organizers and researchers emphasize that while AI tools benefit from vast datasets enabling rapid advancement, humanoid robots struggle to acquire and process comparable real-world data, which hinders their ability to perform complex, dexterous household tasks. Experts predict that humanoid robots may first find practical use in space exploration before becoming common in homes, a transition expected to take over a decade. To address this gap, researchers are exploring reinforcement learning techniques that allow robots to learn from real-time experiences rather than relying solely on pre-programmed actions. Additionally, innovative approaches such as developing biological computer brains using real brain cells on chips aim to enable robots to learn and adapt more like humans. The Olymp

    robothumanoid-robotsartificial-intelligencerobotic-learningreinforcement-learningrobotic-brainrobotics-competition
  • Unique robot welded from online parts walks on two legs with ease

    MEVITA is a newly developed open-source bipedal robot created by engineers at the University of Tokyo's JSK Robotics Laboratory. It addresses common challenges in DIY robotics platforms by combining durability, simplicity, and accessibility. Unlike many existing designs that rely on fragile 3D-printed parts or complex metal assemblies with hard-to-source components, MEVITA uses sheet metal welding to integrate complex shapes into just 18 unique metal parts, four of which are welded. This approach significantly reduces the number of components, making the robot easier to build using parts readily available through online e-commerce. The robot’s control system leverages advanced AI techniques, specifically reinforcement learning trained in simulation environments (IsaacGym and MuJoCo), before transferring the learned behaviors to the physical robot via Python scripts. This Sim-to-Real transfer enables MEVITA to walk effectively across diverse terrains such as uneven indoor floors, grassy fields, dirt, concrete tiles, and gentle slopes. Safety and control are enhanced by features including wireless

    roboticsbipedal-robotopen-source-robotsheet-metal-weldingAI-control-systemreinforcement-learningrobot-assembly
  • Boston Dynamics’ robot dog nails daring backflips in new video

    Boston Dynamics has showcased its robot dog, Spot, performing consistent backflips in a new video, highlighting the robot’s advanced agility and refined design. While these gymnastic feats are unlikely to be part of Spot’s routine tasks, they serve a critical engineering purpose: pushing the robot to its physical limits to identify and address potential balance failures. This helps improve Spot’s ability to recover quickly from slips or trips, especially when carrying heavy payloads in industrial settings, thereby enhancing its reliability and durability. The development of Spot’s backflip capability involved reinforcement learning techniques, where the robot was trained in simulations to optimize its movements by receiving rewards for successful actions, akin to training a dog with treats. This iterative process of simulation and real-world testing allows engineers to fine-tune Spot’s behavior and ensure robust performance. Beyond technological advancements, Spot’s agility has also been demonstrated in entertainment contexts, such as performing dance routines on America’s Got Talent, showcasing its versatility. Looking forward, Spot’s ongoing evolution through

    robotroboticsBoston-Dynamicsrobot-dogreinforcement-learningmachine-learningquadruped-robot
  • Humanoids, robot dogs master unseen terrains with attention mapping

    Researchers at ETH Zurich have developed an advanced control system for legged robots, including the quadrupedal ANYmal-D and humanoid Fourier GR-1, enabling them to navigate complex and previously unseen terrains. This system employs a machine learning technique called attention-based map encoding, trained via reinforcement learning, which allows the robot to focus selectively on the most critical areas of a terrain map rather than processing the entire map uniformly. This focused attention helps the robots identify safe footholds even in challenging environments, improving robustness and generalization across varied terrains. The system demonstrated successful real-time locomotion at speeds up to 2 meters per second, with notably low power consumption relative to the robot’s motors. While the current approach is limited to 2.5D height-map locomotion and cannot yet handle overhanging 3D obstacles such as tree branches, the researchers anticipate extending the method to full 3D environments and more complex loco-manipulation tasks like opening doors or climbing. The attention mechanism also provides

    robothumanoid-robotsquadrupedal-robotsmachine-learningreinforcement-learningattention-mappinglocomotion-control
  • Video: Swiss robot dog plays perfect badminton match with a human

    Researchers at Switzerland’s ETH Zurich have developed a quadruped robot dog named ANYmal, capable of playing badminton with a human at the skill level of a seven-year-old child. ANYmal, created by ANYbotics, uses a sophisticated control system equipped with two cameras to track and predict the shuttlecock’s trajectory. It swings a racket attached to a multi-axis arm to hit the shuttlecock precisely. The robot was trained using reinforcement learning in a virtual environment, where it practiced thousands of rallies to learn positioning, shot accuracy, and anticipatory movement, enabling it to perform with remarkable precision in real-world play. A key challenge addressed in the development was maintaining balance while lunging and moving quickly to return shots. ANYmal’s reinforcement learning algorithm enhances its coordination and stability, allowing it to move with agility and balance comparable to a human player. Originally designed for industrial inspection and navigating rough terrains, including disaster zones, ANYmal’s capabilities have now been extended to dynamic sports environments. Priced at around

    robotroboticsreinforcement-learningquadruped-robotrobot-dogautonomous-robotsrobot-control-systems
  • China’s robot dog sprints 328 feet in 16.33 seconds, breaks record

    China’s Zhejiang University announced that its quadruped robot, White Rhino, set a new Guinness World Record by sprinting 100 meters (328 feet) in 16.33 seconds, surpassing the previous record of 19.87 seconds held by South Korea’s Hound robot. The run took place in Hangzhou and marks a significant advancement in robotic speed, narrowing the gap between machine and human sprint performance (Usain Bolt’s human record is 9.58 seconds). This achievement demonstrates the robot’s explosive power, speed, stability, and precise control during rapid movement. White Rhino was developed through a collaborative effort involving Zhejiang University’s Center for X-Mechanics, School of Aeronautics and Astronautics, and the Hangzhou Global Scientific and Technological Innovation Center. The design process employed a “robot forward design” approach, using comprehensive dynamics simulations and multi-objective optimization to simultaneously refine geometry, motor specifications, and reduction systems. The robot features high-power-density joint actuators

    robotquadruped-robotroboticsactuatorscontrol-algorithmsreinforcement-learningmechanical-design
  • Humanoid robots Adam and Adam-U display lifelike AI movement

    At the World Artificial Intelligence Conference 2025 in Shanghai, Chinese robotics company PNDbotics unveiled two advanced humanoid robots, Adam and Adam-U, showcasing significant strides in AI-driven robotics. Adam is a full-sized, 1.6-meter-tall, 132-pound humanoid robot designed for high agility and precision, featuring 44 degrees of freedom and powered by deep reinforcement learning (DRL) and imitation-learning algorithms. It boasts patented quasi-direct drive actuators that enable smooth, human-like movements, including balanced posture and deft manipulation, even without visual input. Adam’s modular, biomimetic design and real-time control system allow it to perform complex tasks dynamically, such as playing musical instruments and dancing. Adam-U, developed in partnership with Noitom Robotics and Inspire Robots, serves as a high-precision, stationary data acquisition platform with 31 degrees of freedom. It integrates advanced motion capture technology, including Noitom’s PNLink suit and Inspire’s dexterous robotic hand,

    robothumanoid-robotAImotion-capturerobotics-innovationreinforcement-learningimitation-learning
  • Oli: LimX’s new humanoid robot masters gym, warehouse, dance floor

    LimX Dynamics, a Chinese robotics company, has unveiled its full-sized humanoid robot named LimX Oli, designed to advance embodied AI and automation in manufacturing, warehousing, and research. Available in three variants—Lite, EDU, and Super—starting at about $21,800, Oli features a modular arm system with interchangeable attachments such as standard hands, precision grippers, and dexterous robotic hands. This modularity allows the robot to perform a wide range of tasks across different environments, from lifting dumbbells in a gym to sorting items in a warehouse and even performing Chinese kung fu and dancing, showcasing its strength, agility, balance, and full-body disturbance recovery capabilities. Standing 1.65 meters tall with 31 degrees of freedom, Oli is tailored for AI researchers, robotics engineers, and system integrators, offering an open SDK that provides full access to sensor data, joint control, and task scheduling. This flexible hardware-software design and scalable development toolchain make it a powerful

    robothumanoid-robotAI-roboticsmodular-roboticswarehouse-automationreinforcement-learningembodied-intelligence
  • China’s humanoid robot stuns by opening car door in a 'world-first'

    AiMOGA Robotics has achieved a significant breakthrough with its humanoid robot, Mornine, which autonomously opened a car door inside a functioning Chery dealership in China—marking a world-first in embodied AI. Unlike scripted or teleoperated robots, Mornine used only onboard sensors, full-body motion control, and reinforcement learning to identify the door handle, adjust its posture, and apply coordinated force to open the door without any human input. This task, performed in a live commercial setting, demonstrates advanced autonomy and a shift from simulation-based robotics to real-world service applications. Mornine’s sophisticated sensor suite includes 3D LiDAR, depth and wide-angle cameras, and a visual-language model, enabling real-time perception and continuous learning through a cloud-based training loop. The robot was not explicitly programmed to recognize door handles but learned through millions of simulated cycles, with the learned model transferred to real-world operation via Sim2Real methods. Currently deployed in multiple Chery 4S dealerships

    roboticshumanoid-robotautonomous-robotsAI-roboticsservice-robotsreinforcement-learningsensor-technology
  • China’s humanoid robot achieves human-like motion with 31 joints

    Chinese robotics company PND Robotics, in collaboration with Noitom Robotics and Inspire Robots, has introduced the Adam-U humanoid robot platform, which features 31 degrees of freedom (DOF) enabling human-like motion. The robot includes a 2-DOF head, 6-DOF dexterous hands, a 3-DOF waist with a braking system for safety, and a binocular vision system that mimics human sight. Standing adjustable between 1.35 to 1.77 meters and weighing 61 kilograms, Adam-U cannot walk as it uses a stationary platform instead of legs. It is designed for precise, flexible operation in dynamic environments and is particularly suited for reinforcement and imitation learning, making it a valuable tool for AI researchers, robotics engineers, and academic institutions. The Adam-U platform integrates hardware and software into a comprehensive ecosystem, including Noitom’s PNLink full-body wired inertial motion capture suit and Inspire Robots’ RH56E2 tactile dexterous

    roboticshumanoid-robotmotion-captureartificial-intelligencemachine-learningreinforcement-learningdata-acquisition
  • EngineAI raises nearly $140M to develop legged, humanoid robots - The Robot Report

    EngineAI, a Shenzhen-based robotics company, has raised nearly $140 million (RMB 1 billion) through its pre-A++ and A1 funding rounds to advance the development and commercialization of legged humanoid robots. The company plans to use the capital to scale trial production, expand its workforce fivefold, and diversify its product lines, focusing on bipedal and full humanoid robots. EngineAI’s technology combines proprietary joint modules that deliver high power, torque, and precision with a hybrid control system integrating traditional controls and reinforcement learning (RL), enabling lifelike, dynamic movements such as complex dances and sprinting with millimeter-level accuracy. EngineAI aims to penetrate the growing global humanoid robotics market, projected by various analysts to reach anywhere from $15 billion by 2030 to $5 trillion by 2050, driven by demand in manufacturing, logistics, and services. The company employs an “open-source hardware + ecosystem profit-sharing” model to accelerate market adoption through strategic partnerships with

    roboticshumanoid-robotsreinforcement-learningAI-roboticsrobot-hardwarerobot-softwarerobotics-market
  • Robot Adam grooves on keytar at China’s futuristic music festival

    The article highlights the debut of Adam, a full-sized humanoid robot developed by PNDbotics, performing as a keytar player alongside Chinese musician Hu Yutong’s band at the VOYAGEX Music Festival in Changchun, China, on July 12, 2025. Adam impressed the audience with fluid, human-like movements and precise musical timing, showcasing a seamless integration of robotics and live performance art. Standing 1.6 meters tall and weighing 60 kilograms, Adam’s agility and control stem from 25 patented quasi-direct drive (QDD) PND actuators with advanced force control, enabling smooth, coordinated motions that closely mimic human dexterity. Powered by a proprietary reinforcement learning algorithm and supported by a robust control system featuring an Intel i7-based unit, Adam demonstrates sophisticated real-time coordination across its limbs and joints. The robot’s modular design enhances its versatility, maintainability, and adaptability to dynamic environments, including congested or uneven terrain. PNDbotics has continuously

    robothumanoid-robotroboticsartificial-intelligencereinforcement-learningactuatorsrobot-control-systems
  • New quadruped robot climbs vertically 50 times faster than rivals

    Researchers at the University of Tokyo’s Jouhou System Kougaka Laboratory (JSK) have developed KLEIYN, a quadruped robot capable of climbing vertical walls up to 50 times faster than previous robots. Unlike other climbing robots that rely on grippers or claws, KLEIYN uses a chimney climbing technique, pressing its feet against two opposing walls for support. Its flexible waist joint allows adaptation to varying wall widths, particularly narrow gaps. The robot weighs about 40 pounds (18 kg), measures 2.5 feet (76 cm) in length, and features 13 joints powered by quasi-direct-drive motors for precise movement. KLEIYN’s climbing ability is enhanced through machine learning, specifically Reinforcement Learning combined with a novel Contact-Guided Curriculum Learning method, enabling it to transition smoothly from flat terrain to vertical surfaces. In tests, KLEIYN successfully climbed walls spaced between 31.5 inches (80 cm) and 39.4 inches (

    robotquadruped-robotmachine-learningreinforcement-learningclimbing-robotrobotics-innovationautonomous-robots
  • Swiss robot dog can now pick up and throw a ball accurately like humans

    ETH Zurich’s robotic dog ANYmal, originally designed for autonomous operation in challenging environments, has been enhanced with a custom arm and gripper, enabling it to pick up and throw objects with human-like accuracy. The robot’s advanced actuators and integrated sensors allow it to navigate complex terrain while maintaining stability and situational awareness. Unlike traditional factory robots, ANYmal is built to handle unpredictable outdoor conditions, making it suitable for tasks such as industrial inspection, disaster response, and exploration. The research team, led by Fabian Jenelten, trained ANYmal using reinforcement learning within a highly realistic virtual environment that simulated real-world physics. This approach, known as sim-to-real transfer, allowed the robot to practice millions of throws safely and ensured its skills transferred effectively to real-world scenarios. In testing, ANYmal successfully picked up and threw various objects—including balls, bottles, and fruit—across different surfaces and environmental challenges, such as wind and uneven ground, demonstrating adaptability and precise control without pre-programmed steps. This

    roboticsautonomous-robotsreinforcement-learninglegged-robotsrobot-manipulationsim-to-real-transferrobot-perception
  • NBC’s AGT pushes Spot to perform under pressure

    Boston Dynamics showcased its Spot quadruped robots on NBC’s America’s Got Talent (AGT), performing a live, choreographed dance routine to Queen’s “Don’t Stop Me Now.” Five Spots danced synchronously, using their robot arms to “lip-sync” Freddie Mercury’s vocals, impressing all four AGT judges who voted to advance the act. This high-profile appearance was both an entertainment milestone and a rigorous technical stress test for the robots and engineering team. The performance combined autonomous dancing via proprietary choreography software with teleoperated interactions, pushing Spot’s capabilities with aggressive moves like high-speed spins and one-legged balancing. These advanced maneuvers, enabled by recent improvements in reinforcement learning and dynamic behavior modeling, also enhance Spot’s real-world applications, such as maintaining balance on slippery factory floors. The decision to bring Spot to AGT followed successful live performances at the 2024 Calgary Stampede, which built confidence in managing the technical and logistical challenges of a live broadcast. Despite over 100

    roboticsBoston-DynamicsSpot-robothumanoid-robotsrobot-performanceautonomous-robotsreinforcement-learning
  • Robot Talk Episode 126 – Why are we building humanoid robots? - Robohub

    The article summarizes a special live episode of the Robot Talk podcast recorded at Imperial College London during the Great Exhibition Road Festival. The discussion centers on the motivations and implications behind building humanoid robots—machines designed to look and act like humans. The episode explores why humanoid robots captivate and sometimes unsettle us, questioning whether this fascination stems from vanity or if these robots could serve meaningful roles in future society. The conversation features three experts: Ben Russell, Curator of Mechanical Engineering at the Science Museum, Maryam Banitalebi Dehkordi, Senior Lecturer in Robotics and AI at the University of Hertfordshire, and Petar Kormushev, Director of the Robot Intelligence Lab at Imperial College London. Each brings a unique perspective, from historical and cultural insights to technical expertise in robotics, AI, and machine learning. Their dialogue highlights the rapid advancements in humanoid robotics and the ongoing research aimed at creating adaptable, autonomous robots capable of learning and functioning in dynamic environments. The episode underscores the multidisciplinary nature

    roboticshumanoid-robotsartificial-intelligenceautonomous-robotsmachine-learningreinforcement-learningrobot-intelligence
  • Sweater-wearing humanoid robot gets brain upgrade to clean, cook solo

    1X Technologies has introduced Redwood, an advanced AI model powering its humanoid robot NEO, designed to autonomously perform complex household tasks such as laundry, door answering, and home navigation. Redwood is a 160 million-parameter vision-language model that integrates perception, locomotion, and control into a unified system running onboard NEO Gamma’s embedded GPU. This integration enables full-body coordination, allowing NEO to simultaneously control arms, legs, pelvis, and walking commands, which enhances its ability to brace against surfaces, handle higher payloads, and manipulate objects bi-manually. Redwood’s training on diverse real-world data, including both successful and failed task demonstrations, equips NEO with strong generalization capabilities to adapt to unfamiliar objects and task variations, improving robustness and autonomy even in offline or low-connectivity environments. Complementing Redwood, 1X Technologies has developed a comprehensive Reinforcement Learning (RL) controller that expands NEO’s mobility and dexterity for navigating real home environments. This controller supports fluid

    robothumanoid-robotAI-modelrobotics-autonomymotion-controlmobile-manipulationreinforcement-learning
  • Chinese firm eases humanoid, legged robot development with new suite

    EngineAI Robotics, a Shenzhen-based Chinese firm, has launched EngineAI RL Workspace, an open-source, modular reinforcement learning platform tailored specifically for legged robotics development. This comprehensive suite includes dual frameworks—a training code repository and a deployment code repository—that together provide an end-to-end solution from algorithm training to real-world application. The platform is designed to enhance development efficiency through reusable logic structures, a unified single-algorithm executor for both training and inference, and decoupled algorithms and environments that enable seamless iteration without interface changes. The EngineAI RL Workspace integrates the entire development pipeline with four core components: environment modules, algorithm engines, shared toolkits, and integration layers, each independently encapsulated to facilitate multi-person collaboration and reduce communication overhead. Additional features include dynamic recording systems for capturing training and inference videos, intelligent version management to maintain experiment consistency, and detailed user guides to support rapid onboarding. At CES 2025, EngineAI showcased humanoid robots like the SE01, a versatile 5.

    roboticshumanoid-robotsreinforcement-learninglegged-robotsrobot-developmentAI-in-roboticsmodular-robotics-platform
  • Chinese firm achieves agile, human-like walking with AI control

    Chinese robotics startup EngineAI has developed an advanced AI-driven control system that enables humanoid robots to walk with straight legs, closely mimicking natural human gait. This innovative approach integrates human gait data, adversarial learning, and real-world feedback to refine robot movement across diverse environments, aiming to achieve more energy-efficient, stable, and agile locomotion. EngineAI’s lightweight humanoid platform, the PM01, has demonstrated impressive agility, including successfully performing a frontflip and executing complex dance moves from the film Kung Fu Hustle, showcasing the system’s potential for fluid, human-like motion. The PM01 robot features a compact, lightweight aluminum alloy exoskeleton with 24 degrees of freedom and a bionic structure that supports dynamic movement at speeds up to 2 meters per second. It incorporates advanced hardware such as an Intel RealSense depth camera for visual perception and an Intel N97 processor paired with an NVIDIA Jetson Orin CPU for high-performance processing and neural network training. This combination allows the PM01 to interact effectively with its environment and perform intricate tasks, making it a promising platform for research into human-robot interaction and agile robotic assistants. EngineAI’s work parallels other Chinese developments like the humanoid robot Adam, which uses reinforcement learning and imitation of human gait to achieve lifelike locomotion. Unlike traditional control methods such as Model Predictive Control used by robots like Boston Dynamics’ Atlas, EngineAI’s AI-based framework emphasizes adaptability through real-world learning, addressing challenges in unpredictable environments. While still in the research phase, these advancements mark significant progress toward next-generation humanoid robots capable of natural, efficient, and versatile movement.

    robothumanoid-robotAI-controlgait-controlreinforcement-learningrobotics-platformenergy-efficient-robotics
  • Congratulations to the #AAMAS2025 best paper, best demo, and distinguished dissertation award winners - Robohub

    The 24th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2025), held from May 19-23 in Detroit, recognized outstanding contributions in the field with awards for best paper, best demo, and distinguished dissertation. The Best Paper Award went to the team behind "Soft Condorcet Optimization for Ranking of General Agents," led by Marc Lanctot and colleagues. Several other papers were finalists, covering topics such as commitments in BDI agents, curiosity-driven partner selection, reinforcement learning for vehicle-to-building charging, and drone delivery systems. The Best Student Paper Award was given to works on decentralized planning using probabilistic hyperproperties and large language models for virtual human gesture selection. In addition, the Blue Sky Ideas Track honored François Olivier and Zied Bouraoui for their neurosymbolic approach to embodied cognition, while the Best Demo Award recognized a project on serious games for ethical preference elicitation by Jayati Deshmukh and team. The Victor Lesser Distinguished Dissertation Award, which highlights originality, impact, and quality in autonomous agents research, was awarded to Jannik Peters for his thesis on proportionality in selecting committees, budgets, and clusters. Lily Xu was the runner-up for her dissertation on AI decision-making for planetary health under conditions of low-quality data. These awards underscore the innovative research advancing autonomous agents and multiagent systems.

    robotautonomous-agentsmultiagent-systemsdronesreinforcement-learningenergy-storageAI
  • Tesla’s Optimus robot takes out trash, vacuums, cleans like a pro

    robotTeslaOptimusAIautomationhumanoid-robotreinforcement-learning
  • Watch humanoid robots clash in a tug of war, pull cart, open doors

    robothumanoidreinforcement-learningcontrol-systemforce-awareloco-manipulationCMU
  • Robot Talk Episode 121 – Adaptable robots for the home, with Lerrel Pinto

    robotmachine-learningadaptable-robotsroboticsartificial-intelligenceautonomous-machinesreinforcement-learning
  • Shlomo Zilberstein wins the 2025 ACM/SIGAI Autonomous Agents Research Award

    robotautonomous-agentsmulti-agent-systemsdecision-makingreinforcement-learningresearch-awardAI