RIEM News LogoRIEM News

Articles tagged with "AI-models"

  • NVIDIA debuts first open reasoning AI for self-driving vehicles

    NVIDIA has introduced a suite of open-source AI models and tools aimed at advancing autonomous vehicles, robotics, and speech processing. Central to this launch is Alpamayo-R1 (AR1), the world’s first open reasoning vision-language-action (VLA) model designed for self-driving cars. AR1 integrates chain-of-thought reasoning with path planning to navigate complex driving scenarios by evaluating possible trajectories and contextual data, enabling human-like decision-making in challenging environments such as crowded intersections or lane closures. Built on NVIDIA’s Cosmos Reason platform, AR1 is available for customization by researchers for non-commercial use and has demonstrated improved reasoning capabilities through reinforcement learning post-training. Beyond AR1, NVIDIA’s Cosmos platform offers additional tools for physical AI development, including LidarGen for generating lidar data, Omniverse NuRec Fixer for neural reconstruction cleanup, Cosmos Policy for robot behavior creation, and ProtoMotions3 for training humanoid robots in simulated settings. These resources are already being utilized by ecosystem partners and academic

    robotautonomous-vehiclesAI-modelsNVIDIA-Cosmosreinforcement-learninglidar-simulationhumanoid-robots
  • Nvidia announces new open AI models and tools for autonomous driving research

    Nvidia has unveiled new AI infrastructure and models aimed at advancing physical AI applications, particularly in robotics and autonomous vehicles. At the NeurIPS AI conference, the company introduced Alpamayo-R1, described as the first vision-language-action model specifically designed for autonomous driving research. This model integrates visual and textual data to enable vehicles to perceive their environment and make informed decisions, leveraging Nvidia’s existing Cosmos reasoning model family, which was initially launched in January 2025. Alpamayo-R1 is intended to help autonomous vehicles achieve level 4 autonomy—full self-driving capability within defined areas and conditions—by providing them with “common sense” reasoning to handle complex driving scenarios more like humans. In addition to the new model, Nvidia released the Cosmos Cookbook on GitHub, a comprehensive resource including step-by-step guides, inference tools, and post-training workflows to assist developers in customizing and training Cosmos models for various applications. This toolkit covers essential processes such as data curation, synthetic data generation, and model

    robotautonomous-vehiclesAI-modelsNvidiaphysical-AIautonomous-drivingvision-language-models
  • Google's Gemini model lets humanoid robot carry out multimodal tasks

    Google DeepMind has unveiled advancements in its humanoid robots powered by the Gemini Robotics 1.5 AI models, enabling them to perform complex, multi-step tasks through multimodal reasoning. Demonstrated in a recent video, the bi-arm Franka robot successfully completed the "banana test," sorting different fruits by color into separate plates, showcasing improved capabilities over previous models that could only follow single-step instructions. Another test featured Apptronik’s Apollo humanoid sorting laundry by color, even adapting to changes in basket positions mid-task, highlighting the robots' enhanced perception and adaptability. The Gemini Robotics 1.5 family includes two complementary models: one that converts visual inputs and instructions into actions, and another that reasons about the environment to create step-by-step plans. This agentic framework allows robots to autonomously study their surroundings, make decisions, and execute tasks such as sorting waste according to local recycling rules by researching guidelines online and applying them in real time. Google emphasizes safety in these models, incorporating risk assessment

    roboticshumanoid-robotsAI-modelsmultimodal-tasksautonomous-robotsrobot-perceptionrobot-reasoning
  • FieldAI raises $405M to build universal robot brains

    FieldAI, a robotics AI company, announced a $405 million funding raise to develop universal "robot brains" capable of controlling diverse physical robots across varied real-world environments. The latest funding round, including a $314 million tranche co-led by Bezos Expedition, Prysm, and Temasek, adds to backing from investors such as Khosla Ventures and Intel Capital. FieldAI’s core innovation lies in its "Field Foundation Models," which integrate physics-based understanding into embodied AI—AI that governs robots physically navigating environments—enabling robots to quickly learn, adapt, and manage risk and safety in new settings. This physics-informed approach contrasts with traditional AI models that often lack risk awareness, making FieldAI’s robots better suited for complex and potentially hazardous environments. Founder and CEO Ali Agha emphasized that their goal is to create a single, general-purpose robot brain that can operate across different robot types and tasks, with a built-in confidence measure to assess decision reliability and manage safety thresholds. Agha’s decades

    robotartificial-intelligenceembodied-AIrobotics-safetyrobot-learningAI-modelsrobotics-technology
  • Buzzy AI startup Multiverse creates two of the smallest high-performing models ever

    Multiverse Computing, a leading European AI startup based in Spain, has developed two of the smallest yet high-performing AI models, humorously named after animal brain sizes: SuperFly and ChickBrain. These models are designed to be embedded in Internet of Things (IoT) devices and run locally on smartphones, tablets, and PCs without requiring an internet connection. SuperFly, inspired by a fly’s brain, is a compressed version of Hugging Face’s SmolLM2 135 model with 94 million parameters, optimized for limited data and voice-command applications in home appliances. ChickBrain, with 3.2 billion parameters, is a compressed version of Meta’s Llama 3.1 8B model and offers advanced reasoning capabilities, outperforming the original in several benchmarks such as MMLU-Pro, Math 500, GSM8K, and GPQA Diamond. The key technology behind these models is Multiverse’s proprietary quantum-inspired compression algorithm called CompactifAI, which significantly reduces model

    IoTAI-modelsmodel-compressionedge-computingembedded-AIquantum-inspired-algorithmssmart-devices
  • Nvidia unveils new Cosmos world models, infra for robotics and physical uses

    Nvidia has introduced a suite of new AI models and infrastructure aimed at advancing robotics and physical AI applications. The centerpiece is Cosmos Reason, a 7-billion-parameter vision-language model designed to enable robots and AI agents to "reason" by leveraging memory and physics understanding. This capability allows Cosmos Reason to function as a planning model, helping embodied agents determine their next steps, and it can be applied in data curation, robot planning, and video analytics. Alongside Cosmos Reason, Nvidia unveiled Cosmos Transfer-2, which accelerates synthetic data generation from 3D simulations or spatial controls, and a faster, distilled version of Cosmos Transfers optimized for speed. In addition to the AI models, Nvidia announced new neural reconstruction libraries that facilitate 3D simulation of real-world environments using sensor data, with integration into the open-source CARLA simulator. The company also updated its Omniverse software development kit and introduced new hardware solutions, including the Nvidia RTX Pro Blackwell Servers tailored for robotics workflows and the DG

    roboticsAI-modelsNvidia-Cosmossynthetic-data-generation3D-simulationrobot-planningneural-reconstruction
  • Meta’s new AI helps robots learn real-world logic from raw video

    Meta has introduced V-JEPA 2, an advanced AI model trained solely on raw video data to help robots and AI agents better understand and predict physical interactions in the real world. Unlike traditional AI systems that rely on large labeled datasets, V-JEPA 2 operates in a simplified latent space, enabling faster and more adaptable simulations of physical reality. The model learns cause-and-effect relationships such as gravity, motion, and object permanence by analyzing how people and objects interact in videos, allowing it to generalize across diverse contexts without extensive annotations. Meta views this development as a significant step toward artificial general intelligence (AGI), aiming to create AI systems capable of thinking before acting. In practical applications, Meta has tested V-JEPA 2 on lab-based robots, which successfully performed tasks like picking up unfamiliar objects and navigating new environments, demonstrating improved adaptability in unpredictable real-world settings. The company envisions broad use cases for autonomous machines—including delivery robots and self-driving cars—that require quick interpretation of physical surroundings and real

    roboticsartificial-intelligencemachine-learningautonomous-robotsvideo-based-learningphysical-world-simulationAI-models
  • Meta says its Llama AI models have been downloaded 1.2B times

    MetaLlama-AIartificial-intelligencedownloadstechnology-newsmachine-learningAI-models
  • Alibaba unveils Qwen 3, a family of ‘hybrid’ AI reasoning models

    AlibabaQwen-3AI-modelshybrid-AImachine-learningtech-newsopen-source-AI