RIEM News LogoRIEM News

Articles tagged with "AI-models"

  • Google's Gemini model lets humanoid robot carry out multimodal tasks

    Google DeepMind has unveiled advancements in its humanoid robots powered by the Gemini Robotics 1.5 AI models, enabling them to perform complex, multi-step tasks through multimodal reasoning. Demonstrated in a recent video, the bi-arm Franka robot successfully completed the "banana test," sorting different fruits by color into separate plates, showcasing improved capabilities over previous models that could only follow single-step instructions. Another test featured Apptronik’s Apollo humanoid sorting laundry by color, even adapting to changes in basket positions mid-task, highlighting the robots' enhanced perception and adaptability. The Gemini Robotics 1.5 family includes two complementary models: one that converts visual inputs and instructions into actions, and another that reasons about the environment to create step-by-step plans. This agentic framework allows robots to autonomously study their surroundings, make decisions, and execute tasks such as sorting waste according to local recycling rules by researching guidelines online and applying them in real time. Google emphasizes safety in these models, incorporating risk assessment

    roboticshumanoid-robotsAI-modelsmultimodal-tasksautonomous-robotsrobot-perceptionrobot-reasoning
  • FieldAI raises $405M to build universal robot brains

    FieldAI, a robotics AI company, announced a $405 million funding raise to develop universal "robot brains" capable of controlling diverse physical robots across varied real-world environments. The latest funding round, including a $314 million tranche co-led by Bezos Expedition, Prysm, and Temasek, adds to backing from investors such as Khosla Ventures and Intel Capital. FieldAI’s core innovation lies in its "Field Foundation Models," which integrate physics-based understanding into embodied AI—AI that governs robots physically navigating environments—enabling robots to quickly learn, adapt, and manage risk and safety in new settings. This physics-informed approach contrasts with traditional AI models that often lack risk awareness, making FieldAI’s robots better suited for complex and potentially hazardous environments. Founder and CEO Ali Agha emphasized that their goal is to create a single, general-purpose robot brain that can operate across different robot types and tasks, with a built-in confidence measure to assess decision reliability and manage safety thresholds. Agha’s decades

    robotartificial-intelligenceembodied-AIrobotics-safetyrobot-learningAI-modelsrobotics-technology
  • Buzzy AI startup Multiverse creates two of the smallest high-performing models ever

    Multiverse Computing, a leading European AI startup based in Spain, has developed two of the smallest yet high-performing AI models, humorously named after animal brain sizes: SuperFly and ChickBrain. These models are designed to be embedded in Internet of Things (IoT) devices and run locally on smartphones, tablets, and PCs without requiring an internet connection. SuperFly, inspired by a fly’s brain, is a compressed version of Hugging Face’s SmolLM2 135 model with 94 million parameters, optimized for limited data and voice-command applications in home appliances. ChickBrain, with 3.2 billion parameters, is a compressed version of Meta’s Llama 3.1 8B model and offers advanced reasoning capabilities, outperforming the original in several benchmarks such as MMLU-Pro, Math 500, GSM8K, and GPQA Diamond. The key technology behind these models is Multiverse’s proprietary quantum-inspired compression algorithm called CompactifAI, which significantly reduces model

    IoTAI-modelsmodel-compressionedge-computingembedded-AIquantum-inspired-algorithmssmart-devices
  • Nvidia unveils new Cosmos world models, infra for robotics and physical uses

    Nvidia has introduced a suite of new AI models and infrastructure aimed at advancing robotics and physical AI applications. The centerpiece is Cosmos Reason, a 7-billion-parameter vision-language model designed to enable robots and AI agents to "reason" by leveraging memory and physics understanding. This capability allows Cosmos Reason to function as a planning model, helping embodied agents determine their next steps, and it can be applied in data curation, robot planning, and video analytics. Alongside Cosmos Reason, Nvidia unveiled Cosmos Transfer-2, which accelerates synthetic data generation from 3D simulations or spatial controls, and a faster, distilled version of Cosmos Transfers optimized for speed. In addition to the AI models, Nvidia announced new neural reconstruction libraries that facilitate 3D simulation of real-world environments using sensor data, with integration into the open-source CARLA simulator. The company also updated its Omniverse software development kit and introduced new hardware solutions, including the Nvidia RTX Pro Blackwell Servers tailored for robotics workflows and the DG

    roboticsAI-modelsNvidia-Cosmossynthetic-data-generation3D-simulationrobot-planningneural-reconstruction
  • Meta’s new AI helps robots learn real-world logic from raw video

    Meta has introduced V-JEPA 2, an advanced AI model trained solely on raw video data to help robots and AI agents better understand and predict physical interactions in the real world. Unlike traditional AI systems that rely on large labeled datasets, V-JEPA 2 operates in a simplified latent space, enabling faster and more adaptable simulations of physical reality. The model learns cause-and-effect relationships such as gravity, motion, and object permanence by analyzing how people and objects interact in videos, allowing it to generalize across diverse contexts without extensive annotations. Meta views this development as a significant step toward artificial general intelligence (AGI), aiming to create AI systems capable of thinking before acting. In practical applications, Meta has tested V-JEPA 2 on lab-based robots, which successfully performed tasks like picking up unfamiliar objects and navigating new environments, demonstrating improved adaptability in unpredictable real-world settings. The company envisions broad use cases for autonomous machines—including delivery robots and self-driving cars—that require quick interpretation of physical surroundings and real

    roboticsartificial-intelligencemachine-learningautonomous-robotsvideo-based-learningphysical-world-simulationAI-models
  • Meta says its Llama AI models have been downloaded 1.2B times

    MetaLlama-AIartificial-intelligencedownloadstechnology-newsmachine-learningAI-models
  • Alibaba unveils Qwen 3, a family of ‘hybrid’ AI reasoning models

    AlibabaQwen-3AI-modelshybrid-AImachine-learningtech-newsopen-source-AI