Gemini Robotics 1.5 brings AI agents into the physical world at Google DeepMind's blog post at https://lnkd.in/gpC-5jXA
Some really cool developments in robotics from Google.
#Gemini Robotics 1.5 is a VLA that can now think 🧠 while taking action 🦾 , generating an internal sequence of reasoning using natural language 🤯. This makes the robot actions more interpretable (you can literally read the robot's thoughts), and unlocks more useful multi-step tasks.
Learn more at https://lnkd.in/gkNmCieXGoogle DeepMind Robotics
#Gemini Robotics 1.5 is a VLA that can now think 🧠 while taking action 🦾 , generating an internal sequence of reasoning using natural language 🤯. This makes the robot actions more interpretable (you can literally read the robot's thoughts), and unlocks more useful multi-step tasks.
Learn more at https://lnkd.in/gkNmCieXGoogle DeepMind Robotics
Gemini Robotics 1.5 released to developers. Someone put the wrong parts in a parts bin? No problem, tell your robot to sort through them. Need to rework product due to a quality defect? No problem, tell your robot to start picking the items for your ad-hoc bill of materials. A few of its capabilities:
The model excels at generating semantically-precise 2D points, grounded in reasoning about item sizes, weights, and affordances, enabling commands like 'point at anything you can pick up' for accurate, responsive interaction.
Leverage advanced spatial and temporal reasoning, planning, and success detection for reliable long-horizon task execution loops (e.g., "re-organize my desk according to this picture").
It can also natively call tools like the Google Search tool and any third-party user-defined functions (e.g. “separate the trash into the correct bins according to local rules”).
Incredible work by the #Gemini Robotics team.
#Google#Googlecloud#manufacturing#ai
#Gemini Robotics 1.5 is a VLA that can now think 🧠 while taking action 🦾 , generating an internal sequence of reasoning using natural language 🤯. This makes the robot actions more interpretable (you can literally read the robot's thoughts), and unlocks more useful multi-step tasks.
Learn more at https://lnkd.in/gkNmCieXGoogle DeepMind Robotics
🚨 Google DeepMind introduces Gemini Robotics 1.5 – a path towards robotics generalization
Google DeepMind has unveiled Gemini Robotics 1.5 and Gemini Robotics-ER 1.5, new models that bring agentic AI into the physical world.
🤖 What they do
Gemini Robotics 1.5 (VLA): turns visual input + language into robot actions, thinks before acting, and transfers skills across robot types.
Gemini Robotics-ER 1.5 (VLM): plans and reasons at a high level, creates multi-step strategies, and calls external tools (like Google Search) to complete tasks.
🌍 Why it matters
This is a milestone toward general-purpose robots — systems that can understand context, adapt to new environments, and safely perform complex, multi-step tasks.
🔮 What to expect
Gemini Robotics-ER 1.5 available now via the Gemini API.
Gemini Robotics 1.5 accessible to select partners.
Both models lay the foundation for robots that reason, plan, and act more intelligently in the real world.
📌 Read more: https://lnkd.in/d_RigCQC#AI#Robotics#EmbodiedAI#Gemini
Can artificial intelligence and creativity combine to create profitable ventures? One individual has explored the intersection of business, creativity, and AI, transitioning from traditional employment to entrepreneurship and then delving into simulation, robotics, and AI. The emergence of LLMs has democratized AI, making its power accessible to a wider audience. Curious to know how others are leveraging AI in unconventional ways? #artificialintelligence#creativity#businessintelligence#robotics#LLMs
Introducing our first widely available Robotics model: Gemini Robotics-ER 1.5 🤖
It is currently SOTA on a set of Embodied Reasoning tasks and can be used used directly through the Gemini API or Google AI Studio!
It is very clear that robotics is the future, so seeing both the capability transfer from mainline Gemini to this use case but also the pace of research progress specific to embodied reasoning is amazing.
https://lnkd.in/gTqEzuDY
A certain movie and its source material may become increasingly relevant as AI and robotics continue to advance. The debate between training AI in simulated environments versus the natural world mirrors discussions happening in labs worldwide, from Google DeepMind to Boston Dynamics. As AI evolves, these discussions will only intensify. This concept of testing AI in natural settings versus simulation is crucial for the future of robotics. Do you believe real-world testing is essential for AI development? #AI#robotics#technology#innovation#future
Gemini Robotics is not merely another AI innovation—it's a vision-language-action (VLA) model that enables robots to see, reason, and act. From interpreting natural language instructions to executing intricate physical tasks, this tech is a giant leap toward intelligent, adaptable machines.
✅ Connecting perception and action
✅ Industry-agnostic general-purpose utility
✅ Smooth human-AI interaction
?#AI#DeepMind#Robotics#Innovation#FutureOfWork
Robotics milestone!
Figure AI’s humanoid robot Figure 02 can now load a dishwasher with human-like precision — stacking plates, glasses & dishes carefully, even in tight spaces.
Powered by Helix AI, the robot learns tasks by simply watching humans, no coding required. It even recovers from mistakes smoothly.
Since Helix launched in Feb, Figure 02 has been mastering more everyday tasks — bringing us one step closer to practical household robots.
#FigureAI#HelixAI#HumanoidRobots#AI
👀 Ever wondered what happens when you give a robot eyes and a brain?
Well... I’ve been working on exactly that! 🤖✨
My new project (still under development) is an 𝗔𝗜 𝗩𝗶𝘀𝗶𝗼𝗻 𝗔𝘂𝘁𝗼𝗻𝗼𝗺𝗼𝘂𝘀 𝗥𝗼𝗯𝗼𝘁 that can see, detect, and react to its surroundings in real time. 🚀
The best part? Watching it recognize objects, make decisions on its own and navigating objects autonomously is so satisfying.
📹 Sharing a quick snippet of its AI vision in action — this is just the beginning!
🛠️ Tech stack powering it:
🔹𝗥𝗮𝘀𝗽𝗯𝗲𝗿𝗿𝘆 𝗣𝗶 𝟰𝗕
🔹𝗘𝗦𝗣𝟯𝟮
🔹𝗬𝗢𝗟𝗢𝘃𝟴-𝗟𝗶𝘁𝗲 𝗳𝗼𝗿 𝗔𝗜 𝗿𝗲𝗰𝗼𝗴𝗻𝗶𝘁𝗶𝗼𝗻
I want to thank my teammate Ishpreet Singh to add his valuable contributions to the project.
Can’t wait to push this further — would love to hear your thoughts! 🙌
#AI#Robotics#ComputerVision#YOLOv8#RaspberryPi#ESP32#Innovation
🚀 Thrilled to share a quick glimpse of our latest work-in-progress 🌊🤖
This video highlights an early prototype along with a simulated environment. While it’s still at a basic stage, it marks the beginning of something exciting ✨.
Our goal is to explore how 🤖 robotics and 💡 intelligent systems can be applied in real-world problem-solving, especially in critical situations 🌐⚡.
Grateful to be working alongside my amazing collaborators:
Sai Nikhil Reddy Mondeddu 🙌
Eager to keep learning 📚, improving 🔧, and pushing this project forward 🚀!
#Robotics#AI#Innovation#WorkInProgress#Engineering