Navigation
Search
|
Why fast-learning robots are wearing Meta glasses
Friday February 28, 2025. 12:00 PM , from ComputerWorld
AI and robots need data — lots of it. Companies that have millions of users have an advantage in this data collection, because they can use the data of their customers.
A well-known example is Google’s reCAPTCHA v2 service, which I recently lambasted. reCAPTCHA v2 is a cloud-based service operated by Google that is supposed to stop bots from proceeding to a website (it doesn’t). When users “prove they’re human” by clicking on the boxes containing traffic lights, crosswalks, buses and other objects specified in the challenge, that information is used to train models for Waymo’s self-driving cars and improve the accuracy of Google Maps. It’s also used to improve vision capabilities used in Google Photos and image search algorithms. Google isn’t the only company that does this. Microsoft processes voice recordings from Teams and Cortana to refine speech models, for example. Now Meta is using data user to train robots, sort of. Meta’s augmented reality (AR) division and researchers at Georgia Institute of Technology have developed a framework called EgoMimic, which uses video feeds from smartglasses to train robots. Regular robot imitation learning requires painstaking teleoperation of robotic arms — humans wearing special sensor-laden clothing and VR goggles perform the movements robots are being trained for, and the robot software “learns” how to do the task. It is a slow, expensive, and unscalable process. EgoMimic uses Meta’s Project Aria glasses to train robots. Meta announced Project Aria in September 2020 as a research initiative to develop advanced AR glasses. The glasses have five cameras (including monochrome, RGB, and eye-tracking), Inertial Measurement Units (IMUs), microphones, and environmental sensors. The project also has privacy safeguards such as anonymization algorithms, a recording indicator LED, and a physical privacy switch, according to Meta. The purpose of Aria is to enable applications in robotics, accessibility, and 3D scene reconstruction. Meta rolled out the Aria Research Kit (ARK) on Oct. 9. By using Project Aria glasses to record first-person video of humans performing tasks like shirt folding or packing groceries, Georgia Tech researchers built a dataset that’s more than 40 times more “demonstration-rich” than equivalent robot-collected data. The technology acts as a sophisticated translator between human and robotic movement. Using mathematical techniques called Gaussian normalization, the system maps the rotations of a human wrist to the precise joint angles of a robot arm, ensuring natural motions get converted into mechanical actions without dangerous exaggerations. This movement translation works alongside a shared visual understanding — both the human demonstrator’s smartglasses and the robot’s cameras feed into the same artificial intelligence program, creating common ground for interpreting objects and environments. A critical safety layer called “action masking” prevents impossible maneuvers, functioning like an invisible fence that stops robots from attempting any biomechanically implausible actions they observe in humans. The magic happens through the EgoMimic algorithm, which bridges the gap between human demonstration and robotic execution. Here’s the funny part: After collecting video data, researchers mount the same AR glasses onto a robot, effectively giving it “eyes” and sensor data that perceive the tasks from the exact same perspective as the human who demonstrated the task. That’s right. The research involves robots wearing glasses. (You can see what this looks like in the video published by Meta.) The algorithm then translates the human movements from the videos into actionable instructions for the robot’s joints and grippers. This approach reduces the need for extensive robot-specific training data. Training by teaching The approach taken by EgoMimic could revolutionize how robots are trained, moving beyond current methods that require physically guiding machines through each step. The EgoMimic approach basically democratizes robot training, enabling small business, farmers and others who normally wouldn’t even attempt robot training to do the training themselves. (Note that Meta’s and Georgia Tech’s EgoMimic is different from the University of North Carolina at Charlotte’s EgoMimic project, which uses first-person videos for training large vision-language models.) EgoMimic is expected to be publically demonstrated at the 2025 IEEE Engineers’ International Conference on Robotics and Automation in Atlanta, which begins May 19. This general approach is likely to move from the research stage to the public stage, where multimodal AI video feeds from millions of users wearing smartglasses (which in the future will come standard with all the cameras and sensors in Project Aria glasses). AI will identify the user videos where the user is doing something (cooking, tending a garden, operating equipment) and making that available to robot training systems. This is basically how chatbots got their AI for training — billions of people live their lives and in the process generate content of every sort, which is then harvested for AI training. The EgoMimic researchers didn’t invent the concept of using consumer electronics to train robots. One pioneer in the field, a former healthcare-robot researcher named Dr. Sarah Zhang, has demonstrated 40% improvements in the speed of training healthcare robots using smartphones and digital cameras; they enable nurses to teach robots through gestures, voice commands, and real-time demonstrations instead of complicated programming. This improved robot training is made possible by AI that can learn from fewer examples. A nurse might show a robot how to deliver medications twice, and the robot generalizes the task to handle variations like avoiding obstacles or adjusting schedules. The robots also use sensors like depth-sensing cameras and motion detectors to interpret gestures and voice instructions. It’s easy to see how the Zhang approach combined with the EgoMimic system using smartglasses and deployed at a massive scale could dramatically enhance robot training. Here’s a scenario to demonstrate what might be possible with future smartglasses-based robot training. A small business owner who owns a restaurant can’t keep up with delivery pizza orders. So, he or she buys a robot to make pizzas faster, puts on the special smartglasses while making pizzas, then simply puts the same glasses on the robot. The robot can then use that learning to take over the pizza-making process. The revolution here wouldn’t be robotic pizza making. In fact, robots are already being used for commercial pizzas. The revolution is that a pizza cook, not a robot scientist or developer, is training robots with his proprietary and specific recipes. This could provide a small business owner with a huge advantage over buying a pre-programmed robot that performs a task generically and identically to other buyers of that particular robot. You could see similar use cases in homes or factories. If you want a robot to perform a task, you simply teach it how to do that task by demonstrating it. Of course, this real-world, practical use case is years away. But using smartglasses to train robots isn’t something the public, or even technologists, are even thinking about right now. But smartglasses-based training is just one way robotics will be democratized and mainstreamed in the years ahead.
https://www.computerworld.com/article/3832789/why-fast-learning-robots-are-wearing-meta-glasses.html
Related News |
25 sources
Current Date
Mar, Sat 1 - 05:10 CET
|