<ul data-eligibleForWebStory="true">Meta's latest model, V-JEPA 2, aims to enable robots to manipulate objects in unfamiliar environments.V-JEPA 2 learns a "world model" from video and interactions, aiding in predicting outcomes and planning actions.The model focuses on understanding scenes, predicting scene changes, and planning action sequences.It operates efficiently by predicting high-level features of a scene rather than pixel-level details.The two-stage training enables zero-shot robot planning, allowing robots to manipulate new objects without retraining.This advancement benefits industries like logistics and manufacturing by enabling adaptable robots and digital twins.V-JEPA 2's lower training costs and abstract predictions make it suitable for real-world deployment.Meta's release of the model and training code aims to drive progress in developing world models for AI interactions in the physical world.The model's ability to plan and act in new situations can revolutionize business operations and industrial safety.V-JEPA 2's approach aligns with Meta's vision of advanced machine intelligence (AMI) and software-defined robotics.The model's low training overhead reshapes the cost equation for enterprises deploying AI in physical settings.