Agentic AI systems are increasingly bridging the digital and physical realms, allowing AI agents to perceive, reason, and act in real environments. As these systems expand into robotics, autonomous vehicles, and smart infrastructure, a key question arises: how can developers create agents that harness massive cloud computing power for complex reasoning while ensuring millisecond-level responsiveness for physical interactions?
This year has marked a significant evolution for agentic AI at AWS. In May 2025, the company introduced Strands Agents, which simplifies the developer experience and emphasizes a model-driven approach to agent development. Following that, in July, AWS launched version 1.0, featuring multi-agent orchestration capabilities and the introduction of Amazon Bedrock AgentCore to expedite AI agents into large-scale production. At the recent re:Invent 2025 conference, enhancements included the TypeScript SDK and bidirectional streaming for voice agents. Today, AWS is focusing on extending these capabilities into the edge and physical AI, where agents not only process information but also collaborate with humans in real-world scenarios.
Demonstrations at AWS illustrate how physical AI agents control disparate robotic platforms through a unified Strands Agents interface. One example involves a 3D-printed SO-101 robotic arm utilizing the NVIDIA GR00T vision-language-action (VLA) model to perform tasks like identifying and picking fruit. Meanwhile, a Boston Dynamics Spot quadruped robot executes mobility tasks, such as inspecting its sensors by reasoning through its physical state. Both systems operate on NVIDIA Jetson edge hardware, showcasing the ability to execute sophisticated AI tasks directly on embedded systems.
The interplay between edge and cloud computing reveals a fundamental tension in designing intelligent systems. A robotic arm catching a ball must react almost instantaneously—requiring local processing to overcome network latency challenges. Yet, this same system can benefit from cloud capabilities for complex tasks like multi-step assembly, which necessitates the computational power of the cloud. This dichotomy reflects Daniel Kahneman’s concept of dual thinking—quick, instinctual responses at the edge versus slower, deliberate reasoning in the cloud.
Cloud capabilities contribute additional features that edge systems cannot accommodate alone. For instance, AgentCore Memory enables the storage of spatial and temporal context over extended periods. This collaborative memory allows insights gained by one robot to be shared across fleets, fostering a collective learning environment. Furthermore, Amazon SageMaker facilitates the large-scale simulation and training of models, enabling organizations to integrate real-world and simulated learnings to enhance performance across entire fleets.
This hybrid architecture is paving the way for new categories of intelligent systems. For instance, humanoid robots can utilize cloud-based reasoning for planning multi-step tasks while using edge-based models for precise physical movements. Autonomous vehicles can optimize routes and predict traffic using cloud intelligence while maintaining real-time obstacle avoidance locally. This collaborative approach ensures that robots can respond to immediate dangers while leveraging broader analytical insights from the cloud.
Developing edge and physical AI systems need not begin with complex orchestration. Instead, a progressive iteration approach allows developers to start simple and evolve sophistication based on requirements. For example, initial setups can involve installing the Strands Agents Python SDK on edge devices and running local models such as Qwen3-VL.
As developers gain familiarity, they can integrate visual understanding through cameras, allowing the AI to comprehend its surroundings. Agents can also tap into other sensors, enhancing their decision-making capabilities. For instance, reading battery levels can inform the agent about task continuity or the need to recharge.
Moving from sensing to action, physical AI systems follow a continuous cycle of environmental interpretation, reasoning, and action execution. Control of hardware is central to physical interaction, as robots need to coordinate multiple motors and joints to perform precise tasks. Using VLA models like NVIDIA GR00T, robots can seamlessly integrate visual inputs with language instructions, enabling a more intuitive interaction.
The integration of tools such as Hugging Face’s LeRobot and robust motion primitives allows developers to program robotic systems that can operate efficiently within their environments. For example, using SDK-based control, a Boston Dynamics Spot can be commanded to perform complex tasks while ensuring the safety and efficiency of its movements.
Edge agents also have the capability to delegate complex reasoning tasks to cloud-based counterparts as necessary. By employing a layered agent architecture, a local agent can consult a more powerful cloud agent for intricate planning tasks, while the cloud agent can orchestrate multiple edge devices working collaboratively.
As physical AI systems evolve, learning from collective experiences becomes paramount. For instance, in a warehouse scenario, multiple robots encountering similar challenges can share observations, leading to operational optimizations. AgentCore Observability provides a mechanism for continuous improvement, where agents learn from their interactions and adjust strategies accordingly. This loop—comprising inference, observation, evaluation, and optimization—contributes to the overall enhancement of AI capabilities.
The convergence of multimodal reasoning models, edge computing, and open-source robotics is reshaping the capabilities of physical AI systems. AWS aims to make AI agent development accessible, facilitating a transition where agents learn from experience in real environments. As these systems become more adept at reasoning about the physical world, they will increasingly simulate future scenarios, predict outcomes, and coordinate seamlessly within larger operational frameworks.
See also
University of Florida Launches 40,000-Sq-Ft AI Hub to Revolutionize Farming and Combat Food Insecurity
Microsoft Invests $23B in AI, Raises Microsoft 365 Prices Ahead of 2026 Rollout
Google Reveals AI Glasses Coming in 2026, Features Built-In Assistant and Camera
Government Watchdog Warns: AI Tools Risk Discrimination and Higher Housing Costs for Buyers, Renters
Experts Warn: AI’s Bias Threatens Music Industry Integrity, Calls for Ethical Reforms




















































