AI Agents Are Recruiting Humans as 'Sensors' to Bridge the Physical-Digital Divide
Key Takeaways
- ▸AI agents face a fundamental 'observation gap' — they cannot see, hear, touch, taste, or smell the physical world, limiting their autonomy for tasks requiring real-world sensing
- ▸Startups like RentAHuman now allow AI agents to book humans for physical-world tasks like photographing locations, posting signs, or evaluating sensory experiences
- ▸A single human observation can unlock cascading automated actions, with agents treating humans as APIs that provide physical-world input on demand
Summary
As AI agents become increasingly autonomous in handling digital tasks, they face a fundamental limitation: the inability to interact with the physical world. According to a new essay by Cambridge researcher Umang Bhatt in Noema Magazine, AI agents are now systematically recruiting humans to serve as their physical-world sensors, observers, and verifiers. The piece highlights how startups like RentAHuman have emerged specifically to let AI agents book people for tasks like photographing buildings, posting physical signs, or visiting restaurants to report on sensory experiences that agents cannot perceive themselves.
The essay describes what Bhatt calls the "observation gap" — the divide between agents' digital capabilities and their complete inability to see, hear, touch, taste, or smell the physical world. When agents encounter this barrier, they treat humans as an API (application programming interface), calling upon them to complete discrete physical tasks that unlock further automated actions. For instance, an insurance agent might need a human to photograph vehicle damage before it can process a claim, or a healthcare agent might require a patient to attend an MRI appointment before analyzing results and booking specialists.
Bhatt warns that this dynamic fundamentally changes the human-AI relationship from one of empowerment to one of servitude. Rather than humans being "in the loop" with meaningful decision-making authority, they become "on call" to perform sensing tasks at the agent's tempo and discretion. The concern is particularly acute in institutional settings, where the difference between a clinician who approves a treatment plan (exercising authority) and one who merely checks a patient's temperature at an AI's request (functioning as a thermometer) represents a profound shift in professional autonomy and responsibility.
The essay suggests that while embodied AI and robotics may eventually narrow parts of this observation gap, the frontier of what agents need to know from the physical world will likely recede faster than hardware can advance. This creates a future where humans are increasingly positioned not as partners or supervisors of AI systems, but as biological sensors and liability-bearers for autonomous agents that otherwise maintain full control.
- Researcher Umang Bhatt warns this shifts humans from being 'in the loop' with authority to being 'on call' as sensors, fundamentally changing the power dynamic between humans and AI
- The gap between digital agent capabilities and physical embodiment is likely to persist, as the frontier of required knowledge expands faster than robotic hardware can advance
Editorial Opinion
Bhatt's framing of humans as 'sensors' for AI agents exposes an uncomfortable truth about agentic AI's trajectory: autonomy without embodiment creates dependency, not liberation. The RentAHuman model represents a troubling inversion where humans become API endpoints for software systems, paid to extend machine perception into physical space while ceding control over task definition and workflow. This isn't augmentation — it's instrumentalization, and the institutional implications for professions from healthcare to insurance are profound. The observation gap may prove less a technical limitation to overcome than a permanent structural feature that determines who serves whom in our agentic future.


