Back
December 10, 2025
The Next Productivity Engine: Humanoid Robots and Beyond

Jaime Paik
Director of the RRL at EPFL

Yuhao Jiang
Postdoc Researcher at EPFL

Mark Esposito
Chief Economist at micro1

Arian Sadeghi
VP, Robotics Data at micro1
In an era defined by rapid technological advancement, the convergence of artificial intelligence (AI) and robotics is poised to redefine productivity across industries and daily lifeᶦ. As we stand on the brink of widespread automation, humanoid robots emerge as a transformative force, promising to liberate human labor from repetitive and hazardous tasksᶦᶦ. Yet, to fully realize this potential, we must address key challenges and innovate beyond traditional robotic designs. This article explores the evolution of robotics, the limitations of current systems, and the promise of modular reconfigurable robots as the next frontier in automation—drawing parallels to human ingenuity and tool use.
The Dawn of Intelligent Automation
Robotics has evolved dramatically, from specialized single-function machines like industrial arms and grippers (e.g., Universal Robotsᶦᶦᶦ, KUKAᶦᵛ, etc.) to multifunctional systems such as quadrupeds (e.g., Boston Dynamicsᵛ, Unitreeᵛᶦ), drones (e.g., DJIᵛᶦᶦ, Skydioᵛᶦᶦᶦ), and humanoid robots (e.g., Figure AIᶦˣ, Agility Roboticsˣ, and Tesla Optimusˣᶦ). Powered by AI, these machines are becoming more intelligent, capable of performing complex, context-aware tasks across environments — from factory conveyor belts to household chores. According to the International Federation of Robotics, the industrial robot deployments have doubled in 10 yearsˣᶦᶦ. The promise is clear: an automation revolution that frees manual workers from mundane or dangerous labor, allowing them to focus on higher-value activities.
Humanoid robots represent a leap toward general-purpose intelligence. Modeled on the human form, humanoid robots can navigate doorways and ladders, manipulate tools, and interpret social cues, enabling seamless integration into human spaces. This versatility supports smoother collaboration, reduces the need for workplace redesign, and positions them as foundational tools for both daily life and industrial transformation. Goldman Sachs projects the humanoid market could reach $38 billion by 2035ˣᶦᶦᶦ, signaling surging demand and a once-in-a-generation opportunity for transformative growth. However, the real-world deployment of humanoid robots is still modestˣᶦᵛ. Intelligence remains a limiting factor: current systems lack the general reasoning and autonomy needed for complex, open-ended tasks. Safety and regulation also constrain progress: rigid structures and powerful actuators pose injury risks, and there are no humanoid-specific standards to govern design, deployment, or use. Energy is a third bottleneck: continuous balance control and multi-joint actuation drain batteries quickly, keeping runtimes to only a few hours. The trade-off is familiar—broad adaptability and intuitive interaction versus peak performance and durability in specialized or extreme settings.
Preparing humanoid AI for general intelligent automation is now a pressing goal, and it will take progress on both software and hardwareˣᵛ. On the software side, foundational AI models need to mature to deliver reliable, dexterous automation, and the rapid advances in large language models (LLMs) suggest this is within reach. The hardware side is tougher: fixed embodiments with limited joints and rigid structures impose hard limits—wheels outperform legs in many scenarios, and drones can reach places humanoids can’t. To bridge this gap, we envision modular, reconfigurable robotic systems as a complementary path alongside humanoids, providing flexible building blocks that extend their capabilities—much like how humans use tools—to unlock more general, adaptable automation.
The Human Data Layer Behind Embodied Intelligence
To address the intelligence limitations of previous generations, every meaningful jump in modern robotics has come from better data. We are now at the point where robots are leaving controlled lab spaces and entering real homes, offices, and industrial environments. As this transition occurs, the training distribution must reflect the real world: cluttered desks, mixed lighting, awkward object placements, and the corrective motions humans make subconsciously. Three practical data pipelines have emerged as the backbone of this new embodied intelligence.
1. Human Demonstration Data
The most effective method to teach robots real-world physics and logic is to observe human operation. Research in large-scale manipulation models, such as RT-1, has demonstrated that human demonstrations dramatically improve generalization in real tasksˣᵛᶦ. Teams now capture everyday activities from multiple viewpoints—POV and fixed third-person—often incorporating depth or stereo vision to capture fine-grained geometry. The value lies not in sheer volume, but in clarity: full sequences showing how a human begins, adjusts, corrects, and completes a task. Datasets like Ego4D have shown that natural human footage captures variation and nuance that synthetic or simulated data cannot yet reproduceˣᵛᶦᶦ.
2. Teleoperation and Robot-Native Data
Some behaviors can only be taught directly on the robot. Teleoperation streams; joint angles, force feedback, gripper states, end-effector trajectories, synced RGB/RGB-D, create robot-native motion examples that humans can’t physically provide. This approach is becoming more common, especially in systems like Tesla’s Optimus program, where teleop data feeds robot-specific training loopsˣᵛᶦᶦᶦ. It closes the gap between human motion and the robot’s own mechanical constraints.
3. Action Segmentation and Annotation
Raw footage requires structure to be useful. This layer breaks demonstrations into timestamped actions, natural-language descriptions, and object interactions, standardizing them into consistent formats. Recent large-scale efforts like Open X-Embodiment have highlighted how crucial this consistency is for cross-task and cross-environment generalizationˣᶦˣ.
Why Human Demonstrations are the Critical Bottleneck
The integration of these pipelines addresses the fundamental constraint in modern robotics. The field is no longer held back primarily by actuators or compute power, but by grounded understanding—the ability to grasp a cup, stabilize an object, or adjust mid-motion fluidly. Human demonstrations bridge this gap in three specific ways:
- Capturing Reality: Simulators still struggle to accurately model clutter, deformable objects, and occlusions. Real-world demonstrations capture these complexities automatically, along with the micro-adjustments humans make subconsciously.
- Providing Strong Priors: Before a robot can understand its own mechanics, it requires a blueprint of how a task should look. Human motion provides the necessary "priors" to guide the learning process.
- Encoding Human Norms: Robots operating in human spaces must be predictable. Human-collected data naturally embeds pacing, caution, and intent, ensuring safe interaction.
As humanoid platforms approach real deployment, the decisive factor for their utility will not be hardware specs, but the depth and structure of this human data layer. However, while the software "brain" is rapidly maturing through these data strategies, the hardware "body" remains a stubborn bottleneck.
The Robot Scaling Wall: Lessons from AI
Recent results on LLMs show diminishing returnsˣˣ: beyond a point, adding GPUs and more training data yields smaller gains and can even plateau—a manifestation of “AI’s scaling wall.” We propose an analogous “robot scaling wall.” In robotics, more actuators, linkages, and sensors can expand capability, but at mounting costs: higher mass and power demand, and greater control complexity. Simply scaling toward a “general AI robot” often isn’t sensible—practical limits create diminishing or even negative returns and can doom the effort. For example, moving from a quadruped or drone to a full humanoid adds versatility but can make the system impractical in tight spaces or for long-duration missions. Progress toward general humanoid AI isn’t about scaling hardware alone; it requires intelligent co-design that balances capability with efficiency.
Adaptable humanoids need new “tools”: Modular and Reconfigurable Robots
We envision progress toward general humanoid AI that pairs a versatile humanoid platform with modular, reconfigurable robots. The humanoid delivers broad competence in human-centric environmentsˣˣᶦ—navigation, perception, and basic manipulation—while interchangeable, reconfigurable modules provide the specialized “last mile” of capability with greater reach, efficiency, and precision.
Modular reconfigurable robots (e.g., Tribotˣˣᶦᶦ, MORI3ˣˣᶦᶦᶦ, Roombotsˣˣᶦᵛ) function like Lego blocks: interchangeable modules can be reconfigured as grippers, wheels, sensor pods, or specialized limbs that mount to a humanoid base or other platforms. On a factory floor, humanoids can quickly swap reconfigurable modular robots—configured as torque drivers, riveters, welders, vision‑guided inspection heads, or micromanipulators—to pivot across diverse assembly and teardown tasks with lower energy and actuation effort. In clinical settings, they can attach reconfigurable modular robots configured as precision surgical or assistive instruments. Modularity avoids one‑size‑fits‑all designs, enabling tailored solutions without over‑engineering. In this view, the humanoid is the versatile platform that gets you 90% of the way, and modular toolkits provide the specialized final 10%—extending reach, reducing effort, and enabling expert performance where standard hardware falls short.
Economic Implications: Freeing Labor and Creating Opportunities
The economics of the robotic revolution go well beyond workplace convenience; they represent one of the defining productivity shifts of the 21st century. Deploying humanoid robots and modular systems at scale has the potential to reallocate labor from repetitive, low-skill, or high-risk tasks, mining, assembly-line work, or hazardous waste handling into higher-value activities. This transition carries two critical payoffs: reducing costs associated with workplaceˣˣᵛ accidents and inefficiencies and unlocking new growth opportunities by redeploying human capital into creative, supervisory, and analytical functions.
Historically, fears of net job destruction have not materialized; each wave of automation—from mass production in the 1970s–80s to digital automation in the 2010s–20s has ultimately expanded labor markets by creating complementary industries and demand for new skill sets. The 2020s and 2030s are likely to repeat this pattern, with humanoid and modular robotics catalyzing entirely new categories of employmentˣˣᵛᶦ: remote robot fleet supervisors, modular system designers, and trainers of artificial intelligence (AI) models for robotic behavior. These are not marginal gains; the redirection of labor into higher-productivity tasks could add trillions to global GDP as adoption scales.
At the same time, modular robotics amplify human–machine complementaritiesˣˣᵛᶦᶦ rather than replacing them outright. Farmers in developing economies could integrate reconfigurable drones and humanoid assistants to manage crop cycles more efficiently, while logistics operators in advanced economies could partner with robotics to optimize last-mile delivery. These dual benefits, lowering operational costs and widening access to advanced production tools democratize productivity growth. In short, the economic case for embracing this robotic transition is straightforward. The costs of hesitation, lost competitiveness, underutilized labor, and exposure to workplace risks indeed are quantifiableˣˣᵛᶦᶦᶦ. The rewards on the rise of new industries, higher-value jobs, and accelerated growth are equally measurable. This is not only about technological progress; it is about ensuring that productivity gains translate into broad-based prosperityˣˣᶦˣ in the decades ahead.
Looking Ahead: Challenges and the Path Forward
Over the next decade, general humanoid AI paired with reconfigurable modular robots can progress from pilots to a pervasive, human‑centric automation layer. Humanoids will handle safe navigation and baseline manipulation in human spaces, while hot‑swappable modules—spanning mobility, sensing, and task‑specific tools—extend capability on demand without redesigning workplaces or overbuilding hardware. This vision will catalyze an industry flywheel: open standards for mechanical, power, and data interfaces will attract third‑party module makers; marketplaces for modules and skills will lower integration costs; and shared datasets and simulators will accelerate learning. In turn, these dynamics will spur technical and scientific advances—lighter, more efficient actuators; compliant materials; energy‑aware control; self‑calibrating perception; and formal safety methods for reconfiguration. Economically, this architecture can shift labor from repetitive or hazardous tasks into supervisory, creative, and analytical roles, cut accident and downtime costs, and create new occupations in robot operations, module design, and AI training. The result is a compounding productivity engine that broadens participation in innovation and spreads its gains across sectors and economies.
References
ᶦ Billard, A., Albu-Schaeffer, A., Beetz, M. et al. A roadmap for AI in robotics. Nat Mach Intell 7, 818–824 (2025). https://doi.org/10.1038/s42256-025-01050-6
ᶦᶦ Hanbury, P., Dutt, A., & Veratti, E. (2025, April). Humanoid robots at work: What executives need to know. Bain & Company. https://www.bain.com/insights/humanoid-robots-at-work-what-executives-need-to-know/
ᶦᶦᶦ Universal Robots. (n.d.). Collaborative robotic automation. Universal Robots A/S. Retrieved October 15, 2025, from https://www.universal-robots.com/
ᶦᵛ KUKA SE & Co. KGaA. (n.d.). Industrial intelligence 4.0_beyond automation. Retrieved October 15, 2025, from https://www.kuka.com/
ᵛ Boston Dynamics. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://www.bostondynamics.com/
ᵛᶦ Unitree Robotics. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://www.unitree.com/ᵛᶦ
ᵛᶦᶦ DJI. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://www.dji.com/
ᵛᶦᶦᶦ Skydio. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://www.skydio.com/
ᶦˣ Figure AI. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://www.figure.ai/
ˣ Agility Robotics. (n.d.). Home [Website]. Retrieved October 15, 2025, from https://agilityrobotics.com/
ˣᶦ Tesla. (n.d.). Optimus [Website]. Retrieved October 15, 2025, from https://www.tesla.com/optimus
ˣᶦᶦ The Robot Report Staff. (2025, September 25). IFR: Industrial robot deployments have doubled in 10 years. The Robot Report. https://www.therobotreport.com/ifr-industrial-robot-deployments-have-doubled-in-10-years/
ˣᶦᶦᶦ Goldman Sachs Research. (2024, February 27). The global market for humanoid robots could reach $38 billion by 2035. Goldman Sachs. https://www.goldmansachs.com/insights/articles/the-global-market-for-robots-could-reach-38-billion-by-2035
ˣᶦᵛ O'Donnell, J. (2025, May 6). Why the humanoid workforce is running late. MIT Technology Review. https://www.technologyreview.com/2025/05/06/1116108/why-the-humanoid-workforce-is-running-late/
ˣᵛ Kelkar, A., Jansen, C., Webster, M., Vaish, S., & Nepal, S. (2025, July 28). A leap in automation: The new technology behind general-purpose robots. McKinsey & Company. https://www.mckinsey.com/industries/industrials-and-electronics/our-insights/a-leap-in-automation-the-new-technology-behind-general-purpose-robots
ˣᵛᶦ Brohan, A. et al. “RT-1: Robotics Transformer for Real-World Control.” Google DeepMind (2022). https://research.google/blog/rt-1-robotics-transformer-for-real-world-control-at-scale/
ˣᵛᶦᶦ Grauman, K. et al. “Ego4D: Around the World in 3,000 Hours of Egocentric Video.” Meta AI (2021). https://ai.meta.com/research/publications/ego4d-unscripted-first-person-video-from-around-the-world-and-a-benchmark-suite-for-egocentric-perception/
ˣᵛᶦᶦᶦ Tesla AI Day Presentation (2021). https://ironbar.github.io/understanding_intelligence/artificial-intelligence/04-tesla-ai-day-2021/
ˣᶦˣ Shah, D. et al. “Open X-Embodiment: Robotic Learning Datasets and RT-X Models” Stanford, CMU, Google, DeepMind (2023). https://robotics-transformer-x.github.io/
ˣˣ Masood, A. (2025, May 26). Is there a wall? An evidence-based analysis of diminishing returns in large language model scaling: Investigating benchmark plateaus, data scarcity, and computational-economic constraints. Medium. https://medium.com/@adnanmasood/is-there-a-wall-34d02dfd85f3
ˣˣᶦ Riso, S., & Adăscăliței, D. (2024). Human–robot interaction: What changes in the workplace? European Foundation for the Improvement of Living and Working Conditions. https://www.eurofound.europa.eu/en/publications/all/human-robot-interaction-what-changes-workplace
ˣˣᶦᶦ Zhakypov, Z., Mori, K., Hosoda, K. et al. Designing minimal and scalable insect-inspired multi-locomotion millirobots. Nature 571, 381–386 (2019). https://doi.org/10.1038/s41586-019-1388-8
ˣˣᶦᶦᶦ Belke, C.H., Holdcroft, K., Sigrist, A. et al. Morphological flexibility in robotic systems through physical polygon meshing. Nat Mach Intell 5, 669–675 (2023). https://doi.org/10.1038/s42256-023-00676-8
xxiv Hauser, S., Mutlu, M., Léziart, P.-A., Khodr, H., Bernardino, A., & Ijspeert, A. J. (2020). Roombots extended:Challenges in the next generation of self-reconfigurable modular robots and their application in adaptive and assis-tive furniture. Robotics and Autonomous Systems, 127, Article 103467. https://doi.org/10.1016/j.robot.2020.103467
ˣˣᵛ Mecademic. (n.d.). The impact of automation on the global economy. https://mecademic.com/blog/impact-automation-global-economy/
ˣˣᵛᶦ Holzer, H. J. (2022, January 19). Understanding the impact of automation on workers, jobs, and wages. Brookings Institution. https://www.brookings.edu/articles/understanding-the-impact-of-automation-on-workers-jobs-and-wages/
ˣˣᵛᶦᶦ Atkinson, R. D. (2019, October 15). Robotics and the future of production and work. Information Technology and Innovation Foundation. https://itif.org/publications/2019/10/15/robotics-and-future-production-and-work/
ˣˣᵛᶦᶦᶦ Stephens, R. (2015, September 22). Robots at work: The economic effects of workplace automation. The Journalist's Resource. https://journalistsresource.org/economics/robots-at-work-the-economics-effects-of-workplace-automation/
ˣˣᶦˣ Bharadwaj, A., & Dvorkin, M. A. (2019, July 10). The rise of automation: How robots may impact the U.S. labor market. Federal Reserve Bank of St. Louis. https://www.stlouisfed.org/publications/re.g.ional-economist/second-quarter-2019/rise-automation-robots
.webp)
