Figure 1

Figure 1

General-purpose humanoid designed as an “embodied workforce” wedge: start with repeatable factory tasks (parts handling, machine tending), then expand scope via a data + model flywheel.

Figure FIGURE 01 (general-purpose humanoid)

FIGURE 01 is best understood as a systems bet: build a human-scale body that can physically fit existing industrial layouts, then make the intelligence layer general enough to learn new tasks without writing a new control stack for every job. The early proof points aren’t about acrobatics—they’re about whether the robot can run safely in a real production environment, recover from exceptions, and produce measurable throughput with an operator-friendly rollout playbook.

Domain: manufacturing / logistics (first wedge) Job: parts handling + machine-loading style tasks Commercial signal: BMW agreement + plant trials AI posture: Vision-Language-Action (VLA) roadmap Onboard compute claim: low-power embedded GPUs (company) Scaling vector: fleet + data flywheel Confidence: high on partnerships/news, medium on fine-grain FIGURE 01 specs

Humanoid robotics has a hype problem: videos can make anything look “nearly solved” while the real bottleneck is always the same—reliability at scale. FIGURE 01 sits in the middle of that tension. It was introduced as a general-purpose humanoid, but the practical path is a disciplined wedge: pick a small number of industrial tasks that are valuable, repeatable, and safety-verifiable, then widen the skill library over time.

The body’s strategic value is not “being human” for its own sake—it’s compatibility. Factories, warehouses, carts, bins, handles, doors, and workstations are designed around human geometry. A human-shaped robot can, in theory, operate in those spaces without forcing a facility redesign. That facility-compatibility is the real business case behind the form factor: lower integration friction, faster pilots, and fewer custom fixtures.

Where Figure tries to differentiate is the intelligence stack and the iteration loop. The company’s public narrative increasingly emphasizes generalist control: one model that can connect perception, language, and action, so you can describe a goal (“place these parts here,” “load this fixture,” “move this bin”) and the robot produces continuous control rather than a brittle scripted routine. In 2025, Figure introduced Helix as a Vision-Language-Action model aimed at generalist humanoid control, including high-rate upper-body control and claims about running onboard on embedded GPUs. Even if Helix is presented alongside newer hardware generations, it signals the direction: general-purpose behavior should come from a scalable model family, not handcrafted task code.

On the commercialization side, Figure’s early partnership narrative matters because it’s where robotics companies usually break. A commercial agreement with BMW Manufacturing (announced in early 2024) and subsequent public updates from BMW about plant testing are exactly the kind of external validation readers should value: it suggests the robot is being evaluated against real constraints—cycle time, safety procedures, aisle congestion, reflective surfaces, noise, and the relentless “every shift, every day” cadence that kills fragile prototypes.

The honest constraint: the public record for FIGURE 01’s precise hardware specs (DoF breakdown, actuator types, torque density, continuous payload curves, runtime under load, service intervals) is thin and often summarized by third parties. A high-quality analysis should separate what’s strongly confirmed (partnerships, direction of the AI stack, and the “factory wedge” thesis) from what is not reliably published (precise performance limits for FIGURE 01). That separation is not nitpicking—it’s how you keep your readers’ trust while still describing the system at a useful level.


Scoreboard

Mobility (factory realism)

74 / 100

Industrial humanoids must walk, turn, and stop predictably around people and equipment. Public evidence is growing, but long-run “shift after shift” robustness is still being proven.

Dexterity (task utility)

72 / 100

The target is not “perfect hands,” it’s reliable grasp + place on common industrial objects. The company direction (VLA + learned manipulation) is strong; measured throughput is the missing public metric.

Deployment maturity

76 / 100

A BMW agreement and plant testing are meaningful signals, but “maturity” ultimately means repeatable production work with published intervention and uptime metrics.

Safety architecture confidence

71 / 100

Factories demand conservative safety behavior: speed/force limiting, clear stops, audited logs, and site hazard analysis. Public details are limited; external trials increase confidence.

Fleet operations (productization)

69 / 100

The “real product” is deployment tooling: monitoring, staged updates, remote diagnosis, and service logistics. The company story implies it; the public artifacts are still sparse.

Scalability (manufacturing + economics)

67 / 100

Funding and partners help, but scale will be decided by serviceability, cost-per-hour, and whether the AI stack reduces—not increases—operator burden.

Technical specifications

Spec Details
Robot owner Figure (Figure AI Inc.)
Category General-purpose humanoid intended for industrial tasks (manufacturing/logistics wedge)
Height ~1.68–1.70 m (often listed as ~5'6"–5'7"). Fine-grain values vary across sources; treat as approximate unless confirmed by Figure.
Mass Often listed ~60 kg (third-party compilation; not consistently published by Figure as a primary spec).
Payload Often listed ~20 kg (third-party compilation; confirm per deployment task envelope).
Top speed Often listed ~1.2 m/s (third-party compilation; speed in production is usually governed by safety constraints, not mechanical max).
Runtime Often listed “up to ~5 hours” (third-party compilation; real runtime depends heavily on task intensity + payload + walk ratio).
Actuation / DoF disclosure Not consistently published in primary sources for FIGURE 01. Treat detailed actuator/DoF counts as “not publicly confirmed” unless Figure provides a spec sheet.
Intelligence stack direction Company has publicly introduced a generalist Vision-Language-Action model (Helix) for humanoid control, emphasizing unified perception+language+control and onboard execution (presented in 2025).
Commercial agreement Figure announced a commercial agreement with BMW Manufacturing (2024). BMW later publicly discussed testing humanoid robots with Figure at Plant Spartanburg (2024) and Figure published subsequent production results (later generation).
Public spec reliability High on partnerships + strategic direction; medium/low on precise FIGURE 01 hardware performance numbers (often sourced via third-party summaries).

What stands out beyond the scoreboard

Where FIGURE 01 is structurally ahead
  • Factory wedge is correct: Manufacturing tasks can be scoped, measured, and safety-validated faster than open-world household tasks—ideal for a first commercialization loop.
  • External validation: BMW’s engagement forces the system to face real constraints (line cadence, safety procedures, reflective materials, human co-work) rather than demo-stage perfection.
  • Generalist AI posture: The public direction (VLA models like Helix) targets a core robotics problem: scaling behaviors without per-task hand coding.
  • Data flywheel advantage: Industrial deployments produce the kind of dense sensor + action data that improves models—if you instrument it well and close the loop quickly.
  • Capital + ecosystem: Funding and partnerships can accelerate compute, hiring, and manufacturing iteration—helpful in a domain where “time to reliability” is the real competition.
Where the hard problems still live
  • Intervention economics: If a robot needs frequent human resets, the cost-per-hour collapses. The key missing public metric is mean time between interventions (MTBI).
  • Perception in industrial reality: Factories include glare, reflective metal, tight tolerances, and repeated occlusions—conditions that punish brittle perception.
  • Manipulation variance: Real parts vary: small pose changes, flexible packaging, inconsistent bins. “Works in a video” is not “works in production.”
  • Safety as architecture: Speed limits, force limits, safe stops, and audited logs must be independent of “smart” behaviors. This is costly, slow, and unavoidable.
  • Serviceability + uptime: The winner isn’t only the best robot; it’s the best maintenance system: spare parts, diagnostics, repair time, and update discipline.

Provider pick

Priority Pick Why Tradeoff to accept
Best first environment Automotive / manufacturing cells with defined stations and clear human-robot zones Station-based work reduces ambiguity: consistent pickup zones, known fixtures, and predictable routing make safety and reliability achievable earlier. Early success will be scaffolded (marked zones, standardized bins). That’s not “cheating”—it’s how you ship reliability.
Best first workload Parts transfer + simple machine tending (load/unload style steps) High-frequency, measurable tasks that produce clear KPI signals (cycle time, success rate, intervention rate, safe-stop rate). Not glamorous. But boring tasks are exactly what pay and scale.
Fastest scaling loop Repeatable tasks across multiple identical stations Replicated stations produce cleaner data and faster iteration: you can improve one behavior and deploy it across many cells. Scope stays narrow at first; generality is earned over time, not declared.
Long-term bet Generalist humanoid skill library powered by VLA models + strict safety constraints If the model + fleet pipeline matures, the robot can expand from transfers into richer manipulation without per-task software rewrites. The last 10% is brutal: verification, safety validation, and service economics decide winners more than demos.

Real workloads cost table

FIGURE 01 should be evaluated like an operator evaluates automation: not by a highlight reel, but by how often the robot needs help. In early deployments, the hidden bill is exceptions—perception uncertainty, awkward grasps, congestion, safety stops, and “it got stuck” events. The table below frames cost in those terms.

Scenario Input Output What it represents Estimated cost driver
Parts bin transfer (bin ↔ fixture) Standard bins, known pickup pose ranges Parts delivered to a fixture / staging point Core “humanoid wedge”: human-like reach and handling in a human-built cell Regrasp attempts + pose estimation failures
Machine tending (load/unload) Defined machine access + safety interlocks Consistent load/unload cycles High ROI if cycle time is stable and supervision is minimal Recovery time after misalignment + safe-stop frequency
Cart / tote repositioning Known carts/totes, marked zones Materials routed to next station Low-dexterity but high-utility logistics inside the plant Navigation exceptions + traffic coordination with humans
Mixed-traffic aisle operations Humans, forklifts, dynamic obstacles Safe navigation plus task completion The reality check: can the robot behave conservatively without destroying throughput? Near-miss tolerance + conservative speed limiting
Higher-mix manipulation expansion More parts, varied packaging, varied station geometry Broader skill library Where “general purpose” becomes real—or collapses under complexity Training + verification + interventions (tail events)

How to control cost (a practical playbook)

1) Measure intervention economics first, not “task success”

Robots can “succeed” in demos while still being uneconomic in production. Your first KPI should be interventions: how often does a human need to reset, correct, or rescue the robot—and how long does that take?

  • Track mean time between interventions (MTBI) and the tail (rare failures dominate cost).
  • Log every intervention cause: perception uncertainty, grasp slip, path blockage, safety stop, planner dead-end.
  • Refuse scope expansion until MTBI improves across multiple shifts (not just one good day).
2) Design the station so the robot can be boring

The quickest ROI comes from lowering ambiguity. Standardize bins, mark pickup zones, keep heights consistent, and reduce reflective clutter near critical perception areas. Early deployments should be engineered for repeatability.

  • Constrain object pose variation: guides, trays, and consistent bin geometry beat “AI magic.”
  • Separate human lanes and robot lanes where possible during initial pilots.
  • Instrument the environment so you can correlate failures with lighting, congestion, or floor conditions.
3) Keep safety authority independent of “smart” behavior

In industrial settings, safety must be a separate layer with conservative defaults. A robot that is safe-but-slower can be improved; a robot that is fast-but-unsafe gets shut down.

  • Hard limits on speed/force near humans; predictable stop behaviors and clear re-start procedures.
  • Auditable logs for safety events and near-misses; treat safety reviews like release gates.
  • Site-specific hazard analysis as a repeatable template, not a one-off negotiation.
4) Operate like a fleet business: staged updates, telemetry, service loops

Scaling humanoids is not only robotics—it’s operations. You need remote health telemetry, staged rollouts, rollback capability, and service logistics. Without that, every site becomes a fragile science project.

  • Telemetry: actuator temps, sensor health, battery degradation, drift detection, charging success.
  • Rollout discipline: canary robot → small cohort → fleet; fast rollback if anomalies appear.
  • Service plan: spare parts, repair time targets, and a clear process for field triage.

FAQ

Why a humanoid instead of specialized factory automation?

Because the world is already built for people. A humanoid can, in theory, use existing aisles, workstations, carts, bins, and tools without forcing expensive retooling. The value is “compatibility” and flexibility—if reliability and safety are solved well enough to beat the cost of dedicated automation.

Does Helix “power” FIGURE 01?

Figure publicly introduced Helix in 2025 as a generalist Vision-Language-Action model for humanoid control, emphasizing unified perception+language+control and onboard execution. That’s a company-level direction. Whether FIGURE 01 specifically ran Helix in the field is not consistently documented publicly, so treat it as “roadmap signal,” not a guaranteed FIGURE 01 configuration.

What would make FIGURE 01 “proven” to serious operators?

Boring numbers: hours run in production-like conditions, MTBI, mean time to recover, intervention taxonomy, safe-stop rate, throughput distributions (not averages), and service metrics (time to repair, spare parts consumption). When those are public and stable across sites, the category stops being speculative.

What’s the clearest near-term use case?

Station-based parts handling and simple machine-tending steps: consistent bins, consistent geometry, defined zones, and high repetition. These tasks generate clean data and measurable ROI, which accelerates the learning loop while keeping the safety case manageable.

Sources used for this page (primary first)


Also in Robotics

NEO
NEO

Phoenix
Phoenix

Apollo
Apollo

Subscribe