Why Kitchens Are Exceptionally Hard
The kitchen environment combines nearly every open challenge in robot manipulation into a single space. Deformable objects (dough, vegetables, proteins) require models that generalize across continuous shape deformation. Liquids introduce fluid dynamics that current planners cannot reliably predict for pouring tasks. Heat creates safety constraints that limit acceptable failure modes -- a robot that drops a pan of boiling water is a different category of failure than one that drops a dry object. Object variety exceeds 10,000 SKUs in a fully-stocked commercial kitchen, compared to the 50-200 object categories where current policies generalize reliably. And human co-presence requires OSHA-compliant force limits and behavioral predictability that add layers of constraint on top of already-hard manipulation problems.
None of this means kitchen automation is impossible -- it means that successful deployments in 2025 are task-specific, not general. The robots that are actually working are solving carefully bounded subproblems.
Task Decomposition: Breaking Cooking into Solvable Subproblems
The key insight for building any kitchen robot system is that "cooking" is not a single task -- it is a sequence of subtasks with very different difficulty levels. Treating each subtask independently allows you to match the right hardware, control strategy, and training data investment to each challenge level.
The Subtask Pipeline for a Sandwich Assembly Task
Consider a concrete example: assembling a ham and cheese sandwich. This decomposes into approximately 12 subtasks:
- Retrieve bread package (L2 -- rigid grasp, known shelf position)
- Open bread package (L4 -- deformable package, variable seal)
- Extract two slices (L3 -- deformable object, thin sheet grasping)
- Place slices on surface (L1 -- release, known target)
- Retrieve ham container (L2 -- rigid grasp)
- Open container (L3 -- lid removal, variable tightness)
- Grasp and place ham slices (L3 -- deformable, sticky, thin)
- Retrieve cheese (L2)
- Slice or place cheese (L3-L5 depending on whether pre-sliced)
- Apply condiments (L4 -- squeeze bottle control, variable viscosity)
- Stack and align (L2 -- compliant placement)
- Cut sandwich (L5 -- knife manipulation, deformable target)
The practical strategy: automate L1-L3 subtasks first (steps 1, 4, 5, 8, 11), use pre-processed ingredients to eliminate L4-L5 subtasks (pre-sliced bread, pre-sliced cheese, squeeze bottles instead of knives), and keep humans in the loop for the hardest steps until policies mature. This approach lets you deploy a partially-automated system that handles 60-70% of the workflow and is immediately useful, rather than waiting for a fully autonomous solution.
Training Data Requirements by Subtask
| Subtask Category | Demo Count | Collection Time | Expected Success Rate |
|---|---|---|---|
| Rigid object pick-place (L1-L2) | 100-300 | 2-4 hours | 85-95% |
| Container manipulation (L3) | 300-800 | 8-16 hours | 70-85% |
| Deformable grasping (L3-L4) | 500-2,000 | 1-3 days | 60-80% |
| Pouring/dispensing (L4) | 500-1,500 | 1-2 days | 70-85% |
| Tool use -- knife/spatula (L5) | 2,000-5,000+ | 1-2 weeks | 40-60% (research stage) |
Sensor Requirements for Kitchen Manipulation
Kitchen tasks demand more sensing modalities than typical tabletop manipulation. Here is what each sensor contributes and when it is necessary:
Force/Torque Sensing: Essential for Compliant Grasping
A 6-axis force/torque sensor at the wrist is critical for kitchen tasks involving fragile, deformable, or variable-stiffness objects. Without force feedback, the robot has no way to distinguish between a firm tomato and a soft one -- and the grasping force that works for one will crush the other.
- ATI Nano17 (6-axis, 0.012N resolution, ~$5,000): Research gold standard. Excellent sensitivity for egg handling and delicate produce. Overkill for most commercial deployments.
- Robotous RFT40 (6-axis, 0.1N resolution, ~$800): Good balance of cost and performance. Adequate for container handling, bottle grasping, and most kitchen objects.
- Integrated joint torque (Franka, OpenArm): Both Franka Research 3 and the OpenArm have joint-level torque sensing that can infer end-effector forces. Less accurate than a dedicated F/T sensor but sufficient for many tasks and requires no additional hardware.
Tactile Sensing: Grip Confidence and Slip Detection
Tactile sensors on gripper fingertips provide contact-level information that cameras cannot see (the object is occluded by the gripper during grasping). The two leading options:
- GelSight/DIGIT (vision-based tactile, ~$400/sensor): Provides a high-resolution tactile image (deformation field) at 30Hz. Excellent for detecting slip, estimating contact normal, and measuring object hardness. Integration: mount on parallel jaw gripper fingertips, feed tactile images to the policy alongside camera observations.
- Paxini tactile sensor arrays: SVRC stocks Paxini sensors for integration with OpenArm and other platforms. The distributed array format (multiple sensing points across the fingertip rather than a single image) is well-suited for grasping tasks where you need to know if the entire grasp is stable, not just the local deformation pattern.
Thermal Sensing
For tasks involving hot objects (pan handling, oven loading), a FLIR Lepton thermal camera module (~$200, 160x120 resolution, 8Hz) mounted near the wrist camera provides temperature awareness. This is a safety requirement, not a nice-to-have -- the policy needs to know whether a surface is 25C or 250C before making contact.
Current Successful Deployments
- Miso Robotics Flippy ($30K): Operates at commercial fry stations, flipping burgers and managing fryer baskets at approximately 2 orders/minute. Succeeds because the task is spatially constrained (known tray positions, fixed grill geometry), the object deformation is predictable (burger patties deform in known ways), and it operates at an isolated station without general kitchen navigation.
- Nala Robotics (coffee barista, airports ~$80K): Automates coffee and beverage preparation in airport terminal kiosks. Success factors: structured environment with custom-designed equipment, high-repeatability cartridge-based ingredient handling, and tolerance for 30+ second preparation times. Found in Pittsburgh (PIT), Dallas (DFW), and several others as of 2025.
- Conveyor Sushi Systems: Fully automated rice-and-topping placement systems used in chains like Genki Sushi. Technically effective but require purpose-built conveyor infrastructure and do not involve general manipulation -- ingredients are dispensed from cartridges onto moving platforms.
The common thread: every successful kitchen deployment has reduced the environment to a solvable problem by constraining the workspace, limiting object variety, and designing custom equipment around the robot's capabilities rather than trying to make the robot adapt to arbitrary kitchen equipment.
Manipulation Task Difficulty Scale
A practical 5-level difficulty scale for kitchen manipulation, calibrated against current robot capability:
| Level | Example Task | Key Challenge | 2025 State of Art |
|---|---|---|---|
| L1 | Coffee pod insertion | Constrained fit, known object | Solved -- commercial products ship |
| L2 | Cup placement on tray | Rigid object, flat surface | Solved in structured environments |
| L3 | Flipping a burger | Spatula dynamics, partial occlusion | Commercial (Flippy) -- fixed setup only |
| L4 | Pouring liquid | Fluid dynamics, variable fill level | Research -- 70-85% success in lab |
| L5 | Knife skills (dicing) | Deformable object, safety constraints | Research only -- not close to deployment |
Key Engineering Challenges in Detail
- Pan and Pot Handling: A full commercial saute pan with contents weighs 2-4kg at 0.5-0.7m reach. At this moment arm, you need 5kg+ payload at rated reach -- this eliminates most collaborative robot arms (UR5e is rated 5kg but at 850mm reach struggles with pan dynamics). The Fanuc CR-7iA/L and Yaskawa HC10 are the minimum specs for reliable pan manipulation. Beyond payload, thermal considerations require end-effector insulation rated to 200C for direct pan contact.
- Pouring Accuracy: Controlled liquid pouring requires coordinating wrist rotation rate with fluid dynamics to hit a target volume within +/-10ml. Current approaches use pre-calibrated pour curves (rotation angle to volume for specific container fill levels) rather than real-time fluid simulation. Works for standard recipes but fails on partially-used containers. Adding a scale under the target container provides closed-loop volume feedback that improves accuracy to +/-3ml.
- Egg and Delicate Object Compliance: Egg shells fail at approximately 50-75N of compressive force. Safe egg handling requires force control with a 10-20N ceiling -- achievable with modern impedance-controlled arms like Franka Research 3 or OpenArm with joint torque sensing, but requires explicit force monitoring in the controller loop, not just position control with gentle motion profiles.
Failure Analysis: Where Kitchen Robots Break Down
After 6 months of kitchen manipulation data collection at SVRC's Mountain View lab, we have cataloged the failure modes that account for the majority of task failures. Understanding these is essential for prioritizing engineering effort:
| Failure Mode | Frequency | Root Cause | Mitigation |
|---|---|---|---|
| Grasp slip on wet/oily objects | ~25% of failures | Friction coefficient changes with surface moisture | Textured gripper pads; increase grasp force 30% for wet objects; tactile slip detection |
| Object deformation during grasp | ~20% of failures | Policy trained on rigid objects applied to deformables | Separate policies for rigid vs. deformable; force-limited grasping with 10N cap for soft items |
| Occlusion during multi-step tasks | ~18% of failures | Robot arm blocks camera view of workspace during manipulation | Wrist-mounted camera is essential; 3-camera setup minimum |
| Pouring overshoot/undershoot | ~15% of failures | No closed-loop volume feedback | Scale under target container; pre-calibrated pour curves per container type |
| Collision with kitchen fixtures | ~12% of failures | Cluttered workspace geometry not fully perceived | Depth camera for collision avoidance; workspace decluttering protocol between tasks |
Cost Breakdown: What a Kitchen Robot System Actually Costs
Teams frequently underestimate the total cost of a kitchen manipulation system. Here is a realistic breakdown for a single-station system capable of L1-L3 tasks:
| Component | Option A (Budget) | Option B (Performance) |
|---|---|---|
| Robot arm | OpenArm ($4,500) | Franka Research 3 ($22,000) |
| Gripper + tactile | Parallel jaw + GelSight ($800) | Robotiq 2F-85 + Paxini ($3,500) |
| Cameras (3x) | USB ELP modules ($90) | Intel RealSense D435i ($500) |
| F/T sensor | Joint torque (included) | ATI Nano17 ($5,000) |
| Compute (edge) | Jetson Orin Nano ($249) | Jetson AGX Orin ($1,999) |
| Kitchen station + mounts | $500 | $2,000 |
| Data collection (500 demos via SVRC) | $2,500 (pilot) | $8,000 (campaign) |
| Total | ~$8,600 | ~$43,000 |
The budget option using OpenArm is a fully functional kitchen manipulation research platform. Its 1.5kg payload limits it to light objects (cups, utensils, small ingredients), but for data collection and policy development on L1-L3 tasks, it is sufficient. Teams needing heavier payload for pan handling and L4+ tasks should invest in the performance option.
A Practical Recipe for Getting Started
If you want to build a kitchen robot project in 2025-2026, here is the recommended path:
- Week 1-2: Define your task narrowly. Pick one specific food preparation task at L2-L3. "Make a sandwich" is too broad. "Place pre-sliced ingredients on bread in sequence" is the right granularity.
- Week 2-3: Set up hardware. Lease an OpenArm or Mobile ALOHA through SVRC leasing. Install the 3-camera setup. Verify the teleop loop works end-to-end.
- Week 3-5: Collect 200-500 demonstrations. Use SVRC's data services for operator support or train your own team. Focus on consistent, successful episodes. Filter failures before training.
- Week 5-6: Train and evaluate. Train an ACT policy on your dataset. Evaluate on held-out object positions and lighting conditions. Expect 60-80% success rate on first iteration for L2-L3 tasks.
- Week 6-8: Data flywheel. Analyze failure modes. Collect 50-100 targeted demonstrations addressing the top 3 failure categories. Retrain. Iterate until you hit your success rate target.
SVRC Kitchen Task Data Collection
SVRC runs a dedicated kitchen manipulation data collection program at our Mountain View lab, with a purpose-built kitchen station including commercial-grade equipment, load cells on all surfaces, and a multi-camera capture system. We're currently collecting demonstrations for L2-L4 tasks including beverage preparation, ingredient handling, and assembly tasks (sandwich construction, plated dish assembly).
Teams building kitchen robots can access this dataset or commission custom collection through our data services program. For hardware and end-effector recommendations for kitchen manipulation, see our solutions page.
Related Reading
- Getting Started with Teleoperation -- how to set up data collection
- ACT Policy Explained -- the algorithm that powers most kitchen manipulation research
- Common Mistakes in Imitation Learning -- avoid these when training kitchen policies
- OpenArm vs Franka -- which arm to choose for kitchen tasks
- SVRC Data Services -- commission kitchen manipulation demonstrations