Robosuite vs Raw MuJoCo: Which Should You Use for Manipulation RL?

Almost every modern manipulation RL paper runs on MuJoCo physics, but only some use raw MuJoCo. Most instead reach for Robosuite, a task-and-asset framework that sits on top of MuJoCo and gives you Franka, Sawyer, UR5e, and other arms with standardized observations, controllers, and benchmark tasks. This guide explains which layer you actually need — and where Robosuite's opinionated defaults help versus hurt.

TL;DR

  • Robosuite is a higher-level manipulation framework built on MuJoCo. It ships robots, tasks, controllers, and gym-compatible environments.
  • Raw MuJoCo is just the physics engine and Python bindings. Tasks, controllers, and observations are yours to define.
  • If you want a benchmark result that is comparable to published work on LIBERO, RoboMimic, or MimicGen, use Robosuite. Those benchmarks are built on it.
  • If you are building a novel robot, novel controller, or non-standard observation space, raw MuJoCo is lighter and less opinionated.
  • Both are free and Apache 2.0 / MIT licensed. Both run on CPU or GPU (via MJX for raw MuJoCo).

What each project actually is

MuJoCo (Multi-Joint dynamics with Contact) is a physics engine originally created by Emo Todorov and acquired by Google DeepMind in 2021, at which point it was open-sourced under Apache 2.0. It is a C library with Python bindings that takes an MJCF (an XML description of a scene and its robots), simulates contact-rich rigid-body dynamics, and returns state. That is, roughly, everything it does. There are no task definitions, no rewards, no benchmark environments, and no robots shipped by default — those live in the separate MuJoCo Menagerie repository or in MJCF files you write yourself.

Robosuite is a Python framework originally developed at Stanford's Pair lab that provides manipulation-focused scaffolding on top of MuJoCo. Concretely, it ships pre-configured robot models (Franka Panda, Rethink Sawyer, Universal Robots UR5e, Kinova Gen3, and more), standardized controllers (operational space, joint position, joint velocity), a set of benchmark tasks (Lift, Stack, PickPlace, NutAssembly, Door, Wipe, ToolHang, and two-arm variants), and a Gymnasium-compatible environment API. The official site is the best starting point, and the repository at github.com/ARISE-Initiative/robosuite is actively maintained into 2026.

Where they share DNA and where they diverge

Everything that happens inside a Robosuite simulation is still MuJoCo under the hood — the physics, the contacts, the rendering, the MJCF scene format. Robosuite is not a fork of MuJoCo, it is a user of MuJoCo's Python bindings. That means any fix or improvement that lands in MuJoCo flows into Robosuite on its next upstream bump, and anything you could do in raw MuJoCo (custom controllers, unusual joint structures, deformables via composites) is still available from inside a Robosuite env if you subclass the right base classes.

What diverges is everything above the physics layer. Robosuite imposes a particular view of what a manipulation task looks like: a robot plus a task XML plus a controller plus an observation spec plus a reward function. That view is a good fit for 90% of published manipulation research, which is why the framework has the adoption it does. It is a worse fit for research that breaks those assumptions — mobile manipulation, whole-body humanoid control, and dexterous hand tasks with unusual sensor suites are all awkward in Robosuite but natural in raw MuJoCo.

Feature comparison

Dimension Robosuite Raw MuJoCo
LicenseMITApache 2.0
Physics engineMuJoCo (as a dependency)MuJoCo (directly)
Pre-built tasks13+ standardized manipulation tasksNone shipped by default
Pre-built robotsFranka, Sawyer, UR5e, Kinova, Baxter, IIWAVia Menagerie repo, manual import
ControllersOperational space, joint position, joint velocity, IKBYO
GPU accelerationCPU by default; MJX port exists but partialMJX (JAX-native GPU/TPU)
Sensor simRGB, depth, segmentation, proprioception, force-torqueWhatever you configure in MJCF
PhotorealismMuJoCo's PBR renderer; fine, not photorealSame renderer; can integrate Blender offline
Parallel envsGym-style vector env on CPUMJX scales to 4096+ on a GPU
ROS supportNone direct; bridge via mujoco_rosNone direct; same bridge available
Learning curveLow (hours to first training run)Moderate (days to first clean task)
Paper & docsRobosuite paper, docs at robosuite.aiMuJoCo docs, paper arXiv 2106.04134

Where Robosuite pays off

There are three places where Robosuite saves meaningful engineering time. The first is benchmark comparability. LIBERO, RoboMimic, and MimicGen — the three most cited manipulation imitation-learning and RL benchmarks of 2024-2026 — all build on Robosuite. If you want your result to sit alongside published numbers, building on Robosuite means you share task definitions, assets, and controllers with prior work, and the comparison is apples-to-apples. Reimplementing those benchmarks in raw MuJoCo is possible but introduces small divergences that your reviewers will notice.

The second is controllers. Implementing operational-space control well for a 7-DOF arm is a week of engineering; Robosuite gives you a working OSC controller out of the box, tuned against the robots it ships. If your policy outputs delta end-effector poses rather than joint torques, using Robosuite's controller is almost always the right move.

The third is VLA evaluation. OpenVLA, Octo, RT-X, and newer VLA families all publish eval harnesses that either use Robosuite directly or define tasks in Robosuite-compatible MJCF. If your end goal is to measure a VLA policy's success rate on standard suites, Robosuite is already on the critical path.

Where raw MuJoCo pays off

Raw MuJoCo is the right base when Robosuite's assumptions get in your way. Common scenarios: you are simulating a humanoid or mobile manipulator (Robosuite is arm-centric); you are designing a novel gripper or a multi-fingered hand with custom tendons (Robosuite's controller suite does not target this); you want MJX on GPU with thousands of parallel envs (the Robosuite MJX port is useful but less mature than raw MJX); you need full control over the observation schema for an unusual VLA or diffusion-policy architecture; or you are doing physics-centric research where the task scaffolding would just be overhead.

Raw MuJoCo also wins on installation minimalism. pip install mujoco gets you the engine, Python bindings, and a viewer, and you are done. There is no curriculum of robots, tasks, and controllers to learn before you can simulate anything. For single-file reproducible research, this is a real advantage.

Performance considerations

Both paths go through the same physics kernel, so per-step throughput on a single env is comparable. Where they diverge is in parallelism. Raw MuJoCo gives you access to MJX, a JAX-native rewrite that runs the physics on GPU or TPU and scales cleanly to thousands of parallel envs. Robosuite's primary path is CPU vectorization via Gymnasium's vector env API, which tops out below what MJX can do on modern hardware, although the Robosuite team has been steadily porting tasks to MJX and by 2026 the common manipulation benchmarks all have an MJX variant. If your plan is massive-batch PPO on a single GPU, raw MuJoCo (through MJX) is the lower-risk path today; if your plan is moderate-batch RL or imitation learning with Gymnasium-compatible envs, Robosuite is fine.

Asset and ecosystem considerations

Robosuite's asset library is curated and consistent: every shipped robot has verified inertias, collision meshes, controller tunings, and example tasks. This is invaluable when starting out. MuJoCo Menagerie is the closest analog on the raw MuJoCo side — a DeepMind-maintained repository of high-quality robot models (Unitree, Boston Dynamics, Franka, Shadow Hand, ARX, and dozens more) in MJCF format. Menagerie assets tend to be higher-fidelity than Robosuite's (better collision meshes, more accurate inertias), but they ship just the model, with no tasks, controllers, or reward functions attached.

A common pattern in 2026 labs is to use Menagerie for the robot model, raw MuJoCo for the physics, and Robosuite-style task scaffolding (either from Robosuite itself or a custom lightweight clone) for the RL harness. This gives you high-fidelity models, predictable physics, and a familiar RL API without being locked into Robosuite's shipped arms.

VLA evaluation: the decisive factor

For most teams in 2026, the decisive factor is whether their policy will be compared against published VLA numbers. If the answer is yes, the path of least resistance is Robosuite plus LIBERO or MimicGen, because that is where the published numbers live. If you are training in raw MuJoCo and then re-running on Robosuite benchmarks for evaluation, expect a few days of integration work per benchmark suite. It is doable, but it is the kind of work that tends to expand to fill whatever time the deadline allows.

When to use Robosuite

When to use raw MuJoCo

Verdict

Robosuite is raw MuJoCo's opinionated face for manipulation research. If your project fits its mold — a tabletop arm, a handful of objects, a standardized controller, a benchmark comparison — use Robosuite and save yourself weeks of plumbing. If your project lives outside that mold, build on raw MuJoCo, borrow assets from Menagerie, and keep Robosuite as the eval harness for the specific benchmarks you need to publish against.

If you want a manipulation RL stack delivered with calibrated dynamics, VLA eval harnesses, and reproducible task definitions, the SVRC Services team builds them both ways and hands them off. See also our ACT vs Diffusion Policy guide for the policy side of the same question.

Related comparisons