/Genesis

A generative world for general-purpose robotics & embodied AI learning.

A generative and simulated physical realm for general-purpose embodied-AI learning.


What is Genesis?

Genesis is a generative and differentiable physical world for general-purpose robot learning, providing a unified simulation platform supporting diverse range of materials, allowing simulating vast range of robotic tasks, while being fully differentiable.

Genesis is also a next-gen simulation infrastructure that natively supports generative simulation: a future paradigm combining generative AI and physically-grounded simulation, aiming for unlocking infinite and diverse data for robotic agents to learn vast range of skills across diverse environments like never before.

Genesis is still under development, and will be made publicly available soon.

What can Genesis do?

Genesis differs from prior simulation platforms with a number of distinct key features:

✅⚡⚡ Unprecedented lightning fast simulation speed (Million+ FPS)⚡⚡

I can't help updating this here secretly: after a ton of engineering effort streamlining our underlying physics engine, Genesis now can potentially produce over millions of simulation steps per second for rigid body dynamics on a single RTX4090, way faster than any existing simulators. We are getting closer to a public release soon!

✅ 👕 Support for diverse range of materials and their coupling

Genesis supports physics simulation of a wide range of materials encountered in humans and (future) robots' daily life, and coupling (interaction) between these objects, including:

  • 🚪 Rigid and articulated bodies
  • 💦 Liquid: Newtonian, non-Newtonian, viscosity, surface tension, etc.
  • 💨 Gaseous phenomenon: air flow, heat flow, etc.
  • 🥟 Deformable objects: elastic, plastic, elasto-plastic
  • 👕 Thin-shell objects: 🪱 ropes, 👖 cloths, 📄 papers, 🃏 cards, etc.
  • ⏳ Granular objects: sand, beans, etc.

✅ 🤖 Support for various forms of robots

Genesis supports simulating of vast range of robots, including

  • 🦾 Robot arm
  • 🦿 Legged robot
  • ✍️ Dexterous hand
  • 🖲️ Mobile robot
  • 🚁 Drone
  • 🦎 Soft robot

Note that Genesis is the first-ever platform providing comprehensive support for soft muscles and soft robot, as well as their interaction with rigid robots. Genesis also ships with a URDF-like soft-robot configuration system.

✅ 🚀 Support for a wide spectrum of physics backend solvers

Various physics-based solvers tailored to different materials and needs have been developed in the past decades. Some prioritize high simulation fidelity, while others favor performance, albeit sometimes sacrificing accuracy.

Genesis, in contrast to existing simulation platforms, natively supports a wide range of different physics solvers. Users are able to effortlessly toggle between solvers, depending on their specific requirements.

We have built a unified physics engine from the ground up, supporting solvers include:

  • Material Point Method (MPM)
  • Finite Element Method (FEM)
  • Position Based Dynamics (PBD)
  • Smoothed-Particle Hydrodynamics (SPH)
  • Articulated Body Algorithm (ABA)-based Rigid Body Dynamics

We also provide contact resolving via multiple methods:

  • Shape primitives
  • Convex mesh
  • SDF-based non-convex contact
  • Incremental Potential Contact (IPC)

✅ 🤏 Physically-accurate tactile sensor simulation, applicable to diverse material types

Genesis is the first platform that integrates physics-based simulation of GelSight-type tactile sensors, providing dense state-based and RGB-based tactile feedback simulation when handling diverse range of materials.

✅ 📸 Ultra-fast ray-tracing based renderer

Genesis provides both rasterization-based and ray-tracing-based rendering pipelines. Our ray tracer is ultra fast, providing real-time photo-realistic rendering at RL-sufficient resolution. Notably, users can effortlessly switch between different rendering backend with one line of code.

✅ 💥 Massive parallel, GPU-accelerated and fully differentiable

Genesis supports massive parallelization over GPUs, and supports efficient gradient checkpointing for differentiable simulation. Genesis is internally powered by Taichi, but users are shielded from any potential debugging intricacies associated with Taichi. Genesis implements a custom Tensor system (Genesis.Tensor) seamlessly integrated with PyTorch. This integration guarantees that the Genesis simulation pipeline mirrors the operation of a standard PyTorch neural network in both functionality and familiarity. Provided that input variables are Genesis.Tensor, users are able to compute custom loss using simulation outputs, and one simple call of loss.backward() will trigger gradient flow all the way back through time. back to input variables, and optionally back to upstream torch-based policy networks, allowing effortless gradient-based policy optimization and gradient-accelerated RL policy learning.

✅ 👶 Optimized ease of use

The rule of thumb we kept in mind during the development is to make Genesis as user-friendly and intuitive as possible, ensuring a seamless experience even for those new to the realm of simulations. Genesis is fully embedded in Python. We provide a unified interface for loading various entities including URDF, MJCF, various forms of meshes, etc. With one single line of code, users are able to switch between different physics backend, or switch from rasterization to ray-tracing for rendering photorealistic visuals, or define physical properties of loaded objects.

✅ 🌌 Most importantly, native support for Generative Simulation

Genesis, while being a powerful simulation infrastructure, natively embraces the upcoming paradigm of Generative Simulation. Powered by generative AIs and physics-based simulations, generative simulation is a paradigm that automates generation of diverse environments, robots, tasks, skills, training supervisions (e.g. reward functions), thereby generating infinite and diverse training data and scaling up diverse skill learning for embodied AI agents, in a fully automated manner.

Genesis provides a set of APIs for generating diverse components needed for embodied AI learning, ranging from interactable and actionable environments, to different tasks, robots, language descriptions, and ultimately vast skills.

Users will be able to do the following:

import genesis as gs
# generate environments
gs.generate('A robotic arm operating in a typical kitchen environment.')

# generate robot
gs.generate('A 6-DoF robotic arm with a mobile base, a camera attached to its wrist, and a tactile sensor on its gripper.')
gs.generate('A bear-like soft robot walking in the sand.')

# generate skills
gs.generate('A Franka arm tossing trashes into a trashcan')
gs.generate('A UR-5 arm bends a noodle into a U-shape')

# or, let it generate on its own
gs.generate('A random robot learning a random but meaningful skill')
# or simply let it surprise you
gs.generate()

Stay tuned :P

Genesis is a collaborative effort between multiple universities and industrial partners, and is still under active development. The Genesis team is diligently working towards an alpha release soon. Stay tuned for updates!