MuxGel: Simultaneous Dual-Modal Visuo-Tactile Sensing via Spatially Multiplexing and Deep Reconstruction
MuxGel is a novel plug-and-play visuo-tactile gel-pad design that spatially multiplexes visual and tactile sensing, enabling the simultaneous recovery of vision and tactile signals for robotic manipulation.
Clone the repo along with its submodules:
git clone --recursive https://github.com/zhixian/MuxGel.git
cd MuxGelWe recommend using miniforge for dependency management:
# Create the environment from the provided file
conda env create -f conda_environment.yaml
# Activate the environment
conda activate muxgelTo generate your own datasets using the mujoco_scanned_objects library, fetch the 3D assets (~2 GB):
git submodule update --init external/mujoco_scanned_objectsTo train in simulation, you must first set up the Scene Backgrounds, then obtain the Object Patches via one of two methods.
For scene backgrounds, you will need to download the following from Hugging Face and extract them into the data/ folder:
- File:
indoorCVPRBlur_320_240.tar.xz - Note: This dataset was processed using a disk defocus blur, originally from A. Quattoni and A. Torralba, “Recognizing indoor scenes,” CVPR 2009.
You can either generate these patches manually or download our pre-processed version.
-
Option A: Manual Generation
Run the following scripts in order:
- Generate Data:
python scripts/datasetGeneration/mujoco_imageGenerate.py - Clean Folders:
python scripts/datasetGeneration/mujoco_folder_clean.py - Resize Data:
python scripts/datasetGeneration/dataResize.py(Downsamples to 320x240) The results will be stored indata/mujoco_patch_output_320_240.
- Generate Data:
-
Option B: Direct Download
Download and extract the following into the
data/folder:- File:
mujoco_patch_output_320_240.tar.xzfrom Hugging Face.
- File:
To train with real-world data, you will need the calibration assets:
- Download
calibration_data.zipfrom Hugging Face and unzip it into thedata/folder.
The trained model weights for all six architectures (SI, DI-AbsT, DI-ResT for both Simulation and Real-world) are hosted on Hugging Face.
| Repository | Link |
|---|---|
| MuxGel Weights | huggingface.co/datasets/zhixianhu/muxgel |
We provide a helper script to fetch the necessary checkpoints (approx. 636 MB total) directly into your project root.
-
Install requirements:
pip install huggingface_hub
-
Run the download script:
python scripts/download_weights.py
Training scripts are located in scripts/train/. We use the following acronyms for configurations:
| Acronym | Meaning | Acronym | Meaning |
|---|---|---|---|
| si | Single-Input | abst | Absolute Tactile |
| di | Dual-Input | rest | Residual Tactile |
Example run (Dual-Input Residual-Tactile model):
python scripts/train/train_real_di_rest.py --wandbTo run the real-time inference and visualization:
python scripts/test/realtime_vis.pyThis project adapts and modifies several excellent open-source repositories:
- GelSight Mini & GS Robotics: Basis for sensor drivers and integration.
- Taxim: Our tactile simulation is built upon Taxim for example-based rendering.
- TacEx: Source of calibration files for tactile simulation.
- MuJoCo Scanned Objects: Simulation 3D model assets, originally from the Google Scanned Objects collection.
- Indoor CVPR Dataset: Visual backgrounds, based on Quattoni & Torralba, "Recognizing indoor scenes," CVPR 2009.
If you find this work helpful, welcome to cite our paper:
@article{hu2026muxgel,
title={MuxGel: Simultaneous Dual-Modal Visuo-Tactile Sensing via Spatially Multiplexing and Deep Reconstruction},
author={Hu, Zhixian and Xu, Zhengtong and Athar, Sheeraz and Wachs, Juan and She, Yu},
journal={arXiv preprint arXiv:2603.09761},
year={2026}
}