Skip to main content

The Tactile MNIST benchmark for active tactile perception.

Project description

The Tactile MNIST Benchmark

TactileMNIST-v0
TactileMNIST-v0
Starstruck-v0
Starstruck-v0
Toolbox-v0
Toolbox-v0
ABCCenterOfMass-v0
ABCCenterOfMass-v0
TactileMNISTVolume-v0
TactileMNISTVolume-v0
ABCVolume-v0
ABCVolume-v0

Tactile MNIST is a benchmark for learning-based active perception algorithms. It introduces four simulated tactile perception tasks, ranging from classification and counting to pose and volume estimation. Each task comes with a unique set of challenges and, thus, Tactile MNIST requires adaptive algorithms and clever exploration strategies. The aim of Tactile MNIST is to provide an extensible framework for a fair comparison of active tactile perception methods.

Tactile MNIST implements each task as an ap_gym environment and is, thus, very straightforward to set up and use. In each task, the agent controls a single simulated GelSight Mini above a platform with some task-specific objects. The agent's objective is to make a prediction about some property of the objects it is exploring, such as their class, count, pose, or volume.

In addition to the simulated benchmark tasks, this package provides access to a large dataset of real tactile images collected from 3D printed MNIST digits and a couple of synthetic datasets.

Further details can be found on our project page, which also links to the paper.

Installation

Install Tactile MNIST via pip:

pip install tactile-mnist[OPTIONS]

where OPTIONS can be any number of the following (comma separated):

  • examples: installs dependencies for the examples.
  • jax (recommended) , torch, and jax-cpu: installs dependencies for the interactive Tactile MNIST environment with PyTorch, JAX, or JAX without CUDA support, respectively. Without any of those options, the interactive environment cannot be used but the static datasets will still work.

If you encounter problems during the installation or execution, check our troubleshooting section.

Contents

This package provides ap_gym environments for four simulated active tactile classification benchmark tasks and access to the Tactile MNIST datasets. The ap_gym environments can be used to train and evaluate agents on active tactile perception problems on simulated data and are further described in the Benchmark section. In the Tactile MNIST datasets, you find two datasets of 3D CAD models, MNIST 3D and Starstruck, and several datasets of simulated and real tactile images. This package provides an easy way of loading and working with these datasets, as further described in the Datasets section.

Simulated Active Tactile Perception Benchmark

This package provides ap_gym environments for six active tactile perception environments: TactileMNIST, Starstruck, Toolbox, ABCCenterOfMass, TactileMNISTVolume, and ABCVolume. In all environments, the agent must solve a perception problem by actively controlling a GelSight Mini tactile sensor in a simulated environment.

The TactileMNIST environment challenges the agent to find and classify a 3D MNIST model as quickly as possible. aside from finding the object, the main challenge in the TactileMNIST environment is to learn contour following strategies to efficiently classify it once found.

In the Starstruck environment, the agent must count the number of stars in a scene cluttered with other objects. Since all stars look the same, distinguishing stars from other objects is rather straightforward. Instead, the main challenge posed in this environment is to learn an effective search strategy to systematically cover as much space as possible.

The Toolbox environment challenges the agent to locate a wrench positioned randomly on a platform and estimate its precise 2D position and 1D orientation. Unlike the previous classification tasks, Toolbox is poses a regression problem that requires combining multiple touch observations to resolve ambiguities inherent in the wrench’s shape. For example, touching the handle may reveal lateral placement but not longitudinal position or orientation, making it critical for the agent to explore strategically and seek out one of the wrench’s ends to accurately determine its pose. Overall, the Toolbox tests the agent’s ability to both find and precisely localize an object through sequential tactile exploration.

In the ABCCenterOfMass environment, the agent must determine the exact 2D position of the center of mass of an object from the ABC dataset. Unlike in the Toolbox environment, where the object shape is known, in ABCCenterOfMass, the agent has to deal with a large variety of object shapes and learn thorough exploration policies. Due to the larger variety of object shapes compared to TactileMNISTVolume, we give the agent control over the sensor's rotation in ABCVolume, which adds another layer of complexity to the problem.

The TactileMNISTVolume environment poses another regression problem. Here, the agent must determine the exact volume of the 3D MNIST model it is given. Thus, unlike in the TactileMNIST environment, where a couple of touches might already be sufficient for classification, in TactileMNISTVolume, the agent has to make sure to explore the entire object.

Similar to the TactileMNISTVolume environment, the ABCVolume environment challenges the agent to determine the exact volume of an object. However, instead of 3D MNIST models, the objects in ABCVolume are randomly sampled from the ABC dataset, which contains a large variety of 3D CAD models. Similar to ABCCenterOfMass, we give the agent control over the sensor's rotation in ABCVolume.

A detailed description of the environments can be found here.

Datasets

Aside of the simulated benchmark tasks, this package provides access to two classes of static datasets: 3D mesh datasets and touch datasets. Below is an overview of the datasets provided in this package:

  • 3D Mesh Datasets:
    1. MNIST 3D: a dataset of 3D models generated from a high-resolution version of the MNIST dataset.
    2. Starstruck: a dataset in which the number of stars in a scene have to be counted (3 classes, 1 - 3 stars per scene).
    3. ABC Dataset: a variant of the ABC dataset, processed for the use with this benchmark suite.
  • Touch Datasets
    1. Synthetic Tactile MNIST: a dataset of synthetic tactile images generated from the MNIST 3D dataset with the Taxim simulator.
    2. Real Tactile MNIST: a dataset of real tactile images of 3D printed MNIST 3D digits collected with a Franka robot.
    3. Synthetic Tactile Starstruck: a dataset of synthetic tactile images generated from the Starstruck dataset with the Taxim simulator.

A detailed description of the datasets can be found in the Dataset documentation.

Troubleshooting

Torch-Scatter Undefined Symbol Error

If you are seeing errors as such

OSError: [...]/torch_scatter/_version_cpu.so: undefined symbol: _ZN3c1017RegisterOperatorsD1Ev

then there are CUDA version incompatibilities between torch_scatter, PyTorch, and nvcc.

If you are seeing errors as such

RuntimeError: Not compiled with CUDA support

then nvcc was not found when installing torch_scatter.

In both cases, follow the instructions in the official torch_scatter repository to install compatible PyTorch and torch_scatter versions.

License

The project is licensed under the MIT license.

Contributing

If you wish to contribute to this project, you are welcome to create a pull request. Please run the pre-commit hooks before submitting your pull request. To install the pre-commit hooks, run:

  1. Install pre-commit
  2. Install the Git hooks by running pre-commit install or, alternatively, run `pre-commit run --all-files manually.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tactile_mnist-0.12.0.tar.gz (42.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tactile_mnist-0.12.0-py3-none-any.whl (42.9 MB view details)

Uploaded Python 3

File details

Details for the file tactile_mnist-0.12.0.tar.gz.

File metadata

  • Download URL: tactile_mnist-0.12.0.tar.gz
  • Upload date:
  • Size: 42.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for tactile_mnist-0.12.0.tar.gz
Algorithm Hash digest
SHA256 bc82d8f140aea22cbbac11fa2d2907d761f5f3f9d67cbda80925a6e733f64773
MD5 35b153bab0afe61b02bf3cd4ee5900fc
BLAKE2b-256 af81ed772061110cd036e02d6f1780f7f6ee55e3035cbe35ca0200db5a73c503

See more details on using hashes here.

Provenance

The following attestation bundles were made for tactile_mnist-0.12.0.tar.gz:

Publisher: publish.yml on TimSchneider42/tactile-mnist

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file tactile_mnist-0.12.0-py3-none-any.whl.

File metadata

  • Download URL: tactile_mnist-0.12.0-py3-none-any.whl
  • Upload date:
  • Size: 42.9 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for tactile_mnist-0.12.0-py3-none-any.whl
Algorithm Hash digest
SHA256 76efd6d17b7507235ae37b2a1c7998f7e4d7dbd0492dd7e2bfe52d52a3916ae9
MD5 6cc68c260aacd46a67721d8465c4888d
BLAKE2b-256 b3203a5060ece10bfbbd070abe7c1500f5b2d2716d4695bbd770786719c07b93

See more details on using hashes here.

Provenance

The following attestation bundles were made for tactile_mnist-0.12.0-py3-none-any.whl:

Publisher: publish.yml on TimSchneider42/tactile-mnist

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page