Skip to content

MMintLab/touch2touch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

touch2touch

This repository shows the main files for training a cross-modal tactile generation model using the VQ-VAE architecture and the main files to evaluate its performance.

Dataset

We use Touch2Touch Dataset.

Paper: https://www.arxiv.org/abs/2409.08269

Dataset: https://drive.google.com/drive/folders/15vWo5AWw9xVKE1wHbLhzm40ClPyRBYk5?usp=sharing

Checkpoints

Touch2Touch VQ-VAE Model Checkpoint: https://drive.google.com/file/d/10_HR54aKSUuF3hQPTY1zLOgYuBHMITch/view?usp=drive_link

Train VQVAE Model

cd scripts
python train_vq_vae.py --model_type VQ-VAE-small --device cuda:0 --data cross_GB --dataset new_partial --mod 4 --random_sensor --color_jitter --rotation --flipping

Evaluate VQVAE Model

python testing.py

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published