Skip to content

newdexterity/TMC-ViT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Python Version TensorFlow Version MIT License


Temporal Multi-Channel Vision Transformer (TMC-ViT)

This repository implements a Transformer-based model called Temporal Multi-Channel Vision Transformer (TMC-ViT). The TMC-ViT was developed to adapt the Vision Transformer model proposed by Dosovitskiy et al.[1] for processing multi-channel temporal signals as input. In this example, we will predict 18 gestures from the Ninapro DB05 Database. This has also been implemented in a Google Colab project.

Loading the data

The input data must already be divided into training and test sets, with 200 ms samples. Use one separate repetition for testing. More information on the data preprocessing can be found in [2].

References

[1] Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., & Houlsby, N. (2020). An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. 1–21. http://arxiv.org/abs/2010.11929
[2] R. V. Godoy et al., "Electromyography-Based, Robust Hand Motion Classification Employing Temporal Multi-Channel Vision Transformers," in IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10200-10207, Oct. 2022, doi: 10.1109/LRA.2022.3192623.

About

This repository implements a Transformer-based model called Temporal Multi-Channel Vision Transformer (TMC-ViT)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages