Skip to content

2019 term project for Machine Learning Practical course at the University of Edinburgh

Notifications You must be signed in to change notification settings

Jeff-Wu97/Multimodal_Video_Recognition_Co-Trainning_Model

Repository files navigation

Multimodal_Video_Recognition_Co-Trainning_Model

This is a model for MLP course project. The project focuses on the recognition of hand gesture in the video, which is achieved by two co-trained I3D networks. Our model takes two different modalities of frames from the video, RGB and optical, as input. The co-training network is supposed to have the better performance on hand gesture recognition than a single-branch network. In this project, the spatiotemporal semantic alignment optimization method is applied to optimized the video recognition system. The designed model can achieve accuracy more than 99.3% on the EgoGesture dataset,which contains different subjects of hand gesture on various scenes.

Environment

Tensorflow-gpu-1.5

Tensorflow_probability-0.7

Sonnet-1.25

Opencv-3.4.2

Imageio

About

2019 term project for Machine Learning Practical course at the University of Edinburgh

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages