python动作识别库_Pytorch中用于动作识别和目标检测的视频平台

ViP是一个用于视频中识别和检测的深度学习平台,支持C3D和I3D等模型。提供了配置好的数据集,可以方便地训练和评估模型。通过添加新模型和数据集类,可以扩展平台的功能。
摘要由CSDN通过智能技术生成

A platform for quick and easy development of deep learning networks for recognition and detection in videos. Includes popular models like C3D and SSD.

Check out our wiki!

Implemented Models and their performance

Recognition

Model Architecture

Dataset

ViP Accuracy (%)

I3D

HMDB51 (Split 1)

72.75

C3D

HMDB51 (Split 1)

50.14 ± 0.777

C3D

UCF101 (Split 1)

80.40 ± 0.399

Object Detection

Model Architecture

Dataset

ViP Accuracy (%)

SSD300

VOC2007

76.58

Video Object Grounding

Model Architecture

Dataset

ViP Accuracy (%)

DVSA (+fw, obj)

YC2-BB (Validation)

30.09

fw: framewise weighting, obj: object interaction

Citation

Please cite ViP when releasing any work that used this platform: https://arxiv.org/abs/1910.02793

@article{ganesh2019vip,

title={ViP: Video Platform for PyTorch},

author={Ganesh, Madan Ravi and Hofesmann, Eric and Louis, Nathan and Corso, Jason},

journal={arXiv preprint arXiv:1910.02793},

year={2019}

}

Table of Contents

Configured Datasets

Dataset

Task(s)

Activity Recognition

Activity Recognition

Video Object Detection

Object Detection, Keypoints

Object Detection, Classification

Video Object Grounding

Video Saliency Prediction

Models

Model

Task(s)

Activity Recognition

Activity Recognition

Object Detection

Video Object Grounding

Requirements

Python 3.6

Cuda 9.0

(Suggested) Virtualenv

Installation

# Set up Python3 virtual environment

virtualenv -p python3.6 --no-site-packages vip

source vip/bin/activate

# Clone ViP repository

git clone https://github.com/MichiganCOG/ViP

cd ViP

# Install requirements and model weights

./install.sh

Quick Start

Run train.py and eval.py to train or test any implemented model. The parameters of every experiment is specified in its config.yaml file.

Use the --cfg_file command line argument to point to a different config yaml file. Additionally, all config parameters can be overriden with a command line argument.

Testing

Run eval.py with the argument --cfg_file pointing to the desired model config yaml file.

Ex: From the root directory of ViP, evaluate the action recognition network C3D on HMDB51

python eval.py --cfg_file models/c3d/config_test.yaml

Training

Run train.py with the argument --cfg_file pointing to the desired model config yaml file.

Ex: From the root directory of ViP, train the action recognition network C3D on HMDB51

python train.py --cfg_file models/c3d/config_train.yaml

Additional examples can be found on our wiki.

Development

New models and datasets can be added without needing to rewrite any training, evaluation, or data loading code.

Add a Model

To add a new model:

Create a new folder ViP/models/custom_model_name

Create a model class in ViP/models/custom_model_name/custom_model_name.py

Complete __init__, forward, and (optional) __load_pretrained_weights functions

Add PreprocessTrain and PreprocessEval classes within custom_model_name.py

Create config_train.yaml and config_test.yaml files for the new model

Examples of previously implemented models can be found here.

Additional information can be found on our wiki.

Add a Dataset

To add a new dataset:

Convert annotation data to our JSON format

The JSON skeleton templates can be found here

Existing scripts for datasets can be found here

Create a dataset class in ViP/datasets/custom_dataset_name.py.

Inherit DetectionDataset or RecognitionDataset from ViP/abstract_dataset.py

Complete __init__ and __getitem__ functions

Example skeleton dataset can be found here

Additional information can be found on our wiki.

FAQ

A detailed FAQ can be found on our wiki.

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值