xqpinitial / RPCA_actional_recognition

In human action recognition, visual feature descriptors are of vital importance. based on RPCA,we get sparse component and low-rank of each video. Then we use the two-stream architecture.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

RPCA_actional_recognition

In human action recognition, visual feature descriptors are of vital importance. We propose a novel thought of getting the low-rank features and sparse features from the video. This ideal is based on the Robust Principal Component Analysis (RPCA), and obtained through sparse component and low-rank on the raw image pixels of each video. Then we use the two-stream architecture for our robust PCA segmentation descriptors with fine-tuning.

1、using caffe https://github.com/yjxiong/caffe/tree/action_recog

2、git clone https://github.com/xqpinitial/Robust-PCA-RPCA
the v_BoxingPunchingBag_g01_c02.rar is demo of rpca
unrar PROPACK.rar and run video_rpca_main.m to get rpca features

3、train A_hat_C3(low-rank feature) and E_hat_C3(sparse feature)
using model_channels_change.py to revise your model for train

技术交流QQ群action_recognition:333571956

About

In human action recognition, visual feature descriptors are of vital importance. based on RPCA,we get sparse component and low-rank of each video. Then we use the two-stream architecture.


Languages

Language:Python 95.7%Language:Shell 4.3%