I'm currently trying to reproduce the inference in arbitrary videos of the HMDB51 dataset by using the pretrained weights.
So far, I have:
- Compiled OpenCV with the contrib module and cuda support
- Successfully compiled caffe (caffe-tpp-net) with CUDA and python support
- Set up the PYTHONPATH environment variable so python2.7 may find the compiled modules
- Downloaded the pretrained weights with
get_init_models.sh
and get_kinetics_pretraining_models.sh
Since README.md doesn't mention how to perform inference, I read README_old.md and found that there is a script tools/eval_net.py
for that purpose. So I ran:
python tools/eval_net.py hmdb51 1 rgb /var/datasets/hmdb51/ models/hmdb51/flow_tpp_delete_dropout_deploy.prototxt ./init_models/hmdb51_split_1_tsn_flow_reference_bn_inception.caffemodel
But, after a ton of messages, that gave me the following error:
Traceback (most recent call last):
File "tools/eval_net.py", line 125, in
video_scores = map(eval_video, eval_video_list)
File "tools/eval_net.py", line 69, in eval_video
video_frame_path = f_info[0][vid]
KeyError: '20060723sfjffbumblebeesuitman_run_f_cm_np2_ri_med_1'
The path to the dataset is correct, and I have unpacked all rar files from the HMDB dataset. The README_old.md file mentions a script scripts/extract_optical_flow.sh
for preprocessing the video files, but in the master branch, this script doesn't exist.
So, are there any further steps necessary for reproducing inference?