You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Because I saw that all the features you have extracted in your code are a file, which is to directly extract all the features of the video into an HDF5 file, without distinguishing between the training set and the test set. Hope you can extract valuable events to answer the question.
The text was updated successfully, but these errors were encountered:
It is easier than you think. The features are extracted for the whole video regardless of the dataset part. During training, we can simply trim the feature stack according to the start and end timestamps. During test, you can download the predictions of the proposal generator from BAFCG.
Note, the code might not be adapted for the train set; only for train and two validation sets.
How to distinguish? Well, we have separate files for audio, speech, and visual features. These are uni-modal.
Because I saw that all the features you have extracted in your code are a file, which is to directly extract all the features of the video into an HDF5 file, without distinguishing between the training set and the test set. Hope you can extract valuable events to answer the question.
The text was updated successfully, but these errors were encountered: