TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Datasets/KTH

KTH

KTH Action dataset

VideosCustom (non-commercial, attribution)Introduced 2004-01-01

The efforts to create a non-trivial and publicly available dataset for action recognition was initiated at the KTH Royal Institute of Technology in 2004. The KTH dataset is one of the most standard datasets, which contains six actions: walk, jog, run, box, hand-wave, and hand clap. To account for performance nuance, each action is performed by 25 different individuals, and the setting is systematically altered for each action per actor. Setting variations include: outdoor (s1), outdoor with scale variation (s2), outdoor with different clothes (s3), and indoor (s4). These variations test the ability of each algorithm to identify actions independent of the background, appearance of the actors, and the scale of the actors.

Source: Review of Action Recognition and Detection Methods

Benchmarks

Action Recognition/16:9 AccuracyActivity Recognition/16:9 AccuracyVideo/FVDVideo/SSIMVideo/PSNRVideo/LPIPSVideo/CondVideo/TrainVideo/PredVideo/Params (M)Video/MSEVideo/DiversityVideo Prediction/FVDVideo Prediction/SSIMVideo Prediction/PSNRVideo Prediction/LPIPSVideo Prediction/CondVideo Prediction/TrainVideo Prediction/PredVideo Prediction/Params (M)Video Prediction/MSEVideo Prediction/Diversity

Related Benchmarks

KTH 64x64 cond10 pred30/Video/FVDKTH 64x64 cond10 pred30/Video Prediction/FVDKTH-TIPS2/Image Classification/Accuracy (%)

Statistics

Papers
279
Benchmarks
22

Links

Homepage

Tasks

Action RecognitionActivity RecognitionTemporal Action LocalizationVideoVideo Prediction