Can Deep Learning Recognize Subtle Human Activities?

Deep Learning has driven recent and exciting progress in computer vision, instilling the belief that these algorithms could solve any visual task. Yet, datasets commonly used to train and test computer vision algorithms have pervasive confounding factors. Such biases make it difficult to truly estimate the performance of those algorithms and how well computer vision models can extrapolate outside the distribution in which they were trained. In this work, we propose a new action classification challenge that is performed well by humans, but poorly by state-of-the-art Deep Learning models. As a proof-of-principle, we consider three exemplary tasks: drinking, reading, and sitting. The best accuracies reached using state-of-the-art computer vision models were 61.7%, 62.8%, and 76.8%, respectively, while human participants scored above 90% accuracy on the three tasks. We propose a rigorous method to reduce confounds when creating datasets, and when comparing human versus computer vision performance. Source code and datasets are publicly available.

Medienart:

Artikel

Erscheinungsjahr:

2020

Erschienen:

2020

Enthalten in:

Zur Gesamtaufnahme - volume:2020

Enthalten in:

Conference on Computer Vision and Pattern Recognition Workshops. IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Workshops - 2020(2020) vom: 01. Juni

Sprache:

Englisch

Beteiligte Personen:

Jacquot, Vincent [VerfasserIn]
Ying, Zhuofan [VerfasserIn]
Kreiman, Gabriel [VerfasserIn]

Themen:

Journal Article

Anmerkungen:

Date Revised 22.08.2022

published: Print-Electronic

Citation Status PubMed-not-MEDLINE

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

NLM328358916