About this Digital Document
In this paper, we attack the problem of classifying human actions from a single, static image. We propose that leveraging an automatic caption generator for this task will provide extra information when compared to a traditional convolutional neural network based classifier. The architecture consists of two stages, caption generation and caption classification, used sequentially to a proposed human action class label from a single image. Evaluation is performed of our system and it is evident that caption generation is the limiting factor in accuracy. We propose fixes to both the dataset and the caption generator, in order to improve the model. Finally, it is discovered that caption classification is significantly improved by concatenating all captions from a single image together, to produce one input vector.
Full Title
Caption Aided Action Recognition Using Single Images
Member of
Contributor(s)
Creator: Kafka, Adam - Lehigh University
Creator: Chuah, Mooi C. - Lehigh University
Date Issued
2017
Language
English
Type
Genre
Department name
Computer Science
Media type
Subject (LCSH)
Kafka, . A., & Chuah, . M. C. (2017). Caption Aided Action Recognition Using Single Images (1–). https://preserve.lehigh.edu/lehigh-scholarship/graduate-publications-theses-dissertations/theses-dissertations/caption-aided-0
Kafka, Adam, and Mooi C. Chuah. 2017. “Caption Aided Action Recognition Using Single Images”. https://preserve.lehigh.edu/lehigh-scholarship/graduate-publications-theses-dissertations/theses-dissertations/caption-aided-0.
Kafka, Adam, and Mooi C. Chuah. Caption Aided Action Recognition Using Single Images. 2017, https://preserve.lehigh.edu/lehigh-scholarship/graduate-publications-theses-dissertations/theses-dissertations/caption-aided-0.