Automatic Metadata Generation Through Analysis of Narration Within Instructional Videos

Research output: Contribution to journalArticle

2 Citations (Scopus)

Abstract

Current activity recognition based assistive living solutions have adopted relatively rigid models of inhabitant activities. These solutions have some deficiencies associated with the use of these models. To address this, a goal-oriented solution has been proposed. In a goal-oriented solution, goal models offer a method of flexibly modelling inhabitant activity. The flexibility of these goal models can dynamically produce a large number of varying action plans that may be used to guide inhabitants. In order to provide illustrative, video-based, instruction for these numerous actions plans, a number of video clips would need to be associated with each variation. To address this, rich metadata may be used to automatically match appropriate video clips from a video repository to each specific, dynamically generated, activity plan. This study introduces a mechanism of automatically generating suitable rich metadata representing actions depicted within video clips to facilitate such video matching. This performance of this mechanism was evaluated using eighteen video files; during this evaluation metadata was automatically generated with a high level of accuracy.
LanguageEnglish
JournalJournal of Medical Systems
Volume39
Issue number9
DOIs
Publication statusPublished - 8 Aug 2015

Fingerprint

Narration
Metadata
Surgical Instruments

Keywords

  • Assistive living
  • Automated speech recognition
  • Metadata
  • Ontology
  • Parsing
  • Smart environments
  • Video

Cite this

@article{8cbe89a3fc604e21b9807845647fd3f1,
title = "Automatic Metadata Generation Through Analysis of Narration Within Instructional Videos",
abstract = "Current activity recognition based assistive living solutions have adopted relatively rigid models of inhabitant activities. These solutions have some deficiencies associated with the use of these models. To address this, a goal-oriented solution has been proposed. In a goal-oriented solution, goal models offer a method of flexibly modelling inhabitant activity. The flexibility of these goal models can dynamically produce a large number of varying action plans that may be used to guide inhabitants. In order to provide illustrative, video-based, instruction for these numerous actions plans, a number of video clips would need to be associated with each variation. To address this, rich metadata may be used to automatically match appropriate video clips from a video repository to each specific, dynamically generated, activity plan. This study introduces a mechanism of automatically generating suitable rich metadata representing actions depicted within video clips to facilitate such video matching. This performance of this mechanism was evaluated using eighteen video files; during this evaluation metadata was automatically generated with a high level of accuracy.",
keywords = "Assistive living, Automated speech recognition, Metadata, Ontology, Parsing, Smart environments, Video",
author = "Joseph Rafferty and Christopher Nugent and Jun Liu and Liming Chen",
year = "2015",
month = "8",
day = "8",
doi = "10.1007/s10916-015-0295-2",
language = "English",
volume = "39",
journal = "Journal of Medical Systems",
issn = "0148-5598",
number = "9",

}

TY - JOUR

T1 - Automatic Metadata Generation Through Analysis of Narration Within Instructional Videos

AU - Rafferty, Joseph

AU - Nugent, Christopher

AU - Liu, Jun

AU - Chen, Liming

PY - 2015/8/8

Y1 - 2015/8/8

N2 - Current activity recognition based assistive living solutions have adopted relatively rigid models of inhabitant activities. These solutions have some deficiencies associated with the use of these models. To address this, a goal-oriented solution has been proposed. In a goal-oriented solution, goal models offer a method of flexibly modelling inhabitant activity. The flexibility of these goal models can dynamically produce a large number of varying action plans that may be used to guide inhabitants. In order to provide illustrative, video-based, instruction for these numerous actions plans, a number of video clips would need to be associated with each variation. To address this, rich metadata may be used to automatically match appropriate video clips from a video repository to each specific, dynamically generated, activity plan. This study introduces a mechanism of automatically generating suitable rich metadata representing actions depicted within video clips to facilitate such video matching. This performance of this mechanism was evaluated using eighteen video files; during this evaluation metadata was automatically generated with a high level of accuracy.

AB - Current activity recognition based assistive living solutions have adopted relatively rigid models of inhabitant activities. These solutions have some deficiencies associated with the use of these models. To address this, a goal-oriented solution has been proposed. In a goal-oriented solution, goal models offer a method of flexibly modelling inhabitant activity. The flexibility of these goal models can dynamically produce a large number of varying action plans that may be used to guide inhabitants. In order to provide illustrative, video-based, instruction for these numerous actions plans, a number of video clips would need to be associated with each variation. To address this, rich metadata may be used to automatically match appropriate video clips from a video repository to each specific, dynamically generated, activity plan. This study introduces a mechanism of automatically generating suitable rich metadata representing actions depicted within video clips to facilitate such video matching. This performance of this mechanism was evaluated using eighteen video files; during this evaluation metadata was automatically generated with a high level of accuracy.

KW - Assistive living

KW - Automated speech recognition

KW - Metadata

KW - Ontology

KW - Parsing

KW - Smart environments

KW - Video

U2 - 10.1007/s10916-015-0295-2

DO - 10.1007/s10916-015-0295-2

M3 - Article

VL - 39

JO - Journal of Medical Systems

T2 - Journal of Medical Systems

JF - Journal of Medical Systems

SN - 0148-5598

IS - 9

ER -