•  Retrait gratuit dans votre magasin Club
  •  7.000.000 titres dans notre catalogue
  •  Payer en toute sécurité
  •  Toujours un magasin près de chez vous     
  •  Retrait gratuit dans votre magasin Club
  •  7.000.0000 titres dans notre catalogue
  •  Payer en toute sécurité
  •  Toujours un magasin près de chez vous
  1. Accueil
  2. Livres
  3. Sciences humaines
  4. Sciences
  5. Technique
  6. Énergie
  7. Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes

Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes

Roman Seidel
Livre broché | Anglais
21,95 €
+ 43 points
Livraison sous 1 à 4 semaines
Passer une commande en un clic
Payer en toute sécurité
Livraison en Belgique: 3,99 €
Livraison en magasin gratuite

Description

The property of human motion perception is used in this dissertation to infer human activity from data using artificial neural networks. One of the main aims of this thesis is to discover which modalities, namely RGB images, optical flow and human keypoints, are best suited for HAR in omnidirectional data. Since these modalities are not yet available for omnidirectional cameras, they are synthetically generated with a 3D indoor simulation with the result of a large-scale dataset, called OmniFlow. Due to the lack of omnidirectional optical flow data, the OmniFlow dataset is validated using Test-Time Augmentation. Compared to the baseline, which contains Recurrent All-Pairs Field Transforms trained on the FlyingChairs and FlyingThings3D datasets, it was found that only about 1000 images need to be used for fine-tuning to obtain a very low End-point Error. For an evaluation on activity-level, two state-of-the-art convolutional neural networks (CNNs), namely the Temporal Segment Network (TSN) for the modalities RGB images and optical flow and the PoseC3D for the modality human keypoints, were used. Both CNNs were trained and validated on OmniFlow and on the real-world dataset OmniLab. For both networks, TSN and PoseC3D, three hyperparameters were varied and the top-1, top-5 and mean accuracies were reported. In addition, confusion matrices indicating the class-wise accuracy of the 15 activity classes have been given for the modalities RGB images, optical flow and human keypoints.

Spécifications

Parties prenantes

Auteur(s) :
Editeur:

Contenu

Nombre de pages :
180
Langue:
Anglais

Caractéristiques

EAN:
9783961002054
Format:
Livre broché
Dimensions :
148 mm x 12 mm
Poids :
314 g

Les avis

Nous publions uniquement les avis qui respectent les conditions requises. Consultez nos conditions pour les avis.