Modelling scenes using the activity within them

Standard

Modelling scenes using the activity within them. / Dee, Hannah; Fraile, Roberto; Hogg, David C. et al.

2008. 394-408 Paper presented at Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI).

Research output: Contribution to conferencePaperpeer-review

Harvard

Dee, H, Fraile, R, Hogg, DC & Cohn, AG 2008, 'Modelling scenes using the activity within them', Paper presented at Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI), 15 Sept 2008 pp. 394-408. https://doi.org/10.1007/978-3-540-87601-4_28

APA

Dee, H., Fraile, R., Hogg, D. C., & Cohn, A. G. (2008). Modelling scenes using the activity within them. 394-408. Paper presented at Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI). https://doi.org/10.1007/978-3-540-87601-4_28

Vancouver

Dee H, Fraile R, Hogg DC, Cohn AG. Modelling scenes using the activity within them. 2008. Paper presented at Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI). doi: 10.1007/978-3-540-87601-4_28

Author

Dee, Hannah ; Fraile, Roberto ; Hogg, David C. et al. / Modelling scenes using the activity within them. Paper presented at Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI).15 p.

Bibtex - Download

@conference{b6f67de39575489996992fb58055dfa4,
title = "Modelling scenes using the activity within them",
abstract = "This paper describes a method for building visual “maps” from video data using quantized descriptions of motion. This enables unsupervised classification of scene regions based upon the motion patterns observed within them. Our aim is to recognise generic places using a qualitative representation of the spatial layout of regions with common motion patterns. Such places are characterised by the distribution of these motion patterns as opposed to static appearance patterns, and could include locations such as train platforms, bus stops, and park benches. Motion descriptions are obtained by tracking image features over a temporal window, and are then subjected to normalisation and thresholding to provide a quantized representation of that feature{\textquoteright}s gross motion. Input video is quantized spatially into N ×N pixel blocks, and a histogram of the frequency of occurrence of each vector is then built for each of these small areas of scene. Within these we can therefore characterise the dominant patterns of motion, and then group our spatial regions based upon both proximity and local motion similarity to define areas or regions with particular motion characteristics. Moving up a level we then consider the relationship between the motion in adjacent spatial areas, and can characterise the dominant patterns of motion expected in a particular part of the scene over time. The current paper differs from previous work which has largely been based on the paths of moving agents, and therefore restricted to scenes in which such paths are identifiable. We demonstrate our method in three very different scenes: an indoor room scenario with multiple chairs and unpredictable unconstrained motion, an underground station featuring regions where motion is constrained (train tracks) and regions with complicated motion and difficult occlusion relationships (platform), and an outdoor scene with challenging camera motion and partially overlapping video streams.",
author = "Hannah Dee and Roberto Fraile and Hogg, {David C.} and Cohn, {Anthony G.}",
note = "Dee, H. M.; Fraile, R; Hogg, D. C. and Cohn, A. G. Modelling scenes using the activity within them, Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI) pp 394-408, Freiburg, Germany 2008 Sponsorship: EPSRC; Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI) ; Conference date: 15-09-2008",
year = "2008",
doi = "10.1007/978-3-540-87601-4_28",
language = "English",
pages = "394--408",

}

RIS (suitable for import to EndNote) - Download

TY - CONF

T1 - Modelling scenes using the activity within them

AU - Dee, Hannah

AU - Fraile, Roberto

AU - Hogg, David C.

AU - Cohn, Anthony G.

N1 - Dee, H. M.; Fraile, R; Hogg, D. C. and Cohn, A. G. Modelling scenes using the activity within them, Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI) pp 394-408, Freiburg, Germany 2008 Sponsorship: EPSRC

PY - 2008

Y1 - 2008

N2 - This paper describes a method for building visual “maps” from video data using quantized descriptions of motion. This enables unsupervised classification of scene regions based upon the motion patterns observed within them. Our aim is to recognise generic places using a qualitative representation of the spatial layout of regions with common motion patterns. Such places are characterised by the distribution of these motion patterns as opposed to static appearance patterns, and could include locations such as train platforms, bus stops, and park benches. Motion descriptions are obtained by tracking image features over a temporal window, and are then subjected to normalisation and thresholding to provide a quantized representation of that feature’s gross motion. Input video is quantized spatially into N ×N pixel blocks, and a histogram of the frequency of occurrence of each vector is then built for each of these small areas of scene. Within these we can therefore characterise the dominant patterns of motion, and then group our spatial regions based upon both proximity and local motion similarity to define areas or regions with particular motion characteristics. Moving up a level we then consider the relationship between the motion in adjacent spatial areas, and can characterise the dominant patterns of motion expected in a particular part of the scene over time. The current paper differs from previous work which has largely been based on the paths of moving agents, and therefore restricted to scenes in which such paths are identifiable. We demonstrate our method in three very different scenes: an indoor room scenario with multiple chairs and unpredictable unconstrained motion, an underground station featuring regions where motion is constrained (train tracks) and regions with complicated motion and difficult occlusion relationships (platform), and an outdoor scene with challenging camera motion and partially overlapping video streams.

AB - This paper describes a method for building visual “maps” from video data using quantized descriptions of motion. This enables unsupervised classification of scene regions based upon the motion patterns observed within them. Our aim is to recognise generic places using a qualitative representation of the spatial layout of regions with common motion patterns. Such places are characterised by the distribution of these motion patterns as opposed to static appearance patterns, and could include locations such as train platforms, bus stops, and park benches. Motion descriptions are obtained by tracking image features over a temporal window, and are then subjected to normalisation and thresholding to provide a quantized representation of that feature’s gross motion. Input video is quantized spatially into N ×N pixel blocks, and a histogram of the frequency of occurrence of each vector is then built for each of these small areas of scene. Within these we can therefore characterise the dominant patterns of motion, and then group our spatial regions based upon both proximity and local motion similarity to define areas or regions with particular motion characteristics. Moving up a level we then consider the relationship between the motion in adjacent spatial areas, and can characterise the dominant patterns of motion expected in a particular part of the scene over time. The current paper differs from previous work which has largely been based on the paths of moving agents, and therefore restricted to scenes in which such paths are identifiable. We demonstrate our method in three very different scenes: an indoor room scenario with multiple chairs and unpredictable unconstrained motion, an underground station featuring regions where motion is constrained (train tracks) and regions with complicated motion and difficult occlusion relationships (platform), and an outdoor scene with challenging camera motion and partially overlapping video streams.

U2 - 10.1007/978-3-540-87601-4_28

DO - 10.1007/978-3-540-87601-4_28

M3 - Paper

SP - 394

EP - 408

T2 - Springer Lecture notes in Artificial Intelligence (Spatial Cognition VI)

Y2 - 15 September 2008

ER -

View graph of relations
Citation formats