Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2070481.2070525acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article

Detecting F-formations as dominant sets

Published: 14 November 2011 Publication History

Abstract

The first step towards analysing social interactive behaviour in crowded environments is to identify who is interacting with whom. This paper presents a new method for detecting focused encounters or F-formations in a crowded, real-life social environment. An F-formation is a specific instance of a group of people who are congregated together with the intent of conversing and exchanging information with each other. We propose a new method of estimating F-formations using a graph clustering algorithm by formulating the problem in terms of identifying dominant sets. A dominant set is a form of maximal clique which occurs in edge weighted graphs. As well as using the proximity between people, body orientation information is used; we propose a socially motivated estimate of focus orientation (SMEFO), which is calculated with location information only. Our experiments show significant improvements in performance over the existing modularity cut algorithm and indicates the effectiveness of using a local social context for detecting F-formations.

References

[1]
A. French, A. Naeem, E. Dryden, and T. Pridmore. Using social effects to guide tracking in complex scenes. In AVSS, pages 212--217, 2007.
[2]
O. Brdiczka, J. Maisonnasse, and P. Reignier. Automatic detection of interaction groups. In Proceedings of the 7th international conference on Multimodal interfaces, ICMI, pages 32--36, 2005.
[3]
M.-C. Chang, N. Krahnstoever, S. Lim, and T. Yu. Group level activity recognition in crowded environments across multiple cameras. AVSS, pages 56--63, 2010.
[4]
T. M. Ciolek. Location of static gatherings in pedestrian areas: an exploratory study. Man-Environment Systems, 7:41--54, 1977.
[5]
T. M. Ciolek and A. Kendon. Environment and the spatial arrangement of conversational encounters. Sociological Inquiry, 50(3--4):237--271, 1980.
[6]
R. Dunbar, N. Duncan, and D. Nettle. Size and structure of freely forming conversational groups. Human Nature, 6(1):67--78, 1995.
[7]
J. J. Edney and N. L. Jordan-Edney. Territorial spacing on a beach. Sociometry, 37(1):92--104, March 1974.
[8]
S. Fortunato and M. Barthélemy. Resolution limit in community detection. Proceedings of the National Academy of Sciences, 104(1):36, 2007.
[9]
W. Ge, R. Collins, and B. Ruback. Automatically detecting the small group structure of a crowd. In Workshop on Applications of Computer Vision, pages 1--8, 2009.
[10]
E. Goffman. Behavior in Public Places: Notes on the Social Organization of Gatherings. Free Press, 1966.
[11]
E. T. Hall. The silent language. Doubleday, Garden City, N. Y., 1959.
[12]
E. T. Hall. The Hidden Dimension. Doubleday, Garden City, N.Y. :, {1st ed.} edition, 1966.
[13]
H. Hung and G. Chittaranjan. The wolf corpus: Exploring group behaviour in a competitive role-playing game. In ACM Multimedia, 10 2010.
[14]
D. B. Jayagopi, H. Hung, C. Yeo, and D. Gatica-Perez. Modeling dominance in group conversations using nonverbal activity cues. IEEE Transactions on Audio, Speech and Language Processing, 17(3):501--513, 2009.
[15]
E. A. Leicht and M. E. J. Newman. Community structure in directed networks. Phys. Rev. Lett., 100:118703, 2008.
[16]
M. E. J. Newman. Modularity and community structure in networks. Proceedings of the National Academy of Sciences, 103(23):8577--8582, June 2006.
[17]
M. Pavan and M. Pelillo. Dominant sets and pairwise clustering. IEEE Transactions on PAMI, 29, 2007.
[18]
S. Pellegrini, A. Ess, K. Schindler, and L. J. V. Gool. You'll never walk alone: Modeling social behavior for multi-target tracking. In ICCV'09, pages 261--268, 2009.
[19]
R. Ranganath, D. Jurafsky, and D. McFarland. It's not you, it's me: Detecting flirting and its misperception in speed-dates. In EMNLP'09, pages 334--342, 2009.
[20]
T. Yu, S. Lim, K. A. Patwardhan, and N. Krahnstoever. Monitoring, recognizing and discovering social networks. In CVPR, 2009.
[21]
W. Zajdel, J. Krijnders, T. Andringa, and D. Gavrila. Cassandra: audio-video sensor fusion for aggression detection. AVSS, 0:200--205, 2007.
[22]
G. Zen, B. Lepri, E. Ricci, and O. Lanz. Space speaks: towards socially and personality aware visual surveillance. In 1st ACM international workshop on Multimodal pervasive video analysis, pages 37--42, 2010.
[23]
W. Zheng, S. Gong, and T. Xiang. Associating groups of people. In BMVC, September 2009.

Cited By

View all
  • (2024)Enabling Social Robots to Perceive and Join Socially Interacting Groups Using F-formation: A Comprehensive OverviewACM Transactions on Human-Robot Interaction10.1145/368207213:4(1-48)Online publication date: 23-Oct-2024
  • (2024)Towards Automatic Social Involvement EstimationProceedings of the 26th International Conference on Multimodal Interaction10.1145/3678957.3688615(612-616)Online publication date: 4-Nov-2024
  • (2024)Psychology-Guided Environment Aware Network for Discovering Social Interaction Groups from VideosACM Transactions on Multimedia Computing, Communications, and Applications10.1145/365729520:8(1-23)Online publication date: 13-Jun-2024
  • Show More Cited By

Recommendations

Reviews

Xiannong Meng

The authors discuss the method of using F-formations to detect dominant sets in a crowd. According to the authors, "A dominant set is a form of maximal clique which occurs in edge weighted graphs." Note that this research is applied to a real human crowd-a social network of people gathered in a public setting, not the social networks found online. The authors define an F-formation as "a group of people who have easy ... access to the same shared space, around which they can communicate for a prolonged time." The F-formation is modeled as graph G = ( V , E , w ), where V is a vertex set, E is an edge set, and w is a positive edge weight function. People are modeled as the set of vertices; connections among people are modeled as the set of edges; and the affinity is implied by the edge weight function w . Function w measures the overall relative affinity between a vertex i (a person) and the rest of the vertices (the rest of the group), weighted by the average affinity of the group. If the value w for a vertex i is above a certain threshold, the person represented by i is said to be in the dominant group. The function w takes into consideration the distance between any pair and face orientation. The dataset used in the study "consists of real footage of over 50 people who met to present scientific work during a poster session." The three-hour event was videotaped. The authors annotated 82 images (of about 1,700 people) and found different F-formation sizes. The data was analyzed after the singletons (one-person groups) were removed. The authors then used the distance between a pair (two meters) and face orientation as parameters to measure the affinity among the people in the crowd. The results show that F-formations are effective with regard to finding dominant sets. Compared with the modularity cut method, F-formation shows significant and more stable performance improvements. The paper is of interest to researchers in sociology who study crowd behavior. It will also help those in the online social networking field look at the problem from a different perspective. Online Computing Reviews Service

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ICMI '11: Proceedings of the 13th international conference on multimodal interfaces
November 2011
432 pages
ISBN:9781450306416
DOI:10.1145/2070481
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 November 2011

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. F-formations
  2. dominant sets
  3. human behavior
  4. modularity cut

Qualifiers

  • Research-article

Conference

ICMI'11
Sponsor:

Acceptance Rates

Overall Acceptance Rate 453 of 1,080 submissions, 42%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)30
  • Downloads (Last 6 weeks)2
Reflects downloads up to 16 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Enabling Social Robots to Perceive and Join Socially Interacting Groups Using F-formation: A Comprehensive OverviewACM Transactions on Human-Robot Interaction10.1145/368207213:4(1-48)Online publication date: 23-Oct-2024
  • (2024)Towards Automatic Social Involvement EstimationProceedings of the 26th International Conference on Multimodal Interaction10.1145/3678957.3688615(612-616)Online publication date: 4-Nov-2024
  • (2024)Psychology-Guided Environment Aware Network for Discovering Social Interaction Groups from VideosACM Transactions on Multimedia Computing, Communications, and Applications10.1145/365729520:8(1-23)Online publication date: 13-Jun-2024
  • (2024)Exploring User Placement for VR Remote Collaboration in a Constrained Passenger SpaceProceedings of the 30th ACM Symposium on Virtual Reality Software and Technology10.1145/3641825.3687722(1-11)Online publication date: 9-Oct-2024
  • (2024)Toward Grouping in Large Scenes With Occlusion-Aware Spatio–Temporal TransformersIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2023.332486834:5(3919-3929)Online publication date: May-2024
  • (2024)I am Part of the Robot’s Group: Evaluating Engagement and Group Membership from Egocentric Views2024 33rd IEEE International Conference on Robot and Human Interactive Communication (ROMAN)10.1109/RO-MAN60168.2024.10731376(1774-1779)Online publication date: 26-Aug-2024
  • (2024)Indoor Group Identification and Localization Using Privacy-Preserving Edge Computing Distributed Camera NetworkIEEE Journal of Indoor and Seamless Positioning and Navigation10.1109/JISPIN.2024.33542482(51-60)Online publication date: 2024
  • (2024)The Audio-Visual Conversational Graph: From an Egocentric-Exocentric Perspective2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR52733.2024.02493(26386-26395)Online publication date: 16-Jun-2024
  • (2024)Exploiting temporal information to detect conversational groups in videos and predict the next speakerPattern Recognition Letters10.1016/j.patrec.2023.10.002177(164-168)Online publication date: Jan-2024
  • (2024)A graph neural approach for group recommendation system based on pairwise preferencesInformation Fusion10.1016/j.inffus.2024.102343107(102343)Online publication date: Jul-2024
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media