Log In Sign Up

MeetSense: A Lightweight Framework for Group Identification using Smartphones

In an organization, individuals prefer to form various formal and informal groups for mutual interactions. Therefore, ubiquitous identification of such groups and understanding their dynamics are important to monitor activities, behaviours and well-being of the individuals. In this paper, we develop a lightweight, yet near-accurate, methodology, called MeetSense, to identify various interacting groups based on collective sensing through users' smartphones. Group detection from sensor signals is not straightforward because users in proximity may not always be under the same group. Therefore, we use acoustic context extracted from audio signals to infer interaction pattern among the subjects in proximity. We have developed an unsupervised and lightweight mechanism for user group detection by taking cues from network science and measuring the cohesivity of the detected groups in terms of modularity. Taking modularity into consideration, MeetSense can efficiently eliminate incorrect groups, as well as adapt the mechanism depending on the role played by the proximity and the acoustic context in a specific scenario. The proposed method has been implemented and tested under many real-life scenarios in an academic institute environment, and we observe that MeetSense can identify user groups with close to 90 environment.


page 1

page 2

page 3

page 4


GroupSense: A Lightweight Framework for Group Identification using Smartphones

In an organization, individuals prefer to form various formal and inform...

Chemotaxis and Quorum Sensing inspired Device Interaction supporting Social Networking

Conference and social events provides an opportunity for people to inter...

Automated Lane Detection in Crowds using Proximity Graphs

Studying the behavior of crowds is vital for understanding and predictin...

The Anatomy of Leadership in Collective Behaviour

Understanding the mechanics behind the coordinated movement of mobile an...

Anatomy of Leadership in Collective Behaviour

Understanding the mechanics behind the coordinated movement of mobile an...

What Smartphones, Ethnomethodology, and Bystander Inaccessibility Can Teach Us About Better Design?

Smartphones, the ubiquitous mobile screens now normal parts of everyday ...

Detecting F-formations Roles in Crowded Social Scenes with Wearables: Combining Proxemics Dynamics using LSTMs

In this paper, we investigate the use of proxemics and dynamics for auto...

1 Introduction

Workplace meetings and team formation among the individuals are key factors behind organizational efficiency. In organizations and institutions, people formally as well as sporadically meet, interact and form groups for various purposes, which include information sharing [1], teaching and learning [2], problem solving and decision making [3], brainstorming [4], socialization [5] etc. Tracking the dynamics of group formation facilitates various utilities; for instance, organizational leaders may prefer to monitor the formation of teams, which benefit the overall efficiency and activeness of the organization [6, 7]; course instructors in flipped classrooms [8] in an academic environment may like to know how the students form groups among themselves to solve assignments and exercises. Unlike regular & pre-scheduled team meetings, the formation of sporadic and instantaneous groups (often observed in office breaks, conferences etc.) make the problem challenging. On the other hand, increasing availability of sensor-rich smartphones provides a unique opportunity for collecting wide sensor information in a seamless manner. In this backdrop, we investigate the potential of smartphones to develop a lightweight ubiquitous system for identifying and monitoring group formation. Notably, in this paper, we primarily concentrate on the meeting groups where co-located group members occasionally interact with each other. In this line, we capture different types of real-life meeting group scenarios such as outdoor roadside informal meeting; informal outdoor cafe meet, formal and informal laboratory meeting, and classroom interaction as shown in Figure 1.

Identification of a meeting group primarily relies on the location proximity [9, 10] of the group members, which (apparently) can be conceptualized as a localization problem [11, 12]. In that direction, prior art explores the following three modalities – GPS, Bluetooth, and WiFi for identification of the location similarity in supervised [10] as well as unsupervised [13] manner. In our context of group detection, vanilla localization based solutions demand high accuracy, which pushes the system towards complex processing. Notably, location proximity alone is insufficient to correctly discriminate and identify the meeting groups. For instance, consider a large conference hall, where multiple meeting groups get formed simultaneously; here members of different groups may exhibit location similarity among themselves, which makes the group detection challenging. Close inspection reveals that albeit similarity in location proximity, context [14, 15, 16, 17] of the members participating in individual meeting groups play a critical role in identifying groups; for instance all the members of a specific group in the conference hall share a substantial amount of contextual similarity (room illumination, ambience noise, member interactions, magnetic fluctuations) [16, 17, 11]. However, identifying suitable contextual information, which is computationally lightweight as well as carries the signature of a meeting group, is an important problem.

We propose acoustic context, extracted from the audio signals received by individual smartphones, as a key context indicator. In order to compute acoustic context, one can apply standard Mel-frequency Cepstral Coefficients (MFCC) [18] on the recorded audio signals for speaker identification by measuring the tone & pitch. However, this solution comes with multiple challenges. (a) The process of using MFCC usually follows a supervised approach which needs individuals’ pre-trained information. (b) MFCC is computationally expensive, which makes it inappropriate for developing a lightweight system. (c) MFCC technique is quite sensitive to noise, hence most suitable for the unidirectional microphone with the stereo channel. Unfortunately, most of the commercial smartphones are equipped with omnidirectional microphones, which makes them prone to noise and corrupting the speaker identification process111 (last accessed Apr 12, 2018) . In Next2Me [19], Baker and Efstratiou attempted to detect social groups considering WiFi and sound fingerprints. First, WiFi signal strengths are used for detecting the co-located population; next, this filtered population is fed to the audio module for finding out the social groups. The audio module considers the top frequencies of all the co-located individuals and computes the pairwise similarities. However, in the real-life environment, getting the actual top frequencies is challenging, and little variation in the selection of frequencies exerts a huge impact on the similarity computation. Additionally, the audio signals captured on different smartphones can be time drifted, even if a single speaker acts as the audio source, since the clocks of different devices may not be time synchronized, and the subjects (devices) may be at different distances from the speaker. Once the co-located population has been identified and audio based context information has been extracted, state of the art techniques perform naive component analysis [19] and community detection [20] to identify social groups. However, in most of the cases, quality (cohesivity) of the discovered groups have been overlooked, which leads to the detection of incorrect communities (false positives).

Fig. 1: Setup of Different Meeting Group Scenarios

In this paper, we develop MeetSense, a smartphone-driven ubiquitous platform for automatic detection of meeting groups. The proposed method is lightweight, unsupervised, hence equipped to detect instantaneously formed groups, without any pre-training. First, we determine the co-located population using standard localization techniques [10, 19]. In our implementation, we relied on the WiFi-based proximity; nevertheless, this can be extended to Bluetooth and GPS based techniques as well. The crux of the proposed method is the computation of acoustic context of the identified co-located population, which is based on the following key intuition. Interactions between participants of a meeting group switches from one speaker to another; where, at a time, there exists (mostly) one dominating speaker. Hence, power of the dominating tone (say ) captured by the smartphones (subjects222In this paper, we use the term ‘subject’ to indicate a participant, a member or a smartphone, interchangeably.) in one group (say ) is significantly different from the power of the dominating tone () captured by the devices of the another group . If both the groups and are closely located, then all the devices might capture both the tones with varying power. However, for the devices in group , the power of the dominating tone should be higher than , whereas exactly opposite is likely to happen for the group . By discriminating the power of the dominating tone, one can differentiate the acoustic context of the members of different groups. Finally, leveraging on the proximity of the co-located population and their acoustic context, we propose MeetSense, a community-driven group detection model. The advantage of this model is manifold.

  1. The model is unsupervised and lightweight.

  2. This model can perform group detection even in the absence of proximity indicators (say, WiFi etc.).

  3. We take cues from network science and measure the cohesivity of the detected groups with the help of modularity. Taking modularity into consideration, MeetSense can efficiently eliminate incorrect groups (reduce false positives), as well as adapt the algorithm depending on the role played by the proximity and the acoustic context in a specific scenario. For instance, in case of a noisy environment, MeetSense combines both the modalities to identify meeting groups.

The organization of the paper is the following. In Section 2, we formally define the meeting group and state the problem of group detection. We introduce two primary indicators and the related literature in those contexts – (a) proximity, to identify co-located population and (b) audio signal, to compute acoustic context. We also conduct pilot experiments to highlight the challenges in extracting acoustic context amidst noisy environment, device heterogeneity etc. In Section 3, we propose a novel sound signal processing approach that can capture the acoustic context even with low power microphones available with the smartphones. In Section 4, we develop MeetSense, a group detection model leveraging on the community detection algorithms. We have implemented MeetSense in an academic campus scenario, and captured several groups like classroom teaching, lab meetings, seminars, cafeteria gatherings, outdoor meetings etc. In Section 5, we show that MeetSense can detect such groups with more than accuracy while incurring low computation overhead compared to the state of the art group identification methods.

2 Problem Definition and Background Study

In this section, first, we define the meeting group and state the problem of group detection in the context of smartphone-based sensing. Next, we identify the primary indicators (say, proximity, acoustic context etc.) facilitating the group detection and explore their potential in the light of state of the art endeavours. Finally, we concentrate on the acoustic context and conduct a pilot study to highlight the challenges in group detection from audio signatures.

2.1 Problem Statement

We start with the definition of a Meeting Group and subsequently state the problem of group detection.

Definition 1 (Meeting Group)

Given a population of subjects , we define a meeting group for the time period as the collection of co-located individuals sharing similar context.

For instance, two subjects and participate in a group iff and are located in close proximity and share similar context for time duration [13]. In this paper, we hypothesize that sound profile, observed by the group members, defines the acoustic context of a group. For instance, sensing the verbal interactions between the participating members can discriminate one meeting group from another. Notably, in the acoustic context, we only concentrate on the tone from the human voice signal, and categorically disregard the content of interaction to preserve privacy.

Consider each subject carries a smartphone equipped with various sensors. We collect the sensor log from each subject and populate the data in a central repository . The sensor log comprises of the location information and acoustic information . The location information may come from various signals for indoor and outdoor localization techniques based on GPS, wireless signals etc. [21, 17, 22, 12, 23, 24]; similarly acoustic information can be extracted from the audio signals captured by the smartphones [25, 26, 27, 28]. We aim to discover the meeting group formed during the period from the logged sensor repository .

2.2 Primary Indicators and Respective Prior Art

The definition of the meeting group mainly relies on two kinds of sensing modalities – (a) location Proximity and (b) Acoustic context. We explore the recent attempts in this direction and highlight their potential & challenges in group detection.

2.2.1 Location Proximity

Localizing the subjects within their proximity is the initial step towards the group identification. In the line, the past literature explores mainly three modalities – GPS, Bluetooth, and WiFi. GPS [23] is an important modality (albeit energy-hungry) for localization and detecting population within proximity. Although GPS performs well in outdoor environments, its accuracy sharply falls in indoor environments due to the interruption in the signal [21]. On the other side, Bluetooth-based study [29] is one of the earliest attempts for localization in indoor environments. However, Bluetooth scanning is power hungry [10]. Moreover, many of the Android smartphones (starting from versions ) have partial support for Bluetooth Low-Energy (BLE), which are capable of only detecting other BLE devices [16]. Additionally, the Bluetooth signal as a medium of information is considered to be unreliable and noisy.

Recently, attempts have been made to detect proximity from WiFi fingerprint [10]. WiFi-based localization is considered as a promising indicator for identifying the population in proximity. WiFi consumes significantly less power as compared to Bluetooth and GPS. Although BLE appears as an alternative to WiFi in terms of power consumption, nevertheless BLE suffers from data loss and fluctuations with increasing distance [30]. Furthermore, WiFi can work in any environment irrespective of whether the device location is indoor or outdoor. Each modality has its positive and negative aspects in the context of localization. Hence, the selection of modalities is highly dependent on the application for which the proximity is computed. In [10], the authors have developed a supervised based learning approach for person-to-person proximity detection using WiFi fingerprints, like access point (AP) coverage and signal strength measurements. On the other hand, the authors in [19]

have developed an unsupervised learning based approach for proximity detection using a novel WiFi based metric computed using

Manhattan distance, which is the average of the pairwise signal strength difference among the APs, from which the subject receives signals. Any of these existing mechanism can be used for proximity detection for group identification. Once a set of subjects are detected to be in proximity, then the contextual similarity further characterizes the subset that forms a meeting group.

Fig. 2: Impact of Audio Signals in Group Detection - two speakers from two different groups talk simultaneously

2.2.2 Acoustic Context

A microphone is an important indicator to identify the meeting group members. Participants, in general, avoid talking simultaneously in a meeting; although there can be a small overlap when the discussion switches from one speaker to another (utterance duration). Therefore the voice properties, such as pitch and tone of the current speaker in a group dominates in the audio signals captured by individual subjects in that group [31]. Pitch defines the perceived fundamental frequency of the sound [32], whereas tone is the variation or thickness of the pitch, indicating the quality of the sound [33]. Figure 2 explains the intuition behind using human voice characteristics for group identification. The blue audio signal dominates for the subjects of , whereas the red signal dominates for the subjects of . Therefore, human voice characteristics (aka acoustic context) may show a strong feature similarity, if the subjects belong to the same group.

Audio pitch and tone extraction from human voice signal is a well-studied problem in the literature [33, 32]. YIN [32] is a simple time-domain pitch calculation algorithm which is used in many existing applications such as counting the crowd from human voice signals [33]. Although the pitch is a good indicator for speaker identification, however, pitch alone fails to differentiate the relative distance of the speakers from other subjects, since it only concentrates on the central frequency of the audio signal. Therefore, tone information needs to be extracted along with the pitch, and Mel-frequency Cepstral Coefficients (MFCC) based techniques [18]

with Gaussian Mixture Model (GMM) 

[34] can be applied for this purpose. However, in smartphones, the use of unidirectional microphones with the stereo channel is rare. A smartphone may capture the voice signals from the subjects of the other nearby groups, apart from the primary speaker of its group, as shown in Figure 2. Further, the environmental noise generated from the variety of external sources may collude the recorded audio signal. For instance, the humming noise generated from the ACs and other machines (indoor) and vehicles (outdoor) may collude the collected audio signals and make the group detection challenging. Additionally, for instantaneous group detection, there is no apriori knowledge of the group members’ tone information. Therefore, group identification from MFCC based audio processing along with some supervised techniques may pose some additional challenges, although they work well for applications like crowd count [33]. In the following, we explore these challenges from the observations over a pilot experiment.

2.3 Pilot Study: Revealing Challenges with Audio Signals

We launched a pilot study to examine the potential of the acoustic context in identifying the meeting groups amidst challenging scenarios. We developed an Android app for collecting the audio signal log from the smartphones for conducting the pilot experiment. We recruited six subjects in this experiment for two weeks, installed the app on their smartphones and instructed them to occasionally form pre-designed meeting groups (multiple times) for around minutes. Subjects have been asked to record the group formation instances manually for validation. The detailed overview of the formed groups in this study is listed in Table I. During the experiments, we have captured bit audio signal at kHz sampling rate from the in-built microphones of the smartphones. Notably, while forming those controlled groups, we pay special attention towards incorporating two fundamental challenges that may affect the group identification from audio signals – (1) device heterogeneity and (2) noisy environments. To capture device heterogeneity, we have used smartphones from four different makes and models – Moto X, Moto G 2nd Gen, OnePlus3, Samsung Note5. The noise environment can be summarized in the following generic scenarios.

Group ID Member IDs Location
,, SMR Lab
,, Class C-118
,,, Cafeteria
, SMR Lab
,,, Way to Cafeteria
,, Outdoor Roadside
,, Outdoor Roadside
TABLE I: Pilot Study Minutiae
Fig. 3: Impact of Audio Pressure among the Subjects of Same Group: , , , in
Fig. 4: Impact of Audio Pressure among the Subjects of Different Groups: , in and in
Fig. 5: Deviations of Frequencies in Groups

(a) Low noise environment: This includes the formation of meeting groups where the surrounding environmental noise is low (audio amplitude less than dB [35]). Subjects forming groups inside classrooms (group ), during the conference in formal gatherings, while moving inside the laboratory (, ) etc. can be included in this scenario.

(b) Noisy environment: In this scenario, subjects are forming groups in the highly noisy environment (audio amplitude more than dB). Subjects forming groups in cafeterias (group ), in informal gathering, marketplace, outdoor environment () etc. fall in this scenario.

2.3.1 Observations

We first normalize the amplitude of the audio signal and then compute the audio pressure as an indicator of the volume of the audio signal received by individual devices. We concentrate on the meeting group where subject primarily speaks while other group members mostly remain silent. In Figure 5, we plot the audio pressure received from the individual subjects (, , , ) of group . We observe that subjects (, , ), participating in the same group , exhibit similar audio pressure. However, the audio pressure of deviates from the rest of the subjects since the user is moving while speaking. Therefore, the values are slightly different than the other group members.

The scenario gets compounded when we consider two groups (, , ) and (, ) which get formed inside the same laboratory during the similar time period. Figure 5 highlights the fact that although audio pressure of the subjects (, ) participating in same group () exhibit similar behavior, however, the same indicator fails to show clear discrimination between the subjects (say , ) participating in two different groups (, ). For further investigation, we move to frequency domain based analysis. Importantly, Figure 5 shows that the frequency component present in subject exhibits contrasting behaviour from the subject , belonging to a different group. However, the frequency components of subjects and present in the same meeting group () exhibit (albeit minor) difference (due to environmental noise), posing a new challenge. Last but not the least, Figure 6 demonstrates the variation of amplitude (raw version of audio pressure) due to device heterogeneity. The smartphone microphones use automatic gain control (AGC) circuit, which exaggerates the variation of amplitude for the same audio signal captured through different devices. In group , the subjects and carrying same make & model devices whereas another subject carries a different build. Although all three of them belong to the same meeting group, Figure 5(b) exhibits a dissimilarity in amplitude for the subjects and (nevertheless, similarity can be observed for subjects and (Figure 5(a))). The detailed comparison of the devices is listed in Table II, in terms of similarity index for the same audio signal. We observe that the similarity index is sometimes quite low for two different devices from two different makes and models.

(a) Same Build:
and in
(b) Different Build:
and in
Fig. 6: Audio amplitude in devices of same and different builds

2.3.2 Lessons Learnt

We observe that audio signals provide us with a good indicator to capture the acoustic context of a group. However, due to omnidirectional nature of smartphone microphones, a significant audio pressure from the speakers of the nearby groups is also getting captured, as we observe in Figure 5 (formation of two groups and in the same lab). For MFCC based techniques, the separability of cepstral coefficients gets distorted in the presence of multiple speakers and environmental noise [36, 31, 37]. Hence, MFCC may be able to capture the presence of two speakers in the vicinity for groups and

, but it will fail to classify the subjects based on the primary speakers. Device heterogeneity exaggerates this problem further. In summary, although microphone provides important signature uncovering group membership, however, it is inadequate in its current form for handling adverse scenarios.

Audio Amplitude
OnePlus3 MotoG
Moto X 0.3247 0.1178 -0.2781 -0.1138
Samsung Note5 0.1178 0.2977 0.0896 0.1287
Oneplus3 -0.2781 0.0896 0.5671 -0.0653
Moto G -0.1138 0.1287 -0.0653 0.5822
TABLE II: Audio Amplitude Similarity in Heterogeneous Devices

3 Measuring Acoustic Context of Meeting Groups

Fig. 7: Audio Signal Processing Flowchart

From the pilot study, we demonstrate that audio signals are rich sources to capture the context of a meeting group. However, we also comprehend that the naive audio processing techniques are not sufficient to extract reliable information under various complicated scenarios. In this section, we develop a methodology for computing acoustic context from smartphone audio signals, as shown in Figure 7. The different steps in this procedure are as follows.

3.1 Preprocessing of Vanilla Audio Signals

For audio-based feature extraction, we collect the audio data

from all the subjects at a sampling rate of , continuously for units with an interval of units of time, where and are specified by the application developers. We first extract the human speech signal between Hz to Hz using Butterworth bandpass filtering. The human speech signals captured from different smartphones are used for further processing.

3.1.1 Time Drift Adjustment

(a) Two time drifted signals from same audio source
(b) Correlation by shifting the second signal with reference to the first signal
Fig. 8: Computation of Time Drift

The audio signals captured from different devices can be time drifted, even if a single speaker acts as the audio source. There are broadly two reasons for this – (a) the clocks at different devices may not be time synchronized, and (b) the subjects may be at different distances from the speaker, which introduces propagation lag to the signals. Figure 7(a) shows the time drifted signals with a single speaker, captured from two different subjects. To compare two signals, we need to place both the signals at the same time reference frame, and therefore eliminating the time drift is an important task for audio processing.

Although some existing studies have developed techniques for time drift adjustment of audio signals captured in hand-held devices [38], they employ smoothing techniques over the raw signal and thus tend to lose the physical properties of the signal, such as tone and pitch of the signal. However, such physical properties are important to capture the nature of human voice, which are essential for extracting acoustic context. Therefore, we apply a simple technique in this paper to mitigate the time drift introduced in the signals coming from a single audio source.

To eliminate the time drift, we use the concept of similarity measure between the signals in the time domain. Consider the audio signal coming from a single source, but captured at two different devices. Ideally, when both the signals are placed at the same reference frame, at the time domain (drift is zero), the similarity between them should be maximum. To measure the similarity between the signals, we use statistical correlation. The procedure works as follows. We fix one signal as the reference, and then shift another signal for the one-time unit at every step, and measure the correlation between the signals. Figure 7(b) plots the correlation between the signals shown in Figure 7(a), with respect to the amount of time shift applied over the second signal, while considering the first signal as the reference. A positive time shift indicates that the second signal has been shifted towards the time clock, and a negative time shift represents that the signal has been shifted backwards the time clock. In the example, we observe that the correlation is maximum when the second signal is shifted seconds, indicating that the drift is seconds. Once the drift is calculated, one signal is shifted to make the drift zero with respect to the reference signal.

3.2 Audio Tone Extraction

The audio tone of the members of a meeting group should exhibit high similarity among themselves whereas tone dissimilarity indicates different groups. Hence, pairwise tone similarity between the group members should be an important property to determine the acoustic context of that group. Considering that group participants, in general, avoid talking simultaneously in a meeting, intuitively, there exists one dominating tone that gets captured at the smartphones of all the subjects in a meeting group. Audio tone extraction is a well-studied problem [33, 32] and Mel-frequency cepstral coefficients (MFCC) based techniques [32] are widely applied for tone extraction from audio signals. However, we face the following challenges while extracting the tone from smartphone audio signals. (a) Smartphone microphones are omnidirectional, and they capture environmental noise along with the human voice. Moreover, the devices are heterogeneous. MFCC fails in the face of the noisy environment and with device heterogeneity [36]. (b) The device heterogeneity is in general handled through various energy-based normalization techniques [33, 31, 37], however they fail for smartphone microphones due to the nonlinearity gain of amplifiers and the presence of automatic gain control (AGC) circuits1. (c) As MFCC is mostly followed by a supervised scheme, the approach may require the voice samples from each user for the correct identification of pitch and tone. However, most of the members in the instantaneous groups are new and appear for the first time. Hence, pre-training is impossible in most of the scenarios.

In this paper, we apply Complex Cepstrum (CCEP) to perform tone extraction. CCEP of a signal is computed as follows.


where FT(.) is the Fourier transform, IFT(.) is the inverse Fourier transform and

. The imaginary part uses complex logarithmic function, and is an integer which is required to properly unwrap the imaginary part of the complex log function. CCEP uses the log compression of the power spectrum, and therefore is less affected by the environmental noise, the nonlinearity of amplifiers and the effect of AGC circuits. To extract the tone from an audio signal, we segment the signal into one second units, and then compute the CCEP for the audio segments. The CCEP for segment from subject is denoted as . These CCEP values for all the subjects are then used for tone similarity measure, as discussed next.

3.3 Computing Acoustic Context Feature

We compute cross-correlation between the CCEP values to measure tone similarity, thereby high and low cross-correlation indicates similar and dissimilar acoustic context between the pair of subjects, respectively. Let and denote the CCEP for segment from two different subjects and . We compute the segment wise cross-correlation between and as , and then average it over the time span . This audio cepstrum cross-correlation is used as the acoustic context similarity for subject pair and during time duration .

Fig. 9: Audio Similarity Variation in Different Scenarios

To demonstrate the role of tone similarity to compute acoustic context of meeting groups, we consider the groups and formed in outdoor roadside where the group scenario shown in Figure 2. Subjects , , and , , form the Group and , respectively. In the first scenario, subject in group is the dominating speaker, whereas members of Group are mostly silent. Figure 9a shows the pairwise context similarity between the individual subject and the dominating speaker (of group ). We observe that subjects in group (say, and ) exhibit higher similarity with dominating speaker as compared with the members of the group (say, subjects in and ). Next, we consider two dominating speakers and in two respective groups and . We compute the context similarity between any pair of (non-speaking) subjects. In Figure 9b we observe that members belonging to the same group (say and in Group and and in Group ) exhibit higher context similarity compared to non-group pairs. Precisely, the context similarity between the intragroup members is substantially higher (close to ) than the intergroup members (close to ). This result indicates that acoustic context within a single group exhibits substantial similarity.

We also investigate the impact of location of a subject on her acoustic context. We set up two groups and , m apart in the outdoor environment, with two dominating speakers namely and respectively. We consider one moving subject , initially inside the (from Table I) and walks towards group (it takes around sec to reach group from ). Figure 11 shows the variation in the acoustic context similarity between the subject and the dominating speaker over time. We observe that, when the subject is in group , the context similarity between the and the subject is high as compared with the . The reverse behaviour is noticed at the end of the experiment when the subject reaches . However, the context is confusing as the subject located in the middle of both the groups.

Fig. 10: Audio Cross-correlation variation overtime with the moving
Fig. 11: ECDF of Cepstrum Cross-Correlation

Finally, we perform an overall evaluation, considering all the meeting groups formed in the pilot study. In Figure 11, we plot the empirical cumulative distribution (ECDF) of the acoustic context similarity for the pair of subjects. We observe that when the subject pair is in the same group, the acoustic context similarity is high. On the contrary, when one subject is outside the group, context similarity exhibits low value. This establishes the fact that tone similarity, computed from cepstrum cross-correlation, reflects the acoustic context of a group and more importantly, the acoustic context within a single group exhibits substantial similarity.

The aforesaid methodology of extracting acoustic context from smartphone microphone has three broad advantages. First, as the feature is extracted from the dominating tone in an audio signal (captured by cepstrum), it is sufficient if at least one subject in a group talks for a duration. The method can be utilized to consider subjects who belong to a group, but do not prefer to interact (consider a conference presentation); as long as some other subject from that group speaks. Second, the methodology does not violate the privacy of individual subjects. We extract only the tone information from the captured audio signal and do not leverage the exact conversation. Third, the proposed model is unsupervised. Therefore, there is no need of pre-training of the tone information of the group members.

4 Design of GroupSense

MeetSense is an unsupervised framework for detecting meeting groups based on subject proximity and acoustic context. Figure 12 shows the flow outline of the MeetSense framework. First of all, the sensor logger module records the microphone data along with the proximity indicators followed by the pairwise feature computation. Leveraging on the proximity and acoustic context features, we develop MeetSense for meeting group detection.

Fig. 12: MeetSense Model [: Proximity Feature, : Acoustic Context Feature]

4.1 Feature Construction

In this module, we compute acoustic context similarity between the subject pair and at time from the collected microphone log; the detailed computation procedure of acoustic context feature has already been explained in section 3. The pairwise proximity similarity feature at time can be extracted from any of the state of the art techniques [19, 10]. Now considering a subject pair and , we need to compute the aggregated features and respectively for time duration . One simplest way of aggregation is computing the mean and from the instantaneous features and respectively for time duration . However, the signal sample collected from proximity indicator and microphone may suffer from sensitivity and fluctuations. Additionally, the audio signals can also get muffled by obstacles, clothing materials, and are also impacted by the interference. Evidently, the colluded mean features, computed from all the feature points and for the time duration , may not provide a clear indication of feature similarities between the subject pair and . Hence, we compute the refined mean features and , by eliminating the low-frequency noise component. Here we split all the features points (say, for proximity feature

) into two clusters (via k-means clustering, with

). Eliminating the minor cluster as the noisy component, we compute the mean from the feature points in the major cluster (see Algorithm 1). However, in case of , we compute the mean considering all the feature points in the single cluster. Similarly, we compute the refined mean acoustic context feature from Algorithm 1.


2:if  then Single Cluster Scenario
5:       if  then Major Cluster Scenario
7:       else Major Cluster Scenario
9:       end if
10:end if
Algorithm 1 Feature_Construction

4.2 Model Development

Finally, leveraging on the aforementioned features, we develop an unsupervised model for meeting group detection. The model executes Part A (Algorithm 3) or Part B (Algorithm 4) depending on the availability of the location information (WiFi, Bluetooth, GPS etc). If the subject possesses location information, the model exploits both proximity as well as acoustic features in Part A. Otherwise, the model only relies on the acoustic information in Part B. The outcome of the model is all the groups detected by both the individual parts. The outline of the model is described in Algorithm 2.


1:if  then
3:end if
5:if  then Proximity Available Scenario
6:        ProximityAvailable_Function ()
7:end if
8:if  then Proximity Not Available Scenario
9:        ProximityNotAvailable_Function ()
10:end if
Algorithm 2 MeetSense: Group_Dection_Algorithm

Part A: In this part, we first attempt to extract the cluster of co-locating subjects only based on the pair-wise proximity similarity. If we identify a highly cohesive cluster based on proximity only, we consider as a potential meeting group and execute the second step. In the second step, we leverage only on the acoustic context features to detect meeting group(s) from the identified proximity clusters . However, if we identify moderately cohesive cluster from the proximity features, the model abandons the cluster , considering proximity as a critical albeit weak signal, and moves to the third step. In the third step, we combine both the proximity and acoustic context similarity features together to detect cohesive cluster on the complete proximity available population . If exhibits high cohesivity, we assert the cluster as the meeting group. Poor cohesivity in any step rejects the existence of any group in the population. The overall procedure is illustrated in Algorithm 3. In the following, we introduce cohesive cluster detection, which is the core of the model.

Detection of cohesive cluster: Consider a weighted network , where is a subject and denotes the weighted link between the subject pair and . We apply community detection algorithm [39] on to obtain a partition on population . Essentially the community detection algorithm partitions the network into communities ensuring dense connections within a community and sparser connections between communities. We consider the detected community as a cluster of population . The cohesivity of the partition can be measured with modularity index as , which reflects the fraction of the links that fall within a given community, compared to the expected fraction if links are distributed at random [39]. The , , , and represent the sum of all of the edge weights in the network, sum of the edge weight attached to node , the community of node , and delta function, respectively. Notably, modularity of a weighed fully connected graph becomes if all the nodes form a single large community [40]. In this paper, we apply Walktrap algorithm [20]; however, our methodology is not sensitive to any specific (weighted) community detection algorithm.

Algorithm 3 comprises of the following three steps.

Step 1: We construct a complete proximity graph where denotes the complete proximity available population and is a link between the subject pair and weighted by the proximity feature computed over the time . We apply community detection algorithm on the proximity graph to discover the cluster with modularity . If the is above a threshold , we consider as the candidate meeting group and move to step 2. If falls below a threshold , we reject the existence of any meeting group in population . Otherwise, we move to step 3.

Step 2: We construct the complete acoustic context graphs where links between subject pair and . Essentially, in , the link weight depicts the acoustic context similarity over the time . Similar to step 1, we apply the community detection on to discover the cluster with modularity . If the is above a threshold , we confirm as the detected meeting groups. Otherwise, we reject the existence of meeting groups in population .

Step 3: We construct a complete proximity-acoustic context graph where links between subject pair and weighted by . Essentially, in , the link weight carries the information from both acoustic context and proximity feature. Similar to step 1, we apply the community detection on to discover the cluster with modularity . If the is above a threshold , we confirm as the detected meeting groups. Otherwise, we reject the presence of any group in population .


1:Compute Feature Generation
2: Feature_Construction ()
3: Community_Detection ()
4:if  then Proximity Dominating Scenario
5:        Feature_Construction ()
6:        Community_Detection ()
7:       if  then Proximity & Audio Influence Scenario
9:       else Proximity Influence & Audio Insignificance Scenario
10:             Failure
11:       end if
13:       if  then
14:              Feature_Construction ()
15:              Community_Detection () Weighted Features
17:             if  then Proximity Confused & Audio Influence Scenario
19:             else Proximity Confused & Audio Insignificance Scenario
20:                    Failure
21:             end if
22:       else Proximity Insignificance Scenario
23:             Failure
24:       end if
25:end if
Algorithm 3 ProximityAvailable_Function

Part B: Due to the unavailability of location data, in this part, we completely rely on the acoustic context similarity between the subjects. We first construct a complete acoustic context graph where links between subject pair and . Essentially, in , the link weight carries the information of the acoustic context feature over the time . Similar to part A, we apply the community detection on to discover the clusters with modularity . If the is above a threshold , we confirm as the detected meeting groups. Otherwise, we reject the existence of meeting groups in population . The outline of the overall mechanism is portrayed in Algorithm 4.


1:Compute Feature Generation
2: Feature_Construction ()
3: Community_Detection()
4:if  then Audio Influence Scenario
6:else Audio Insignificance Scenario
7:       Failure
8:end if
Algorithm 4 ProximityNotAvailable_Function

5 Performance Evaluation

We evaluate MeetSense by developing a smartphone-based application and deploying it over IIT Kharagpur campus spreading square kilometres, consisting of administrative blocks, approximately academic departments along with campus residential, hostels and market areas. We first discuss the implementation of MeetSense followed by the field study and performance comparison with different baselines.

5.1 Field Study and Data Collection

The data collection for MeetSense is done initially through an Android app, DataGatherer, which has been launched over the smartphones of subjects consisting of undergraduate and postgraduate students, summer interns, research scholars and faculties of the institute. In our implementation of MeetSense, we have considered , that means if an interaction continues for at least minutes, we consider it as a group. Nevertheless, this is an application specific tunable parameter. We have used different models of smartphones, where costs per phone range from USD to USD approximately. We primarily gather WiFi (BSSID and signal strength) and audio data from smartphones through DataGatherer which sends the data to a central server. The WiFi data is used for proximity detection based on existing methodologies [10, 19], and then audio data is used to detect the groups among the subjects in proximity. The app scans the available WiFi access points once in a minute time interval, and continuous audio signals are tracked at a sampling rate of kHz for a minute time span followed by an interval of three minutes. Moreover, we have discarded the details of the access points having the signal strength less than dBm which is the minimum signal strength for basic connectivity333 (Accessed on Apr 12, 2018). The data has been collected for approximately six months.

We collect the ground truth meeting group information from the participants for validation. In ground truth data collection, a questionnaire app periodically probes from the participants regarding the (a) start time of the meeting, (b) end time of the meeting, (c) meeting venue and (d) details of the other participants of the meeting. In some cases where a participant misses to provide the ground truth information, we validate the detected meeting groups from the participants by forwarding an email at every two hours of each day. Based on the field study collected data, we identify seven typical meeting group scenarios, which repeatedly occurred (at least once in a week) during the six months field study. These situations are highlighted keeping in mind the critical conditions of group formation that were developed in the pilot study (Section 2

); thus reflect realistic meeting group scenarios with high probability. We evaluate the performance of

MeetSense and compare it with other baselines considering these typical scenarios, as well as different scenarios observed from the collected data. These scenarios are as follows.

S1 (Indoor: Two groups at neighbouring rooms): subjects attend a lecture in classroom C-, and subjects have another meeting in the FV Lab opposite to C- at the same instance of time.

S2 (Indoor: Three groups at different rooms at the same department): subjects interact in the faculty office in the second floor, subjects are in a meeting at the departmental library opposite to that faculty office, and subjects are in another meeting at the SMR Lab in the first floor.

S3 (Outdoor: Cafeteria interactions): Two different groups at the cafeteria, one with subjects in front of the cafeteria and another one with subjects at the back of the cafeteria.

S4 (Indoor: Large single group): subjects attend a presentation at the departmental conference room.

S5 (Indoor: Two different groups at a large lab): subjects meet at cubicle K- and another subjects meet in the cubicle K- of the SMR Lab.

S6 (Indoor: Two roaming groups): subjects together and subjects together roam around the corridor of the department, and moves from one room to another, forming two non-static groups.

S7 (Outdoor: Two roaming groups): subjects together and subjects together roam within the campus maintaining a certain distance from each other, forming two non-static groups.

5.2 Preprocessing – Proximity Computation for Group Detection

As we discussed earlier, group detection first requires to find out the subjects in proximity, and in MeetSense, we utilize the existing proximity detection mechanisms that have been well studied in the literature. We focus on the two approaches of proximity detection based on WiFi data, as follows.

(a) Supervised learning with WiFi-based proximity sensing (

SLWP[10]: Sapiezynski et al. developed a WiFi access point based supervised

proximity detection mechanism, where Bluetooth data is considered as the ground truth. In this approach, a set of WiFi-based features has been computed, such as overlapping access points, signal strength from different access points etc., and then a support vector machine (SVN) is used to classify whether two subjects are in proximity or not.

(b) Next2Me [19]: This is a smartphone-based system for capturing social interactions within close proximity users. Next2Me uses WiFi signal information for measuring the pairwise co-located Manhattan distance between the users, and then a threshold over the distance function is used to find out the subjects in proximity. It can be noted that this is an unsupervised approach.

5.3 Baselines for Audio Based Interaction Detection

We have analyzed the performance of MeetSense with following three baselines, which utilize audio signals for acoustic context detection. We use the proximity followed by audio based acoustic context to detect various meeting groups.

(a) Next2Me [19]: After determining the subjects in proximity, Next2Me utilizes Jaccard similarity over top audio frequencies to capture the audio fingerprints of various subjects. Finally, they generate social community by applying Louvain community detection algorithm.

(b) AudioMatch [13]: Casagranda et al. implemented a smartphone based group detection system based on the joint usage of GPS and audio fingerprints. The GPS information is used for filtering out nearby devices. On top of the GPS based clusters, the audio module is executed for identifying the groups. AudioMatch uses short time Fourier transform (STFT) with overlapping Hamming window. Finally, it computes the Hamming distance between the pair of devices for detecting the nearby pairs.

Next, we discuss the experimental procedure by combing the WiFi-based proximity detection and audio based acoustic context detection together.

Next2Me GroupSense AudioMatch Next2Me GroupSense AudioMatch
S1 1.0000 0.1124 1.0000 0.2879 0.7273 0.0000 1.0000 0.0412 1.0000 0.2907 0.7273 0.0000
S2 0.9000 0.2030 1.0000 0.1760 0.6667 0.0000 0.9000 0.2030 1.0000 0.1760 0.6667 0.0000
S3 0.5333 0.1261 1.0000 0.3642 0.7273 0.0000 0.5333 0.1261 1.0000 0.3642 0.7273 0.0000
S4 0.8326 0.0772 1.0000 0.0000 1.0000 0.0000 0.8326 0.0772 1.0000 0.0000 1.0000 0.0000
S5 0.8571 0.0732 1.0000 0.3801 0.7273 0.0000 0.8571 0.0732 1.0000 0.3801 0.7273 0.0000
S6 1.0000 0.0000 1.0000 0.0000 1.0000 0.0000 1.0000 0.0000 1.0000 0.0000 1.0000 0.0000
S7 0.5833 0.1942 0.6500 0.0976 0.8333 0.0000 0.5833 0.1942 0.6500 0.0976 0.8333 0.0000
ALL 0.7971 0.0866 0.9421 0.2114 0.8212 0.0000 0.7971 0.0826 0.9421 0.2116 0.8212 0.0000
TABLE III: Performance Comparison

5.4 Experimental Procedure

For comparing the performance of MeetSense under the different environment with different baselines, we consider following combinations of proximity (P) and acoustic context (I) detection mechanisms. It can be noted that MeetSense primarily focuses on capturing acoustic context, whereas the proximity module is borrowed from existing methodologies. The different combinations of proximity and acoustic context detection mechanisms used in our experiments are as follows.

I. SLWP (P) + Next2Me (I): In this arrangement, we extract the pairwise proximity information from SLWP, and the outcome is directly fed to the Next2Me audio model for group detection.

II. SLWP (P) + MeetSense (I): This arrangement uses the pairwise proximity information from SLWP. Then, MeetSense audio centric context detection is applied on top of the proximity outcome.

III. SLWP (P) + AudioMatch (I): In this arrangement, we apply SLWP for pairwise proximity detection. After that, AudioMatch is applied on the outcome of the proximity clusters for detecting the pairwise acoustic context from the audio signals. It can be noted that we have not used GPS for proximity detection as used in AudioMatch, as GPS gives very poor signal in the indoor scenarios. However, the audio module is used as it is, and finally, the community detection algorithm is used for group detection.

IV. Next2Me (P) + Next2Me (I): This arrangement is analogous with the actual Next2Me system, where both the WiFi based proximity detection and the audio based acoustic context detection, as done in Next2Me, are used for group detection.

V. Next2Me (P) + MeetSense (I): In this arrangement, we compute the proximity-based pairwise distance following Next2Me proximity module. The pairwise similarity is computed by reversing the pairwise distance value. Then, we apply MeetSense Feature Construction algorithm 1 followed by community detection module on the pairwise similarity value. Finally, MeetSense acoustic context module is employed on top of the proximity outcome.

VI. Next2Me (P) + AudioMatch (I): This arrangement uses the proximity information from Next2Me like the previous setup. After that, AudioMatch is applied on the outcome of the proximity clusters. The pairwise acoustic context information is finally fed to the community detection algorithm for meeting group detection.

5.5 MeetSense Performance

We first evaluate the overall performance of MeetSense in terms of -Score [41] defined as follows. Let and be the sets of meeting groups in the ground truth data and the ones detected by MeetSense, respectively. Then where and . This parameter captures the accuracy of the detected group in terms of membership overlap with ground truth for the meeting duration . Now, to obtain the final accuracy of MeetSense considering all the detected meeting groups, we compute the average -Score as .

Table III summarizes the performance of MeetSense in terms of -Score and modularity () for seven representative scenarios as well as for all the observed scenarios combined. We experimentally set up the model thresholds () based on the best performance of the overall scenarios. The modularity index indicates the cohesiveness of the detected groups; hence even a low -Score with high modularity contributes more to identify maximum participants in a meeting group. Although MeetSense performs marginally worse for certain scenarios, like during outdoor mobile groups due to the high environmental noise, we observe that accuracy is more than for most of the cases.

5.5.1 Baseline Comparison

As mentioned earlier, we have set up six different arrangements proposed as well as existing schemes for investigating the performance at different scenarios. Table III compares the performance of MeetSense with AudioMatch and Next2Me with two different proximity schemes. For all the three scenarios when SLWP is used for proximity measure, we observe that MeetSense outperforms the other baselines. Although Next2Me uses audio based features to capture the social interaction among the subjects (which is similar to the acoustic context for MeetSense), it uses Jaccard similarity among top audio frequencies, which is susceptible to environmental noise. For example, in an outdoor environment, the sound frequencies originated from external entities, such as moving vehicles, can fall within the top frequency components. As a consequence, we observe that although Next2Me manages to perform well in indoor scenarios, it poorly performs in the outdoor environment. On the other side, AudioMatch applies hamming distance measure of the logarithmic amplitude of the audio signal for suppressing the noise component. Although this scheme works for artificially generated Gaussian noise, it poorly performs in the presence of real environmental noise. The impact is visible in the outdoor scenarios.

Next, we have tested the scenarios with three baselines for acoustic context measurement along with Next2Me proximity measure. The results of these are apparently similar to the results from the scenarios with SLWP proximity measure, except for the proximity dominant scenario in MeetSense. The similar results also claim that the audio features are more dominant compared to the proximity features for meeting group detection. Moreover, this validates the importance of Algorithm 4 in the overall meeting group detection mechanism.

5.5.2 Robustness of Acoustic Context Measure

For investigating the variations in the performance of different schemes, we report the box-plot of the pairwise feature similarity values for the acoustic context, shown in Figure 13. The box plot depicts that there are significant mean differences between the various schemes. In the box plot, the medians for different schemes are shown in red lines. Focusing on the upper and the lower halves from the median, the results show that MeetSense captures significant variations in the pairwise similarity between the subjects. As we consider multiple meeting group scenarios, the variation of the pairwise similarity between the subjects is justified. It can be noted that the median is biased towards the lower values because the pairwise feature similarity becomes very close to zero whenever the two subjects in the pair are from different groups. However, a wide variation of similarity values greater than is observed when both the subjects are in the same group. On the contrary, the results for other baselines depict that Next2Me and AudioMatch show the minimal difference in the upper and the lower halves from the median. Therefore, the constructed feature is incapable of distinguishing between the acoustic context when the subjects are in the same group or different groups. Hence, the

score significantly drops for those baselines. Additionally, we also observe that the median value is closer to the first quartile. As we capture the proximity and audio signatures of the subjects in various environments, the similarity values between each pair of subjects significantly varies over the different meeting groups, causing the dense zone towards the lower halves from the median. The wide variation of the pairwise similarity values in different groups further interprets that the simple thresholding based scheme is not suitable for detecting various types of meeting groups in the diverse environment. Hence, it justifies the requirement of the complex

MeetSense scheme (Algorithm 2).

Fig. 13: Mean difference of similarity obtained through different methodologies

5.5.3 Dissecting the Methodologies

Next, we look into the performance of the various competing methodologies by exploring their internals. From the above experiments, we observe that baselines perform poorly for the scenario . Therefore, we further study the top frequency based similarity, thresholding based hamming distance measure, and cepstrum similarity for that scenario. From Figure 13(a), we found that audio cross-correlation for ‘same group’ (target subjects are within a group) and ‘different groups’ (target subjects are from different groups) pairs of subjects are more distinct in MeetSense as compared with the other two baseline methods. The outdoor environment like the cafeteria (scenario ) are noisy due to the presence of the non-member group voice and noise from the environment. As Next2Me considers only the top frequency components (as per our implementation ), it unknowingly considers those frequencies, resulting in the similar audio correlation values for ‘same group’ and ‘different groups’. On the other side, AudioMatch compares the logarithmic amplitude of STFT of the audio signal with its neighbouring points to generate -bit fingerprint. Therefore, the -bit fingerprint generation completely relies on the center comparing amplitude value. If the center value is corrupted due to the environmental noise, the entire -bit fingerprint is prone to be corrupted. Those spurious fingerprints are further used for computing the Hamming distance between the pair of subjects, resulting in identical behaviour for the audio correlation values in ‘same group’ and ‘different groups’. As MeetSense considers cepstrum containing the tone information for computing the audio correlation, the correlation values are more close to one for ‘same group’ and close to zero for ‘different groups’. Therefore, the audio features of MeetSense can distinguish this scenario. Consequently, although Next2Me and AudioMatch fail to separate out the groups based on the audio features, MeetSense can correctly differentiate the groups.

We further evaluate Next2Me, AudioMatch and MeetSense in varying noise environment. As simulating complete random noise is nearly impossible, we generate Gaussian noise at different levels and superimpose the noise with the captured audio signal. Figure 13(b) shows that MeetSense is more noise resistant than Next2Me, whereas AudioMatch is as noise resistant as MeetSense. Analogous to noisy environmental scenarios, Next2Me performs poorly in the presence of statistically generated Gaussian noise due to the improper selection of top frequencies. In case of AudioMatch the generation of -bit fingerprint causes the drop though it is much less prone to noise as compared to Next2Me because of considering the logarithmic amplitude of STFT.

(a) Audio correlation
comparison for scenario
(b) Effect of Noise in
MeetSense and Next2Me
Fig. 14: Performance analysis at different environments

Next, we compare the three acoustic context detection mechanisms in terms of computational resource requirements, as shown in Figure 15. We measure these performance statistics in a standard Linux (Kernel version: 4.4.0) based workstation (Dell Precision Tower 7810) using the free command to obtain the primary memory consumption of the different methodologies. We compute the total execution time and the overall memory consumption during the execution of the three methods. We observe that (i) MeetSense takes very less time per iteration during the computation process compared to Next2Me and AudioMatch (Figure 15a); (ii) the memory consumption for MeetSense is less than Next2Me (Figure 15b). MeetSense enjoys the benefit of lower resource consumption primarily because it computes only cepstrum component for a few segment over the entire interaction time, whereas Next2Me uses several windowing operations along with smoothing and FFT computations. AudioMatch calculates audio spectrogram using short time fourier transform with a highly overlapping hamming window, causing higher elapsed time than that of MeetSense. In a nutshell, we observe that MeetSense can detect various meeting groups generically and in a device independent way, however, can provide better group detection accuracy with less resource usage compared to the baseline mechanisms.

Fig. 15: Performance in terms of computational cost for the unsupervised mechanisms

5.5.4 MeetSense Internals

In this subsection, we discuss the importance of the modularity value in MeetSense, and how the proximity and acoustic context features improve the modularity of the proposed group detection mechanism.

We plot the -Score with respect to the modularity, as shown in Figure 16a. We observe that the -Score converges to when the modularity is more than . Hence a group is detected with high accuracy when the cohesiveness is also high. This indicates the importance of modularity index in MeetSense. Therefore, the community detection algorithm used in MeetSense tries to optimize the modularity in successive iterations.

Fig. 16: MeetSense insights

In this line, Figure 16b highlights the importance of Step 2 of MeetSense model, where we plot -Score with respect to the weight () of the audio feature. The figure indicates that the maximum modularity is obtained when both the proximity and the acoustic context attain non-zero weights, indicating that both the features are important for correct detection of meeting groups. However, the importance of acoustic context is more prominent over the proximity feature.

Next, we look into the convergence property of MeetSense. As mentioned earlier, modularity of a weighted fully connected graph converges to zero when all the nodes form a single large community [40]. Accordingly, the group detection algorithm converges with two cases of the modularity () value – (a) , when there are multiple groups in the population of subjects () and (b) , when there is a single large group consisting of all the subjects from the population (). Figure 16c plots the change in modularity value with respect to the number of iterations performed in the algorithm, for these two cases. We observe that for a group corresponding to scenario , we get the maximum modularity close to with iterations, whereas for a group corresponding to scenario , the modularity starts with a negative value and converges to with iteration .

6 Conclusion

In this paper, we have developed MeetSense, a smartphone based light-weight methodology to infer various meeting groups by sensing the acoustic context around the users in proximity. From the pilot study, we have observed that although audio levels captured by a smartphone give a good indication of the acoustic context of the environment, a significant audio pressure from speakers of the nearby groups also gets captured due to the omnidirectional nature of smartphone microphones. We have developed a novel unsupervised methodology to process audio signals to capture the context and used the concept of cohesivity from network science to identify the groups based on context information. The implementation and thorough testing of MeetSense shows that it can significantly improve group detection accuracy compared to other baselines, and the method is independent of scenarios or devices used to capture signals. However, our understanding is that MeetSense can perform well when the underline groups are sufficiently cohesive; it may fail in the scenarios when multiple groups are overlapped in space, or a group is spatially overlapped with individuals who are not part of that group, for example, small groups in a crowded space. Nevertheless, the proposed methodology has the advantages of device independence, unsupervised modelling and light-weight computation, which can be utilized to develop the wide range of applications that require user group identification and group behaviour analysis.


  • [1] K. A. McComas, “Citizen satisfaction with public meetings used for risk communication,” Journal of Applied Communication Research, vol. 31, no. 2, pp. 164–184, 2003.
  • [2] T. Clark, “Teaching students to enhance the ecology of small group meetings,” Business Communication Quarterly, vol. 61, no. 4, pp. 40–52, 1998.
  • [3] K. McComas, L. S. Tuite, L. Waks, and L. A. Sherman, “Predicting satisfaction and outcome acceptance with advisory committee meetings: The role of procedural justice,” Journal of Applied Social Psychology, vol. 37, no. 5, pp. 905–927, 2007.
  • [4] B. A. Reinig and B. Shin, “The dynamic effects of group support systems on group meetings,” Journal of Management Information Systems, vol. 19, no. 2, pp. 303–325, 2002.
  • [5] A. P. Horan, “An effective workplace stress management intervention: Chicken soup for the soul at employee groups,” Work, vol. 18, no. 1, pp. 3–13, 2002.
  • [6] J. A. Allen, T. Beck, C. W. Scott, and S. G. Rogelberg, “Understanding workplace meetings: A qualitative taxonomy of meeting purposes,” Management Research Review, vol. 37, no. 9, pp. 791–814, 2014.
  • [7] K. Jayarajah, Y. Lee, A. Misra, and R. K. Balan, “Need accurate user behaviour?: pay attention to groups!” in Proceedings of the ACM conference on Pervasive and Ubiquitous Computing, 2015, pp. 855–866.
  • [8] M. B. Gilboy, S. Heinerichs, and G. Pazzaglia, “Enhancing student engagement using the flipped classroom,” Journal of nutrition education and behavior, vol. 47, no. 1, pp. 109–114, 2015.
  • [9]

    J. Weppner and P. Lukowicz, “Bluetooth based collaborative crowd density estimation with mobile phones,” in

    Proceedings of the IEEE conference on Pervasive computing and communications, 2013, pp. 193–200.
  • [10] P. Sapiezynski, A. Stopczynski, D. K. Wind, J. Leskovec, and S. Lehmann, “Inferring person-to-person proximity using wifi signals,” Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, vol. 1, no. 2, p. 24, 2017.
  • [11] H. Wang, S. Sen, A. Elgohary, M. Farid, M. Youssef, and R. R. Choudhury, “No need to war-drive: Unsupervised indoor localization,” in Proceedings of the IEEE conference on Mobile systems, applications, and services, 2012, pp. 197–210.
  • [12] C. Wu, J. Xu, Z. Yang, N. D. Lane, and Z. Yin, “Gain without pain: Accurate wifi-based localization using fingerprint spatial gradient,” Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, vol. 1, no. 2, p. 29, 2017.
  • [13] P. Casagranda, M. L. Sapino, and K. S. Candan, “Audio assisted group detection using smartphones,” in Proceedings of IEEE Conference on Multimedia & Expo Workshops, 2015, pp. 1–6.
  • [14] D. Santani and D. Gatica-Perez, “Loud and trendy: Crowdsourcing impressions of social ambiance in popular indoor urban places,” in Proceedings of the ACM conference on Multimedia, 2015, pp. 211–220.
  • [15] F. Calabrese, L. Ferrari, and V. D. Blondel, “Urban sensing using mobile phone network data: a survey of research,” ACM Computing Surveys, vol. 47, no. 2, p. 25, 2015.
  • [16] R. Sen, Y. Lee, K. Jayarajah, A. Misra, and R. K. Balan, “Grumon: Fast and accurate group monitoring for heterogeneous urban spaces,” in Proceedings of the ACM conference on Embedded Network Sensor Systems, 2014, pp. 46–60.
  • [17] M. Azizyan, I. Constandache, and R. Roy Choudhury, “Surroundsense: mobile phone localization via ambience fingerprinting,” in Proceedings of the ACM conference on Mobile computing and networking, 2009, pp. 261–272.
  • [18] S. Davis and P. Mermelstein, “Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 28, no. 4, pp. 357–366, 1980.
  • [19] J. Baker and C. Efstratiou, “Next2me: Capturing social interactions through smartphone devices using wifi and audio signals,” in EAI Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, 2017.
  • [20] P. Pons and M. Latapy, “Computing communities in large networks using random walks,” Journal of Graph Algorithms and Applications, vol. 10, no. 2, pp. 191–218, 2006.
  • [21] K. Chintalapudi, A. Padmanabha Iyer, and V. N. Padmanabhan, “Indoor localization without the pain,” in Proceedings of the ACM conference on Mobile computing and networking, 2010, pp. 173–184.
  • [22] H. Abdelnasser, R. Mohamed, A. Elgohary, M. F. Alzantot, H. Wang, S. Sen, R. R. Choudhury, and M. Youssef, “Semanticslam: Using environment landmarks for unsupervised indoor localization,” IEEE Transactions on Mobile Computing, vol. 15, no. 7, pp. 1770–1782, 2016.
  • [23] H. Aly, A. Basalamah, and M. Youssef, “Accurate and energy-efficient gps-less outdoor localization,” ACM Transactions on Spatial Algorithms and Systems, vol. 3, no. 2, p. 4, 2017.
  • [24] J. Paek and R. Kim, Joongheon Hand Govindan, “Energy-efficient rate-adaptive gps-based positioning for smartphones,” in Proceedings of the ACM conference on Mobile systems, applications, and services, 2010.
  • [25] T. M. T. Do and D. Gatica-Perez, “Groupus: Smartphone proximity data and human interaction type mining,” in Proceedings of the Annual IEEE Symposium on Wearable Computers, 2011, pp. 21–28.
  • [26] H. Hong, C. Luo, and M. C. Chan, “Socialprobe: understanding social interaction through passive wifi monitoring,” in Proceedings of the ACM Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, 2016, pp. 94–103.
  • [27] Y. Lee, C. Min, C. Hwang, J. Lee, I. Hwang, Y. Ju, C. Yoo, M. Moon, U. Lee, and J. Song, “Sociophone: Everyday face-to-face interaction monitoring platform using multi-phone sensor fusion,” in Proceeding of the ACM conference on Mobile systems, applications, and services, 2013, pp. 375–388.
  • [28] S. Zhang, Y. Zhu, and A. Roy-Chowdhury, “Tracking multiple interacting targets in a camera network,” Computer Vision and Image Understanding, vol. 134, pp. 64–73, 2015.
  • [29] N. Eagle and A. S. Pentland, “Reality mining: sensing complex social systems,” Personal and ubiquitous computing, vol. 10, no. 4, pp. 255–268, 2006.
  • [30] R. Friedman, A. Kogan, and Y. Krivolapov, “On power and throughput tradeoffs of wifi and bluetooth in smartphones,” IEEE Transactions on Mobile Computing, vol. 12, no. 7, pp. 1363–1376, 2013.
  • [31] Z. Liu, Z. Zhang, L. W. He, and P. Chou, “Energy-based sound source localization and gain normalization for ad hoc microphone arrays,” in Proceedings of the IEEE Conference on Acoustics, Speech and Signal Processing, vol. 2, 2007, pp. II–761–764.
  • [32] A. De Cheveigné and H. Kawahara, “YIN, a fundamental frequency estimator for speech and music,” The Journal of the Acoustical Society of America, vol. 111, no. 4, pp. 1917–1930, 2002.
  • [33] C. Xu, S. Li, G. Liu, Y. Zhang, E. Miluzzo, Y.-F. Chen, J. Li, and B. Firner, “Crowd++: Unsupervised speaker count with smartphones,” in Proceedings of the ACM Joint Conference on Pervasive and Ubiquitous Computing, 2013, pp. 43–52.
  • [34] T. Song, X. Cheng, H. Li, J. Yu, S. Wang, and R. Bie, “Detecting driver phone calls in a moving vehicle based on voice features,” in Proceedings of the IEEE conference on Computer Communications, 2016, pp. 1–9.
  • [35] W. Passchier-Vermeer and W. F. Passchier, “Noise exposure and public health,” Environmental Health Perspectives, vol. 108, pp. 123–131, 2000.
  • [36] M. L. Narayana and S. K. Kopparapu, “Effect of noise-in-speech on mfcc parameters,” in Proceedings of the WSEAS conference on signal, speech and image processing, and WSEAS conference on Multimedia, internet & video technologies, 2009, pp. 39–43.
  • [37] M. Chen, Z. Liu, L. W. He, P. Chou, and Z. Zhang, “Energy-based position estimation of microphones and speakers for ad hoc microphone arrays,” in Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2007, pp. 22–25.
  • [38] M. Guggenberger, M. Lux, and L. Böszörmenyi, “An analysis of time drift in hand-held recording devices,” in International Conference on Multimedia Modeling, 2015, pp. 203–213.
  • [39] A. Lancichinetti, S. Fortunato, and F. Radicchi, “Benchmark graphs for testing community detection algorithms,” Physical review E, vol. 78, no. 4, p. 046110, 2008.
  • [40] M. E. Newman, “Modularity and community structure in networks,” Proceedings of the national academy of sciences, vol. 103, no. 23, pp. 8577–8582, 2006.
  • [41] L. R. Dice, “Measures of the amount of ecologic association between species,” Ecology, vol. 26, no. 3, pp. 297–302, 1945.