- 1IN3 – Universitat Oberta de Catalunya, Castelldefels, Spain
- 2Universität Klagenfurt, Universitätszentrum für Frauen- und Geschlechterstudien, Klagenfurt, Austria
Sensor-based data are becoming increasingly widespread in social, behavioral, and organizational sciences. Far from providing a neutral window on “reality,” sensor-based big-data are highly complex, constructed data sources. Nevertheless, a more systematic approach to the validation of sensors as a method of data collection is lacking, as their use and conceptualization have been spread out across different strands of social-, behavioral-, and computer science literature. Further debunking the myth of raw data, the present article argues that, in order to validate sensor-based data, researchers need to take into account the mutual interdependence between types of sensors available on the market, the conceptual (construct) choices made in the research process, and the contextual cues. Sensor-based data in research are usually combined with additional quantitative and qualitative data sources. However, the incompatibility between the highly granular nature of sensor data and the static, a-temporal character of traditional quantitative and qualitative data has not been sufficiently emphasized as a key limiting factor of sensor-based research. It is likely that the failure to consider the basic quality criteria of social science measurement indicators more explicitly may lead to the production of insignificant results, despite the availability of high volume and high-resolution data. The paper concludes with recommendations for designing and conducting mixed methods studies using sensors.
Introduction
Sensors are becoming increasingly widespread in social science research. Speculations regarding the potential applications and implications of “big data” are relatively frequent across the entire spectrum of scientific disciplines. However, despite the growing number of debates on the challenges and prospects of big data, and notwithstanding dedicated discussions on “digital methods” (Rogers, 2013) and “digital sociology” (Lupton, 2015; Marres, 2017), more targeted accounts of the methodological implications of using sensors are lacking. This is especially true in the case of mixed methods research (MMR) literature, in which general references to big data do exist (Mertens et al., 2016), but specific engagements from a focused methodological point of view are mostly limited to qualitative contextualization of GPS data (Fielding, 2012; Taylor and Horst, 2013; Remijn et al., 2015).
The fact that MMR debates have failed to systematically address the challenges of the “data deluge” inherent to big data research is unfortunate, since the collection and analysis of sensor data is fundamentally indebted—often without a conscious recognition of such—to key principles and concepts from the MMR field. Drawing upon the MMR literature effectively deepens our understanding of the methodological issues involved in sensor-based research. As a result we contribute to existing knowledge across several disciplines in important ways. First, we alert engineers and computer scientists developing sophisticated models with sensor measures (Fu et al., 2015; Sapru and Bourlard, 2015) to consider the quality of the underlying “raw” data from a social science perspective. Physical sensor measures are too easily equated with social or psychological constructs, thereby underestimating the context-sensitive nature of most social phenomena. Second, we provide organizational researchers concerned with the validation of sensor data (Chaffin et al., 2017; Kayhan et al., 2018) a conceptual framework that situates sensors in relation to established social science research methods. Existing validation efforts combine sensor measures with quantitative as well as qualitative data sources without clearly understanding the inherent limitations of “mixing” methods. MMR debates are here pivotal for adjusting contemporary social science research practice to the new volume and high granularity of sensor-derived data. Third, the present article goes beyond many abstract conceptual reflections regarding “big data” in the social sciences (Tonidandel et al., 2018; Wenzel and Van Quaquebeke, 2018). Our literature review of wearable sensor-based research identifies advantages and disadvantages of different sensor platforms as well as the main social and psychological constructs explored up to date. In combination with our review of the most popular devices this will allow interested scholars to gauge more precisely the relative strengths, weaknesses, and practical implications of each sensor device, for their own research.
In order to advance toward a consistent methodological framework of (wearable) sensors as an instrument for social science research, our argument develops as follows. We suggest first, to clearly separate between two measurement dimensions of sensors, namely the physical data, on the one hand, and the potential social or psychological constructs, on the other hand. On a very basic level, sensors capture changes in the physical environment such as a change of temperature, the relative strength of Bluetooth (BT) signals or the acceleration of objects among others. These physical measures are of little interest to social scientists. Only when (wearable) sensors come to represent social and psychological constructs, do they really constitute an exciting new research instrument. However, as we will argue, the potential fit between the basic physical measures and the higher-level constructs needs to be carefully considered. As our literature review will show, these two levels are often conflated, which has a negative impact on the generation of relevant research insights.
Distinguishing between the physical measurement level and higher-level constructs enables us to focus on the fit between these two levels in a second step. We therefore argue that sensor data, like any other social science indicator, involves three distinct elements: the measurement (or indicator) itself, the unobservable construct to which it refers, and the correspondence between the measurement and the construct (Meyer, 2017). The challenge that needs addressing when using sensors concerns the fact that one and the same physical measurement is expected to represent very different social constructs. Thus, “proximity” between BT enabled devices has been used as an indicator of “friendship,” “advice seeking,” or subjective “well-being” (Sekara and Lehmann, 2014; Matusik et al., 2018; Zhang et al., 2018). It is a key contribution of this paper to show, however, that the quality of fit between the physical measure and the social construct depends to a large degree on how sensor data is combined with additional, complementary data sources. Yet, the crucial methodological question on how to carefully plan for “complementary” data needs at the research design stage with wearable sensors has been largely ignored in the literature.
In a third step we therefore advocate that considering the literature on “mixed methods research” enables a better understanding of the ways in which the quality of fit between the sensor measurement and the targeted social/psychological constructs can be controlled and improved. While the mainstream MMR literature restricts the definition of MMR to the mere combination of qualitative and quantitative data, in this article we use a broader definition of MMR that includes “within-paradigm” combinations. This is, the mixture of several quantitative methods/data as well as the integrating of diverse qualitative methods/data (Morse and Niehaus, 2016; Flick, 2017). At the same time, the MMR literature distinguishes between two broad rationales for “mixing” quantitative and qualitative data, namely “convergent validation” and “complementarity” (Greene et al., 1989; Creswell and Plano Clark, 2007). By reviewing the existing literature on sensor research through the lens of these two MMR rationales, as summarized in Table 1, two important methodological issues start to emerge. First, combining sensor data with other quantitative data sources feeds into “convergent validation” as a result of which the biased nature of sensor measures is revealed. Sensor data in this context is subject to “triangulation” in order to achieve a higher validity of field efforts (Denzin, 1978; Erzberger and Kelle, 2003). Although the potential for validity by convergence has been criticized on various grounds (Fielding and Fielding, 1986; Flick, 2008), and is relatively less frequent in actual research practice (Bryman, 2006), when applied to the context of sensor data, it yields crucial insights into the bias inherent in apparently “raw,” physical sensor measures. Second, important challenges emerge when combining sensor data with qualitative methods, which can follow not only the rationale of “convergent validation” but also “complementarity.” Our analysis of existing approaches of “ethno-mining” (Anderson et al., 2009), “blending” (Bornakke and Due, 2018), or “stitching together” (Blok et al., 2017) sensor data with qualitative data shows that the validity of research results is often hampered. This is because the purpose for “mixing” data is left ambiguous, neither being concerned with “convergent validation” nor “complementarity.” By insufficiently conceptualizing the gap between physical measures of behavior and the corresponding layer of social meaning, research efforts fail to deliver significant insights. The negative effects of ignoring the MMR literature for sensor-based research are accentuated by the diverging granularity of the involved data sources. As a further result of our literature review, we argue that the incompatibility between the highly granular nature of sensor data and the static, a-temporal character of traditional quantitative and qualitative data has to date not been sufficiently emphasized as a key limiting factor for sensor-based research.
Table 1. Rationales and challenges for integrating qualitative and quantitative data with sensor-based measurements and derived measures.
Methods
This article grows out of the need to deal with- and reflect upon a concrete research experience with Sociometric badges in R&D teams (Müller, 2018). We agree largely with the growing body of critical literature that fundamental issues regarding construct validity need to be clarified before sensors can fulfill the promise of opening up new and exciting research avenues (Chaffin et al., 2017; Kayhan et al., 2018; Parker et al., 2018). In order to advance toward this goal, we carried out a scoping review of the literature on the use of wearable sensors in organizational research. The methodology for the scoping review followed the steps suggested by Arksey and O'Malley (2005). First, the two research questions guiding the review were identified: What are the strengths and the weaknesses of the different sensor platforms available in the market? What are the main methodological challenges associated with the combination of sensor data with other complementary data sources? Second, the SCOPUS database was searched for relevant publications. The search was performed in the title, abstract and keywords of the publications using terms associated with the concepts of sensor-based research (i.e., wearable sensor*, bluetooth, sociomet*, sensor-based*, wearable comput*) and MMR (mixed method*). A total of 449 publications were generated from the database search. Third, these publications were screened by title and abstract according to the following three exclusion criteria: (1) health related texts dealing with fitness monitoring, elderly care or injury rehabilitation, (2) purely technical and engineering related articles besides research targeting Human Activity Recognition (HAR) with sensors, and (3) studies related to stationary and centralized sensor systems as deployed in smart-homes. From this screening, 419 references were excluded, reducing the number of references to 30, whose eligibility was assessed by reading the full text of the publications. Using the same exclusion criteria stated above, 7 publications were discarded, leaving 23 references for inclusion in the review. Fourth, data from these references such as the types of sensors used, the challenges addressed by the authors or the validity of the reported measurements, were extracted, charted, and summarized. Additional references were identified by consulting key authors CVs, websites of relevant research groups and the tracking of publications that cite key articles on wearable sensors.
A Framework for Conceptualizing Sensor Data
The current interest in using sensors for studying social phenomena is embedded in wider debates regarding big data in the social sciences (Savage and Burrows, 2007; Tinati et al., 2014; George et al., 2016; Tonidandel et al., 2018). Sensors constitute one source of big data when tracking, for example, mobility patterns in cities or monitoring health parameters via fitness trackers and smart-phones. Sensors are increasingly pervasive in all aspects of human life. They are widely used in health related applications as already mentioned (rehab, fitness, elderly care, occupational safety) as well as in “smart” cities and homes, tracking of consumer behavior (retail, tourism) or in the social signal processing community (Vinciarelli et al., 2009; Imani et al., 2016; Alcaraz et al., 2017; Goonawardene et al., 2017; Jiang et al., 2017; Oosterlinck et al., 2017). Out of the considerable variety of sensor types and their divergent application fields, the present article concentrates on a relatively well-defined sub-set, namely wearable sensors used in organizational research. Table 2 provides an overview of the most widely used sensor platforms. All these platforms have in common that sensors are usually worn on the body of individual (research) participants as opposed to operating from one (or several) centrally installed, stationary sensor system(s). The platforms differ however in terms of how many individual sensors are integrated into the same system. This has consequences in terms of cost, battery life, size of devices, and data storage. Importantly, the type of sensors or combinations of sensors also conditions which social or psychological phenomena can be measured. Currently, the OpenBeacon badges deployed by the Sociopatterns project (Cattuto et al., 2010) as well as Sociometric badges developed initially by MIT (Olguin et al., 2009; Kim et al., 2012) are the most widely used wearable sensor systems in social science research. Although other technical solutions do exist, their uptake in actual research is rather limited.
Measurement Dimensions of Wearable Sensors
In order to provide a solid foundation for assessing the reliability and validity of sensor measures in social- and behavioral research, a distinction between the physical measurement level and the social and psychological constructs needs to be drawn. At the first and fundamental level, sensors measure physical phenomena. This fact, although self-evident, cannot be over-emphasized given the tendency within the literature to conflate the physical sensor measures with social phenomena. BT sensors, for example, produce a quantity called Radio Signal Strength Indicator (RSSI). Values usually range from −40 to −90, where higher numbers indicate a stronger signal which is usually produced by devices being closer together (Liu and Striegel, 2011). Captured with a certain periodicity, RSSI is a “moderate” indicator of varying physical proximity. It is “moderate” because obstacles such as walls can lower the BT signal strength, even though devices are close to each other. Table 3 summarizes the existing literature dedicated to the validation of the physical measurement level of wearable sensors. The more or less stringent fit between the actual measurement and the physical construct it represents is indicated in the “Relational Quality” column in Table 3. As illustrated in the next section, the quality of sensors for certain physical constructs is not as tight as one would expect at this basic level of measurement.
Table 3. Overview of sensor types and scientific literature regarding validation of physical constructs.
At a second level, the physical indicator provides the basis for more complex social and psychological constructs. Table 4 lists the higher level concepts that have been addressed so far with wearable sensors. The measurement of such constructs introduces further complexities into the research process, since the quality of research results not only depends upon the quality of the sensor, but also on the degree of consensus around the definition of the social concept under study. How we conceive for example notions of “creativity” or “dominance” on a purely conceptual level is not straightforward but subject to often heated debates within and across the scientific communities (Piffer, 2012). The column “Construct Quality” in Table 4 indicates how disputed the different constructs are. “Stress” for example is relatively well-defined by the level of cortisol in saliva (Taylor et al., 2016). The same holds for “contagion” or “infection,” as defined by the presence or absence of a virus after physical contact. Now the crucial question is to what degree the chosen physical indicators do actually measure these different social concepts. Similar to a medical diagnosis where symptoms are more or less strong indicators of a certain pathology, different sensor-derived measurements are more or less stringent indicators of higher-level constructs. The BT RSSI value is a moderate indicator of physical proximity, which is, in turn, a good indicator of “contagion” but a poor indicator of different types of social relations such as “friendship” or professional “advice.” A microphone, to give another example, provides a numerical record of dominant frequencies (pitch) of the voice which then can be used as an indicator of the persons “sex” (men usually having a lower voice than women) or “stress” (Taylor et al., 2016). Therefore, it is clear that the quality of the overall sensor-based indicator is not just dependent upon the precision of its measurement, but also upon the degree of consensus around the underlying construct and the degree of correspondence between the indicator and the higher-level social- and psychological constructs. Given the variety of constructs for one and the same sensor as described in Table 4, one should expect different levels of validity due to different levels of fit. The fact that we can collect now BT or any other sensor-based data relatively cheaply and with unprecedented detail does not imply that the quality criteria for the measurement properties of these data (i.e., the existence of a solid correspondence between the indicator and the concept it represents) can be ignored.
Table 4. Overview of sensor types and scientific literature regarding validation of social and psychological constructs.
The following sections will delve into some detail in summarizing the state of the art regarding the validity of sensors on the level of physical measures as well as in relation to higher level constructs. The combination of sensor-based data with qualitative and quantitative research approaches as summarized in Table 1 is key for this next step.
Combining Sensor Data With Quantitative Methods
Sensor data are often combined with other quantitative data sources and analytic techniques to validate the given metrics. As mentioned, this concerns both the physical measurement level as well as the level of social- and psychological constructs. While the exploration of the former reveals the constructed nature of sensor measures, the latter emphasizes the ambiguities that exist on the level of the social constructs themselves even before sensor devices are deployed for measurement. Work with sensors in the social sciences should not be blinded by a belief in “big data” measurements where the quantity of data is often thought to miraculously compensate for a lack of a well-argued correspondence rule between the social phenomena of interest and the available indicator.
Scrutinizing “Raw” Sensor Data in the Lab
Much of the excitement around using sensors in the social sciences probably has to do with the promise of providing “objective” measurements of the phenomena under study. Capturing data automatically without direct human intervention addresses the fundamental issue that measurements should not be affected by researchers' bias. As a mechanical form of “observation,” sensors are incorruptible regarding when or what they measure and thus supposedly capable of generating “valid” data that will lead to more solid and far-reaching scientific insights. This promise is even more persuasive within the social sciences where sensors are poised to capture almost imperceptible but elementary behaviors that underlie human interaction. Facilitated by decreasing size and cost, sensors keep track of our behavioral “footprints” on a global scale (Golder and Macy, 2014), underscoring the dominance of numbers as “the modern fact”—simple, unbiased descriptors of phenomena that are only subject to the invariable rules of mathematics (Poovey, 2004, p. xii).
However, although sensors provide a mechanical means of measuring the social, they do not necessarily generate more objective data. Sensors need to be calibrated, have error rates, and are influenced by environmental conditions while data gets sampled, aggregated, and filtered by different algorithms before being exported as “raw” observations for downstream analysis. In fact, as some commentators have remarked, “raw data is an oxymoron” (Gitelman, 2013; Marres, 2017). As shown below, far from being a simple by-product of our activities, sensor-derived data are often an expression of the theories and instruments required for building them in the first place. Therefore, it is increasingly hard to distinguish whether we measure a social phenomenon or the underlying technological devices that mediate it (Marres and Gerlitz, 2015; Marres, 2017). One might ask if we are studying society or technology, RSSI thresholds or knowledge networks, actual friendship ties or algorithmic effects of recommender systems on Facebook? Or, to put it the other way around: to what degree are contemporary social phenomena (such as “friendship”) the effect of technological devices (such as the underlying recommender algorithms in Facebook which suggest new “friends”) rather than genuine social exchange?
The fact that data never “speaks for itself” (Lewis, 2015) has inspired a host of critical studies within behavioral and health sciences, which are summarized in Table 3. These studies are predominantly laboratory experiments where tightly controlled conditions establish the ground-truth on the basis of which sensor-derived measurements are assessed. For instance, Chaffin et al. (2017) strap Sociometric badges onto panels placed into increasing distance to each other in order to see how variability of RSSI metrics correlate with changing physical distance. As the authors report, up to 60% of the variance that BT detects is due to the experimental conditions (physical distance) with 8% of the variance being due to systematic bias of single sensors (Chaffin et al., 2017, p. 9). Next, face-to-face detects are consistently under-reported down to 50% even when placing Sociometric badges in optimal, i.e., manufacturer-specified conditions, for face-to-face detection (Yu et al., 2016; Chaffin et al., 2017). Concerning OpenBeacon sensors, only about half of the actual interactions were recorded by these tested RFID devices (Elmer et al., 2019). Microphones exhibit similar problems: in Yu et al. (2016) sociometers underestimate the duration of speech by 30–40 s while having problems to correctly identify speakers as such (Yu et al., 2016, p. 7). Kayhan et al. (2018) show, through extensive microphone tests, that badges tend to capture changes in volume and frequency accurately, but that differences exist between badges for the same experimental conditions, due to variable sensitivity of each sensor. Building further upon speech detection capacities, turn-taking has a specifically low validity, where sociometers overestimate the ground-truth (actual turns: 6) by a large margin (counted turns: +50) (Chen and Miller, 2017; Kayhan et al., 2018; Müller, 2018). Accelerometer readings provide, on the other hand, more reliable measurements, as reported by both Yu et al. (2016) and Kayhan et al. (2018). On the whole, results are less precise than one would expect at this simple level of physical measurement.
The discrepancy and variability of sensor data with respect to experimental conditions might be further exacerbated by the influence of intermediate processing and data aggregation decisions. Before any data is actually exported for downstream analysis, audio signals are processed by sophisticated algorithms to filter out “noise” from actual “speech” signals—which can substantially alter detected speaking time (Chen and Miller, 2017). Other underlying issues, such as the synchronization of the Sociometric badges internal clock, also play a fundamental role in determining the precision with which badges can measure speaking turns or mirroring activities and, therefore, the extent to which they are able to provide valid measurements. Without a precise synchronization of timestamps between badges, the ability to identify the “same” event across badges is severely impaired (Kayhan et al., 2018).
Thus, although quantitative sensor data enjoy the aura of being “objective” measurements that liberate us from any interpretative effort, upon closer inspection, BT, Infrared, microphone and accelerometer data incorporate a host of technical and measurement biases that need to be taken into account. Instead of providing an exact indicator, it seems more plausible to conceive sensors as probabilistic indicators, even on the level of physical constructs where one intuitively would expect a much more reliable functioning of sensors. This already poses an important limitation to consider before advancing to situations in which sensor data constitute indicators of more complex, social and psychological constructs.
Convergent Validation I: Social Construct Validity via Quantitative Data
While the basic, critical evaluation of sensor metrics presented during the preceding paragraphs is necessary and certainly has a sobering effect regarding their validity as indicators on a physical level, it only provides a first step in the research process. Combinations with established, quantitative measurement scales also aim to validate sensor-based measurements in relation to higher-level constructs, including “types of social relations” or “creativity,” to name just two (see again Table 4 for further examples). Since validity concerns now go beyond the pure physical construct level, the consensus regarding the underlying (social) construct, as well as the correspondence between the indicator and the construct, do enter into the equation, as the following examples will show.
Given the widespread availability of BT sensors, convergent validation of proximity data is relatively common. Sekara and Lehmann (2014) argue, for example, that by selecting a suitable RSSI threshold (at −80) one can distinguish between strong and weak links that correlate with friendship ties on Facebook (Sekara and Lehmann, 2014). The validation efforts in this case already demonstrate the problematic assumption that social network “friends” are reliable indicators of actual friendship. Matusik et al. (2018), in contrast, use self-reports among leaders in a large-scale research facility to assess the convergent and discriminant validity of BT RSSI thresholds for friendship and advice seeking networks (Matusik et al., 2018). As the authors show, self-reported “friendship” ties correlate to a certain extent with lower RSSI values indicating closer spatial proximity, while advice seeking/receiving ties map best onto more liberal RSSI signals. A further related study concentrates on the correlation between face-to-face detects and derived measures such as the diversity of social communication with positive and negative affect and thus subjective well-being (Alshamsi et al., 2016). Although correlations are found between these two elements, they are quite low (Alshamsi et al., 2016, p. 5). Parker et al., to cite another study, show that the number of speaking segments as measured by Sociometric badges and self-turns in conversations are correlated with higher perceived individual and group creativity as measured by the KEYS survey—“the best-established survey instrument for studying creativity in working environments” (Parker et al., 2018, p. 13).
While these studies provide some evidence regarding the possibilities to map various sensor types of Sociometric badges onto social constructs, they also highlight some critical issues. Sekara and Lehmann rightly contend that “proximity” is a questionable indicator of social relations: “[m]ultiple scenarios exist where people are in close contact but are not friends, one obvious example is queuing” (Sekara and Lehmann, 2014, p. 7). As already discussed, the validity of sensor-derived metrics is not just determined by the measurement precision or the fine-tuning of RSSI thresholds, but it is also framed by the relative strength of the correspondence between the indicator and the chosen construct. Now the crucial point is that in cases where this relationship is rather weak, additional contextual information can improve the quality of the indicator. In this scenario, convergent validation is not only carried out by validating sensor data with other quantitative measurement scales, but also by combining such data with qualitative data sources.
Combining Sensor Data With Qualitative Methods
Despite the prevalence of the “complementarity” rationale in literature on integration of sensor data with qualitative methods, as Table 1 suggests, other motivations for combining big data with qualitative sources do exist. However, researchers, when reporting their sensor-based findings rarely distinguish between these underlying MMR rationales and consequently underestimate the methodological implications of obtaining well-founded research results. Based upon Table 1, in the following paragraphs, we will describe three clearly distinct rationales for combining qualitative- with sensor-based data. The resulting typology of approaches provides the first in-road to improved research planning using sensors, and improved collection- and interpretation of the obtained data.
Convergent Validation II: Construct Validity via Qualitative Data
Qualitative data sources can be combined with sensor data for the purpose of validation, i.e., to examine if measurements of the same construct with different instruments converge (or not). A recent exemplary case is available in Parker et al. (2018) where ethnographic observations are used to validate Sociometric speech measurements and body movement in relation to “creativity” in group processes. By closely reading body activity metrics of team members side by side with field notes of the actual working sessions among the environmental scientists, similarities between the two data sources are identified and matched. Thus, specific incidents where the group became “excited and engaged” produce a higher variability in the corresponding body- and speech measurements for the given time slots. Creative moments within the group “are louder on average than any other portion of the group's working day and even louder than their lively lunches and coffee breaks” (Parker et al., 2018, p. 16). Ethnographic observations do not provide complementary insights but rather sufficient contextual detail that make the body- and speech metrics accessible to a specific interpretation—in the current case in terms of a creative flow.
A similar convergent validation logic can be applied to the highly-cited BT studies described in the preceding paragraphs. Eagle et al. (2009) observe, for example, that inferring the friendship network structure through BT signals can be vastly improved when contextual information regarding work and leisure times and places is taken into consideration. Since the ratio of proximity detects outside work (hours) is much higher for friends than for non-friends, “it was possible to predict 96% of symmetric reports of non-friendship and 95% of symmetric friendship” (Eagle et al., 2009, p. 15,275). By observing broad contextual cues of the overall situations under study, qualitative accounts can improve the overall quality of the indicator. Indeed, by providing complementary data on the times and places where “friendship” is more likely to occur, the fit between the indicator and the construct can be improved.
Qualitative Data for Complementary Insights
Contrary to the “convergent validation” rationale, qualitative data might also be used to complement sensor data, i.e., to examine different facets of the same phenomena by using different (complementary) methods. A recent, exemplary case, is a study published by Bornakke and Due (2018) on the utility of bike signs in the city of Copenhagen. Researchers tracked mobility patterns of cyclists via GPS and combined these with participant observation, direct inquiries and a short questionnaire. The sensor-based information on actual bicycle journeys was thus contextualized with the cyclists' stories explaining their choice of one trajectory or another. The qualitative material explored the “why” question, based upon “what” had happened, i.e., the difference between morning vs. afternoon routes. Or, to put it the other way round, the GPS data “extends the thick observations with knowledge on the generalizability of the behavior of using multiple routes to and from one's home.” (Bornakke and Due, 2018, p. 12).
The way qualitative and sensor data are combined in this example follows a strong “complementarity” rationale where qualitative observations provide a new layer of meaning to the quantitative mobility tracks. This is possible because GPS is a relatively reliable indicator of actual positions in the city; the coordinates are valid as indicators of physical routes and deliver an “autonomous” result to be used for further analysis, questioning and interpretation by researchers. Qualitative data adds here a complementary layer of meaning rather than being preoccupied with untangling the precise social situation represented by the data. Therefore, what this example shows is that only when sensor data has been established as an independent and valid source or measurement, does it make sense to “complement” it with qualitative data. The importance of the dialectic between “complementarity” and “validity” is illustrated more clearly in the example presented in the next section.
Anchoring Sensor Data via Qualitative Data?
A third approach to combine qualitative data with sensor data does not aim for convergent validation of certain constructs, nor does it provide complementary insights. Rather, it constitutes a more problematic account where qualitative insights aim to make the data interpretable as such. This approach can be illustrated by the following example. In their study at the Danish Technical University, Blok et al. (2017) tracked an entire freshman class of 800 people using BT sensors in smartphones. Ethnographic observations are part of the overall methods, used to contrast the recorded “big” interaction traces with first-person accounts of “thick” descriptions during selected events such as a student party for example. Given that BT is a relatively fuzzy indicator, the analytic work involved probing different aggregation and visualization techniques of the BT signals while constantly verifying the corresponding time slots with the ethnographers' descriptions. Digital data, as the authors explain, “allow for great plasticity” requiring a constant oscillation between extracting “interpretable occasions” from the data and cross-checking with observational accounts (Blok et al., 2017, p. 6). However, what Blok et al. identify as the “great plasticity” of digital data, is nothing but a lack of specificity and hence low quality of proximity-based indicators for social phenomena. In Blok's research, it is the fuzzy nature of proximity data that leads researchers into an endless regress of probing different aggregations levels, filters or visualizations in a hunt for significant patterns of behavior. Exploring the data without any fixed social concept in mind, these patterns mostly fail to emerge because “proximity” in itself is a very ambiguous indicator that can signify anything and nothing. When a distinct proximity/distance pattern finally does emerge and can be correctly labeled with the help of ethnographic field notes—as people leaving the party through a tunnel, at the end of which they “hug goodbye” (Blok et al., 2017, p. 8)—it is not clear that this insight contributes to the wider research question. Although satisfying to “see” this pattern of behavior in the data, it is, somehow, a rather shallow research result because it simply mirrors a broad contextual observation of the event, namely the end of the party. As Doreian and Conti (2012) argue, it is a common mistake of researchers to conflate organizational or spatial context variables with genuine social structures and individual preferences. Without a clear construct that sensor data is supposed to measure, its decontextualized nature leads to interpretations that mirror more the resources and capabilities of the observers than genuine patterns of social behavior.
Therefore, combining qualitative data sources with sensor data is unproductive when the two data sources are simply “stitched” together, without clear theoretical motivation. Although this statement is understandable from an ethnographic and inductively-based approach to data collection, it causes problems when applied to sensor data. And there is a precise reason for that, which will be largely discussed in the next section: inductively exploring sensor data easily falls prey to reproducing broad contextual observations that are first and foremost an expression of the limited observational capacities of the ethnographer rather than an expression of the inherent patterns in high-resolution, time-based sensor data.
Validation and the Problem of “Matching the Resolution”
The preceding examples of combinations of sensor data with quantitative and qualitative data sources reveal an underlying, but nevertheless important, problem of mismatch between the scale (or resolution) of the sensor data and that of more conventional data collection methods. BT sensors register interaction between persons on a continuous basis. Monitoring a group of 11 people over 5 days can produce up to +100,000 detects. Although data collection methods such as participant observations, interviews, or questionnaires are able to contextualize a number of data points, its focus is, however, limited to events that are more notable and to coarse aggregation levels. Consequently, these methods fail to capture a large part of the continuous temporal information that can be obtained through sensors. Returning to Blok et al. (2017) example, it is clear from this study that the minute details of physical proximity registered by sensors between several hundred participants could not be matched with the observational accounts generated by a well-trained ethnographer. The ethnographic insights in Blok et al. (2017) study (e.g., the “high” and “low” energy of the overall party) remain coarse approximations to the phenomenon under study and, therefore, are ineffective in grasping more subtle events such as, for example, how the overall “energy of the party” emerges out of specific micro-dynamics on the group level. If big data stands for data velocity and volume, qualitative and quantitative accounts have a hard time to keep up the pace. The possibilities of addressing more fine-grained questions are indeed directly dependent on available resources which, most frequently, tend to be limited by “time, research funds, and human coding hours” (Lehmann-Willenbrock et al., 2017, p. 523).
Quantitative, survey-based approaches share a similar fate. The previously cited study by Matusik et al. (2018) ignores the temporal dimension of the proximity data as it correlates averaged RSSI values across a 9-day field period with advice- and friendship networks. This high aggregation level could, of course, be broken down into shorter time spans, such as daily- or hourly RSSI mean values. However, the shorter the time-slots over which the sensor data is averaged, the higher the efforts to collect the increasing number of corresponding survey measures. The cited study of Alshamsi et al. (2016) faced similar challenges when participants were asked to respond three times per day to the same questionnaire over a period of 30 days in order to validate their “affect states” with slices of sociometric face-to-face networks. The number of times that these fixed snapshots of affective moods are collected is, to a certain extent, somewhat arbitrary and clearly conditioned by the researchers' ambition to push forward the limits of taxing data collection. What all these validation studies fail to address is the fundamental problem of a mismatch of resolution between sensor-derived data and more traditional, static data types.
As with any measurement instrument, using sensors within the social sciences requires a conscious calibration effort that examines closely how tightly physical- and social constructs are bound to the available sensor metrics. Combining sensor data with other qualitative and quantitative sources confronts the dilemma of either addressing only a fragment of the behavioral measures or having to invest disproportional efforts to collect data with the comparable amount of detail using conventional methods. To say it somewhat figuratively: conventional qualitative and quantitative methods are the bottleneck through which the validation of big data has to pass. This methodological challenge is increasingly being addressed, for example by Luciano et al. (2018) who work toward new approaches to assess “measurement fit” with dynamic, time-based data.
Concluding Remarks and Recommendations: Toward Mixed Method Research With Sensors
In this article we argue, that in order to reliably use sensor data in organizational research, complementary data needs to be collected. The necessity of integrating sensor data with other, quantitative and qualitative methods sharply contrasts with the predominant tendency in the literature to use (wearable) sensors as a “stand-alone” research instrument (Schmid Mast et al., 2015; George et al., 2016; Tonidandel et al., 2018). The exciting new research questions made possible by the continuous, high resolution monitoring of human behavior remains dependent upon the variable “fit” between physical measurement and the targeted social or psychological constructs. Since the gap between physical sensor data and social construct level is non-negotiable, researchers need to address the validity of their sensor data in the context of their theoretical framework, research questions and by incorporating a mixed methods perspective. As Parker et al. (2018) argue: “From a sociological perspective, this research is intriguing and potentially generative, but its exclusive focus on non-verbal behavior, tendency to use sensors without triangulating methods or confirmatory data, and minimal grounding in sociological theory raise questions about its reliability, validity, and explanatory power.” (p. 9). The following paragraphs provide concrete recommendations for research practice that address the inter-related nature between sensor platform, analytic interest and contextual data needs. We thereby focus on broader methodological considerations that apply to wearable sensors at large; others have provided practical tips when using Sociometric Badges (Chaffin et al., 2017; Kayhan et al., 2018; Parker et al., 2018) or Sociopatterns/OpenBeacons (Elmer et al., 2019) in research. Ethical considerations including privacy issues are discussed in Stopczynski et al. (2014); Metcalf and Crawford (2016).
First, the choice of a sensor platform needs to be adjusted in relation to the research question and a coherent theory that argues for the correspondence between the sensor-based indicator and the relevant constructs. No amount of big data can, by itself, leverage a potential misfit between sensor metrics and higher-level constructs. The overview of Tables 3, 4 provide a first orientation regarding diverse social and psychological concepts explored with different sensor types. “Creativity” for example has been addressed with three types of sensors, namely accelerometers, microphone, and proximity-based sensors (BT or RFID). Researchers interested in exploring “creativity” or other affect-related concepts are well-advised to use an integrated solution such as Sociometric badges or smartphones that can record proximity between participants as well as body activity and speech features. The importance of these data dimensions for monitoring “creative” behavioral markers can justify the relatively higher cost and complex field logistics involved with Sociometric badges (which require daily recharging and data download). On the other hand, researchers interested in monitoring the spread of information or contact and collaboration patterns more generally are well-advised to use simpler systems such as the Sociopatterns/OpenBeacons platform. Although this platform “only” incorporates a proximity sensor, it is the most efficient solution in terms of cost, device size and field logistics for delivering the required “contact” data. Other sensor platforms mentioned in Table 2 cannot be recommended at this point, either because they have been discontinued (TelosB), they are not available for researchers (HBM), or they are still under active development (Rhythm badges).
A second point concerns the recommendations for increasing the validity of sensor measures. As Parker et al. remark, “most current sociometric research conducted by computer scientists and engineers accepts sociometric data at face value” thus risking to create “large datasets and performing sophisticated analysis on data of questionable quality, yielding incomplete or incorrect understandings of small group structure and process” (Parker et al., 2018, p. 25). As the following examples will show, researchers have the choice between different methodological approaches to control the coupling of physical measures, human behavior, and their social interpretation. A tighter coupling of behavior to social meaning is usually dependent upon a strict control of the experimental situation in laboratory settings. More open and complex situations in field research require the recollection of complementary, contextual cues to guarantee the validity of sensor data.
It is important to realize that much of the initial interest surrounding sensors as powerful new research instrument is based upon research carried out in the laboratory. The tight control of the experimental situation eases the interpretative burden of the sociometric data to a considerable degree, fueling hopes for capturing “honest signals” (Pentland, 2008). As sensors can capture a number of elements of body language (i.e., body posture and movement, vocal behavior such as pitch or volume) they provide access to subtle, non-verbal behaviors beyond the literal meaning of words (Hall et al., 2013; Bonaccio et al., 2016). Since these semi-automated behaviors nevertheless steer and structure social interactions on a fundamental level, it becomes conceivable to interpret the physical measurement of behavioral as “honest” signals of social phenomena. Impressive results have been reported by Alex “Sandy” Pentland's group at MIT using Sociometric badges to predict the outcome of speed dating events, elevator pitches, or salary negotiations with surprising precision (see Appendix B of Pentland, 2008, p. 113ff). However, it's worth remembering that the tight coupling of physical sensor measures to social phenomena in these initial studies has been achieved by not only setting up quasi-experimental situations but also by limiting the scope of the dependent variables of interest. By using simple, binary outcome variables (win/lose, trade business card/not, higher salary/lower), relatively high correlations between the sensor measurements and the dependent variables could be achieved. In short, it is the control of “context” that “improves” the validity of the sensor measures as a probabilistic indicator of “honest signals.”
However, as soon as the outcome variables become more complex and the context of social interaction is less defined, theoretical choices regarding the social constructs of interest need to be fined tuned in relation to the type of sensors deployed and their dependency on contextual cues. As soon as sensors are embedded into real-world field settings, the defining feature of “big data”—namely, having a high (temporal) resolution while lacking contextual information (Cai and Zhu, 2015)—comes to the fore. Sensor data is “big” but also “thin” data that needs to be combined with qualitatively grounded “small data” derived from interviews, focus groups or participant observation to unlock their context and hence social meaning (Burrell, 2012; Curran, 2013; Ford, 2014). Eagle et al. (2009) study comes here to mind where the interpretation of physical proximity patterns in terms of “friendship” could be made much more precise by distinguishing between work- or leisure contexts. Complementary observations regarding people's choices when and where to be near each other provides the contextual cues to qualify their physical proximity in terms of specific social relations such as “friendship.” While the observation of context is key for research about “friendship,” complementary observations might be less important when exploring the spread of a contagious disease. In order to study the transmission of a pathogenic germ in a hospital ward, the observation of the physical contact pattern is enough because the validity of the physical proximity measure is a good indicator for contact-related phenomena. Researchers thus have to decide on the relation between the physical measure, the targeted construct and the contextual cues to be observed in order to achieve a better fit and validity of data.
A further example concerns the contextual observations necessary for the study of “creativity.” “Creativity” is usually associated with higher agitation levels in body movement as well as speech (volume) data (Yano et al., 2015; Parker et al., 2018). However, as Parker at al. argue, “having an ethnographer in the room was critical for distinguishing between an exciting scientific episode and a coffee break or photo opportunity” (Parker et al., 2018, p. 25). The observation of the wider context anchors the interpretation of the sensor data in relation to the theoretical interest of the research. The fact that sensor-based methods apparently reduce field efforts in terms of the amount and continuous collection of data needs to be balanced in relation to the additional efforts necessary for contextualizing the collected data. In this regard, under tight research budgets, it is crucial to take into account the resources and skills needed for gathering complementary data by selecting specific events and episodes of observation or for soliciting comments from research participants.
The elaboration of a more precise understanding of observational cues necessary for working with different social and psychological constructs is an important challenge for the future of wearable sensor research. How to best distinguish “knowledge sharing” behavior in proximity data of shared office spaces or small groups for example is such a problem (Génois and Barrat, 2018). From the data it is not directly deducible if the proximity of colleagues is a product of seating order, actual collaboration or even the product of other sources of “bias” such as “people's personality, cultural background, or substance consumption” (Elmer et al., 2019, p. 16). However, addressing validity concerns of sensor data collected in the field is a first, necessary step before considering the true potential of this type of new research instruments, namely to study the temporal dimension of social phenomena (Leenders et al., 2016). Sensors can contribute for example to the study of hitherto marginally explored temporal dimension of “creativity” in organizations by providing a window on the micro-sequencing of events that are responsible of “flow” experiences (Csikszentmihalyi, 1975; Gaggioli et al., 2013). Or, to take the study of leadership emergence as another example: in combination with new analytical Relational Event Models (Butts, 2008), highly granular sensor data enables a much closer look at the “microdynamic relational processes” (Carter et al., 2015) or “micro-origins” (Cook et al., 2019) that govern leadership emergence. Given the predominance of classical, a-temporal accounts of many classical research methods, a vast landscape of sociological and psychological constructs are awaiting to be rethought by organizational researchers in terms of their genuine temporal grounding, including the path dependency between events, their duration, frequencies and cyclicality (Quintane et al., 2013; Ubaldi et al., 2017).
In this article, we have outlined the rather complex processes involved when using wearable sensors in social science research. It counters what is often blind faith in data volume, speed, and variety. As we have argued, the pitfalls of an uncritical acceptance of sensor-based data become apparent by contextualizing the usage of sensor data with MMR rationales as well as basic notions of social science indicators. To the best of our knowledge, the current article is the first to explicitly address issues of wearable sensor research in conjunction with MMR. A careful examination of available sensor types and their corresponding social and psychological constructs as outlined in Tables 3, 4 provides the first steps to a more beneficial deployment of sensors in field research settings. A clear alignment of research question, theoretical constructs and complementary data needs is key for successful wearable sensor research. We hope to have contributed some conceptual clarifications that will provide the framework for a more realistic assessment of these new, existing instruments for social science the potential applications of which are nothing short of exhilarating.
Author Contributions
JM: conceptual development and first draft of paper. SF: methodological development and major revisions/comments. MR and EG: contribution to conceptual development, empirical fieldwork and analysis of data underlying the paper, and revisions/comments to drafts.
Funding
The present work has received funding from the European Commission H2020 programme under grant agreement number 665851.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
Alcaraz, J. C., Moghaddamnia, S., and Peissig, J. (2017). “Mobile quantification and therapy course tracking for gait rehabilitation,” in 2017 22nd International Conference on Digital Signal Processing (DSP), 1–5, (London, UK). doi: 10.1109/ICDSP.2017.8096106
Alshamsi, A., Pianesi, F., Lepri, B., Pentland, A., and Rahwan, I. (2016). Network diversity and affect dynamics: the role of personality traits. PLoS ONE 11:e0152358. doi: 10.1371/journal.pone.0152358
Anderson, K., Nafus, D., Rattenbury, T., and Aipperspach, R. (2009). Numbers have qualities too: experiences with ethno-mining. Ethnogr. Praxis Industry Conf. Proc. 2009, 123–140. doi: 10.1111/j.1559-8918.2009.tb00133.x
Arksey, H., and O'Malley, L. (2005). Scoping studies: towards a methodological framework. Int. J. Soc. Res. Methodol. 8, 19–32. doi: 10.1080/1364557032000119616
Atzmueller, M., Thiele, L., Stumme, G., and Kauffeld, S. (2018). “Analyzing group interaction on networks of face-to-face proximity using wearable sensors,” in 2018 IEEE International Conference on Future IoT Technologies (Future IoT), 1–10, (Eger). doi: 10.1109/FIOT.2018.8325593
Barrat, A., Cattuto, C., Tozzi, A. E., Vanhems, P., and Voirin, N. (2014). Measuring contact patterns with wearable sensors: methods, data characteristics and applications to data-driven simulations of infectious diseases. Clin. Microbiol. Infect. 20, 10–16. doi: 10.1111/1469-0691.12472
Blok, A., Carlsen, H. B., Jørgensen, T. B., Madsen, M. M., Ralund, S., and Pedersen, M. A. (2017). Stitching together the heterogeneous party: a complementary social data science experiment. Big Data Soc. 4, 1–15. doi: 10.1177/2053951717736337
Bonaccio, S., OReilly, J., OSullivan, S. L., and Chiocchio, F. (2016). Nonverbal behavior and communication in the workplace: a review and an agenda for research. J. Manage. 42, 1044–1074. doi: 10.1177/0149206315621146
Boonstra, T. W., Larsen, M. E., Townsend, S., and Christensen, H. (2017). Validation of a smartphone app to map social networks of proximity. PLoS ONE 12:e0189877. doi: 10.1371/journal.pone.0189877
Bornakke, T., and Due, B. L. (2018). Big–Thick Blending: a method for mixing analytical insights from big and thick data sources. Big Data Soc. 5, 1–16. doi: 10.1177/2053951718765026
Bryman, A. (2006). Integrating quantitative and qualitative research: how is it done? Qualit. Res. 6, 97–113. doi: 10.1177/1468794106058877
Burrell, J. (2012). “The Ethnographer's complete guide to big data: small data people in a big data World,” in Ethnography Matters. Available online at: http://ethnographymatters.net/blog/2012/05/28/small-data-people-in-a-big-data-world/ (accessed May 7, 2018).
Butts, C. T. (2008). A relational event framework for social action. Sociol. Methodol. 38, 155–200. doi: 10.1111/j.1467-9531.2008.00203.x
Cai, L., and Zhu, Y. (2015). The challenges of data quality and data quality assessment in the big data era. Data Sci. J. 14:2. doi: 10.5334/dsj-2015-002
Carter, D. R., DeChurch, L. A., Braun, M. T., and Contractor, N. S. (2015). Social network approaches to leadership: an integrative conceptual review. J. Appl. Psychol. 100, 597–622. doi: 10.1037/a0038922
Cattuto, C., Van den Broeck, W., Barrat, A., Colizza, V., Pinton, J.-F., and Vespignani, A. (2010). Dynamics of person-to-person interactions from distributed RFID sensor networks. PLoS ONE 5:e11596. doi: 10.1371/journal.pone.0011596
Chaffin, D., Heidl, R., Hollenbeck, J. R., Howe, M., Yu, A., Voorhees, C., et al. (2017). The promise and perils of wearable sensors in organizational research. Organiz. Res. Methods 20, 3–31. doi: 10.1177/1094428115617004
Chancellor, J., Layous, K., Margolis, S., and Lyubomirsky, S. (2017). Clustering by well-being in workplace social networks: homophily and social contagion. Emotion 17, 1166–1180. doi: 10.1037/emo0000311
Chen, H.-E., and Miller, S. R. (2017). “Can wearable sensors be used to capture engineering design team interactions? An investigation into the reliability of wearable sensors,” in Proceedings of the ASME 2017 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. (Cleveland, OH). doi: 10.1115/DETC2017-68183
Cook, A., Meyer, B., Gockel, C., and Zill, A. (2019). Adapting leadership perceptions across tasks: micro-origins of informal leadership transitions. Small Group Res. 50, 227–265. doi: 10.1177/1046496418810437
Creswell, J. W., and Plano Clark, V. L. (2007). Designing and Conducting Mixed Methods Research. Thousand Oaks, CA: SAGE Publications.
Csikszentmihalyi, M. (1975). Beyond Boredom and Anxiety, 1st Edn. San Francisco, CA: Jossey-Bass Publishers.
Curran, J. (2013). “Big Data or ‘Big Ethnographic Data'?” in Positioning Big Data Within the Ethnographic Space. Ethnographic Praxis in Industry Conference Proceedings 2013, 62–73, (London, UK). doi: 10.1111/j.1559-8918.2013.00006.x
Daggett, M., O'Brien, K., Hurley, M., and Hannon, D. (2017). “Predicting team performance through human behavioral sensing and quantitative workflow instrumentation,” in Advances in Human Factors and System Interactions Advances in Intelligent Systems and Computing (Cham: Springer), 245–258. doi: 10.1007/978-3-319-41956-5_22
Denzin, N. K. (1978). The Research Act: A Theoretical Introduction to Sociological Methods. 2nd Edn. New York, NY: McGraw-Hill.
Dietzel, J., Francu, E. R., Lucas, B., and Zaki, M. (2018). Contextually Defined Postural Markers Reveal Who's in Charge: Evidence from Small Teams Collected with Wearable Sensors. Available online at: https://cambridgeservicealliance.eng.cam.ac.uk/resources/Downloads/Monthly%20Papers/2018JanuaryPaperLucas.pdf (accessed May 13, 2019).
Doherty, S. T., Lemieux, C. J., and Canally, C. (2014). Tracking human activity and well-being in natural environments using wearable sensors and experience sampling. Soc. Sci. Med. 106, 83–92. doi: 10.1016/j.socscimed.2014.01.048
Dong, W., Lepri, B., and Pentland, A. (2012). “Automatic prediction of small group performance in information sharing tasks,” in Collective Intelligence Conference. Available online at: http://arxiv.org/abs/1204.3698
Doreian, P., and Conti, N. (2012). Social context, spatial structure and social network structure. Soc. Netw. 34, 32–46. doi: 10.1016/j.socnet.2010.09.002
Eagle, N., Pentland, A., and Lazer, D. (2009). Inferring friendship network structure by using mobile phone data. Proc. Natl. Acad. Sci. U.S.A. 106, 15274–15278. doi: 10.1073/pnas.0900282106
Elmer, T., Chaitanya, K., Purwar, P., and Stadtfeld, C. (2019). The validity of RFID badges measuring face-to-face interactions. Behav Res, 1–19. doi: 10.3758/s13428-018-1180-y
Endedijk, M., Hoogeboom, M., Groenier, M., de Laat, S., and van Sas, J. (2018). Using sensor technology to capture the structure and content of team interactions in medical emergency teams during stressful moments. Educ. Sci. Tech. Med. 6, 123–147. doi: 10.14786/flr.v6i3.353
Erzberger, C., and Kelle, U. (2003). “Making inferences in mixed methods: the rules of integration,” in Handbook of Mixed Methods in Social and Behavioral Research, eds A. Tashakkori and C. Teddlie (Thousand Oaks, CA: SAGE Publications), 457–490.
Fielding, N., and Fielding, J. L. (1986). Linking Data. Beverly Hills, CA: Sage Publications. doi: 10.4135/9781412984775
Fielding, N. G. (2012). Triangulation and mixed methods designs: data integration with new research technologies. J. Mix. Methods Res. 6, 124–136. doi: 10.1177/1558689812437101
Flick, U. (2008). Managing Quality in Qualitative Research. Los Angeles, CA: London: SAGE Publications. doi: 10.4135/9781849209441
Flick, U. (2017). Mantras and myths: the disenchantment of mixed-methods research and revisiting triangulation as a perspective. Qual. Inquiry 23, 46–57. doi: 10.1177/1077800416655827
Ford, H. (2014). Big Data and Small: collaborations between ethnographers and data scientists. Big Data Soc. 1, 1–4. doi: 10.1177/2053951714544337
Fu, E. Y., Leong, H. V., Ngai, G., and Chan, S. (2015). “Automatic fight detection based on motion analysis,” in 2015 IEEE International Symposium on Multimedia (ISM) (Miami, FL: IEEE), 57–60. doi: 10.1109/ISM.2015.98
Gaggioli, A., Cipresso, P., Serino, S., and Riva, G. (2013). Psychophysiological correlates of flow during daily activities. Annu. Rev. CyberTher. Telemed. 11, 65–69. doi: 10.3233/978-1-61499-282-0-65
Génois, M., and Barrat, A. (2018). Can co-location be used as a proxy for face-to-face contacts? EPJ Data Sci. 7:11. doi: 10.1140/epjds/s13688-018-0140-1
George, G., Osinga, E. C., Lavie, D., and Scott, B. A. (2016). Big Data and Data science methods for management research. Acad. Manag. J. 59, 1493–1507. doi: 10.5465/amj.2016.4005
Gitelman, L. ed. (2013). “Raw Data” Is an Oxymoron. Cambridge, MA; London: The MIT Press. doi: 10.7551/mitpress/9302.001.0001
Gloor, P. A., Fischbach, K., Fuehres, H., Lassenius, C., Niinimäki, T., Olgúin-Olgúin, D., et al. (2011). Towards “Honest Signals” of creativity – identifying personality characteristics through microscopic social network analysis. Proc. Soc. Behav. Sci. 26, 166–179. doi: 10.1016/j.sbspro.2011.10.573
Golder, S. A., and Macy, M. W. (2014). Digital footprints: opportunities and challenges for online social research. Annu. Rev. Sociol. 40, 129–152. doi: 10.1146/annurev-soc-071913-043145
Gong, J., Huang, Y., Chow, P. I., Fua, K., Gerber, M. S., Teachman, B. A., et al. (2019). Understanding behavioral dynamics of social anxiety among college students through smartphone sensors. Inform. Fusion 49, 57–68. doi: 10.1016/j.inffus.2018.09.002
Goonawardene, N., Toh, X., and Tan, H.-P. (2017). “Sensor-Driven detection of social isolation in community-dwelling elderly,” in Human Aspects of IT for the Aged Population. Applications, Services and Contexts Lecture Notes in Computer Science, eds. J. Zhou and G. Salvendy (Vancouver, BC: Springer International Publishing), 378–392. doi: 10.1007/978-3-319-58536-9_30
Greene, J. C., Caracelli, V. J., and Graham, W. F. (1989). Toward a conceptual framework for mixed-method evaluation designs. Educ. Eval. Policy Anal. 11, 255–274. doi: 10.3102/01623737011003255
Guclu, H., Read, J., Vukotich, C. J., Galloway, D. D., Gao, H., Rainey, J. J., et al. (2016). Social contact networks and mixing among students in K-12 schools in pittsburgh, PA. PLoS ONE 11:e0151139. doi: 10.1371/journal.pone.0151139
Hall, J. A., and Knapp, M. L. eds. (2013). Nonverbal Communication. Berlin; Boston, MA: Walter de Gruyter. doi: 10.1515/9783110238150
Holding, B. C., Sundelin, T., Lekander, M., and Axelsson, J. (2019). Sleep deprivation and its effects on communication during individual and collaborative tasks. Sci. Rep. 9:3131. doi: 10.1038/s41598-019-39271-6
Huang, W., Kuo, Y.-S., Pannuto, P., and Dutta, P. (2014). “Opo: a wearable sensor for capturing high-fidelity face-to-face interactions,” in Proceedings of the 12th ACM Conference on Embedded Networked Sensor Systems SenSys'14 (Memphis, TN: ACM).
Imani, S., Bandodkar, A. J., Mohan, A. M. V., Kumar, R., Yu, S., Wang, J., et al. (2016). A wearable chemical–electrophysiological hybrid biosensing system for real-time health and fitness monitoring. Nat. Commun. 7:11650. doi: 10.1038/ncomms11650
Jiang, J., Pozza, R., Gunnarsdóttir, K., Gilbert, N., and Moessner, K. (2017). Using Sensors to Study Home Activities. J. Sensor Actuat. Netw. 6:32. doi: 10.3390/jsan6040032
Kayhan, V. O., Chen, Z., French, K. A., Allen, T. D., Salomon, K., and Watkins, A. (2018). How honest are the signals? A protocol for validating wearable sensors. Behav. Res. Methods 50, 57–83. doi: 10.3758/s13428-017-1005-4
Kim, T., McFee, E., Olguin, D. O., Waber, B., Pentland, A., and “Sandy,” (2012). Sociometric badges: using sensor technology to capture new forms of collaboration. J. Organ. Behav. 33, 412–427. doi: 10.1002/job.1776
Leenders, R. T. A. J., Contractor, N. S., and DeChurch, L. A. (2016). Once upon a time. Understanding team processes as relational event networks. Organiz. Psychol. Rev. 6, 92–115. doi: 10.1177/2041386615578312
Lehmann-Willenbrock, N., Hung, H., and Keyton, J. (2017). New frontiers in analyzing dynamic group interactions: bridging social and computer science. Small Group Res. 48, 519–531. doi: 10.1177/1046496417718941
Lepri, B., Staiano, J., Shmueli, E., Pianesi, F., and Pentland, A. (2016). The role of personality in shaping social networks and mediating behavioral change. User Model. User Adapt. Interact. 26, 143–175. doi: 10.1007/s11257-016-9173-y
Lewis, K. (2015). Three fallacies of digital footprints. Big Data Soc. 2, 1–4. doi: 10.1177/2053951715602496
Liu, S., and Striegel, A. (2011). “Accurate extraction of face-to-face proximity using smartphones and bluetooth,” in 2011 Proceedings of 20th International Conference on Computer Communications and Networks (ICCCN) (Maui, HI: IEEE), 1–5. doi: 10.1109/ICCCN.2011.6006081
Luciano, M. M., Mathieu, J. E., Park, S., and Tannenbaum, S. I. (2018). A fitting approach to construct and measurement alignment: the role of big data in advancing dynamic theories. Organiz. Res. Methods 21, 592–632. doi: 10.1177/1094428117728372
Marres, N., and Gerlitz, C. (2015). Interface Methods: Renegotiating relations between digital social research, STS and sociology. Sociol. Rev.64, 21–46. doi: 10.1111/1467-954X.12314
Martinez Mozos, O., Sandulescu, V., Andrews, S., Ellis, D., Bellotto, N., Dobrescu, R., et al. (2016). Stress detection using wearable physiological and sociometric sensors. Int. J. Neural Syst. 27:1650041. doi: 10.1142/S0129065716500416
Matusik, J. G., Heidl, R., Hollenbeck, J. R., Yu, A., Lee, H. W., and Howe, M. (2018). Wearable bluetooth sensors for capturing relational variables and temporal variability in relationships: a construct validation study. J. Appl. Psychol. 104, 357–387. doi: 10.1037/apl0000334
Mertens, D. M., Bazeley, P., Bowleg, L., Fielding, N., Maxwell, J., Molina-Azorin, J. F., et al. (2016). Expanding thinking through a kaleidoscopic look into the future: implications of the mixed methods international research association's task force report on the future of mixed methods. J. Mix. Methods Res. 10, 221–227. doi: 10.1177/1558689816649719
Metcalf, J., and Crawford, K. (2016). Where are human subjects in Big Data research? The emerging ethics divide. Big Data Soc. 3, 1–14. doi: 10.1177/2053951716650211
Meyer, W. (2017). “Einführung in die Grundlagen der Entwicklung von Indikatoren,” in Gleichstellung Messbar Machen Grundlagen und Anwendungen von Gender- und Gleichstellungsindikatoren, eds A. Wroblewski, U. Kelle, and F. Reith (Wiesbaden: Springer VS), 15–38. doi: 10.1007/978-3-658-13237-8_2
Morse, J. M., and Niehaus, L. (2016). Mixed Method Design: Principles and Procedures. London; New York, NY: Routledge. doi: 10.4324/9781315424538
Müller, J. (2018). Using Wearable Sensors In Gender Research. Comparative Case Study Report. GEDII Project Deliverable D2.4.
Olguin, D. O., Waber, B. N., Taemie, K.im, Mohan, A., Ara, K., and Pentland, A. (2009). Sensible organizations: technology and methodology for automatically measuring organizational behavior. IEEE Trans. Syst. Man Cybernet. Part B 39, 43–55. doi: 10.1109/TSMCB.2008.2006638
Onnela, J.-P., Waber, B. N., Pentland, A., Schnorf, S., and Lazer, D. (2014). Using sociometers to quantify social interaction patterns. Sci. Rep. 4:5604. doi: 10.1038/srep06278
Oosterlinck, D., Benoit, D. F., Baecke, P., and Van de Weghe, N. (2017). Bluetooth tracking of humans in an indoor environment: an application to shopping mall visits. Appl. Geogr. 78, 55–65. doi: 10.1016/j.apgeog.2016.11.005
Parker, J. N., Cardenas, E., Dorr, A. N., and Hackett, E. J. (2018). Using sociometers to advance small group research. Sociol. Methods Res. doi: 10.1177/0049124118769091
Pentland, A. (2008). Honest Signals How They Shape Our World. Cambridge, MA: MIT Press. doi: 10.7551/mitpress/8022.001.0001
Piffer, D. (2012). Can creativity be measured? An attempt to clarify the notion of creativity and general directions for future research. Think. Skills Creat. 7, 258–264. doi: 10.1016/j.tsc.2012.04.009
Poovey, M. (2004). A History of the Modern Fact: Problems of Knowledge in the Sciences of Wealth and Society. Chicago, IL: University of Chicago Press.
Psylla, I., Sapiezynski, P., Mones, E., and Lehmann, S. (2017). The role of gender in social network organization. PLoS ONE 12:e0189873. doi: 10.1371/journal.pone.0189873
Quintane, E., Pattison, P. E., Robins, G. L., and Mol, J. M. (2013). Short- and long-term stability in organizational networks: temporal structures of project teams. Soc. Netw. 35, 528–540. doi: 10.1016/j.socnet.2013.07.001
Remijn, L., Stembert, N., Mulder, I. J., and Choenni, S. (2015). “Emergent technologies in multimethod and mixed methods research: incorporating mobile technologies,” in The Oxford Handbook of Multimethod and Mixed Methods Research Inquiry, eds S. Hesse-Biber and B. Johnson (Oxford; New York, NY: Oxford University Press), 548–560. doi: 10.1093/oxfordhb/9780199933624.013.33
Rogers, R. (2013). Digital Methods. The MIT Press. Available online at: http://www.amazon.com/Digital-Methods-Richard-Rogers/dp/0262018837. doi: 10.7551/mitpress/8718.001.0001
Salathé, M., Kazandjieva, M., Lee, J. W., Levis, P., Feldman, M. W., and Jones, J. H. (2010). A high-resolution human contact network for infectious disease transmission. Proc. Natl. Acad. Sci. U.S.A. 107, 22020–22025. doi: 10.1073/pnas.1009094108
Sapru, A., and Bourlard, H. (2015). Automatic recognition of emergent social roles in small group interactions. IEEE Trans. Multimedia 17, 746–760. doi: 10.1109/TMM.2015.2408437
Savage, M., and Burrows, R. (2007). The coming crisis of empirical sociology. Sociology 41, 885–899. doi: 10.1177/0038038507080443
Schmid Mast, M., Gatica-Perez, D., Frauendorfer, D., Nguyen, L., and Choudhury, T. (2015). Social sensing for psychology: automated interpersonal behavior assessment. Curr. Dir. Psychol. Sci. 24, 154–160. doi: 10.1177/0963721414560811
Sekara, V., and Lehmann, S. (2014). The strength of friendship ties in proximity sensor data. PLoS ONE 9:e100915. doi: 10.1371/journal.pone.0100915
Smieszek, T., Castell, S., Barrat, A., Cattuto, C., White, P. J., and Krause, G. (2016). Contact diaries versus wearable proximity sensors in measuring contact patterns at a conference: method comparison and participants' attitudes. BMC Infect. Dis. 16:341. doi: 10.1186/s12879-016-1676-y
Stehlé, J., Charbonnier, F., Picard, T., Cattuto, C., and Barrat, A. (2013). Gender homophily from spatial behavior in a primary school: a sociometric study. Soc. Netw. 35, 604–613. doi: 10.1016/j.socnet.2013.08.003
Stopczynski, A., Pietri, R., Pentland, A., “Sandy,”, Lazer, D., and Lehmann, S. (2014). Privacy in Sensor-Driven Human Data Collection: A Guide for Practitioners. Available online at: http://arxiv.org/abs/1403.5299
Taylor, C. J., Freeman, L., Olgúin-Olgúin, D., and Kim, T. (2016). “Deviation in voice pitch as a measure of physiological stress response to group processes,” in Advances in Group Processes, eds S. R. Thye and E. J. Lawler (Bingley: Emerald Group Publishing, Limited), 211–242. doi: 10.1108/S0882-614520160000033008
Taylor, E. B., and Horst, H. A. (2013). From street to satellite: mixing methods to understand mobile money users. Ethnogr. Praxis Industry Conf. Proc. 2013, 88–102. doi: 10.1111/j.1559-8918.2013.00008.x
Tinati, R., Halford, S., Carr, L., and Pope, C. (2014). Big Data: methodological challenges and approaches for sociological analysis. Sociology 48, 663–681. doi: 10.1177/0038038513511561
Tonidandel, S., King, E. B., and Cortina, J. M. (2018). Big Data Methods: leveraging modern data analytic techniques to build organizational science. Organ. Res. Methods 21, 525–547. doi: 10.1177/1094428116677299
Tripathi, P., and Burleson, W. (2012). “Predicting creativity in the wild: experience sample and sociometric modeling of teams,” in Proceedings of the ACM 2012 Conference on Computer Supported Cooperative Work - CSCW'12 (New York, NY: ACM Press), 1203. doi: 10.1145/2145204.2145386
Ubaldi, E., Vezzani, A., Karsai, M., Perra, N., and Burioni, R. (2017). Burstiness and tie activation strategies in time-varying social networks. Sci. Rep. 7:46225. doi: 10.1038/srep46225
U. S. Government GPS.gov. GPS Accuracy. Available online at: https://www.gps.gov/systems/gps/performance/accuracy/ (accessed March 28, 2019).
Vanhems, P., Barrat, A., Cattuto, C., Pinton, J.-F., Khanafer, N., Régis, C., et al. (2013). Estimating potential infection transmission routes in hospital wards using wearable proximity sensors. PLoS ONE 8:e73970. doi: 10.1371/annotation/b20d3cec-62b7-44ec-9150-8a06a9b30a9b
Vinciarelli, A., Pantic, M., and Bourlard, H. (2009). Social signal processing: survey of an emerging domain. Image Vis. Comput. 27, 1743–1759. doi: 10.1016/j.imavis.2008.11.007
Wang, R., Chen, F., Chen, Z., Li, T., Harari, G., Tignor, S., et al. (2014). “StudentLife: assessing mental health, academic performance and behavioral trends of college students using smartphones,” in Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing UbiComp'14 (New York, NY: ACM), 3–14. doi: 10.1145/2632048.2632054
Wenzel, R., and Van Quaquebeke, N. (2018). The double-edged sword of big data in organizational and management research. Organ. Res. Methods. 21, 548–591. doi: 10.1177/1094428117718627
Wu, L., Waber, B. N., Aral, S., Brynjolfsson, E., and Pentland, A. (2008). Mining face-to-face interaction networks using sociometric badges: predicting productivity in an IT configuration task. SSRN Electron. J. doi: 10.2139/ssrn.1130251
Wyatt, D., Bilmes, J., Choudhury, T., and Kitts, J. A. (2008). “Towards the automated social analysis of situated speech data,” in UbiComp '08 Proceedings of the 10th International Conference on Ubiquitous Computing (New York, NY: ACM Press), 168–171. doi: 10.1145/1409635.1409658
Wyatt, D., Choudhury, T., Bilmes, J., and Kitts, J. A. (2011). Inferring colocation and conversation networks from privacy-sensitive audio with implications for computational social science. ACM Trans. Intell. Syst. Technol. 2, 1–41. doi: 10.1145/1889681.1889688
Yano, K., Akitomi, T., Ara, K., Watanabe, J., Tsuji, S., Sato, N., et al. (2015). Measuring happiness using wearable technology. Hitachi Rev. 64, 97–104. Available online at: http://www.hitachi.com/rev/pdf/2015/r2015_08_116.pdf (accessed May 13, 2019).
Yu, D., Blocker, R. C., Sir, M. Y., Hallbeck, M. S., Hellmich, T. R., Cohen, T., et al. (2016). Intelligent emergency department: validation of sociometers to study workload. J. Med. Syst. 40:53. doi: 10.1007/s10916-015-0405-1
Zhang, Y., Olenick, J., Chang, C.-H., Kozlowski, S. W. J., and Hung, H. (2018). “TeamSense: assessing personal affect and group cohesion in small teams through dyadic interaction and behavior analysis with wearable sensors,” in Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (New York, NY: ACM), 2, 1–22. doi: 10.1145/3264960
Keywords: mixed methods research, wearable sensors, Bluetooth, indicator, qualitative research, quantitative research, big data
Citation: Müller J, Fàbregues S, Guenther EA and Romano MJ (2019) Using Sensors in Organizational Research—Clarifying Rationales and Validation Challenges for Mixed Methods. Front. Psychol. 10:1188. doi: 10.3389/fpsyg.2019.01188
Received: 21 December 2018; Accepted: 06 May 2019;
Published: 24 May 2019.
Edited by:
M. Teresa Anguera, University of Barcelona, SpainReviewed by:
Stefan Volk, University of Sydney, AustraliaNuria Codina, University of Barcelona, Spain
Copyright © 2019 Müller, Fàbregues, Guenther and Romano. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jörg Müller, jmuller@uoc.edu