Academia.eduAcademia.edu

Measuring Interpersonal Communication

2016, International encyclopedia of interpersonal communication

Measurement is a core activity and challenge in all social sciences, and the complex nature of interpersonal communication makes it particularly challenging to measure. This entry provides a discussion of basic concepts in measurement and an introduction to the myriad considerations to be made in measuring constructs in interpersonal communication. It includes a review of the advantages and disadvantages of the most common techniques for measuring interpersonal communication, including observational assessments, self-reports, diaries and logs, and physiological assessments. In addition to the perennial challenges of measuring in interpersonal communication scholarship, this entry notes some of the promises and difficulties implicit in the rise of new communication technologies in interpersonal interaction.

Measuring Interpersonal Communication John P. Caughlin Erin Basinger University of Illinois at Urbana-Champaign Caughlin, J. P., & Basinger, E. D. (2016). Measuring interpersonal communication. In C. R. Berger & M. E. Roloff (General Eds.) & S. R. Wilson, J. P. Dillard, J. P. Caughlin, & D. H. Solomon (Assoc. Eds.), International encyclopedia of interpersonal communication (pp. 1-13). New York, NY: Wiley. doi: 10.1002/9781118540190.wbeic0229 Final Draft: This version may differ from published version due to copy editing. Abstract Measurement is a core activity and challenge in all social sciences, and the complex nature of interpersonal communication makes it particularly challenging to measure. This entry provides a discussion of basic concepts in measurement and an introduction to the myriad considerations to be made in measuring constructs in interpersonal communication. It includes a review of the advantages and disadvantages of the most common techniques for measuring interpersonal communication, including observational assessments, self-reports, diaries and logs, and physiological assessments. In addition to the perennial challenges of measuring in interpersonal communication scholarship, this entry notes some of the promises and difficulties implicit in the rise of new communication technologies in interpersonal interaction. Measuring concepts is essential to the systematic study of interpersonal communication. Although sound measurement is a foundation of good social science, it can be extremely complex. Communication scholars are typically interested in abstract concepts such as affinity-seeking, attributions, uncertainty, and many other constructs explicated in this encyclopedia. Although communication theory and research focus on abstract constructs, measurement of such concepts requires the use of empirical variables that provide some indication about the amount or levels of the constructs of interest. What researchers study directly are the variables that are meant to represent constructs, rather than the constructs themselves (Capella, 1991). Finding empirical variables to assess concepts is far from simple. Because “abstract constructs do not have one-to-one correspondence with empirical indicants…abstract concepts can only be approximated by empirical indicants” (Zeller & Carmines, 1980, p. 3). The notion that measurement is an approximation of constructs is important to keep in mind when considering research on communication. Scholars sometimes claim that their measures are objective, and this can be true in the sense that some variables are based on observable and discrete information that anyone could see (or be trained to see). For instance, one behavior that has been coded in interpersonal conflict is eye rolling, and with high quality video, coders can discern a precise and objective number of times people roll their eyes during a particular segment. Yet, readers should remember that having an objective behavior to count does not change the fact that the empirical indicant is still an approximation (or, more likely, one indicator among several that collectively serve as an approximation). Few communication scholars would be interested in counting eye rolls as an end in itself; instead, examining how often people roll their eyes nearly always would reflect an attempt to assess some broader theoretical construct. Pointing out that the scholarly interest is usually in some broader construct, rather than in the empirical variables themselves, is important because it highlights the fact that even when communication behaviors can be counted objectively, the meaning of those behaviors is not necessarily clear or straightforward. For instance, the meaning of behaviors during communication is greatly influenced by context; thus, even if one determines that a particular behavior in a particular situation provides useful information about a specific concept, that same behavior could mean something else in another context. Consider eye rolling behavior, for example. If it occurs in response to a relational complaint that one has heard many times (e.g., “you never do your fair share around the house”), rolling one’s eyes is probably an indicator of contempt or exasperation. Yet if the same behavior is in response to a complaint about a third party (e.g., “can you believe your mother said that to me”), the same eye roll probably would indicate solidarity with the partner. Assessing measures Because sound measurement is so crucial to systematic studies of communication, determining whether measures are appropriate for the purposes at hand is one of the fundamental aspects of the research process. Given that measures usually are merely indications or approximations of constructs, no measure is perfect. Nevertheless, some measures are better than others for particular purposes, and there are a number of basic concepts and considerations that help scholars decide whether a measure is suitable. One key aspect of measurement is conceptualization, which involves thinking about a potential construct and making it clear and precise enough that the concept can be assessed. This involves not only defining what the concept is but also distinguishing it from other potentially related constructs so that any resulting measure can be a good assessment of that construct alone (and not confounded with other constructs). Conceptualization is an inherently interpretive process that is often implicit, but it can be made more explicit and systematic if it is based on a clear theoretical perspective. Operationalization is the process of moving from conceptualization to variables that one can measure. It involves specifying qualities or characteristics that can actually be assessed and determining what will comprise the actual measure of a construct. Although any operationalization will be an approximation of the construct, the goal is to design a measure that captures as much of the breadth of the concept as possible without including indicators that are closely associated with other constructs. Once a measure is developed and used in research, it can be evaluated in terms of reliability and validity. Technically, reliability refers to the probability that if a measure is a concept is taken repeatedly, it will yield the same or a similar value. The desirability of reliability is obvious; if an assessment can yield many wildly different values for the same construct when there is no reason to believe the underlying construct has changed, it would be unwise to trust the results of the measurement. If a thermometer produced temperatures ranging from 96 to 104 degrees Fahrenheit on the same patient in repeated measurements of body temperature, one likely conclusion is that thermometer is broken (that is, no longer reliable). It is theoretically possible, of course, that a human’s temperature could cycle rapidly through extreme temperatures, but if a new thermometer showed steady readings, it would be safe to conclude that the first thermometer is unreliable and not to be trusted. Of course, the case of a thermometer varying by as much as 8 degrees is an extreme one; social scientific measures generally do not have perfect reliability. Some degree of unreliability must be tolerated. The extent to which a measure can show unreliability before it should not be trusted is a judgment call that is informed by some rules of thumb in the research community, but there is no definitively correct answer to the question of exactly how much unreliability is too much for a particular measure in a particular study. Although the notion of reliability involves getting similar results in repeated measurements, reliability is usually inferred in research by comparing values on related variables that were measured at roughly the same time. In a multi-item scale, for example, if all the items are highly correlated with each other or if subsets of the items correspond well with other subsets of items, one can infer that the scale has reliability. If a measure is based on people rating or coding communication behaviors, then the scores from multiple people should correspond highly to indicate reliability in the assessment. Even if a measure is sufficiently reliable, it is not necessarily a good instrument. Reliability is about consistency, and a measure can be consistently wrong. That is, even a measure with high reliability can lack validity. Validity concerns the correspondence between how a construct is conceptualized and how it is operationalized (Levine, 2011). The operationalizations of a construct are incomplete or imperfect reflections of the underlying construct, but the connections between construct and operationalization are based on inferences of varying degrees and soundness (Cappella, 1991). Because operationalization involves a translation of a concept into a measurable indication of the construct, validity is not a binary notion. Measures are not either valid or invalid; instead, they vary in the extent to which they are valid. Evaluating the validity of a measure involves gathering evidence that can allow one to make inferences about validity. For example, one can infer that a measure is sufficiently valid if it is correlated in the expected ways with measures of other constructs or if it is correlated with other measures of the same construct that are thought to be valid enough for particular research purposes (Singleton & Straits, 2005). In addition to comparing a measure to other measures for indications of validity, researchers also attempt to design measures that seek to minimize various factors that are known to threaten validity. One well-known threat to some measures is the set of biases that influence how people recall communication phenomena. Such threats can be minimized (although not necessarily eliminated) by taking steps such as asking people about recent events rather than ones that happened long ago and by phrasing questions in ways that reflect how people perceive phenomena rather than asking them to engage in a unfamiliar mental calculus (Huston & Robins, 1982). A number of textbooks and scholarly articles provide thorough lists and descriptions of common threats to validity and ways to minimize those threats (e.g., Frey, Botan, & Kreps, 2000; Huston & Robins, 1982). Common measurement techniques in interpersonal communication research In response to the fact that measures of communication constructs are imperfect translations of the underlying constructs of interest, researchers have developed a number of different measurement techniques. Each technique has strengths and weaknesses and is suited better for some purposes than others (Feeney & Noller, 2012). Because no technique is perfect and all have different strengths, the ideal is for research to utilize multiple measurement techniques for assessing interpersonal communication constructs (Caughlin & Basinger, 2014; Feeney & Noller 2012). This need not occur in all or even most specific studies, but there is value to using diverse measurement techniques to address important research questions. If findings based on multiple measurement techniques coalesce around similar conclusions, the combined evidence is more compelling than would be possible with strict replications using limited techniques. If, on the other hand, findings are consistently different depending on techniques, it can highlight previously unrecognized questions and insights. It is possible, for example, that one type of measure would reflect one aspect of a theoretical construct whereas another type of measure would be indicative of a distinct facet of the same construct. In such an instance, the systematic examination of a construct with multiple techniques could lead researchers to learn that a concept is more complex or multifaceted than previously understood. Observations One common technique for assessing interpersonal communication is using observations. To obtain measures based on observations, researchers must be able to observe a communication segment of interest, and they must assess what they observe in a way that allows them to derive variables from those observations. Although these two steps can happen simultaneously (e.g., by having judges watch ongoing conversations and rate them on some characteristics), it is more common for researchers to create an audio or video record of the interaction, allowing for a careful assessment of the interactions later. Although there is a wide range of possibilities for observing interaction, observational research frequently involves having participants in a study report to a laboratory and asking them to engage in some kind of interaction. The tasks people are given are intended to stimulate communication that is relevant to the research questions of interest. Scholars interested in relational conflict, for example, probably would find it exceedingly inefficient to simply wait for couples in a laboratory to spontaneously have conflicts; thus, conflict researchers provide instructions that result in a conversation that simulates or reflects conflicts that couples might have in their everyday lives. For instance, countless investigations have been based on having couples identify issues of conflict in their relationship and then discussing those issues in a laboratory. There are a number of advantages of observations in a laboratory. Researchers can control the environment and ensure that the participants produce a segment of interaction that is relevant to the research agenda. It also allows researchers to record the interactions in a way that allows for extremely precise or detailed analysis later. Because of these features, laboratory observations are often considered the “gold standard” for data about interpersonal communication. In spite of the important advantages of observing interpersonal communication in a laboratory setting, such methods also have some drawbacks. Because the usual procedure involves stimulating a conversation, it is not clear how much the conversations are representative of interpersonal communication that occurs spontaneously outside the laboratory. Moreover, in laboratory settings individuals know that they are being observed, which can influence their behavior in a number of ways, such as heightening their desire to manage a positive impression. Reis (1994) has even argued that the typical laboratory setting is so distinct from natural experiences that behavior enacted in such settings is best thought of as exemplary rather than typical. Although not everyone in a laboratory setting engages in the optimal behavior, Reis noted that people do try to optimize their behavior when they know they are being observed, even if they are not always successful at making their behavior seem ideal. The potential issues with laboratory observations are important because so many observational studies occur in such settings, but it is important to recognize that some observations can be made in ways that minimize some of the problems of laboratories. For example, Sillars (1991) reported that in direct comparisons of video recorded laboratory observations to audio recordings done at home, the audio recordings at home were much more “spontaneous and realistic” (p. 206). Moreover, even scholars best known for tightly controlled studies of couples’ conflicts in laboratories have recognized the limits of such studies and have augmented laboratory research with studies in which people were observed in apartment settings for long periods of time (e.g., 24 hours), which diminishes the effects of observation (e.g., Hawkins, Carrere, & Gottman, 2002). Another promising development in observational research on interpersonal communication involves the rise of new communication technologies that people are using for their everyday interaction. Many newer modes of communication are recorded by default. Communication by text, via social media, or by email is often conducted in a manner in which the entire interaction is documented by default. If people are willing to provide researchers access to these records, it can be a rich source of data that was generated entirely apart from the influences that researchers’ observations often have. Of course, this rich new source of observational data has its own drawbacks; for instance, if generated naturally, it may or may not evince the phenomenon of interest to the researchers. Researchers can, of course, give participants tasks that would stimulate them to engage in the types of interactions under investigation, but as soon as they are directed to interact via technologies in particular ways, the interactions become influenced by the research design, just as is the case with traditional laboratory studies. In sum, interaction data can be collected in a variety of ways, ranging in how naturalistic it is and how focused it is on the type of interaction that is of interest to the researchers. Because there are many research questions and purposes, there is no single best way to gather observational data. Scholars may find it useful to use different approaches at different times, even within the same research program. Regardless of how the interaction data are collected, gaining access to such data is only the first step in observational research. Next, the researchers must find some way to code or rate the data so that variables can be drawn from the conversations. As with any measurement issue, the researcher is usually interested in constructs that might be evident in the data rather than an overt behavior in and of itself. For example, a researcher studying conflict might be interested in the construct of avoidance. The notion of avoidance is not something that can be directly observed, but there are behaviors that scholars have inferred reflect avoidance during conflict. For instance, if a person changes the subject during a conflict discussion, it often indicates that this person is trying to avoid the topic. This is a reasonable inference, but notice that it is still an inference and not a direct measure of avoidance. Because the meaning and importance of communication behaviors are contextual, one can make mistakes assuming that there is a direct one-to-one correspondence between communication constructs and behavioral indicants of those constructs. For instance, Sillars (1986) noted that in some conversations about potential conflict topics, there are circumstances in which the conflict has already been resolved or is simply not an issue for certain couples. In such cases, changing the subject simply reflects a lack of need to talk about that issue, and it would be a mistake to call that conflict avoidance. It is important to be cognizant of the interpretive aspect of observational assessment of interpersonal communication. Because observational records of data allow for objective counting of the indicants, some scholars assume that observational measures are also objective. If a researcher is truly only interested in the directly observable manifestation of communication (e.g., how many times a person changes the subject, how many times a person rolls his or her eyes), then the measure can fairly be called objective. Yet, typically researchers wish to study underlying constructs (e.g., conflict avoidance, expression of contempt) that are not directly observable and therefore must be inferred from indicants. Once such inferences are required, the measure is no longer entirely objective, and it is important to scrutinize the extent to which the inferences are sound. A careful reader of observational studies will notice that sometimes the inferences researchers (or other readers) make about the meaning of behaviors are actually quite debatable. Moreover, because the meaning of communication behaviors is contextual, there is always some risk in making inferences about what those behaviors mean from an outside perspective, particularly when the interpersonal communication is between longtime relational partners who have had time to develop idiosyncratic patterns of interacting and a history of discussing issues that imbues the discussion of current topics with meaning developed over previous encounters. In short, for a variety of reasons, one should not confuse collecting an objective sample of interaction data with being able to create objective measures of interpersonal communication. Retrospective self-reports. Retrospective self-reports are another common method for gathering data about interpersonal communication. In this technique, research participants are asked to reflect on communication events or experiences that have already occurred and then provide information about those occurrences. Retrospective self-reports can be used to gather data about a wide variety of phenomena, including reports about the frequency or quality of one’s own or another person’s behavior, the thoughts or experiences one had during communication episodes, and evaluations of communication or the people involved in a conversation. Retrospective reports are also quite flexible in terms of timeframe. Researchers have asked participants to report on events that happened recently as well as ones that happened long ago; for instance, a number of studies have asked adults to report on the communication patterns that they experienced in their family of origin when they were growing up. Researchers also have asked about very specific episodes (e.g., a memorable conversation) as well as broad patterns of communication that occur over longer periods of time (e.g., family communication patterns). Retrospective reports, in short, can be used to assess a very wide range of variables related to interpersonal communication. The flexibility and range allowed by retrospective self-reports is a key advantage of this technique. There are many important communication phenomena that are difficult to observe, and many more that are altered so much by being observed that it is not clear what the observations actually indicate. Research using retrospective self-reports in the two hours following sexual activity, for example, has shown that women who report experiencing orgasm report disclosing more about themselves afterwards than do women who do not report orgasm (Denes, 2012). Although the same research questions could conceivably be examined with observational methods, there would be numerous practical obstacles; for example, the number of people willing to participate in a study that observed sexual activity as it unfolds (especially one that included assessments of orgasm other than self-reports ones) may be so limited that one would wonder about the generalizability of the results. Even if one did find a representative sample of participants willing to be observed, it is likely that the sexual activity would need to be planned in such a way that it would introduce concerns about whether the activity was representative of sexual activity. Even apart from the sexual activity, there are reasons to believe that the self-disclosures would be altered if participants knew that their disclosures were being recorded. Yet it is possible to gather self-report data about such phenomena, especially if anonymity or confidentiality is assured. According to Feeney and Noller (2012), “the limitations of self-report questionnaires are well known (in fact, they have been more widely acknowledged than the limitations of other methodologies such as observation)” (p. 30). Most often, scholars focus on problems that can lead to inaccurate or biased responses, such as those caused by the inability to remember what happened or the desire to make a good impression. There are also known and systematic errors in judgment that lead people frequently to be more negative in their reports of others’ behaviors than they are with reports of their own behaviors. Such limitations are inherent to all self-report measures. Yet just because self-report measures are inherently limited does not mean that they are all equally problematic. There are certain threats to validity that can be minimized—or amplified—when using self-report measures. The ability to report accurately on communication behavior can be diminished in a number of circumstances, such as when participants are asked to report on events from long ago, when the behaviors they are asked to report on or the metrics used for the items are vague and unspecific, or when answering the question requires mental calculations that deviate from how participants spontaneously think about such events. Consider, for instance, the difference between the following two questions: “Did you send a text message to your partner since you came to the lab five minutes ago?” and “Over the past year, how often did you communicate with your partner each week on average?” The former question asks about a very specific behavior, a short and recent timeframe, and does not require any calculations. Unless they have some reason to overtly lie to the researchers, most participants could answer the former question accurately. The latter question, in contrast, requires participants to think back over the past year, reflect on an average across any waxing and waning in communication, and figure out how to define what the word often means (e.g., if they see each other twice daily but send hundreds of texts back and forth all day, how many times have they communicated?). It is unlikely that this question would be a particularly accurate reflection of how many times the participants communicated no matter how the researchers were defining what counted as communicating. Because retrospective self-reports vary considerably in how biased they are, it is important to examine the specific measures and goals of the research before deciding whether a retrospective report is appropriate. For instance, there are times when researchers believe that people’s perceptions of their communication behaviors are more important than an accurate count of such behaviors. Moreover, if the communication phenomenon is one that cannot be easily inferred from observable behaviors, a self-report can be more accurate than an observational measure. If one is studying the keeping of secrets, for example, it may be possible to infer that some observable behaviors indicate that someone is keeping a lot of secrets, but those behaviors undoubtedly would be open to multiple interpretations and would be influenced by context. In contrast, an individual asked to report on secret keeping needs to make far less challenging inferences to be able to provide data on the subject. To put this another way, one strength of many retrospective report measures is that the inferential leap between indicants (such as items on a scale) and a construct can be much smaller than is the case with observational methods. Thus, retrospective measures can address one of observational measures’ biggest weaknesses, just as observational measures address the most obvious limitation of retrospective measures: bias. Because they have complementary strengths and weaknesses, retrospective self-reports and observations often can be used productively together to provide different viewpoints on important research questions. Diaries and logs Diaries and logs are self-report measures, but they are usually treated as a separate type of measure because they involve getting reports of the same experiences or behaviors soon after they occur (or even during), and the reports are gathered repeatedly over time. Diary and log data tend to refer to relatively short periods of time that are presumably accessible to participants. For example, rather than asking about behaviors over a two week period in general, a diary approach could entail having participants report on what happened on a particular day and then repeat the diary each day for two weeks. Because participants are being asked to consider a fairly short and recent time period, some of the biases associated with retrospective self-reports are minimized. Not all of such biases are reduced or eliminated, however. If the questions are about stigmatized behaviors, for instance, there is no reason to believe that people would be less likely to obfuscate in diary reports than they are with retrospective ones. By aggregating across days, researchers can create summary variables that are conceptually equivalent to retrospective reports of frequency but are probably more accurate because the recall time is minimized and because the researchers do the calculations rather than asking participants to make estimates such as the average number of times per day that an event occurred. Moreover, collecting data on the same experience repeatedly allows for an examination of questions that are difficult to assess with retrospective reports. For example, diary studies have shown that the patterns of marital interaction vary according to the day of the week, but it is not clear that such temporal rhythms would be apparent to participants, even if the researchers could pose clear retrospective questions about such temporal rhythms. Physiological measures Perhaps the fastest growing category of assessment technique in the study of interpersonal communication is physiological measurement. Because there is a separate entry on such measures, they are covered only briefly here to contextualize them within other types of measures. There is a wide variety of possible physiological measures, ranging from cardiovascular characteristics, hormones, brain activity, and so forth. At least one reason for the rapid rise in the use of physiological measures is the technologies to assess physiology have improved greatly in terms of lower cost and intrusiveness. One of the biggest advantages of physiological measures in general is they can provide data about communicators that would be very difficult for participants to report on or to observe without biological markers. For instance, cortisol levels and other hormones can indicate physiological stress, even when participants report not experiencing stress and there are no clearly discernable visible indicators of stress. Because they provide information about individuals who are communicating that cannot be discerned by other means, physiological measures are likely to continue becoming more prominent over time. Yet like any type of measure, it is important to keep in mind what such measures can and cannot do. Writing specifically about fMRI data, Aue, Lavelle, and Cacioppo (2009) argued that scholars often are too extreme in their views of physiological data, with some dismissing it entirely and others believing it will reveal all the mysteries of psychological experiences. In reality, physiological measures are very much like observations of behavioral data. The information that is gathered is different, but it usually is gathered directly by researchers, rather than via reports from participants. The link between the assessments and broader theoretical constructs is one that must be inferred by the researchers. Thus, although the indicators in physiological data are different than with traditional observational data about communication, some of the strengths and weaknesses are analogous. Physiological assessments can be made without respondent biases, and this lack of bias in the assessment itself creates a sense of objectivity to the research. Yet what the assessments actually mean must be inferred, and in many instances the inferences are unclear, or at least not as unequivocal as implied by some researchers (Aue et al., 2009). A particular physiological response, for instance, may be an indicator of a particular experience, but that same response often can indicate other experiences as well. That is, the connection between physiological data and what it actually means to people and their communication is not simplistic, which makes the interpretation of such data complex. Researchers can address some of these challenges by carefully examining their inferences to make sure they are as sound as possible; for instance, inferences can be improved by examining multiple physiological indicators together and by recognizing that the connection between physiology and psychology is not “an invariant relationship” (Aue et al., 2009, p. 13). Key challenges to measuring interpersonal communication The discussion to this point has focused on various measurement techniques that are common in the study of interpersonal communication. Given that each has strengths and weaknesses and that the strengths often complement each other, one obvious implication is that multiple types of measures should be used to address important questions in interpersonal communication research. Beyond the consideration of classes of measures, however, there are some larger conceptual issues that are important to consider. Implications of new communication technologies It is now widely recognized that interpersonal communication happens frequently via mediated channels. Whether texting, talking on mobile phones, or using social media, it is undeniable that some people frequently use technologies as they interact with people close to them. For instance, people now utilize communication technologies in a variety of ways during their relational conflicts (Caughlin, Basinger, & Sharabi, in press). Some individuals claim that they look up information on their smartphones to help resolve differences; others report that they consult their phones for evidence in text messages that can prove their point (e.g., “see, you DID say that you would definitely be there”). Some individuals claim that by looking at a phone or other screen, they are able to gather their thoughts so they can say the right thing during conflict; others say they look at their phones to avoid having to look at the other person. Although we are just beginning to understand what the ubiquity of such devices means for interpersonal communication, it is clear that many people are using them as tools for communicating, and this has potentially important implications for assessing interpersonal communication. The traditional “gold standard” for studies of conflict, for instance, has been to have dyads come to a laboratory and discuss conflict issues. Whether explicitly proscribed or not, the nature of the studies makes it exceedingly unlikely that people would incorporate their communication technologies into their discussions. Yet if that is something they typically do when conflicts arise, not encouraging them to have their communication tools makes the traditional laboratory study even less naturalistic to these participants than it was before such communication devices became widespread. Because communication technologies have become part of interpersonal communication, our assessments of interpersonal communication need to do more to take this into account. Although we still know very little about the use of communication technologies in everyday interpersonal communication, the lessons of research on observations of other behaviors provide some insight into what we should consider when trying to assess the use of communication technologies in interpersonal communication. We know, for example, that the connection between behaviors and meaning is not one-to-one, and we would expect that the same would be the case for the use of technologies. For example, understanding the role of the communication technologies is likely to involve more nuanced conceptualization than would be implied by simply counting behaviors, such as how often one looks at a smartphone during a discussion. As the examples above illustrate, one can look at a phone to facilitate attempts at constructive conflict, one can look at a phone to gain evidence for a distributive strategy, or one can look at the phone to avoid the conflict. It would be fairly easy to count an objective behavior such as how often a person looked at a smartphone, but such counts probably would not tell us much without careful consideration of what that behavior means (and can mean). Consequently, as with other measures of interpersonal communication, before researchers start counting behaviors related to technology use, it will be important to do the theoretical work that helps identify key constructs. Measure selection and justification Measures are never inherently valid; they are instead more or less valid for a given purpose. As such, evaluating the relative validity of a survey measure, coding scheme, or any measure requires at least implicitly evaluating an argument about the suitability of the measure for a specific purpose. A word of caution is warranted, then, about the selection and/or adaption of measures for one’s own study that were designed for other purposes. For instance, psychologists have designed a number of assessments of personality or relationship traits that are of interest in interpersonal communication research (e.g., narcissism, depression). Although these measures may be valid for clinical assessment, they may not be appropriate for non-clinical samples. Or, measures developed to assess relationship characteristics between spouses may not make sense to use in studies of parent-child relationships. In cases like these, the importance of face validity, carefully assessing whether a measure actually seems to capture the concept of interest (and nothing else), cannot be overestimated. There must be a logic behind the selection of measurement tools beyond their prolific use in previous research. A separate, but related issue, is that researchers sometimes operationalize the same concept in multiple ways. Conflict, for example, may be operationalized by assessing frequency, intensity, management, satisfaction, or a number of other characteristics of conflict-laden interactions. Certainly, each of these may be “valid” representations of conflict, but it is worth noticing that if different scholars are conceptualizing the same word in different ways, they might actually be assessing slightly different constructs or different aspects of a larger construct. Thus, even when different scholars purport to be assessing the same construct, it may not be possible to compare the results of the various studies because the conceptualizations or operationalizations may be vastly different. Consequently researchers should be careful about which measure of a construct they choose; a case should be made for choosing one assessment tool over the other, based on the purpose of the study. The goals of one’s own research and the purpose of the measure in question should always be of primary interest in making an argument for the validity of a measurement tool. Another frequently cited piece of evidence for validity involves statistical tools. Although communication researchers tend to report and interpret statistics in similar ways, our rules of thumb for providing “proof of validity” should not be followed blindly. Rather, researchers should pay attention to the theoretical meaning behind their statistically-based decisions. There is sometimes a good theoretical justification for retaining items in a factor that do not load cleanly with other items or for separating items that do correlate highly. For instance, Thompson and Walker (1982) have pointed out that spouses’ attitudes generally correlate highly, but they should still be treated as separate scores because attitudes are individually formed, rather than characteristics of a couple. Statistical tools sometimes provide us with insight into our data, but basing decisions purely on the math behind these tools may lead us to make illogical or invalid decisions. In sum, the goal in the selection of a measurement tool ought to be related to the purpose of the study, and the argument for the validity should include a balance of reason, theory, research, and available statistical analyses. SEE ALSO: Attributions in Conflict; Conflict Avoidance; Conflict Styles and Strategies; Family Typologies; Measurement Bias; Physiological Measurement References Aue, T., Lavelle, L. A., & Cacioppo, J. T. (2009). Great expectations: What can fMRI research tell us about psychological phenomena? International Journal of Psychophysiology, 73, 10-16. doi: 10.1016/j.ijpsycho.2008.12.017 Cappella, J. N. (1991). Mutual adaptation and relativity of measurement. In B. M. Montgomery & S. Duck (Eds.), Studying interpersonal interaction (pp. 103-117). New York, NY: Guilford Press. Caughlin, J. P., Basinger, E. D., & Sharabi, L. L. (in press). The connections between communication technologies and relational conflict: A multiple goals and communication interdependence perspective. In J. A. Samp (Ed.), Communicating interpersonal conflict in close relationships: Contexts, challenges and opportunities. New York: NY: Routledge. Caughlin, J. P., & Basinger, E. D. (2014). Measuring social interaction. In In P. J. Schultz and P. Cobley (Series Eds.) & C. R. Berger (Vol. Ed.), Handbooks of communication science: Vol 6: Interpersonal Communication. (pp. 103-126). Berlin, Germany: De Gruyter Mouton. Denes, A. (2012). Pillow talk: Exploring disclosures after sexual activity. Western Journal of Communication, 76, 91-108. doi: 10.1080/10570314.2011.651253 Feeney, J. A., & Noller, P. (2012). Perspectives on studying family communication: Multiple methods and multiple sources. In: Anita L. Vangelisti (Ed.), Routledge handbook of family communication (2nd ed., pp. 29-45). New York, NY: Routledge. Frey, L. R., Botan, C. H., & Kreps, G. L. (2000). Investigating communication: An introduction to research methods (2nd ed.). Boston, MA: Allyn and Bacon. Hawkins, M. W., Carrere, S., & Gottman, J. M. (2002). Marital sentiment override: Does it influence couples’ perceptions? Journal of Marriage and Family, 64, 193-201. doi: 10.1111/j.1741-3737.2002.00193.x Huston, T. L., & Robins, E. (1982). Conceptual and methodological issues in studying close relationships. Journal of Marriage and the Family, 44, 901-925. doi: 10.2307/351454 Levine, T. R. ( 2011). Quantitative social science methods of inquiry. In M. L. Knapp & J. A. Daly (Eds.), The SAGE handbook of interpersonal communication (4th ed., pp. 25-57). Los Angeles, CA: Sage. Reis, H. T. (1994). Domains of experience: Investigating relationship processes from three perspectives. In R. Erber & R. Gilmour (Eds.), Theoretical frameworks for personal relationships (pp. 87-110). Hillsdale, NJ: Erlbaum. Sillars, A. L. (1986, April). Procedures for coding interpersonal conflict (revised) (Manual). Missoula: University of Montana, Department of Interpersonal Communication. Sillars, A.L. (1991). Behavioral observation. In B.M. Montgomery & S. Duck (Eds.), Studying interpersonal interaction (pp. 197-218). New York, NY: Guilford Press. Singleton Jr., R. A., & Straits, B. C. (2005). Approaches to social research (4th ed.). New York, NY: Oxford University Press. Thompson, L. & Walker, A. J. (1982). The dyad as a unit of analysis: Conceptual and methodological issues. Journal of Marriage and the Family, 44, 889-900. doi: 10.2307/351453 Zeller, R. A., & Carmines, E. G. (1980). Measurement in the social sciences. New York, NY: Cambridge University Press. Further reading Baesler, E. J., & Burgoon, J. (1987). Measurement and reliability of nonverbal behavior. Journal of Nonverbal Behavior, 11, 205-233. doi: 10.1007/BF00987254 Metts, S., Sprecher, S. & Cupach, W. R. (1991). Retrospective self-reports. In B. M. Montgomery & S. Duck (Eds.), Studying interpersonal interaction (pp. 162-178). New York, NY: Guilford Press.