A Rigidity Detection System for Automated Credibility Assessment

30
Journal of Management Information Systems / Summer 2014, Vol. 31, No. 1, pp. 173–201. © 2014 M.E. Sharpe, Inc. All rights reserved. Permissions: www.copyright.com ISSN 0742–1222 (print) / ISSN 1557–928X (online) DOI: 10.2753/MIS0742-1222310108 A Rigidity Detection System for Automated Credibility Assessment NATHAN W. TWYMAN, AARON C. ELKINS, JUDEE K. BURGOON, AND JAY F. NUNAMAKER JR. NATHAN W. TWYMAN is a postdoctoral research scientist in the MIS Department at the University of Arizona, where he received his Ph.D. in MIS. His research interests span human–computer interaction, group support systems, virtual communities, health IS, and leveraging human and organizational factors in auditing, security, and forensic investigation systems. He has published articles in the Journal of Manage- ment Information Systems, Journal of the Association for Information Systems, and Information & Management. AARON C. ELKINS is a postdoctoral researcher in both the MIS Department at the Uni- versity of Arizona and the Department of Computing at Imperial College London. He holds a Ph.D. in MIS from the University of Arizona. He investigates how the voice, face, body, and language reveal emotion, deception, and cognition for advanced human–computer interaction and artificial intelligence applications. Complementary to the development of advanced artificial intelligence systems is their impact on the people using them to make decisions. Dr. Elkins also investigates how human deci- sion makers are psychologically affected by, use, perceive, and incorporate the next generation of screening technologies into their lives. JUDEE K. BURGOON is a professor of communications, a professor of family studies and human development, the director of human communication research for the Center for the Management of Information, and the site director of the Center for Identification Technology Research at the University of Arizona. She holds a Ph.D. in communica- tion and educational psychology from West Virginia University. Her research interests are in deception, trust, interpersonal interaction, and new technologies. JAY F. NUNAMAKER JR. is Regents and Soldwedel Professor of MIS, Computer Science and Communication and Director of the Center for the Management of Information and the National Center for Border Security and Immigration at the University of Arizona. He received his Ph.D. in operations research and systems engineering from Case Institute of Technology, an M.S. and B.S. in engineering from the University of Pittsburgh, and a B.S. from Carnegie Mellon University. He received his professional engineer’s license in 1965. Dr. Nunamaker was inducted into the Design Science Hall of Fame in May 2008 and received the LEO Award for Lifetime Achievement from the Association for Information Systems (AIS) in December 2002 and was elected a fellow of the AIS in 2000. He was featured in the July 1997 issue of Forbes Magazine on technology as one of eight key innovators in information technology. He is widely published, with an H index of 60. His specialization is in the fields of system analysis and design, collaboration technology, and deception detection. The commercial product GroupSystems’ ThinkTank, based on Dr. Nunamaker’s research, is often referred to

Transcript of A Rigidity Detection System for Automated Credibility Assessment

Journal of Management Information Systems / Summer 2014, Vol. 31, No. 1, pp. 173–201.

© 2014 M.E. Sharpe, Inc. All rights reserved. Permissions: www.copyright.com

ISSN 0742–1222 (print) / ISSN 1557–928X (online)

DOI: 10.2753/MIS0742-1222310108

A Rigidity Detection System for Automated Credibility Assessment

NAthAN W. tWyMAN, AArON C. ElkINS, JuDEE k. BurgOON, AND JAy F. NuNAMAkEr Jr.

NathaN W. tWymaN is a postdoctoral research scientist in the MIS Department at the university of Arizona, where he received his Ph.D. in MIS. his research interests span human–computer interaction, group support systems, virtual communities, health IS, and leveraging human and organizational factors in auditing, security, and forensic investigation systems. he has published articles in the Journal of Manage-ment Information Systems, Journal of the Association for Information Systems, and Information & Management.

aaroN C. ElkiNs is a postdoctoral researcher in both the MIS Department at the uni-versity of Arizona and the Department of Computing at Imperial College london. he holds a Ph.D. in MIS from the university of Arizona. he investigates how the voice, face, body, and language reveal emotion, deception, and cognition for advanced human–computer interaction and artificial intelligence applications. Complementary to the development of advanced artificial intelligence systems is their impact on the people using them to make decisions. Dr. Elkins also investigates how human deci-sion makers are psychologically affected by, use, perceive, and incorporate the next generation of screening technologies into their lives.

JudEE k. BurgooN is a professor of communications, a professor of family studies and human development, the director of human communication research for the Center for the Management of Information, and the site director of the Center for Identification technology research at the university of Arizona. She holds a Ph.D. in communica-tion and educational psychology from West Virginia university. her research interests are in deception, trust, interpersonal interaction, and new technologies.

Jay F. NuNamakEr Jr. is regents and Soldwedel Professor of MIS, Computer Science and Communication and Director of the Center for the Management of Information and the National Center for Border Security and Immigration at the university of Arizona. he received his Ph.D. in operations research and systems engineering from Case Institute of technology, an M.S. and B.S. in engineering from the university of Pittsburgh, and a B.S. from Carnegie Mellon university. he received his professional engineer’s license in 1965. Dr. Nunamaker was inducted into the Design Science hall of Fame in May 2008 and received the lEO Award for lifetime Achievement from the Association for Information Systems (AIS) in December 2002 and was elected a fellow of the AIS in 2000. he was featured in the July 1997 issue of Forbes Magazine on technology as one of eight key innovators in information technology. he is widely published, with an h index of 60. his specialization is in the fields of system analysis and design, collaboration technology, and deception detection. the commercial product groupSystems’ thinktank, based on Dr. Nunamaker’s research, is often referred to

174 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

as the gold standard for structured collaboration systems. he was a research assistant funded by the ISDOS project at the university of Michigan and an associate profes-sor of computer science at Purdue university. he founded the MIS Department at the university of Arizona in 1974 and served as department head for 18 years.

aBstraCt: Credibility assessment is an area in which information systems research can make a major impact. this paper reports on two studies investigating a system solution for automatic, noninvasive detection of rigidity for automated interviewing. kinesic rigidity has long been a phenomenon of interest in the credibility assessment literature, but until now was infeasible as a veracity indicator in practical use cases. An initial study unexpectedly revealed the occurrence of rigidity in a highly controlled concealed information test setting, prompting the design and implementation of an automated rigidity detection system for interviewing. A unique experimental evalu-ation supported the system concept. the results of the second study confirmed the kinesic rigidity found in the first, and provided further theoretical insights explaining the rigidity phenomenon. Although additional research is needed, the evidence from this investigation suggests that credibility assessment can benefit from a rigidity detection system.

kEy Words aNd phrasEs: automated interviewing systems, computer vision, concealed information test, credibility assessment, deception detection, freeze response, kinesic rigidity.

CrEdiBility assEssmENt is a maJor CoNCErN iN maNy orgaNizatioNs and is an area in which information systems (IS) research can have a major impact. kPMg Integrity Surveys report that nearly three-quarters of employees have firsthand knowledge of wrongdoing in their organization and half state that if such wrongdoing were made pub-lic, a significant loss of trust would result [42, 43]. the u.S. government estimates that less than 1 percent of drug trafficking proceeds were detected in a two-year span [74]. In these and many other examples that could be cited, noncredible information has proven difficult to detect, spurring interest among researchers in criminal justice, cognitive psychology, and more recently, IS. unaided human judgment consistently performs near chance levels [9] in spite of chronic overconfidence by decision mak-ers [22, 31], while current veracity system aids are cumbersome, criticized for validity problems, and labor intensive, thus limiting ubiquity.

the most well-known and widely used methods of veracity assessment rely on skilled professionals using interviewing techniques that require time and specialized equipment. Extensive training, invasive sensors, and time limitations are among the factors that have limited the application of these traditional methods mostly to criminal investigations. From an academic standpoint, the validity of current techniques and the reliability of results have been questioned [51, 58]. In particular, lack of procedural control and potential for human error have been cited as potential concerns [58].

IS research can have a major impact in this area by integrating theory, methods, and technology to generate useful and creative solutions [59, 61]. Many efforts have

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 175

already begun. Proposed alternative approaches range from monitoring text-based communication in search of linguistic indicators [46, 90] to identifying telling vocalic or eye movement patterns in human screening [63, 79]. these methods differ from traditional approaches employed to assess veracity in the sensors used, questioning protocols administered and cues identified. the present paper builds on this work by investigating the veracity assessment potential of a body movement cue termed rigidity via the Nunamaker approach [59, 60, 61].

this paper has four objectives. First, it reports on the unique discovery of a rigidity effect in a concealed information test (CIt) protocol via a realistic exploratory mock crime experiment. Second, it proposes a system design for automatic recognition of rigidity in credibility assessment interviews. third, it reports on an implementation of the proposed system design, and evaluation of the instance via a mock screen-ing experiment. Fourth, it summarizes theoretical insights gained throughout this investigation.

reported here are proof-of-concept iterations, beginning by examining (1) results of an experiment that informed (2) a conceptual system design for rigidity detection, of which an instance was (3) built and evaluated to (4) generate valuable knowledge for advancing credibility assessment. the results of this work help establish the proof of concept for a system for automated rigidity detection and also feed back into a larger program of research investigating solutions for enhancing the accuracy, validity, ubiq-uity, and management of automated credibility assessment [23, 63, 64, 78].

Background

CrEdiBility assEssmENt aNd dECEptioN dEtECtioN are gaining increasing interest in IS. Some IS research has focused on how decision makers interact with credibility assessment decision aids [6, 37, 38]. From an e-commerce perspective, IS research has begun to develop a deeper understanding of the key factors involved in e-commerce deception [8]. Automated extraction of linguistic cues to deception has been explored in computer-mediated communication [46, 90, 91], written criminal statements [28], and financial reports [32, 35]. Systems for exploring oculometric indicators of hid-den knowledge have also been an emerging interest [63, 78, 79]. Particularly relevant to the current study, there exists a stream of research investigating the use of certain computer vision techniques to identify movement variables that may have relevance to deception detection [14, 39, 56]. Outside of IS circles, recent research has exam-ined systems for credibility assessment using blinking patterns [27] as well as more invasive systems such as functional magnetic resonance imaging (fMrI) [29, 30, 45] and electroencephalography (EEg) [1].

Some IS credibility assessment research has emphasized the need for noninvasive, autonomic system solutions to increase the ubiquity and reliability of credibility assessment, allowing it to create value in nontraditional contexts such as employment screening, auditing, and physical security screening [23, 64, 78]. this forward-looking approach to credibility assessment inspired the concept of an automated conversa-tional agent that monitors psychophysiological and behavioral indicators relevant to

176 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

credibility assessment [63, 66]. For such an approach to work, many research ques-tions need to be addressed. Among these include the need for identification of valid, reliable cues to deception that can be automatically generated in real or near-real time in a noninvasive manner. kinesic rigidity is one cue that has the potential to meet these criteria.

kinesic rigidity in Credibility Assessment

Rigidity is oNE oF sEvEral kiNEsiC (i.e., body movement) cues that have been identified in communication and psychology research as potential indicators of veracity. kine-sic rigidity is a temporal period of constricted body movement. During high-stakes deception, a liar tends to exhibit fewer noncommunicative movements, such as fewer instances of rubbing hands together or bouncing a leg. Expressive or illustrative ges-tures that do occur tend to be more confined and appear forced, as if they are being resisted [12, 85, 88]. rigidity has been discovered in several studies featuring open-ended questioning protocols [17, 86, 87].

Despite decades of research into bodily rigidity and related kinesic cues to decep-tion, the contextual boundaries of these phenomena are still not well understood, and defining the nuanced interrelationships among nonverbal behaviors and veracity is an active area of investigation in psychology, communication, criminology, and IS research. Several theories have been proposed as explanations for rigidity during periods of low veracity.

Common theoretical explanations include cognitive load [24, 89] and behavioral control [20]; the results of the current study help make a case that hypervigilance may also be a plausible explanation. Proponents of a cognitive load explanation propose that lying takes more cognitive effort than telling the truth, and assumes that fabricat-ing events requires more cognitive resources than simply recalling events. Because more cognitive resources are allocated to creating a plausible deception, it is thought that other activities, including movement, are given less attention, leading to fewer illustrative or communicative gestures [21].

the second common explanation is overt behavioral control. Proponents of this theory emphasize that the general population holds to a false belief that liars show increased nervousness in their body movements. however, while the average person believes a person shows increased body movement when lying, the opposite tends to be the case [75, 76]. According to behavioral control proponents, a deceiver therefore either reflexively or perhaps purposely becomes more rigid in an attempt to mimic his or her own false perception of what a truthful communication should look like [13, 92].

A third possible explanation of rigidity may be more basic, and a precursor to cog-nitive excitation or overt behavior. rigidity may result from the body entering a state of hypervigilance during the biologically driven “stop, look, and listen” response to a perceived threat [10, 33]. When examinees perceive that a line of inquiry has the potential to expose their deception, their body may naturally gravitate toward this

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 177

hypervigilant state, which is characterized partially by bodily rigidity. rigidity in hypervigilance is explained in more depth in the Study 1 Discussion section.

traditionally, rigidity has been measured using human coders, who review video recordings and subjectively rate interview segments according to the appearance of forced versus natural gesturing given the type of gesture and the context in which it was made. human coding is limited to the major movement that can be perceived by a given coder, and it remains subject to intercoder error. Minute changes in movement can be imperceptible to human coders.

Beyond natural human bias and limitation, the biggest restriction to wider adoption of subjective rigidity coding for deception detection is that of the large amount of time and labor costs involved. Every hour of a recorded interview can take two to six hours of expert post-process coding. An automated solution will have the potential to greatly decrease the time and labor cost. Operator training costs can be eliminated altogether if sensors do not require attachment to or manual calibration with each examinee. A noninvasive, automated measurement method is thus a key contribution of this study and an integral component of the proposed rigidity detection system design.

A second important contribution is the exploration of rigidity in CIt interviews, which previously has not been investigated. In addition to automated and noninvasive measurement, an effective system design requires a reliable questioning protocol. Several potential protocols were investigated in the experimental phase of Study 1, and a CIt structure was ultimately selected as the foundational questioning protocol in the system design. the CIt is detailed further in the Study 1 Discussion section.

Study 1: Initial Investigation into Automated rigidity Detection

BasEd oN thE oBsErvatioNs NotEd iN thE BaCkgrouNd sECtioN, general requirements for an automated rigidity detection system were clear from the beginning of the inves-tigation (table 1). the design requirements were necessarily at a high-level stage, given the novelty of the knowledge space. Additional requirements were added as the investigation progressed, and key considerations were revealed through prototyping and experimentation.

Our initial efforts to track and measure movement in credibility assessment inter-views involved using automated techniques for detecting the location of the face and hands in video and tracking two-dimensional changes in location over time. An initial experiment using a mock crime paradigm led to the discovery of rigidity in a CIt paradigm, driving further understanding and development of an automated screening and rigidity detection system design.

tracking Movement in Credibility Assessment Interviews

the context first selected for investigating movement was a standard interview set-ting with the examiner and examinee sitting across from one another, with cameras recording the examinee throughout the interaction. to measure movement, we

178 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

adapted existing computer vision algorithms for recognizing the hands and face in images [14, 56].

For the detection of face and hand/arm locations in video, we applied a skin blob tracking (SBt) technique recently introduced to deception detection research [14, 55, 56, 70]. the SBt technique involves analyzing video frame by frame. For each frame, the face is detected using the Viola–Jones algorithm [84]. Once the face is detected, hand/arm “blobs” are identified by searching for areas of similar (skin) color. the centroid of the face and each hand/arm blob is identified for each frame.

Compared to hand/arm movement, minor changes in head movement ultimately proved more difficult to detect using a full-body frame, standard-definition video. As an alternative method of collecting data for head movement, a close-up video recording of the face was processed using the software suite ASM Face tracker [41]. this software tracks the two-dimensional Cartesian coordinates of many points on a face. the computer vision technique is built on active shape modeling (ASM), which uses spatial-statistical models of shapes to match identified points on an object in one image to points on an object in a new image. the ASM algorithm tries to match the statistical model parameters to the image. thus, the model can deform (e.g., stretch), but not beyond what would be naturally seen in a real-world object of similar features, given properly defined model parameters [19]. For faces, this means that identified facial points must represent the image of a face as a whole. For instance, a point on the chin cannot be accidentally identified as immediately adjacent to a point on the eye as this would be outside the bounds of statistically normal model parameters.

Experiment 1

the SBt and ASM body point location tracking algorithms were used to generate movement data through postprocessing of video recordings of interviews that were part of a realistic mock crime experiment. Mock crime experiments are appropriate for veracity assessment research because the realism involved can elicit reactions that closely mirror real-world scenarios [16]. this mock crime experiment was designed to explore many sensor and questioning technique combinations. the current paper emphasizes that portion of the experiment relevant to automated rigidity detection for credibility assessment.

table 1. Initial Design requirements for Automated rigidity Detection

requirement number Description

1 Automatic tracking of overall movement 2 Noninvasive measurement apparatus 3 Automatic identification of rigidity during deception

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 179

Participants

Participants (n = 164) were recruited from the local community of a large university in the southwestern united States via newspaper and Craigslist (www.craigslist.org) listings and paper flyers placed in community centers. We recruited from the local community in order to obtain a sample of participants more representative than students alone, which we felt was important for this more exploratory phase. the participants received $15 per hour for participation, plus a $50 bonus if they successfully convinced the examiner that they were innocent of the mock crime. Qualitative observations of the participants noted a broad diversity in economic and social status. Of the 164 enrolled participants, 134 (82 percent) followed instructions and completed the experiment. the remaining 18 percent were disqualified because they did not follow instructions, failed to consent to participate, or confessed during the interview. Because of technical problems with the video recording and analysis system, only 107 of the initial 134 cases produced usable data for analysis. Of these 107 participants, 40 participants “committed” the crime, leaving 67 who did not. In this subset, 63 percent were female, and the average age was 39.5 (standard deviation = 14.0).

Experimental Procedures

Participants in a simple two-treatment mock crime experiment were instructed to arrive at a room in an upper floor of an old apartment complex. A prerecorded set of instruc-tions was waiting for them. After listening to the instructions and signing a consent form, the participants left the apartment complex and walked to a nearby building.

Per the instructions, the participants reported to a room on the top floor and asked for a Mr. Carlson. A confederate acting as a new receptionist who did not know Mr. Carl-son asked the participant to wait while he went to locate Mr. Carlson. A camera in the room verified the participants’ activities while they were waiting for the receptionist’s return. Participants in the Innocent condition simply waited, while those in the Guilty condition stole a diamond ring from the desk. Guilty participants took a key from a mug on the top of a desk and used it to open a blue cash box in the desk drawer that was hidden underneath a tissue box. they removed the ring from the cash box and hid it somewhere on their person.

upon returning, the receptionist directed the participants to another room on the bottom floor of the building, the layout of which is depicted in Figure 1. there, the participants were told that a crime had occurred in the building that day and that they would be interviewed to assess their possible involvement in the crime. All the par-ticipants were interviewed by one of four professional polygraph examiners provided by the National Center for Credibility Assessment (NCCA). the interviewers were trained and experienced, and were familiar with the purpose and procedure involved in administering various interviewing techniques, including CIt, a veracity assessment technique highly regarded in academic circles [5, 58, 80] but rarely used in practice, Japanese criminal investigations being the notable exception [57, 65]. the participants were offered a $50 bonus if they successfully convinced the interviewer that they were

180 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

innocent. this large monetary reward together with the realism of the experiment was important to induce behavioral effects and motivate participants to appear innocent in ways that would closely mirror real-world scenarios.

two studio-quality video cameras were placed directly in front of the chair in which each participant sat during the interview. the chair had a low back and did not have armrests. No other furniture or objects were within reach. this setup ensured that inactive arms and hands would rest on legs during the CIt portion of the interview. Other cameras and sensors were also present in the room, to examine their potential for credibility assessment (to be reported elsewhere). the location of each hand/arm and the head were identified frame by frame using the SBt and ASM computer vision techniques.

the interview consisted of several questioning techniques, including a CIt. the CIt was a major portion of the interview and became the focal procedural component of the rigidity detection system, as explained in later sections. to our knowledge, rigidity has never been investigated in a CIt format prior to this study, and exploring rigid-ity in the CIt was not initially a primary consideration. rather, we sought to detect rigidity in alternative questioning techniques similar to previous work. however, the control and simplicity of the CIt, together with its potential for an automated system

Figure 1. layout of Interviewing room for Experiment 1

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 181

prompted an exploratory rigidity analysis. the three CIt questions together with their associated target and nontarget items are included in table 2.

Video from the two cameras recorded during the interview were processed to gen-erate overall movement data, as explained in the next section. A final questionnaire followed the interview portion of the experiment, and contained simple manipulation check questions, together with a question about perceived behavioral control and measures of arousal and motivation levels.

Measuring rigidity

this study took a novel approach to measuring movement, designed to circumvent the need for post hoc, manual subjective judgments. For the mock crime experiment, the centroid coordinates of each SBt-generated blob and the center coordinates of the ASM face model were generated for each frame between the end of an interviewer question and the beginning of the next question. Once data for each frame were gener-ated, overall movement for the left and right hands/arms for each video segment was calculated by determining the average Euclidean distance between centroid position changes frame by frame during a given response in the following manner:

M y y x x js j

i= −( ) + −( )

∑ 2 1

2

2 1

2/ .

table 2. Questions used in the Study 1 Concealed Information test

QuestionWords repeated

by suspect

If you are the person who stole the ring, you are familiar with details of the cash box it was stored in. Repeat after me these cash box colors:

GreenBeigeWhiteBlue*BlackRed

If you are the person who stole the ring, you moved an object in the desk drawer to locate the cash box containing the ring. Repeat after me these objects:

NotepadTelephone bookWoman’s sweater Laptop bagTissue box*Brown purse

If you are the person who stole the ring, you know what type of ring it was. Repeat after me these types of rings:

Emerald ringTurquoise ringAmethyst ringDiamond ring*Ruby ringGold ring

* target items (i.e., correct answers).

182 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

this produced an average overall movement score for each response for each indi-vidual. however, average overall movement during a response is certain to be affected by more than just veracity level. Culture, personality, mood, gender, and question type are example factors that may also affect overall movement or lack thereof. For instance, qualitative observations of the participants revealed that, on average, those from Western cultures tended to exhibit more movement overall when “sitting still” than those from Eastern cultures. Identifying, automatically measuring, and integrat-ing all such potential global moderating factors that influence movement is a difficult and complex task, and well beyond the scope of this study.

however, a repeated-measures interviewing protocol provides the possibility of an alternative approach. Individuals can be compared to an individual baseline rather than an overall population average [2, 82], sidestepping the need to account for factors such as gender, culture, or mood. the movement averages for each segment were thus standardized as within-subject z-scores. the z-scores were also body point specific, because natural variance is expected in the amount of movement that each point on the body will exhibit (e.g., a little movement of the head can be just as meaningful as a relatively large movement of a hand). In the case of the CIt questions, z-scores were also question specific to control for the possibility of question effects.

results

As part of the postinterview questionnaire, the respondents self-reported their levels of motivation, effort, and tension, each on a seven-point scale (see the Appendix). Participants reported high levels of motivation and effort, and moderate levels of ten-sion. Summary statistics are in table 3.

Within-subject comparisons of interquestion overall movement did not produce significant results for any tested interviewing protocols except the CIt. the rigidity results of tests other than the CIt are omitted for succinctness.

For the CIt questions, a multilevel regression model was specified for overall move-ment during the response time for each foil item. Multilevel regression models use adjusted standard errors to reflect the uncertainty that arises from variation within a subject. the summation of standardized movement scores for right hand, left hand, and head was used as the dependent variable. the independent variables included Condition (dummy coded: 1 = guilty, 0 = Innocent), Participant, and target Item (dummy coded: 1 = Correct Answer, 0 = Incorrect Answer). Question and interviewer were initially included as covariates but were found to not be significant predictors and were subsequently dropped from the model. the effect of greatest interest was the Condition and target Item interaction, which reflected overall movement when guilty participants responded to the correct answer. the results of the multilevel regression model are shown in table 4, with Condition labeled “guilty” to help facilitate interpretation.

the significant estimate of –0.624 in the guilty × target Item interaction can be interpreted to mean that when guilty participants were asked to repeat the correct answer to a CIt question, they tended to be approximately 0.624 standard deviations

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 183

below their own personal average. to test if the guilty and target Item interaction provided a significant improvement to the fit of the data, the model was compared to an unconditional model, omitting any fixed effects and using deviance-based hypoth-esis tests. the fit of the current model was significantly better than the unconditional model, χ2(1, N = 1,887) = 17.15, p < 0.001.

Study 1 Discussion

using the SBt and ASM tracking techniques and the automated movement measure-ment method, rigidity was successfully detected automatically in only one interview-ing protocol—the CIt. there were several important lessons learned from this initial investigation, including the discovery of rigidity in the CIt and important observations and refinements necessary to advance the initial concept of a rigidity detection design closer toward a successful proof of concept.

the rigidity effect was consistently significant for three successive CIt questions. rigidity has previously been identified in open-ended interviewing where various types of movement are common, each measured in frequency or duration [73, 85, 88], but the CIt offers little opportunity for communicative movement during short answers. No story fabrication occurs in a CIt; any movement that does occur in a CIt is either natural movement that occurs even during a state of stillness, or self-adaptors or similar movements that are non-communicative in nature. thus, in a CIt context, the previ-ously referenced cognitive load theory is not plausible. Behavioral control remains a

table 3. Self-reported Motivation, Effort, and tension

Self-report (seven-point scale) Condition Mean

Standard deviation

Motivation to succeed InnocentGuilty

6.126.15

1.311.29

Effort InnocentGuilty

5.75 6.13

1.541.24

Tension InnocentGuilty

3.00 3.21

1.561.75

Note: No comparisons were significantly different between groups.

table 4. Overall Movement: Multilevel regression Model results

Fixed effects β β standard error

Intercept 0.044n.s. 0.069Guilty 0.102n.s. 0.109Target Item –0.197n.s. 0.171Guilty × Target Item –0.624* 0.267

Notes: N = 1,887. Model fit using maximum likelihood. * p < 0.05; n.s. = not significant.

184 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

possibility, and participants did self-report high motivation and effort. however, the rigidity in head movement even though no nodding occurred suggests that the rigid-ity detected in the CIt might have been at a more granular level than what has been traditionally attributed to behavioral control. these unanticipated findings prompted a more in-depth investigation into the driving factors behind rigidity and the CIt, in order to better understand how a rigidity detection system could best be designed to take advantage of this unique discovery.

rigidity in the Concealed Information test

the CIt is a questioning method that seeks to minimize potential interviewer effects while generating a strong individual baseline for analysis. the CIt is similar to a multiple choice exam: An interviewer asks multiple-choice questions specific to an illicit act, with each question followed by a series of possible answers, each stated verbally by the interviewer [50]. When each possible answer is stated, the examinee is required to either repeat each possible answer or respond with “yes” or “no” [52, 83]. Physiological measurements such as skin electrical conductivity are tradition-ally measured throughout this exercise, and analyzed for abnormalities during target items (correct answers) as compared to nontarget items (incorrect answers) [48]. the CIt has been praised as the most valid of current credibility assessment interviewing techniques, and calls have been made for its more widespread use [4, 36], perhaps even to applications other than criminal interviewing. Some evidence even suggests that the CIt format can be used to measure and predict behavioral intent [54].

traditional measures in the CIt gauge levels of the psychophysiological orienting response, the sympathetic nervous system activation triggered by novel or personally significant stimuli [48, 51, 72]. A commonly cited example of the orienting response depicts a loud cocktail party where an individual is oblivious to peripheral conversa-tions [18]. yet the same individual will naturally orient attention toward a peripheral conversation when his or her name is spoken, because of its personal significance. the physiological effects of this natural, autonomic orienting range from variations in pupil dilation and respiration to skin conductivity and heart rate, with skin conductivity being the most commonly measured effect in the CIt [1, 4, 25].

there is evidence to suggest that CIt rigidity may also stem from the orienting reflex, if the stimulus is associated with a potential threat. Perceived threats cause an individual to enter a “stop, look, and listen” hypervigilant state characterized by physiological and behavioral modifications designed to better recognize and respond to perceived threats. Although the term hypervigilance may seem to imply increased movement, when a non-immediate threat is involved the opposite tends to be the case. In such cases an individual exhibits less overall movement, at the same time experiencing heightened sensitivity to cues that may require a defensive reaction, such as fight or flight [15, 40]. this “freeze response” has often been grouped with what is termed a defensive response rather than with the orienting reflex. however, the freeze response is associated with bradycardia (i.e., decreased heart rate) [7, 69], a distinguishing characteristic of the orienting reflex [81, 83]. this apparent contradiction may be resolved by the defense cascade model, which describes the defensive response as a

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 185

temporal sequence of events, of which one event is an orienting response. the defense cascade model conceptualizes the freeze response as part of an orienting reflex [11, 44], “possibly serving to facilitate detection of information relevant for a subsequent fight-or-flight response involving whole-body movements” [68, p. 1580]. the freeze response occurs as part of the orienting reflex when a perceived threat is not immedi-ate, such as when an animal senses a predator from afar [26, 68].

recent evidence indicates that the freeze reflex is also evident when the nature of the perceived threat is social, rather than physical [68]. to the extent, then, that a CIt stimulus is perceived as having potential to expose a deception and subject an individual to consequences, the natural reaction should include a freeze response, or whole-body rigidity. In such a case, the threat is not imminent, but can have serious social repercussions, and we would therefore expect the orienting reflex to include a freeze response.

Design Performance and revised requirements

Several key observations and results from Study 1 generated the insight necessary for a revised and more specific design for a rigidity detection system for credibility assessment. these insights affected the questioning protocol as well as the interaction design and measurement technology.

Effective credibility assessment systems will necessarily synthesize questioning protocol with technology and interaction design requirements, because the manner in which questions are asked can be just as important a factor as the deception cues measured [34, 47]. that rigidity was discovered only in the CIt indicated that a reliable system questioning protocol and interview design will need to minimize or control for many potentially confounding factors. In a credibility assessment interview, the amount of movement appeared to be strongly affected by many factors beyond veracity level, including culture, interview style, and context. this likely caused difficulty automati-cally detecting rigidity in open-ended interviewing protocols, counter to prior research. For instance, one individual response to a question may involve minor communicative gestures depicting “small,” while a similar gesture later in the discussion may be more expansive when depicting the concept “large.” Also, individuals from certain cultural backgrounds were less inclined to gesture compared to others. the interviewer’s style and demeanor influenced the response style and demeanor of the examinee.

the effects of these cultural and contextual factors were likely large enough to significantly impact overall movement in non-CIt question responses. traditional manual coding can subjectively account for context and movement type, whereas the proposed automated method treated all movement equally. the analysis revealed that traditional human coding may be more useful for detecting rigidity in certain types of movement, such as a behavioral control–induced decrease in hand and finger move-ments, while the automated approach used in this study may work best for detecting rigidity resulting from the freeze response, which affects the whole body (table 5).

In addition to discovering the advantage of a CIt-like questioning protocol, the results of Study 1 have interaction design and measurement technology implications. If the rigidity effect in a CIt stems more from an overall, whole-body freeze response,

186 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

a better system design may be one that captures movement from more body points, rather than focusing mostly on hand and head movement as in Study 1. Also, a stand-ing interview may be more effective for capturing body movement, as opposed to a sitting interview where movement at a resting state is inhibited by a chair. SBt and ASM tracking provided usable data for every participant for whom video was properly recorded, but movement in the third dimension (i.e., depth) was not possible to capture using standard two-dimensional video. the SBt and ASM techniques required time-consuming post-processing to complete, which in practice would limit usefulness to contexts that require less rapid decision making, such as criminal investigations, or, in order to be useful in rapid contexts, would require improvement in hardware or software to generate near real-time data for analysis. Also, the beginning and end points for each response had to be manually marked, as the SBt and ASM procedures by themselves provide no video segmentation or tagging capability.

In sum, the observations and findings from Study 1 helped to better specify requirements for a system for automated rigidity detection for credibility assessment (table 6).

Study 2: Development of a rigidity Detection System for Credibility Assessment

BuildiNg oN thE rEsults oF study 1, we developed an autonomous screening system for automated rigidity detection that would operationalize the design requirements. the underlying software for automated interviewing was adapted from earlier efforts at automated interviewing systems designed for a kiosk-like interaction [63, 77]. these

table 5. Comparison of rigidity Measurement Methods

traditional manual movement coding

Proposed automated method

Rigidity measurement method

Human judgments of frequencies and durations of predefined movements

Automatically captured movement measured by detecting body point position changes

Granularity of measurement

Major gestures to minor finger movements

Pixel-level changes (assuming computer vision approach)

Time requirement 2 to 3 man-hours per hour of recorded interview (estimated)

Near real-time

Categorization of movement

Can reliably classify movement (e.g., self-adaptor, illustrator gesture)

All movement for a given body point treated equally; no differentiation of movement types

What can be detected in the CIT

Frequencies and durations of non-functional movements (e.g., fidgeting)

Overall body movement rates

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 187

automated screening technologies use an automated agent to ask questions and record responses. the same virtual interview can be used for every examinee, controlling for interviewer effects. the screening systems also enable automatic segmentation of interviews into relevant portions. the current project built on and extended these system designs by implementing the CIt-based questioning protocol and adding a method for movement detection.

Interviewing Protocol

Because rigidity in a CIt was a new discovery, it was important to determine whether rigidity is a robust effect and whether it translates to alternative contexts. From a practi-cal standpoint, it was important to design and evaluate a system capable of delivering an automated CIt and detecting rigidity with minimal user input. Instantiating the findings of Study 1, the CIt served as the foundational questioning protocol for the rigidity system design. the multiple-choice question format provided a means for generating a strong individual- and question-specific baseline, by comparing behavior displayed during the response to the correct answers to behavior during incorrect answers. this questioning format makes it possible to control for individual and question effects.

rigidity Detection and Measurement

We adapted a process for detecting rigidity using an automated screening approach. to capture data for movement analysis, we used a Microsoft® kinect® sensor rather than standard cameras. the kinect uses stereoscopic imaging to identify the

table 6. Design requirements for an Automated rigidity Detection System

requirement number Description

1 Conduct an interview which includes potentially threatening stimuli (for guilty individuals) mixed with non-threatening stimuli in a multiple choice–like format

2 Ensure the interviewing protocol that minimizes potentially confounding communicative movement

3 Conduct the interview automatically to control for or minimize the effects of a human interviewer

4 Preferably use a standing interview5 Employ non-invasive, near–real time or better raw movement data

capture, such as computer vision algorithms6 Preferably track movement in three dimensions7 Automatically segment interviews into relevant time periods (e.g., tag

start and end points for responses)8 Leverage the questioning protocol to generate a strong individual

baseline for comparison9 Compare overall movement during key questions to that during

irrelevant questions

188 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

three-dimensional (3D) location of 20 major body points in real time at roughly 30 samples per second in real time. the kinect was chosen over SBt and ASM because of its ability to capture the position of more body points (potentially important for a standing interview), track in three dimensions, and minimize the delay in data generation.

the kinect sensor was placed approximately 6 feet away from a platform where an interviewee was told to stand during the interview. this distance was just enough to allow the sensor to capture a full body view. A computer screen serving as the interface for the automated interview was mounted on a reticulating arm rather than a desk or table so that the sensor’s line of sight would be unobstructed. Custom lighting was kept constant and designed to eliminate shadows similar to the first iteration. Figure 2 depicts this configuration.

Custom algorithms were scripted to capture the 3D Cartesian coordinates (in meters) of each of the 20 points in real time and segment and analyze the data. Data from an entire interview were segmented into relevant portions, then summarized and standard-ized in the same manner as were the SBt and ASM data in Study 1, the only difference being that there were 20 body points to standardize rather than 3.

Experiment 2

A laboratory experiment was designed to evaluate the system. rather than a mock crime scenario, a building security screening context was chosen. this choice stemmed

Figure 2. Physical System Configuration used in Experimental Evaluation

Note: the kinect sensor (circled) was place approximately 6 feet from the platform where interviewees stood.

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 189

from a desire to identify potential contextual nuances for the rigidity effect for cred-ibility assessment. Important to this evaluation from a design science perspective was identification of the extent to which the system fulfilled design requirements as well as the most promising areas for improvement. From a theoretical perspective, it was important to confirm the rigidity effect discovered in the first experiment, and perhaps gain further insight into its origins.

Participants

Students from a large southwestern university participated in the experiment as part of a course requirement. While the ideal population to test would be individuals who intend to commit an actual crime, such a population would be prohibitively difficult to find. Students are an appropriate alternative because the main drivers behind the standard CIt have been found to be equally valid among adult populations and anti-social or hyporesponsive criminals [82], and kinesic rigidity itself has been observed in criminal behavior during conditions of low veracity [86] as well as students [88]. thus, while students are not the optimal test population, the observations seen in this population likely have a useful level of external validity.

the participants were offered $20 if the system judged them as truthful. this was a mild deception designed to promote engagement and jeopardy; in reality, all the par-ticipants received $20. In both experiments, human subjects protocols were followed, and approval was obtained from the university’s human subject institutional review board. Of the initial 175 participants, 18 were disqualified because they either did not follow instructions (e.g., failed to construct/pack the device) or failed both manipulation check questions.1 Of the remaining 157 participants, 43 percent were female.

Experimental Procedures

As in experiment 1, there were two conditions in this study: Guilty and Innocent. the protocol mirrored the experiment detailed in twyman et al. [77], except in this case, the Guilty condition was broken into four groups, with three of the groups trying dif-ferent countermeasures (i.e., methods for defeating a system). the effectiveness of the countermeasures was an important consideration, but is beyond the scope of this paper and will be reported elsewhere.

Participants packed a bag with standard travel items (e.g., clothes, books) with the intention to deliver the bag to a contact ostensibly in a secure portion of the build-ing. the guilty participants in this experiment also constructed a mock improvised explosive device (IED) and packed it with the innocuous items. the participants were told that to get to the secure portion of the building, they would have to pass through a security screening checkpoint. A separate room was set aside for the screening (Figure 3) to ensure no distractions would be present. Participants interacted with the second-generation ASk system as a prerequisite for entering the secure portion of the building where they were to deliver the bag. All of the participants had been introduced to the ASk system during the experiment sign-up, and on arrival, they were offered $20

190 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

if they were able to successfully pass through the screening without being flagged by the system. the task to deliver the bag was a mild deception to encourage realism; in reality, no deliveries took place. After the automated screening, the participants filled out a questionnaire and were debriefed to ensure they understood that their task was complete in spite of not having delivered the bag, and that the experiment was only a simulation and they had not done anything illegal.

the interview phase proceeded in the following manner: A participant stepped up to the ASk system, and the automated system provided audible initial instructions and guided the individual through a 15-second calibration process. After this, a realistic, on-screen animated avatar face asked the question, “Are you carrying any of these items?” A screen appeared immediately after each question was asked, displaying four images representing banned items for five seconds. Figure 4 shows an example of a screen displaying the images. this question was asked five times. An image represent-ing an IED appeared only once during this period. the sequence of five questions was asked three more times, rotating the location of the images and changing the temporal point at which the image representing an IED was displayed. the entire 20-question process took about four minutes, after which the participants proceeded through the exit, filled out a questionnaire, and were debriefed.

Figure 3. Screening room layout for Experiment 2

Note: Participants brought their packed bag into the room and stood on the platform for the brief screening interview.

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 191

results

two manipulation check questions ensured the participants had understood and fol-lowed the instructions. Additional manipulation check questions in the postexperiment questionnaire captured self-reported motivation, effort, and tension (see the Appendix). As in Study 1, the participants reported high levels of motivation and effort as well as moderate levels of tension (table 7).

As in Study 1, summations of the major body points’ standardized movement scores served as a measure of overall movement in a target segment (when an IED appeared on the screen) relative to control segments (where no IED was present). A multilevel regression model using these standardized movement scores as the dependent variable was specified in a manner similar to that reported in the first experiment. Age, gender, education level, and English fluency were initially included but, as expected, because of the standardization procedure, produced no significant results and did not add to the fit of the model. Although the Guilty condition contained four distinct groups, the overall trend of rigidity was similar for all groups. Because the nuances between the Guilty groups will be investigated in detail in a future study, and for the sake of simplicity, we collapsed all Guilty conditions into a single group for this analysis. table 8 summarizes the overall movement results.

the significant estimate of –0.109 in the guilty × target Item interaction can be interpreted to mean that when Guilty participants viewed the image of an explosive on the screen, they tended to be approximately 0.109 standard deviations below their own personal average. to test if the effects seen provide a significant improvement to the fit of the data, the model was compared to an unconditional model, which omits any fixed effects, using deviance-based hypothesis tests. the fit of the current model

Figure 4. Sample Stimuli Screen

Note: twenty of these stimuli were displayed during each interview.

192 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

was significantly better than the unconditional model, χ2(1, N = 62,780) = 181.26, p < 0.001.

A follow-up analysis was undertaken to examine stillness on a more granular level. Overall movement was broken down by major body points and similar variables were identified using principal components analysis (PCA). using the traditional cutoff point of eigenvalues > 1, five factors were generated. table 9 shows the PCA factor loadings; table 10 shows the variance explained by each factor.

the first factor and most of the explained movement came from the core part of the body, with the remaining four components each centered on a hand or foot. the results of separate multilevel regression models for each component were generated to help locate the area(s) of greatest rigidity during CIt target items (table 11).

the significant estimate of –0.250 in the guilty × target Item interaction can be interpreted to mean that when Guilty participants viewed the image of an explosive on the screen, their movement tended to be approximately 0.250 standard deviations below their own personal average. this represents a significant drop in overall movement for the core body, which is ostensibly at a state of rest throughout the entire interview.

Discussion

iN praCtiCE, rigidity is Not a CommoNly CoNsidErEd variaBlE for credibility assess-ment. to the extent rigidity is considered, it is only in an ad hoc, subjective matter. Such usage is unreliable and prone to error. Person-to-person credibility assessment

table 7. Self-reported Motivation, Effort, and tension

Self-report (five-point scale) Condition Mean

Standard deviation

Motivation to succeed InnocentGuilty

4.294.33

0.760.74

Effort InnocentGuilty

3.173.52

1.141.15

Tension InnocentGuilty

2.713.15

1.141.14

Note: No comparisons were significantly different between groups.

table 8. Overall Movement: Multilevel regression Model results

Fixed effects β β standard error

Intercept 0.013n.s. 0.014Target item –0.024n.s. 0.019Guilty 0.022* 0.010Guilty × Target item –0.109* 0.028

Notes: N = 62,780. Model fit using maximum likelihood. * p < 0.05; n.s. = not significant.

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 193

in interviews, negotiations, and other meetings could benefit from a credibility assess-ment IS that includes a rigidity detection system.

given the potential impact, the purpose of this project was to use prototyping and experimentation to investigate the potential of automated rigidity detection for cred-ibility assessment and advance a system design toward a proof of concept. these goals were realized in Studies 1 and 2, as new understanding emerged with regard to rigidity itself in a credibility assessment context as well as to a requirement set that could be feasibly instantiated in a system solution.

the results of Study 1 unexpectedly revealed that rigidity was present in the highly controlled CIt, which suggested the possibility of exploiting this finding in an auto-mated interviewing solution. the results from Study 2 verified the initial discovery of CIt rigidity, and provided evidence that the automated screening system design

table 9. PCA Factor loadings for Overall Movement

Body point C1 C2 C3 C4 C5

Head 0.56Hip center 0.93Left hip 0.88Right hip 0.88Shoulder center 0.71Left shoulder 0.71Right shoulder 0.65Spine 0.94Right hand 0.88Right wrist 0.91Left hand 0.87Left wrist 0.89Left ankle 0.87Left foot 0.86Right ankle 0.85Right foot 0.84Left elbow 0.42 0.49Right elbow 0.35 0.49Left knee 0.42Right knee 0.34

Notes: Movement values are standardized as noted in the Measuring rigidity section. Varimax rotation factor loadings less than 0.3 are omitted.

table 10. PCA Factors Variance Explained

C1 C2 C3 C4 C5

Proportion of variance explained 0.28 0.10 0.09 0.08 0.07Cumulative variance explained 0.28 0.38 0.48 0.56 0.63

194 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

table 11. results by Principal Component

Fixed effects

Core body

(C1) β

right hand/arm

(C2) β

left hand/arm

(C3) βleft foot (C4) β

right foot

(C5) β

Intercept 0.013 0.009 –0.005 –0.002 –0.004Target Item –0.063 –0.046 0.026 –0.008 0.018Guilty 0.050 –0.002 0.010 0.001 0.014Guilty × Target Item –0.250* 0.012 –0.048 –0.003 –0.070

Notes: N = 3,139. Model fit using maximum likelihood. * p < 0.05.

can work. the investigation also provides theoretical insight with regard to the rigid-ity phenomenon. Previous work investigating rigidity has focused on conversational interactions where movement is a natural, integral part of the interaction. rigidity has been thought to stem from an increase in cognitive load or misapplied behavioral control. In the CIt, there is no message production or similar demand for increased cognitive activity when lying. It is difficult to argue, then, that the nervous system lacks sufficient resources to attend to natural movement. Behavioral control remains a plausible explanation for rigidity in a CIt context; however, if most rigidity is centered in the core body, as was seen in Study 2, the psychophysiological freeze response may be a more likely explanation. It is important to note that these insights are limited to a highly controlled format such as the CIt, and thus do not necessarily contradict prior rigidity research. rigidity in alternative, more communicative, and open-ended interviewing techniques may have different underlying drivers.

It is also important to note also the differences in measurement. Prior research has focused on manually measuring frequencies of distinct, visible movements, whereas this study collapsed all movement within a given segment. Manual coding differen-tiates between movement types and can consider semantic meaning of movement types. this may be one reason why rigidity was not found in alternative questioning protocols in Study 1, where open-ended questions encourage semantically driven movements. While it lacks the ability to recognize intended semantic meaning, the movement-tracking design in this study has advantages of automation and ability to capture fine-grained movement, even at a natural resting state.

the second study replicated the first, but with several variations, including a system-driven rather than a human-driven interview, movement detection using an alternative approach, and a screening paradigm rather than a post-crime investigative interview. the rigidity effect appeared to be robust to these differences.

Contributions

Nunamaker and Briggs [59] have called for an expanded vision of IS research that includes inventing new systems that address information needs. the overall contribu-tion of this investigation is a design for a proof-of-concept prototype for automated

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 195

rigidity detection for credibility assessment—a novel and previously unexplored con-cept. Although the domain and design are somewhat novel to IS research, the contribu-tion lies at the heart of IS: the greatest value of this prototype design is not the process or the technology chosen, but in the synthesis of technology and process [71].

this investigation generated knowledge through exploration, prototyping, theory development, and experimentation. Whereas initially the requirements for a feasible system for rigidity detection for credibility assessment were an unknown or lacked evidence, this investigation specified a process and technology combination that can achieve this goal. In so doing, new insights into rigidity were revealed. Prior to this investigation, rigidity had been observed in open-ended responses. rigidity was discovered in a CIt setting, a scenario not previously investigated, likely because of the expectation of little movement in a CIt. however, it was this very controlled and simple technique that suggested not all rigidity is due to cognitive overload, and may derive from the natural hypervigilance state—more directly hardwired into human genetic makeup than was previously thought.

this new discovery for credibility assessment systems and CIt research has potential for high impact not only to IS researchers developing credibility assessment systems but also to criminal justice, communication, and psychology research on CIt and kinesic rigidity. research in reference disciplines bemoans the fact that unaided human credibility assessment is consistently poor [9], uncorrelated with decision makers’ confidence [22], and the most common methods employed to help are scientifically unsound [3, 58]. the CIt was initially proposed in part to add validity to credibility assessment interviewing. this systems approach to conducting a CIt furthers that ideal via greater control of interviewer effects and noninvasive sensors, while at the same time increasing the potential for ubiquitous application. Minimizing the human skill and specialized technology requirements by developing noninvasive, automated system solutions such as these may facilitate the adoption of CIt examinations on a much broader scale, whereas current usage is limited to a very small percentage of criminal investigations, and very few other applications. As this line of research advances, rigidity detection may also be useful as an additional indicator for existing and alternative use cases, such as security screening, employment interviews, pre-audit-screening interviews, insider threat detection, or group collaboration sessions.

limitations and Future Directions

the instantiation of the design requirements performed as anticipated for all of the interviews. thus, the results of Study 2 provide support for a proof of concept. Never-theless, several observations revealed areas for improvement on the conceptual design. A potential drawback of the kinect sensor was its low resolution, which increases the probability that minor differences in movement patterns will go undetected, and may therefore have affected the movement detection ability of the overall system. An evaluation of the SBt and ASM techniques in Study 1 revealed that rigidity in head movement was indiscernible in a full-body standard-definition video frame, but easily detected in a facial close-up frame. this indicates that points on the body that

196 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

naturally have little movement may need higher resolution to discover rigidity effects stemming from a freeze response.

the findings in this investigation serve as an important piece of a large research program focused on systems solutions for modern credibility assessment needs. the research program mirrors the university of Arizona group support systems research program in that high-impact knowledge is discovered and integrated over an extended period of time using many system iterations and explorations [62]. Many additional important research topics and critical research questions will be investigated through the exploration, design, and evaluation of automated credibility assessment systems. Some immediate examples include prediction, sensor and indicator fusion, and countermeasures.

While this study focused on feasibility and proof of concept, future research will need to investigate the predictive capability of the rigidity phenomenon alone and in combination with other CIt deception cues usable in automated credibility assessment. there are many prediction algorithms and scoring methods investigated in the extant CIt research [49, 53], and an investigation into the usefulness of each of these and novel approaches will be important for establishing proof of value. Investigating the most appropriate prediction techniques and identifying key system design changes they will warrant will be an important topic for follow-up work.

In many contexts it is unlikely that a single indicator or sensor will be adequately predictive of veracity or credibility risk [63, 67]. Future systems for border check-points, airport security terminals, employment screenings, internal investigations, or financial audits will likely require many predictive indicators to better triangulate and make counter measures more difficult. thus, future work will seek to integrate the contributions of this paper into systems that will impact both practical outcomes and understanding of the interrelationships of cognitive and behavioral correlates of veracity.

Conclusion

this proJECt has takEN importaNt stEps toWard dEFiNiNg a feasible system design for automated rigidity detection for credibility assessment. the initial study helped identify important technological and process constraints for the system, at the same time generating insight into the rigidity phenomenon. Future work will emphasize prediction and integration of additional noninvasive sensors and automated solutions in an effort to improve on the conceptual design and move it closer to proof of value. As noninvasive, automated credibility assessment systems such as these advance and integrate, the resulting increased understanding and solutions will have potential to revolutionize approaches to managing security and integrity.

Acknowledgments: the Department of homeland Security’s (DhS) National Center for Bor-der Security and Immigration (BOrDErS), the National Center for Credibility Assessment (NCCA), and the Center for Identification technology research (CIter), a National Science Foundation (NSF) Industry/university Cooperative research Center (I/uCrC), provided fund-

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 197

ing for this research. Statements provided herein do not necessarily represent the opinions of the funding organizations.

NotE

1. Manipulation check questions involved asking the participants to state whether they had carried anything illicit in their bag and to identify the person to whom they were supposed to deliver the bag.

rEFErENCEs

1. Ambach, W.; Bursch, S.; Stark, r.; and Vaitl, D. A concealed information test with mul-timodal measurement. International Journal of Psychophysiology, 75, 3 (2010), 258–267.

2. Ben-Shakhar, g. Standardization within individuals: A simple method to neutralize indi-vidual differences in skin conductance. Psychophysiology, 22, 3 (1985), 292–299.

3. Ben-Shakhar, g. A critical review of the control question test (CQt). In M. kleiner (ed.), Handbook of Polygraph Testing. San Diego: Academic Press, 2002, pp. 103–126.

4. Ben-Shakhar, g., and Elaad, E. the validity of psychophysiological detection of informa-tion with the guilty knowledge test: A meta-analytic review. Journal of Applied Psychology, 88, 1 (2003), 131–151.

5. Ben-Shakhar, g.; Bar-hillel, M.; and kremnitzer, M. trial by polygraph: reconsidering the use of the guilty knowledge technique in court. Law and Human Behavior, 26, 5 (2002), 527–541.

6. Biros, D.P.; george, J.F.; and Zmud, r.W. Inducing sensitivity to deception in order to improve decision making performance: A field study. MIS Quarterly, 26, 2 (2002), 119–144.

7. Blanchard, r.J.; Flannelly, k.J.; and Blanchard, D.C. Defensive behaviors of laboratory and wild Rattus norvegicus. Journal of Comparative Psychology, 100, 2 (1986), 101–107.

8. Bo, X., and Benbasat, I. Product-related deception in e-commerce: A theoretical perspec-tive. MIS Quarterly, 35, 1 (2011), 169–95.

9. Bond, C.F., and DePaulo, B.M. Accuracy of deception judgments. Personality and Social Psychology Review, 10, 3 (2006), 214–234.

10. Bracha, h.S.; ralston, t.C.; Matsukawa, J.M.; Williams, A.E.; and Bracha, A.S. Does “fight or flight” need updating? Psychosomatics, 45, 5 (2004), 448–449.

11. Bradley, M.M.; Codispoti, M.; Cuthbert, B.N.; and lang, P.J. Emotion and motivation I: Defensive and appetitive reactions in picture processing. Emotion, 1, 3 (2001), 276–298.

12. Buller, D.B., and Aune, r.k. Nonverbal cues to deception among intimates, friends, and strangers. Journal of Nonverbal Behavior, 11, 4 (1987), 269–290.

13. Buller, D.B., and Burgoon, J.k. Interpersonal deception theory. Communication Theory, 6, 3 (1996), 203–242.

14. Burgoon, J.k.; Jensen, M.; twyman, N.W.; Meservy, t.O.; Metaxas, D.N.; Michael, N.; Elder, k.; and Nunamaker, J.F., Jr. Automated kinesic analysis for deception detection. In Pro-ceedings of the HICSS-43 Credibility Assessment and Information Quality in Government and Business Symposium. koloa, hI: IEEE Computer Society, 2010, pp. 31–40.

15. Campbell, B.A.; Wood, g.; and McBride, t. Origins of orienting and defensive responses: An evolutionary perspective. In P.J. lang, r.F. Simons, and M. Balaban (eds.), Attention and Orienting: Sensory and Motivational Processes. Mahwah, NJ: lawrence Erlbaum, 1997, pp. 137–164.

16. Carmel, D.; Dayan, E.; Naveh, A.; raveh, O.; and Ben-Shakhar, g. Estimating the validity of the guilty knowledge test from simulated experiments: the external validity of mock crime studies. Journal of Experimental Psychology: Applied, 9, 4 (2003), 261–269.

17. Caso, l.; Maricchiolo, F.; Bonaiuto, M.; Vrij, A.; and Mann, S. the impact of deception and suspicion on different hand movements. Journal of Nonverbal Behavior, 30, 1 (2006), 1–19.

18. Cherry, E.C. Some experiments on the recognition of speech, with one and with two ears. Journal of the Acoustic Society of America, 25, 5 (1953), 975–979.

198 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

19. Cootes, t.F.; taylor, C.J.; Cooper, D.h.; and graham, J. Active shape models—their training and application. Computer Vision and Image Understanding, 61, 1 (1995), 38–59.

20. DePaulo, B.M., and kirkendol, S.E. the motivational impairment effect in the commu-nication of deception. In J.C. yuille (ed.), Credibility Assessment. Dordrecht: kluwer, 1989, pp. 51–70.

21. DePaulo, B.M.; kirkendol, S.E.; tang, J.; and O’Brien, t.P. the motivational impairment effect in the communication of deception: replications and extensions. Journal of Nonverbal Behavior, 12, 3 (1988), 177–201.

22. DePaulo, B.M.; Charlton, k.; Cooper, h.; lindsay, J.J.; and Muhlenbruch, l. the accuracy-confidence correlation in the detection of deception. Personality and Social Psychology Review, 1, 4 (1997), 346–357.

23. Derrick, D.C.; Elkins, A.C.; Burgoon, J.k.; Nunamaker, J.F., Jr.; and Zeng, D.D. Border security credibility assessments via heterogeneous sensor fusion. IEEE Intelligent Systems, 25, 3 (2010), 41–49.

24. Ekman, P., and Friesen, W.V. hand movements. Journal of Communication, 22, 4 (1972), 353–374.

25. Elaad, E., and Ben-Shakhar, g. Countering countermeasures in the concealed informa-tion test using covert respiration measures. Applied Psychophysiology and Biofeedback, 34, 3 (2009), 197–208.

26. Fanselow, M.S. Neural organization of the defensive behavior system responsible for fear. Psychonomic Bulletin & Review, 1, 4 (1994), 429–438.

27. Fukuda, k. Eye blinks: New indices for the detection of deception. International Journal of Psychophysiology, 40, 3 (2001), 239–245.

28. Fuller, C.M.; Biros, D.P.; Burgoon, J.k.; and Nunamaker, J.F., Jr. An examination and validation of linguistic constructs for studying high-stakes deception. Group Decision and Negotiation, 22, 1 (2013), 117–134.

29. gamer, M.; Bauermann, t.; Stoeter, P.; and Vossel, g. Covariations among fMrI, skin conductance, and behavioral data during processing of concealed information. Human Brain Mapping, 28, 12 (2007), 1287–1301.

30. ganis, g.; kosslyn, S.M.; Stose, S.; thompson, W.l.; and yurgelun-todd, D.A. Neural correlates of different types of deception: An fMrI investigation. Cerebral Cortex, 13, 8 (2003), 830–836.

31. garrido, E.; Masip, J.; and herrero, C. Police officers’ credibility judgments: Accuracy and estimated ability. International Journal of Psychology, 39, 4 (2004), 254–275.

32. glancy, F.h. and yadav, S.B. A computational model for financial reporting fraud detec-tion. Decision Support Systems, 50, 3 (2011), 595–601.

33. gray, J.A. The Psychology of Fear and Stress, 2d ed. Cambridge: Cambridge university Press, 1988.

34. hartwig, M.; granhag, P.A.; Stromwall, l.A.; and kronkvist, O. Strategic use of evidence during police interviews: When training to detect deception works. Law and Human Behavior, 30, 5 (2006), 603–619.

35. humpherys, S.l.; Moffitt, k.C.; Burns, M.B.; Burgoon, J.k.; and Felix, W.F. Identifica-tion of fraudulent financial statements using linguistic credibility analysis. Decision Support Systems, 50, 3 (2011), 585–594.

36. Iacono, W.g. Encouraging the use of the guilty knowledge test (gkt): What the gkt has to offer law enforcement. In B. Verschuere, g. Ben-Shakhar, and E. Meijer (eds.), Memory Detection: Theory and Application of the Concealed Information Test. New york: Cambridge university Press, 2011, pp. 12–24.

37. Jensen, M.l.; lowry, P.B.; and Jenkins, J.l. Effects of automated and participative deci-sion support in computer-aided credibility assessment. Journal of Management Information Systems, 28, 1 (Summer 2011), 201–233.

38. Jensen, M.l.; lowry, P.B.; Burgoon, J.k.; and Nunamaker, J.F., Jr. technology dominance in complex decision making: the case of aided credibility assessment. Journal of Management Information Systems, 27, 1 (2010), 175–201.

39. Jensen, M.l.; Meservy, t.O.; Burgoon, J.k.; and Nunamaker, J.F., Jr. Automatic, multimod-al evaluation of human interaction. Group Decision and Negotiation, 19, 4 (2010), 367–389.

40. kalin, N.h. the neurobiology of fear. Scientific American, 268, 5 (1993), 94–101.

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 199

41. kanaujia, A.; huang, y.; and Metaxas, D. tracking facial features using mixture of point distribution models. Computer Vision, Graphics and Image Processing, 4338 (2006), 492–503.

42. kPMg. Integrity Survey 2005–2006. Amsterdam, 2006.43. kPMg. Integrity Survey 2008–2009. Amsterdam, 2009.44. lang, P.J. the emotion probe: Studies of motivation and attention. American Psycholo-

gist, 50, 5 (1995), 371–385.45. langleben, D.D. Detection of deception with fMrI: Are we there yet? Legal and Crimi-

nological Psychology, 13, 1 (2008), 1–9.46. lee, C.C.; Welker, r.B.; and Odom, M.D. Features of computer-mediated, text-based

messages that support automatable, linguistics-based indicators for deception detection. Journal of Information Systems, 23, 1 (2009), 5–24.

47. levine, t.r.; Shaw, A.; and Shulman, h.C. Increasing deception detection accuracy with strategic questioning. Human Communication Research, 36, 2 (2010), 216–231.

48. lykken, D.t. the gSr in the detection of guilt. Journal of Applied Psychology, 43, 6 (1959), 385–388.

49. lykken, D.t. the validity of the guilty knowledge technique: the effects of faking. Journal of Applied Psychology, 44, 4 (1960), 258–262.

50. lykken, D.t. Psychology and the lie detector industry. American Psychologist, 29, 10 (1974), 725–739.

51. lykken, D.t. A Tremor in the Blood: Uses and Abuses of the Lie Detector. New york: Plenum trade, 1998.

52. Maclaren, V.V. A quantitative review of the guilty knowledge test. Journal of Applied Psychology, 86, 4 (2001), 674–683.

53. Matsuda, I.; hirota, A.; Ogawa, t.; takasawa, N.; and Shigemasu, k. Within-individual discrimination on the concealed information test using dynamic mixture modeling. Psychophysi-ology, 46, 2 (2009), 439–449.

54. Meijer, E.; Verschuere, B.; and Merckelbach, h. Detecting criminal intent with the con-cealed information test. Open Criminology Journal, 3 (2010), 44–47.

55. Meservy, t.O.; Jensen, M.l.; kruse, J.; Burgoon, J.k.; and Nunamaker, J.F., Jr. Automatic extraction of deceptive behavioral cues from video. In P. kantor, g. Muresan, F. roberts, D. Zeng, F.-y. Wang, h. Chen, and r. Merkle (eds.), Proceedings of the IEEE International Conference on Intelligence and Security Informatics (ISI 2005). Berlin: Springer, pp. 198–208.

56. Meservy, t.O.; Jensen, M.l.; kruse, J.; Burgoon, J.k.; Nunamaker, J.F., Jr.; twitchell, D.P.; tsechpenakis, g.; and Metaxas, D.N. Deception detection through automatic, unobtrusive analysis of nonverbal behavior. IEEE Intelligent Systems, 20, 5 (2005), 36–43.

57. Nakayama, M. Practical use of the concealed information test for criminal investigation in Japan. In M. kleiner (ed.), Handbook of Polygraph Testing. San Diego: Academic Press, 2002, pp. 49–86.

58. National research Council. The Polygraph and Lie Detection. Committee to review the Scientific Evidence on the Polygraph (ed.). Washington, DC: National Academies Press, 2003.

59. Nunamaker, J.F., Jr., and Briggs, r.O. toward a broader vision for information systems. ACM Transactions on Management Information Systems, 2, 4 (2011), article 20.

60. Nunamaker, J.F., Jr.; Chen, M.; and Purdin, t.D.M. Systems development in informa-tion systems research. Journal of Management Information Systems, 7, 3 (Winter 1990–91), 89–106.

61. Nunamaker, J.F., Jr.; twyman, N.W.; and giboney, J.S. Breaking out of the design sci-ence box: high-value impact through multidisciplinary design science programs of research. In Proceedings of the Nineteenth Americas Conference on Information Systems. Chicago: As-sociation for Information Systems, 2013.

62. Nunamaker, J.r., Jr.; Briggs, r.O.; Mittleman, D.D.; Vogel, D.r.; and Balthazard, P.A. lessons from a dozen years of group support systems research: A discussion of lab and field findings. Journal of Management Information Systems, 13, 3 (Winter 1996–97), 163–207.

63. Nunamaker, J.F., Jr.; Derrick, D.C.; Elkins, A.C.; Burgoon, J.k.; and Patton, M.W. Em-bodied conversational agent–based kiosk for automated interviewing. Journal of Management Information Systems, 28, 1 (Summer 2011), 17–48.

200 tWyMAN, ElkINS, BurgOON, AND NuNAMAkEr

64. Nunamaker, J.F., Jr.; Burgoon, J.k.; twyman, N.W.; Proudfoot, J.g.; Schuetzler, r.; and giboney, J.S. Establishing a foundation for automated human credibility screening. In Proceed-ings of the 2012 IEEE International Conference on Intelligence and Security Informatics. los Alamitos, CA: IEEE Computer Society, 2012, pp. 202–211.

65. Osugi, A. Daily application of the concealed information test: Japan. In B. Verschuere, g. Ben-Shakhar, and E. Meijer (eds.), Memory Detection: Theory and Application of the Con-cealed Information Test. Cambridge: Cambridge university Press, 2011, pp. 253–275.

66. Patton, M.W. Decision support for rapid assessment of truth and deception using automated assessment technologies and kiosk-based embodied conversational agents. Ph.D. dissertation, university of Arizona, tucson, 2009.

67. Porter, S., and ten Brinke, l. the truth about lies: What works in detecting high-stakes deception? Legal and Criminological Psychology, 15, 1 (2010), 57–75.

68. roelofs, k.; hagenaars, M.A.; and Stins, J. Facing freeze: Social threat induces bodily freeze in humans. Psychological Science, 21, 11 (2010), 1575–1581.

69. Schenberg, l.C.; Vasquez, E.C.; and Dacosta, M.B. Cardiac baroreflex dynamics during the defense reaction in freely moving rats. Brain Research, 621, 1 (1993), 50–58.

70. Shan, l.; tsechpenakis, g.; Metaxas, D.N.; Jensen, M.l.; and kruse, J. Blob analysis of the head and hands: A method for deception detection. In r.h. Sprague (ed.), Proceedings of the 38th Annual Hawaii International Conference on System Sciences. Washington, DC: IEEE Computer Society, 2005.

71. Silver, M.S.; Markus, M.l.; and Beath, C.M. the information technology interaction model: A foundation for the MBA core course. MIS Quarterly, 19, 3 (1995), 361–390.

72. Sokolov, E.N. higher nervous functions—Orienting reflex. Annual Review of Physiology, 25, 1 (1963), 545–580.

73. Sporer, S.l., and Schwandt, B. Moderators of nonverbal indicators of deception—A meta-analytic synthesis. Psychology, Public Policy, and Law, 13, 1 (2007), 1–34.

74. Stana, r.M. Moving illegal proceeds: Opportunities exist for strengthening the federal government’s efforts to stem cross-border currency smuggling. testimony before the Senate Caucus on International Narcotics Control, Washington, DC, 2011.

75. Strömwall, l.A., and granhag, P.A. how to detect deception? Arresting the beliefs of police officers, prosecutors and judges. Psychology, Crime & Law, 9, 1 (2003), 19–36.

76. Strömwall, l.A.; granhag, P.A.; and hartwig, M. Practitioners’ beliefs about deception. In P.A. granhag and l.A. Strömwall (eds.), The Detection of Deception in Forensic Contexts. New york: Cambridge university Press, 2004, pp. 229–250.

77. twyman, N.W.; Burgoon, J.k.; and Nunamaker, J.F., Jr. Non-invasive screening for con-cealed information. In r.h. Sprague (ed.), Proceedings of the 44th Annual Hawaii International Conference on System Sciences. koloa, hI: IEEE Computer Society, 2011.

78. twyman, N.W.; Burgoon, J.k.; Elkins, A.C.; and Proudfoot, J.g. Alternative cues in concealed information testing. In r.h. Sprague (ed.), Proceedings of the 46th Annual Hawaii International Conference on System Sciences. Maui, hI: IEEE Computer Society, 2013.

79. twyman, N.W.; Moffitt, k.; Burgoon, J.k.; and Marchak, F. using eye tracking technology as a concealed information test. In r.h. Sprague (ed.), Proceedings of the 43rd Annual Hawaii International Conference on System Sciences. koloa, hI: IEEE Computer Society, 2010.

80. Verschuere, B.; Ben-Shakhar, g.; and Meijer, E. Memory Detection: Theory and Applica-tion of the Concealed Information Test. Cambridge: Cambridge university Press, 2012.

81. Verschuere, B.; Crombez, g.; de Clercq, A.; and koster, E.h.W. Autonomic and behav-ioral responding to concealed information: Differentiating orienting and defensive responses. Psychophysiology, 41, 3 (2004), 461–466.

82. Verschuere, B.; Crombez, g.; de Clercq, A.; and koster, E.h.W. Psychopathic traits and autonomic responding to concealed information in a prison sample. Psychophysiology, 42, 2 (2005), 239–245.

83. Verschuere, B.; Crombez, g.; Smolders, l.; and de Clercq, A. Differentiating orienting and defensive responses to concealed information: the role of verbalization. Applied Psychophysiol-ogy and Biofeedback, 34, 4 (2009), 237–244.

84. Viola, P., and Jones, M.J. robust real-time face detection. International Journal of Com-puter Vision, 57, 2 (2004), 137–154.

rIgIDIty DEtECtION SyStEM FOr CrEDIBIlIty ASSESSMENt 201

85. Vrij, A. Behavioral correlates of deception in a simulated police interview. Journal of Psychology, 129, 1 (1995), 15–28.

86. Vrij, A., and Mann, S. telling and detecting lies in a high-stake situation: the case of a convicted murderer. Applied Cognitive Psychology, 15, 2 (2001), 187–203.

87. Vrij, A.; Mann, S.; and Fisher, r.P. An empirical test of the behaviour analysis interview. Law and Human Behavior, 30, 3 (2006), 329–345.

88. Vrij, A.; Semin, g.r.; and Bull, r. Insight into behavior displayed during deception. Human Communication Research, 22, 4 (1996), 544–562.

89. Vrij, A.; granhag, P.A.; Mann, S.; and leal, S. Outsmarting the liars: toward a cognitive lie detection approach. Current Directions in Psychological Science, 20, 1 (2011), 28–32.

90. Zhou, l., and Zhang, D. Following linguistic footprints: Automatic deception detection in online communication. Communications of the ACM, 51, 9 (2008), 119–122.

91. Zhou, l.; Burgoon, J.k.; Nunamaker, J.F., Jr.; and twitchell, D.P. Automating linguistics-based cues for detecting deception in text-based asynchronous computer-mediated communica-tion. Group Decision and Negotiation, 13, 1 (2004), 81–106.

92. Zuckerman, M.; DePaulo, B.M.; and rosenthal, r. Verbal and nonverbal communication of deception. Advances in Experimental Social Psychology, 14, 1 (1981), 1–59.

Appendix

thE ExpErimENts iN this iNvEstigatioN WErE dEsigNEd With a lEvEl oF rEalism. As such, it was important to gauge whether participants approached the interviews with a reasonable level of motivation, effort, and tension. the manipulation check questions used are included in tables A1 and A2.

table A1. Self-report Questions for Experiment 1 (Seven-Point Scale)

Measure Question

Motivation to succeed During the interview, how important was it to you to succeed in making the interviewer believe you?

Effort How hard did you try to convince the interviewer that you were telling the truth?

Tension How tense did you feel during the interview?

table A2. Self-report Questions for Experiment 2 (Five-Point Scale)

Measure Question

Motivation to succeed During the interview, how important was it to you to succeed in making the interviewer believe you?

Effort How hard did you try to convince the interviewer that you were telling the truth?

Tension How much do you agree (or disagree) with the following statement? I was tense during the screening process.