Systems Design Thinking - Deep Blue Repositories

154
Systems Design Thinking: Identification and Measurement of Attitudes for Systems Engineering, Systems Thinking, and Design Thinking by Melissa T. Greene A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy (Design Science) in the University of Michigan 2019 Doctoral Committee: Professor Richard Gonzalez, Co-Chair Professor Panos Y. Papalambros, Co-Chair Associate Professor Eytan Adar Dr. Anna-Maria Rivas McGowan, National Aeronautics and Space Administration Professor Oscar Ybarra

Transcript of Systems Design Thinking - Deep Blue Repositories

Systems Design Thinking: Identification and Measurement of Attitudes for Systems

Engineering, Systems Thinking, and Design Thinking

by

Melissa T. Greene

A dissertation submitted in partial fulfillment

of the requirements for the degree of

Doctor of Philosophy

(Design Science)

in the University of Michigan

2019

Doctoral Committee:

Professor Richard Gonzalez, Co-Chair

Professor Panos Y. Papalambros, Co-Chair

Associate Professor Eytan Adar

Dr. Anna-Maria Rivas McGowan, National Aeronautics and Space Administration

Professor Oscar Ybarra

Melissa T. Greene

[email protected]

ORCID iD: 0000-0002-7063-2500

© Melissa T. Greene 2019

ii

DEDICATION

To Dad

Love, Tess

iii

ACKNOWLEDGMENTS

First and foremost, my sincerest thanks to Anna McGowan for recognizing my potential,

introducing me to like minds at NASA Langley, and connecting me with the thought leaders in

Design Science at the University of Michigan. Without you, I’d have never found “my people.”

To Panos Papalambros, my advisor, chair and mentor: you have helped me grow

intellectually, professionally, and personally. Thank you for challenging me to be the best

version of myself, for encouraging me to be confident, and for your empathy and kindness

through it all. It has been an honor and a privilege to learn from you.

To Rich Gonzalez: learning to leverage the incredible amount of knowledge, experience,

and skill available to me at Michigan was and continues to be a challenge. Thank you for

allowing me the time and space to figure it out. I sincerely appreciate your intellectual

contributions to the work and your patience and support as I learn how to communicate.

To DESCI and ODE, past and present: thank you all for being wonderful friends and

colleagues. I have many fond memories and have learned so much from each of you. To Vignesh

and Sanjana, thank you for your intellectual contributions. Aria, thank you for sharing this

journey with me – what a long, strange trip it’s been.

Finally, and most importantly, my deepest gratitude to my family, for believing in me

every step of the way. Mom and Dad, thank you for your unwavering support through every

decision, pivot, change of plan, and change of location over the last ten years. You have enabled

me to realize my talents and fulfill my potential. I will be forever grateful. To Doug and Steve,

my brothers and best friends, thank you for always reminding me what’s truly important, for

iv

keeping me grounded, and for being a constant source of fun and light in my life. To Patty and

Ryanne, my soul sisters, thank you for being the wonderful women you are. I’m so thankful to

know you.

To Andy, my boots on the ground: thank you for living the “every day” with me, for truly

understanding what it took to get here, and for helping me make it happen. You are the best

teammate and partner and I am so grateful for you.

To all: thanks again. I couldn’t have done it without you.

v

TABLE OF CONTENTS

DEDICATION ................................................................................................................................ ii

ACKNOWLEDGMENTS ............................................................................................................. iii

LIST OF TABLES ......................................................................................................................... ix

LIST OF FIGURES ....................................................................................................................... xi

LIST OF APPENDICES ............................................................................................................... xii

ABSTRACT ................................................................................................................................. xiii

CHAPTER

I. What is “Systems Design Thinking?” ......................................................................................... 1

1.1 Introduction ......................................................................................................................... 1

1.2 Research Questions and Methodology................................................................................ 7

1.2.1 Understanding Systems Design Thinking Attitudes .................................................... 8

1.2.2 Measuring Systems Design Thinking Attitudes ........................................................ 10

1.3 Dissertation Overview ...................................................................................................... 10

II. Understanding Systems Design Thinking Attitudes ................................................................ 13

2.1 Introduction ......................................................................................................................... 13

2.2 Literature Review................................................................................................................ 14

2.2.1 An Introduction to the “Systems Approach” for Dealing with Complexity .............. 14

vi

2.2.2 Systems Engineering .................................................................................................. 16

2.2.3 Systems Thinking....................................................................................................... 19

2.2.4 Design Thinking......................................................................................................... 25

2.3 Developing a Codebook for Interview Analysis ............................................................... 27

2.3.1 Systems Engineering Codes ....................................................................................... 28

2.3.2 Systems Thinking Codes............................................................................................ 30

2.3.3 Design Thinking Codes.............................................................................................. 32

2.4 From Frameworks to Attitudes: Interviews with Systems Engineers ............................... 34

2.4.1 Method ....................................................................................................................... 35

2.4.2 Analysis...................................................................................................................... 35

2.4.3 Findings...................................................................................................................... 38

2.5 Summary ........................................................................................................................... 41

III. Modeling Systems Design Thinking Attitudes ....................................................................... 43

3.1 Introduction ......................................................................................................................... 43

3.2 Study 1: Technical, Organizational, and Social Systems Thinking .................................. 45

3.2.1 Scale Development .................................................................................................... 45

3.2.2 Pilot Test, Factor Analysis, and Results for Study 1 ................................................. 48

3.2.3 Discussion .................................................................................................................. 52

3.3 Study 2: Systems Thinking and Design Thinking ............................................................ 55

3.3.1 Comparing Systems Thinking and Design Thinking ................................................. 56

vii

3.3.2 Scale Development .................................................................................................... 59

3.3.3 Pilot Test, Factor Analysis, and Results for Study 2 ................................................. 62

3.3.4 Discussion .................................................................................................................. 64

3.4 Study 3: Systems Engineering and Design Thinking ....................................................... 65

3.4.1 Comparing Systems Engineering and Design Thinking ............................................ 66

3.4.2 Scale Development and Pilot Test ............................................................................. 70

3.4.3 Exploratory Factor Analyses...................................................................................... 77

3.4.4 Confirmatory Factor Analyses ................................................................................... 83

3.4.5 Multigroup CFA......................................................................................................... 87

3.4.6 Tests for Measurement Invariance ............................................................................. 91

3.4.7 Additional Qualitative Findings ................................................................................. 92

3.4.8 Discussion .................................................................................................................. 94

3.5 Summary ........................................................................................................................... 96

IV. Validating the Systems Design Thinking Scale...................................................................... 98

4.1 Introduction ....................................................................................................................... 98

4.2 Behavioral Research in Systems Engineering and Design Thinking ................................ 99

4.3 Pilot Validation Study ..................................................................................................... 101

4.3.1 Overview and Objectives ......................................................................................... 101

4.3.2 Methods.................................................................................................................... 101

4.3.3 Behavioral Task Selection ....................................................................................... 102

viii

4.3.4 Study Population and Recruitment Strategy ............................................................ 104

4.3.5 Pilot Test, Factor Analysis, and Results for Validation Study ................................ 104

4.3.6 Findings and Lessons Learned ................................................................................. 107

4.4 Validation Opportunities ................................................................................................. 109

4.5 Summary ......................................................................................................................... 110

V. Conclusion ............................................................................................................................. 111

5.1 Summary of Dissertation ................................................................................................ 111

5.2 Contributions to Design Science ..................................................................................... 112

5.3 Limitations and Opportunities for Future Work ............................................................. 113

Appendices .................................................................................................................................. 116

Bibliography ............................................................................................................................... 126

ix

LIST OF TABLES

Table

2.1 Five themes were coded in all ten interviews…………………………………………….36

2.2 Design thinking themes, number of interviews including each theme, and total number of

references to each theme……..…………………………………………………………………..37

2.3 Systems thinking themes, number of interviews including each theme, and total number of

references to each theme…………………………………………………………………………37

2.4 Systems engineering themes, number of interviews including each theme, and total

number of references to each theme...…………………………………………………………...38

3.1 Technical systems thinking attitude items tested in Study 1……………………………..46

3.2 Organizational systems thinking attitude items tested in Study 1………………………..47

3.3 Social systems thinking attitude items tested in Study 1…………………………………47

3.4 Post-hoc exploratory factor analysis results for technical systems thinking attitude items in

Study 1…………………………………….……………………………………………………..49

3.5 Post-hoc exploratory factor analysis results for social systems thinking attitude items in

Study 1 …………………………………………………………………………………………..50

3.6 Post-hoc exploratory factor analysis results for organizational systems thinking attitude

items in Study 1 …………………………………………………………………………………51

3.7 Summary of factor structure after post-hoc exploratory factor analysis (Study 1)………51

3.8 Example systems thinking attitude items and themes from Study 2 …………………….61

3.9 Example design thinking attitude items and themes from Study 2………………………62

x

3.10 Two-factor EFA results with varimax rotated loadings (Study 2)………………………63

3.11 Systems engineering attitude items from Study 3.………………………………………71

3.12 Design thinking attitude items from Study 3……………………………………………72

3.13 Posting the Systems Design Thinking Scale on Reddit……...………………………….77

3.14 Varimax rotated loadings for four factors (Study 3)…………………………………….80

3.15 Post-hoc EFA results: Varimax rotated loadings for three factors (Study

3)……..………..............................................................................................................................81

3.16 Two-factor EFA results (Study 3)…….………………………………………………...82

3.17 Final factor loadings (Study 3)….………………………………………………………83

3.18 Model results from confirmatory factor analysis with 12 items…..…………………….85

3.19 Model results from confirmatory factor analysis with 9 items...………………………..86

3.20 Multigroup CFA results: Reddit vs. known expert sample……………………………...88

3.21 CFA for Reddit group…………………………………………………………………...88

3.22 CFA for known expert sample………………………………………………………......89

3.23 Multigroup CFA results: Entry vs. senior level (combined sample)……………………89

3.24 CFA for entry-level group………………………………………………………………90

3.25 CFA for senior-level group……………………………………………………………...90

4.1 Model results from confirmatory factor analysis (validation study)...………………….105

4.2 Comparison in factor loadings between validation study and Study 3 for items DT7, DT9,

and DT10……………………………………………………………………….………………105

4.3 Covariances of factors and tasks………………………………………………………...106

4.4 Regression model results………………………………………………………………..106

xi

LIST OF FIGURES

Figure

2.1 Design thinking process models from IDEO, Stanford d-school, and Google Design

Sprint….…………………………………………...……………………………………………..27

2.2 Systems engineering process models. On the left is the INCOSE systems engineering

“vee” and the NASA “systems engineering engine” is on the right….………………………….29

2.3 Systems design thinking attitude model…...……………………………………………..40

3.1 A graphical representation of the three types of systems thinking—technical, social, and

organizational………………………………………………………..…………………………...45

3.2 Findings from Study 1 suggest that social systems thinking items seem to overlap with the

design thinking framework………………………………………………………………………55

3.3 In Study 2, systems thinking items from Study 1 are redistributed and a new model is

tested……………………………………………………………………………………………..56

3.4 Systems engineering and design thinking frameworks each include elements of systems

thinking…………………………………………………………………………………………..64

3.5 Two-factor model with parameter values and standard error…………………………….84

3.6 Systems Design Thinking Classification of 458 survey participants……………………..87

xii

LIST OF APPENDICES

Appendix

A. Systems Design Thinking Codebook…………………………………………………...117

B. Semi-Structured Interview Questions.………………………………………………….122

xiii

ABSTRACT

Systems engineering, systems thinking, and design thinking are frameworks for

understanding complex problems and developing effective, holistic solutions. Each framework is

comprised of assumptions, concepts, values, and practices that affect the design of products,

systems, and services. In this dissertation, we explore the assumptions, concepts, values, and

practices that define systems engineering, systems thinking, and design thinking, and compare

them using a mixed methods approach. This dissertation also explores the existence and

definition of systems design thinking—an integrated framework for systems engineering, systems

thinking, and design thinking—along with the development of the Systems Design Thinking

Scale. The Systems Design Thinking Scale is a 5-point Likert scale survey that measures

attitudes about systems engineering, systems thinking, and design thinking, and is used to

provide insight about potential relationships between these attitudes. Such a scale may be used

for categorizing individuals based on these attitudes, which could be useful for informing

teaming and other management decisions in design organizations.

The development of the Systems Design Thinking Scale in this dissertation was

conducted as follows. First, thematic analysis of the systems engineering, systems thinking, and

design thinking literature was used to generate codes that reflect core assumptions, concepts,

values, and practices of each framework. These codes were then compiled into a systems design

thinking codebook, and used to analyze data from semi-structured interviews with experienced

systems engineers who were also recognized as strong systems thinkers by a technical leader

xiv

within their organization. Interview data was used to identify common attitudes reflecting

systems engineering, systems thinking, and design thinking in practice, and to generate

hypotheses about how the frameworks are related. These attitudes were represented as statements

on a 5-point Likert scale and distributed to a diverse sample of engineers and designers.

Exploratory and confirmatory factor analysis were used to determine how well the attitudes

reflected systems engineering, systems thinking, and design thinking; and to test the

hypothesized relationships between these frameworks quantitatively. Ethnography informs the

research throughout.

Findings suggest several nuances that distinguish systems engineering, systems thinking,

and design thinking. Findings also suggest that systems thinking attitudes exist within both

systems engineering and design thinking frameworks. Results from the factor analyses suggests

that systems engineering and design thinking attitudes are independent, and individuals may

have systems engineering attitudes, design thinking attitudes, or both. A higher correlation

between these attitudes is observed for experts in engineering design.

The final version of the scale is a 9-item questionnaire about systems engineering and

design thinking attitudes. An exploratory study for validating the scale is described, in which

correlations between scale scores and performance on analytical reasoning and divergent

thinking tasks are examined. While no significant correlation was observed between the

subscales and performance on the analytical reasoning task, some correlation between the design

thinking subscale and divergent thinking measure suggests that the Systems Design Thinking

Scale may be useful for predicting certain behaviors. Further validation through gamification and

other opportunities for future work are discussed.

1

CHAPTER I

What is “Systems Design Thinking?”

1.1 Introduction

Systems engineering and design thinking are methodologies for developing products,

systems, and services. INCOSE, the International Council for Systems Engineering, defines

systems engineering as “an interdisciplinary approach and means to enable the realization of

successful systems” that “focuses on defining customer needs and required functionality early in

the development cycle, documenting requirements, then proceeding with design synthesis and

system validation (INCOSE, 2015).” At NASA, emphasis is on “the satisfaction of stakeholder

functional, physical, and operational performance requirements in the intended use environment

over the planned life cycle within cost and schedule constraints (Conner, 2015).” Systems

engineering arose with the increase in complexity of military-industrial systems in the 1940s. As

defense projects increased in size and scope, computational tools for requirements analysis,

modeling and simulation, coordination, and scheduling were necessary for successful system

design, implementation, and decommission (Yassine and Braha, 2003; Braha and Bar-Yam,

2007).

Design thinking, now commonly applied as a product development framework, also arose

in response to increasing complexity. Design thinking is “a human-centered approach to

innovation that draws from the designer’s toolkit to integrate the needs of the people, the

possibilities of technology, and the requirements for business success (Brown, 2008).” Design

2

thinking is a methodology for defining and solving problems. It is particularly useful for tackling

product design problems that are ill-defined or unknown, by “understanding the human needs

involved, by reframing the problem in human-centric ways, by creating many ideas in

brainstorming sessions, and by adopting a hands-on approach in prototyping and testing (Dam

and Siang, 2019).” The recognition of “wicked problems” in social planning and policy

(Buchanan, 1992), where solutions required a great number of people to change their mindsets

and behavior, helped first draw attention to the fact that comprehensive needs assessment and

problem definition are critical first steps in designing successful solutions. Design thinking

evolved as a way to accomplish these goals and maintain human-centered values from

conceptual design through to embodiment.

Systems engineering and design thinking were developed with different applications,

approaches, and goals in mind. However, the value of integrating their principles and processes

is becoming increasingly recognized in both communities (Dym et al., 2005; IndustryWeek,

2018; Liedtka and MacLaren, 2018). Systems engineering organizations like governmental

mission agencies and global manufacturing corporations are exploring opportunities for applying

design thinking principles in systems engineering projects (Souza and Barnhöfer, 2015; Darrin

and Devereux, 2017; McGowan et al., 2013; McGowan et al., 2017). Short courses where

engineers can learn design thinking methods through a hands-on design process and reflection

are common. Some organizations have dedicated research groups for developing

multidisciplinary approaches that include design thinking (Grace et al., 2017). Similarly, the

product design community has recognized the need for a more systematic approach as the

complexity of design tasks at hand has increased (Greene et al., 2017). With the pervasiveness of

"smart" technology, mechatronic products requiring a combination of mechanical, electronics,

3

and computer engineering also require the performance of systems engineering sub-processes for

successful integration. The expansion of product-service systems, in which products and the

services that support them are integrated and delivered simultaneously, has created a further need

for a systematic approach to product design (Baines et al., 2007; Cavalieri and Pezzotta, 2012).

Many companies that are producing highly complex products are already doing systems

engineering, although some do this more formally than others and with varying levels of

completeness (IndustryWeek, 2018).

Both approaches have value, but integrating the two is not without its challenges. For

some systems engineers, design thinking represents a non-traditional and unfamiliar approach.

The emphasis on defining the problem over defining requirements for a solution creates

dissimilar interpretations of the problem scope and project objectives (McGowan et al., 2017).

Likewise, pursuing the fully integrated scope of systems engineering processes from design to

operation to decommission may seem overwhelming or unnecessary to many product designers

designing on a smaller scale with less complexity and risk. However, the ultimate goal of both

types of designers is the same: developing fully verified and validated products that result in

higher quality and customer acceptance (IndustryWeek, 2018).

The challenge in integrating systems engineering and design thinking processes may be

attributed in part to a difference in attitudes held by engineers and designers. Attitudes are sets of

beliefs, emotions, and behaviors toward a particular object, person, process, or event. Attitudes

develop over time as the result of experience, education, and upbringing, and can have a

powerful influence over behavior (Allport, 1935). Attitudes can serve a variety of functions for

individuals, such as providing general approach or avoidance tendencies, helping organize and

interpret new information, protecting self-esteem, and expressing central values or beliefs (Eagly

4

and Chaiken, 1998; Katz, 1960). An individual’s attitudes about systems engineering and design

thinking can therefore influence both the implementation of these processes and their outcomes.

Why might these attitudes be different and how? According to Doob (1947), learning can

account for most of an individual's attitudes. Identifying the differences between engineering

education and design education can be a useful starting point for understanding the different

attitudes held by each professional community. A typical undergraduate engineering science

curriculum in the United States consists of mathematics, general chemistry, physics, computer

science, and other introductory engineering courses (Panel on Undergraduate Engineering

Education, 1986; Tribus, 2005), while design curricula are more likely to include courses like

psychology, sociology, anthropology, marketing, and other social and behavioral sciences (Ilhan,

2017; Self and Baek, 2017). Students in each curriculum regularly address different types of

problems and research questions, using different theoretical frameworks, knowledge, and

methods. Students in each curriculum likely get some exposure to courses in the other, but

courses and curricula are often not well-integrated. Design programs typically offer few

engineering courses. More engineering programs are expanding to include rigorous design

courses, but engagement with other disciplines is still limited, and design as it is taught in

engineering curricula differs significantly from design as it is taught in design school curricula.

Engineering students learn functional design and design optimization, nearly to the exclusion of

social, cultural, economic, and other design needs. Design students learn aesthetic design, design

processes, and theory, but often do not learn the engineering and manufacturing processes

required for embodiment.

After graduation, job function, responsibilities, and training are also different. Systems

engineering is concerned with assessing the technical and logistical feasibility of potential

5

solutions to given problems. Responsibilities include defining system requirements and

developing verification and validation plans for those requirements, with the goal of delivering a

fully integrated product that is reliable and safe in all operational conditions. Product designers

are more likely to consider business viability first. In the product design context, identifying the

right problem to solve is important and can be very lucrative. For this reason, product design

processes are more likely to include market research, user needs assessment, and other human-

centered analysis, and product design organizations are more likely to encourage an expansive,

socially and economically rich problem analysis and problem definition.

Because of these differences, systems engineers are often characterized as methodical,

analytical, and data-driven, while designers are often described as having a more flexible,

creative, and human-centered perspective. With the increased awareness that both systems

engineering and design thinking need each other, the effects of a possibly persisting distinction

on systems engineers’ attitudes toward design, and designers’ attitudes towards systems

engineering, are not well understood. This dissertation seeks to explore these attitudes as they

relate to a concept we will call systems design thinking. Systems design thinking is a

hypothesized set of attitudes that reflect an integrated perspective on systems engineering, design

thinking, and systems thinking, a related but distinct framework. Systems engineers who employ

a human-centered approach during complex systems design are believed to have systems design

thinking attitudes. Product designers who intuitively follow systems engineering processes to

integrate, verify, and validate complex consumer products may also have systems design

thinking attitudes.

In this work, the term ‘design thinking’ is used to represent the plurality of human-

centered processes that drive design decision-making. These include creativity, intuition, and

6

empathy, and also include design activities such as prototyping and testing with users and

collaborating and communicating with other designers. Systems engineering attitudes include

some fundamental attitudes about engineering that may be generalizable to any engineering

discipline. These include attitudes about mathematical modeling, simulation, analysis, and other

engineering processes. This is because systems engineers are discipline engineers first, trained to

analyze and solve problems using engineering science and mathematics. In their systems design

work, they learn to make analogies and draw connections to other disciplines, and to pursue

social interactions with other engineers in those disciplines, strengthening these connections and

gaining additional knowledge. This “philosophy” of systems engineering, described as the

“systems perspective” or “systems view,” is what distinguishes systems engineering from other

engineering disciplines (Frank, 2012). The “systems view” describes the ability to identify and

manage interactions between sub-systems in the technical system, as well as interactions

between individuals, disciplinary working groups, and organizations, to facilitate complex

systems design.

In this work, systems design thinking is explored through the development and validation

of a new instrument that we will call the Systems Design Thinking Scale. The Systems Design

Thinking Scale measures attitudes about systems engineering, systems thinking, and design

thinking, and is used to provide insight about potential relationships between these attitudes.

Scale development in this dissertation was conducted as follows. First, thematic analysis of the

systems engineering, systems thinking, and design thinking literature was used to generate codes

that reflect core assumptions, concepts, values, and practices of each framework. These codes

were then used to analyze data from semi-structured interviews with experienced systems

engineers, who were also recognized as strong systems thinkers by a technical leader within their

7

organization. Interview data was used to identify common attitudes reflecting systems

engineering, systems thinking, and design thinking in actual practice, and to generate hypotheses

about how the frameworks are related. These attitudes were represented as statements on a 5-

point Likert scale and distributed to a diverse sample of engineers and designers. Factor analysis

was used to determine how well the attitudes reflected systems engineering, systems thinking,

and design thinking; and to test the hypothesized relationships between these frameworks

quantitatively. Ethnography informs the research throughout.

In the remainder of this chapter, an overview of the research questions and methodology

is presented. We describe systems engineering, systems thinking, and design thinking

frameworks; how we get from frameworks to attitudes; and the approach for quantifying and

measuring them. Then, the larger dissertation overview is presented.

1.2 Research Questions and Methodology

This dissertation seeks to define better both the boundaries between systems engineering,

systems thinking, and design thinking frameworks, and their intersection at systems design

thinking. The following research questions are addressed:

• What are the core assumptions, concepts, values, and practices of systems

engineering, systems thinking, and design thinking frameworks?

• How are these frameworks related?

• What attitudes reflect these frameworks and their relationships in practice?

• Can these attitudes be represented and measured in a meaningful way?

By framework, we mean a set of assumptions, concepts, values, and practices that constitutes

a way of viewing reality (American Heritage Dictionary, 2016). The goal of the dissertation is to

8

identify some of these assumptions, concepts, values and practices for systems engineering,

systems thinking, and design thinking, and compare them using a mixed methods approach

(Crede and Borrego, 2013; Watkins and Gioia, 2015).

We hypothesize that systems thinking and design thinking are or can be integrated within

the context of systems engineering practice, based on our interpretation of relevant literature and

observation and interviews with systems engineering practitioners. We create and describe the

systems design thinking framework based on this hypothesis. Systems design thinking is an

integrated framework that situates systems thinking and design thinking within the context of

systems engineering practice. Our motivation for creating and studying this framework is to

expand existing systems engineering process models to include relevant cognitive and social

processes from systems thinking and design thinking frameworks. This is an important first step

towards appropriately applying and maximizing the potential of systems thinking and design

thinking as fully integrated systems engineering subprocesses.

1.2.1 Understanding Systems Design Thinking Attitudes

Qualitative methods are useful for conducting exploratory research. This dissertation

describes qualitative research for exploring the relationship between systems engineering,

systems thinking, and design thinking attitudes in the systems engineering context.

Informal ethnographic research informed many aspects of the dissertation (Griffin and

Bengry-Howell, 2017). The research questions were developed after spending several months in

a government laboratory and participating in complex systems research with engineers and

technologists with varying education, experience, and training. Relationships with experienced

systems engineering and design professionals were cultivated over the course of the research.

9

These relationships were useful for “member-checking,” to ensure accuracy, credibility, validity,

and transferability of the findings throughout, and to guide the development of the dissertation.

Thematic analysis of systems engineering, [engineering] systems thinking, and design

thinking literature was used to identify key assumptions, concepts, values, and practices of each

framework (Braun and Clarke, 2006). These assumptions, concepts, values, and practices were

integrated into a codebook for understanding systems design thinking attitudes (DeCuir-Gunby,

Marshall, and McCulloch, 2011). A codebook is a set of codes, definitions, and examples used as

a guide to help analyze interview data. Then, semi-structured, individual interviews were

conducted with a small, non-random sample of systems engineers. The codebook was used to

identify attitudes that reflect systems engineering, systems thinking, design thinking, and their

relationships in practice. Interviews also informed ideas and hypotheses about systems design

thinking for the quantitative study described in the next section.

The individuals selected for interview were recruited for their experience as systems

engineers, and also for their expertise in systems thinking as recognized by a technical leader in

the organization. Although the sample was small, the experience represented was rich and

diverse. Participants related to and were engaged with the research questions, and they described

complex cognitive, behavioral, and organizational processes and their relationships. They

articulated their thoughts, feelings, and strategies clearly, and direct quotations were often used

to reflect systems engineering, systems thinking, and design thinking attitudes in the Systems

Design Thinking Scale. Generalizability and transferability of these attitudes were tested

quantitatively when the scale was distributed to a larger, balanced sample.

10

1.2.2 Measuring Systems Design Thinking Attitudes

Quantitative analysis was used to test hypotheses about systems engineering, systems

thinking, and design thinking derived from the qualitative study findings. Theory and methods

from psychometrics were used to model systems engineering, systems thinking, and design

thinking as latent psychological constructs, measured by observable attitudes identified in

interviews. Attitudes are declarative statements that reflect key assumptions, concepts, values,

and practices of systems engineering, systems thinking, and design thinking. For example, the

statement “system requirements should be flexible and expected to change” is an attitude that

reflects systems thinking and the core value “flexibility.”

Systems engineering, systems thinking, and design thinking constructs were represented

by networks of survey items (attitude statements) in a 5-point Likert scale. Numerals were

assigned to each attitude statement to measure the degree to which participants agree or disagree

with each attitude. Structural Equation Modeling was used to determine how well each attitude

statement reflected the underlying construct, and to explore relationships between constructs in a

quantitative way. We hypothesize that systems design thinkers will endorse most systems

engineering, systems thinking, and design thinking attitudes. This would provide support for the

existence of systems design thinking, an integrated framework consisting of systems

engineering, systems thinking, and design thinking attitudes.

1.3 Dissertation Overview

In this chapter we introduced systems engineering, systems thinking, and design thinking

frameworks, and proposed a new integrated framework we call systems design thinking. In

Chapter 2, Understanding Systems Design Thinking Attitudes, “understanding” begins with a

review of systems engineering, systems thinking, and design thinking literature. Thematic

11

analysis is used to develop a codebook for interpreting differences between systems engineering,

systems thinking, and design thinking throughout the research. An exploratory qualitative study

is also described, in which semi-structured interviews are conducted with a small sample of

systems engineers. The purpose of the exploratory qualitative study is to identify attitudes that

reflect systems engineering, systems thinking, and design thinking frameworks, and how they are

related and integrated to reflect systems design thinking in practice.

In Chapter 3, Measuring Systems Design Thinking Attitudes, findings from the

qualitative studies guide quantitative comparison of systems engineering, systems thinking, and

design thinking attitudes. Three factor analysis studies are described. Study 1 explores systems

thinking attitudes and their relationship. Systems thinking is defined along three dimensions—

technical, social, and organizational— as suggested in interview findings. Attitudes statements

reflecting these dimensions are derived from literature and interview findings and represented as

three factors in a 5-point Likert scale. This scale was distributed to a small sample of practicing

engineers and engineering researchers in industry, academia, and government. Data was

analyzed using exploratory factor analysis. The data did not support the three-factor model of

technical, social, and organizational systems thinking, but provided other informative results.

Significant items in the social systems thinking factor closely resembled design thinking. This

comparison was explored further in Study 2, in which systems thinking and design thinking are

compared. In Study 2, significant systems thinking attitudes from Study 1 are retained and

reorganized, and new statements are added to reflect design thinking better. Technical and

organizational systems thinking attitudes grouped into one factor, and design thinking attitudes

grouped into a second factor. Significant technical and organizational systems thinking closely

resemble a systems engineering framework when the systems design thinking codebook is used

12

to interpret results. This finding is explored further in Study 3. In Study 3, systems engineering

attitudes are studied as they relate to design thinking attitudes. Significant items from Study 2 are

retained and improved and additional systems engineering items are added. This model was

tested on a larger sample recruited through the social media platform Reddit and yielded

promising results for measuring systems design thinking.

Findings suggest many nuances that differentiate systems engineering, systems thinking

and design thinking. Findings also suggest that systems thinking attitudes exist within both

systems engineering and design thinking frameworks. Results from the factor analyses suggests

that systems engineering and design thinking attitudes are independent, and individuals may

have systems engineering attitudes, design thinking attitudes, or both. A higher correlation

between these attitudes is observed for experts in engineering design, suggesting that the

integrated systems design thinking perspective may develop with education and experience.

Chapter 4 describes a first attempt at validating the Systems Design Thinking Scale by

studying the relationship between scale scores and performance on divergent thinking and

analytical reasoning tasks. While no significant correlation was observed between the subscales

and performance on the analytical reasoning task, some correlation between design thinking

subscale scores and performance on the divergent thinking measure suggests that the Systems

Design Thinking Scale may be useful for predicting some behaviors. Additional tasks and

methods are identified for their relevance and potential usefulness in validating the scale, and a

validation plan is outlined and discussed. Chapter 5 summarizes the main contributions of the

work and highlights some opportunities for future research.

13

CHAPTER II

Understanding Systems Design Thinking Attitudes

2.1 Introduction

This chapter describes a literature review and exploratory qualitative study. The goal is to

begin to understand systems design thinking by understanding systems engineering, systems

thinking, design thinking, and their relationship. A two-step research process is described. First,

a literature review and thematic analysis are conducted with the objective of producing a

codebook for systems design thinking. This codebook describes differentiating features of

systems engineering, systems thinking, and design thinking frameworks, for the purpose of

informing the rest of the research. In the second step, individual semi-structured interviews are

conducted with a small sample of practicing systems engineers. The codebook is used to identify

and interpret attitudes about each of the three frameworks and their relationships in practice.

The work described in this chapter is intended to address a gap in the existing literature.

While considerable effort is dedicated to understanding the individual nuances of systems

engineering, systems thinking, and design thinking frameworks, discussion about the

relationships between them is limited (Greene et al., 2017). Relatively little work has been done

to understand what distinguishes a “systems engineering” framework from a “systems thinking”

framework, or “systems thinking” from “design thinking.” These questions will be explored

throughout.

14

2.2 Literature Review

2.2.1 An Introduction to the “Systems Approach” for Dealing with Complexity

The history of the “systems approach” for dealing with design complexity is long and

diverse. The following section provides a brief overview of several major schools of systems

theory, beginning with Ludwig von Bertalanffy’s General Systems Theory (von Bertalanffy,

1940). von Bertalanffy’s work was quickly expanded to describe cybernetic systems (Wiener,

1948; Ashby, 1956) and dynamic systems (Forrester, 1961; Boulding, 1964; Meadows, 2000;

Sterman, 2000), and has also informed operations research, systems engineering, and

management science (Checkland, 1981). Later work examined the application of general systems

concepts to human social systems (Barker, 1968; Bateson, 1972; Luhmann, 1984; Parsons,

1951). General systems theory was somewhat useful for analyzing social and organizational

systems, but traditional physics-based models did not always adequately represent human factors

that influence system performance (Checkland, 1981; Forrester, 1994). Bateson, for example,

argues that cybernetic principles have direct mappings in social systems (1972). Barker’s work,

however, suggests that human behavior is radically situated, and that predictions about human

behavior can only be made if the situation, context, and environment in which the human is

operating can be sufficiently understood (1968).

The term “systems theory” coined by Ludwig von Bertalanffy in 1937 describes the

interdisciplinary study of systems in general. Systems theory emerged as an attempt to uncover

patterns and principles common to all levels of all types of systems, again with an emphasis on

generality. The primary goal in developing systems theory was to provide a useful framework for

describing a broad range of systems using the same terminology, in contrast to existing

discipline-specific systems models in biology, engineering, and psychology (von Bertalanffy,

15

1940). von Bertalanffy divided systems inquiry into three major domains: philosophy, science,

and technology (1968). Similar domains were also explored by Béla H. Bánáthy (1967), whose

discussion of systems theory included philosophy, theory, methodology, and application.

Philosophy refers to the ontology, epistemology, and axiology of systems; theory refers to a set

of interrelated concepts and principles that apply to all systems; methodology refers to the set of

models, tools, and strategies that operationalize systems theory and philosophy; and application

refers to the context and interaction of the domains. Systems inquiry is described as

“knowledgeable action:” philosophy and theory are integrated as knowledge, and method and

application are integrated as action.

A central tenet of systems theory is self-regulation, and systems theory is often applied to

describe systems that self-correct through feedback. These types of systems are found in nature,

in local and global ecosystems, and in human learning processes at both the individual and

organizational level (Laszlo, Levine, and Milsum, 1974). Early work in self-regulating systems

eventually led to the development of cybernetics – the formal study of communication and

control of regulatory feedback (Wiener, 1948). Cybernetics offers approaches for exploring

natural systems, mechanical systems, physical systems, cognitive systems, and social systems.

Cybernetics is applicable in the analysis of any system that incorporates a closed signaling loop.

While the cybernetics framework can be applied to analyze non-engineering systems,

engineering theory and methods are still used to represent the system.

Systems dynamics emerged in the mid-1950s as application of electrical control theory to

the analysis of business systems. Forrester (1961) developed a mathematical modeling technique

to help corporate managers improve their understanding of industrial practices. By simulating the

stock-flow-feedback structure of an organization, Forrester demonstrated that instability in

16

organizational employment was due to internal structure of the firm, and not to an external force

such as a business cycle. From the late 1950s to the late 1960s, system dynamics was applied

almost exclusively to corporate and managerial problems (Radzicki and Taylor, 2008). In 1969,

Forrester extended the system dynamics model beyond its corporate application in Urban

Dynamics. In this book, Forrester presents a simulation model that describes the major internal

forces controlling the balance of population, housing, and industry within an urban area

(Forrester, 1969).

2.2.2 Systems Engineering

Many decades ago, systems dynamics formed the core of some systems engineering

concepts. Systems engineering today describes an interdisciplinary field of formalized

approaches for designing and managing large-scale, complex engineered systems (LSCES)

throughout the life cycle. Systems engineering methodology offers a process for technical

management of LSCES. Sophisticated quantitative techniques are used to organize and

coordinate work activities, evaluate technical systems interactions, and assure system quality and

performance. To address personnel issues that influence LSCES design, systems engineering has

drawn from operations research and management science. Management involves identifying “the

mission, objective, procedures, rules and manipulation of the human capital of an enterprise to

contribute to the success of the enterprise” (P. Frank, 2006; Oliver, 1997).

Systems engineering considers issues such as requirements development and verification,

work-process management, and system safety/reliability, and utilizes methods such as

probabilistic risk assessment, modeling and simulations, and design optimization (Goode and

Machol, 1957; Papalambros and Wilde, 2017). The systems engineering approach is generally

reductionist in nature, and offers several different tools and methods for decomposing large,

17

complex systems into smaller, more manageable subsystems (Altus, Kroo, and Gage, 1996;

Browning, 2001). An aircraft, for example, can be deconstructed into subsystems such as

structures, controls, propulsion, etc., and strategies for arriving at this particular partitioning fall

within the scope of systems engineering. This type of decomposition-based approach to the

design of engineered systems requires significant forethought, as different partitioning strategies

can determine the effectiveness of the design process (Allison, 2008). Likewise, the way in

which subsystems are reintegrated, or coordinated, can similarly impact project success

(Forsberg and Mooz, 1992; Lake, 1992).

Systems engineering methods for identifying and minimizing the effects of interactions

between technical system elements are of particular interest in this work. Interactions between

technical system components are widely studied in systems engineering and systems design

optimization, most notably in the multidisciplinary design and optimization (MDO) literature,

through concepts and tools such as design structure matrices (Eppinger and Browning, 2012;

Steward, 1981), global sensitivity equations (Hajela, Bloebaum, and Sobieszczanski-Sobieski,

1990; Sobieszczanski-Sobieski, 1990), coupling metrics (Alyaqout et al., 2011; Kannan,

Bloebaum, and Mesmer, 2014) and partitioning and coordination methods for decomposition-

based design optimization (Allison, 2008; Lasdon, 1970).

Decomposition-based design optimization has been established as a valuable tool for

systems engineering design (Papalambros and Wilde, 2017). Complicated products or systems

can be simulated and designed using optimization algorithms, which accelerates product

development and drastically reduces the need for expensive physical prototypes (Allison, 2008).

Because the modeled products are inherently complicated, a single optimization algorithm

cannot usually accommodate the large number of design variables and constraints operating

18

simultaneously. The simulation-based design process must first be partitioned into smaller and

easier to solve subproblems, and the solutions of these subproblems must all be consistent and

system-optimal. Because optimization requires a parsing of the simulation into subproblems,

decisions about system decomposition are made before any formal system design activities

actually commence (Allison, 2008). Decomposition-based design optimization therefore depends

on a priori definition of partitioning and coordination strategies.

In design optimization, subproblems are linked through common design variables and

interactions, but generic approaches for specifying a partitioned problem are rare (Tosserams et

al. 2010). Seasoned engineers concede that a certain “element of artistry” is required for the

process to be successful (Buede, 2009). Systems engineers are expected to apply systems

thinking, and some aspire to become big picture “visionary designers” who manage technical

processes as well as social and organizational interactions in dynamic environments (Brooks et

al., 2011). The complexity of this task and its implications for system performance, cost, and

schedule led to the development of “soft” operations research (OR)— an extension of traditional

operations research that places less emphasis on mathematical modelling of business and social

systems and more on thoroughly defining system boundaries and problems, resolving conflicting

viewpoints, and reaching consensus on future action (Forrester, 1994). Soft OR methods

characterize systems on a variety of qualitative dimensions — e.g., physical vs. social, causal vs.

probabilistic, degree of complexity, or susceptibility to control— and utilize discussion and

intuition rather than quantitative methods to analyze systems engineering and design processes

(Forrester, 1994). While soft operations research and systems engineering are related in theory,

the two are rarely related in practice today.

19

2.2.3 Systems Thinking

Early realizations of the systems thinking concept in management and organization are

Karl Weick’s sensemaking framework and Peter Senge’s “learning organization” (Weick, 1979;

Senge, 1990). Sensemaking is the process of “creating situational awareness and understanding

in situations of high complexity in order to make decisions,” or, more simply, making sense of

the world in order to act in it (Klein, 2006). Sensemaking is a powerful bridging concept that

connects individual cognitive and affective processes with organizational structures and

behavior, to create a ‘systems model’ of an individual’s behavior within the organization

(Manning, 2013).

Sensemaking has seven individual and social properties: personal identity, retrospection,

enactment, social action, projection, cue extraction, and plausibility (Weick, 1995). These seven

elements interact to form the narrative individuals use to interpret events. In the organizational

context, understanding the individual narrative is critical for understanding “how organizational

structures, processes, and practices are constructed, and how these, in turn, shape social relations

and create institutions that ultimately influence people" (Clegg and Bailey, 2008).

Peter Senge’s ‘learning organization’ describes “a group of people working together

collectively to enhance their capabilities, to create results they really care about (Fulmer and

Keys, 1998).” The learning organization has five characteristics; similar to the properties of

sensemaking, these characteristics describe both individual and social factors that combine to

explain human behavior in organizations. The properties of the learning organization are systems

thinking, personal mastery, mental models, shared vision, and team learning.

In this dissertation, special attention is given to Engineering Systems Thinking (EST),

which describes the study of systems thinking in the systems engineering context. Today’s

systems engineers and LSCES designers often find themselves in highly complex, highly

20

ambiguous situations, technically and socially, and are responsible for making decisions about

both. Technical problems are becoming more challenging, as new materials, technologies, and

regulatory environments influence design capabilities, and solving these problems requires

engineers and scientists from different disciplines, organizations, etc., to quickly overcome

substantial cultural differences and become productive with one another (Cummings, 2002). The

modern systems engineering context describes both physical engineering systems as well as the

logical human organization of data, and systems engineering methodology has expanded over

time to include work-process models along with optimization methods, risk management tools,

etc. Contemporary descriptions categorize systems engineering as both a technical process and a

management process, in which the goal of the management process is to organize technical

efforts throughout the life cycle (Oliver et al., 1997).

While this divide-and-conquer approach to systems engineering was once sufficient, a

holistic view of LSCES and their operational environments has become increasingly important to

engineering decision-making (McDermott and Freeman, 2016). Systems engineers in the 21st

century must understand technical models of work flow, element couplings, interactions,

uncertainty, risk, etc., but must also appreciate the social context in which these models are

created, interpreted, and acted upon. To do so requires a unique cognitive and social skill set—

engineering systems thinking—that has attracted the attention of systems engineering researchers

in more recent years. Systems thinking has been described as “what makes systems engineering

different from other kinds of engineering” and as “the underpinning skill required to do systems

engineering (Beasley & Partridge, 2011).”

Research efforts in recent decades have explored this activity, attempting to elucidate

both individual traits, skills, and attitudes as well team properties that contribute to the “capacity

21

for engineering systems thinking (CEST) (Frank, 2000; Frank, 2006; Frank, 2012; Williams and

Derro, 2008; Brooks et al., 2011).” Frank’s studies in engineering systems thinking first

appeared in 2000. A substantial body of research followed (Frank, 2000; Frank and Waks, 2001;

Frank, 2002; Frank, 2006; Frank, 2007; Frank and Kordova, 2009; Frank, Sadeh, and Ashkenasi,

2011; Frank, 2012; Kordova, Frank, and Nissel, 2018), focusing on engineering systems thinking

as it exists in both education and practice. This work is important in its early advocacy of

updating the engineering curricula to incorporate systems thinking skills as part of standard

engineering education.

In the professional context, the work by Frank, Sadeh, and Ashkenasi (2011)

demonstrates a correlation between CEST and project success. This work is valuable for

conveying the potential for the improvement of engineering practice that could result from a

rigorous study of systems thinking. However, this work is yet to be substantially tested and

validated, and the traits that Frank identifies as the “capacity” for or “cognitive characteristics

of” engineering systems thinking are rather loosely defined. For example, in describing the

cognitive characteristics of engineering systems thinkers in his 2012 paper, each one of the ten

characteristics Frank identifies begins with “understanding.” What constitutes this understanding

and how engineering systems thinkers come to understand things in this way remain open

questions.

Frank describes engineering systems thinking in terms of behaviors that result from

systems thinking or demonstrations of the act of engineering systems thinking, rather than the

underlying psychological processes required for doing engineering systems thinking. Frank’s

work is beneficial in recognizing the behaviors or tendencies of systems thinking that can be

22

useful in practice, but there is still much work to define ways to develop the skills required to

become an engineering systems thinker.

Frank and Waks (2001) refer to the capacity for engineering systems thinking as a

distinct personality trait. Psychology, however, defines personality traits and thinking abilities as

distinctly different aspects. Personality traits are relatively stable over time, and are not expected

to change as a function of experience (Doob, 1947). Cognitive skills and strategies like systems

thinking can be developed through education, training, and experience. By describing the

capacity for systems thinking as an innate ability or personality characteristic, the potential for

methodically teaching systems thinking is lost, reducing a rather sophisticated concept to a

simple interaction of personality and experience. The psychological distinction is important if the

study and advancement of systems thinking is to progress in beneficial ways.

Work by Rhodes and co-workers Lamb, Nightingale, and Davidz is also important for

opening up the discussion about systems thinking in engineering, but may be subject to a similar

critique. In their 2008 paper, Davidz and Nightingale suggest that enabling systems thinking is a

critical step in advancing the development of senior systems engineers. They recognize at the

same time that fundamental questions still remain about how systems thinking develops in

engineers (Davidz and Nightingale, 2008). The authors attempted to answer these questions

through field studies and interviews with 205 engineers across 10 host companies in the US

aerospace sector (Davidz, Nightingale, and Rhodes, 2004). Engineers of various levels of

expertise and experience and with varying levels of proficiency in systems thinking were asked

how they define systems thinking, and were also given a definition of systems thinking and were

then asked to comment on what aspects of the definition they agreed and disagreed with. This

approach resulted in divergent definitions of systems thinking, which did not help in developing

23

a single, unified framework from which to advance its study. The authors organized their

findings into five broad foundational elements that explain what perspectives and behaviors

constitute systems thinking, but do not address the underlying commonalities or constructs.

Only one element, deemed the “modal” element, describes how an individual performs systems

thinking, but the authors described this “how” in the context of tools, methods, models, and

simulations and do not address the actual cognitive processes required to do systems thinking.

Despite this shortcoming, the work identifies some important enablers to the development of

systems thinking, such as experiential learning, education, interpersonal interactions, and training

in a supportive environment. Other research by Rhodes, Lamb, and Nightingale (2008) also

describes methods for studying systems thinking empirically. As in the work by Davidz and

Nightingale (2008), the authors seek to uncover the enablers, barriers, and precursors to

engineering systems thinking. The authors recognize that both an in-depth understanding of

engineering practice coupled with an orientation in social science is necessary to properly

capture the essence of engineering systems thinking.

In other work, Lamb, Nightingale, and Rhodes (2008) offered a different explanation for

engineering systems thinking by suggesting that it is perhaps not something that can be evaluated

at the individual level at all. Instead, this paper and others by the same group (Lamb and Rhodes,

2008; Lamb and Rhodes, 2010) suggest that systems thinking may be better understood as “an

emergent behavior of teams resulting from the interactions of the team members… utilizing a

variety of thinking styles, design processes, tools, and languages to consider system attributes,

interrelationships, context, and dynamics towards executing systems design.”

While social context is certainly a relevant and important factor in systems thinking, one

can argue against describing systems thinking as an emergent behavior of teams (Greene and

24

Papalambros, 2016). The study by Davidz and Nightingale (2008) relied on testimony from

“proven stellar systems thinkers.” If systems thinking were simply an emergent property of

teams, these individuals could not exist independent of the teams in which they work. Clearly,

certain individuals have a more refined systems thinking skill set than others, and understanding

why and how it is that this occurs is important. A cognitive psychological approach at the

individual level of analysis is one first step in this direction (Cagan, 2007; Greene and

Papalambros, 2016). This dissertation represents such an approach. Offering additional support

for this strategy, Davidz and Nightingale recognized the importance of addressing systems

thinking at the level of the individual (2008). They argue that understanding how systems

thinking develops in an individual is important for subsequently understanding how systems

thinking develops in a team. If systems thinking is to be described in terms of emergence, it is

more appropriate to summarize systems thinking as an emergent feature of a highly refined set of

individual cognitive processes or attitudes rather than an emergent feature of teams (Greene and

Papalambros, 2016).

Contemporary research in EST explores some human-centered assumptions, concepts,

values, and practices (Hutchison, Henry, and Pyster, 2016). Successful engineering systems

thinkers are consistently recognized as being good leaders and communicators and as naturally

"curious" or "innovative." Studies suggest that systems thinkers can see and define boundaries;

understand system synergy; and balance reductionist and holistic viewpoints (Frank, 2012). They

think creatively, overcome fixation, and tolerate ambiguity. ESTs ask "good questions"; can

understand new systems and concepts quickly; can consider non-engineering factors that

influence system performance; and understand analogies and parallelism between systems

(Davidz and Nightingale, 2008; Davidz et al., 2008; Frank, 2012; Madni, 2015; McGowan,

25

2014; Rhodes et al., 2008; Williams and Derro, 2008). This prior research provides valuable

contributions and strong evidence for the importance of studying systems thinking in

engineering, but contributions from extant knowledge in psychology and cognitive science are

not included. This dissertation seeks to address this gap.

2.2.4 Design Thinking

Design thinking is a recent popular topic in academic research, pedagogy and education,

engineering, and business. Rapid technological development throughout the twentieth century

generated a need for formal academic study of "the science of design (Simon, 1969)." Two

important periods in the modern history of design science are identified by Cross (1982). The

first, the "design products movement" of the 1920s, sought to "produce works of art and design

based on objectivity and rationality;" that is, on the values of science. The second, the "design

methods movement" of the 1960s, sought to establish design processes—in addition to the

products of design—based on similar scientific principles. Despite some backlash against design

methodology in the 1970s, the tradition continued to flourish in engineering and industrial

design, and several prominent academic journals for design research, theory, and methodology

emerged during the 1980s and 1990s.

Design methodology is defined by Cross (1982) as "the study of the principles, practices,

and procedures of design" and "includes the study of how designers work and think." Over the

past several decades, engineering researchers have successfully leveraged cognitive and social

science approaches to study how designers think through engineering design problems, exploring

a breadth of topics including creativity, ideation in early conceptual design, the role of analogies

in creative problem solving, differences between novices and experts, and strategies for

overcoming fixation and mental blocking. Verbal protocol analysis, cognitive ethnography,

26

controlled laboratory experiments, and other formal methods from cognitive science have been

rigorously applied to the study of designer thinking in engineering (Dinar et al., 2015; Shah et

al., 2012). Results of these studies and others suggest that design thinking approaches use

solution-based methods to explore human-centered values throughout the engineering design

process. This finding is reflected in many applications of design thinking: prototyping, a

solution-based method, is often cited as a useful way to encourage inspiration, ideation, and

organizational learning, all human-centered values (Brown, 2009; McGowan et al., 2017).

Design thinking emerged as a formalism of successful practice in identifying the right

design problem, generating creative solutions, and making design decisions through rapid

prototyping and user testing (Papalambros, 2018). Design thinking, as a formalism of user-

centered design practice, emerged in stark contrast to engineering design obsessed with

functionality. Design thinking frameworks typically do not include analytical-quantitative

methods, dealing instead with sociocognitive processes such as creativity and empathy. Today’s

industrial designers and product designers are relatively free from the complexity of engineering,

including analysis, and are clear to focus instead on user needs, desires, and experience with

designed artifacts (Papalambros, 2018).

Many applications of the design thinking framework exist in the literature and in practice.

Examples include Herbert Simon's design thinking process (1969), which suggests seven stages

of design thinking for product design, including defining the problem, researching, ideating,

prototyping, choosing a solution, implementing the solution, and learning. Plattner, Meinel, and

Leifer (2011) propose a five-step version of the design thinking process that includes redefining

the problem, need finding and benchmarking, ideating, building, and testing. International design

and consulting firm IDEO applies a four-phase process that includes gathering inspiration,

27

generating ideas, making ideas tangible, and sharing your story (Brown, 2008). Graphic

representations of these processes are included in Figure 2.1.

Figure 2.1 Design Thinking Process Models from IDEO, Stanford d-school, and Google Design

Sprint

While each interpretation differs slightly from the others, important foundational values

of design thinking persist. First, while design thinking frameworks emphasize the importance of

problem definition, the process is solution-driven, and most design thinking methods include

prototyping and iteration phases for generating solutions that meet customer needs. These

solutions are human-centered products or services, developed through designers' personal

experiences, empathy, and engagement with stakeholders. The design thinking process itself is

also human-centered, offering methods for inspiration, ideation, and learning to designers

(Brown, 2008). Design thinking has been described as a "high order intellectual activity" that

"requires practice and is learnable" (Plattner et al., 2011).

2.3 Developing a Codebook for Interview Analysis

In this section, themes from the literature are integrated and described in depth to develop

a “systems design thinking codebook.” This codebook will be used for informing interview

analysis, as well as quantitative analyses in later chapters. The codebook will be used to

understand individuals’ attitudes about systems engineering, systems thinking, and design

thinking; how they are done/demonstrated in practice; how they are related; and the perceived

effect each has on projects and outcomes. This codebook is a contribution to the interdisciplinary

28

study of these concepts, with the goal of addressing the gap in the literature comparing these

concepts.

2.3.1 Systems Engineering Codes

Systems engineering codes reflect organizational and process elements, such as planning,

documenting, and managing. Systems engineering often begins with system requirements, and

describes all processes related to verifying and validating that those requirements have been met

on time and within budget. This process is large-scale and complex, requiring the coordination of

many individuals over long periods of time, and extensive documentation to retain and

communicate information effectively. Systems engineering codes reflect an organized, detail-

oriented process for accomplishing these tasks.

Documentation

A key function of systems engineers is ensuring proper documentation throughout

systems design. This includes documenting requirements, verification and validation plans,

design changes, etc., in accordance with ISO 9000 requirements (Recker, 2002). This can be

very challenging, as there are many engineering groups working together on systems projects,

and each group manages and maintains their documentation in their own way. Systems engineers

are responsible for integrating these documents and delivering them on time.

Systems engineers also exhibit informal documentation behaviors, such as personal note-

taking, recording meeting minutes, etc. This reflects attention to detail and enables information

retention over time and space.

29

Planning and scheduling

Systems engineers are responsible for delivering products on time and within budget

constraints. This includes long-term as well as short-term project planning, and scheduling

meetings between individuals. This can be done formally and informally.

Methodical/process-driven

Systems engineering is methodical and follows the same general process for every

project. This is captured in the “systems vee,” although many organizations have a slightly

modified/more personalized version that reflects the unique values of the organization. The

INCOSE systems engineering process model and NASA “systems engineering engine” are

pictured in Figure 2.2 below.

Figure 2.2 Systems engineering process models. On the left is the INCOSE systems engineering

“vee” and the NASA “systems engineering engine” is on the right.

Requirements

Requirements definition, verification, and validation are key elements in systems

engineering. Much of the systems engineering process is devoted to ensuring that requirements

are clearly defined and easily interpreted by designers and customers alike. Systems engineering

30

processes typically begin at requirements definition, and requirements then serve as a contract

between engineers and customers through the remainder of the design process.

Management

Differences exist between systems engineers and project managers, but technical

management is at the heart of the systems engineering process (see Figure 2.2). Technical

management includes ensuring that all subsystems are developed in a cohesive way that allows

for easy integration and alignment.

2.3.2 Systems Thinking Codes

Engineering systems thinking shares a foundation with systems science; thus,

assumptions, concepts, values, and practices bear some resemblance to those of general systems

theory, cybernetics, and systems dynamics. The systems thinking framework states that a system

is composed of parts, the system is greater than the sum of its parts, and all parts of the system

are interrelated. Systems receive inputs from the environment, execute processes that transform

these inputs into outputs, and send these outputs back into the environment in feedback loops.

Systems are dynamic and complex, interactions may be difficult to identify or quantify, and

emergence is common. Systems thinking themes reflect an appreciation for system complexity,

with specific emphasis on understanding interactions between system elements, and integrating

and aligning these elements. Flexibility, adaptability, and a tolerance for ambiguity reflect the

systems thinking approach for dealing with complexity.

Contemporary research in engineering systems thinking seeks to make the approach more

human-centered. Successful engineering systems thinkers are consistently recognized as being

good leaders and communicators and as naturally "curious" or "innovative." They think

creatively, overcome fixation, and tolerate ambiguity. ESTs ask "good questions"; can

31

understand new systems and concepts quickly; can consider non-engineering factors that

influence system performance; and understand analogies and parallelism between systems

(Frank, 2012; Williams and Derro, 2008; Davidz and Nightingale, 2008; Davidz et al., 2008;

Rhodes et al., 2008; McGowan, 2014). Systems thinking codes also capture these human-

centered elements.

Big picture view

Systems thinking codes reflect the “big picture view,” but also emphasize an appreciation

for details about individual system elements and interactions between them. Studies suggest that

systems thinkers can see and define boundaries; understand system synergy; and balance

reductionist and holistic viewpoints (Frank, 2006).

Interactions

Systems thinking involves the constant search for interactions; technical, social, and

organizational. This is due in part to the systems thinker’s natural curiosity about how technical

systems work and how individual elements are related. Systems thinkers are also interested in

social interactions and their effects on systems design, and leverage social relationships to

deepen their understanding of the technical system. They do this within organizational

constraints, following organizational cultural norms, processes, and practices.

Integration/alignment

This code emphasizes the importance of mitigating interactions and preventing

emergence. Systems design activities and their resulting artifacts must be aligned and integrated

throughout the systems engineering process. Each element in a system must be designed to

function holistically. Individual optimization of subsystems is not enough. Socially, this involves

32

“getting everyone on the same page,” as integrating technical subsystems requires integrating the

social systems that work on them.

Flexibility/adaptability

The systems thinking framework addresses complexity through flexibility and tolerance

of ambiguity. This differs from the systems engineering framework, where the approach involves

reducing ambiguity and risk through analysis and management.

2.3.3 Design Thinking Codes

Design thinking codes are human-centered and related to understanding user and other

stakeholder needs through interaction and engagement. Shared experience is a common theme in

design thinking, for its ability to generate empathy and insight into uncovered needs. Design

thinking also includes prototyping as a form of shared experience and communication, as well as

an opportunity to test ideas and move from conceptual design to embodiment.

Empathetic [human-centered]

Various design thinking processes cited in literature start with identifying and immersing

oneself in the end user’s position. This phase has also been described as “immersion,”

“awareness,” and “inspiration,” etc. (Fleury, Stabile, and Carvalho, 2016). The core function of

the designer in this phase is to build empathy for their end users. Empathy in this context refers

to imagining the world from the perspective of multiple stakeholders and prioritizing the latent

needs of the people (Brown, 2008). Plattner et al. (2011) claims that empathy in design thinking

is the process of ‘needfinding,’ or discovering both implicit and explicit needs before working on

a design problem. Empathy forms a core component of the design process regardless of the steps

taken to achieve the solution.

33

Intuitive/experience-driven

This code shares some overlaps with empathy (e.g., intuition, awareness, immersion).

Immersive experiences like co-design, prototyping/user testing (frequent customer interaction &

feedback), etc., are useful for developing empathy and design intuition (Tonetto & Tamminen,

2015). An important function of the designer is to intuitively shape the problem at hand by

identifying the view of participants and the underlying issues that concern them (Buchanan,

1992). Gasparini claims emotional empathy as being an instinctive, affective, shared and

mirrored experience where one feels what other people experience (Gasparini, 2015).

Ambiguous problems & solutions

Design problems are now widely recognized as ill-defined or “wicked” (Cross, 1982;

Rittel and Webber, 1973, 1974; Kuhn, 1962; Buchanan, 1992). Cross examines scientific

problems to be ‘tame’ and analogous to puzzles which can be solved by applying well known

rules to given data while design problems are ill-structured with ambiguous requirements.

Buchanan (1992) supports this claim by stating that the underlying reason for the ‘wickedness’

of design problem lies in the fact that design has no subject matter of its own other than what the

designer conceives it to be. In light of this, design thinking requires the designer to embrace and

preserve ambiguity (Plattner et al., 2011). Plattner claims that when the constraints are explicitly

defined, it provides minimal chance for the discovery of latent needs leading to innovative

solutions. Modifying the nature of the problem to find a solution is a challenging yet imperative

aspect of the designer’s role and this in turn states the ambiguous nature of problems in design

thinking (Jones, 1992).

34

Innovative/creative

Design thinking is widely claimed to be a driver of innovation (Brown, 2008; Cross,

2001; Plattner, Meinel, and Leifer, 2012). Plattner et al. (2012) claim that creative design takes

the center stage of all design thinking activities. Anderson et al. (2014) defines creative design as

something that is novel and useful, or appropriate and adaptive. Shah (2012) claims that the

discipline of design requires abductive reasoning, divergent thinking, and creative thinking, and

that this differs from the science-based regimen that promotes convergent thinking and deductive

reasoning through closed-end problem solving. Shah also claims that while these skills are core

to engineers, they may be insufficient to the field of design. Buchanan (1992) claims that design

thinking requires designers to think beyond their personal preferences and visualize novel

possibilities by conceptual placements.

2.4 From Frameworks to Attitudes: Interviews with Systems Engineers

This section describes an informal pilot study, in which semi-structured interviews are

conducted with practicing systems engineers. The goal is to further explore the relationship

between “systems engineering” and “systems thinking,” and understand if/how design thinking

themes are discussed, based on our understanding of the literature. The interview sample is

small and not well-balanced, and findings are neither generalizable nor transferable from a

research methodology perspective, nor are they meant to be. The interviews added depth to

literature findings, provided useful insights and clarifications about systems engineering, systems

thinking, and design thinking, and generated “attitude statements” for use in Likert scales to test

generalizability and transferability quantitatively, as described in the next chapter.

35

2.4.1 Method

The interview setting was a government laboratory for large-scale systems design. The

subject population for the study included ten experienced adult engineers (more than ten years

work experience) working in organizations that are responsible for the design and management

of large-scale, complex engineered systems. The subject population included senior systems

engineers, chief engineers, project managers, and related roles. Participants were identified as

exceptional systems thinkers and recruited by a technical leader within the organization, who

asked them to participate in “an interview about systems thinking.”

The cognitive work analysis framework was used to structure the interview (Vicente,

1999). Interview questions roughly followed the critical decision method (Klein and Armstrong,

2005; Klein, Calderwood, and MacGregor, 1989). This method uses cognitive probes to elicit

information regarding naturalistic expert decision-making. In this work, cognitive probes were

used to elicit attitudes, beliefs, and approaches for systems engineering and systems thinking.

Interviews also included questions about communication preferences, based on theories that

suggest social coordination is closely related to technical coordination (Cataldo, Herbsleb, and

Carley, 2008; Colfer and Baldwin, 2016; Conway, 1968; de Souza et al., 2007). Interview

questions can be found in the Appendix.

The codebook from Section 2.3 is imported into NVivo 12 and used to analyze interview

data. Additional codes and themes are also generated from the data. Several word frequency

queries and matrix coding methods are used to understand significant themes and relationships

between these themes.

2.4.2 Analysis

Thematic analysis is used to identify patterned meaning across a dataset (Braun and

Clarke, 2006). In the first step, a deductive approach is used, where coding and theme

36

development are directed by existing concepts and ideas. The codebook described in Section 2.3

is used for deductive coding. Inductive coding was also conducted within the results of the

deductive coding to organize emergent concepts. The goal of interviews is to identify attitudes

that reflect systems engineering, systems thinking, and design thinking in practice, and

understand the relationships between them.

The first analysis conducted was a general count of number of interviews including each

theme and number of references to each theme. Five themes were coded in all ten interviews.

These are in Table 2.1 below. Communication was first, referenced in all ten interviews with a

total of 96 references. Interactions and empathy were next, both referenced in all ten interviews,

56 and 42 times respectively.

Table 2.1 Five themes were coded in all ten interviews. These five themes, with total number of

references in parentheses, are: communication (96), interactions (56), empathy (42), complexity

(30), and learning and information (29).

Theme Interviews References

Communication 10 96

Interactions 10 56

Empathy 10 42

Complexity 10 30

Learning and

information 10 29

37

Table 2.2 represents design thinking themes, and the number of interviews including each

theme and number of references to each theme:

Table 2.2 Design thinking themes, number of interviews including each theme, and total number

of references to each theme. Design thinking themes had highest number of total references

(297).

Theme Interviews References

Communication 10 96

Empathy 10 42

Human-centered 10 40

People & personalities 8 45

Experience 8 39

Awareness & intuition 8 35

Total 297

Design thinking themes had highest number of references overall. Communication, a sub-theme

of “human-centeredness,” and empathy together had 138 references.

Systems thinking themes in Table 2.3 had second highest number of overall references.

Interactions and complexity had 86 combined.

Table 2.3. Systems thinking themes, number of interviews including each theme, and total

number of references to each theme. Systems thinking themes had the second highest number of

total references (238).

Theme Interviews References

Interactions 10 56

Complexity 10 30

Integrate & align 9 56

Big picture 9 53

Flexibility/adaptability 8 26

Ambiguity/uncertainty 8 17

Total 238

Learning and information, an inductive code referenced in all ten interviews, refers to the

approach for gathering and managing information about the technical system. Learning and

38

information includes themes from systems thinking and systems engineering, such as “asking

questions” (about relationships, interactions) and “gathering data”. To do this effectively

requires design thinking skills. Empathy is helpful but not always required; “awareness” (of tone,

reception, etc.), emotional intelligence, and approaching conversations as “qualitative data” are

all important precursors to learning.

Systems engineering themes are in Table 2.4 below:

Table 2.4 Systems engineering themes, number of interviews including each theme, and total

number of references to each theme. Systems engineering themes had the fewest number of total

references (190).

Theme Interviews References

General SE 8 31

Management 8 23

Document 7 42

Requirements 7 25

Risk 7 17

Planning/scheduling 6 29

Process 6 23

Total 190

“Document” had the highest number of references overall. “Planning/scheduling” was the

second most specifically referenced systems engineering theme. “Requirements” and “process”

were third and fourth most referenced, respectively. Systems engineering themes and attitudes

were mentioned fewest overall.

2.4.3 Findings

Systems engineering reflects necessary occupational processes; systems thinking reflects

the underlying cognitive processes that support them. One participant describes the relationship

between systems engineering and systems thinking attitudes in the following way:

39

“People think systems engineering, they [think] this guy does requirements,

does configuration management, does this paperwork. But a systems engineer

is really—to be one—you’ve got to have command of how the system interacts

with each other, what the sensitivities are, and be able to carve that system up

and manage those interactions.”

Another interview offers a similar perspective on this, suggesting that the focus on policy

and procedures is more of an “academic perspective” to systems engineering, rather than a good

representation of systems engineering practice, which relies more heavily on systems thinking.

However, these processes make up the fundamental responsibilities of systems engineers, which

influences their assumptions and values (Elsbach, Barr, and Hargadon, 2005). We therefore

define “systems engineering attitudes” as attitudes about requirements, scheduling, planning, and

documentation. This overlaps with technical and organizational systems thinking (e.g.,

scheduling, planning, and documentation are organizational systems thinking processes).

Systems thinking is organizational, in that it focuses on identifying and understanding

relationships and interactions, and systems engineers see relationships between humans as

equally important as relationships between technical system elements. Systems thinking is used

for identifying what the problems are, where they are, and who can help. Solutions mostly

involve getting people in a room to discuss, which has human-centered themes in common with

the design thinking framework.

Design thinking attitudes reflect empathy and understanding, the act of hearing and

listening, and other human-centered beliefs and approaches. There appears to be some overlap

with the social element of systems thinking. However, in social systems thinking as described in

interviews, human-centeredness is not limited to understanding user needs as it does in design

40

thinking/product design, but also includes understanding the personalities, needs, concerns, etc.

of other engineers and designers within the organization in order to get design done.

Interviews also offered insights into the relationship between engineering and design

thinking. Data captured differences in interpretation and preference for design thinking and

systems engineering processes:

“Most people think the design, analysis, test and build is the cool stuff, so,

let's just get to the cool stuff right away. So, I think having some rigor in the

system from a systems engineering perspective forces us to make more

deliberate design development decisions that you might not make otherwise.”

Interview findings are summarized in the attitude model in Fig. 2.3 below.

Figure 2.3 Systems Design Thinking Attitude Model. Systems engineering attitudes describe

feelings about requirements, analysis, documentation, scheduling, and management processes.

Design thinking attitudes are human-centered, describing feelings about people/personalities,

41

empathy, and communication. Systems Design Thinking uses the holistic systems thinking

approach (appreciation for ambiguity, complexity, interactions, and integration) for

understanding both technical systems and social systems.

Systems thinking is positioned between systems engineering and design thinking, as

systems design thinkers will apply a “systems philosophy” to technical as well as human-

centered systems, and understand the relationships between them.

2.5 Summary

This chapter presented a review of relevant literature on systems engineering, systems

thinking, and design thinking. The outcome of this review was a set of codes that reflect core

assumptions, concepts, values, and practices of systems engineering, systems thinking, and

design thinking frameworks. These codes were compiled into a codebook for systems design

thinking, and used to analyze data from semi-structured interviews with experienced systems

engineers.

Consistent with the literature, interview findings suggest that engineering systems

thinking is an integral skill for systems engineering. The systems engineering perspective is

based on systems thinking (INCOSE, 2015). Systems thinking is necessary for identifying and

understanding interactions between technical system elements, and for identifying and

coordinating the corresponding social units within the engineering organization. This

multidimensional concept is formalized in interview findings, where three “types” of systems

thinking—technical, social, and organizational—are identified and described.

While design thinking is not mentioned explicitly in the interviews, systems engineers

make references to concepts, values, and practices from design thinking. These include human-

centered practices such as active listening and communication, empathy, and shared experience.

42

“Communication,” a subset of the design thinking code “human-centered,” was the most

frequently referenced code in the qualitative analysis.

Systems engineering, systems thinking, and design thinking frameworks overlap in the

context of complex systems design. All of these frameworks are useful at different stages and for

different activities within the systems engineering process. Not surprisingly, systems engineers’

attitudes reflect different assumptions, concepts, values, and practices from these frameworks. In

the next chapter, these attitudes are explored quantitatively. Factor analysis is used to determine

whether these attitudes can be clearly differentiated, according to systems engineering, systems

thinking, and design thinking frameworks.

43

CHAPTER III

Modeling Systems Design Thinking Attitudes

3.1 Introduction

Findings from interviews suggest different ways to define and relate systems engineering,

systems thinking, and design thinking frameworks. In this chapter, theory and methods from

psychometrics are used to make quantitative comparisons between systems engineering, systems

thinking, and design thinking frameworks. Psychometrics is the field of study concerned with the

theory and technique of objective psychological measurement (Furr and Bacharach, 2013). This

includes the assessment of skills, knowledge, abilities, attitudes, personality traits, and

educational achievement. Psychometrics includes the construction and validation of assessment

instruments such as questionnaires, scales, and tests.

Some work in recent years has explored psychometric approaches for modeling systems

thinking and design thinking (Castelle and Jaradat, 2016; Chesson, 2017; Davis and Stroink,

2016; Davis et al. 2018; Dosi, Rosati, and Vignoli, 2018; Jaradat, 2014; Thibodeau, Frantz, and

Stroink, 2016). Few studies explore the psychology of systems engineering. None of the

identified studies have attempted to integrate these frameworks into a single measure. In this

chapter, psychometrics is used to identify the key assumptions, concepts, values, and practices of

systems engineering, systems thinking, and design thinking, and integrate them through the

development of the Systems Design Thinking Scale. A series of three iterative studies is

described. In Study 1, factor analysis is used to test a model of systems design thinking

44

consisting of technical systems thinking attitudes, social systems thinking attitudes, and

organizational systems thinking attitudes. In Study 2, a two-factor model of systems design

thinking consisting of systems thinking and design thinking attitudes is tested. In Study 3, we test

a two-factor model of systems engineering and design thinking attitudes.

Structural Equation Modeling (SEM) is used to fit systems engineering, systems thinking,

and design thinking frameworks to data (Kline, 2015). SEM is a popular technique in the social

sciences, where unobservable constructs such as intelligence or self-esteem are more commonly

studied than directly measurable variables such as volume or mass. SEM can be described as a

two-step hypotheses testing technique. First, social scientists develop hypotheses about a

construct (e.g., intelligence), and write measurement instruments (e.g., an IQ test) with questions

designed to measure intelligence according to their hypotheses. Then, statistical methods are

used to assess the validity of the hypotheses, using data gathered from people who took the

intelligence test. In this example, “intelligence” is the latent variable, and test questions, referred

to as ‘items,’ are the observed variables.

“Intelligence” in this example can be replaced with “systems design thinking.” Systems

design thinking is a latent construct that is believed to exist, but it is not directly measurable in

the same way that volume and mass are directly measurable. Thus, in order to “measure”

systems design thinking, the network of constructs that comprise systems design thinking must

first be decomposed into items that are directly measurable. In this case, items reflect

assumptions, concepts, values, and practices in systems engineering, systems thinking, and

design thinking, and the degree to which subjects agree or disagree is measured. Several network

models are developed based on findings from the qualitative research described in Chapter II.

45

These models are tested using exploratory and confirmatory factor analysis as described in the

following sections.

3.2 Study 1: Technical, Organizational, and Social Systems Thinking

3.2.1 Scale Development

Interviews with systems engineers suggested three “types” of systems thinking, depicted

in Figure 3.1. These are technical, organizational, and social systems thinking. Items

representing technical, social, and organizational systems thinking are organized into three

respective factors. This three-factor model of systems design thinking is tested first.

Figure 3.1 A graphical representation of the three types of systems thinking—technical, social,

and organizational. This categorization was suggested in interviews with professional systems

engineers. These individuals were recruited to participate in the study based on their designation

as “exceptional systems thinkers” by a technical leader within the organization.

Tables 3.1, 3.2, and 3.3 include 57 attitude statements that were included in the pilot test.

These statements are grouped into three factors according to systems thinking types. It is

important to note that many of the items are intended to be “reverse worded”, i.e., meant to

represent the ‘opposite’ of a systems design thinking attitude. This is indicated in the table with a

(-) following each reverse-worded statement. There are issues with this choice, which will be

discussed in the findings, Section 3.2.3.

46

Table 3.1 Technical systems thinking attitude items tested in Study 1. Items marked with a (-) are

reverse-worded, i.e., meant to represent the ‘opposite’ of a systems design thinking attitude.

T1. I would prefer to design & manufacture a

single part rather than analyze interactions

between two parts of a system. (-)

T2. I tend to focus on the nuances of a

problem, rather than the big picture. (-)

T3. For best system performance, subsystems

should be as independent as possible. (-)

T4. I test several ways to solve a problem

before choosing the best one.

T5. Modifications and adjustments made

after a system is deployed indicate that the

design was inadequate. (-)

T6. Design decisions should be made based

on the problem a system was designed to

address, rather than the system that currently

exists.

T7. A system will perform optimally if each

of its subsystems is designed optimally. (-)

T8. I prefer to work on problems with

objective solutions. (-)

T9. Comprehensive understanding of a

system can be achieved by analyzing each

individual subsystem. (-)

T10. I like to receive a detailed set of

requirements before beginning a project.

T11. Once desired performance is achieved,

a system should be left alone. (-)

T12. When designing a system, plans should

be tentative and expected to change.

T13. Once successful, a technical solution

should result in similar success in other

applications.

T14. I prefer to work on technical problems

rather than non-technical problems (e.g., fix

a part vs. negotiate a contract). (-)

T15. It is better to try proven solutions before

pursuing new solutions to a problem.

T16. It is important that I understand how my

work contributes to the larger system or

mission.

T17. Prototyping speeds up the process of

innovation.

T18. I break problems into tasks/steps before

beginning work.

T19. Problems at technical interfaces take

longer to resolve than problems in other

areas.

T20. Engineering and design are different. T21. I can recall at least one “a-ha moment”

I’ve had at work.

T22. Design decisions should be made based

on data.

47

Table 3.2 Organizational systems thinking attitude items tested in Study 1

O1. Engineering organizations should ensure

that their divisions are integrated, even if that

means limiting each division's freedom to

make decisions.

O2. Working in geographically dispersed

teams is harder than working in co-located

teams.

O3. Planning is wasteful in uncertain

situations.

O4. My organization's values are important

to me.

O5. Organizations should follow the same

design process for every project.

O6. I believe that control of my work

environment is possible.

O7. Task-focused individuals are just as

valuable to organizations as innovators are.

O8. I ask myself if what I'm learning is

related to what I already know.

O9. Engineering organizations should

distribute decision authority equally between

discipline engineers and project

management.

O10. Long-term planning and short-term

planning are equally important.

O11. I delay making plans rather than make

plans I know will change later.

O12. It is important that my job offers

flexible scheduling.

O13. Organizations should support

interdisciplinary collaboration.

Table 3.3 Social systems thinking attitude items tested in Study 1

S1. It is important to consider stakeholders’

values and emotions in addition to technical

system requirements.

S2. When I have a question about my work, I

try to figure it out by myself before asking

anyone for help.

S3. I enjoy using group creativity techniques

such as mind mapping/brainstorming.

S4. I seek out others' opinions when deciding

how to approach a problem.

S5. I prefer to have my own office rather

than shared workspace.

S6. I enjoy working with people outside my

discipline.

S7. I like to know what my colleagues are

doing, even if it doesn't relate directly to my

work.

S8. I prefer to meet as needed rather than

attend regularly scheduled meetings.

S9. Collaborating and working independently

are equally important.

S10. I use storytelling and/or analogies to

describe my work to others.

S11. It is important that I know my

colleagues personally as well as

professionally.

S12. I am comfortable applying my technical

knowledge and skills in unfamiliar situations.

S13. Engineering is a creative process. S14. I enjoy taking on leadership roles. S15. My colleagues know about my life

outside of work.

S16. I enjoy sharing my ideas with others. S17. I feel most creative around other people. S18. I prefer to work with colleagues that

have more experience than I do.

S19. I wish I had more time to learn about

my colleagues' work.

S20. It is difficult to establish a common

vocabulary with colleagues from other

disciplines.

S21. Team-building exercises are valuable.

S22. I prefer to work with people with whom

I've worked successfully before.

48

3.2.2 Pilot Test, Factor Analysis, and Results for Study 1

Exploratory factor analysis (EFA) was used to study all 57 items in Tables 3.1, 3.2, and

3.3 in Section 3.2.1 (Mulaik, 2009). These items were arranged in a 5-point Likert scale and

evaluated by a sample of 20 professional engineers and engineering researchers from industry,

academia, and government. It is important to note that sample size of 20 is very small for this

type of study. The recommended sample size for conducting exploratory factor analysis is at

least 100 subjects (Kline, 2015). Minimum ratios of sample size to the number of variables have

also been proposed (Pearson and Mundform, 2010). However, Study 1 was intended to serve

only as a pilot test for identifying and removing items that function particularly poorly, and to

identify preliminary patterns in the data for further analysis.

The EFA is specified to explore up to three factors in Study 1. Maximum likelihood

estimation and varimax (orthogonal) rotation were used in the analysis, as it is expected that an

individual can have any combination of systems design thinking attitudes. Based on the

interview findings, in which systems thinking was partitioned into technical, social, and

organizational systems thinking, the EFA was expected to produce a corresponding 3-factor

solution. However, no interpretable results were observed for three-factor model.

Additional analysis was conducted on the 2-factor model, to interpret each factor and

refine the hypotheses for further study. Variables with loadings less than 0.300 were dropped

from the model due to poor fit, and the data was analyzed again. Variables included in this

second analysis can be seen with their varimax rotated loadings in Tables 3.4, 3.5, and 3.6.

49

Table 3.4 Post-hoc exploratory factor analysis results for technical systems thinking attitude

items in Study 1

Item ID F1 F2 Attitude Statement

T2 -0.028 0.670 I tend to focus on the nuances of a problem, rather than the big picture.

T3 0.344 -0.184 For best system performance, subsystems should be as independent as possible.

T4 0.149 -0.309 I test several ways to solve a problem before choosing the best one.

T5 -0.334 0.145 Modifications and adjustments made after a system is deployed indicate that the design

was inadequate.

T7 -0.148 -0.382 A system will perform optimally if each of its subsystems is designed optimally.

T8 -0.718 0.375 I prefer to work on problems with objective solutions.

T9 -0.578 -0.217 Comprehensive understanding of a system can be achieved by analyzing each

individual subsystem.

T11 0.314 -0.065 Once desired performance is achieved, a system should be left alone.

T13 0.623 0.153 Once successful, a technical solution should result in similar success in other

applications.

T14 -0.072 0.407 I prefer to work on technical problems rather than non-technical problems (e.g., fix a

part vs. negotiate a contract).

T17 -0.327 0.341 Prototyping speeds up the process of innovation.

T18 -0.351 0.339 I break problems into tasks/steps before beginning work.

T22 -0.304 0.301 Design decisions should be made based on data.

50

Table 3.5 Post-hoc exploratory factor analysis results for social systems thinking attitude items in

Study 1

Item ID F1 F2 Attitude Statement

S2 0.258 -0.674 When I have a question about my work, I try to figure it out by myself before asking

anyone for help.

S3 -0.869 0.039 I enjoy using group creativity techniques such as mind mapping/brainstorming.

S5 0.670 -0.093 I prefer to have my own office rather than shared workspace.

S7 -0.142 -0.764 I like to know what my colleagues are doing, even if it doesn't relate directly to my

work.

S8 0.721 -0.057 I prefer to meet as needed rather than attend regularly scheduled meetings.

S10 0.065 -0.863 I use storytelling and/or analogies to describe my work to others.

S13 -0.793 0.249 Engineering is a creative process.

S14 -0.645 0.196 I enjoy taking on leadership roles.

S15 0.433 0.013 My colleagues know about my life outside of work.

S17 -0.556 0.188 I feel most creative around other people.

S18 0.789 -0.219 I prefer to work with colleagues that have more experience than I do.

S20 -0.160 0.465 It is difficult to establish a common vocabulary with colleagues from other disciplines.

S21 -0.729 -0.336 Team-building exercises are valuable.

S22 0.850 0.019 I prefer to work with people with whom I've worked successfully before.

51

Table 3.6 Post-hoc exploratory factor analysis results for organizational systems thinking

attitude items in Study 1

Item ID F1 F2 Attitude Statement

O3 -0.359 0.101 Planning is wasteful in uncertain situations.

O4 -0.309 0.308 My organization's values are important to me.

O5 0.503 0.220 Organizations should follow the same design process for every project.

O7 -0.216 -0.756 Task-focused individuals are just as valuable to organizations as innovators are.

O8 0.387 0.180 I ask myself if what I'm learning is related to what I already know.

O10 0.078 -0.682 Long-term planning and short-term planning are equally important.

O12 0.055 -0.473 It is important that my job offers flexible scheduling.

O13 -0.119 -0.906 Organizations should support interdisciplinary collaboration.

Again, the 3-factor model did not produce interpretable results. The 2-factor model

demonstrated slightly better than the 1-factor model, although both models result in poor fit

according to conventional model fit indices (Hooper, Coughlan, and Mullen, 2008). The 2-factor

model is summarized in Table 3.7, and is analyzed and interpreted as follows.

Table 3.7 Summary of factor structure after post-hoc exploratory factor analysis (Study 1)

Factor Technical ST Social ST Organizational ST

1 3, 5, 8, 9, 11, 13, 17, 18 3, 5, 8, 13, 14, 15, 17, 18,

21, 22 3, 4, 5, 8

2 2, 7, 14, 17, 18 2, 7, 19, 20 4, 7, 10, 12, 13

Factor 1 seems to represent systems design thinking attitudes. Technical systems thinking

items in Factor 1 reflect a balance between reductionism and holism; support for iteration and

improvement after system deployment; and comfort with ambiguity and subjective problem-

solving. Social attitudes in Factor 1 reflect a similar balance. Collaboration and independent

work are both important, and individual preferences reflect the need for both shared time and

52

space, and personal time and space. Organizational systems thinking items are not readily

interpretable in this factor.

Factor 2 seems to represent general discipline engineering attitudes. These include things

like appreciation for nuance and detail, a preference for technical work (versus non-technical

work like negotiation), and a reductionist approach to design based on organizational

partitioning. Social attitudes in this factor are also less representative of systems design thinking,

reflecting introversion and difficulty establishing common vocabulary across disciplines despite

interest and effort. Organizational systems thinking items are also difficult to interpret in this

factor.

3.2.3 Discussion

EFA results for the three-factor model of technical, social, and organizational systems

thinking, as written, seem to reflect the style and quality of the attitude statements rather than the

structure of the systems design thinking framework. Almost half of the items in this analysis

were reverse-worded, and meant to represent the ‘opposite’ of a systems design thinking attitude.

But, these items didn’t represent the opposite of a systems design thinker, and some reverse

items were consistently endorsed. This is interpreted to mean that systems design thinkers are

cognitively flexible and context-sensitive, and alternate between a holistic and reductionist view

as necessary. This is consistent with literature findings about engineering systems thinking

(Brooks, Carroll, and Beard, 2011).

Study 1 also highlighted other important methodological issues. First, 57 attitude

statements are too many to include on a scale of this type. The average survey response time for

this survey was 15 minutes, more than double the target time of 7 minutes. Data indicates that

surveys longer than 9 minutes start to see substantial levels of respondent break-off (Qualtrics

53

Support, 2018). Second, the item display may have affected response time and user experience in

a meaningful way. Items were presented in 10 consecutive matrices, each containing 5-6 items.

Research has shown that response quality and completion rates both decline when questions use

the matrix format (Qualtrics Support, 2018). The high number of matrix rows contained in the

survey may have negatively impacted the data quality and quantity.

Study 1 yielded additional valuable theoretical insights. First, the study did not provide

any evidence for a model of systems thinking with technical, social, and organizational factors.

Technical and social items that reflect the systems design thinking framework grouped into the

first factor, and technical and social items that represent a more localized disciplinary perspective

grouped into the second factor. Organizational systems thinking items were not easily interpreted

within either factor. This could be because technical, social, and organizational systems thinking

function similarly, although the subject is different; i.e., “systems thinking” is a big-picture,

holistic framework for analyzing technical, social, and organizational systems alike. There may

not be any functional difference between how each type of system is conceptualized.

This could be because technical subsystems and the social subsystems that design and

deliver them are two sides of the same coin. To effectively integrate technical subsystems,

systems engineers rely on methods for proactively engaging with and organizing corresponding

individuals and groups. Interview findings suggest that systems engineers build personal

connections with discipline engineers through empathy, questioning techniques, and active

listening. They facilitate meetings and interactions between disciplines to support additional

information exchange. Additionally, systems engineers have formal methods for recording,

integrating, and distributing this information according to organizational processes. This includes

54

requirements definition and validation plans, schedules, change documentation, etc. (Collopy,

2019).

To explain the lack of support for “organizational systems thinking” as an independent

factor, we interpret the findings to mean that systems thinking in systems engineering is a

framework for organizing technology, people, and information. In other words, organizational

systems thinking does not exist as an independent construct; rather, systems thinking is itself an

organizational framework. Systems, whether technical or social, require organization. Technical

systems thinking and social systems thinking both include organizational themes: e.g.,

partitioning subsystems and delegating work; identifying interactions and coordinating

discipline engineers. The systems thinking framework can be applied to organize, understand,

and influence the relationship between both elements in systems and individuals in design teams

and organizations. This is related to the “learning and information” theme from the thematic

analysis in Chapter 2.

Another major finding is that significant social systems thinking items bear strong

resemblance to design thinking concepts, as depicted in Figure 3.2. Interview data suggested

many instances of design thinking concepts in systems engineering. “Empathy/understanding”

and “human-centered” were among the most common codes in the thematic analysis in Chapter

II. Design thinking attitudes include many of the significant social and social/organizational

systems thinking attitudes, such as “I enjoy using group creativity techniques such as mind

mapping/brainstorming.” and “Engineering is a creative process.”

55

Figure 3.2 Findings from Study 1 suggest that social systems thinking items seem to overlap

with the design thinking framework. This finding is explored further in Study 2. Technical and

organizational systems thinking items from Study 1 are used again in a new systems thinking

factor. Social systems thinking items from Study 1 are reorganized into a design thinking factor

that includes several new items for testing in Study 2.

This finding is explored further in Study 2. The goal of Study 2 is to understand the

relationship between systems thinking and design thinking, which includes social systems

thinking. Many items from Study 1 were used again in Study 2. Additional thematic analysis of

the design thinking literature is discussed in Section 3.3. The goal of this analysis was to identify

and include in Study 2 additional design thinking items which reference design thinking

practices, such as prototyping, that were not included as a part of social or organizational

systems thinking in Study 1.

3.3 Study 2: Systems Thinking and Design Thinking

The previous section described an exploratory study designed to test the quantitative

research methodology and inform preliminary hypotheses about the latent factors underlying a

construct we called systems design thinking. A three-factor model derived from interview data

was tested, consisting of technical, social, and organizational factors. Findings suggested that

this model was not supported in the quantitative analysis. In Study 2, systems thinking items are

redistributed as depicted in Figure 3.3. The social systems thinking factor is restructured as a

design thinking factor based on the findings in Section 3.2.3. The design thinking factor also

56

includes themes like interpersonal relationships, empathy, understanding, questioning, curiosity,

prototyping etc., not included as part of social or organizational systems thinking in Study 1.

Figure 3.3 In Study 2, systems thinking items from Study 1 are redistributed and a new model is

tested. The social systems thinking factor from Study 1 is included as part of a new design

thinking factor. The design thinking factor also includes additional themes like empathy,

questioning, prototyping, etc. not included as a part of social organizational systems thinking

items in Study 1.

3.3.1 Comparing Systems Thinking and Design Thinking

Chapter 2 described codes that capture major themes in design thinking and engineering

systems thinking. This section expands on these themes, describing similarities and differences

between themes in design thinking and engineering systems thinking. These similarities and

differences are represented as items and explored quantitatively in subsequent sections.

Differences Between Systems Thinking and Design Thinking

Systems thinking and design thinking have different theoretical backgrounds and goals,

which influence problem framing, approach, and solution methods. Systems thinking is based on

systems science, systems theory, cybernetics, management science, and operations research;

design thinking is based initially on experience from practice and later on social and behavioral

science, art, and the humanities. These differences result in some of the stereotypes described in

earlier sections.

57

Systems thinking is more associated with mathematical modeling and analysis than

design thinking, and objectivity drives the problem framing approach. In systems thinking, focus

is on the technical solution and deals with quantifiable relationships between elements. Because

of the complexity of these models, necessary information isn’t always available, and accepting

this ambiguity is part of the process. Emergence may happen, and managing related risks is

imperative.

In design thinking, the problem solving approach is informed by industrial design,

product design, psychology, art and architecture, and many other disciplines. These disciplines

make use of qualitative analysis, where subjectivity is sought out and embraced, especially as is

reflects the human experience. Focus is on problem definition in design thinking. Uncovering

latent needs by sharing and understanding diverse personal experiences is of particular

importance. Then, innovative, human-centered solutions are generated based on these rich

descriptions. Methods for design thinking are also human-centered, and include practices such as

storytelling, empathy-building, and creativity techniques. Generally speaking, there is less cost

and risk involved in consumer product design, so less emphasis is placed on planning,

scheduling, coordinating, and managing in the design thinking framework.

Systems thinking and design thinking are also found to have different goals and

processes, based off the analysis in Chapter 2. For example, systems thinking requires

partitioning and coordination processes. The systems thinking framework is holistic, but a key

practice is understanding elements both individually and also as they interact and relate to the

larger system. The different processes can be summarized as follows:

In systems thinking:

• Identify system elements and their interactions/interdependencies

58

• Partition solution

• Coordinate design

• Verify/validate requirements have been met

In design thinking:

• Identify needs, wants, and desires of users, stakeholders, and beneficiaries

• Synthesize

• Define problem

• Generate alternative solutions

• Prototype and test

Similarities Between Systems Thinking and Design Thinking

The “human element” is a key feature of both systems thinking and design thinking

frameworks. Both systems thinking and design thinking deal with people, but different people

and in different ways. In systems thinking, the people are the systems designers – those

responsible for designing the technical system elements. Coordinating technical systems requires

coordinating the related individuals and groups. This is done formally and informally, and

requires empathy and understanding, communication skills, etc., similar to design thinking. In

design thinking (for product design), the people of interest are users of the product being

designed. Communication, empathy, and understanding are still necessary for engaging them,

but the nature of the user/designer relationship is different than the designer/designer

relationship. This adds a “leadership and management layer” to engineering systems thinking.

Leadership and management skills are not necessary for engaging and empathizing with users as

described in the design thinking framework, but leadership and management skills are necessary

to organize designers and their work activities.

59

Another common theme is the exploration of relationships and interactions. In systems

thinking, the focus is on each system element and how it affects and is affected by the others.

Change propagation is a particularly salient concern, as the effects of change in one subsystem

may be disruptive to the functioning of other subsystems. Relationships and interactions also

have implications for partitioning and coordination approaches, and can dictate the best approach

for integrating elements to produce a cohesive, fully-functioning system. The design thinking

framework seeks to uncover important factors in the problem space, and understand how their

interactions create problems and unmet needs. Design thinking also attempts to address these

factors holistically, through the development of a single, unified solution.

3.3.2 Scale Development

The scale for comparing systems thinking and design thinking in Study 2 was developed

using codebook findings from Chapter 2 and insights from Study 1. Systems thinking is

represented as a single factor, consisting of the following subthemes:

• “Big picture” thinking

• Tolerant of change/uncertainty

• Managing risk

• Coordination/communication

In Study 1, systems thinking was deconstructed into technical, social, and organizational

factors. In Study 2, technical, social, and organizational systems are treated as attitude objects,

while attitudes themselves are reflected above. I.e., regardless of whether the technical, social, or

organizational system is being discussed, systems design thinkers should take a “big picture”

approach, tolerate ambiguity/uncertainty, manage risk, and coordinate.

Design thinking themes include the following:

60

• Human-centered

• Creativity

• Prototyping/testing

• Iteration

These concepts, values, and practices are more human-centered and more related to

understanding and working within social systems. Empathy, creativity, and storytelling enable

communication across disciplines, organizations, and levels of education and experience.

Prototyping, testing, and iterating with users deepens the relationship between designers, and

between designers and users. These practices focus on the human-centered process of designing,

rather than the technical, social, and organizational systems on and in which design happens.

In Study 2, 40 items were examined in a second exploratory factor analysis. A total of 29

items were retained from Study 1, and 11 new items were added. Example items are included in

Tables 3.8 and 3.9. New items are listed with their corresponding systems thinking/design

thinking theme. Items that were adapted from Study 1 and reused are labelled

technical/social/organizational in parentheses underneath the new systems thinking/design

thinking themes to reflect their categorization in Study 1.

61

Table 3.8 Example systems thinking attitude items and themes from Study 2. A total of 29 items

were retained from Study 1, and 11 new items were added. New items are listed with their

corresponding systems thinking theme. Items that were adapted from Study 1 and reused are

labelled technical/social/organizational in parentheses underneath the new systems thinking

themes to reflect their categorization in Study 1.

Systems Thinking Attitude Statements Themes

I need to know how my technical decisions affect the bigger system architecture. "Big picture" thinking

(Technical)

I like to know what my colleagues are working on, even if it isn't directly related to my work. "Big picture" thinking

(Social)

I am comfortable working with flexible/changing system requirements. Manage uncertainty/risk

(Technical)

I am comfortable making technical decisions with incomplete information. Manage uncertainty/risk

I always have a backup plan in case something goes wrong. Manage uncertainty/risk

I enjoy taking on leadership roles. Coordination

(Social)

It is easy for me to establish a common vocabulary with colleagues from other disciplines. Coordination

(Social)

I am comfortable delegating tasks to others. Coordination

I prefer to resolve minor work issues in person rather than over email. Coordination

62

Table 3.9 Example design thinking attitude items and themes from Study 2. A total of 29 items

were retained from Study 1, and 11 new items were added. New items are listed with their

corresponding design thinking theme. Items that were adapted from Study 1 and reused are

labelled technical/social/organizational in parentheses underneath the new design thinking

themes to reflect their categorization in Study 1.

Design Thinking Attitude Statements Themes

I consider myself to be an empathetic person. Human-centered

I enjoy co-designing products/systems with customers. Human-centered

Workplace mentoring programs are important. Human-centered

I use storytelling and/or analogies to describe my work to others. Human-centered

(Social)

I consider myself to be a creative person. Creativity

I enjoy using group creativity techniques such as mind mapping/brainstorming with my team. Creativity

(Social)

Prototyping speeds up the design process. Prototyping/testing

(Technical)

Higher-fidelity prototypes are always best. Prototyping/testing

(Technical)

I see iteration to be an improvement of an idea rather than a setback. Iteration

3.3.3 Pilot Test, Factor Analysis, and Results for Study 2

Exploratory factor analysis (EFA) was used to study 40 attitude statements in Study 2.

These items were arranged in a 5-point Likert scale and evaluated by a second sample of 16

professional engineers and engineering researchers from industry, academia, and government.

Again, a sample size of 16 is very small for this type of study, but Study 2 is also considered to

be a pilot study in these analyses. We use maximum likelihood estimation and varimax

63

(orthogonal) rotation again in this analysis, as it is expected that an individual can have any

combination of systems design thinking attitudes. We specify the model to explore solutions with

up to three factors. We expect to see support for a two-factor model, with distinct systems

thinking and design thinking factors.

The three-factor model did not produce any interpretable results. The 2-factor model fit

slightly better than the 1-factor model, although both models are considered poor according to

standard goodness of fit indices. Following the same protocol from Study 1, all variables in the

2-factor model with loadings below .300 were dropped, and the data reanalyzed. Variables with

loadings above .300, retained in the post-hoc analysis, are listed with varimax rotated loadings in

Tables 3.10.

Table 3.10 Two-factor EFA results with varimax rotated loadings (Study 2)

Item ID F1 F2 Attitude Statement

T2 0.383 0.267 It is important that I understand how my work contributes to the overall system design.

T4 0.672 0.263 I like to receive a detailed set of requirements before beginning a project.

T5 0.201 0.532 I have received formal education in more than one discipline.

T10 0.328 -0.469 Higher-fidelity prototypes are always best.

S1 -0.030 0.662 I consider myself to be an empathetic person.

S2 -0.908 0.380 I enjoy co-designing products/systems with stakeholders outside my organization.

S3 -0.123 0.860 I use storytelling and/or analogies to describe my work to others.

S5 0.001 0.362 I enjoy using creativity techniques such as mind mapping/brainstorming with my team.

O2 0.510 0.112 Decision authority should be equally distributed between engineering and project

management.

O6 0.512 -0.138 I prefer to resolve minor work issues in person rather than over email.

O7 0.426 -0.134 I feel comfortable delegating tasks to others.

64

In this analysis, Factor 2 can clearly be identified as design thinking. Empathy, co-design,

storytelling, and creativity are key themes, as identified in the codebook in Chapter II. Many of

the social systems thinking items from Study 1 loaded significantly onto the Design Thinking

factor in Study 2. Items in Factor 1 appear to represent technical and organizational themes of

systems thinking. These include requirements definition, delegation, partitioning, etc., and seem

to more closely reflect systems engineering, as described in the codebook in Chapter II. This

finding, and findings from Study 1, are combined in Figure 3.4.

Figure 3.4 Systems engineering and design thinking frameworks each include elements of

systems thinking. Study 1 indicated that social and some organizational systems thinking items

had much in common with the design thinking framework. Study 2 suggested that technical and

some organizational systems thinking items may be closely related to the systems engineering

framework. This finding is explored further in Study 3.

3.3.4 Discussion

As described in Section 3.3.1, systems thinking and design thinking share some

fundamental similarities. Both are holistic approaches for understanding problems, engaging

with stakeholders, and generating and realizing potential solutions. While design thinking

originally evolved in business as a method for developing consumer products, it is now more

generally applied as an innovation framework for technical, social, and economic problems. In

today’s landscape, in which the internet, big data, and globalization have made complexity the

norm, it is difficult to do design thinking without also doing systems thinking. While design

thinking alone provides a compelling process for innovation, consideration of systemic

65

complexity and systems dynamics is necessary to ensure that potential solutions to these

problems are viable and feasible. Thus, there exists an opportunity to integrate systems

engineering and design thinking more formally. Complex social challenges like education and

healthcare require systems engineering to design and diffuse innovations at scale (Tjendra,

2018). The relationship between systems engineering and design thinking attitudes is explored in

Study 3.

3.4 Study 3: Systems Engineering and Design Thinking

Section 3.3 described development and testing of a two-factor model of systems design

thinking, consisting of a systems thinking subscale and a design thinking subscale. This model

was not supported by the data, possibly due to the similarities between systems thinking and

design thinking frameworks. Systems thinking and design thinking are both holistic problem

solving approaches that emphasize careful analysis of the problem space, with specific focus on

interactions between elements of the system. Systems thinking is concerned with partitioning and

coordination of these elements, while design thinking focuses primarily on synthesis between

elements (Spacey, 2016). Significant systems thinking items from Study 2 more closely reflected

the processes of systems engineering, focusing on modeling, analysis, requirements definition,

and planning rather than the epistemological representation of systems.

Big differences exist between systems engineering and design thinking, especially in the

problem definition phase (whereas systems thinking and design thinking have a similar approach

to problem definition). In design thinking, it is assumed that the problem is not well-understood,

and uncovering the problem is a key part of the process. Systems engineering processes do not

formally begin until the project, problem, and requirements are already defined. These processes

require different skills; in design thinking, ideas are integrated; in systems engineering, hardware

66

and software are integrated. Design thinking is an open ended, iterative process; systems

engineering is a linear, stepwise process with clear answers and next steps. These processes ask

different questions: design thinking asks, “what is the need?” to generate requirements; and

systems engineering asks, “what can we build?” to meet the given requirements.

This section explores systems engineering attitudes and their relationship to design

thinking attitudes, structured as follows. First, a literature review is conducted to understand

contemporary descriptions of the SE/DT distinction- the “stereotypes” of engineers vs. designers.

Then, scale development and factor analysis are discussed, in which we test a two-factor model

of systems design thinking consisting of systems engineering and design thinking subscales. A

pilot test is conducted first using Amazon M-Turk and produces good results. The scale is then

distributed to broader audiences through snowball sampling and on Reddit, a social media

platform. Reddit’s usefulness for survey research is discussed in Section 3.4.4. Then, good

model results are confirmed, first quantitatively using confirmatory factor analysis, and then

qualitatively, through additional analysis of feedback from Reddit users.

3.4.1 Comparing Systems Engineering and Design Thinking

Thematic analysis of the literature reflecting the traditional systems engineering vs.

design thinking distinction is used to derive conceptual models of the relationship between

systems engineering and design thinking attitudes. Systems engineering and design thinking have

been widely seen as distinctly different processes, systems engineering being more data-driven

and analytical, and design thinking being more human-centered and creative.

Classical systems engineering and operations research by Checkland (1981) describes

this duality as systems practice and systems thinking. The application of systems thinking in

systems engineering practice is described as part of soft systems methodology (Checkland and

67

Scholes, 1990). Soft systems methodology is a sense-making approach to solving engineering

problems that addresses human and social aspects of problem situations in addition to

engineering objectives. Holwell (1997) describes soft systems methodology as “classic systems

engineering with the transforming addition of human activity systems modelling.” Our themes

reflect this characterization by capturing “classic systems engineering” (i.e., methodical,

analytical, and data-driven) attitudes and attitudes about human-centered (i.e., empathetic,

innovative, and creative) modelling and design processes.

Pidd (1996) summarizes the differences between “hard” engineering approaches and

“soft” analytical approaches like design thinking along four dimensions: problem definition,

model, the organization, and outcomes. From Pidd’s work we take the following description of

differences related to problem definition:

“Problems are social or psychological constructs that are the result of framing and

naming (Schön, 1982). This contrasts with the view, common in engineering, that work

begins once a need is established. Thus...in soft analysis, the work focuses on ends as

well as means to those ends. In hard systems engineering, the idea is to provide

“something to meet the need” and the concern is with “how [do we meet the need]...not

what [is the need]?” (Checkland and Scholes, 1990)

A common attitude in engineering is that design problems are taken as given, along with

their parameters, constraints, objectives, and requirements. These problems are unambiguous,

involve known physical variables, and mathematical and simulation-based analyses are

necessary and sufficient for solving them. In “soft analysis,” problems, parameters, and solutions

are ambiguous, and qualitative methods are used to consider social and psychological contexts

while defining the design problem and engineering parameters.

68

A similar discussion of requirements definition approaches appears as part of a research

agenda on the “top 10 illusions of systems engineering” (Pennock and Wade 2015). This work

describes classical assumptions of systems engineering, and ways in which these assumptions are

illusory in today’s systems environment. Requirements definition is described as traditionally

absolute and unambiguous:

“Traditional systems engineering assumes that there is a “right’ or optimal

answer....this is one of the major assumptions that often separates traditional

systems engineering from systems thinking which embraces the soft, and often

inconclusive, nature of systems. Systems engineering operates under the

illusion that it is possible to specify unambiguous requirements using human

language. Of course, experience tells us that it is quite common for reasonable

people to hold differing interpretations of a requirement.” (Pennock and Wade

2015)

The authors go on to describe systems engineering as traditionally mechanistic:

“Traditional systems engineering is blind to this human element, including

culture and history. [In systems engineering] it is implicitly assumed that these

factors will not substantially influence outcomes. However, norms and values

can affect both what potential solutions are considered and how a systems

engineering program is executed.” (Pennock and Wade 2015)

“Empathetic” and “relationship-driven” are included as complementary dimensions of

the design thinking process. In addition to empathy-driven processes, design thinking also

includes strategic and practical processes for the innovation of products and services within

69

business contexts. Inspiration, creativity, iteration, and prototyping are human-centered

processes for stimulating innovation and increasing business viability (Brown and Katz, 2011).

Strategic and practical processes in systems engineering (we refer to these as

“methodical/systematic”) more closely resemble project management processes such as

scheduling, budgeting, and change documentation.

Other Differences Between Systems Engineering and Design Thinking

Like systems thinking and design thinking, systems engineering and design thinking have

different theoretical backgrounds and methodologies. Systems engineering includes theory and

methods for design optimization, operations research, and other processes, which are often

mathematics-based and model-driven. Systems engineering is concerned with “solving the

problem right.” Design thinking is intended to address other considerations, and ensure that

designers are “solving the right problem.” Design thinking moves away from mathematical

modeling, favoring theory from disciplines such as psychology, sociology, and anthropology and

methods such as field studies, interview, ethnography, etc.

Systems engineering and design thinking also have different starting points. While design

thinking can be applied in different ways throughout a design process, it is typically considered

to be a part of front-end design. Again, the goal of design thinking is to provide insight into

stakeholder values for the purpose of problem definition; thus, design thinking begins with needs

assessment and “desirability” of the proposed solution. Design thinking goes from this stage

through to concept generation, prototyping, and testing.

Systems engineering starts with “solving the problem right.” The problem has already

been defined and the design concept has largely been selected; systems engineering is concerned

with optimizing and implementing the desired solution.

70

3.4.2 Scale Development and Pilot Test

The following themes were selected to represent systems engineering attitudes in Study

3:

· Unambiguous problem definition, objectives, and requirements

· Analytical/data-driven

· Methodical/systematic

The following themes were selected to represent design thinking attitudes in Study 3:

· Ambiguous problem definition, objectives, and requirements

· Empathetic/relationship-driven

· Innovative/creative

Ethnographic findings suggest that these characteristics are not necessarily “opposites” or

mutually exclusive, although they are often represented as such. The “methodical/systematic”

engineering process is not the opposite of an “innovative/creative” design process and many

systems engineers are both methodical/systematic and innovative/creative. Similarly, one can be

both empathic/relationship-driven and analytical/data-driven, and research suggests that a

human-centered approach is actually necessary for coordinating complex sociotechnical systems

(Williams and Derro, 2008). Experienced systems engineers are also able to challenge

assumptions about problems and constraints while working within process boundaries.

Statements reflecting these six themes were written using language from professional,

academic, and open-source materials on systems engineering and design thinking (INCOSE,

2015; Shea, 2017). These statements were included as items on a five-point Likert scale ranging

from “1- strongly disagree” to “5- strongly agree”. Systems engineering items and themes are

listed in Table 3.11. Design thinking items and themes are listed in Table 3.12.

71

Table 3.11 Systems engineering attitude items from Study 3

Assume unambiguous problems, objectives

and solutions Analytical/data driven Methodical/systematic

I like to receive a detailed set of requirements before beginning a project.

I build simulations and/or models to test my ideas. I document every change I make to my designs.

I generate better ideas when I have a defined

problem statement and objectives.

I use quantitative methods to compare different

ideas.

I always compare my final design to

the initial project goals.

I can infer a customer's expectations based

on the project goals.

I use mathematical modelling/analysis to predict

whether my designs will meet customer

expectations.

I evaluate designs based on cost and

schedule.

I make design decisions based on data/analytical results.

I evaluate the success of my designs using

quantifiable performance measures.

Table 3.12 Design thinking attitude items from Study 3

Assume ambiguous problems/solutions Empathetic/relationship-driven Innovative/creative

I am comfortable working with changing

project requirements.

I am an empathetic person. Iteration is an improvement of a

design rather than a setback.

I like to redefine or restructure the problems I

am given to work on.

I like to speak directly with my

customers to ensure that my design

meets expectations.

I am a creative person.

I like to find unconventional ways to solve

problems instead of relying on past methods.

I like to interact with customers

frequently throughout the design

process.

I find inspiration for my work in

my everyday life.

I am a curious person. I use storytelling techniques to

understand the problems I am

given to work on.

I use intuition to make design

decisions.

72

A pilot test was conducted through Amazon Mechanical Turk (MTurk), a crowdsourcing

marketplace for distributing processes and jobs to a decentralized human workforce (Amazon,

2018). MTurk enables “Requesters” to coordinate the use of human intelligence to perform tasks

that computers are unable to do. These tasks include things like simple data validation and

research, in addition to more subjective tasks like survey participation, content moderation, and

more (Amazon, 2018).

We chose to use MTurk to recruit participants to limit demands on our known pool of

experts after the first two pilot studies. Amazon Mechanical Turk allows for qualifying users

before they work in their tasks; for this study, we requested that all users have degrees in

engineering. Data was collected from 32 individuals with bachelor’s degrees in engineering.

Results were promising, so we proceeded to recruit a larger panel of experts for model testing as

described in Section 3.4.3.

3.4.3 Subject populations & recruitment

The survey was designed for an expert sample working in a professional systems

engineering or design context. The survey was developed using interview data from experts,

expert materials such as professional systems engineering handbooks, and published research on

experts conducted in professional settings. For this study, we recruited a small sample of known

experts in systems engineering and design thinking research. This panel consisted of 88 self- and

peer- identified experts, recruited through personal referral, listserv data, and from participation

in engineering design conferences and events. Of 88 expert participants, 66 held doctorate

degrees and 15 held masters degrees. Because a sample size of 88 is still too small to draw

meaningful conclusions using exploratory and confirmatory factor analysis (Kline, 2015), a

second group of participants was also recruited. The second group consisted of a panel of 369

73

participants recruited through the social media site Reddit. We use multigroup confirmatory

factor analysis to compare the two recruited panels as a consistency check that the same model

emerges across both samples, even if one sample is relatively small.

3.4.4 Crowdsourcing for Data Collection: Recruiting From Reddit

In this study we explore the use of crowdsourcing in the context of data collection.

Crowdsourcing is a term that refers to tasks performed by an undefined network of people

(Howe, 2006). Survey distribution and data collection over the internet can be viewed as

crowdsourcing, but modern definitions of crowdsourcing have shifted away from the large

untargeted network of people contributing to a body of knowledge to a targeted group of

individuals that participate in organizational decision making processes (Kietzmann, 2016).

There are several contextual variations of the term “crowdsourcing,” such as Open Innovation

Challenges, Data Collection, and Analysis (Hill, Dean, and Murphy, 2014).

While there are popular options such as Amazon Mechanical Turk in the crowdsourcing

marketplace (Landers and Behrend, 2015), measuring engineering and design attitudes requires

a targeted distribution strategy to individuals with appropriate background experience. We seek

other convenient distribution strategies and look to Reddit as a source for study participants with

appropriate background in engineering. Reddit is an online forum, organized into smaller

communities called subreddits. There are several engineering and design-focused subreddits to

which the survey recruitment script was posted, described in detail in the results section.

There currently exist too few published articles where participants were recruited through

Reddit to gauge reliably the quality of data collected through it. However, using Reddit to

recruit study participants is potentially advantageous for several reasons (Shatz, 2017). First, it

is possible to recruit large samples in a short amount of time. Second, Reddit’s community of

74

subforums, or subreddits, makes it is possible to recruit participants from specific demographics

and special interest groups. Distributing a survey via Reddit also adds the following layers to the

process of distributing a survey:

• Upvotes and downvotes: An upvote (positive vote) or downvote (negative vote) adds to a

score that indicates overall quality of a post on Reddit. Early votes determine the quality

of the post for future readers and hence influences the popularity of that post on a

particular subreddit (Birman, 2018).

• Academic discourse with users: The incentive for the user to engage in quality academic

research is to gain an understanding of the topic that they did not have before. Reddit has

evolved by relying on external information to satisfy an ever increasing need for original

and self-referential content (Singer et al., 2014). Users look to provide feedback on their

personal experiences via feedback in the survey or in the comments section. The quality

of these comments and feedback could span a wide range based on the disinhibition

effect (Koivu, 2015).

• Targeted distribution: Surveys can be presented to selected subreddits. Selection of

subreddits is based on factors such as quality of content on a subreddit, day-to-day

activity, number of subscriptions for the subreddit, number of Redditors online who are

subscribed to the subreddit, age of the subreddit, and other similar metrics.

There are general guidelines for posting content to Reddit informally known as Reddiquette

(Reddit, 2018). The following Reddiquette guidelines should be adhered to while posting to

Reddit:

75

• Having a Reddit account with sufficient age/activity to begin posting: Creating

several posts with hyperlinks in them on a new account will trigger auto-moderation

and may even result in a ban on the account.

• Removing personally identifying information from the survey recruitment script:

This includes names and email addresses that link the Reddit user account to a real

person. This is done to avoid bots/scavenging algorithms from collecting sensitive

information. This can also be grounds for automatic deletion of a post. This

information should only be presented inside the survey itself if required.

• Using full length links: Using shortened links will hide the original URL. This will

cause automatic deletion of the post in some subreddits.

• Assigning appropriate flair: Flair for a post on a subreddit indicates the nature of the

post to users that read the post. These are tags attached to the original post that are

necessary for Redditors on a subreddit to determine the nature of a post. Default

flairs are unique to every subreddit. Flair can be used to indicate a post intended for

discussion, or a post that conveys theoretical or factual knowledge.

• Search for appropriate subreddits: Allow sufficient time to engage with a subreddit

and determine its true nature. Spamming several unrelated subreddits will lead to

backlash from the community through comments and downvoting.

• Tagging: Title of the post should include a tag such as [Survey] and [Intended

Demographic] to clearly indicate the purpose of the post. This is to help users avoid

reading a post about a survey when they are not interested in taking a survey.

• Time for completion: The script should also include accurate representation of the

amount of time for completion of the survey. Underestimation could lead to

76

frustrated users and overestimation could lead to decrease in number of survey

attempts (Galesic and Bosnjak, 2009).

• Prepare to engage actively with Redditors over the life of a post: A post is most

active in the first 2-3 hours. Tips on maximizing the utility of a Reddit post and

maximizing engagement are summarized in the article by Shatz (2017).

The Systems Design Thinking Scale on Reddit

The Systems Design Thinking Scale was posted to 25 unique subreddits. A total of 369

Redditors provided scores for the 23 variables in Table 3.12. In this sample, 35% reported their

job title as senior level or above (management, senior management, director, and professorship

positions); 37% as entry-level or analyst/associate; and 17% as student/intern. The survey

completion rate was ~40%.

Table 3.13 summarizes the results from the top 3 posts created on Reddit. Although it is

difficult to determine what constitutes a successful post, posts that had positive community

engagement early on saw higher view counts. The best posts are those in which users leave

highly personalized feedback in the comments section of the post, and researchers respond to

these comments individually.

77

Table 3.13 Posting the Systems Design Thinking Scale on Reddit. The “r/” preceding Subreddit

names indicates “Subreddit.” “Subscriptions” refers to the number of users following the

Subreddit. “Upvotes” refers to the number of positive votes the survey post received on each

subreddit.

Subreddit Subscriptions Upvotes % Upvoted Views

r/MechanicalEngineering 20000 27 89 1500

r/userexperience 34700 24 100 1500

r/ElectricalEngineering 23000 16 86 1400

All the subreddits in the above table indicated interest in viewing/learning the results

from the survey. There was little to no engagement from other subreddits due to poor initial

reaction possibly due to the survey not resonating well with the nature of the subreddit.

The mean response time in seconds for a 5% trimmed sample (to eliminate extremities)

of the subset of fully completed responses was 291 seconds with standard deviation of 138

seconds. This is in line with advertised survey duration of 300 seconds. The total percentage of

all survey attempts that were fully completed is ~40%.

3.4.3 Exploratory Factor Analyses

Partial responses with missing data were deleted from the data set. All 23 items in Tables

3.11 and 3.12 were included in the EFA. The analysis was conducted on a data set consisting of

457 observations of these 23 variables (the expert and Reddit data were combined).

Factors were extracted using maximum likelihood estimation. An orthogonal rotational

method (varimax) was used because we did not expect the obtained factors to be correlated based

on the theoretical evidence described in Section 3.1.

A factor solution was obtained by considering Kaiser’s criterion (retaining factors with

eigenvalues greater than one), the interpretability of obtained factor solutions, the internal

78

consistency of the obtained factors, and model fit indices (Worthington and Whittaker, 2006).

Items were removed if they did not load onto a distinct factor consisting of at least three items, or

if they did not have a primary factor loading of 0.40 or above with no significant cross-loadings

onto other factors.

Four-, three-, two- and one- factor models were compared in Table 3.14. In the 4-factor

model, systems engineering variables SE1 and SE2 group together (with loadings of 0.844 and

0.516); these are the only two items loading greater than 0.400 on Factor 1. Factor 2 includes

design thinking items 5-10 and 12, providing strong evidence for a “Design Thinking Attitudes”

factor. Factor 3 includes systems engineering items 5-8, providing evidence for a “Systems

Engineering Attitudes” factor. Factor 4 includes design thinking items 2 and 3; these are the only

two items loading greater than 0.400 on Factor 4. Systems engineering items 3, 4, 9, 10, and 11

and design thinking items 1, 4, and 11 didn’t load onto any factor with a loading higher than

0.400. These results can be seen in Table 3.14.

A second analysis was conducted. First, systems engineering items SE 1 and SE 2 are

dropped from the model. These two items grouped together into a single factor in the first

analysis. These items describe attitudes about problem statements, objectives, and requirements.

Preferences for these vary considerably among systems engineers; more qualitative analysis on

this is required to fully understand the perception of requirements in systems engineering and to

draft good survey questions about it. Design thinking items DT 2 and 3 are also dropped in the

post-hoc analysis for similar reasons. These items grouped together in a single factor describing

preferences for speaking and working directly with customers. A second EFA suggests high

eigenvalues for 1 and 2 factor models (3.303 & 2.574), and lower values for the third factor and

beyond (1.222, 1.160). A clear interpretation exists for the 2-factor model. There is no clear

79

interpretation of 3-factor model, and no items in this factor have loadings greater than 0.400.

Results are captured in Tables 3.15 and 3.16.

Model fit indices suggested that the two-factor solution is a relatively good fit to the data.

The Root Mean Square Residual (RMR) value of 0.046 was below the suggested 0.08 cut-off for

very good fit. The Root Mean Square Error of Approximation (RMSEA) value of 0.047 was

below the recommended 0.06 cut-off value for “very good” fit (Hu & Bentler, 1999; Kline,

2015). While there is some disagreement regarding exact cut-offs for fit indices and the RMR

estimate of model fit, the two-factor EFA model was determined to be the best model. This

solution suggests two conceptually meaningful factors, reflecting the underlying systems

engineering and design thinking frameworks.

80

Table 3.14 Varimax rotated loadings for four factors (Study 3)

Item ID Factor1 Factor2 Factor3 Factor4 Attitude Statement

SE1 0.844 0.187 0.025 0.032 I like to receive a detailed set of requirements before

beginning a project.

SE2 0.516 -0.012 0.000 -0.021 I generate better ideas when I have a defined

problem statement and objectives.

SE3 0.237 -0.123 0.167 -0.005 I can infer a customer's expectations based on the

project goals.

SE4 0.015 -0.218 0.359 -0.008 I build simulations and/or models to test my ideas.

SE5 -0.057 -0.066 0.660 -0.001 I use quantitative methods to compare different

ideas.

SE6 0.128 0.040 0.684 0.089

I use mathematical modeling/analysis to predict

whether my designs will meet customer

expectations.

SE7 -0.012 0.027 0.588 -0.050 I make design decisions based on data/analytical

results.

SE8 0.052 -0.046 0.553 0.059 I evaluate the success of my designs using

quantifiable performance measures.

SE9 0.168 0.009 0.268 -0.035 I document every change I make to my designs.

SE10 0.236 -0.124 0.239 -0.051 I always compare my final design to the initial

project goals.

SE11 0.265 0.028 0.284 -0.150 I evaluate ideas based on cost and schedule.

DT1 0.050 -0.297 -0.092 -0.186 I am an empathetic person.

DT2 0.086 -0.212 0.050 -0.697 I like to speak directly with my customers to ensure

that my design meets expectations.

DT3 0.070 -0.218 0.021 -0.808 I like to interact with customers frequently

throughout the design process.

DT4 -0.107 -0.367 0.031 -0.247 I am comfortable working with changing project

requirements.

DT5 -0.060 -0.486 -0.014 -0.152 I like to redefine or restructure the problems I am

given to work on.

DT6 -0.016 -0.512 0.153 -0.008 I like to find unconventional ways to solve problems

instead of relying on past methods.

DT7 0.135 -0.570 0.089 -0.068 I am a curious person.

DT8 -0.029 -0.402 0.137 -0.086 Iteration is an improvement of a design rather than a

setback.

DT9 0.026 -0.589 0.049 -0.074 I am a creative person.

DT10 -0.044 -0.554 0.079 -0.219 I find inspiration for my work in my everyday life.

DT11 -0.082 -0.384 -0.141 -0.313 I use storytelling techniques to understand the

problems I am given to work on.

DT12 0.089 -0.423 -0.047 0.075 I use intuition to make design decisions.

81

Table 3.15 Post-hoc EFA results: Varimax rotated loadings for three factors (Study 3)

Item ID Factor1 Factor2 Factor3 Attitude Statement

SE3 0.212 0.027 0.097 I can infer a customer's expectations based on the project

goals.

SE4 0.383 0.085 0.172 I build simulations and/or models to test my ideas.

SE5 0.636 -0.031 0.025 I use quantitative methods to compare different ideas.

SE6 0.690 -0.043 -0.107 I use mathematical modeling/analysis to predict whether

my designs will meet customer expectations.

SE7 0.565 -0.187 -0.035 I make design decisions based on data/analytical results.

SE8 0.571 0.049 -0.012 I evaluate the success of my designs using quantifiable

performance measures.

SE9 0.281 -0.098 -0.006 I document every change I make to my designs.

SE10 0.281 0.059 0.099 I always compare my final design to the initial project

goals.

SE11 0.287 -0.274 0.005 I evaluate ideas based on cost and schedule.

DT1 -0.079 -0.095 0.354 I am an empathetic person.

DT4 0.003 -0.394 0.470 I am comfortable working with changing project

requirements.

DT5 -0.006 -0.145 0.523 I like to redefine or restructure the problems I am given to

work on.

DT6 0.188 0.052 0.471 I like to find unconventional ways to solve problems

instead of relying on past methods.

DT7 0.140 -0.058 0.570 I am a curious person.

DT8 0.154 -0.063 0.410 Iteration is an improvement of a design rather than a

setback

DT9 0.129 0.387 0.590 I am a creative person.

DT10 0.113 0.101 0.568 I find inspiration for my work in my everyday life.

DT11 -0.124 0.071 0.447 I use storytelling techniques to understand the problems I

am given to work on.

DT12 0.007 0.064 0.380 I use intuition to make design decisions.

82

Table 3.16 Two-factor EFA results (Study 3)

Item ID SYSENG DESIGN Attitude Statement

SE3 0.206 0.103 I can infer a customer's expectations based on the project goals.

SE4 0.374 0.184 I build simulations and/or models to test my ideas.

SE5 0.638 0.034 I use quantitative methods to compare different ideas.

SE6 0.697 -0.100 I use mathematical modeling/analysis to predict whether my designs will

meet customer expectations.

SE7 0.570 -0.032 I make design decisions based on data/analytical results.

SE8 0.561 -0.003 I evaluate the success of my designs using quantifiable performance

measures.

SE9 0.288 -0.012 I document every change I make to my designs.

SE10 0.272 0.115 I always compare my final design to the initial project goals.

SE11 0.301 -0.002 I evaluate ideas based on cost and schedule.

DT1 -0.076 0.350 I am an empathetic person.

DT4 0.031 0.419 I am comfortable working with changing project requirements.

DT5 0.001 0.519 I like to redefine or restructure the problems I am given to work on.

DT6 0.181 0.484 I like to find unconventional ways to solve problems instead of relying

on past methods.

DT7 0.137 0.563 I am a curious person.

DT8 0.153 0.410 Iteration is an improvement of a design rather than a setback

DT9 0.090 0.568 I am a creative person.

DT10 0.099 0.586 I find inspiration for my work in my everyday life.

DT11 -0.137 0.454 I use storytelling techniques to understand the problems I am given to

work on.

DT12 -0.002 0.386 I use intuition to make design decisions.

Systems engineering items SE3, 4, 9, 10 and 11 and design thinking items DT1 and 12

did not load on either factor with a loading greater than 0.400. These factors were dropped from

83

the model after the second analysis. A total of 11 items were removed from the measure after

failing to meet the minimum criteria. The final factor loadings for 12 items are provided in Table

3.17. These 12 items were included in a confirmatory factor analysis, described in Section 3.4.6.

Table 3.17 Final factor loadings (Study 3)

Item ID SYSENG DESIGN Attitude Statement

SE5 0.638 0.034 I use quantitative methods to compare different ideas.

SE6 0.697 -0.100 I use mathematical modeling/analysis to predict whether my designs will

meet customer expectations.

SE7 0.570 -0.032 I make design decisions based on data/analytical results.

SE8 0.561 -0.003 I evaluate the success of my designs using quantifiable performance

measures.

DT4 0.031 0.419 I am comfortable working with changing project requirements.

DT5 0.001 0.519 I like to redefine or restructure the problems I am given to work on.

DT6 0.181 0.484 I like to find unconventional ways to solve problems instead of relying on

past methods.

DT7 0.137 0.563 I am a curious person.

DT8 0.153 0.410 Iteration is an improvement of a design rather than a setback

DT9 0.090 0.568 I am a creative person.

DT10 0.099 0.586 I find inspiration for my work in my everyday life.

DT11 -0.137 0.454 I use storytelling techniques to understand the problems I am given to work

on.

3.4.4 Confirmatory Factor Analyses

A confirmatory factor analysis was conducted with 12 items and two factors. The systems

engineering factor includes items SE5, 6, 7, and 8 from Table 3.17. The design thinking factor

includes items DT4-11. The model is shown in Figure 3.5.

84

Figure 3.5 Two-factor model with parameter values and standard error

85

Results of this analysis, reported in Table 3.18, suggest an acceptable model fit:

Table 3.18 Model results from confirmatory factor analysis with 12 items

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.702 0.039 18.035 0.000

SE6 0.633 0.040 15.787 0.000

SE7 0.602 0.042 14.449 0.000

SE8 0.557 0.043 12.959 0.000

DESIGN by Estimate S.E. Est./S.E. P-value

DT4 0.423 0.047 8.967 0.000

DT5 0.509 0.044 11.594 0.000

DT6 0.505 0.044 11.559 0.000

DT7 0.585 0.041 14.355 0.000

DT8 0.431 0.047 9.271 0.000

DT9 0.566 0.042 13.625 0.000

DT10 0.601 0.040 15.016 0.000

DT11 0.413 0.047 8.735 0.000

DESIGN WITH

SYSENG 0.118 0.064 1.862 0.063

ꭓ2 df P RMSEA CFI/TLI SRMR

131.777 53 0.000 0.057 0.912/0.890 0.050

Modification indices suggest some issues with variables DT4 and DT11, in addition to

low loadings. DT 4, DT 11, and DT 8 were all dropped due to low loadings (less than .500) and

the analysis was run again with 9 total items: systems thinking items SE5, 6, 7, and 8, and design

thinking items 5, 6, 7, 9, and 10. Model results are reported in Table 3.19.

86

Table 3.19 Model results from confirmatory factor analysis with 9 items

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.705 0.039 18.122 0.000

SE6 0.634 0.040 15.854 0.000

SE7 0.598 0.042 14.333 0.000

SE8 0.557 0.043 12.949 0.000

Two-tailed

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.478 0.048 10.029 0.000

DT6 0.551 0.045 12.167 0.000

DT7 0.575 0.044 12.953 0.000

DT9 0.585 0.044 13.227 0.000

DT10 0.586 0.044 13.311 0.000

DESIGN WITH

SYSENG 0.149 0.065 2.285 0.022

ꭓ2 df P RMSEA CFI/TLI SRMR

43.128 26 0.019 0.038 0.974/0.964 0.034

As shown in Table 3.19, the CFA shows a good fit to the data. The RMSEA of 0.038 is

below the suggested 0.08 cut-off value. The CFI and TLI values of 0.974 and 0.964,

respectively, are above the .90 cut-off value for good fit suggested by Kline (2015). The SRMR

value of 0.034 is below the 0.08 cut-off for good fit. The low correlation (0.149) between factors

suggests that systems engineering attitudes and design thinking attitudes are independent. An

individual can hold engineering attitudes, design attitudes, or both.

The distribution of individuals examined is summarized in Figure 3.6. Mean scores were

calculated for systems engineering and design thinking subscales (scores of 15 and 20,

respectively). Individuals were categorized as “high” in a category if they scored above the mean

on that subscale, and “low” if they scored below the mean.

87

Design Thinking

Low High

Syst

ems

Engin

eeri

ng

Low

58

116

Hig

h

96

188

Figure 3.6 Systems Design Thinking Classification of 458 survey participants. Mean scores

were calculated for systems engineering and design thinking subscales. Individuals were

categorized as “high” in a category if they scored above the mean, and “low” if they scored

below the mean.

Overall, model fit indices indicated that the hypothesized relationships between observed

variables and their corresponding latent constructs were a good fit to the data. All variables

significantly loaded onto the same factor in the CFA as they had in the EFA, which provides

psychometric support for the systems engineering and design thinking Scale and its factor

structure using an alternative modelling approach.

3.4.5 Multigroup CFA

The survey was designed for an expert sample working in a professional setting The

survey was developed using interview data from experts, expert materials (e.g., systems

engineering handbooks), and published research from expert settings. An expert sample was

recruited for the study; however, this sample was small as experts are difficult to find. To survey

the recommended number of participants for EFA/CFA, the sample was expanded to include the

Reddit group. Demographic information was recorded for both groups. Information collected

included “job title,” with the following choices: student, intern, entry-level, analyst/associate,

88

senior level, management, senior management, director, professor, or other with option to write-

in.

Two multigroup CFAs were conducted. First, the Reddit sample was compared to the

expert participants we contacted directly. Then, both samples were combined, and entry-level

participants were compared to expert participants from both samples.

Reddit vs. Known Sample

In this analysis, the Reddit sample (n=371) and known expert sample (n=87) are

compared. Model results are reported in Tables 3.20, 3.21, and 3.22 below:

Table 3.20 Multigroup CFA results: Reddit vs. known expert sample

ꭓ2

Reddit

sample

Known

sample Df p RMSEA CFI/TLI SRMR

126.354 51.819 74.535 66 0.000 0.063 0.915/0.907 0.074

Table 3.21 CFA for Reddit group

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.683 0.043 16.063 0.000

SE6 0.621 0.040 15.352 0.000

SE7 0.585 0.042 13.833 0.000

SE8 0.548 0.043 12.631 0.000

Two-tailed

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.489 0.049 9.942 0.000

DT6 0.554 0.042 13.099 0.000

DT7 0.600 0.047 12.895 0.000

DT9 0.555 0.044 12.687 0.000

DT10 0.534 0.046 11.654 0.000

DESIGN WITH

SYSENG 0.086 0.075 1.143 0.253

89

Table 3.22 CFA for known expert sample

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.717 0.053 13.492 0.000

SE6 0.769 0.057 13.384 0.000

SE7 0.670 0.062 10.790 0.000

SE8 0.597 0.067 8.973 0.000

Two-tailed

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.463 0.060 7.762 0.000

DT6 0.691 0.069 10.051 0.000

DT7 0.478 0.061 7.818 0.000

DT9 0.749 0.059 12.711 0.000

DT10 0.723 0.056 12.827 0.000

DESIGN WITH

SYSENG 0.349 0.124 2.808 0.005

The model performs well for both groups. Systems engineering items have higher loadings

among known experts. Also, there is a higher correlation between systems engineering and

design thinking attitudes in the known expert group.

Entry vs. Senior Level: Both Groups

In this analysis, Reddit data was combined with data collected through snowball

sampling. Entry-level and senior (expert) level were compared. Entry level includes student,

intern, entry-level, and analyst/associate. Senior level includes the following job titles: senior

level, management, senior management, director, and professor. The entry-level group included

213 observations. The senior-level group included 245 observations. Total sample size was 458

observations. Model results are presented in Tables 3.23, 3.24, and 3.25 below:

Table 3.23 Multigroup CFA results: Entry vs. senior level (combined sample)

ꭓ2 Entry-level Senior Df p RMSEA CFI/TLI SRMR

60.469 31.464 29.005 50 0.1475 0.030 0.981/0.979 0.054

90

Table 3.24 CFA for entry-level group

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.681 0.050 13.597 0.000

SE6 0.638 0.047 13.667 0.000

SE7 0.563 0.050 11.233 0.000

SE8 0.549 0.049 11.273 0.000

Two-tailed

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.480 0.056 8.599 0.000

DT6 0.528 0.057 0.199 0.000

DT7 0.595 0.051 11.601 0.000

DT9 0.540 0.047 11.437 0.000

DT10 0.543 0.048 11.277 0.000

DESIGN WITH

SYSENG 0.182 0.097 1.867 0.062

Table 3.25 CFA for senior-level group

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.711 0.043 16.391 0.000

SE6 0.632 0.047 13.309 0.000

SE7 0.627 0.047 13.447 0.000

SE8 0.574 0.049 11.615 0.000

Two-tailed

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.483 0.049 9.885 0.000

DT6 0.543 0.046 11.705 0.000

DT7 0.545 0.050 10.798 0.000

DT9 0.670 0.050 13.527 0.000

DT10 0.641 0.049 12.973 0.000

DESIGN WITH

SYSENG 0.117 0.087 1.342 0.180

The model performs well for both groups, but fits the expert group slightly better. Biggest

91

differences are observed on design thinking items DT9 (0.525 for entry-level vs 0.702 for senior

level, ∆=0.177) and DT10 (0.511 for entry-level vs 0.658 for senior level, ∆=0.147).

3.4.6 Tests for Measurement Invariance

The extent to which this model exhibited measurement and structural invariance between

novices and experts was examined using Mplus v. 8 (Muthen & Muthen, 2017), following the

method by Hoffman (2018). Robust maximum likelihood (MLR) estimation was used for all

analyses. Novices were used as the reference group in all invariance models. A configural

invariance model was initially specified, in which single-factor models were estimated

simultaneously within each group. Factor mean was fixed to 0 and the factor variance was fixed

to 1 for identification within each group. The configural model had good fit. A series of model

constraints were then applied in successive models to examine potential decreases in fit due to

measurement or structural non-invariance.

Equality of the unstandardized item factor loadings across all groups was then examined

in a metric invariance model in which the factor variance was fixed to 1 in novices but was freely

estimated in experts. Factor means were fixed to 9 in both groups. All factor loadings were

constrained to be equal across groups. All intercepts and residual variances were permitted to

vary across groups. The metric invariance model fit well and did not result in a significant

decrease in fit relative to the configural model. The modification indices suggested no points of

localized strain among the constrained loadings. The fact that metric invariance held indicates

that the items were related to the latent factors equivalently across groups; i.e, the same latent

factors were being measured in each group.

Equality of the unstandardized item intercepts across groups was then examined in a

scalar invariance model. The factor mean and variance were fixed to 0 and 1 respectively for

92

identification in the novice group. The factor mean and variance were estimated in the expert

group. Factor loadings and item intercepts were constrained to be equal across groups. All

residual variances were permitted to differ across groups. The scalar invariance model fit well

and did not result in significant decrease in fit relative to the metric invariance model.

Equality of the unstandardized residual variances across groups was then examined in a

residual variance invariance model. The factor mean and variance were fixed to 0 and 1

respectively for identification in the novice group. The factor mean and variance were estimated

in the expert group. All factor loadings, item intercepts, and residual variances were constrained

to be equal across groups. The residual variance invariance model fit well, and did not result in

significant decrease in fit relative to other models.

After achieving measurement invariance as described, structural invariance was then

tested with two additional models. First, the factor variance in experts, which had been estimated

freely, was constrained to 1 to be equal to the factor variance in novices. Second, the factor mean

in experts which had been estimated freely, was constrained to 0.

These analyses showed that full measurement and structural invariance was obtained

between novices and experts.

3.4.7 Additional Qualitative Findings

In addition to quantitative results, the Reddit analysis yielded additional qualitative data

for validating the underlying theory and hypotheses. A feedback box was included in the

Qualtrics form for both samples. Redditors were also able to post comments on the survey

thread. The following responses were received from Reddit and are organized by hypothesis

supported.

93

On systems engineering:

• “When working with engineers (I am trained as a designer) I often find their

process very linear and practical. They often will get the job done, but will also

sometimes miss opportunities.”

On design thinking:

• “In my experience the major difference between a designer and an engineer is that

a designer has a better vision and understanding of how the product will impact

the consumer in their use.”

• “I often create new goals for a project because my clients often have a vague

sense of it. Thinking outside my client’s box is why I get a lot more work,

compared to other designers I work with. And when I do my best work, I’m also

presenting clients something unique, and much needed.”

On engineering and design being perceived as [stereotypically] different:

• “Some questions are strange - like quantitative / data analysis for design.”

• “My perspective is not all things need to look good. They just need to work.

Where needed aesthetics are extra time and need to be discussed with the

customer beforehand.”

On possible relationships between frameworks:

• “I like to think of myself not as an engineer, but as a technical designer.”

• “As a mechanical design engineer, I use engineering tools (computer analysis,

calculations, etc.) to solve design problems (form, fit, function, manufacturability,

cost). They are two sides of a coin.”

94

• “I personally have worked quite well in collaboration with several actual

engineers; while our skills are different, our mindsets both point to analysis and

testing uncharted waters.”

• “In my world engineers ARE designers. There is nearly 100% overlap. We do not

have anyone employed to do design that isn't either a degreed engineer or a senior

technician with decades of experience.”

• “I think design and even art are very important to engineering in providing a base

for creativity and expression. And it goes both ways. Knowing the analytical

aspects of engineering and something as simple as a design of experiments

process is highly useful for ID and product designers.”

In summary:

• “Similarities: Engineering and Design both employ creativity, input and

interaction with user/customer, and an investment of personal passion.

Differences: Engineering makes use of math, science, and rigor to balances

multiple and sometimes conflicting requirements to achieve a successful result.

Design may employ engineering methods but often is driven by other factors. An

engineering solution may be successful but not viewed as an elegant design. A

design solution may appear elegant but not necessarily be a sound engineering

solution.”

3.4.8 Discussion

The goal of the studies described in this chapter was to understand systems engineering

attitudes, design thinking attitudes, and the relationship between these two constructs through the

development and validation of the systems design thinking scale. Traditional representations of

95

systems engineering and design thinking were used to develop structural models of systems

engineering and design thinking attitudes, which were evaluated quantitatively using exploratory

and confirmatory factor analyses.

Findings support the traditional representation of systems engineering and design

attitudes as two distinct latent constructs, but do not support the stereotype that these two

constructs are mutually exclusive. Consistent with contemporary observation and experience,

systems engineering and design thinking attitudes can be complementary. Design thinking can be

used to innovate new solutions based on a "bottom-up" human-centered approach, while systems

engineering processes support change management and integration by maintaining a “top-down,”

big-picture view. This is especially important when applying design thinking to systems-level

problems. While human-centered design processes often generate innovations that meet human

needs, there is no guarantee that their diffusion into a large-scale, complex system will mirror

diffusion into consumer markets. Systems engineering, which includes technical and

organizational elements of systems thinking, is an approach for designing and deploying these

types of solutions and ensuring that they perform optimally in their intended environment.

Systems engineering adds key values and practices synchronicity, consistency, integration, and

optimization to the design thinking process (Tjendra, 2018).

The Systems Design Thinking Scale shows promise as a tool for capturing systems

engineering and design thinking attitudes along a spectrum. Subscale scores for engineering and

design attitudes may be useful for identifying and balancing perspectives within engineering

design teams. This possibility presents an opportunity for an observational study in the future.

Another promising direction for this work is a behavioral study, in which scores on the Systems

Design Thinking Scale are used to predict behaviors with known implications for the success of

96

systems engineering projects. Understanding the relationship between attitudes and behaviors in

this context would be useful for education and training.

Using Reddit to collect survey data had the unintended benefit of enabling concurrent

collection of qualitative data for validating underlying theory and hypotheses, and also for

improving them and refining vocabulary. Through user feedback, we received the following

recommendations for improving the questionnaire:

• For some items, frequency (e.g., rarely → often) would have been a better

indicator than agree/disagree;

• Additional items describing delegation of responsibilities would have been useful

(e.g., “I direct people to do X simulation/analysis.”)

Additional feedback about the questionnaire shared through open-response survey items, along

with the Reddit comment sections, provided information that would be interesting to include in

future qualitative analysis.

3.5 Summary

In this chapter, quantitative analysis guided several iterations of the Systems Design

Thinking Scale. In Study 1, systems design thinking was represented by three factors: technical

systems thinking, social systems thinking, and organizational systems thinking. Findings

suggested that technical and social systems thinking are important factors in systems design

thinking. Findings related to the organizational systems thinking factor were difficult to interpret.

Significant social systems thinking items appeared similar to design thinking concepts,

when the codebook from Chapter II was used to interpret the findings. In Study 2, the difference

between systems thinking and design thinking was explored. Systems thinking items included

significant technical and organizational items from Study 1. Design thinking items included

97

significant social systems thinking items, and several additional items reflecting the design

thinking themes from Chapter II.

Significant systems thinking items appeared similar to systems engineering concepts,

when the codebook from Chapter II was used to interpret the findings. Almost all of the design

thinking items were significant. Systems engineering and design thinking subscales were

analyzed in Study 3. Results suggested that this model of systems design thinking was a good fit.

In the next chapter, we attempt to validate the systems design thinking scale by studying

correlations between scale scores and performance on analytical reasoning and divergent

thinking tasks.

98

CHAPTER IV

Validating the Systems Design Thinking Scale

4.1 Introduction

Chapter 2 described the development of systems design thinking theory, and Chapter 3

described the development of a scale for measuring systems design thinking. In this chapter, a

pilot validation study is described, in which the scale’s ability to predict performance on systems

engineering and design tasks is explored.

The validation study has several goals. The first is to establish construct validity; i.e., that

the systems design thinking scale measures what it claims to be measuring. The second is to

establish criterion validity to determine the extent to which the systems design thinking measure

is related to an outcome. The systems design thinking scale should be operationally useful for

identifying occupational strengths, like the Jung Typology Profiler for Workplace (Kerbel &

Wainstein, 2013), Clifton StrengthsFinder (Rath, 2007), and other similar scales. It should mean

something to be high on systems engineering attitudes, high on design thinking attitudes, or both.

Therefore, it is necessary to identify what skills and behaviors are correlated with these self-

reported attitudes, and their impact on work, in order for the scale to be operationally useful.

Then, the scale can be used to identify or categorize individuals based on their perspectives and

strengths in a meaningful way, as “playing to strengths” is a time-effective way to improve

performance and engagement at work (Rath, 2007).

99

Identifying the right behaviors to observe and the right way to measure them is a major

challenge. In this chapter, a first attempt is made, guided by related work. Scale scores were used

to predict performance on analytical reasoning and divergent thinking tasks. While no correlation

was observed between scale scores and performance on the analytical reasoning task, some

correlation was observed between design thinking subscale scores and performance on the

divergent thinking task. This study yielded other important findings, which are interpreted,

summarized as lessons learned, and used to contribute a validation plan for the Systems Design

Thinking Scale.

4.2 Behavioral Research in Systems Engineering and Design Thinking

Controlled experiments, including laboratory and field experiments, are becoming

increasingly important in systems engineering. While design of experiments is not unique to

systems engineering research, certain features of the systems engineering context pose unique

experimental design challenges that require special consideration (Panchal and Szajnfarber,

2017). First, complex systems are designed and developed by large, geographically-dispersed

teams, over spans of years or even decades. Controlled experiments are typically performed in

brief sessions with one or few subjects involved. It is difficult to recreate complex systems

problems in an experimental setting without significant loss of relevant information. Second, for

experienced systems engineers in real organizations, transactional knowledge of how things get

done in their particular organization is an important aspect of expertise. Recreating this

knowledge in a short laboratory session can be difficult.

Cognitive and behavioral research in systems engineering studies topics such as problem

solving, mental model formation, teams and team effectiveness (Avnet, 2016; DeFranco et al.,

2011; de Graaf and Loonen, 2018). In these studies, surveys, interviews, observation, and

100

documentation are used to study engineering design teams working in aerospace laboratories,

academia, construction, and other settings. Avnet (2016) explores team coordination and shared

cognition in engineering design teams using a network-based approach. DeFranco et al. (2011)

describes the importance of shared mental models for team performance. De Graaf and Loonen

(2018) explore team effectiveness, and the extent to which differences in team effectiveness can

be explained based on characteristics of systems engineering teams and organizations.

Other work in systems research and behavioral science describes a similar approach to

the development and validation of a systems thinking scale (Davis and Stroink, 2016; Randle and

Stroink, 2018; Thibodeau, Frantz, and Stroink 2016). This work describes the development of a

psychometric instrument for measuring systems thinking, and a study of this instrument in

relation to well-studied constructs (e.g., holistic and relational thinking; creativity) and decision-

making tasks in the psychological literature. Other studies of systems thinking list decision-

making tasks, measures of holistic thinking (Maddux and Yuki, 2006; Choi et al., 2007; Chiu et

al., 2000), and measures of relational reasoning (Thibodeau, Frantz, and Stroink, 2016; Vendetti,

Wu, and Holyoak, 2014) as options for validating measures of systems thinking.

Design thinking research studies many aspects of design thinking, including

psychological and cognitive processes of design (Dinar et al. 2015; Plattner, Meinel, and Leifer,

2011; Shah, 2012; Toh and Miller, 2016). These processes include creativity, divergent thinking,

analogical reasoning, sketching, prototyping, and visual representation, among others. Case

studies, think-aloud protocols, controlled experiments, psychometric measurement, and, more

recently, physiological measurement techniques are used to study these processes (Dinar et al.,

2015).

101

4.3 Pilot Validation Study

4.3.1 Overview and Objectives

This study represents the first attempt at understanding the relationship between systems

design thinking, behavior, and related psychological constructs. The goal of the study is to

validate the Systems Design Thinking Scale by identifying correlations between scale scores and

observable behaviors. This is the first step in operationalizing the Systems Design Thinking

Scale.

4.3.2 Methods

In this study, the Systems Engineering and Design Scale from Chapter 3 was delivered to

participants on Reddit, along with analytical reasoning and divergent thinking tasks. The

analytical reasoning tasks used in this study were adapted from Frederick’s Cognitive Reflection

Test (2005). This task is designed to measure a person’s tendency to override an incorrect “gut”

response and engage in higher-level analysis to find the correct answer to a problem. This task is

taken here to represent the analytical and mathematical systems engineering approach to problem

solving, versus the more intuitive approach of design. Design thinking is measured by a classical

test of divergent thinking (Guilford et al., 1960). Divergent thinking is an integral process for

generating many possible solutions in creative problem solving. Linear regression modeling is

used to analyze the correlation between scores on the Systems Design Thinking Scale and the

number of correct responses and unique responses on the analytical reasoning and divergent

thinking tasks.

102

4.3.3 Behavioral Task Selection

Systems Engineering Tasks

Problem solving lies at the core of engineering practice and education alike. Word

problems are a traditional instructional mechanism for learning how to apply mathematics to

solving problems in the educational setting (Salado, Chowdhury, and Norton, 2018). In this

study, word problems are used to explore the relationship between scores on the Systems Design

Thinking Scale, analytical reasoning, and mathematical problem solving. Participants complete

three short word problems, based on the cognitive reflection test by Frederick (2005). The

cognitive reflection test is a task designed to measure a person’s tendency to override an

incorrect “gut” response and engage in further analysis to find the correct answer to a problem.

It is expected that systems engineers are more likely to activate what Frederick and others

describe as “System 2”: a deliberate and analytical cognitive process. System 1 describes an

intuitive, immediate response that is executed quickly and without reflection. As designers are

often categorized as more intuitive and less analytical, we expect designers to be more likely to

answer these questions more intuitively, while systems engineers are more likely to apply an

analytical process for writing mathematical formulas and finding the correct solution. The

following questions are used for the analytical reasoning task in this study:

• If Rhonda is the 50th fastest and 50th slowest runner in her school, how many

students are there in her school? [Correct answer: 99; Intuitive answer: 100]

• You have 100 pounds of potatoes, which are 99% water by weight. You let them

dehydrate until they are 98% water by weight. Now how much do they weigh?

[Correct answer: 50 pounds; Intuitive answer: 99 pounds]

103

• At a party, everyone shook hands with everybody else. There were 6 handshakes.

How many people were at the party? [Correct answer: 4 people; Intuitive answer:

5 people]

Performance on these tasks is measured by the number of correct responses. Correlation is

expected between systems engineering subscale scores and number of responses. Individuals

high in systems engineering subscale scores are expected to generate more correct responses on

these questions.

Design Thinking Tasks

Scores on the design thinking subscale should correlate with design thinking behaviors.

In systems engineering, these include creating and generating divergent opinions and ideas,

encouraging differing opinions, driving convergence on decisions, and others (Williams and

Derro, 2008). Studies in systems thinking, which shares commonalities with design thinking

described in Chapter 3, have explored the relationship between systems thinking and creativity.

One study uses a simple measure of creativity in which participants are presented with three

hypothetical situations and are asked to list as many consequences of the situations as possible

(Randle and Stroink, 2018; Furnham and Nederstrom, 2010). Participants who scored higher on

Randle and Stroink’s Systems Thinking Scale tended to generate more creative responses on the

consequence measure.

In this study, a classic measure of divergent thinking (Guilford et al., 1960) is used to

evaluate the design thinking subscale of the Systems Design Thinking Scale. Divergent thinking

is the cognitive process used to generate creative ideas by exploring many possible solutions

(Furnham and Nederstrom, 2010). In this study, participants were asked to generate as many uses

as they could think of for two common household items, a newspaper and a coffee mug.

104

4.3.4 Study Population and Recruitment Strategy

In this study, the process of recruiting from Reddit was refined. The research strategy

now included the best time of day to post, based on the analysis by Candocia (2017). A similar

process was applied to timing postings on Facebook and LinkedIn as well (Hootsuite, 2018;

Kolowich, 2019)

The best time to post to Reddit was estimated to be 8:00 AM EST (Candocia, 2017); for

this study, the process began at 6 AM EST because there is a 10-minute delay to posting on

multiple subreddits and eleven total subreddits were targeted. This is also to allow for resolution

of mistakes that result in auto-moderators removing posts. Posts are made on the following

subreddits: r/architecture; r/designthought; r/ComputerEngineering; r/aerospaceengineering;

r/chemicalenigneering; r/SoftwareEngineering; r/MechanicalEngineering; r/userexperience;

r/ElectricalEngineering; r/productdesign; and r/EngineeringStudents. The posts were monitored

for 24 hours, during which time researchers were interacting with participants via forum

comments to answer questions and increase the visibility of the post.

4.3.5 Pilot Test, Factor Analysis, and Results for Validation Study

A pilot study was conducted on Reddit using Qualtrics. This study enabled testing of the

Systems Design Thinking Scale on a new population, and a test for correlation between the scale

and analytical reasoning and divergent thinking tasks. 136 observations were recorded. Model

results are reported in Table 4.1 and summarized below.

105

Table 4.1 Model results from confirmatory factor analysis (validation study)

SYSENG by Estimate S.E. Est./S.E. P-value

SE5 0.672 0.039 18.035 0.000

SE6 0.713 0.040 15.787 0.000

SE7 0.588 0.042 14.449 0.000

SE8 0.577 0.043 12.959 0.000

DESIGN by Estimate S.E. Est./S.E. P-value

DT5 0.460 0.047 8.967 0.000

DT6 0.575 0.044 11.594 0.000

DT7 0.466 0.044 11.559 0.000

DT9 0.450 0.041 14.355 0.000

DT10 0.379 0.047 9.271 0.000

ꭓ2 Df P RMSEA CFI/TLI SRMR

55.629 42 0.077 0.049 0.918/0.893 0.067

Findings suggest that the specified model is a good fit to the data for the new population.

Factor loading coefficients in the validation study are similar to previous studies in which items

were selected. However, coefficients for DT7, 9, and 10 are slightly lower in the validation study

than in Study 3, as indicated in Table 4.2. The p-value for these items are less than .001 in both

models. Factors and task variables are independent and do not vary together, as indicated in

Table 4.3.

Table 4.2 Comparison in factor loadings between validation study and Study 3 for items DT7,

DT9, and DT10

Item Study 3 Loading

Validation

Study Loading Δ

DT7 0.575 0.475 0.100

DT9 0.585 0.442 0.143

DT10 0.586 0.383 0.203

106

Table 4.3 Covariances of factors and tasks

Estimate Std. Err z-value P(>|z|) Std.lv Std.all

SYSENG ~~

on DESIGN -0.012 0.031 -0.396 0.692 -0.051 -0.051

Analytical Reasoning ~~

Divergent Thinking 1.142 0.893 1.278 0.201 1.142 0.115

Participants were given scores on both the analytical reasoning and divergent thinking

tasks. The tasks were then treated as observed variables in the model and regressed on the

systems engineering and design thinking factors. The mean score for the analytical reasoning

task is 1.686 correct answers (out of 3 possible correct answers), with a standard deviation of

1.034. The mean score for the divergent thinking task is 13.730 unique answers, with a standard

deviation of 10.376. The regression model results are reported in Table 4.4:

Table 4.4 Regression model results

Estimate Std. Err z-value P(>|z|) Std.lv Std.all

Analytical Reasoning

on SYSENG 0.237 0.188 1.262 0.207 0.13 0.127

on DESIGN -0.05 0.259 -0.193 0.847 -0.022 -0.021

Divergent Thinking

on SYSENG 2.481 1.866 1.33 0.184 1.359 0.132

on DESIGN 6.946 2.988 2.325 0.02 3.063 0.297

We expected scores on the systems engineering subscale to correlate with scores on the

analytical reasoning task. We also expected scores on the design thinking subscale to correlate

with scores on the divergent thinking tasks. We did not expect to see any correlation between

scores on the systems engineering subscale and scores on the divergent thinking task; nor did we

expect to see any correlation between scores on the design thinking subscale and scores on the

analytical reasoning task.

107

Scores on the systems engineering subscale were not correlated with scores on either

task. No correlation was observed between scores on the divergent thinking subscale and

analytical reasoning task as expected. Some correlation (0.288, p<.05) between the design

thinking subscale and divergent thinking measure suggests that the Systems Design Thinking

Scale may be useful for predicting certain behaviors.

Most participants were able to solve the analytical reasoning questions correctly. Most

participants provided some answers to the divergent thinking task, although some indicated that

they did not feel incentivized to complete this task. Others put in considerable effort, generating

as many as 50 unique answers.

4.3.6 Findings and Lessons Learned

No significant correlation was observed between either subscale and performance on the

analytical reasoning task. The mean score on this task was 1.68 correct answers, indicating that

of 3 possible correct answers, most individuals got close to 2 correct. This task may have been

too easy, making it difficult to predict performance using the systems engineering subscale. No

correlation was expected between the analytical reasoning task and design thinking subscale

scores.

Some correlation between the design thinking subscale and divergent thinking measure

was observed. Divergent thinking is a thought process used to generate creative ideas by

exploring a large number of possible solutions. While creativity is a part of design thinking,

design thinking includes many additional key features, such as empathy and human-

centeredness. The low correlation observed in this analysis could be due to the fact that creativity

is only part of design thinking. A multi-dimensional measure that captures more than one feature

of design thinking would be useful for fully validating the scale. Additionally, feedback from

108

participants suggested that the divergent thinking tasks were boring/not engaging and

participants had no incentive to put effort in (participants were not compensated for completing

the study). It is likely that task performance captured participants’ level of interest and

engagement, rather than their divergent thinking abilities.

The lack of desired results is probably also due in part to the artificiality of the tasks and

information provided to the participants more generally. The tasks were one-dimensional and did

not require any interaction, which does not adequately represent systems engineering and design

tasks. The analytical reasoning and divergent thinking tasks were not complex, little information

was given/required, and skills for those tasks are different from the skills required for systems

design. Similarly, artificiality of the incentives and environment likely contributed to the lack of

correlation. These tasks did not include similar incentive conditions to most systems engineering

and design tasks. In this experimental setting, subjects were not rewarded for high performance,

nor were they penalized for poor performance. In a real systems engineering environment, cost

of poor performance can be very high.

Construct validity seems to be the major issue with the pilot study. Operationalized

measures need to proxy the measures of interest in validation studies. In this case, the

operationalized measures—the analytical reasoning and divergent thinking tasks—do not seem

to accurately proxy systems design thinking behavior. A systems problem would be a better

choice for future work, and the problem should depend on the specific aspect of systems

engineering being investigated, such as problem partitioning, concept generation, decision-

making, etc. (Panchal and Szajnfarber, 2017).

109

4.4 Validation Opportunities

Like the work by Thibodeau, Frantz, and Stroink (2016), this work attempted to situate

the Systems Design Thinking Scale in the landscape of existing psychological constructs and

measurement instruments. An attempt was made to study systems design thinking relative to

analytical reasoning and divergent thinking. Findings did not demonstrate any correlation

between scores on the systems engineering subscale and the selected tasks. Findings suggested

some correlation between scores on the design thinking subscale and divergent thinking task.

Additional findings suggest that a more interactive and engaging, systems-level task

would be more appropriate for studying complex behaviors like systems design thinking. This

section suggests additional opportunities for validation. The goal is to still be able to use Reddit

as a way of recruiting large number of participants quickly. An interactive and engaging task that

can be distributed to a large number of people online is needed.

Gamification is a good approach for doing this. According to Farber (2017), “all games

are systems: the rules (or constraints), components, space, and goal interconnect. A game’s

interconnected system is driven by player actions; as players learn a game, they also learn its

system.” Games encourage players to think about relationships; not isolated events, facts, and

skills (Goodwin and Franklin, 1994). Games like Rise of Nations, for example, require players

to think of how each action taken might impact their future actions and the actions of the other

players playing against them. Similarly, the city management series SimCity tasks players with

balancing a complex urban system. Plague, Inc., also requires systems thinking. These games can

be used to teach systems thinking, and could also be used to measure it.

Similarly, “play is essential to the design thinking process,” and “a playful and

exploratory attitude leads to move innovative, competitive, and breakthrough ideas (Silvers,

110

2016).” There is a great deal of academic research describing the value of play and its

importance not just to childhood development, but to adult life. Play, games, and the principles

that underlie them have vital roles in “building critical skills like systems thinking, creative

problem solving, collaboration, empathy and innovation,” according to the National Institute of

Play. Gamification could also be used to improve interest and effort on design thinking tasks.

4.5 Summary

In this chapter, a study for validating the Systems Design Thinking Scale was described.

Scale scores were compared with performance on analytical reasoning and divergent thinking

tasks. While no correlations were observed between subscales and performance on the analytical

reasoning task, some correlation was observed between scores on the design thinking subscale

and performance on the divergent thinking task. This suggests that the Systems Design Thinking

Scale is useful for measuring certain behaviors, but additional work is required to find a suitable

multi-dimensional measure for validating the scale fully.

Because the Systems Design Thinking Scale was only partially validated, some ideas for

additional validation were offered. We suggest validating the Systems Design Thinking Scale

through gamification, which will present more realistic environments, tasks, and incentives.

Several games were identified for their potential usefulness in validating the scale, and

performance metrics for these games were also identified. Correlation between these

performance metrics and subscale scores will be useful for further validating the scale.

111

CHAPTER V

Conclusion

5.1 Summary of Dissertation

This dissertation explored systems engineering, systems thinking, design thinking, and

their relationships using a mixed methods approach. Qualitative analysis was used to identify key

assumptions, concepts, values, and practices for each framework, and to identify individual

attitudes that reflect each of these frameworks. Quantitative analysis was used to develop an

instrument for measuring these attitudes, called the Systems Design Thinking Scale.

The scale was developed in three major iterations. The first iteration captured systems

thinking attitudes along three dimensions: technical, social, and organizational. This hypothesis

was based off interview findings and tested. While technical and social items grouped together in

meaningful ways, organizational items did not appear to follow any clear pattern. We explain

this finding by defining systems thinking itself as an organizational framework. Systems thinking

is useful for organizing both technical and social system elements.

“Social systems thinking” in Study 1 demonstrated many similarities with design

thinking, when the codebook in Chapter II was used to interpret findings. The relationship

between systems thinking and design thinking was explored further in Study 2. Systems thinking

and design thinking were not clearly distinguishable. However, interesting patterns were

discovered in the systems thinking factor. Significant items appeared to reflect systems

engineering more than systems thinking.

112

In Study 3, the relationship between systems engineering and design thinking was

explored further. This model was a good fit to the data. While the scale functioned well

theoretically, a goal of the work was to develop an instrument with practical use. We attempted

to link scale scores with performance outcomes on analytical reasoning and divergent thinking

tasks in Chapter 4. While some correlation was observed between divergent thinking subscale

scores and performance on divergent thinking tasks, additional work is required before the

Systems Design Thinking Scale can be considered as fully validated.

Several important lessons were learned throughout the research process. First, the

application of psychometrics within the domain of systems engineering represented a unique

challenge, as few precedents existed for identifying and measuring many relevant concepts.

Consistent qualitative and quantitative data collection and analysis were required for interpreting

findings and advancing hypotheses throughout the dissertation. Methodological lessons were

related to the use of reverse-coding techniques, as well as survey design and user experience.

5.2 Contributions to Design Science

The main contribution of this work to the discipline of Design Science is the development

and demonstration of an integrated framework of systems engineering, systems thinking, and

design thinking we call “systems design thinking.” Systems design thinking describes systems

engineers who use a human-centered approach to complex systems design. Systems design

thinking also describes product designers who have a natural inclination to follow systematic,

analytical processes during design. The work presented in this dissertation suggests that the

analytical and systematic attitudes of engineering and the holistic and human-centered attitudes

of design can coexist, contrary to existing stereotypes that suggest an individual can only have

113

one set of attitudes or the other. The use of psychometric methods to make this claim is a novel

approach.

Another major contribution of this work is the Systems Design Thinking Scale. The

Systems Design Thinking Scale is a 9-item questionnaire that measures systems engineering and

design thinking attitudes in two subscales. In this work, we demonstrate that this model of

systems design thinking is a good fit to two different samples. Additional work is required to

demonstrate the usefulness of the scale in practice, as described in Section 5.3; however, it is

believed that the scale could be useful for categorizing individuals based on their attitudes, with

the ultimate goal of developing effective management strategies, for example, in systems design

teams.

In this work, psychometric methods were used to solve engineering problems. It is

difficult for organizations to identify systems design thinkers, despite their known value. By

making theory and methods from psychology available for use in systems engineering

organizations, we hope to offer an innovative method for identifying valuable individuals within

an organization, allowing for the maximization of human potential.

5.3 Limitations and Opportunities for Future Work

Several limitations exist in the current study. First, throughout the research, items were

dropped from the Systems Design Thinking scale in order to converge on a solution. Revisiting

certain concepts, items and potential factor structures would be useful for building a more

representative model. A good example of this occurs during the scale development process in

Study 3. In the exploratory factor analysis, systems engineering items SE1 & SE2 group together

in a single factor statistically and theoretically. Both items ask about preferences for

requirements, a core concept in systems engineering. However, because conventions of structural

114

equation modeling suggest that a factor should consist of a minimum of three items, these two

items were dropped from the final model. The same happens for design thinking items DT2 and

DT3, which ask about preferences for working with customers, reflecting the relationship-driven

practice of design thinking. In future work, these items should be reintroduced into the final

model, with additional test items included to explore these concepts further. Also, it would be

worth investigating small differences in phrasing of the items and the impact on results. In the

systems engineering subscale, the attitude items are heavily behavior focused. Examples include

“I use quantitative methods” and “I use mathematical modeling/analysis.” Few affective or

cognitive items were included (e.g., I prefer X; I think X). For the behavioral items as written,

frequency is another metric worth exploring (i.e., I use quantitative methods

frequently/infrequently rather than agree/disagree).

More work is needed to validate the Systems Design Thinking Scale, and to determine

that the scale is actually measuring systems design thinking. The scale could be capturing

differences between creative and analytical thinkers more generally, for example. There is

evidence in neuroscience suggesting this difference is real and due to differences in neural

activity that can be observed even when people are not working on a problem (Erickson et al.,

2018).

A related scale is the Rational and Intuitive Decision Styles Scale (Epstein et al. 1996;

Hamilton, Shih, and Mohammed, 2014). Our scale appears different, as it is context-rich and

specifically focused on quantitative methods, data, and analysis, instead of a broad “rational”

approach to problem solving (e.g., “I gather all necessary information” & “I thoroughly evaluate

alternatives.”) Also, while design thinking includes intuition, design thinking is not totally

intuitive. Design thinking is also creative (involves restructuring problems, trying new

115

approaches, and finding inspiration in everyday life) and social (involves empathy, codesign, and

collaboration). An intuitive item was included in pilot testing (“I make [design] decisions based

on intuition”), but was dropped in final scale due to low loadings. Another related scale is the

Systems Thinking Scale by Randle and Stroink (2018). Additional work that compares the

Systems Design Thinking Scale to this and other psychological constructs would be useful for

understanding what makes systems design thinking unique.

It is anticipated that the Systems Design Thinking Scale will be useful for identifying

individual strengths in practice after full validation is achieved. This could be useful for building

teams with desired skill compositions. The scale could then also be used as a way of

individualizing training and skill-building programs based on interests, preferences, and abilities.

The games described in the validation plan could be used for both measurement and training

purposes, to record and improve an individual’s performance over time.

116

Appendices

A. Systems Design Thinking Codebook…………………………………………………...117

B. Semi-Structured Interview Questions.………………………………………………….122

117

Appendix A: Systems Design Thinking Codebook

Name Files References

Analysts 3 4

Collaboration 4 5

Design Thinking 5 8

Awareness&Intuition 8 35

Creative&Innovative 5 14

Brainstorming 2 3

Mental blocks 1 1

Curiosity 3 18

Human-centered 10 40

Communication 10 96

Face to face 6 10

Formal&Informal 6 11

Hearing&listening 8 21

118

Meetings 9 23

Questions 7 38

Culture 8 23

Empathy&Understanding 10 42

Experience 8 39

People&Personalities 8 45

Conflict 4 5

Relationships 3 9

Team 2 3

Trust 5 14

Insight 2 3

Problem definition 5 9

Prototyping 0 0

Storytelling&analogy 8 13

Visual 1 1

Leadership 5 9

Leadership&techskills 1 1

119

Learning&Information 10 29

Mediate&Negotiate 8 21

Observation&Participation 2 8

Organization 4 7

Colocation 2 3

Proactive 2 5

Problem solving 5 11

Alternative methods and

approaches

2 6

Ambiguous problems &

solutions

0 0

Unambiguous problems &

solutions

3 3

SE v DTorDiscipline 7 21

Systems Engineering 8 31

Analysis 6 15

Data 7 13

Tools&Methods 2 4

120

Trade studies 2 3

Coordinate 5 21

Cost 6 12

Delegate&Assign 6 14

Document 7 42

Elements&Subsystems 2 3

Goals&objectives 2 4

Interface 7 16

Managing 8 23

Methodical (process) 6 26

MinimizeReducePartition 6 15

Organize 2 4

Planning&Scheduling 6 29

Requirements 7 25

Risk 7 17

Robust 2 3

Strategy 7 14

121

Technical 8 18

Systems Thinking 7 17

AmbiguityUncertainty 8 17

Big picture 9 53

Complexity 10 30

Flexibility&adaptability 8 26

Holistic 7 12

Integrate&align 9 56

Interactions 10 56

Large-scale 8 18

Synergy 2 3

Training 6 8

Education 5 10

Mentorship 4 10

Translate 4 7

Understanding details 6 9

122

Appendix B: Semi-Structured Interview Questions

Please consider your first-hand experiences with designing and maintaining large-scale, complex

engineered systems.

1. Please provide some background context for your experience. Where do you work? What is

your formal title? How many years of work experience do you have? How many years have you

had your current job?

2. Walk me through a typical workday. In a typical week, what are the top 3-5 tasks you spend

the most time on?

3. Which of the following would you say has contributed most to your ability to complete these

tasks/do your job effectively: formal education, on-the-job training, mentoring, or something

else? Could you elaborate?

4. Please describe a specific project in which you participated in the design and management of a

complex system. Can you draw the general architecture of the system? Using this sketch, can

you tell me which groups, teams, or divisions within the organization work on which parts of the

technical system? Is this consistent with the original work breakdown structure for the project?

How many engineers were involved in the project? How many engineers were in each group you

drew in the sketch? Where was each group physically located?

5. How did you arrive at this partitioning (in the sketch)? Is there a typical partitioning common

to your organization, or is each project broken down differently? Can you describe it? Is this

partitioning reflected in the structure of your organization? Who (what title) is responsible for

123

deciding how the work gets done/what the subsystems are? Can you describe how these

decisions are made? How are the design teams selected? Are you directly involved in making

these decisions?

5a. (If participant makes partitioning decisions): Generally speaking, what heuristics or processes

did you use to decide how to distribute the work for this project? At what stage did you finalize

this breakdown? What information did you have available to you at the time you were making

decisions about how to break down the project? Did you use all of the information available to

you when making decisions about how best to distribute the work?

5b. (If participant does not make partitioning decisions): How was this work breakdown structure

presented or communicated to you? Who delivered it to you? Do you feel that this was the best

way to break down the project/distribute the work? To communicate the project structure? Why

or why not? Would you have broken things down differently? How? Why? Would you have

communicated this information differently?

6. Going back to your sketch, what was your role in the project? Did your role change

throughout the design of this system? Can you indicate which subsystem(s) you worked on, and

their relationship to the other systems in the project? How would you characterize your work on

these subsystems (design, interface management, analysis, something else)? Which of the groups

you indicated did you feel you “belonged” to?

7. Which groups in the sketches often found your work relevant to theirs? Which groups in the

sketches rarely found your work relevant? How did you know?

8 Which groups did you work with frequently (e.g., several times per week)? Infrequently (e.g., a

few times per month)? How would you characterize your interactions with these groups, e.g.

requesting information, or providing information? Would you characterize these interactions as

124

primarily formal or informal? Did you routinely anticipate any requests for information from

other groups or the need to provide information to other groups? How do you work this into your

personal process?

9. Did subsystems communicate with one another consistently throughout the design process

(early conceptual stages through to final design)? Which ones? Why? Did patterns of

communication change during different design stages? How? Is there a single person or group

with which all subsystems regularly communicated? What is the role of direct communication

between groups as compared to communication with this central individual/group? Do these

interactions tend to be through scheduled meetings or informal conversation? Something else?

10. What methods of communication does your organization use (email, meetings/face-to-

face/documents/coffee breaks/other)? What role do each of these communication methods have?

Do you find that the methods of communication you mentioned promote or inhibit your ability to

do your job? In what way? How could these communication channels be improved? Is there a

common technology or software that you use to keep track of documentation or host meetings?

Do you feel that this technology or software is useful/effective? Why/why not? Is there

something you would do differently, or another tool that you would use?

11. At what stage in the design process do systems engineers/integrators attempt to coordinate

the design of the subsystems? What about the integration of the subsystems? Can you describe

how this was done in this project? Who was involved? What information was available to [the

systems engineer or systems integrator] in each of these cases? How was this information

presented to the systems engineer? Was this information presented to design groups?

12. At any stage in system design or subsystem coordination, were you uncertain about either the

reliability or the relevance of the information that you had available? At any stage, were you

125

uncertain about the appropriateness of the decisions you made based on this information? How

did you handle this situation?

13. Was there any stage during the system design process in which you found it difficult to

process and integrate the information available? Describe precisely the nature of the situation.

14. Were you reminded of similar experiences/projects at any point during your work on this

project? Were you at any point reminded of different experiences/projects? Were you at any

point reminded of a project that succeeded? Were you at any point reminded of a project that

failed? Did these experiences affect the decisions you made or actions that you took? How?

15. Do you think that you could develop a rule, based on your experience, which could assist

another person to make the same design decisions successfully? Why/why not? What advice

would you give to someone new to the role you had on this project?

Is there anything I might have missed? Do you have any other thoughts about systems design

that you’d like to share?

126

Bibliography

Allison, J. T. (2008). Optimal Partitioning and Coordination Decisions in Decomposition-based

Design Optimization (Doctoral dissertation). University of Michigan, Ann Arbor, USA.

Retrieved from ResearchGate database.

Allport, G. (1935). "Attitudes," in A Handbook of Social Psychology, ed. C. Murchison.

Worcester, MA: Clark University Press, 789–844.

Altus, S.S., Kroo, I.M., and Gage, P.J. (1996). A genetic algorithm for scheduling and

decomposition of multidisciplinary design problems. Trans. ASME (118), 486-489.

Alyaqout, S. F., Peters, D. L., Papalambros, P. Y., and Ulsoy, A.G. (2011). Generalized coupling

management in complex engineering systems optimization. Journal of Mechanical Design,

133(9), 1864-1869.

Amazon. (2018). Amazon Mechanical Turk. Retrieved April 3, 2019 from

https://www.mturk.com/.

American Heritage Dictionary. (2016). “Framework.” Retrieved April 1, 2019 from

https://ahdictionary.com/word/search.html?q=framework.

Anderson, N., Potočnik, K., & Zhou, J. (2014). Innovation and creativity in organizations: A

state-of-the-science review, prospective commentary, and guiding framework. Journal of

Management, 40(5), 1297–1333.

Ashby, W. R. (1965). An Introduction to Cybernetics. Chapman & Hall.

Avnet, M. S. (2016). A network-based analysis of team coordination and shared cognition in

systems engineering. Systems Engineering and Electronics, 19(5), 395–408.

Baines, T. S., Lightfoot, H. W., Evans, S., Neely, A., Greenough, R., Peppard, J., … Wilson, H.

(2007). State-of-the-art in product-service systems. In Proceedings of the Institution of

Mechanical Engineers, Part B: Journal of Engineering Manufacture, 221(10), 1543–1552.

Banathy, B. H. (1967). The systems approach. The Modern Language Journal, 51(5), 281-289.

127

Barker, R. G. (1968). Ecological Psychology: Concepts and Methods for Studying the

Environment of Human Behavior. Stanford, CA.: Stanford University Press.

Bateson, G. (1972). Steps to an Ecology of Mind: Collected Essays in Anthropology, Psychiatry,

Evolution, and Epistemology. University of Chicago Press.

Beasley, R., and Partridge, R. (2011). The three T’s of systems engineering – trading, tailoring,

and thinking. In Proceedings of the 21st Annual Symposium of the International Council on

Systems Engineering (INCOSE). Denver, CO, USA. June 20-23, 2011.

Birman, I. (2018). Moderation in Different Communities on Reddit – A Qualitative Analysis

Study (Undergraduate thesis). Georgia Institute of Technology, Atlanta, GA, USA. Retrieved

from Georgia Institute of Technology SMARTech Repository.

Boulding, K. E. (1964). The Meaning of the Twentieth Century: the Great Transition. Harper &

Row.

Braha, D. and Bar-Yam, Y. (2007). The statistical mechanics of complex product development:

Empirical and analytical results. Management Science, 53(7), 1127-1145.

Braun, V., & Clarke, V. (2006). Using thematic analysis in psychology. Qualitative Research in

Psychology, 3(2), 77–101.

Brooks, J. M., Carroll, J. S., & Beard, J. W. (2011). Dueling stakeholders and dual-hatted

systems engineers: Engineering challenges, capabilities, and skills in government infrastructure

technology projects. IEEE Transactions on Engineering Management, 58(3), 589–601.

Brown, T. (2008). Design thinking. Harvard Business Review, 86(6), 84–92, 141.

Brown, T. and Katz, B. (2011). Change by design. The Journal of Product Innovation

Management, 28(3), 381-383.

Browning, T. (2001). Applying the design structure matrix to system decomposition and

integration problems: A review and new directions. IEEE Transactions on Engineering

Management, 48(3), 292-306.

Buchanan, R. (1992). Wicked Problems in Design Thinking. Design Issues, 8(2), 5–21.

Buede, D.M. (2009). The Engineering Design of Systems: Models and Methods, Second Edition.

John Wiley & Sons.

Candocia, M. (2017). What Time Should You Post to Reddit (Part 2). Retrieved April 4, 2019,

128

from https://maxcandocia.com/article/2017/Oct/12/what-time-should-you-post-to-reddit-pt-2/

Cagan, J. (2007). The cognition of engineering design—An opportunity of impact. Cognitive

Science, 31, 193–195.

Castelle, K. M., & Jaradat, R. M. (2016). Development of an instrument to assess capacity for

systems thinking. Procedia Computer Science, 95, 80–86.

Cataldo, M., Herbsleb, J., & Carley, K. (2008). Socio-technical congruence: A framework for

assessing the impact of technical and work dependencies on software development productivity.

In Proceedings of the 2nd International Symposium on Empirical Software Engineering and

Measurement, Kaiserslautern, Germany, Oct. 9-10, 2008.

Cavalieri, S., & Pezzotta, G. (2012). Product–service systems engineering: State of the art and

research challenges. Computers in Industry, 63(4), 278–288.

Checkland, P.B. (1981). Systems Thinking, Systems Practice. John Wiley & Sons Ltd., West

Sussex, England, UK.

Checkland, P.B. and Scholes, J. (1990). Soft Systems Methodology in Action. John Wiley &

Sons, West Sussex, England, UK.

Chesson, D. (2017). Design Thinker Profile: Creating and Validating a Scale for Measuring

Design Thinking Capabilities (Doctoral dissertation). Antioch University, Yellow Springs, OH.

Chiu, C. Y., Morris, M. W., Hong, Y. Y., & Menon, T. (2000). Motivated cultural cognition: the

impact of implicit cultural theories on dispositional attribution varies as a function of need for

closure. Journal of Personality and Social Psychology, 78(2), 247–259.

Choi, I., Koo, M., & Jong An Choi. (2007). Individual differences in analytic versus holistic

thinking. Personality & Social Psychology Bulletin, 33(5), 691–705.

Clegg, S. and Bailey, J.R. (2008). International Encyclopedia of Organization Studies. Sage

Publications.

Colfer, L.J. and Baldwin, C.Y. (2016). The mirroring hypothesis: Theory, evidence, and

exceptions. Industrial and Corporate Change, 25(5), 709-738.

Collopy, A.X. (2019). Coordination Strategies and Individual Behavior in Complex Engineered

Systems Design (Doctoral dissertation). University of Michigan, Ann Arbor, USA.

Conner, M. (2015). Systems Engineering and Integration. Retrieved from

http://www.nasa.gov/centers/armstrong/capabilities/CodeR/flight/systems_engineering.html

129

Conway, M.E. (1968). How do committees invent? Datamation, 14(5), 28-31.

Crede, E., & Borrego, M. (2013). From ethnography to items: A mixed methods approach to

developing a survey to examine graduate engineering student retention. Journal of Mixed

Methods Research, 7(1), 62–80.

Cross, N. (1982). Designerly ways of knowing. Design Studies, 3(4), 221–227.

Cross, N. (2001). Designerly ways of knowing: Design discipline versus design science. Design

Issues, 17(3), 49–55.

Cumming, M. (2002). Flexible and distributed coordination models for collaborative design. In

Proceedings of the 20th eCAADe Conference, Warsaw, Poland.

Dam, R., & Siang, T. (2019). 5 Stages in the Design Thinking Process. Retrieved April 9, 2019,

from https://www.interaction-design.org/literature/article/5-stages-in-the-design-thinking-

process

Darrin, M. A. G., & Devereux, W. S. (2017). The agile manifesto, design thinking and systems

engineering. In Proceedings of the 2017 Annual IEEE International Systems Conference

(SysCon), 1–5.

Davidz, H. L., & Nightingale, D. J. (2008). Enabling systems thinking to accelerate the

development of senior systems engineers. Systems Engineering, 11(1), 1–14.

Davidz, H. L., Nightingale, D. J., & Rhodes, D. H. (2004). Enablers, barriers, and precursors to

systems thinking development: The urgent need for more information. In Proceedings of the

2004 Conference on Systems Engineering Research, Los Angeles, CA.

Davis, A. C., Leppanen, W., Mularczyk, K. P., Bedard, T., & Stroink, M. L. (2018). Systems

thinkers express an elevated capacity for the allocentric components of cognitive and affective

empathy: Systems thinking and empathy. Systems Research: The Official Journal of the

International Federation for Systems Research, 35(2), 216–229.

Davis, A. C., & Stroink, M. L. (2016). The relationship between systems thinking and the new

ecological paradigm: Systems thinking and environmental worldview. Systems Research: The

Official Journal of the International Federation for Systems Research, 33(4), 575–586.

DeCuir-Gunby, J.T., Marshall, P.L., and McCulloch, A.W. (2011). Developing and using a

codebook for the analysis of interview data: An example from a professional development

research project. Field Methods, 23(2), 136-155.

de Graaf, R. S., & Loonen, M. L. A. (2018). Exploring team effectiveness in systems engineering

construction projects: Explanations why some SE teams are more effective than others. Systems

Research: The Official Journal of the International Federation for Systems Research, 35(6),

687–702.

130

de Souza, C. R., Quirk, S., Trainer, E., & Redmiles, D. F. (2007). Supporting collaborative

software development through the visualization of socio-technical dependencies. In Proceedings

of the 2007 International ACM Conference on Supporting Group Work. New York, NY, USA.

DeFranco, J. F., Neill, C. J., & Clariana, R. B. (2011). A cognitive collaborative model to

improve performance in engineering teams-A study of team outcomes and mental model sharing.

Systems Engineering, 14(3), 267–278.

Dinar, M., Shah, J. J., Cagan, J., Leifer, L., Linsey, J., Smith, S. M., & Hernandez, N. V. (2015).

Empirical studies of designer thinking: Past, present, and future. Journal of Mechanical Design,

137(2), 021101.

Doob, L. W. (1947). The behavior of attitudes. Psychological Review, 54(3), 135–156.

Dosi, C., Rosati, F., and Vignoli, M. (2018). Measuring design thinking mindset. In Proceedings

of the DESIGN 2018 15th International Design Conference. Dubrovnik, Croatia, May 21-24,

2018.

Dym, C. L., Agogino, A. M., Eris, O., Frey, D. D., & Leifer, L. J. (2005). Engineering design

thinking, teaching, and learning. Journal of Engineering Education, 94(1), 103–120.

Eagly, A., & Chaiken, S. (1995). Attitude strength, attitude structure and resistance to change. In

R. Petty and J. Kosnik (Eds.), Attitude Strength. 413–432. Mahwah, NJ: Erlbaum.

Elsbach, K. D., Barr, P. S., & Hargadon, A. B. (2005). Identifying situated cognition in

organizations. Organization Science, 16(4), 422-433.

Eppinger, S.D. and Browning, T.R. (2012). Design Structure Matrix Methods and Applications.

MIT Press, Boston, MA, USA.

Epstein, S., Pacini, R., Denes-Raj, V., & Heier, H. (1996). Individual differences in intuitive-

experiential and analytical-rational thinking styles. Journal of Personality and Social

Psychology, 71(2), 390–405.

Farber, M. (2017). Games, hacking, and the 21st century skill of systems thinking. [Blog post].

Retrieved February 12, 2019 from http://info.thinkfun.com/stem-education/blog/games-hacking-

and-the-21st-century-skill-of-systems-thinking.

Fleury, A., Stabile, H., and Carvalho, M. (2016). An overview of the literature on design

thinking: Trends and contributions. International Journal of Engineering Education, 32(4),

1704-1718.

Forrester, J. W. (1994). System dynamics, systems thinking, and soft OR. System Dynamics

Review, 10(2-3), 245–256.

131

Erickson, B., Truelove-Hill, M., Oh, Y., Anderson, J., Zhang, F.Z., & Kounios, J. (2018).

Resting-state brain oscillations predict trait-like cognitive styles. Neuropsychologia, 120, 1-8.

Forrester, J. (1961). Industrial Dynamics. Pegasus Communications, Waltham, MA.

Forrester, J. (1969). Urban Dynamics. Pegasus Communications, Waltham, MA.

Forrester, J. (1994). System dynamics, systems thinking, and soft OR. System Dynamics Review,

10(2), 1-14.

Forsberg, K., & Mooz, H. (1992). The relationship of systems engineering to the project cycle.

Engineering Management Journal, 4(3), 36–43.

Frank, M. (2000). Engineering systems thinking and systems thinking. Systems Engineering,

3(3), 163–168.

Frank, M. (2002). Characteristics of engineering systems thinking - a 3D approach for

curriculum content. IEEE Transactions on Systems, Man and Cybernetics. Part C, Applications

and Reviews: A Publication of the IEEE Systems, Man, and Cybernetics Society, 32(3), 203–214.

Frank, M. (2006). Knowledge, abilities, cognitive characteristics and behavioral competences of

engineers with high capacity for engineering systems thinking (CEST). Systems Engineering,

9(2), 91–103.

Frank, M. (2007). Towards a quantitative tool for assessing the capacity for engineering systems

thinking. International Journal of Human Resources Development and Management, 7(3-4),

240–253.

Frank, M. (2012). Engineering systems thinking: Cognitive competencies of successful systems

engineers. Procedia Computer Science, 8, 273–278.

Frank, M., & Kordova, S. (2009). Developing the capacity for engineering systems thinking

(CEST) of senior engineering management students: Learning in a project-based learning (PBL)

environment. In Proceedings of the 7th Annual Conference on Systems Engineering Research,

Loughborough, England.

Frank, M., Sadeh, A., & Ashkenasi, S. (2011). The relationship among systems engineers’

capacity for engineering systems thinking, project types, and project success. Project

Management Journal, 42(5), 31–41.

Frank, M., & Waks, S. (2001). Engineering systems thinking: A multifunctional definition.

Systemic Practice and Action Research, 14(3).

Frank, P. (2006). People Manipulation: A Positive Approach (2 ed.). New Delhi: Sterling

Publishers Pvt. Ltd.

132

Frederick, S. (2005). Cognitive reflection and decision making. The Journal of Economic

Perspectives, 19(4), 25–42.

Fulmer, R.M. and Keys, J. B. (1998). A conversation with Peter Senge: New developments in

organizational learning. Organizational Dynamics, 27(2), 33-42.

Furnham, A., & Nederstrom, M. (2010). Ability, demographic and personality predictors of

creativity. Personality and Individual Differences, 48(8), 957–961.

Galesic, M., & Bosnjak, M. (2009). Effects of questionnaire length on participation and

indicators of response quality in a web survey. Public Opinion Quarterly, 73(2), 349-360.

Gasparini, A. A. (2015). Perspective and use of empathy in design thinking. In Proceedings of

the Eighth International Conference on Advances in Computer-Human Interactions. Lisbon,

Portugal.

Goode, H.H. and Machol, R.E. (1957). System Engineering: An Introduction to the Design of

Large-scale Systems. McGraw-Hill.

Goodwin, J. S., & Franklin, S. G. (1994). The beer distribution game: Using simulation to teach

systems thinking. International Journal of Management & Enterprise Development, 13(8), 7–15.

Grace, M., Smith, D. L., Juhnke, L., & Dalton, S. (2017). Lean by Design: The Synthesis of Lean

and Design Thinking. Boeing.

Greene, M. T., Gonzalez, R., Papalambros, P. Y., and McGowan, A. R., (2017). Design thinking

vs. systems thinking for engineering design: What's the difference? In Proceedings of the 21st

International Conference on Engineering Design, Vancouver, British Columbia, Canada, Aug.

21-25, 2017.

Greene, M.T., and Papalambros, P.Y. (2016). A cognitive framework for engineering systems

thinking. In Proceedings of the 2016 Conference on Systems Engineering Research, Huntsville,

AL, Mar 22-Mar 24, 2016.

Griffin, C. & Bengry-Howell, A. (2017). Ethnography. In Willig, C., & Rogers, W. The SAGE

Handbook of qualitative research in psychology (pp. 38-54). 55 City Road, London: SAGE

Publications Ltd.

Guilford, J.P., Christensen, P.R., Merrifield, P.R., and Wilson, R.C. (1960). Alternative Uses

Manual. Sheridan Supply Co.

Furr, R. M., & Bacharach, V. R. (2013). Psychometrics: An Introduction (Second edition).

SAGE Publications, Inc.

133

Hajela, P., Bloebaum, C.L., and Sobieszczanski-Sobieski, J. (1990). Application of global

sensitivity equations in multidisciplinary aircraft synthesis. Journal of Aircraft, 27(12), 1002-

1010.

Hamilton, K., Shih, S-I, and Mohammed, S. (2014). The development and validation of the

rational and intuitive decision styles scale. Journal of Personality Assessment, 98(5), 523-535.

Hill, C. A., Dean, E., & Murphy, J. (2014). Social Media, Sociality, and Survey Research.

Hoboken, NJ: John Wiley & Sons.

Hoffman, L. (2018). “Multiple-Group Measurement Invariance in CFA using Mplus.” Retrieved

April 9, 2019, from http://www.lesahoffman.com/CLDP948/index.html

Holwell, S.E. (1997). Soft Systems Methodology and its Role in Information Systems (Doctoral

dissertation). Lancaster University, Lancaster, Lancashire, England.

Hootsuite. (2018, March 5). The Best Time to Post on Instagram, Facebook, Twitter, and

LinkedIn. Retrieved April 6, 2019, from https://blog.hootsuite.com/best-time-to-post-on-

facebook-twitter-instagram/

Hu, L., & Bentler, P. , (1999). Cutoff criteria for fit indexes in covariance structure analysis:

conventional criteria versus new alternatives. Structural Equation Modeling, 6, 1-55.

Hutchison, N., Henry, D., & Pyster, A. (2016). Atlas: Understanding what makes systems

engineers effective in the U.S. defense community. Systems Engineering and Electronics, 19(6),

510–521.

Ilhan, A. O. (2017). Growth of undergraduate education in design in the United States, 1988–

2012. Design Issues, 33(4), 17–29.

INCOSE. (2015). INCOSE Systems Engineering Handbook: A Guide for System Life Cycle

Processes and Activities (4th edition). Wiley.

IndustryWeek. (2018). Who Needs Systems Engineering? You Likely Do. Retrieved April 3,

2019, from https://www.industryweek.com/technology-and-iiot/who-needs-systems-engineering-

you-likely-do.

Jaradat, R.M. (2014). An Instrument to Assess Individual Capacity for Engineering Systems

Thinking (Doctoral dissertation). Old Dominion University, Norfolk, VA.

Jones, J. C. (1992). Design Methods. John Wiley & Sons.

Kannan, H., Bloebaum, C. L., and Mesmer, B. (2014). Incorporation of coupling strength models

in decomposition strategies for value-based MDO. In Proceedings of the 15th Annual

AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Atlanta, Georgia, June

16-20, 2014.

134

Katz, D. (1960). The functional approach to the study of attitudes. Public Opinion Quarterly,

24(2), 163–204.

Kerbel, S. and Wainstein, A. (2013). Jung Typology Profiler for Workplace Assessment User

Handbook. Humanmetrics.

Kietzmann, J. (2016). Crowdsourcing: A revised definition and introduction to new research.

Business Horizons, 60(2), 151-153.

Klein, G. and Armstrong, A-A. (2005). Critical decision method. In Handbook of Human Factors

and Ergonomics Methods, 58-1—58-6, Boca Raton, FL:CRC Press.

Klein, G. A., Calderwood, R., & MacGregor, D. (1989). Critical decision method for eliciting

knowledge. IEEE Transactions on Systems, Man, and Cybernetics, 19(3), 462–472.

Klein, G., Moon, B., and Hoffman, R.R. (2006). Making sense of sensemaking. IEEE Intelligent

Systems, 21(4), 70-73.

Kline, R. B. (2015). Principles and Practice of Structural Equation Modeling, Fourth Edition.

The Guilford Press.

Koivu, S. (2015). Why Do People Share Online? Online Disinhibition Effect in the Context of

the Virtual Community of Reddit (Master’s thesis). Aalto University, Finland.

Kolowich, L. (2019). The Best Time to Post on Instagram, Facebook, Twitter, LinkedIn, &

Pinterest. Retrieved April 6, 2019, from https://blog.hubspot.com/marketing/best-times-post-pin-

tweet-social-media-infographic

Kordova, S., Frank, M., & Nissel Miller, A. (2018). Systems thinking education—Seeing the

forest through the trees. Systems, 6(3), 29.

Kuhn, T.S. (1962). The Structure of Scientific Revolutions. University of Chicago Press,

Chicago.

Lake, J. G. (1992). Systems engineering re-energized: Impacts of the revised DoD acquisition

process. Engineering Management Journal, 4(3), September, pp. 8-14.

Lamb, C.T. and Rhodes, D.H. (2008). Systems thinking as an emergent team property: Ongoing

research into the enablers and barriers to team-level systems thinking. In Proceedings of

SysCon2008: IEEE International Systems Conference.

Lamb, C. T., & Rhodes, D. H. (2010). Collaborative systems thinking: Uncovering the rules.

IEEE Aerospace and Electronic Systems Magazine, 25(11), 4–10.

135

Landers, R. N. and Behrend, T.S. (2015). An inconvenient truth: Arbitrary distinctions between

organizational, mechanical turk, and other convenience samples. Industrial and Organizational

Psychology, 8(2), 142-164.

Lasdon, L.S. 1970. Optimization Theory for Large Systems. London: MacMillan

Laszlo, C.A., Levine, M.D., and Milsum, J.H. 1974. A general systems framework for social

systems. Behavioral Science, 19(2), 79-92.

Liedtka, J., & MacLaren, E. (2018, November 7). How Children’s Health System of Texas Is

Improving Care with Design Thinking. Harvard Business Review. Retrieved from

https://hbr.org/2018/11/how-childrens-health-system-of-texas-is-improving-care-with-design-

thinking

Luhmann, N. (1984). Soziale Systeme. Frankfurt/Main: Suhrkamp.

Maddux, W. W., & Yuki, M. (2006). The “ripple effect”: cultural differences in perceptions of

the consequences of events. Personality & Social Psychology Bulletin, 32(5), 669–683.

Madni, A. M. (2015). Expanding stakeholder participation in upfront system engineering through

storytelling in virtual worlds. Systems Engineering, 18(1), 16–27.

Manning, P. K. (2013). Sensemaking. In Encyclopedia of Management Theory. Sage

Publications.

McDermott, T. and Freeman, D. (2016). Systems thinking in the systems engineering process:

New methods and tools. In Systems Thinking: Foundation, Uses, and Challenges. Nova

Publishing.

McGowan, AM. (2014), Interdisciplinary Interactions During R&D and Early Design of Large

Engineered Systems. Doctoral dissertation, University of Michigan, Ann Arbor, MI.

McGowan, A.-M. R., Bakula, C., & Castner, R. S. (2017). Lessons learned from applying design

thinking in a NASA rapid design study in aeronautics. In Proceedings of the 58th

AIAA/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference.

McGowan, A.-M. R., Daly, S., Baker, W., Papalambros, P., & Seifert, C. (2013). A socio-

technical perspective on interdisciplinary interactions during the development of complex

engineered systems. Procedia Computer Science, 16, 1142–1151.

Meadows, D. H. (2008). Thinking in Systems: A Primer. White River Junction, VT: Chelsea

Green Publishing.

136

Mulaik, S. A. (2009). Foundations of Factor Analysis. Boca Raton, FL: Chapman & Hall/CRC.

Muthén, L. K., & Muthén, B. O. (2017). Statistical Analysis With Latent Variables (Version

Mplus User’s Guide. , Eighth Edition). Los Angeles, CA: Muthén & Muthén. Retrieved from

https://www.statmodel.com/download/usersguide/MplusUserGuideVer_8.pdf.

Oliver, D.W., Kelliher, T.P., and Keegan, J. G. (1997). Engineering Complex Systems with

Models and Objects. McGraw-Hill.

Panchal, J. H., & Szajnfarber, Z. (2017). Experiments in systems engineering and design

research. Systems Engineering and Electronics, 20(6), 529–541.

Panel on Undergraduate Engineering Education. (1986). Engineering Undergraduate Education.

Washington, D.C., United States: National Academies Press.

Papalambros, P. Y. (2018). From design optimization to design science: An evolution in design

thinking. In “Design Research: The Sociotechnical Aspects of Quality, Creativity, and

Innovation”, D. Marjanovic, M. Storya, S. Skec (eds.), Springer-Verlag, Berlin, 2019 (in print).

Papalambros, P.Y. and Wilde, D.J. (2017). Principles of Optimal Design. Cambridge University

Press.

Parsons, T. (1951). The Social System. Abingdon, Oxon: Routledge.

Pearson, R. H., & Mundfrom, D. J. ( 2010). Recommended sample size for conducting

exploratory factor analysis on dichotomous data. Journal of Modern Applied Statistical Methods,

9(2), 359– 368.

Pennock, M. J., & Wade, J. P. (2015). The top 10 illusions of systems engineering: A research

agenda. Procedia Computer Science, 44, 147–154.

Pidd, M., (1996). Tools for Thinking: Modelling in Management Science. Wiley, Chichester.

Plattner, H., Meinel, C., & Leifer, L. (Eds.). (2011). Design Thinking Research: Measuring

Performance in Context. Springer, Berlin, Heidelberg.

Qualtrics Support. (2018). “Survey Methodology and Compliance Best Practices.” Retrieved 01

May 2019 from https://www.qualtrics.com/support/.

Radzicki, M.J. and Taylor, R.A. (2008). "Origin of System Dynamics: Jay W. Forrester and the

History of System Dynamics". In: U.S. Department of Energy's Introduction to System

Dynamics.

137

Randle, J. M., & Stroink, M. L. (2018). The development and initial validation of the paradigm

of systems thinking: Development and validation of systems thinking. Systems Research, (6),

645–657.

Rath, T. (2007). StrengthsFinder 2.0. Simon and Schuster.

Recker, S. A. (2002). Architecting an engineering documentation system. In Proceedings of the

2002 INCOSE International Symposium, Las Vegas, NV, July 28-Aug. 1, 2002.

Reddit. (2018). “Reddiquette.” Retrieved 01 May 2019 from

https://www.reddit.com/wiki/reddiquette.

Rhodes, D. H., Lamb, C. T., & Nightingale, D. J. (2008). Empirical research on systems thinking

and practice in the engineering enterprise. In Proceedings of the 2008 2nd Annual IEEE Systems

Conference.

Rittel, H. W. J., & Webber, M. M. (1973). Dilemmas in a general yheory of planning. Design:

Critical and Primary Sources.

Rittel, H. W. J., & Webber, M. M. (1974). Wicked problems. Man-Made Futures, 26(1), 272–

280.

Salado, A., Chowdhury, A. H., & Norton, A. (2019). Systems thinking and mathematical

problem solving. School Science and Mathematics, 119(1), 49–58.

Self, J. A., & Baek, J. S. (2017). Interdisciplinarity in design education: Understanding the

undergraduate student experience. International Journal of Technology and Design Education,

27(3), 459–480.

Senge, P. (1990). The Fifth Discipline: The Art and Practice of the Learning Organization.

Doubleday, New York, NY, USA.

Shah, J. J., Millsap, R. E., Woodward, J., & Smith, S. M. (2012). Applied tests of design skills—

Part 1: divergent thinking. Journal of Mechanical Design, 134(2).

Shatz, I. (2017). Fast, free, and targeted: Reddit as a source for recruiting participants online.

Social Science Computer Review, 35(4), 537–549.

Shea, G. (2017). NASA Systems Engineering Handbook Revision 2. Retrieved from

http://www.nasa.gov/connect/ebooks/nasa-systems-engineering-handbook.

Silvers, D. M. (2016, October 31). Why play is essential to the design thinking process.

Retrieved April 3, 2019, from https://designthinkingformuseums.net/2016/10/31/why-play-is-

138

essential-to-the-design-thinking-process/.

Simon, H. A. (1996). The Sciences of the Artificial. MIT Press.

Singer, P., Flock, F., Meinhart, Cl., Zietfogel, E., & Strohmaier, M. (2014). Evolution of Reddit:

From the front page of the internet to a self-referential community? In Proceedings of the 23rd

World Wide Web Conference, Seoul, South Korea.

Sobieszczanski-Sobieski, J. (1990). Sensitivity of complex, internally coupled systems. AIAA

Journal, 28(1), 153-160.

Souza, J., & Barnhöfer, U. (2015). Design thinking: It’s the flare that adds another dimension to

systems engineering. Insight, 18(3), 25–27.

Spacey, J. (2016). “Design Thinking vs Systems Thinking”. Retrieved April 3, 2019, from

https://simplicable.com/new/design-thinking-vs-systems-thinking

Sterman, J. D. (2000). Business dynamics: Systems thinking and modeling for a complex world.

Boston: Irwin/McGraw-Hill.

Steward, D.V. (1981). Systems analysis and management: Structure, strategy, design. Petrocelli

Books, San Francisco, CA.

Thibodeau, P. H., Frantz, C. M., & Stroink, M. L. (2016). Situating a measure of systems

thinking in a landscape of psychological constructs. Systems Research, 33(6), 753–769.

Tjendra, J. (2018, April 25). “Systems Thinking is the New Design Thinking.” Retrieved April 3,

2019, from http://businessinnovation.design/blog/2018/4/25/systems-thinking-is-the-new-design-

thinking.

Toh, C. A., & Miller, S. R. (2016). Creativity in design teams: the influence of personality traits

and risk attitudes on creative concept selection. Research in Engineering Design, 27(1), 73–89.

Tonetto, L. M., & Tamminen, P. (2015). Understanding the role of intuition in decision-making

when designing for experiences: contributions from cognitive psychology. Theoretical Issues in

Ergonomics Science, 16(6), 631–642.

Tosserams, S., Hofkamp, A.T., Etman, L.F.P., and Rooda, J.E. (2010). A specification language

for problem partitioning in decomposition-based design optimization. Structural and

Multidisciplinary Optimization, 42(1), 707-723.

Tribus, M. (2005). Some remarks on the improvement of engineering education. Journal of

Science Education and Technology, 14(1), 1.

139

Vendetti, M. S., Wu, A., & Holyoak, K. J. (2014). Far-out thinking: Generating solutions to

distant analogies promotes relational thinking. Psychological Science, 25(4), 928–933.

Vicente, K. J. (1999). Cognitive Work Analysis: Toward Safe, Productive, and Healthy

Computer-Based Work. CRC Press.

von Bertalanffy, L. (1940). An outline of general systems theory. British Journal for the

Philosophy of Science, 134-165.

Watkins, D., & Gioia, D. (2015). Mixed Methods Research. USA: Oxford University Press.

Weick, K. E. (1979). The Social Psychology of Organizing. New York, NY: Random House.

Weick, K. E. (1995). Sensemaking in Organizations. Thousand Oaks, CA: Sage.

Wiener, N. (1948). Cybernetics: Or Control and Communication in the Animal and the Machine.

Cambridge, MA: MIT Press.

Williams, C. and Derro, M. (2008). NASA systems engineering behavior study. NASA Office of

the Chief Engineer, Washington, DC, USA.

Worthington, R., & Whittaker, T. (2006). Scale development research: A content analysis and

recommendations for best practices. The Counseling Psychologist, 34, 806-838.

Yassine, A. and Braha, D. (2003). “Complex Concurrent Engineering and the Design Structure

Matrix Approach. Concurrent Engineering: Research and Applications, 11(3), 165-177.