SlideShare a Scribd company logo
1 of 145
https://doi.org/10.1177/0963721418763931
Current Directions in Psychological
Science
2018, Vol. 27(3) 183 –187
© The Author(s) 2018
Reprints and permissions:
sagepub.com/journalsPermissions.nav
DOI: 10.1177/0963721418763931
www.psychologicalscience.org/CDPS
Compared with Whites, African Americans are 2 times
more likely to die as infants, 3.6 times more likely to expe-
rience childhood poverty, and 2 times more likely to have
not completed high school (Centers for Disease Control
and Prevention, 2017; National Center for Education Sta-
tistics, 2017; Patten & Krogstad, 2015). Disturbing as these
disparities are, they are not as extreme as disparities in the
criminal justice system, in which African Americans are 5.1
times more likely than Whites to be incarcerated (Nellis,
2016). Many scholars and activists alike assume the general
public would be motivated to fight racial inequality if only
they knew the full extent of existing disparities. In
response, activists bombard people with shocking statis-
tics. Focusing specifically on policing and incarceration,
we explore why simply presenting evidence of extreme
racial disparities in the criminal justice system can backfire.
We propose three potential strategies that may mitigate
this paradoxical effect and provide important avenues for
future research on how to reduce racial inequities.
Evidence of Racial Disparities in the
Criminal Justice System
In cities across the United States, racial disparities have
been documented throughout the criminal justice system,
from routine police stops to long-term imprisonment.
Analyzing traffic and pedestrian stops made in Oakland,
California, for example, our team uncovered a consistent
pattern of racial disparities. We found that 60% of police
stops were of Africans Americans, though they make up
only 28% of the population of Oakland. Once stopped,
African Americans were significantly more likely to be
handcuffed, searched, and arrested (Hetey, Monin,
Maitreyi, & Eberhardt, 2016). These disparities remained
statistically significant even after we controlled for more
than two dozen factors relevant to officer decision mak-
ing, including crime rates and the underlying racial and
socioeconomic demographics where the stop was made.
A similar pattern has emerged in other places, including
Boston; Greensboro, North Carolina; Los Angeles; and
New York City (American Civil Liberties Union Founda-
tion of Massachusetts, 2014; Ayres & Borowsky, 2008;
763931CDPXXX10.1177/0963721418763931Hetey,
EberhardtRacial Disparities in the Criminal Justice System
research-article2018
Corresponding Authors:
Jennifer L. Eberhardt, Department of Psychology, Stanford
University,
450 Serra Mall, Stanford, CA 94305
E-mail: [email protected]
Rebecca C. Hetey, Department of Psychology, Stanford
University,
450 Serra Mall, Stanford, CA 94305
E-mail: [email protected]
The Numbers Don’t Speak for Themselves:
Racial Disparities and the Persistence of
Inequality in the Criminal Justice System
Rebecca C. Hetey and Jennifer L. Eberhardt
Stanford University
Abstract
Many scholars and activists assume the public would be
motivated to fight inequality if only they knew the full
extent of existing disparities. Ironically, exposure to extreme
disparities can cause people to become more, not less,
supportive of the very policies that create those disparities
(Hetey & Eberhardt, 2014). Here, we focus on the criminal
justice system—policing and incarceration in particular. We
argue that bringing to mind racial disparities in this domain
can trigger fear and stereotypic associations linking Blacks with
crime. Therefore, rather than extending an invitation
to reexamine the criminal justice system, the statistics about
disparities may instead provide an opportunity to justify
and rationalize the disparities found within that system. With
the goals of spurring future research and mitigating this
paradoxical and unintended effect, we propose three potential
strategies for more effectively presenting information
about racial disparities: (a) offer context, (b) challenge
associations, and (c) highlight institutions.
Keywords
race, crime, disparities, mass incarceration, policing
Article
http://www.psychologicalscience.org/cdps
mailto:[email protected]
https://sagepub.com/journalsPermissions.nav
http://crossmark.crossref.org/dialog/?doi=10.1177%2F09637214
18763931&domain=pdf&date_stamp=2018-05-03
184 Hetey, Eberhardt
Dunn, 2013; Goel, Rao, & Shroff, 2016; LaFraniere &
Lehren, 2015).
Blacks feel they are treated with less respect than
Whites during these routine stops (Epp, Maynard-Moody,
& Haider-Markel, 2014). Using footage from officers’
body-worn cameras, our team developed computational
linguistic methods to analyze the respectfulness of the
language that police officers used with White and Black
community members during traffic stops in Oakland. We
found that officers’ language was less respectful when
directed at Blacks than when directed at Whites, even
after controlling for factors such as the race of the officer,
the severity of the infraction, the location of the stop,
and the outcome of the stop (Voigt et al., 2017).
In incarceration as well there are striking racial dis-
parities. Though only 13% of the U.S. population, African
Americans make up nearly 40% of the nation’s inmates
(Guerino, Harrison, & Sabol, 2012; U.S. Census Bureau,
2011). Black men are incarcerated in state or federal
prison at a rate 6 times that of White men (Carson, 2014).
During their lifetime, 1 in 3 African American males can
expect to be imprisoned compared with 1 in 17 White
males (Bonczar, 2003).
Do the Numbers Speak for Themselves?
People trying to motivate social change frequently
expose the public to grim statistics revealing gross
racial disparities. Ironically, researchers have found that
being presented with evidence of extreme racial dis-
parities in the criminal justice system can cause the
public to become more, not less, supportive of the
punitive criminal justice policies that produce those
disparities. When White participants were informed
about racial disparities in executions, 52% strongly
favored the death penalty compared with 36% in a
baseline condition (Peffley & Hurwitz, 2007). In another
study, we exposed White California voters to more or
less extreme racial disparities in the prison population
by having them view a set of photographs of incarcer-
ated people that showed either a higher (45%) or lower
(25%) percentage of Black inmates (Hetey & Eberhardt,
2014). When the prison population was “more Black,”
voters became significantly less likely to sign a real
petition aimed at lessening the severity of California’s
harsh three-strikes law. At the time we conducted the
study, California had the harshest three-strikes law in
the nation—a person with two prior felony convictions
could receive a life sentence for stealing a dollar in
loose change from a parked car. When we described
California’s three-strikes law and presented participants
with a prison population that was less Black, 52% of
them signed the petition compared with only 27% in
the more-Black condition.
Similarly, when White New York City residents read
that New York’s prison population was 60% Black, com-
pared with 40% Black, they became less likely to report
that they would sign a petition to end New York City’s
stop-and-frisk policy. The study was run at a time when
the number of Blacks detained by the New York Police
Department was dropping from a recent all-time high.
When exposed to a more-Black prison population, only
12% of participants said they would sign the petition,
compared with 33% of participants exposed to a less-
Black prison population. Taken together, these studies
suggest that informing White Americans of racially dis-
proportionate incarceration may paradoxically bolster
support for the very policies that perpetuate those
disparities.
These experiments were conducted with White par-
ticipants. While an open question, it is unlikely that
exposure to evidence of disparities has the same poten-
tial to backfire with people from other racial back-
grounds. People of color, and African Americans in
particular, are more likely than Whites to have direct
experience with, perceive, and engage in discussions
about racial inequality and discrimination. According to
a national Pew Research Center (2016) survey, 71% of
Blacks reported having personally experienced discrimi -
nation or having been treated unfairly because of their
race or ethnicity, compared with 30% of Whites. By wide
margins, Black respondents were more likely than White
respondents to say Blacks are treated less fairly across
a wide variety of settings, including the criminal justice
system. Eighty-four percent of Blacks, compared with
50% of Whites, say Blacks are treated less fairly in deal-
ing with the police. Indeed, 18% of Black Americans say
they have been unfairly stopped by the police over the
past year, compared with 3% of Whites.
In the absence of direct experience and engagement
with racial inequality, one reason disparities do not
speak for themselves is because there is more than one
interpretation of what those differences reveal. To some
people, the disparities in police stops and incarceration
are so extreme they can be interpreted only as signs of
systemic racial bias (e.g., Alexander, 2010). An alterna-
tive interpretation is that members of particular racial
groups must be doing something—namely committing
crime—to capture the attention of police and be impris-
oned at higher rates (e.g., B. Adams, 2016). By focusing
on group traits, the possibility that structural bias is at
play in creating disparities falls out of view. Further-
more, research shows that White Americans are sus-
ceptible to denying claims of bias (G. Adams, Tormala,
& O’Brien, 2006), in part to preserve views of the world
as fair and meritocratic—views that can blind Whites
to racial inequality and the role of institutional practices
that harm minorities (O’Brien et al., 2009).
Racial Disparities in the Criminal Justice System 185
Fear and stereotypic associations also contribute.
Indeed, we found that when Whites were exposed to
a “Blacker” prison population, they became significantly
more fearful of crime, which, in turn, increased their
support of punitive crime policies (Hetey & Eberhardt,
2014). Evidence of racial disparities in the criminal jus-
tice system, then, may activate implicit stereotypical
associations linking Blackness with crime, violence,
threat, and aggression (Correll, Park, Judd, & Wittenbrink,
2002; Eberhardt, Goff, Purdie, & Davies, 2004; Payne,
2001; Richeson & Trawalter, 2008). If learning that Black
men are 6 times more likely than White men to be
imprisoned triggers the stereotype that Blacks are crimi -
nals and criminals are Black, then such information is
no longer concerning on its face. Rather, it becomes
expected or even justifiable. Motivated to maintain the
status quo, people take evidence of what is (such as
the overrepresentation of Blacks in prison) and justify
it as how things should be (Kay et al., 2009).
What Should We Do?
Given these ironic effects and what we know more
generally about people’s use of information to support
what they already believe or expect (Darley & Gross,
1983; Nickerson, 1998), what should people working
for social change do? We propose three potential strate-
gies that may mitigate the unintended consequences of
exposing Whites to evidence of racial inequalities: (a)
offer context, (b) challenge associations, and (c) high-
light institutions.
Offer context
Future research should test whether a more effective
way to frame inequality might be to present informa-
tion about disparities alongside a description of how
those disparities came to be. To contextualize is to
convey that racial disparities are not natural or due to
fixed stereotypical traits. Civil rights activists are aware
of the history of racial disparities and the larger forces
and processes by which these disparities are main-
tained, yet that history is not always highlighted in the
messages aimed at the broader public. Stripped of con-
text, stand-alone statistics may simply be used as “evi-
dence” of the stereotype that Blacks are prone to
criminality. Rather than leading the public to examine
the systemic forces involved, decontextualized infor-
mation about racial inequality can lead people to
engage in defensive processes that instead blame the
victim (Kaiser & Miller, 2003) while protecting the self
and the status quo (G. Adams et al., 2006; Lowery,
Knowles & Unzueta, 2007).
Challenge associations
Exposure to information about racial disparities in the
criminal justice system may activate and strengthen
implicit associations linking Blacks with crime, vio-
lence, threat, and aggression. Black Americans surely
feel the effects of these stereotypical associations:
Nearly half of Blacks (47%) say someone in the past
year has acted as if they were suspicious of them (Pew
Research Center, 2016). More research should investi -
gate the consequences of educating the public about
implicit bias and illustrating how implicit associations
linking Blacks with crime can profoundly affect our
perceptions, decision making, and actions without con-
scious awareness or malicious intentio n. Our team has
created an intervention to educate people on implicit
bias and where it comes from, how it operates, and
how it can be challenged. We are currently rolling out
this training to law enforcement agencies across Cali-
fornia. Preliminary results suggest the training is effec-
tive in increasing knowledge of implicit bias, and most
encouragingly, it leads law enforcement personnel to
desire better relations with the community, to have
more sympathy for community members, and to be
more likely to believe they personally can make a dif-
ference (Stanford SPARQ & California Department of
Justice, 2016).
Highlight institutions
Another contributing factor to the perpetuation of
inequality is the tendency to prioritize the role of indi -
viduals over institutions. We, as a field, are guilty of
this. Adams and colleagues describe how “as psycholo-
gists, social psychologists tend to understand their task
to be the study of individual experience . . . [suggesting]
a standard approach to racism as a problem of biased
individuals rather than a systemic force embedded in
American society” (G. Adams, Edkins, Lacka, Pickett, &
Cheryan, 2008, p. 350, emphasis in original). This con-
ception or portrayal of inequality as an individual—
rather than systemic—phenomenon can blind White
Americans to racism and can guide policy attitudes and
preferences in ways that perpetuate racial disparities.
The individualistic portrayal is problematic because
institutions and systems greatly contribute to reinforc-
ing and reproducing inequality. In the criminal justice
system, numerous distinct actors must be endowed by
the system with the power to investigate, arrest, charge,
prosecute, judge, sentence, and deprive people of their
freedom, if not their lives.
In the abstract, it can be difficult to conceptualize of
vast systems and institutions. Lived experience grounds
186 Hetey, Eberhardt
an understanding and recognition of structural racism.
African Americans who say they have personally expe-
rienced discrimination are equally divided over whether
institutional racism or individual prejudice is the bigger
problem for Black people today (44% each; Pew
Research Center, 2016). African Americans who say they
have never experienced discriminatio n, however, are
nearly twice as likely to see individual racism as the
bigger problem (59% vs. 32%). This lack of personal
experience might explain why 70% of Whites in the
same survey point to individual prejudice as the bigger
problem, compared with only 19% of Whites who
believe racism built into laws and institutions is the big-
ger problem. Awareness of the power of institutions can
be taught. Tutorials about structural or sociocultural
racism have been shown to decrease the extent to which
White Americans define racism in terms of individual
biases, increase perceptions and acknowledgment of
systemic racism, and increase endorsement of policies
designed to promote equality (G. Adams et al., 2008).
One specific way to highlight the role of institutions
in perpetuating inequality may be to demonstrate how
disparities can be attenuated by changing policy. Racial
disparities in the use of consent searches (LaFraniere
& Lehren, 2015; Palomino, 2016), for instance, have led
to policy changes that mandate officers to obtain writ-
ten consent or explicitly tell community members they
have the right to deny an officer’s request to search
them (LaFraniere, 2016). In Oakland, this policy change
led to a massive reduction in consent searches without
increasing crime. As another example, consensus to end
mass incarceration has been building. Over 100 of the
nation’s police chiefs and law enforcement officials
joined together to call for a reduction in incarceration
rates and the end of certain “tough-on-crime” policies
(Williams, 2015). States including Texas, New York,
Georgia, and California have eased minimum sentenc-
ing and repeat offender laws. The result has been a
reduction in both incarceration and crime (Chettiar,
2015), thus illustrating the role of laws and policies to
actively shape, rather than passively respond to, crime
and punishment.
Conclusion
To make the most meaningful change, we as social
psychologists need to become more willing to intervene,
not only with individuals but also by working to under-
stand and change institutions, policies, and practices.
Such a broader view could only enrich our science.
Recommended Reading
Adams, G., Edkins, V., Lacka, D., Pickett, K. M., & Cheryan, S.
(2008). (See References). An article illustrating that the
standard portrayal of racism in social psychology is as
an individual prejudice but that, as the authors show,
an awareness of structural or sociocultural racism can
be taught.
Eberhardt, J. L., Goff, P. A., Purdie, V. J., & Davies, P. G.
(2004). (See References). A series of studies showing
that stereotypical associations linking Blacks and crime
are bidirectional and can act as visual tuning devices; for
instance, exposure to Black faces facilitates the detection
of weapons and activating the concept of crime creates
an attentional bias toward Black faces.
Hetey, R. C., & Eberhardt, J. L. (2014). (See References). Two
experiments showing that exposure to extreme disparities
can cause people to become more, not less, supportive
of the very policies that create those disparities.
Peffley, M., & Hurwitz, J. (2007). (See References). An article
showing how informing Whites about racial disparities in
executions bolsters support of the death penalty.
Western, B. (2006). Punishment and inequality in America.
New York, NY: Russell Sage Foundation. An exploration
of the scope and consequences of mass incarceration in
America, in particular the role of imprisonment in exac-
erbating inequality.
Declaration of Conflicting Interests
The author(s) declared that there were no conflicts of interest
with respect to the authorship or the publication of this
article.
References
Adams, B. (2016, July 11). CNN panel blows up after guest sug-
gests Blacks are prone to crime. Washington Examiner.
Retrieved from http://www.washingtonexaminer.com/
cnn-panel-blows-up-after-guest-suggests-blacks-are-
prone-to-crime/article/2596167
Adams, G., Edkins, V., Lacka, D., Pickett, K. M., & Cheryan, S.
(2008). Teaching about racism: Pernicious implications of
the standard portrayal. Basic and Applied Social Psy-
chology, 30, 349–361.
Adams, G., Tormala, T. T., & O’Brien, L. T. (2006). The effect
of self-affirmation on perception of racism. Journal of
Experimental Social Psychology, 42, 616–626.
Alexander, M. (2010). The new Jim Crow: Mass incarceration in
the age of colorblindness. New York, NY: The New Press.
American Civil Liberties Union Foundation of Massachusetts.
(2014). Black, brown and targeted: A report on Boston Police
Department street encounters from 2007–2010. Retrieved
from https://aclum.org/wp-content/uploads/2015/06/
reports-black-brown-and-targeted.pdf
Ayres, I., & Borowsky, J. (2008). A study of racially disparate
outcomes in the Los Angeles Police Department. Retrieved
from https://www.aclusocal.org/sites/default/files/wp-
content/uploads/2015/09/11837125-LAPD-Racial-Profiling-
Report-ACLU.pdf
Bonczar, T. P. (2003). Prevalence of imprisonment in the U.S.
population, 1974-2001. Bureau of Justice Statistics. Retrieved
from https://www.bjs.gov/content/pub/pdf/piusp01.pdf
Carson, E. A. (2014). Prisoners in 2013. Bureau of Justice
Statistics. Retrieved from https://www.bjs.gov/content/
pub/pdf/p13.pdf
http://www.washingtonexaminer.com/cnn-panel-blows-up-after-
guest-suggests-blacks-are-prone-to-crime/article/2596167
http://www.washingtonexaminer.com/cnn-panel-blows-up-after-
guest-suggests-blacks-are-prone-to-crime/article/2596167
http://www.washingtonexaminer.com/cnn-panel-blows-up-after-
guest-suggests-blacks-are-prone-to-crime/article/2596167
https://aclum.org/wp-content/uploads/2015/06/reports-black-
brown-and-targeted.pdf
https://aclum.org/wp-content/uploads/2015/06/reports-black-
brown-and-targeted.pdf
https://www.aclusocal.org/sites/default/files/wp-
content/uploads/2015/09/11837125-LAPD-Racial-Profiling-
Report-ACLU.pdf
https://www.aclusocal.org/sites/default/files/wp-
content/uploads/2015/09/11837125-LAPD-Racial-Profiling-
Report-ACLU.pdf
https://www.aclusocal.org/sites/default/files/wp-
content/uploads/2015/09/11837125-LAPD-Racial-Profiling-
Report-ACLU.pdf
https://www.bjs.gov/content/pub/pdf/piusp01.pdf
https://www.bjs.gov/content/pub/pdf/p13.pdf
https://www.bjs.gov/content/pub/pdf/p13.pdf
Racial Disparities in the Criminal Justice System 187
Centers for Disease Control and Prevention. (2017). Sudden
unexpected infant death and sudden infant death syn-
drome: Data and statistics. Retrieved from https://www
.cdc.gov/sids/data.htm
Chettiar, I. M. (2015, April 27). A national agenda to reduce
mass incarceration. Brennan Center for Justice. Retrieved
from https://www.brennancenter.org/analysis/national-
agenda-reduce-mass-incarceration
Correll, J., Park, B., Judd, C. M., & Wittenbrink, B. (2002). The
police officer’s dilemma: Using ethnicity to disambiguate
potentially threatening individuals. Journal of Personality
and Social Psychology, 83, 1314–1329.
Darley, J. M., & Gross, P. H. (1983). A hypothesis-confirming
bias in labeling effects. Journal of Personality and Social
Psychology, 44, 20–33.
Dunn, C. (2013). Stop-and-frisk 2012: NYCLU briefing. New
York Civil Liberties Union. Retrieved from https://www
.nyclu.org/sites/default/files/publications/2012_Report_
NYCLU_0.pdf
Eberhardt, J. L., Goff, P. A., Purdie, V. J., & Davies, P. G.
(2004).
Seeing Black: Race, crime, and visual processing. Journal
of Personality and Social Psychology, 87, 876–893.
Epp, C. R., Maynard-Moody, S., & Haider-Markel, D. P. (2014).
Pulled over: How police stops define race and citizenship.
Chicago, IL: The University of Chicago Press.
Goel, S., Rao, J. M., & Shroff, R. (2016). Precinct or preju-
dice? Understanding racial disparities in New York City’s
stop-and-frisk policy. The Annals of Applied Statistics, 10,
365–394.
Guerino, P, Harrison, P. M., & Sabol, W. J. (2012). Prisoners in
2010. Bureau of Justice Statistics. Retrieved from https://
www.bjs.gov/content/pub/pdf/p10.pdf
Hetey, R. C., & Eberhardt, J. L. (2014). Racial disparities in
incarceration increase acceptance of punitive policies.
Psychological Science, 25, 1949–1954.
Hetey, R. C., Monin, B., Maitreyi, A., & Eberhardt, J. L.
(2016).
Data for change: A statistical analysis of police stops,
searches, handcuffings, and arrests in Oakland, Calif.,
2013-2014. Stanford SPARQ. Retrieved from https://stan
ford.app.box.com/v/Data-for-Change
Kaiser, C. R., & Miller, C. T. (2003). Derogating the victim:
The interpersonal consequences of blaming events on
discrimination. Group Processes & Intergroup Relations,
6, 227–237.
Kay, A. C., Gaucher, D., Peach, J. M., Laurin, K., Friesen, J.,
Zanna, M. P., & Spencer, S. J. (2009). Inequality, discrimi -
nation, and the power of the status quo: Direct evidence
for a motivation to see the way things are as the way they
should be. Journal of Personality and Social Psychology,
97, 421–434.
LaFraniere, S. (2015, November 11). Greensboro puts focus on
reducing racial bias. The New York Times. Retrieved from
https://www.nytimes.com/2015/11/12/us/greensboro-
puts-focus-on-reducing-racial-bias.html
LaFraniere, S., & Lehren, A. W. (2015, October 24). The
dispro-
portionate risks of driving while Black. The New York Times.
Retrieved from https://www.nytimes.com/2015/10/25/us/
racial-disparity-traffic-stops-driving-black.html
Lowery, B. S., Knowles, E. D., & Unzueta, M. M. (2007).
Framing inequity safely: Whites’ motivated perceptions of
racial privilege. Personality and Social Psychology Bulletin,
33, 1237–1250.
National Center for Education Statistics. (2017). Public high
school 4-year adjusted cohort graduation rate (ACGR), by
race/ethnicity and selected demographic characteristics
for the United States, the 50 states, and the District of
Columbia: School year 2015–16 [Data table]. Retrieved
from https://nces.ed.gov/ccd/tables/ACGR_RE_and_char
acteristics_2015-16.asp
Nellis, A. (2016). The color of justice: Racial and ethnic dis -
parity in state prisons. The Sentencing Project. Retrieved
from https://www.sentencingproject.org/wp-content/
uploads/2016/06/The-Color-of-Justice-Racial-and-Ethnic-
Disparity-in-State-Prisons.pdf
Nickerson, R. S. (1998). Confirmation bias: A ubiquitous phe-
nomenon in many guises. Review of General Psychology,
2, 175–220.
O’Brien, L. T., Blodorn, A., Alsbrooks, A., Dube, R., Adams,
G.,
& Nelson, J. C. (2009). Understanding White Americans’
perceptions of racism in hurricane Katrina-related events.
Group Processes & Intergroup Relations, 12, 431–444.
Palomino, J. (2016, April 8). Racial disparities in SF traffic
searches raise concerns of bias. San Francisco Chronicle.
Retrieved from http://www.sfchronicle.com/crime/article/
Racial-disparities-in-SF-traffic-searches-raise-7235690.php
Patten, E., & Krogstad, J. M. (2015). Black child poverty
rate holds steady, even as other groups see declines.
Pew Research Center. Retrieved from http://www.pewre
search.org/fact-tank/2015/07/14/black-child-poverty-rate-
holds-steady-even-as-other-groups-see-declines/
Payne, B. K. (2001). Prejudice and perception: The role of
automatic and controlled processes in misperceiving a
weapon. Journal of Personality and Social Psychology,
81, 181–192.
Peffley, M., & Hurwitz, J. (2007). Persuasion and resistance:
Race and the death penalty in America. American Journal
of Political Science, 51, 996–1012.
Pew Research Center. (2016). On views of race and inequality,
Blacks and Whites are worlds apart. Retrieved from http://
www.pewsocialtrends.org/2016/06/27/on-views-of-race-
and-inequality-blacks-and-whites-are-worlds-apart/
Richeson, J. A., & Trawalter, S. (2008). The threat of appear-
ing prejudiced and race-based attentional biases.
Psychological Science, 19, 98–102.
Stanford SPARQ & California Department of Justice. (2016).
Principled policing training: Procedural justice and
implicit bias. Retrieved from https://stanfor d.app.box
.com/s/e3ik3z20be2k9k6p6pluo9o42wdz5cni
U.S. Census Bureau. (2011). Overview of race and Hispanic
origin: 2010. Retrieved from https://www.census.gov/
prod/cen2010/briefs/c2010br-02.pdf
Voigt, R., Camp, N. P., Prabhakaran, V., Hamilton, W. L.,
Hetey, R. C., Griffiths, C. M., . . . Eberhardt, J. L. (2017).
Language from police body camera footage shows racial
disparities in officer respect. Proceedings of the National
Academy of Sciences, USA, 114, 6521–6526.
Williams, T. (2015, October 20). Police leaders join call to
cut prison rosters. The New York Times. Retrieved from
http://www.nytimes.com/2015/10/21/us/police-leaders-
join-call-to-cut-prison-rosters.html?_r=0
https://www.cdc.gov/sids/data.htm
https://www.cdc.gov/sids/data.htm
https://www.brennancenter.org/analysis/national-agenda-reduce-
mass-incarceration
https://www.brennancenter.org/analysis/national-agenda-reduce-
mass-incarceration
https://www.nyclu.org/sites/default/files/publications/2012_Rep
ort_NYCLU_0.pdf
https://www.nyclu.org/sites/default/files/publications/2012_Rep
ort_NYCLU_0.pdf
https://www.nyclu.org/sites/default/files/publications/2012_Rep
ort_NYCLU_0.pdf
https://www.bjs.gov/content/pub/pdf/p10.pdf
https://www.bjs.gov/content/pub/pdf/p10.pdf
https://stanford.app.box.com/v/Data-for-Change
https://stanford.app.box.com/v/Data-for-Change
https://www.nytimes.com/2015/11/12/us/greensboro-puts-focus-
on-reducing-racial-bias.html
https://www.nytimes.com/2015/11/12/us/greensboro-puts-focus-
on-reducing-racial-bias.html
https://www.nytimes.com/2015/10/25/us/racial-disparity-traffic-
stops-driving-black.html
https://www.nytimes.com/2015/10/25/us/racial-disparity-traffic-
stops-driving-black.html
https://nces.ed.gov/ccd/tables/ACGR_RE_and_characteristics_2
015-16.asp
https://nces.ed.gov/ccd/tables/ACGR_RE_and_characteristics_2
015-16.asp
https://www.sentencingproject.org/wp-
content/uploads/2016/06/The-Color-of-Justice-Racial-and-
Ethnic-Disparity-in-State-Prisons.pdf
https://www.sentencingproject.org/wp-
content/uploads/2016/06/The-Color-of-Justice-Racial-and-
Ethnic-Disparity-in-State-Prisons.pdf
https://www.sentencingproject.org/wp-
content/uploads/2016/06/The-Color-of-Justice-Racial-and-
Ethnic-Disparity-in-State-Prisons.pdf
http://www.sfchronicle.com/crime/article/Racial-disparities-in-
SF-traffic-searches-raise-7235690.php
http://www.sfchronicle.com/crime/article/Racial-disparities-in-
SF-traffic-searches-raise-7235690.php
http://www.pewresearch.org/fact-tank/2015/07/14/black-child-
poverty-rate-holds-steady-even-as-other-groups-see-declines/
http://www.pewresearch.org/fact-tank/2015/07/14/black-child-
poverty-rate-holds-steady-even-as-other-groups-see-declines/
http://www.pewresearch.org/fact-tank/2015/07/14/black-child-
poverty-rate-holds-steady-even-as-other-groups-see-declines/
http://www.pewsocialtrends.org/2016/06/27/on-views-of-race-
and-inequality-blacks-and-whites-are-worlds-apart/
http://www.pewsocialtrends.org/2016/06/27/on-views-of-race-
and-inequality-blacks-and-whites-are-worlds-apart/
http://www.pewsocialtrends.org/2016/06/27/on-views-of-race-
and-inequality-blacks-and-whites-are-worlds-apart/
https://stanford.app.box.com/s/e3ik3z20be2k9k6p6pluo9o42wdz
5cni
https://stanford.app.box.com/s/e3ik3z20be2k9k6p6pluo9o42wdz
5cni
https://www.census.gov/prod/cen2010/briefs/c2010br-02.pdf
https://www.census.gov/prod/cen2010/briefs/c2010br-02.pdf
http://www.nytimes.com/2015/10/21/us/police-leaders-join-call-
to-cut-prison-rosters.html?_r=0
http://www.nytimes.com/2015/10/21/us/police-leaders-join-call-
to-cut-prison-rosters.html?_r=0
Information Systems Success: The Quest for
the Dependent Variable .
William H. DeLone Department of Management
The American University
Washington, D.C. 20016
Ephraim R. McLean Computer Information Systems
Georgia State University
Atlanta, Georgia 30302-4015
A large number of studies have been conducted during the last
decade and a
half attempting to identify those factors that contribute to
information sys-
tems success. However, the dependent variable in these
studies—I/S success
—has been an elusive one to define. Different researchers have
addressed
different aspects of success, making comparisons difficult and
the prospect of
building a cumulative tradition for I/S research similarly
elusive. To organize
this diverse research, as well as to present a more integrated
view of the
concept of I/S success, a comprehensive taxonomy is
introduced. This taxon-
omy posits six major dimensions or categories of I/S success —
SYSTEM
OUALITY, INFORMATION QUALITY, USE, USER
SATISFACTION,
INDIVIDUAL IMPACT, and ORGANIZATIONAL IMPACT.
Using these
dimensions, both conceptual and empirical studies are then
reviewed (a total
of 180 articles are cited) and organized according to the
dimensions of the
taxonomy. Finally, the many aspects of I/S success are drawn
together into a
descriptive model and its implications for future I/S research
are discussed.
Information s)'!)(em« •iucce^s—Inrormition systems
•ssessmenl—Measurement
Introduction
At the first meeting of the International Conference on
Information System (ICIS)in 1980, Peter Keen identified five
issues which he felt needed to be resolved in
order for the field of management information systems to
establish itself as a coherent
research area. These issues were:
(1) What are the reference disciplines for MIS?
(2) What is the dependent variable?
(3) How does MIS establish a cumulative tradition?
(4) What is the relationship of MIS research to computer
technology and to MIS
practice?
(5) Where should MIS Tesearchers publish their findings?
I047.7047/92/0301/0O6O/SOI.25
Copyright © 1992. The Inslilulc of Management Stienees
60 Information Systems Research 3 : I
Information Systems Success
Of the five, the second item, the dependent variable in MIS
research, is a particu-
larly important issue. If information systems research is to make
a contribution to the
world of practice, a well-defined outcome measure (or
measures) is essential. It does
littie good to measure various independent or input variables,
such as the extent of
user participation or the level of I/S investment, if the
dependent or output variable
—I/S success or MIS effectiveness—cannot be measured with a
similar degree of
accuracy.
The importance of defining the I/S dependent variable cannot be
overemphasized.
The evaluation of I/S practice, policies, and procedures requi res
an I/S success mea-
sure against which various strategies can be tested. Without a
well-defined dependent
variable, much of I/S research is purely speculative.
In recognition of this importance, this paper explores the
research that has been
done involving MIS success since Keen first issued his
challenge to the field and
attempts to synthesize this research into a more coherent body
of knowledge. It
covers the formative period 1981 -87 and reviews all those
empirical studies that have
attempted to measure some aspects of "MIS success" and which
have appeared in
one of the seven leading publications in the I/S field. In
addition, a number of other
articles are included, some dating back to 1949. that make a
theoretical or conceptual
contribution even though they may not contain any empirical
data. Taken together,
these 180 references provide a representative review of the
work that has been done
and provide the basis for formulating a more comprehensive
model of I/S success
than has been attempted in the past.
A Taxonomy of Information Systems Success
Unfortunately, in searching for an I/S success measure, rather
than finding none,
there are nearly as many measures as there are studies. The
reason for this is under-
standable when one considers that "information." as the output
of an information
system or the message in a communication system, can be
measured at different
levels, including the technical level, the semantic level, and the
effectiveness level. In
their pioneering work on communications. Shannon and Weaver
(1949) defined the
technical level as the accuracy and efficiency of the system
which produces the infor-
mation, the semantic level as the success of the information in
conveying the in-
tended meaning, and the effectiveness level as the effect of the
information on the
receiver.
Building on this. Mason (1978) relabeled "effectiveness" as
"influence" and de-
fined the influence level of information to be a "hierarchy of
events which take place
at the receiving end of an information system which may be
used to identify the
various approaches that might be used to measure output at the
influence level"
(Mason 1978. p. 227). This scries of influence events includes
the receipt of the
information, an evaluation of the information, and the
application of the informa-
tion, leading to a change in recipient behavior and a change in
system performance.
The concept of levels of output from communication theory
demonstrates the
serial nature of information (i.e., a form of communication).
The information system
creates information which is communicated to the recipient who
is then influenced
(or not!) by the information. In this sense, information flows
through a series of stages
from its production through its use or consumption to its
influence on individual
and/or organizational performance. Mason's adaptation of
communication theory
March 1992
DeLone •
Shannon
Weaver
(1949)
Mason
(1978)
Categories
ofVS
Success
McLean
Technical
Level
Production
System
Quality
Semantic
Level
Product
Information
Quality
Receipt
Use
Effectiveness or hifluence
Level
Influence
on
Recipent
User Individual
Satisfaction Impact
Influence
on
System
Organizational
Impact
FIGURE 1. Categories of [/S Success.
to the measurement of information systems suggests therefore
that there may need to
be separate success measures for each of the levels of
information.
In Figure 1, the three levels of information of Shannon and
Weaver are shown,
together with Mason's expansion of the effectiveness or
influence level, to yield six
distinct categories or aspects of information systems. They are
SYSTEM QUALITY,
INFORMATION QUALITY, USE. USER SATISFACTION,
INDIVIDUAL IM-
PACT, and ORGANIZATIONAL IMPACT.
Looking at the first of these categories, some I/S researcher s
have chosen to focus
on the desired characteristics of the information system itself
which produces the
information (SYSTEM QUALITY). Others have chosen to study
the information
product for desired characteristics such as accuracy, meaningful
ness, and timeliness
(INFORMATION QUALITY). In the influence level, some
researchers have ana-
lyzed the interaction of the information product with its
recipients, the users and/or
decision makers, by measuring USE or USER SATISFACTION.
Still other re-
searchers have been interested in the influence which the
information product has on
management decisions (INDIVIDUAL IMPACT). Finally, some
I/S researchers,
and to a larger extent I/S practitioners, have been concerned
with the eifect of
the information product on organizational performance
(ORGANIZATIONAL
IMPACT).
Once this expanded view of I/S success is recognized, it is not
surprising to find that
there are so many different measures of this success in the
literature, depending upon
which aspect of I/S the researcher has focused his or her
attention. Some of these
measures have been merely identified, but never used
empirically. Others have been
used, but have employed different measurement instruments,
making comparisons
among studies difficult.
Two previous articles have made extensive reviews of the
research literature and
have reported on the measurement of MIS success that had been
used in empirical
studies up until that time. In a review of studies of user
involvement, Ives and Olson
(1984) adopted two classes of MIS outcome variables: system
quality and system
acceptance. The system acceptance category was defined to
include system usage,
system impact on user behavior, and information satisfaction.
Haifa decade earlier,
in a review of studies of individual differences. Zmud (1979)
considered three catego-
ries of MIS success: user performance, MIS usage, and user
satisfaction.
Eoth of these literature reviews made a valuable contribution to
an understanding
of MIS success, but both were more concerned with
investigating independent
62 Information Systems Research 3 : 1
Information Systems Success
variables (i.e., user involvement in the case of Ives and Olson
and individual differ-
ences in the case of Zmud) than with the dependent variable
(i.e.. MIS success). In
contrast, this paper has the measurement of the dependent
variable as its primary
focus. Also, over five years have passed since the Ives and
Olson study was published
and over ten years since Zmud's article appeared. Much work
has been done since
these two studies, justifying an update of their findings.
To review this recent work and to put the earlier research into
perspective, the six
categories of I/S success identified in Figure I—SYSTEM
QUALITY. INFORMA-
TION QUALITY. INFORMATION USE, USER
SATISFACTION, INDIVIDUAL
IMPACT. AND ORGANIZATIONAL IMPACT—are used in the
balance of this
paper to organize the I/S research that has been done on I/S
success.
In each of the six sections which follow, both conceptual and
empirical studies are
cited. While the conceptual citations are intended to be
comprehensive, the empirical
studies are intended to be representative, not exhaustive. Seven
publications, from
the period January 1981 to January 1988. were selected as
reflecting the mainstream
of I/S research during this formative period. Additional studies,
from other publica-
tions, as well as studies from the last couple of years, could
have been included; but
after reviewing a number of them, it became apparent that they
merely reinforced
rather than modified the basic taxonomy of this paper.
In choosing the seven publications to be surveyed. fyQ
(Management Science, MIS
Quarterly, Communications of the ACM, Decision Sciences, and
Information & Man-
agement) were drawn from the top six journals cited by
Hamilton and Ives (1983) in
their study of the journals most respected by MIS researchers.
(Their sixth journal.
Transactions on Database Systems, was omitted from this study
because of its special-
ized character.) To these five were added the Journal of MIS. a
relatively new but
important journal, and the ICIS Proceedings, which is not a
journal per se but repre-
sents the published output of the central academic conference in
the I/S field. A total
of 100 empirical studies are included from these seven sources.
As with any attempt to organize past research, a certain degree
of arbitrariness
occurs. Some studies do not fit neatly into any one category and
others fit into
several. In the former case, every effort was made to make as
close a match as possible
in order to retain a fairly parsimonious framework. In the latter
case, where several
measures were used which span more than one category (e.g..
measures of informa-
tion quality and extent of use and user satisfaction), these
studies are discussed in
each of these categories. One consequence of this multiple
listing is that there appear
to be more studies involving I/S success than there actually are.
To decide which empirical studies should be included, and
which measures fit in
which categories, one of the authors of this paper and a doctoral
student (at another
university) reviewed each of the studies and made their
judgments independently.
The interrater agreement was over 90%. Conflicts over selection
and measure assign-
ment were resolved by the second author.
In each of the following sections, a table is included which
summarizes the empiri-
cal studies which address the particular success variable in
question. In reporting the
success measures, the specific description or label for each
dependent variable, as
used by the author(s) of the study, is reported. In some cases the
wording of these
labels may make it appear that the study would be more
appropriately listed in
another table. However, as was pointed out earlier, all of these
classification decisions
March 1992 63
DeLone • McLean
are somewhat arbitrary, as is true of almost al! attempts to
organize an extensive body
of research on a retrospective basis.
System Quality: Measures of the Information Processing System
Itself
In evaluating the contribution of information systems to the
organization, some
I/S researchers have studied the processing system itself.
Kriebel and Raviv (1980,
1982) created and tested a productivity model for computer
systems, including such
performance measures as resource utilization and investment
utilization. Alloway
(1980) developed 26 criteria for measuring the success of a data
processing operation.
The efficiency of hardware utilization was among Alloway's
system success criteria.
Other authors have developed multiple measures of system
quality. Swanson
(1974) used several system quality items to measure MIS
appreciation among user
managers. His items included the reliability of the computer
system, on-line response
time, the ease of terminal use. and so forth. Emery (1971) also
suggested measuring
system characteristics, such as the content of the data base,
aggregation of details,
human factors, response time, and system accuracy. Hamilton
and Chervany (1981)
proposed data currency, response time, turnaround time, data
accuracy, reliability,
completeness, system flexibility, and ease of use among others
as part of a "formative
evaluation" scheme to measure system quality.
In Table I are shown the empirical studies which had explicit
measures of system
quality. Twelve studies were found within the referenced
journals, with a number of
distinct measures identified. Not surprisingly, most of these
measures are fairly
straightforward, reflecting the more engineering-oriented
performance characteris-
tics of the systems in question.
Information Quality: Measures of Information System Output
Rather than measure the quality of the system performance,
other 1/S researchers
have preferred to focus on the quality of the information system
output, namely, the
quality of the information that the system produces, primarily in
the form of reports.
Larcker and Lessig (1980) developed six questionnaire items to
measure the per-
ceived importance and usableness of information presented in
reports. Bailey and
Pearson (1983) proposed 39 system-related items for measuring
user satisfaction.
Among their ten most important items, in descending order of
importance, were
information accuracy, output timeliness, reliability,
completeness, relevance, preci-
sion, and currency.
In an early study, Ahituv (1980) incorporated five information
characteristics into
a multi-attribute utility measure of information value: accuracy,
timeliness, rele-
vance, aggregation, and formatting. Gallagher (1974) develoiJed
a semantic differen-
tial instrument to measure the value of a group of I/S reports.
That instrument
included measures of relevance, informativeness, usefulness,
and importance.
Munro and Davis (1977) used Gallagher's instrument to measure
a decision maker's
perceived value of information received from information
systems which were cre-
ated using different methods for determining information
requirements. Additional
information characteristics developed by Swanson (1974) to
measure MIS apprecia-
tion among user managers included uniqueness, conciseness,
clarity, and readability
measures. Zmud (1978) included report format as an
information quality measure in
his empirical work. Olson and Lucas (1982) proposed report
appearance and accu-
racy as measures of information quality in office automation
information systems.
Lastly, King and Epstein (1983) proposed multiple information
attributes to yield a
64 Information Systems Research 3 : 1
Information Systems Success
Authors
Bailey and Pearson
(1983)
BartiandHuff(l985)
Belardo. Kanvan. and
Wallace (1982)
Con kiln. Gotterer.
and Rick man
(1982)
Franz and Robey
(1986)
Goslar(1986)
Hiltz and Turoff
(1981)
Kxiebel and Raviv
(1982)
Lehman (1986)
Mahmood(1987)
Morey(1982)
Srinivasan(1985)
TABLE 1
Empirical Measures of System Quality
Description of Study
Overall I/S; 8 organizations,
32 managers
DSS: 9 organizations. 42
decision makers
Emergency management
DSS; 10 emergency
dispatchers
Transaction processing; one
organization
Specific I/S; 34 organizations.
118 user managers
Marketing DSS; 43 marketers
Electronic information
exchange system; 102 users
Academic information
system; one university
Overall I/S; 200 I/S directors
Specific I/S; 61 I/S managers
Manpower management
system; one branch of the
military
Computer-based modeling
systems; 29 firms
Type
Field
Field
Lab
U b
Field
Lab
Field
Case
Field
Field
Case
Field
Description of Measure(s)
(1) Convenience of access
(2) Flexibility of system
(3) Integration of systems
(4) Response time
Realization of user expectations
(1) Reliability
(2) Response time
(3) Ease of use
(4) Ease of learning
Response time
Perceived usefulnessofl/S
(12 items)
LJsefulness of DSS features
Usefulness of specific functions
(1) Resource utilization
(2) Investment utilization
I/S sophistication (use of new
technology)
Flexibility of system
Stored record error rate
(1) Response time
(2) System reliability
(3) System accessibility
composite measure of information value. The proposed
information attributes in-
cluded sufficiency, understandability, freedom from bias,
reliability, decision rele-
vance, comparability, and quantitativeness.
More recently, numerous information quality criteria have been
included within
the broad area of "User Information Satisfaction" (Iivari 1987;
Iivari and Koskela
1987). The Iivari-Koskela satisfaction measure included three
information quality
constructs: "informativeness" which consists of relevance,
comprehensiveness, re-
centness. accuracy, and credibility; "accessibility" which
consists of convenience,
timeliness, and interpretability; and "adaptability."
In Table 2. nine studies which included information quality
measures are shown.
Understandably, most measures of information quality are from
the perspective of
the user of this information and are thus faidy subjective in
character. Also, these
March 1992 65
DeLone • McLean
measures, while shown here as separate entities, are often
included as part of the
measurers of user satisfaction. The Bailey and Pearson (1983)
study is a good exam-
ple of this cross linkage.
Information Use: Recipient Consumption of the Output of an
Information System
The use of information system reports, or of management
science/operations re-
search models, is one of the most frequently reported measures
of the success of an
information system or an MS/OR model. Several researchers
(Lucas 1973; Schultz
andSlevin 1975; Ein-Dor and Segev 1978; Ives, Hamilton, and
Davis 1980; Hamil-
ton and Chervany 1981) have proposed I/S use as an MIS
success measure in concep-
tual MIS articles. Ein-Dor and Segev claimed that different
measures of computer
success are mutually interdependent and so they chose system
use as the primary
criterion variable for their I/S research framework. "Use of
system" was also an
integral part of Lucas's descriptive model of information
systems in the context of
organizations. Schultz and Slevin incorporated an item on the
probability of MS/OR
model use into their five-item instrument for measuring model
success.
In addition to these conceptual studies, the use of an
information system has often
been the MIS success measure of choice in MIS empirical
research (Zmud 1979). The
broad concept of use can be considered or measured from
several p>erspectives. It is
clear that actual use, as a measure of I/S success, only makes
sense for voluntary or
discretionary users as opposed to captive users (Lucas 1978;
Weike and Konsynski
1980). Recognizing this, Maish (1979) chose voluntary use of
computer terminals
and voluntary requests for additional reports as his measures of
I/S success. Similarly,
Kim and Lee (1986) measured voluntariness of use as part of
their measure of
success.
Some studies have computed actual use (as opposed to reported
use) by managers
through hardware monitors which have recorded the number of
computer inquiries
(Swanson 1974; Lucas 1973, 1978; King and Rodriguez 1978.
1981), or recorded the
amount of user connect time (Lucas 1978; Ginzberg 1981a).
Other objective mea-
sures of use were the number of computer functions utilized
(Ginzberg 1981 a), the
number of client records processed (Robey 1979), or the actual
charges for computer
use (Gremillion 1984). Still other studies adopted a subjective
or perceived mea-
sure of use by questioning managers about their use of an
information system
(Lucas 1973, 1975. 1978; Maish 1979; Fuerst and Cheney 1982;
Raymond 1985;
DeLone 1988).
Another issue concerning use of an information system is "Use
by whom?" (Huys-
mans 1970). In surveys of MIS success in small manufacturing
firms, DeLone (1988)
considered chief executive use of information systems while
Raymond (1985) consid-
ered use by company controllers. In an earlier study. Culnan
(1983a) considered both
direct use and chaufFeured use (i.e.. use through others).
There are also different levels of use or adoption. Ginzberg
(1978) discussed the
following levels of use, based on the earlier work by Huysmans;
(1) use that results in
management action, (2) use that creates change, and (3)
recurring use of the system.
Earlier, Vanlommel and DeBrabander (1975) proposed four
levels of use: use for
getting instructions, use for recording data, use for control, and
use for planning.
Schewe (1976) introduced two forms of use: general use of
"routinely generated
computer reports" and specific use of "personally initiated
requests for additional
66 Information Systems Research 3 :
Information Systems Success
TABLE 2
Empirical Measures of Information Quality
Authors Description of Study Type Description of Measure(s)
Bailey and Pearson
(1983)
BlaylockandRees(l984)
Jones and McLeod
(1986)
King and Epstein (1983)
Mahniood(1987)
Mahmood and Medewitz
(1985)
Mitlerand Doyle (1987)
RivardandHuff(l985)
Srinivasan (1985)
Overall I/S; 8 organizations.
32 managers
Financial; one university. 16
MBA students
Several information sources;
5 senior executives
Overall I/S; 2 firms. 76
managers
Specific I/S; 61 I/S managers
DSS; 48 graduate students
Overall 1/S; 21 financial
firms. 276 user managers
User-developed I/S; 10
firms, 272 users
Computer-based modeling
systems; 29 firms
Field Output
(1) Accuracy
(2) Precision
(3) Currency
(4) Timeliness
(5) Reliability
(6) Completeness
(7) Conciseness
(8) Format
(9) Relevance
Lab Perceived usefulness of specific
report items
Field Perceived importance of each
information item
Field Information
(1) Currency
(2) Sufficiency
(3) Understandability
(4) Freedom from bias
(5) Timeliness
(6) Reliability
(7) Relevance to decisions
(8) Comparability
(9) Quantitativeness
Field (I) Report accuracy
(2) Report timeliness
Lab Report usefulness
Field (1) Completeness of information
(2) Accuracy ofinformation
(3) Relevance of reports
(4) Timeliness of report
Field Usefulness ofinformation
Field (1) Report accuracy
(2) Report relevance
(3) Underslandability
(4) Report timeliness
information not ordinarily provided in routine reports." By this
definition, specific
use reflects a higher level of system utilization. Fuerst and
Cheney (1982) adopted
Schewe's classification of general use and specific use in the ir
study of decision sup-
port in the oil industry.
Bean et al. (1975); King and Rodriguez {1978, 1981), and
DeBrabander and Thiers
(1984) attempted to measure the nature of system use by
comparing this use to the
March 1992 67
DeLone • McLean
decision-making purpose for which the system was designed.
Similarly, livari {1985)
suggested appropriate use or acceptable use as a measure of
MIS success. In a study by
Robey and Zeller (1978), I/S success was equated to the
adoption and extensive use of
an information system.
After reviewing a number of empirical studies involving use,
Trice and Treacy
(1986) recommend three classes of utilization measures based
on theories from refer-
ence disciplines: degree of MIS institutionalization, a binary
measure of use vs. non-
use, and unobtrusive utilization measures such as connect time
and frequency of
computer acce^. The degree of institutionalization is to be
determined by user de-
pendence on the MIS, user feelings of system ownership, and
the degree to which MIS
is routinized into standard operating procedures.
Table 3 shows the 27 empirical studies which were found to
employ system use as
at least one of their measures of success. Of all the measures
identified, the system use
variable is probably the most objective and the easiest to
quantify, at least concep)-
tually. Assuming that the organization being studied is {1)
regularly monitoring such
usage patterns, and (2) willing to share these data with
researchers, then usage is a
fairly accessible measure of I/S success. However, as pointed
out earlier, usage, either
actual or perceived, is only pertinent when such use is
voluntary.
User Satisfaction: Recipient Response to the Use ofthe Output
of an
Information System
When the use of an information system is required, the
preceding measures be-
come less useful; and successful interaction by management
with the information
system can be measured in terms of user satisfaction. Several
I/S researchers have
suggested user satisfaction as a success measure for their
empirical I/S research (Ein-
Dor and Segev 1978; Hamilton and Chervany 1981). These
researchers have found
user satisfaction as especially appropriate when a specific
information system was
involved. Once again a key issue is whose satisfaction should be
measured. In at-
tempting to determine the success ofthe overall MIS effort,
McKinsey & Company
(1968) measured chief executives' satisfaction.
In two empirical studies on implementation success, Ginzberg
(1981a, b) chose
user satisfaction as his dependent variable. In one of those
studies (1981 a), he
adopted both use and user satisfaction measures. In a study by
Lucas (1978), sales
representatives rated their satisfaction with a new computer
system. Later, in a differ-
ent study, executives were asked in a laboratory setting to rate
their enjoyment and
satisfaction with an information system which aided decisions
relating to an inven-
tory ordering problem (Lucas 1981).
In the Powers and Dickson study on MIS project success (1973),
managers were
asked how well their information needs were being satisfied.
Then, in a study by King
and Epstein (1983), I/S value was imputed based on managers"
satisfaction ratings.
User satisfaction is also recommended as an appropriate success
measure in experi-
mental I/S research (Jarvenpaa, Dickson, and DeSanctis 1985)
and for researching
the effectiveness of group decision support systems (DeSanctis
and Gallupe 1987).
Other researchers have developed multi-attribute satisfaction
measures rather than
relying on a single overall satisfaction rating. Swanson (1974)
used 16 items to mea-
sure I/S appreciation, items which related to the characteristics
of reports and ofthe
underlying information system itself. Pearson developed a 39-
item instrument for
measuring user satisfaction. The full instrument is presented in
Bailey and Pearson
68 lnfortnation Systems Research 3 : 1
Information Systems Success
(1983), with an earlier version reviewed and evaluated by
KHebel (1979) and by Ives,
Olson, and Baroudi (1983). Raymond (1985) used a subset of 13
items from Pear-
son's questionnaire to measure manager satisfaction with MIS in
small manufactur-
ing firms. More recently, Sanders (1984) developed a
questionnaire and used it
(Sanders and Courtney 1985) to measure decision support
systems (DSS) success.
Sanders' overall success measure involves a number of measures
of user and decision-
making satisfaction.
Finally, studies have found that user satisfaction is associated
with user attitudes
toward computer systems (Igerhseim 1976; Lucas 1978) so that
user-satisfaction
measures may be biased by user computer attitudes. Therefore,
studies which include
user satisfaction as a success measure should ideally also
include measures of user
attitudes so that the potentially biasing effects of those attitudes
can be controlled for
in the analysis. Goodhue (1986) further suggests "information
satisfactoriness" as an
antecedent to and surrogate for user satisfaction. Information
satisfactoriness is de-
fined as the degree of match between task characteristics and
I/S functionality.
As the numerous entries in Table 4 make clear, user satisfaction
or user informa-
tion satisfaction is probably the most widely used single
measure of I/S success. The
reasons for this are at least threefold. First, "satisfaction" has a
high degree of face
validity. It is hard to deny the success ofa system which its
users say that they like.
Second, the development ofthe Bailey and Pearson instrument
and its derivatives
has provided a reliable tool for measuring satisfaction and for
making comparisons
among studies. The third reason for the appeal of satisfaction as
a success measure is
that most of the other measures are so poor; they are either
conceptually weak or
empirically difficult to obtain.
Individual Impact: The Effect ofinformation on the Behavior
ofthe Recipient
Of all the measures of I/S success, "impact" is probably the
most difficult to define
in a nonambiguous fashion. It is closely related to performance,
and so "improving
my—or my department's—performance" is certainly evidence
that the information
system has had a positive impact. However, "impact" could also
be an indication
that an information system has given the user a better
understanding ofthe decision
context, has improved his or her decision-making productivity,
has produced a
change in user activity, or has changed the decision maker's
perception ofthe impor-
tance or usefulness ofthe information system. As discussed
earlier. Mason (1978)
proposed a hierarchy of impact (influence) levels from the
receipt ofthe information,
through the understanding ofthe information, the application
ofthe information to a
specific problem, and the change in decision behavior, to a
resultant change in organi-
zational performance. As Emery (1971, p. I) states:
"Information has no intrinsic
value; any value comes only through the influence it may have
on physical events.
Such influence is typically exerted through human decision
makers."
In an extension ofthe traditional statistical theory ofinformation
value. Mock
(1971) argued for the importance ofthe "learning value
ofinformation." In a labora-
tory study ofthe impact ofthe mode ofinformation presentation,
Lucas and Nielsen
(1980) used learning, or rate of performance improvement, as a
dependent variable.
In another laboratory setting, Lucas (1981) tested participant
understanding ofthe
inventory problem and used the test scores as a measure of I/S
success. Watson and
Driver (1983) studied the impact of graphical presentation on
information recall.
Meador, Guyote, and Keen (1984) measured the impact of a
DSS design
March 1992
DeLone • McLean
TABLE 3
Empirical Measures ofinformation System Use
Authors Description of Study Type Description of Measure(s)
Alavi and Henderson
(1981)
Baroudi, Olson, and Ives
(1986)
BartiandHuff(1985)
Bell (1984)
Work force and production Lab
scheduling DSS; one
university. 45 graduates
Overall I/S; 200 finns, 200
production managers
DSS; 9 organizations, 42 Field
decision makers
Financial; 30 financial Lab
Use or nonuse of computer-based
decision aids
Field Use of 1/S to support production
Percentage of time DSS is used in
decision making situations
Use of numerical vs.
nonnumerical information
Benbasat. Dexter, and
Masulis(1981)
Bergeron (1986b)
Chandrasekaran and Kirs
(1986)
Culnan (1983a)
Culnan (1983b)
DeBrabander and Thiers
(1984)
DeSanctis (1982)
Ein-Dor. Segev, and
Steinfeld(l98l)
Green and Hughes(1986)
Fuerst and Cheney (1982)
Glnzberg(198la)
Hogue(1987)
Gremillion (1984)
Pricing: one university. 50
students and faculty
Overall I/S; 54
organizations. 471 user
managers
Reporting systems; MBA
students
Overall I/S; one
organization, 184
professionals
Overall I/S; 2
organizations, 362
professionals
Specialized DSS: one
university, 91 two-person
teams
DSS; 88 senior level
students
PERT: one R & D
organization, 24
managers
DSS; 63 city managers
DSS; 8 oil companies, 64
users
On-line portfolio
management system;
U.S. bank, 29 portfolio
managers
DSS; 18 organizations
Overall I/S; 66 units of tbe
National Forest system
Lab
Field
Field
Field
Field
U b
Lab
Field
Lab
Field
Field
Field
Field
Frequency of requests for specific
reports
Use of chargeback information
Acceptance of report
(I) Direct use ofl/S vs.
chaufieured use
(2) Number of requests for
information
Frequency of use
Use vs. nonuse of data sets
Motivation to use
(I) Frequency of past use
(2) Frequency of intended use
Number of DSS features used
(I) Frequency of general use
(2) Frequency of specific use
(1) Number of minutes
(2) Number of sessions
(3) Number of functions used
Frequency of voluntary use
Expenditures/charges for
computing use
70 Information Systems Research 3 : 1
Information Systems Success
Authors
Kim and Lee (1986)
King and Rodriguez
(1981)
Mahmood and Medewitz
(1985)
Nelson and Cbeney
(1987)
Perry (1983)
Raymond (1985)
Snitkin and King (1986)
Srinivasan (1985)
Swanson (1987)
Zmud. Boynton. and
Jacobs(1987)
TABLE 3 {cont d)
Description of Study
Overall [/S; 32
organizations, 132 users
Strategic system; one
university. 45 managers
DSS; 48 graduate students
Overall I/S; 100 top/middle
managers
Office 1/S; 53 firms
Overall 1/S; 464 small
manufacturing Brms
Personal DSS; 31 users
Computer-based modeling
systems; 29 firms
Overall I/S; 4
organizations, 182 users
Overall I/S; Sample A: 132
firms
Sample B: one firm
Type
Field
Lab
Lab
Field
Field
Field
Field
Field
Field
Field
Description of Measure(s)
(1) Frequency of use
(2) Voluntariness of use
(1) Number of queries
(2) Nature of queries
Extent of use
Extent of use
Use at anticipated level
(1) Frequency of use
(2) Regularity of use
Hours per week
(1) Frequency of use
(2) Time per computer session
(3) Number of reports generated
Average frequency with which
user discussed report
information
Use in support of
(a) Cost reduction
(b) Management
(c) Strategy planning
(d) Competitive thrust
methodology using questionnaire items relating to resulting
decision effectiveness.
For example, one questionnaire item referred specifically to the
subject's perception
of the improvement in his or her decisions.
In the information system framework proposed by Chervany,
Dickson, and Kozar
(1972), which served as the model for the Minnesota
Experiments (Dickson. Cher-
vany. and Senn 1977). the dependent success variable was
generally defined to be
decision effectiveness. Within the context of laboratory
experiments, decision effec-
tiveness can take on numerous dimensions. Some of these
dimensions which have
been reported in laboratory studies include the average time to
make a decision
(Benbasat and Dexter 1979, 1985; Benbasat and Schroeder
1977; Chervany and
Dickson 1974; Taylor 1975), the confidence in the decision
made (Chervany and
Dickson 1974; Taylor 1975), and the number of reports
requested (Benbasat and
Dexter 1979; Benbasat and Schroeder 1977). DeSanctis and
Gallupe (1987) sug-
gested member participation in decision making as a measure of
decision effective-
ness in group decision making.
In a study which sought to measure the success of user-
developed applications,
Rivard and Huff (1984) included increased user productivity in
their measure of
success. DeBrabander and Thiers (1984) used efficiency of task
accomplishment
(time required to find a correct answer) as the dependent
variable in their laboratory
March 1992 71
DeLone • McLean
TABLE 4
Empirical Measures of User Satisfaction
Author(s) Description of Study Type Description of Measure(s)
Alavi and Henderson
(1981)
Baitey and Pearson
(1983)
Baroudi, Olson, and
Ives(1986)
Barti and Huff(l985)
Work force and production
scheduling DSS; one
university; 45 graduate
students
Overall I/S; 8 organizations.
32 managers
Overall I/S; 200 firms. 200
production managers
DSS; 9 organizations, 42
decision makers
Lab Overall satisfaction witb DSS
Field User satisfaction (39-item instrument)
Field User information satisfaction
Field User information satisfaction
(modified Bailey & Pearson
instrument)
Bruwer(1984)
Cats-Baril and Huber
(1987)
DeSanctis (1986)
Doll and Ahmed (1985)
Edmundson and JefFery
(1984)
Ginzberg (1981a)
Ginzberg(l981b)
Hogue(1987)
Ives. Olson, and
Baroudi(1983)
Jenkins, Naumann, and
Wetherbe (1984)
King and Epstein (1983)
Langle, Leitheiser, and
Naumann (1984)
Lehman, Van Wetering.
and Vogel (1986)
Lucas(1981)
Overall I/S; one organization,
114 managers
DSS; one university, 101
students
Human resources 1/S; 171
human resource system
professionals
Specificl/S; 55 firms, 154
managers
Accounting software package;
12 organizations
On-line portfolio
management system; U.S.
bank. 29 portfolio managers
Overall I/S; 35 I/S users
DSS; 18 organizations
Overall I/S; 200 firms, 200
production managers
A specific I/S; 23
corporations, 72 systems
development managers
Overall I/S; 2 firms. 76
managers
Overall 1/S; 78 oi^nizations,
I/S development managers
Business graphics; 200
organizations, DP managers
Inventory ordering system;
one university, 100
executives
Field
Lab
Field
Field
Field
Field
Field
Field
Field
Field
Field
Field
Field
U b
User satisfaction
Satisfaction with a DSS (multi-item
scale)
(1) Top management satisfaction
(2) Personal management satisfaction
User satisfaction (11 -item scale)
User satisfaction (1 question)
Overall satisfaction
Overall satisfaction
User satisfaction (1 question)
User satisfaction (Bailey & Pearson
instrument)
User satisfaction (25-item instrument)
User satisfaction (1 item: scale 0 to
100)
User satisfaction (1 question)
(1) Software satisfaction
(2) Hardware satisfaction
(1) Enjoyment
(2) Satisfaction
72 Information Systems Research 3 : 1
Information Systems Success
TABLE 4
Author(s) Description of Study Type Description of Measure(s)
Mahmood (1987)
Mahmood and Becker
(1985-1986)
Mahmood and
Medewitz (1985)
McKeen(l983)
Nelson and Cheney
(1987)
Olson and Ives (1981)
Olson and Ives (1982)
Raymond (1985)
Raymond (1987)
Rivard and Huff (1984)
Rushinek and Rushinek
(1985)
Rushinek and Rushinek
(1986)
Sanders and Courtney
(1985)
Sanders, Courtney, and
Uy(I984)
Taylor and Wang(1987)
Specificl/S; 61 I/S managers Field Overall satisfaction
Overall I/S; 59 firms. 118 Field User satisfaction
managers
DSS; 48 graduate students Lab User satisfaction (multi-Item
scale)
Application systems; 5
organizations
Overall I/S; 100 top/middle
managers
Field Satisfaction with the development
project (Powers and Dickson
instrument)
Field User satisfaction (Bailey & Pearson
instrument)
Overall I/S; 23 manufacturing Field Information dissatisfaction
difference
firms. 83 users between information needed and
amount ofinformation received
Overall I/S; 23 manufacturing Field Information satisfaction,
difference
firms, 83 users between information needed and
information received
Overall I/S; 464 small
manufacturing firms
Overall I/S; 464 small-firm
finance managers
User-developed applications;
10 large companies
Accounting and billing
system; 4448 users
Overall I/S; 4448 users
Financial DSS; 124
organizations
Field Controller satisfaction (modified
Bailey & Pearson instrument)
Field User satisfaction (modified Bailey &
Pearson instrument)
Field User complaints regarding
Information Center services
Field Overall user satisfaction
Field Overall user satisfaction
Field (1) Overall satisfaction
(2) Decision-making satisfaction
Interactive Financial Planning Field (I) Decision-making
satisfaction
System (IFPS); 124 (2) Overall satisfaction
oi^nizations, 373 users
DBMS with multiple dialogue Lab User satisfaction with
interface
modes; one university, 93
students
experiment. Finally, Sanders and Courtney (1985) adopted the
speed of decision
analysis resulting from DSS as one item in their DSS success
measurement in-
strument.
Mason (1978) has suggested that one method of measuring I/S
impact is to deter-
mine whether the output ofthe system causes the receiver (i.e.,
the decision maker) to
change his or her behavior. Ein-Dor, Segev, and Steinfeld
(1981) asked decision
makers: "Did use of PERT [a specific information system] ever
lead to a change in a
March 1992 73
DeLone • McLean
decision or to a new decision?" Judd, Paddock, and Wetherbe
(1981) measured
whether a budget exception reporting system resulted in
managers' taking investiga-
tive action.
Another approach to the measurement ofthe impact of an
information system is to
ask user managers to estimate the value of the information
system. Cerullo (1980)
asked managers to rank the value of their computer-based MIS
on a scale of one to
ten. Ronen and Falk (1973) asked participants to rank the value
ofinformation
received in an experimental decision context. Using success
items developed by
Schultz and Slevin (1975), King and Rodriguez (1978. 1981)
asked users of their
"Strategic Issue Competitive Information System" to rate the
worth of that 1/S.
Other researchers have gone a step further by asking
respondents to place a dollar
value on the information received. Gallagher (1974) asked
managers about the maxi-
mum amount they would be willing to pay for a particular
report. Lucas (1978)
reported using willingness to pay for an information system as
one of his success
measures. Keen (1981) incorporated willingness to pay
development costs for im-
proved DSS capability in his proposed "Value Analysis" for
justification ofa DSS. In
an experiment involving MBA students, Hilton and Swieringa
(1982) measured what
participants were willing to pay for specific information which
they felt would lead to
higher decision payoffs. Earlier. Garrity (1963) used MIS
expenditures as a percent-
age of annual capital expenditures to estimate the value ofthe
MIS effort.
Table 5, with 39 entries, contains the largest number of
empirical studies. This in
itself is a healthy sign, for it represents an attempt to move
beyond the earlier inward-
looking measures to those which offer the potential to gauge the
contribution of
information systems to the success ofthe enterprise. Also worth
noting is the predomi-
nance of laboratory studies. Whereas most ofthe entries in the
preceding tables have
been field experiments, 24 ofthe 39 studies reported here have
used controlled labora-
tory experiments as a setting for measuring the impact
ofinformation on individuals.
The increased experimental rigor which laboratory studies offer,
and the extent to
which they have been utilized at least in this success category,
is an encouraging sign
for the maturing of the field.
Organizational Impact: The Effect of Information on
Organizational
Performance
In a survey by Dickson, Leitheiser. Wetherbe, and Nechis
(1984). 54 information
systems professionals ranked the measurement ofinformation
system effectiveness
as the fifth most important I/S issue for the 1980s. In a recent
update of that study by
Brancheau and Wetherbe (1987). I/S professionals ranked
measurement ofinforma-
tion system effectiveness as the ninth most important I/S issue.
Measures of individ-
ual performance and, to a greater extent, organization
performance are of consider-
able importance to I/S practitioners. On the other hand. MIS
academic researchers
have tended to avoid performance measures (except in
laboratory studies) because of
the difficulty of isolating the effect ofthe I/S effort from other
effects which influence
organizational performance.
As discussed in the previous section, the effect of an
information system on individ-
ual decision performance has been studied primarily in
laboratory experiments using
students and computer simulations. Many of these experiments
were conducted at
the University of Minnesota (Dickson, Chervany. and Senn
1977). Among these
"Minnesota Experiments" were some that studied the effects of
different information
74 Information Systems Research 3 : 1
Information Systems Success
formats and presentation modes on decision performance as
measured in terms of
lower production, inventory, or purchasing costs. King and
Rodriguez (1978. 1981)
measured decision performance by evaluating participant tests
responses to various
hypothesized strategic problems.
In another laboratory study. Lucas and Nielsen (1980) measured
participant perfor-
mance (and thus, indirectly, organizational performance) in
terms of profits in a
logistics management game. In a later experiment, Lucas (1981)
investigated the
efiect of computer graphics on decisions involving inventory
ordering. Finally,
Remus (1984) used the costs of various scheduling decisions to
evaluate the effects of
graphical versus tabular displays.
Field studies and case studies which have dealt with the
influence ofinformation
systems have chosen various organizational performance
measures for their depen-
dent variable. In their study, Chervany, Dickson, and Kozar
(1972) chose cost reduc-
tions as their dependent variable. Emery (1971. p. 6) has
observed that: "Benefits
from an information system can come from a variety of sources.
An important one is
the reduction in operating costs of activities external to the
information processing
system."
Several researchers have suggested that the success of the MIS
department is re-
flected in the extent to which the computer is applied to critical
or major problem
areas ofthe firm (Garrity 1963; Couger and Wergin 1974; Ein-
Dor and Segev 1978;
Rockart 1979; Senn and Gibson 1981). In Garrity's early article
(1963), company I/S
operations were ranked partly on the basis o f t h e range and
scope of its computer
applications. In a later McKinsey study (1968), the authors used
the range of "mean-
ingful, functional computer applications" to distinguish between
more or less success-
ful MIS departments. In a similar vein, Vanlommel and
DeBrabander (1975) used a
weighted summation ofthe number of computer applications as a
measure of MIS
success in small firms. Finally, Cerullo (1980) ranked MIS
success on the basis o f a
firms's ability to computerize high complexity applications.
In a survey of several large companies, Rivard and Huff (1984)
interviewed data
processing executives and asked them to assess the cost
reductions and company
profits realized from specific user-developed application
programs. Lucas (1973) and
Hamilton and Chervany (1981) suggested that company
revenues can also be im-
proved by computer-based information systems. In a study o f a
clothing manufac-
turer, Lucas (1975) used total dollar bookings as his measure of
organizational perfor-
mance. Chismar and Kriebel (1985) proposed measuring the
relative efficiency ofthe
information systems effort by applying Data Envelopment
Analysis to measure the
relationship of corporate outcomes such as total sales and return
on investment to I/S
inputs.
More comprehensive studies ofthe effect of computers on an
organization include
both revenue and cost issues, within a cost/benefit analysis
(Emery 1971). McFadden
(1977) developed and demonstrated a detailed computer
cost/benefit analysis using a
mail order business as an example. In a paper entitled "What is
the Value of Invest-
ment in Information Systems?," Matlin (1979) presented a
detailed reporting system
for the measurement ofthe value and costs associated with an
information system.
Cost/benefit analyses are often found lacking due to the
difficulty of quantifying
"intangible benefits." Building on Keen's Value Analysis
approach (1981). Money,
Tromp. and Wegner (1988) proposed a methodology for
identifying and quantifying
March 1992 75
DeLone • McLean
Authors)
Aldag and Power
(1986)
Belardo, Kanvan,
and Wallace
(1982)
Benbasat and
Dexter (1985)
Benbasat and
Dexter (1986)
Benbasat. Dexter,
and Masulis
(1981)
Bergeron (1986a)
Cats-Baril and
Huber (1987)
Crawford (1982)
DeBrabanderand
Thiers(l984)
DeSanctis and
Jarvenpaa
(1985)
Dickson,
DeSanctis, and
McBride
(1986)
Drury(1982)
Ein-Dor, Segev,
and Steinfeld
(1981)
TABLE 5
Empiricat Measures of Individual Impact
Description of Study
DSS; 88 business
students
Emergency
management DSS;
10 emergency
dispatchers
Financiai; 65
business students
Financial; 65
business students
Pricing; one
university, 50
students and
factuly
DP chargeback
system; 54
organizations, 263
user managers
DSS; one university,
101 students
Electronic mail;
computer vendor
organization
Specialized DSS; one
university, 91 two-
person teams
Tables vs. graphs; 75
MBA students
Graphics system; 840
undei^raduate
students
Chargeback system;
173 organizations,
senior DP
managers
PERT; one R & D
organization. 24
managers
Type
Lab
U b
Lab
U b
Lab
Field
Lab
Case
Lab
U b
U b
Field
Field
Description of Measure(s)
(1) User confidence
(2) Quality of decision
analysis
(1) Efficient decisions
(2) Time to arrive at a
decision
Time taken to complete a task
Time taken to complete a task
Time to make pricing
decisions
Extent to which users analyze
charges and investigate
budget variances
(1) Quality of career plans
(2) Number of objectives and
• alternatives generated
Improved personal
productivity, hrs/wk/
manager
(1) Time efficiency of task
accomplishment
(2) User adherence to plan
Decision quality, forecast
accuracy
(1) Interpretation accuracy
(2) Decision quality
(1) Computer awareness
(2) Cost awareness
Change in decision behavior
76 Information Systems Research 3 : 1
Information Systems Success
TABLE 5 ( a w / y )
Author(s) Description of Study Type Description of Measure(s)
Fuerst and
Cheney(1982)
Goslar, Green,
and Hughes
(1986)
Goul, Shane, and
Tonge(1986)
Green and
Hughes (1986)
DSS; 8 oil
companies, 64
users
DSS: 19
organizations, 43
sales and
marketing
personnel
Knowledge-based
DSS; one
university, 52
students
DSS; 63 city
managers
Field Value in assisting decision
making
Lab (I) Number of altematives
considered
(2) Time to decision
(3) Confidence in decision
(4) Ability to identify
solutions
L ^ Ability to identify strategic
opoortunities or problems
Lab (I) Time to decision
(2) Number of alternatives
considered
(3) Amount of data
considered
Grudmtski(I98l)
Gueutal.
Surprenant,
and Bubeck
(1984)
Hilton and
Swieringa
(1982)
Hughes (1987)
Judd. Paddock,
and Wetherbe
(1981)
Kaspar(l985)
King and
Rodriguez
(1981)
Lee,
MacLachlan.
and Wallace
(1986)
Lucas(1981)
Planning and control
system: 65
business students
Computer-aided
design system; 69
students
General !/S; one
university, 56
MBA students
DSS generator; 63
managers
Budget exception
reporting system;
116 MBA students
DSS; 40 graduate
students
Strategic system; one
university. 45
managers
Performance I/S; 45
naarketing students
Inventory ordering
system; one
university, 100
executives
Lab
Lab
Lab
Field
Lab
U b
U b
U b
U b
Precision of decision maker's
forecast
(I) Task performance
(2) Confidence in
performance
Dollar value ofinformation
(1) Time to reach decision
(2) Number of alternatives
considered
Management takes
investigative action
Ability to forecast firm
performance
(1) Worth of information
system
(2) Quality of policy decisions
(I) Accuracy of information
interpretation
(2) Time to solve problem
User understanding of
inventory problem
March 1992 77
DeLone • Mcl^an
Author(s)
Lucas and Palley
(1987)
Luzi and
Mackenzie
(1982)
Meador, Guyote,
and Keen
(1984)
Millman and
Hanwick
(1987)
Rivard and Huff
(1984)
Rivard and Huff
(1985)
Sanders and
Courtney
(1985)
Snitkin and King
(1986)
Srinivasan (1985)
Vogel. Lehman,
and Dickson
(1986)
Watson and
Driver (1983)
Zmud (1983)
Zmud. Blocher.
and Moffie
(1983)
TABLE
Description of Study
Overall 1/S; 3
manufacturing
firms, 87 plant
managers
Performance
information
system; one
university, 200
business students
DSS; 18 firms, 73
users
Office I/S; 75 middle
managers
User-developed
applications; 10
large companies
User-developed I/S;
10 firms 272 users
Financial DSS; 124
oi^nizations
Personal DSS; 31
users
Computer-based
modeling systems;
29 firms
Graphical
Presentation
System; 174
undergraduate
students
Graphical
presentation of
information; 29
undergraduate
business students
External information
channels; 49
software
development
managers
Invoicing system; 51
internal auditors
5 (cont'd)
Type
Field
U b
Field
Field
Field
Field
Field
Field
Field
U b
U b
Field
U b
Description of Measure(s)
(1) Power of I/S department
(2) Influence of I/S
department
(1) Time to solve problem
(2) Accuracy of problem
solution
(3) Ffficiency of effort
(1) Effectiveness in
supporting decisions
(2) Time savings
Personal effectiveness
User productivity
Productivity improvement
Decision-making efficiency
and effectiveness
Effectiveness of personal DSS
(1) Problem identification
(2) Generation of alternatives
Change in commitment of
time and money
(1) Immediate recall of
information
(2) Delayed recall of
information
Recognition and use of
modern software practices
(1) Decision accuracy
(2) Decision confidence
78 Information Systems Research 3 ; 1
Information Systems Success
intangible benefits. The proposed methodology then applied a
statistical test to deter-
mine whether "significant" value can be attached to a decision
support system.
With the corporate "bottom line" in mind, several MIS
frameworks have proposed
that MIS effectiveness be determined by its contribution to
company profits (Cher-
vany, Dickson. and Kozar 1972; Lucas 1973; Hamilton and
Chervany 1981), but few
empirical studies have attempted to measure actual profit
contribution. Ferguson
and Jones (1969) based their evaluation of success on more
profitable job schedules
which resulted from decision-maker use ofthe information
system. Ein-Dor, Segev,
and Steinfeld (1981) attempted to measure contribution to profit
by asking users ofa
PERT system what savings were realized from use of PERT and
what costs were
incurred by using PERT.
Another measure of organizational performance which might be
appropriate for
measuring the contribution of MIS is return on investment. Both
Garrity (1963) and
the McKinsey study (1968) reported using return on investment
calculations to as-
sess the success of corporate MIS efforts. Jenster (1987)
included nonfinancial mea-
sures of organizational impact in a field study of 124
organizations. He included
productivity, innovations, and product quality among his
measures of I/S success. In
a study of 53 firms, Perry (1983) measured the extent to which
an office information
system contributed to meeting organizational goals.
Strassmann, in his book Information Payoffil9S5), presented a
particularly com-
prehensive view of the role of information systems with regards
to performance,
looking at it from the perspective ofthe individual, the
organization, the top execu-
tive, and society. His measure of performance was a specially
constructed "Return on
Management" (ROM) metric.
In nonprofit organizations, specifically government agencies,
Danziger (i 977) pro-
posed using productivity gains as the measure ofinformation
systems impact on the
organization. He explained that productivity gains occur when
the "functional out-
put ofthe government is increased at the same or increased
quality with the same or
reduced resources inputs" (p. 213). In a presentation of several
empirical studies
conducted by the University of California, Irvine, Danziger
included five productiv-
ity measures: staff reduction, cost reduction, increased work
volume, new informa-
tion, and increased effectiveness in serving the public.
The success of informati on systems in creating competitive
advantage has
prompted researchers to study I/S impacts not only on firm
performance but also on
industry structure (Clemons and Kimbrough 1986). Bakos
(1987) reviewed the litera-
ture on the impacts of information technology on firm and
industry-level perfor-
mance from the perspective of organization theory and
industrial economics. At the
firm level, he suggested measures of changes in organizational
structure and of im-
provements in process efficiency using Data Envelopment
Analysis (Chismar and
Kriebel 1985) as well as other financial measures. At the
industry level, he found
impact measures (e.g., economies of scale, scope, and market
concentration) harder
to identify in any readily quantifiable fashion and suggested
that further work is
needed.
Johnston and Vitale (1988) have proposed a modified
cost/benefit analysis ap-
proach to measure the effects of interorganizational systems.
Traditional cost/benefit
analysis is applied to identify quantifiable benefits such as cost
reductions, fee reve-
nues, and increased product sales. Once the quantifiable costs
and benefits have been
identified and compared, Johnston and Vitale suggest that top
management use
March 1992 79
DeLone • McLean
judgment to assess the value ofthe benefits which are more
difficult to quantify such
as reduction of overhead, increases in customer switching costs,
barriers to new firm
entry, and product differentiation.
Table 6 is the last ofthe six tables summarizing the I/S success
measures identified
in this paper. Somewhat surprisingly, 20 empirical studies were
found, with 13 using
field-based measures (as opposed to the laboratory experiments
characterizing the
individual impacts) to get at the real-world effects of the impact
of information
systems on organizational performance. However, this is only a
beginning; and it is in
this area, "assessing the business value of information systems,"
where much work
needs to be done.
Discussion
In reviewing the various approaches that I/S researcher s have
taken in measuring
MIS success, the following observations emerge.
1. As these research studies .show, the I/S researcher has a
broad list of individual
dependent variables from which to choose.
It is apparent that there is no consensus on the measure of
information systems
success. Just as there are many steps in the production and
dissemination ofinforma-
tion, so too are there many variables which can be used as
measures of "I/S success."
In Table 7, all ofthe variables identified in each ofthe six
success categories discussed
in the preceding sections are listed. These include success
variables which have been
suggested but never used empirically as well as those that have
actually been used in
experiments.
In reviewing these variables, no single measure is intrinsically
better than another;
so the choice ofa success variable is often a function ofthe
objective ofthe study, the
organizational context, the aspect ofthe information system
which is addressed by
the study, the independent variables under investigation, the
research method, and
the level of analysis, i.e., individual, organization, or society
(Markus and Robey
1988). However, this proliferation of measures has been
overdone. Some consolida-
tion is needed.
2. Progress toward an MIS cumidative tradition dictates a
significant reduction in
the number of different dependent variable measures so that
research results can be
compared.
One of the major purposes of this paper is the attempt to reduce
the myriad of
variables shown in Table 7 to a more manageable taxomony.
However, within each
of these major success categories, a number of variables still
exist. The existence of so
many different success measures makes it difficult to compare
the results of similar
studies and to build a cumulative body of empirical knowledge.
There are, however,
examples of researchers who have adopted measurement
instruments developed in
earlier studies.
Ives, Olson, and Baroudi (1983) have tested the validity and
reliability ofthe
user-satisfaction questionnaire developed by Bailey and Pearson
(1983) and used that
instrument in an empirical study of user involvement (Baroudi,
Olson and Ives
1986). Raymond (1985. 1987) used a subset ofthe Bailey and
Pearson user-satisfac-
tion instrument to study MIS success in small manufacturing
firms. Similarly, Mah-
mood and Becker (1986) and Nelson and Cheney (1987) have
used the Bailey and
Pearson instrument in empirical studies. In another vein,
McKeen (1983) adopted
80 Information Systems Research 3 : 1
Information Systems Success
the Powers and Dickson (1973) satisfaction scale to measure the
success of I/S devel-
opment strategies.
King and Rodriguez (1978, 1981), Robey (1979), Sanders
(1984). and Sanders and
Courtney (1985) have adopted parts ofa measurement
instrument which Schultz and
Slevin (1975) developed to measure user attitudes and
perceptions about the value of
operations research models. Munro and Davis (1977) and Zmud
(1978) utilized
Gallaghef s questionnaire items (1974) to measure the perceived
value of an informa-
tion system. Finally, Blaylock and Rees (1984) used Larcker
and Lessig's 40 informa-
tion items (1980) to measure perceived information usefulness.
These are encouraging trends. More MIS researchers should
seek out success mea-
sures that have been developed, validated, and applied in
previous empirical re-
search.
3. Not enough MIS field study research attempts to measure the
influence ofthe
MIS effort on organizational performance.
Attempts to measure MIS impact on overall organizatio nal
performance are not
often undertaken because ofthe difficulty of isolating the
contribution ofthe infor-
mation systems function from other contributors to
organizational performance.
Nevertheless, this connection is of great interest to information
system practitioners
and to top corporate management. MIS organizational
performance measurement
deserves further development and testing.
Cost/benefit schemes such as those presented by Emery (1971),
McFadden (1977),
and MatHn (1979) offer promising avenues for further study.
The University of Cali-
fornia, Irvine, research on the impact ofinformation systems on
government activity
(Danziger 1987) suggests useful impact measures for public as
well as private organi-
zations. Lucas (1975) included organizational performance in
his descriptive model
and then operationalized this variable by including changes in
sale revenues as an
explicit variable in his field study ofa clothing manufacturer.
Garrity (1963) and the
McKinsey & Company study (1968) reported on early attempts
to identify MIS
returns on investment. McLean (1989), however, pointed out the
difficulties with
these approaches, while at the same time attempting to define a
framework for such
analyses. Strassmann (1985) has developed his "Return on
Management" metric as a
way to assess the overall impact ofinformation systems on
companies.
These research efforts represent promising beginnings in
measuring MIS impact
on performance.
4. The six success categories and the many specific I/S
measures within each of
these categories clearly indicaie that MIS success is a
multidimensional construct and
that it should he measured as such.
Vanlomme! and DeBrabander (1975) early pointed out that the
success ofa com-
puter-based information system is not a homoge neous concept
and therefore the
attempt should not be made to capture it by a simple measure.
Ein-Dor and Segev
(1978) admitted that their selection of MIS use as their
dependent variable may not
be ideal. They stated that "A better measure of MIS success
would probably be some
weighted average for the criteria mentioned above" (i.e.. use,
profitability, applica-
tion to major problems, performance, resulting quality decision,
and user satis-
faction).
In reviewing the empirical studies cited in Tables 1 through 6. it
is clear that most
of them have attempted to measure I/S success in only one or
possibly two success
March 1992 81
DeLone • McLean
TABLE 6
Measures of Organizational Impact
Author(s) Description of Study Type Description of Measure(s)
Benbasat and IDexter
(1985)
Benbasat and Dexter
(1986)
Benbasat. Dexter, and
Masulis(198l)
Bender(1986)
CronandSobol(1983)
Edelman(I981)
Ein-Dor. Segev. and
Steinfeld (1981)
Griese and Kurpicz
(1985)
Jenster (1987)
Kaspar and Cerveny
(1985)
Lincoln (1986)
Lucas(I981)
Miller and Doyle
(1987)
Miilman and Hartwick
(1987)
Perry (1983)
Remus (1984)
Financial; 65 business
students
Financial; 65 business
students
Pricing; one university. 50
students and faculty
Overall I/S; 132 life
insurance companies
Overall I/S; 138 small to
medium-sized
wbolesalers
Industrial relations; one
firm, 14 operating units
PERT; one R & D
organization 24 managers
Overall I/S; 69 firms
I/S which monitors critical
success factors; 124
organizations
End user systems; 96 MBA
students
Specific 1/S applications; 20
organizations. 167
applications
Inventory ordering system:
one university, 100
executives
Overall I/S; 21 financial
firms, 276 user managers
Office I/S; 75 middle
managers
Office 1/S; 53 firms
Production scheduling
system; one university,
53 junior business
students
Lab Profit performance
Lab Profit performance
Lab Profit
Field Ratio of total general expense to total
premium income
Field (I) Pretax return on assets
(2) Return on net worth
(3) Pretax profits (% of sales)
(4) Average 5-year sales growth
Field Overall manager productivity (cost of
information per employee)
Field Profitability
Field Number of computer applications
Field (I) Economic performance
(2) Marketing achievements
(3) Productivity in production
(4) Innovations
(5) Product and management quality
Lab (1) Return on assets
(2) Market share
(3) Stock price
Field (I) Internal rate of return
(2) Cost-benefit ratio
L^b Inventory ordering costs
Field Overall cost-effectiveness of I/S
Field Organizational effectiveness
Field I/S contribution to meeting goals
Lab Production scheduling costs
82 Information Systems Research 3 :
Information Systems Success
TABLE 6 (com "rf)
Author(s) Description of Study Type Description of Measure(s)
Rivard andHuff(1984)
Turner (1982)
Vasarhelyi(l981)
Yap and Walsham
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions
httpsdoi.org10.11770963721418763931Current Directions

More Related Content

Similar to httpsdoi.org10.11770963721418763931Current Directions

Aleph 2015-2016 text (1)
Aleph 2015-2016 text (1)Aleph 2015-2016 text (1)
Aleph 2015-2016 text (1)Min Woo Sun
 
A Race And Power Perspective On Police Brutality In America
A Race And Power Perspective On Police Brutality In AmericaA Race And Power Perspective On Police Brutality In America
A Race And Power Perspective On Police Brutality In AmericaLisa Muthukumar
 
PO 450 Final Research Project - Alejandro Luna
PO 450 Final Research Project - Alejandro LunaPO 450 Final Research Project - Alejandro Luna
PO 450 Final Research Project - Alejandro LunaAlejandro Luna
 
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docx
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docxRunning head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docx
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docxcharisellington63520
 
Spencer GrantPhotolibrary.docx
Spencer GrantPhotolibrary.docxSpencer GrantPhotolibrary.docx
Spencer GrantPhotolibrary.docxwhitneyleman54422
 
RUNNING HEAD ROUGH DRAFT .docx
RUNNING HEAD ROUGH DRAFT                                 .docxRUNNING HEAD ROUGH DRAFT                                 .docx
RUNNING HEAD ROUGH DRAFT .docxtoltonkendal
 
1Writing Activity 4 Final DraftShaland.docx
1Writing Activity 4  Final DraftShaland.docx1Writing Activity 4  Final DraftShaland.docx
1Writing Activity 4 Final DraftShaland.docxlorainedeserre
 
Sociology Final Paper
Sociology Final PaperSociology Final Paper
Sociology Final PaperSonja Oswald
 
Term Paper Essay Assignment Rubric Due by Saturday May 12th.docx
Term Paper Essay Assignment Rubric  Due by Saturday May 12th.docxTerm Paper Essay Assignment Rubric  Due by Saturday May 12th.docx
Term Paper Essay Assignment Rubric Due by Saturday May 12th.docxjacqueliner9
 
Running head RACE AND CRIME .docx
Running head RACE AND CRIME                                    .docxRunning head RACE AND CRIME                                    .docx
Running head RACE AND CRIME .docxjeanettehully
 
RUNNING HEAD Racial bias and double standards .docx
RUNNING HEAD Racial bias and double standards                    .docxRUNNING HEAD Racial bias and double standards                    .docx
RUNNING HEAD Racial bias and double standards .docxcharisellington63520
 
Racial Profiling in the Black Community
Racial Profiling in the Black CommunityRacial Profiling in the Black Community
Racial Profiling in the Black CommunityPATRICK MAELO
 
BShyenna Vega Professor Artist Writing for Crimina.docx
BShyenna Vega Professor Artist Writing for Crimina.docxBShyenna Vega Professor Artist Writing for Crimina.docx
BShyenna Vega Professor Artist Writing for Crimina.docxrichardnorman90310
 
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docx
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docxJones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docx
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docxpriestmanmable
 
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docx
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docxRunning head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docx
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docxtodd581
 
Ethnicity (Crime and Deviance)
Ethnicity (Crime and Deviance)Ethnicity (Crime and Deviance)
Ethnicity (Crime and Deviance)Beth Lee
 

Similar to httpsdoi.org10.11770963721418763931Current Directions (20)

Aleph 2015-2016 text (1)
Aleph 2015-2016 text (1)Aleph 2015-2016 text (1)
Aleph 2015-2016 text (1)
 
Criminal Justice Commission Chair Statement
Criminal Justice Commission Chair StatementCriminal Justice Commission Chair Statement
Criminal Justice Commission Chair Statement
 
A Race And Power Perspective On Police Brutality In America
A Race And Power Perspective On Police Brutality In AmericaA Race And Power Perspective On Police Brutality In America
A Race And Power Perspective On Police Brutality In America
 
PO 450 Final Research Project - Alejandro Luna
PO 450 Final Research Project - Alejandro LunaPO 450 Final Research Project - Alejandro Luna
PO 450 Final Research Project - Alejandro Luna
 
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docx
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docxRunning head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docx
Running head PUBLIC OPINION OF POLICE BY DIFFERENT ETHNIC GROUPS .docx
 
Spencer GrantPhotolibrary.docx
Spencer GrantPhotolibrary.docxSpencer GrantPhotolibrary.docx
Spencer GrantPhotolibrary.docx
 
SociologyExchange.co.uk Shared Resource
SociologyExchange.co.uk Shared ResourceSociologyExchange.co.uk Shared Resource
SociologyExchange.co.uk Shared Resource
 
RUNNING HEAD ROUGH DRAFT .docx
RUNNING HEAD ROUGH DRAFT                                 .docxRUNNING HEAD ROUGH DRAFT                                 .docx
RUNNING HEAD ROUGH DRAFT .docx
 
1Writing Activity 4 Final DraftShaland.docx
1Writing Activity 4  Final DraftShaland.docx1Writing Activity 4  Final DraftShaland.docx
1Writing Activity 4 Final DraftShaland.docx
 
Sociology Final Paper
Sociology Final PaperSociology Final Paper
Sociology Final Paper
 
Term Paper Essay Assignment Rubric Due by Saturday May 12th.docx
Term Paper Essay Assignment Rubric  Due by Saturday May 12th.docxTerm Paper Essay Assignment Rubric  Due by Saturday May 12th.docx
Term Paper Essay Assignment Rubric Due by Saturday May 12th.docx
 
Running head RACE AND CRIME .docx
Running head RACE AND CRIME                                    .docxRunning head RACE AND CRIME                                    .docx
Running head RACE AND CRIME .docx
 
RUNNING HEAD Racial bias and double standards .docx
RUNNING HEAD Racial bias and double standards                    .docxRUNNING HEAD Racial bias and double standards                    .docx
RUNNING HEAD Racial bias and double standards .docx
 
Death Penalty Persuasive Essay
Death Penalty Persuasive EssayDeath Penalty Persuasive Essay
Death Penalty Persuasive Essay
 
Racial Profiling in the Black Community
Racial Profiling in the Black CommunityRacial Profiling in the Black Community
Racial Profiling in the Black Community
 
BShyenna Vega Professor Artist Writing for Crimina.docx
BShyenna Vega Professor Artist Writing for Crimina.docxBShyenna Vega Professor Artist Writing for Crimina.docx
BShyenna Vega Professor Artist Writing for Crimina.docx
 
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docx
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docxJones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docx
Jones 1Jones 7Kyle JonesMatthew ZimmermanEnglish 10222 N.docx
 
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docx
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docxRunning head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docx
Running head RACIAL INEQUALITY JUVENILE JUSTICE SYSTEM1RACI.docx
 
Ethnicity (Crime and Deviance)
Ethnicity (Crime and Deviance)Ethnicity (Crime and Deviance)
Ethnicity (Crime and Deviance)
 
Death Penalty Essay Against
Death Penalty Essay AgainstDeath Penalty Essay Against
Death Penalty Essay Against
 

More from PazSilviapm

Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docx
Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docxCase Study Clinical LeadersDavid Rochester enjoys his role as a C.docx
Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docxPazSilviapm
 
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docx
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docxCASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docx
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docxPazSilviapm
 
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docx
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docxCASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docx
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docxPazSilviapm
 
Case Study 2A 40 year-old female presents to the office with the c.docx
Case Study 2A 40 year-old female presents to the office with the c.docxCase Study 2A 40 year-old female presents to the office with the c.docx
Case Study 2A 40 year-old female presents to the office with the c.docxPazSilviapm
 
Case Study Horizon Horizon Consulting Patti Smith looked up at .docx
Case Study Horizon  Horizon Consulting Patti Smith looked up at .docxCase Study Horizon  Horizon Consulting Patti Smith looked up at .docx
Case Study Horizon Horizon Consulting Patti Smith looked up at .docxPazSilviapm
 
Case Study EvaluationBeing too heavy or too thin, having a disabil.docx
Case Study EvaluationBeing too heavy or too thin, having a disabil.docxCase Study EvaluationBeing too heavy or too thin, having a disabil.docx
Case Study EvaluationBeing too heavy or too thin, having a disabil.docxPazSilviapm
 
Case Study Disney Corporation1, What does Disney do best to connec.docx
Case Study Disney Corporation1, What does Disney do best to connec.docxCase Study Disney Corporation1, What does Disney do best to connec.docx
Case Study Disney Corporation1, What does Disney do best to connec.docxPazSilviapm
 
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docx
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docxCase Study 3  Exemplar of Politics and Public Management Rightly Un.docx
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docxPazSilviapm
 
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docx
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docxCase Study 2 Structure and Function of the Kidney Rivka is an ac.docx
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docxPazSilviapm
 
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docx
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docxCase Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docx
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docxPazSilviapm
 
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docx
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docxCase Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docx
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docxPazSilviapm
 
Case FormatI. Write the Executive SummaryOne to two para.docx
Case FormatI. Write the Executive SummaryOne to two para.docxCase FormatI. Write the Executive SummaryOne to two para.docx
Case FormatI. Write the Executive SummaryOne to two para.docxPazSilviapm
 
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docx
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docxCase Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docx
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docxPazSilviapm
 
case scenario being used for this discussion postABS 300 Week One.docx
case scenario being used for this discussion postABS 300 Week One.docxcase scenario being used for this discussion postABS 300 Week One.docx
case scenario being used for this discussion postABS 300 Week One.docxPazSilviapm
 
Case Study #2Alleged improper admission orders resulting in mor.docx
Case Study #2Alleged improper admission orders resulting in mor.docxCase Study #2Alleged improper admission orders resulting in mor.docx
Case Study #2Alleged improper admission orders resulting in mor.docxPazSilviapm
 
Case Study 1Denise is a sixteen-year old 11th grade student wh.docx
Case Study 1Denise is a sixteen-year old 11th grade student wh.docxCase Study 1Denise is a sixteen-year old 11th grade student wh.docx
Case Study 1Denise is a sixteen-year old 11th grade student wh.docxPazSilviapm
 
Case AssignmentI. First read the following definitions of biodiver.docx
Case AssignmentI. First read the following definitions of biodiver.docxCase AssignmentI. First read the following definitions of biodiver.docx
Case AssignmentI. First read the following definitions of biodiver.docxPazSilviapm
 
Case and questions are In the attchmentExtra resources given.H.docx
Case and questions are In the attchmentExtra resources given.H.docxCase and questions are In the attchmentExtra resources given.H.docx
Case and questions are In the attchmentExtra resources given.H.docxPazSilviapm
 
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docx
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docxCase C Hot GiftsRose Stone moved into an urban ghetto in order .docx
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docxPazSilviapm
 
Case Assignment must be 850 words and use current APA format with a .docx
Case Assignment must be 850 words and use current APA format with a .docxCase Assignment must be 850 words and use current APA format with a .docx
Case Assignment must be 850 words and use current APA format with a .docxPazSilviapm
 

More from PazSilviapm (20)

Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docx
Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docxCase Study Clinical LeadersDavid Rochester enjoys his role as a C.docx
Case Study Clinical LeadersDavid Rochester enjoys his role as a C.docx
 
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docx
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docxCASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docx
CASE STUDY Clinical  Journal Entry  1 to 2 pages A 21 month .docx
 
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docx
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docxCASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docx
CASE STUDY 5Exploring Innovation in Action The Dimming of the Lig.docx
 
Case Study 2A 40 year-old female presents to the office with the c.docx
Case Study 2A 40 year-old female presents to the office with the c.docxCase Study 2A 40 year-old female presents to the office with the c.docx
Case Study 2A 40 year-old female presents to the office with the c.docx
 
Case Study Horizon Horizon Consulting Patti Smith looked up at .docx
Case Study Horizon  Horizon Consulting Patti Smith looked up at .docxCase Study Horizon  Horizon Consulting Patti Smith looked up at .docx
Case Study Horizon Horizon Consulting Patti Smith looked up at .docx
 
Case Study EvaluationBeing too heavy or too thin, having a disabil.docx
Case Study EvaluationBeing too heavy or too thin, having a disabil.docxCase Study EvaluationBeing too heavy or too thin, having a disabil.docx
Case Study EvaluationBeing too heavy or too thin, having a disabil.docx
 
Case Study Disney Corporation1, What does Disney do best to connec.docx
Case Study Disney Corporation1, What does Disney do best to connec.docxCase Study Disney Corporation1, What does Disney do best to connec.docx
Case Study Disney Corporation1, What does Disney do best to connec.docx
 
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docx
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docxCase Study 3  Exemplar of Politics and Public Management Rightly Un.docx
Case Study 3  Exemplar of Politics and Public Management Rightly Un.docx
 
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docx
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docxCase Study 2 Structure and Function of the Kidney Rivka is an ac.docx
Case Study 2 Structure and Function of the Kidney Rivka is an ac.docx
 
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docx
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docxCase Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docx
Case Study 2 Plain View, Open Fields, Abandonment, and Border Searc.docx
 
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docx
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docxCase Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docx
Case Study 2 Collaboration Systems at Isuzu Australia LimitedDue .docx
 
Case FormatI. Write the Executive SummaryOne to two para.docx
Case FormatI. Write the Executive SummaryOne to two para.docxCase FormatI. Write the Executive SummaryOne to two para.docx
Case FormatI. Write the Executive SummaryOne to two para.docx
 
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docx
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docxCase Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docx
Case Study #2 Diabetes Hannah is a 10-year-old girl who has recentl.docx
 
case scenario being used for this discussion postABS 300 Week One.docx
case scenario being used for this discussion postABS 300 Week One.docxcase scenario being used for this discussion postABS 300 Week One.docx
case scenario being used for this discussion postABS 300 Week One.docx
 
Case Study #2Alleged improper admission orders resulting in mor.docx
Case Study #2Alleged improper admission orders resulting in mor.docxCase Study #2Alleged improper admission orders resulting in mor.docx
Case Study #2Alleged improper admission orders resulting in mor.docx
 
Case Study 1Denise is a sixteen-year old 11th grade student wh.docx
Case Study 1Denise is a sixteen-year old 11th grade student wh.docxCase Study 1Denise is a sixteen-year old 11th grade student wh.docx
Case Study 1Denise is a sixteen-year old 11th grade student wh.docx
 
Case AssignmentI. First read the following definitions of biodiver.docx
Case AssignmentI. First read the following definitions of biodiver.docxCase AssignmentI. First read the following definitions of biodiver.docx
Case AssignmentI. First read the following definitions of biodiver.docx
 
Case and questions are In the attchmentExtra resources given.H.docx
Case and questions are In the attchmentExtra resources given.H.docxCase and questions are In the attchmentExtra resources given.H.docx
Case and questions are In the attchmentExtra resources given.H.docx
 
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docx
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docxCase C Hot GiftsRose Stone moved into an urban ghetto in order .docx
Case C Hot GiftsRose Stone moved into an urban ghetto in order .docx
 
Case Assignment must be 850 words and use current APA format with a .docx
Case Assignment must be 850 words and use current APA format with a .docxCase Assignment must be 850 words and use current APA format with a .docx
Case Assignment must be 850 words and use current APA format with a .docx
 

Recently uploaded

Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structuredhanjurrannsibayan2
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...pradhanghanshyam7136
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the ClassroomPooky Knightsmith
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.MaryamAhmad92
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin ClassesCeline George
 
ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701bronxfugly43
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Vishram Singh - Textbook of Anatomy Upper Limb and Thorax.. Volume 1 (1).pdf
Vishram Singh - Textbook of Anatomy  Upper Limb and Thorax.. Volume 1 (1).pdfVishram Singh - Textbook of Anatomy  Upper Limb and Thorax.. Volume 1 (1).pdf
Vishram Singh - Textbook of Anatomy Upper Limb and Thorax.. Volume 1 (1).pdfssuserdda66b
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024Elizabeth Walsh
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...christianmathematics
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - Englishneillewis46
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17Celine George
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...ZurliaSoop
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfNirmal Dwivedi
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...Poonam Aher Patil
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxAreebaZafar22
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 

Recently uploaded (20)

Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structure
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the Classroom
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Vishram Singh - Textbook of Anatomy Upper Limb and Thorax.. Volume 1 (1).pdf
Vishram Singh - Textbook of Anatomy  Upper Limb and Thorax.. Volume 1 (1).pdfVishram Singh - Textbook of Anatomy  Upper Limb and Thorax.. Volume 1 (1).pdf
Vishram Singh - Textbook of Anatomy Upper Limb and Thorax.. Volume 1 (1).pdf
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - English
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 

httpsdoi.org10.11770963721418763931Current Directions

  • 1. https://doi.org/10.1177/0963721418763931 Current Directions in Psychological Science 2018, Vol. 27(3) 183 –187 © The Author(s) 2018 Reprints and permissions: sagepub.com/journalsPermissions.nav DOI: 10.1177/0963721418763931 www.psychologicalscience.org/CDPS Compared with Whites, African Americans are 2 times more likely to die as infants, 3.6 times more likely to expe- rience childhood poverty, and 2 times more likely to have not completed high school (Centers for Disease Control and Prevention, 2017; National Center for Education Sta- tistics, 2017; Patten & Krogstad, 2015). Disturbing as these disparities are, they are not as extreme as disparities in the criminal justice system, in which African Americans are 5.1 times more likely than Whites to be incarcerated (Nellis, 2016). Many scholars and activists alike assume the general public would be motivated to fight racial inequality if only they knew the full extent of existing disparities. In response, activists bombard people with shocking statis- tics. Focusing specifically on policing and incarceration, we explore why simply presenting evidence of extreme racial disparities in the criminal justice system can backfire. We propose three potential strategies that may mitigate this paradoxical effect and provide important avenues for future research on how to reduce racial inequities. Evidence of Racial Disparities in the
  • 2. Criminal Justice System In cities across the United States, racial disparities have been documented throughout the criminal justice system, from routine police stops to long-term imprisonment. Analyzing traffic and pedestrian stops made in Oakland, California, for example, our team uncovered a consistent pattern of racial disparities. We found that 60% of police stops were of Africans Americans, though they make up only 28% of the population of Oakland. Once stopped, African Americans were significantly more likely to be handcuffed, searched, and arrested (Hetey, Monin, Maitreyi, & Eberhardt, 2016). These disparities remained statistically significant even after we controlled for more than two dozen factors relevant to officer decision mak- ing, including crime rates and the underlying racial and socioeconomic demographics where the stop was made. A similar pattern has emerged in other places, including Boston; Greensboro, North Carolina; Los Angeles; and New York City (American Civil Liberties Union Founda- tion of Massachusetts, 2014; Ayres & Borowsky, 2008; 763931CDPXXX10.1177/0963721418763931Hetey, EberhardtRacial Disparities in the Criminal Justice System research-article2018 Corresponding Authors: Jennifer L. Eberhardt, Department of Psychology, Stanford University, 450 Serra Mall, Stanford, CA 94305 E-mail: [email protected] Rebecca C. Hetey, Department of Psychology, Stanford University, 450 Serra Mall, Stanford, CA 94305 E-mail: [email protected]
  • 3. The Numbers Don’t Speak for Themselves: Racial Disparities and the Persistence of Inequality in the Criminal Justice System Rebecca C. Hetey and Jennifer L. Eberhardt Stanford University Abstract Many scholars and activists assume the public would be motivated to fight inequality if only they knew the full extent of existing disparities. Ironically, exposure to extreme disparities can cause people to become more, not less, supportive of the very policies that create those disparities (Hetey & Eberhardt, 2014). Here, we focus on the criminal justice system—policing and incarceration in particular. We argue that bringing to mind racial disparities in this domain can trigger fear and stereotypic associations linking Blacks with crime. Therefore, rather than extending an invitation to reexamine the criminal justice system, the statistics about disparities may instead provide an opportunity to justify and rationalize the disparities found within that system. With the goals of spurring future research and mitigating this paradoxical and unintended effect, we propose three potential strategies for more effectively presenting information about racial disparities: (a) offer context, (b) challenge associations, and (c) highlight institutions. Keywords race, crime, disparities, mass incarceration, policing Article http://www.psychologicalscience.org/cdps mailto:[email protected] https://sagepub.com/journalsPermissions.nav http://crossmark.crossref.org/dialog/?doi=10.1177%2F09637214
  • 4. 18763931&domain=pdf&date_stamp=2018-05-03 184 Hetey, Eberhardt Dunn, 2013; Goel, Rao, & Shroff, 2016; LaFraniere & Lehren, 2015). Blacks feel they are treated with less respect than Whites during these routine stops (Epp, Maynard-Moody, & Haider-Markel, 2014). Using footage from officers’ body-worn cameras, our team developed computational linguistic methods to analyze the respectfulness of the language that police officers used with White and Black community members during traffic stops in Oakland. We found that officers’ language was less respectful when directed at Blacks than when directed at Whites, even after controlling for factors such as the race of the officer, the severity of the infraction, the location of the stop, and the outcome of the stop (Voigt et al., 2017). In incarceration as well there are striking racial dis- parities. Though only 13% of the U.S. population, African Americans make up nearly 40% of the nation’s inmates (Guerino, Harrison, & Sabol, 2012; U.S. Census Bureau, 2011). Black men are incarcerated in state or federal prison at a rate 6 times that of White men (Carson, 2014). During their lifetime, 1 in 3 African American males can expect to be imprisoned compared with 1 in 17 White males (Bonczar, 2003). Do the Numbers Speak for Themselves? People trying to motivate social change frequently expose the public to grim statistics revealing gross racial disparities. Ironically, researchers have found that
  • 5. being presented with evidence of extreme racial dis- parities in the criminal justice system can cause the public to become more, not less, supportive of the punitive criminal justice policies that produce those disparities. When White participants were informed about racial disparities in executions, 52% strongly favored the death penalty compared with 36% in a baseline condition (Peffley & Hurwitz, 2007). In another study, we exposed White California voters to more or less extreme racial disparities in the prison population by having them view a set of photographs of incarcer- ated people that showed either a higher (45%) or lower (25%) percentage of Black inmates (Hetey & Eberhardt, 2014). When the prison population was “more Black,” voters became significantly less likely to sign a real petition aimed at lessening the severity of California’s harsh three-strikes law. At the time we conducted the study, California had the harshest three-strikes law in the nation—a person with two prior felony convictions could receive a life sentence for stealing a dollar in loose change from a parked car. When we described California’s three-strikes law and presented participants with a prison population that was less Black, 52% of them signed the petition compared with only 27% in the more-Black condition. Similarly, when White New York City residents read that New York’s prison population was 60% Black, com- pared with 40% Black, they became less likely to report that they would sign a petition to end New York City’s stop-and-frisk policy. The study was run at a time when the number of Blacks detained by the New York Police Department was dropping from a recent all-time high. When exposed to a more-Black prison population, only 12% of participants said they would sign the petition, compared with 33% of participants exposed to a less-
  • 6. Black prison population. Taken together, these studies suggest that informing White Americans of racially dis- proportionate incarceration may paradoxically bolster support for the very policies that perpetuate those disparities. These experiments were conducted with White par- ticipants. While an open question, it is unlikely that exposure to evidence of disparities has the same poten- tial to backfire with people from other racial back- grounds. People of color, and African Americans in particular, are more likely than Whites to have direct experience with, perceive, and engage in discussions about racial inequality and discrimination. According to a national Pew Research Center (2016) survey, 71% of Blacks reported having personally experienced discrimi - nation or having been treated unfairly because of their race or ethnicity, compared with 30% of Whites. By wide margins, Black respondents were more likely than White respondents to say Blacks are treated less fairly across a wide variety of settings, including the criminal justice system. Eighty-four percent of Blacks, compared with 50% of Whites, say Blacks are treated less fairly in deal- ing with the police. Indeed, 18% of Black Americans say they have been unfairly stopped by the police over the past year, compared with 3% of Whites. In the absence of direct experience and engagement with racial inequality, one reason disparities do not speak for themselves is because there is more than one interpretation of what those differences reveal. To some people, the disparities in police stops and incarceration are so extreme they can be interpreted only as signs of systemic racial bias (e.g., Alexander, 2010). An alterna- tive interpretation is that members of particular racial groups must be doing something—namely committing
  • 7. crime—to capture the attention of police and be impris- oned at higher rates (e.g., B. Adams, 2016). By focusing on group traits, the possibility that structural bias is at play in creating disparities falls out of view. Further- more, research shows that White Americans are sus- ceptible to denying claims of bias (G. Adams, Tormala, & O’Brien, 2006), in part to preserve views of the world as fair and meritocratic—views that can blind Whites to racial inequality and the role of institutional practices that harm minorities (O’Brien et al., 2009). Racial Disparities in the Criminal Justice System 185 Fear and stereotypic associations also contribute. Indeed, we found that when Whites were exposed to a “Blacker” prison population, they became significantly more fearful of crime, which, in turn, increased their support of punitive crime policies (Hetey & Eberhardt, 2014). Evidence of racial disparities in the criminal jus- tice system, then, may activate implicit stereotypical associations linking Blackness with crime, violence, threat, and aggression (Correll, Park, Judd, & Wittenbrink, 2002; Eberhardt, Goff, Purdie, & Davies, 2004; Payne, 2001; Richeson & Trawalter, 2008). If learning that Black men are 6 times more likely than White men to be imprisoned triggers the stereotype that Blacks are crimi - nals and criminals are Black, then such information is no longer concerning on its face. Rather, it becomes expected or even justifiable. Motivated to maintain the status quo, people take evidence of what is (such as the overrepresentation of Blacks in prison) and justify it as how things should be (Kay et al., 2009). What Should We Do?
  • 8. Given these ironic effects and what we know more generally about people’s use of information to support what they already believe or expect (Darley & Gross, 1983; Nickerson, 1998), what should people working for social change do? We propose three potential strate- gies that may mitigate the unintended consequences of exposing Whites to evidence of racial inequalities: (a) offer context, (b) challenge associations, and (c) high- light institutions. Offer context Future research should test whether a more effective way to frame inequality might be to present informa- tion about disparities alongside a description of how those disparities came to be. To contextualize is to convey that racial disparities are not natural or due to fixed stereotypical traits. Civil rights activists are aware of the history of racial disparities and the larger forces and processes by which these disparities are main- tained, yet that history is not always highlighted in the messages aimed at the broader public. Stripped of con- text, stand-alone statistics may simply be used as “evi- dence” of the stereotype that Blacks are prone to criminality. Rather than leading the public to examine the systemic forces involved, decontextualized infor- mation about racial inequality can lead people to engage in defensive processes that instead blame the victim (Kaiser & Miller, 2003) while protecting the self and the status quo (G. Adams et al., 2006; Lowery, Knowles & Unzueta, 2007). Challenge associations Exposure to information about racial disparities in the
  • 9. criminal justice system may activate and strengthen implicit associations linking Blacks with crime, vio- lence, threat, and aggression. Black Americans surely feel the effects of these stereotypical associations: Nearly half of Blacks (47%) say someone in the past year has acted as if they were suspicious of them (Pew Research Center, 2016). More research should investi - gate the consequences of educating the public about implicit bias and illustrating how implicit associations linking Blacks with crime can profoundly affect our perceptions, decision making, and actions without con- scious awareness or malicious intentio n. Our team has created an intervention to educate people on implicit bias and where it comes from, how it operates, and how it can be challenged. We are currently rolling out this training to law enforcement agencies across Cali- fornia. Preliminary results suggest the training is effec- tive in increasing knowledge of implicit bias, and most encouragingly, it leads law enforcement personnel to desire better relations with the community, to have more sympathy for community members, and to be more likely to believe they personally can make a dif- ference (Stanford SPARQ & California Department of Justice, 2016). Highlight institutions Another contributing factor to the perpetuation of inequality is the tendency to prioritize the role of indi - viduals over institutions. We, as a field, are guilty of this. Adams and colleagues describe how “as psycholo- gists, social psychologists tend to understand their task to be the study of individual experience . . . [suggesting] a standard approach to racism as a problem of biased individuals rather than a systemic force embedded in American society” (G. Adams, Edkins, Lacka, Pickett, &
  • 10. Cheryan, 2008, p. 350, emphasis in original). This con- ception or portrayal of inequality as an individual— rather than systemic—phenomenon can blind White Americans to racism and can guide policy attitudes and preferences in ways that perpetuate racial disparities. The individualistic portrayal is problematic because institutions and systems greatly contribute to reinforc- ing and reproducing inequality. In the criminal justice system, numerous distinct actors must be endowed by the system with the power to investigate, arrest, charge, prosecute, judge, sentence, and deprive people of their freedom, if not their lives. In the abstract, it can be difficult to conceptualize of vast systems and institutions. Lived experience grounds 186 Hetey, Eberhardt an understanding and recognition of structural racism. African Americans who say they have personally expe- rienced discrimination are equally divided over whether institutional racism or individual prejudice is the bigger problem for Black people today (44% each; Pew Research Center, 2016). African Americans who say they have never experienced discriminatio n, however, are nearly twice as likely to see individual racism as the bigger problem (59% vs. 32%). This lack of personal experience might explain why 70% of Whites in the same survey point to individual prejudice as the bigger problem, compared with only 19% of Whites who believe racism built into laws and institutions is the big- ger problem. Awareness of the power of institutions can be taught. Tutorials about structural or sociocultural racism have been shown to decrease the extent to which
  • 11. White Americans define racism in terms of individual biases, increase perceptions and acknowledgment of systemic racism, and increase endorsement of policies designed to promote equality (G. Adams et al., 2008). One specific way to highlight the role of institutions in perpetuating inequality may be to demonstrate how disparities can be attenuated by changing policy. Racial disparities in the use of consent searches (LaFraniere & Lehren, 2015; Palomino, 2016), for instance, have led to policy changes that mandate officers to obtain writ- ten consent or explicitly tell community members they have the right to deny an officer’s request to search them (LaFraniere, 2016). In Oakland, this policy change led to a massive reduction in consent searches without increasing crime. As another example, consensus to end mass incarceration has been building. Over 100 of the nation’s police chiefs and law enforcement officials joined together to call for a reduction in incarceration rates and the end of certain “tough-on-crime” policies (Williams, 2015). States including Texas, New York, Georgia, and California have eased minimum sentenc- ing and repeat offender laws. The result has been a reduction in both incarceration and crime (Chettiar, 2015), thus illustrating the role of laws and policies to actively shape, rather than passively respond to, crime and punishment. Conclusion To make the most meaningful change, we as social psychologists need to become more willing to intervene, not only with individuals but also by working to under- stand and change institutions, policies, and practices. Such a broader view could only enrich our science.
  • 12. Recommended Reading Adams, G., Edkins, V., Lacka, D., Pickett, K. M., & Cheryan, S. (2008). (See References). An article illustrating that the standard portrayal of racism in social psychology is as an individual prejudice but that, as the authors show, an awareness of structural or sociocultural racism can be taught. Eberhardt, J. L., Goff, P. A., Purdie, V. J., & Davies, P. G. (2004). (See References). A series of studies showing that stereotypical associations linking Blacks and crime are bidirectional and can act as visual tuning devices; for instance, exposure to Black faces facilitates the detection of weapons and activating the concept of crime creates an attentional bias toward Black faces. Hetey, R. C., & Eberhardt, J. L. (2014). (See References). Two experiments showing that exposure to extreme disparities can cause people to become more, not less, supportive of the very policies that create those disparities. Peffley, M., & Hurwitz, J. (2007). (See References). An article showing how informing Whites about racial disparities in executions bolsters support of the death penalty. Western, B. (2006). Punishment and inequality in America. New York, NY: Russell Sage Foundation. An exploration of the scope and consequences of mass incarceration in America, in particular the role of imprisonment in exac- erbating inequality. Declaration of Conflicting Interests The author(s) declared that there were no conflicts of interest
  • 13. with respect to the authorship or the publication of this article. References Adams, B. (2016, July 11). CNN panel blows up after guest sug- gests Blacks are prone to crime. Washington Examiner. Retrieved from http://www.washingtonexaminer.com/ cnn-panel-blows-up-after-guest-suggests-blacks-are- prone-to-crime/article/2596167 Adams, G., Edkins, V., Lacka, D., Pickett, K. M., & Cheryan, S. (2008). Teaching about racism: Pernicious implications of the standard portrayal. Basic and Applied Social Psy- chology, 30, 349–361. Adams, G., Tormala, T. T., & O’Brien, L. T. (2006). The effect of self-affirmation on perception of racism. Journal of Experimental Social Psychology, 42, 616–626. Alexander, M. (2010). The new Jim Crow: Mass incarceration in the age of colorblindness. New York, NY: The New Press. American Civil Liberties Union Foundation of Massachusetts. (2014). Black, brown and targeted: A report on Boston Police Department street encounters from 2007–2010. Retrieved from https://aclum.org/wp-content/uploads/2015/06/ reports-black-brown-and-targeted.pdf Ayres, I., & Borowsky, J. (2008). A study of racially disparate outcomes in the Los Angeles Police Department. Retrieved from https://www.aclusocal.org/sites/default/files/wp- content/uploads/2015/09/11837125-LAPD-Racial-Profiling- Report-ACLU.pdf Bonczar, T. P. (2003). Prevalence of imprisonment in the U.S.
  • 14. population, 1974-2001. Bureau of Justice Statistics. Retrieved from https://www.bjs.gov/content/pub/pdf/piusp01.pdf Carson, E. A. (2014). Prisoners in 2013. Bureau of Justice Statistics. Retrieved from https://www.bjs.gov/content/ pub/pdf/p13.pdf http://www.washingtonexaminer.com/cnn-panel-blows-up-after- guest-suggests-blacks-are-prone-to-crime/article/2596167 http://www.washingtonexaminer.com/cnn-panel-blows-up-after- guest-suggests-blacks-are-prone-to-crime/article/2596167 http://www.washingtonexaminer.com/cnn-panel-blows-up-after- guest-suggests-blacks-are-prone-to-crime/article/2596167 https://aclum.org/wp-content/uploads/2015/06/reports-black- brown-and-targeted.pdf https://aclum.org/wp-content/uploads/2015/06/reports-black- brown-and-targeted.pdf https://www.aclusocal.org/sites/default/files/wp- content/uploads/2015/09/11837125-LAPD-Racial-Profiling- Report-ACLU.pdf https://www.aclusocal.org/sites/default/files/wp- content/uploads/2015/09/11837125-LAPD-Racial-Profiling- Report-ACLU.pdf https://www.aclusocal.org/sites/default/files/wp- content/uploads/2015/09/11837125-LAPD-Racial-Profiling- Report-ACLU.pdf https://www.bjs.gov/content/pub/pdf/piusp01.pdf https://www.bjs.gov/content/pub/pdf/p13.pdf https://www.bjs.gov/content/pub/pdf/p13.pdf Racial Disparities in the Criminal Justice System 187 Centers for Disease Control and Prevention. (2017). Sudden unexpected infant death and sudden infant death syn- drome: Data and statistics. Retrieved from https://www
  • 15. .cdc.gov/sids/data.htm Chettiar, I. M. (2015, April 27). A national agenda to reduce mass incarceration. Brennan Center for Justice. Retrieved from https://www.brennancenter.org/analysis/national- agenda-reduce-mass-incarceration Correll, J., Park, B., Judd, C. M., & Wittenbrink, B. (2002). The police officer’s dilemma: Using ethnicity to disambiguate potentially threatening individuals. Journal of Personality and Social Psychology, 83, 1314–1329. Darley, J. M., & Gross, P. H. (1983). A hypothesis-confirming bias in labeling effects. Journal of Personality and Social Psychology, 44, 20–33. Dunn, C. (2013). Stop-and-frisk 2012: NYCLU briefing. New York Civil Liberties Union. Retrieved from https://www .nyclu.org/sites/default/files/publications/2012_Report_ NYCLU_0.pdf Eberhardt, J. L., Goff, P. A., Purdie, V. J., & Davies, P. G. (2004). Seeing Black: Race, crime, and visual processing. Journal of Personality and Social Psychology, 87, 876–893. Epp, C. R., Maynard-Moody, S., & Haider-Markel, D. P. (2014). Pulled over: How police stops define race and citizenship. Chicago, IL: The University of Chicago Press. Goel, S., Rao, J. M., & Shroff, R. (2016). Precinct or preju- dice? Understanding racial disparities in New York City’s stop-and-frisk policy. The Annals of Applied Statistics, 10, 365–394. Guerino, P, Harrison, P. M., & Sabol, W. J. (2012). Prisoners in
  • 16. 2010. Bureau of Justice Statistics. Retrieved from https:// www.bjs.gov/content/pub/pdf/p10.pdf Hetey, R. C., & Eberhardt, J. L. (2014). Racial disparities in incarceration increase acceptance of punitive policies. Psychological Science, 25, 1949–1954. Hetey, R. C., Monin, B., Maitreyi, A., & Eberhardt, J. L. (2016). Data for change: A statistical analysis of police stops, searches, handcuffings, and arrests in Oakland, Calif., 2013-2014. Stanford SPARQ. Retrieved from https://stan ford.app.box.com/v/Data-for-Change Kaiser, C. R., & Miller, C. T. (2003). Derogating the victim: The interpersonal consequences of blaming events on discrimination. Group Processes & Intergroup Relations, 6, 227–237. Kay, A. C., Gaucher, D., Peach, J. M., Laurin, K., Friesen, J., Zanna, M. P., & Spencer, S. J. (2009). Inequality, discrimi - nation, and the power of the status quo: Direct evidence for a motivation to see the way things are as the way they should be. Journal of Personality and Social Psychology, 97, 421–434. LaFraniere, S. (2015, November 11). Greensboro puts focus on reducing racial bias. The New York Times. Retrieved from https://www.nytimes.com/2015/11/12/us/greensboro- puts-focus-on-reducing-racial-bias.html LaFraniere, S., & Lehren, A. W. (2015, October 24). The dispro- portionate risks of driving while Black. The New York Times. Retrieved from https://www.nytimes.com/2015/10/25/us/ racial-disparity-traffic-stops-driving-black.html
  • 17. Lowery, B. S., Knowles, E. D., & Unzueta, M. M. (2007). Framing inequity safely: Whites’ motivated perceptions of racial privilege. Personality and Social Psychology Bulletin, 33, 1237–1250. National Center for Education Statistics. (2017). Public high school 4-year adjusted cohort graduation rate (ACGR), by race/ethnicity and selected demographic characteristics for the United States, the 50 states, and the District of Columbia: School year 2015–16 [Data table]. Retrieved from https://nces.ed.gov/ccd/tables/ACGR_RE_and_char acteristics_2015-16.asp Nellis, A. (2016). The color of justice: Racial and ethnic dis - parity in state prisons. The Sentencing Project. Retrieved from https://www.sentencingproject.org/wp-content/ uploads/2016/06/The-Color-of-Justice-Racial-and-Ethnic- Disparity-in-State-Prisons.pdf Nickerson, R. S. (1998). Confirmation bias: A ubiquitous phe- nomenon in many guises. Review of General Psychology, 2, 175–220. O’Brien, L. T., Blodorn, A., Alsbrooks, A., Dube, R., Adams, G., & Nelson, J. C. (2009). Understanding White Americans’ perceptions of racism in hurricane Katrina-related events. Group Processes & Intergroup Relations, 12, 431–444. Palomino, J. (2016, April 8). Racial disparities in SF traffic searches raise concerns of bias. San Francisco Chronicle. Retrieved from http://www.sfchronicle.com/crime/article/ Racial-disparities-in-SF-traffic-searches-raise-7235690.php
  • 18. Patten, E., & Krogstad, J. M. (2015). Black child poverty rate holds steady, even as other groups see declines. Pew Research Center. Retrieved from http://www.pewre search.org/fact-tank/2015/07/14/black-child-poverty-rate- holds-steady-even-as-other-groups-see-declines/ Payne, B. K. (2001). Prejudice and perception: The role of automatic and controlled processes in misperceiving a weapon. Journal of Personality and Social Psychology, 81, 181–192. Peffley, M., & Hurwitz, J. (2007). Persuasion and resistance: Race and the death penalty in America. American Journal of Political Science, 51, 996–1012. Pew Research Center. (2016). On views of race and inequality, Blacks and Whites are worlds apart. Retrieved from http:// www.pewsocialtrends.org/2016/06/27/on-views-of-race- and-inequality-blacks-and-whites-are-worlds-apart/ Richeson, J. A., & Trawalter, S. (2008). The threat of appear- ing prejudiced and race-based attentional biases. Psychological Science, 19, 98–102. Stanford SPARQ & California Department of Justice. (2016). Principled policing training: Procedural justice and implicit bias. Retrieved from https://stanfor d.app.box .com/s/e3ik3z20be2k9k6p6pluo9o42wdz5cni U.S. Census Bureau. (2011). Overview of race and Hispanic origin: 2010. Retrieved from https://www.census.gov/ prod/cen2010/briefs/c2010br-02.pdf Voigt, R., Camp, N. P., Prabhakaran, V., Hamilton, W. L., Hetey, R. C., Griffiths, C. M., . . . Eberhardt, J. L. (2017). Language from police body camera footage shows racial
  • 19. disparities in officer respect. Proceedings of the National Academy of Sciences, USA, 114, 6521–6526. Williams, T. (2015, October 20). Police leaders join call to cut prison rosters. The New York Times. Retrieved from http://www.nytimes.com/2015/10/21/us/police-leaders- join-call-to-cut-prison-rosters.html?_r=0 https://www.cdc.gov/sids/data.htm https://www.cdc.gov/sids/data.htm https://www.brennancenter.org/analysis/national-agenda-reduce- mass-incarceration https://www.brennancenter.org/analysis/national-agenda-reduce- mass-incarceration https://www.nyclu.org/sites/default/files/publications/2012_Rep ort_NYCLU_0.pdf https://www.nyclu.org/sites/default/files/publications/2012_Rep ort_NYCLU_0.pdf https://www.nyclu.org/sites/default/files/publications/2012_Rep ort_NYCLU_0.pdf https://www.bjs.gov/content/pub/pdf/p10.pdf https://www.bjs.gov/content/pub/pdf/p10.pdf https://stanford.app.box.com/v/Data-for-Change https://stanford.app.box.com/v/Data-for-Change https://www.nytimes.com/2015/11/12/us/greensboro-puts-focus- on-reducing-racial-bias.html https://www.nytimes.com/2015/11/12/us/greensboro-puts-focus- on-reducing-racial-bias.html https://www.nytimes.com/2015/10/25/us/racial-disparity-traffic- stops-driving-black.html https://www.nytimes.com/2015/10/25/us/racial-disparity-traffic- stops-driving-black.html https://nces.ed.gov/ccd/tables/ACGR_RE_and_characteristics_2 015-16.asp https://nces.ed.gov/ccd/tables/ACGR_RE_and_characteristics_2 015-16.asp
  • 20. https://www.sentencingproject.org/wp- content/uploads/2016/06/The-Color-of-Justice-Racial-and- Ethnic-Disparity-in-State-Prisons.pdf https://www.sentencingproject.org/wp- content/uploads/2016/06/The-Color-of-Justice-Racial-and- Ethnic-Disparity-in-State-Prisons.pdf https://www.sentencingproject.org/wp- content/uploads/2016/06/The-Color-of-Justice-Racial-and- Ethnic-Disparity-in-State-Prisons.pdf http://www.sfchronicle.com/crime/article/Racial-disparities-in- SF-traffic-searches-raise-7235690.php http://www.sfchronicle.com/crime/article/Racial-disparities-in- SF-traffic-searches-raise-7235690.php http://www.pewresearch.org/fact-tank/2015/07/14/black-child- poverty-rate-holds-steady-even-as-other-groups-see-declines/ http://www.pewresearch.org/fact-tank/2015/07/14/black-child- poverty-rate-holds-steady-even-as-other-groups-see-declines/ http://www.pewresearch.org/fact-tank/2015/07/14/black-child- poverty-rate-holds-steady-even-as-other-groups-see-declines/ http://www.pewsocialtrends.org/2016/06/27/on-views-of-race- and-inequality-blacks-and-whites-are-worlds-apart/ http://www.pewsocialtrends.org/2016/06/27/on-views-of-race- and-inequality-blacks-and-whites-are-worlds-apart/ http://www.pewsocialtrends.org/2016/06/27/on-views-of-race- and-inequality-blacks-and-whites-are-worlds-apart/ https://stanford.app.box.com/s/e3ik3z20be2k9k6p6pluo9o42wdz 5cni https://stanford.app.box.com/s/e3ik3z20be2k9k6p6pluo9o42wdz 5cni https://www.census.gov/prod/cen2010/briefs/c2010br-02.pdf https://www.census.gov/prod/cen2010/briefs/c2010br-02.pdf http://www.nytimes.com/2015/10/21/us/police-leaders-join-call- to-cut-prison-rosters.html?_r=0 http://www.nytimes.com/2015/10/21/us/police-leaders-join-call- to-cut-prison-rosters.html?_r=0
  • 21. Information Systems Success: The Quest for the Dependent Variable . William H. DeLone Department of Management The American University Washington, D.C. 20016 Ephraim R. McLean Computer Information Systems Georgia State University Atlanta, Georgia 30302-4015 A large number of studies have been conducted during the last decade and a half attempting to identify those factors that contribute to information sys- tems success. However, the dependent variable in these studies—I/S success —has been an elusive one to define. Different researchers have addressed different aspects of success, making comparisons difficult and the prospect of building a cumulative tradition for I/S research similarly elusive. To organize this diverse research, as well as to present a more integrated view of the concept of I/S success, a comprehensive taxonomy is introduced. This taxon- omy posits six major dimensions or categories of I/S success — SYSTEM OUALITY, INFORMATION QUALITY, USE, USER SATISFACTION, INDIVIDUAL IMPACT, and ORGANIZATIONAL IMPACT. Using these dimensions, both conceptual and empirical studies are then
  • 22. reviewed (a total of 180 articles are cited) and organized according to the dimensions of the taxonomy. Finally, the many aspects of I/S success are drawn together into a descriptive model and its implications for future I/S research are discussed. Information s)'!)(em« •iucce^s—Inrormition systems •ssessmenl—Measurement Introduction At the first meeting of the International Conference on Information System (ICIS)in 1980, Peter Keen identified five issues which he felt needed to be resolved in order for the field of management information systems to establish itself as a coherent research area. These issues were: (1) What are the reference disciplines for MIS? (2) What is the dependent variable? (3) How does MIS establish a cumulative tradition? (4) What is the relationship of MIS research to computer technology and to MIS practice? (5) Where should MIS Tesearchers publish their findings? I047.7047/92/0301/0O6O/SOI.25 Copyright © 1992. The Inslilulc of Management Stienees 60 Information Systems Research 3 : I
  • 23. Information Systems Success Of the five, the second item, the dependent variable in MIS research, is a particu- larly important issue. If information systems research is to make a contribution to the world of practice, a well-defined outcome measure (or measures) is essential. It does littie good to measure various independent or input variables, such as the extent of user participation or the level of I/S investment, if the dependent or output variable —I/S success or MIS effectiveness—cannot be measured with a similar degree of accuracy. The importance of defining the I/S dependent variable cannot be overemphasized. The evaluation of I/S practice, policies, and procedures requi res an I/S success mea- sure against which various strategies can be tested. Without a well-defined dependent variable, much of I/S research is purely speculative. In recognition of this importance, this paper explores the research that has been done involving MIS success since Keen first issued his challenge to the field and attempts to synthesize this research into a more coherent body of knowledge. It covers the formative period 1981 -87 and reviews all those empirical studies that have attempted to measure some aspects of "MIS success" and which have appeared in one of the seven leading publications in the I/S field. In addition, a number of other
  • 24. articles are included, some dating back to 1949. that make a theoretical or conceptual contribution even though they may not contain any empirical data. Taken together, these 180 references provide a representative review of the work that has been done and provide the basis for formulating a more comprehensive model of I/S success than has been attempted in the past. A Taxonomy of Information Systems Success Unfortunately, in searching for an I/S success measure, rather than finding none, there are nearly as many measures as there are studies. The reason for this is under- standable when one considers that "information." as the output of an information system or the message in a communication system, can be measured at different levels, including the technical level, the semantic level, and the effectiveness level. In their pioneering work on communications. Shannon and Weaver (1949) defined the technical level as the accuracy and efficiency of the system which produces the infor- mation, the semantic level as the success of the information in conveying the in- tended meaning, and the effectiveness level as the effect of the information on the receiver. Building on this. Mason (1978) relabeled "effectiveness" as "influence" and de- fined the influence level of information to be a "hierarchy of events which take place
  • 25. at the receiving end of an information system which may be used to identify the various approaches that might be used to measure output at the influence level" (Mason 1978. p. 227). This scries of influence events includes the receipt of the information, an evaluation of the information, and the application of the informa- tion, leading to a change in recipient behavior and a change in system performance. The concept of levels of output from communication theory demonstrates the serial nature of information (i.e., a form of communication). The information system creates information which is communicated to the recipient who is then influenced (or not!) by the information. In this sense, information flows through a series of stages from its production through its use or consumption to its influence on individual and/or organizational performance. Mason's adaptation of communication theory March 1992 DeLone • Shannon Weaver (1949) Mason
  • 27. on Recipent User Individual Satisfaction Impact Influence on System Organizational Impact FIGURE 1. Categories of [/S Success. to the measurement of information systems suggests therefore that there may need to be separate success measures for each of the levels of information. In Figure 1, the three levels of information of Shannon and Weaver are shown, together with Mason's expansion of the effectiveness or influence level, to yield six distinct categories or aspects of information systems. They are SYSTEM QUALITY, INFORMATION QUALITY, USE. USER SATISFACTION, INDIVIDUAL IM- PACT, and ORGANIZATIONAL IMPACT. Looking at the first of these categories, some I/S researcher s have chosen to focus on the desired characteristics of the information system itself which produces the
  • 28. information (SYSTEM QUALITY). Others have chosen to study the information product for desired characteristics such as accuracy, meaningful ness, and timeliness (INFORMATION QUALITY). In the influence level, some researchers have ana- lyzed the interaction of the information product with its recipients, the users and/or decision makers, by measuring USE or USER SATISFACTION. Still other re- searchers have been interested in the influence which the information product has on management decisions (INDIVIDUAL IMPACT). Finally, some I/S researchers, and to a larger extent I/S practitioners, have been concerned with the eifect of the information product on organizational performance (ORGANIZATIONAL IMPACT). Once this expanded view of I/S success is recognized, it is not surprising to find that there are so many different measures of this success in the literature, depending upon which aspect of I/S the researcher has focused his or her attention. Some of these measures have been merely identified, but never used empirically. Others have been used, but have employed different measurement instruments, making comparisons among studies difficult. Two previous articles have made extensive reviews of the research literature and have reported on the measurement of MIS success that had been used in empirical
  • 29. studies up until that time. In a review of studies of user involvement, Ives and Olson (1984) adopted two classes of MIS outcome variables: system quality and system acceptance. The system acceptance category was defined to include system usage, system impact on user behavior, and information satisfaction. Haifa decade earlier, in a review of studies of individual differences. Zmud (1979) considered three catego- ries of MIS success: user performance, MIS usage, and user satisfaction. Eoth of these literature reviews made a valuable contribution to an understanding of MIS success, but both were more concerned with investigating independent 62 Information Systems Research 3 : 1 Information Systems Success variables (i.e., user involvement in the case of Ives and Olson and individual differ- ences in the case of Zmud) than with the dependent variable (i.e.. MIS success). In contrast, this paper has the measurement of the dependent variable as its primary focus. Also, over five years have passed since the Ives and Olson study was published and over ten years since Zmud's article appeared. Much work has been done since these two studies, justifying an update of their findings.
  • 30. To review this recent work and to put the earlier research into perspective, the six categories of I/S success identified in Figure I—SYSTEM QUALITY. INFORMA- TION QUALITY. INFORMATION USE, USER SATISFACTION, INDIVIDUAL IMPACT. AND ORGANIZATIONAL IMPACT—are used in the balance of this paper to organize the I/S research that has been done on I/S success. In each of the six sections which follow, both conceptual and empirical studies are cited. While the conceptual citations are intended to be comprehensive, the empirical studies are intended to be representative, not exhaustive. Seven publications, from the period January 1981 to January 1988. were selected as reflecting the mainstream of I/S research during this formative period. Additional studies, from other publica- tions, as well as studies from the last couple of years, could have been included; but after reviewing a number of them, it became apparent that they merely reinforced rather than modified the basic taxonomy of this paper. In choosing the seven publications to be surveyed. fyQ (Management Science, MIS Quarterly, Communications of the ACM, Decision Sciences, and Information & Man- agement) were drawn from the top six journals cited by Hamilton and Ives (1983) in their study of the journals most respected by MIS researchers. (Their sixth journal. Transactions on Database Systems, was omitted from this study
  • 31. because of its special- ized character.) To these five were added the Journal of MIS. a relatively new but important journal, and the ICIS Proceedings, which is not a journal per se but repre- sents the published output of the central academic conference in the I/S field. A total of 100 empirical studies are included from these seven sources. As with any attempt to organize past research, a certain degree of arbitrariness occurs. Some studies do not fit neatly into any one category and others fit into several. In the former case, every effort was made to make as close a match as possible in order to retain a fairly parsimonious framework. In the latter case, where several measures were used which span more than one category (e.g.. measures of informa- tion quality and extent of use and user satisfaction), these studies are discussed in each of these categories. One consequence of this multiple listing is that there appear to be more studies involving I/S success than there actually are. To decide which empirical studies should be included, and which measures fit in which categories, one of the authors of this paper and a doctoral student (at another university) reviewed each of the studies and made their judgments independently. The interrater agreement was over 90%. Conflicts over selection and measure assign- ment were resolved by the second author. In each of the following sections, a table is included which
  • 32. summarizes the empiri- cal studies which address the particular success variable in question. In reporting the success measures, the specific description or label for each dependent variable, as used by the author(s) of the study, is reported. In some cases the wording of these labels may make it appear that the study would be more appropriately listed in another table. However, as was pointed out earlier, all of these classification decisions March 1992 63 DeLone • McLean are somewhat arbitrary, as is true of almost al! attempts to organize an extensive body of research on a retrospective basis. System Quality: Measures of the Information Processing System Itself In evaluating the contribution of information systems to the organization, some I/S researchers have studied the processing system itself. Kriebel and Raviv (1980, 1982) created and tested a productivity model for computer systems, including such performance measures as resource utilization and investment utilization. Alloway (1980) developed 26 criteria for measuring the success of a data processing operation. The efficiency of hardware utilization was among Alloway's
  • 33. system success criteria. Other authors have developed multiple measures of system quality. Swanson (1974) used several system quality items to measure MIS appreciation among user managers. His items included the reliability of the computer system, on-line response time, the ease of terminal use. and so forth. Emery (1971) also suggested measuring system characteristics, such as the content of the data base, aggregation of details, human factors, response time, and system accuracy. Hamilton and Chervany (1981) proposed data currency, response time, turnaround time, data accuracy, reliability, completeness, system flexibility, and ease of use among others as part of a "formative evaluation" scheme to measure system quality. In Table I are shown the empirical studies which had explicit measures of system quality. Twelve studies were found within the referenced journals, with a number of distinct measures identified. Not surprisingly, most of these measures are fairly straightforward, reflecting the more engineering-oriented performance characteris- tics of the systems in question. Information Quality: Measures of Information System Output Rather than measure the quality of the system performance, other 1/S researchers have preferred to focus on the quality of the information system output, namely, the
  • 34. quality of the information that the system produces, primarily in the form of reports. Larcker and Lessig (1980) developed six questionnaire items to measure the per- ceived importance and usableness of information presented in reports. Bailey and Pearson (1983) proposed 39 system-related items for measuring user satisfaction. Among their ten most important items, in descending order of importance, were information accuracy, output timeliness, reliability, completeness, relevance, preci- sion, and currency. In an early study, Ahituv (1980) incorporated five information characteristics into a multi-attribute utility measure of information value: accuracy, timeliness, rele- vance, aggregation, and formatting. Gallagher (1974) develoiJed a semantic differen- tial instrument to measure the value of a group of I/S reports. That instrument included measures of relevance, informativeness, usefulness, and importance. Munro and Davis (1977) used Gallagher's instrument to measure a decision maker's perceived value of information received from information systems which were cre- ated using different methods for determining information requirements. Additional information characteristics developed by Swanson (1974) to measure MIS apprecia- tion among user managers included uniqueness, conciseness, clarity, and readability measures. Zmud (1978) included report format as an information quality measure in
  • 35. his empirical work. Olson and Lucas (1982) proposed report appearance and accu- racy as measures of information quality in office automation information systems. Lastly, King and Epstein (1983) proposed multiple information attributes to yield a 64 Information Systems Research 3 : 1 Information Systems Success Authors Bailey and Pearson (1983) BartiandHuff(l985) Belardo. Kanvan. and Wallace (1982) Con kiln. Gotterer. and Rick man (1982) Franz and Robey (1986) Goslar(1986) Hiltz and Turoff (1981) Kxiebel and Raviv
  • 36. (1982) Lehman (1986) Mahmood(1987) Morey(1982) Srinivasan(1985) TABLE 1 Empirical Measures of System Quality Description of Study Overall I/S; 8 organizations, 32 managers DSS: 9 organizations. 42 decision makers Emergency management DSS; 10 emergency dispatchers Transaction processing; one organization Specific I/S; 34 organizations. 118 user managers Marketing DSS; 43 marketers Electronic information exchange system; 102 users
  • 37. Academic information system; one university Overall I/S; 200 I/S directors Specific I/S; 61 I/S managers Manpower management system; one branch of the military Computer-based modeling systems; 29 firms Type Field Field Lab U b Field Lab Field Case Field Field
  • 38. Case Field Description of Measure(s) (1) Convenience of access (2) Flexibility of system (3) Integration of systems (4) Response time Realization of user expectations (1) Reliability (2) Response time (3) Ease of use (4) Ease of learning Response time Perceived usefulnessofl/S (12 items) LJsefulness of DSS features Usefulness of specific functions (1) Resource utilization (2) Investment utilization I/S sophistication (use of new technology) Flexibility of system Stored record error rate
  • 39. (1) Response time (2) System reliability (3) System accessibility composite measure of information value. The proposed information attributes in- cluded sufficiency, understandability, freedom from bias, reliability, decision rele- vance, comparability, and quantitativeness. More recently, numerous information quality criteria have been included within the broad area of "User Information Satisfaction" (Iivari 1987; Iivari and Koskela 1987). The Iivari-Koskela satisfaction measure included three information quality constructs: "informativeness" which consists of relevance, comprehensiveness, re- centness. accuracy, and credibility; "accessibility" which consists of convenience, timeliness, and interpretability; and "adaptability." In Table 2. nine studies which included information quality measures are shown. Understandably, most measures of information quality are from the perspective of the user of this information and are thus faidy subjective in character. Also, these March 1992 65 DeLone • McLean
  • 40. measures, while shown here as separate entities, are often included as part of the measurers of user satisfaction. The Bailey and Pearson (1983) study is a good exam- ple of this cross linkage. Information Use: Recipient Consumption of the Output of an Information System The use of information system reports, or of management science/operations re- search models, is one of the most frequently reported measures of the success of an information system or an MS/OR model. Several researchers (Lucas 1973; Schultz andSlevin 1975; Ein-Dor and Segev 1978; Ives, Hamilton, and Davis 1980; Hamil- ton and Chervany 1981) have proposed I/S use as an MIS success measure in concep- tual MIS articles. Ein-Dor and Segev claimed that different measures of computer success are mutually interdependent and so they chose system use as the primary criterion variable for their I/S research framework. "Use of system" was also an integral part of Lucas's descriptive model of information systems in the context of organizations. Schultz and Slevin incorporated an item on the probability of MS/OR model use into their five-item instrument for measuring model success. In addition to these conceptual studies, the use of an information system has often been the MIS success measure of choice in MIS empirical research (Zmud 1979). The
  • 41. broad concept of use can be considered or measured from several p>erspectives. It is clear that actual use, as a measure of I/S success, only makes sense for voluntary or discretionary users as opposed to captive users (Lucas 1978; Weike and Konsynski 1980). Recognizing this, Maish (1979) chose voluntary use of computer terminals and voluntary requests for additional reports as his measures of I/S success. Similarly, Kim and Lee (1986) measured voluntariness of use as part of their measure of success. Some studies have computed actual use (as opposed to reported use) by managers through hardware monitors which have recorded the number of computer inquiries (Swanson 1974; Lucas 1973, 1978; King and Rodriguez 1978. 1981), or recorded the amount of user connect time (Lucas 1978; Ginzberg 1981a). Other objective mea- sures of use were the number of computer functions utilized (Ginzberg 1981 a), the number of client records processed (Robey 1979), or the actual charges for computer use (Gremillion 1984). Still other studies adopted a subjective or perceived mea- sure of use by questioning managers about their use of an information system (Lucas 1973, 1975. 1978; Maish 1979; Fuerst and Cheney 1982; Raymond 1985; DeLone 1988). Another issue concerning use of an information system is "Use by whom?" (Huys-
  • 42. mans 1970). In surveys of MIS success in small manufacturing firms, DeLone (1988) considered chief executive use of information systems while Raymond (1985) consid- ered use by company controllers. In an earlier study. Culnan (1983a) considered both direct use and chaufFeured use (i.e.. use through others). There are also different levels of use or adoption. Ginzberg (1978) discussed the following levels of use, based on the earlier work by Huysmans; (1) use that results in management action, (2) use that creates change, and (3) recurring use of the system. Earlier, Vanlommel and DeBrabander (1975) proposed four levels of use: use for getting instructions, use for recording data, use for control, and use for planning. Schewe (1976) introduced two forms of use: general use of "routinely generated computer reports" and specific use of "personally initiated requests for additional 66 Information Systems Research 3 : Information Systems Success TABLE 2 Empirical Measures of Information Quality Authors Description of Study Type Description of Measure(s) Bailey and Pearson (1983)
  • 43. BlaylockandRees(l984) Jones and McLeod (1986) King and Epstein (1983) Mahniood(1987) Mahmood and Medewitz (1985) Mitlerand Doyle (1987) RivardandHuff(l985) Srinivasan (1985) Overall I/S; 8 organizations. 32 managers Financial; one university. 16 MBA students Several information sources; 5 senior executives Overall I/S; 2 firms. 76 managers Specific I/S; 61 I/S managers DSS; 48 graduate students Overall 1/S; 21 financial
  • 44. firms. 276 user managers User-developed I/S; 10 firms, 272 users Computer-based modeling systems; 29 firms Field Output (1) Accuracy (2) Precision (3) Currency (4) Timeliness (5) Reliability (6) Completeness (7) Conciseness (8) Format (9) Relevance Lab Perceived usefulness of specific report items Field Perceived importance of each information item Field Information (1) Currency (2) Sufficiency (3) Understandability (4) Freedom from bias (5) Timeliness (6) Reliability (7) Relevance to decisions (8) Comparability (9) Quantitativeness
  • 45. Field (I) Report accuracy (2) Report timeliness Lab Report usefulness Field (1) Completeness of information (2) Accuracy ofinformation (3) Relevance of reports (4) Timeliness of report Field Usefulness ofinformation Field (1) Report accuracy (2) Report relevance (3) Underslandability (4) Report timeliness information not ordinarily provided in routine reports." By this definition, specific use reflects a higher level of system utilization. Fuerst and Cheney (1982) adopted Schewe's classification of general use and specific use in the ir study of decision sup- port in the oil industry. Bean et al. (1975); King and Rodriguez {1978, 1981), and DeBrabander and Thiers (1984) attempted to measure the nature of system use by comparing this use to the March 1992 67 DeLone • McLean
  • 46. decision-making purpose for which the system was designed. Similarly, livari {1985) suggested appropriate use or acceptable use as a measure of MIS success. In a study by Robey and Zeller (1978), I/S success was equated to the adoption and extensive use of an information system. After reviewing a number of empirical studies involving use, Trice and Treacy (1986) recommend three classes of utilization measures based on theories from refer- ence disciplines: degree of MIS institutionalization, a binary measure of use vs. non- use, and unobtrusive utilization measures such as connect time and frequency of computer acce^. The degree of institutionalization is to be determined by user de- pendence on the MIS, user feelings of system ownership, and the degree to which MIS is routinized into standard operating procedures. Table 3 shows the 27 empirical studies which were found to employ system use as at least one of their measures of success. Of all the measures identified, the system use variable is probably the most objective and the easiest to quantify, at least concep)- tually. Assuming that the organization being studied is {1) regularly monitoring such usage patterns, and (2) willing to share these data with researchers, then usage is a fairly accessible measure of I/S success. However, as pointed out earlier, usage, either actual or perceived, is only pertinent when such use is voluntary.
  • 47. User Satisfaction: Recipient Response to the Use ofthe Output of an Information System When the use of an information system is required, the preceding measures be- come less useful; and successful interaction by management with the information system can be measured in terms of user satisfaction. Several I/S researchers have suggested user satisfaction as a success measure for their empirical I/S research (Ein- Dor and Segev 1978; Hamilton and Chervany 1981). These researchers have found user satisfaction as especially appropriate when a specific information system was involved. Once again a key issue is whose satisfaction should be measured. In at- tempting to determine the success ofthe overall MIS effort, McKinsey & Company (1968) measured chief executives' satisfaction. In two empirical studies on implementation success, Ginzberg (1981a, b) chose user satisfaction as his dependent variable. In one of those studies (1981 a), he adopted both use and user satisfaction measures. In a study by Lucas (1978), sales representatives rated their satisfaction with a new computer system. Later, in a differ- ent study, executives were asked in a laboratory setting to rate their enjoyment and satisfaction with an information system which aided decisions relating to an inven- tory ordering problem (Lucas 1981).
  • 48. In the Powers and Dickson study on MIS project success (1973), managers were asked how well their information needs were being satisfied. Then, in a study by King and Epstein (1983), I/S value was imputed based on managers" satisfaction ratings. User satisfaction is also recommended as an appropriate success measure in experi- mental I/S research (Jarvenpaa, Dickson, and DeSanctis 1985) and for researching the effectiveness of group decision support systems (DeSanctis and Gallupe 1987). Other researchers have developed multi-attribute satisfaction measures rather than relying on a single overall satisfaction rating. Swanson (1974) used 16 items to mea- sure I/S appreciation, items which related to the characteristics of reports and ofthe underlying information system itself. Pearson developed a 39- item instrument for measuring user satisfaction. The full instrument is presented in Bailey and Pearson 68 lnfortnation Systems Research 3 : 1 Information Systems Success (1983), with an earlier version reviewed and evaluated by KHebel (1979) and by Ives, Olson, and Baroudi (1983). Raymond (1985) used a subset of 13 items from Pear- son's questionnaire to measure manager satisfaction with MIS in
  • 49. small manufactur- ing firms. More recently, Sanders (1984) developed a questionnaire and used it (Sanders and Courtney 1985) to measure decision support systems (DSS) success. Sanders' overall success measure involves a number of measures of user and decision- making satisfaction. Finally, studies have found that user satisfaction is associated with user attitudes toward computer systems (Igerhseim 1976; Lucas 1978) so that user-satisfaction measures may be biased by user computer attitudes. Therefore, studies which include user satisfaction as a success measure should ideally also include measures of user attitudes so that the potentially biasing effects of those attitudes can be controlled for in the analysis. Goodhue (1986) further suggests "information satisfactoriness" as an antecedent to and surrogate for user satisfaction. Information satisfactoriness is de- fined as the degree of match between task characteristics and I/S functionality. As the numerous entries in Table 4 make clear, user satisfaction or user informa- tion satisfaction is probably the most widely used single measure of I/S success. The reasons for this are at least threefold. First, "satisfaction" has a high degree of face validity. It is hard to deny the success ofa system which its users say that they like. Second, the development ofthe Bailey and Pearson instrument and its derivatives
  • 50. has provided a reliable tool for measuring satisfaction and for making comparisons among studies. The third reason for the appeal of satisfaction as a success measure is that most of the other measures are so poor; they are either conceptually weak or empirically difficult to obtain. Individual Impact: The Effect ofinformation on the Behavior ofthe Recipient Of all the measures of I/S success, "impact" is probably the most difficult to define in a nonambiguous fashion. It is closely related to performance, and so "improving my—or my department's—performance" is certainly evidence that the information system has had a positive impact. However, "impact" could also be an indication that an information system has given the user a better understanding ofthe decision context, has improved his or her decision-making productivity, has produced a change in user activity, or has changed the decision maker's perception ofthe impor- tance or usefulness ofthe information system. As discussed earlier. Mason (1978) proposed a hierarchy of impact (influence) levels from the receipt ofthe information, through the understanding ofthe information, the application ofthe information to a specific problem, and the change in decision behavior, to a resultant change in organi- zational performance. As Emery (1971, p. I) states: "Information has no intrinsic value; any value comes only through the influence it may have
  • 51. on physical events. Such influence is typically exerted through human decision makers." In an extension ofthe traditional statistical theory ofinformation value. Mock (1971) argued for the importance ofthe "learning value ofinformation." In a labora- tory study ofthe impact ofthe mode ofinformation presentation, Lucas and Nielsen (1980) used learning, or rate of performance improvement, as a dependent variable. In another laboratory setting, Lucas (1981) tested participant understanding ofthe inventory problem and used the test scores as a measure of I/S success. Watson and Driver (1983) studied the impact of graphical presentation on information recall. Meador, Guyote, and Keen (1984) measured the impact of a DSS design March 1992 DeLone • McLean TABLE 3 Empirical Measures ofinformation System Use Authors Description of Study Type Description of Measure(s) Alavi and Henderson (1981) Baroudi, Olson, and Ives
  • 52. (1986) BartiandHuff(1985) Bell (1984) Work force and production Lab scheduling DSS; one university. 45 graduates Overall I/S; 200 finns, 200 production managers DSS; 9 organizations, 42 Field decision makers Financial; 30 financial Lab Use or nonuse of computer-based decision aids Field Use of 1/S to support production Percentage of time DSS is used in decision making situations Use of numerical vs. nonnumerical information Benbasat. Dexter, and Masulis(1981) Bergeron (1986b) Chandrasekaran and Kirs
  • 53. (1986) Culnan (1983a) Culnan (1983b) DeBrabander and Thiers (1984) DeSanctis (1982) Ein-Dor. Segev, and Steinfeld(l98l) Green and Hughes(1986) Fuerst and Cheney (1982) Glnzberg(198la) Hogue(1987) Gremillion (1984) Pricing: one university. 50 students and faculty Overall I/S; 54 organizations. 471 user managers Reporting systems; MBA students Overall I/S; one organization, 184
  • 54. professionals Overall I/S; 2 organizations, 362 professionals Specialized DSS: one university, 91 two-person teams DSS; 88 senior level students PERT: one R & D organization, 24 managers DSS; 63 city managers DSS; 8 oil companies, 64 users On-line portfolio management system; U.S. bank, 29 portfolio managers DSS; 18 organizations Overall I/S; 66 units of tbe National Forest system Lab Field
  • 55. Field Field Field U b Lab Field Lab Field Field Field Field Frequency of requests for specific reports Use of chargeback information Acceptance of report (I) Direct use ofl/S vs. chaufieured use (2) Number of requests for information Frequency of use
  • 56. Use vs. nonuse of data sets Motivation to use (I) Frequency of past use (2) Frequency of intended use Number of DSS features used (I) Frequency of general use (2) Frequency of specific use (1) Number of minutes (2) Number of sessions (3) Number of functions used Frequency of voluntary use Expenditures/charges for computing use 70 Information Systems Research 3 : 1 Information Systems Success Authors Kim and Lee (1986) King and Rodriguez (1981) Mahmood and Medewitz
  • 57. (1985) Nelson and Cbeney (1987) Perry (1983) Raymond (1985) Snitkin and King (1986) Srinivasan (1985) Swanson (1987) Zmud. Boynton. and Jacobs(1987) TABLE 3 {cont d) Description of Study Overall [/S; 32 organizations, 132 users Strategic system; one university. 45 managers DSS; 48 graduate students Overall I/S; 100 top/middle managers Office 1/S; 53 firms Overall 1/S; 464 small
  • 58. manufacturing Brms Personal DSS; 31 users Computer-based modeling systems; 29 firms Overall I/S; 4 organizations, 182 users Overall I/S; Sample A: 132 firms Sample B: one firm Type Field Lab Lab Field Field Field Field Field Field Field
  • 59. Description of Measure(s) (1) Frequency of use (2) Voluntariness of use (1) Number of queries (2) Nature of queries Extent of use Extent of use Use at anticipated level (1) Frequency of use (2) Regularity of use Hours per week (1) Frequency of use (2) Time per computer session (3) Number of reports generated Average frequency with which user discussed report information Use in support of (a) Cost reduction (b) Management (c) Strategy planning (d) Competitive thrust methodology using questionnaire items relating to resulting decision effectiveness.
  • 60. For example, one questionnaire item referred specifically to the subject's perception of the improvement in his or her decisions. In the information system framework proposed by Chervany, Dickson, and Kozar (1972), which served as the model for the Minnesota Experiments (Dickson. Cher- vany. and Senn 1977). the dependent success variable was generally defined to be decision effectiveness. Within the context of laboratory experiments, decision effec- tiveness can take on numerous dimensions. Some of these dimensions which have been reported in laboratory studies include the average time to make a decision (Benbasat and Dexter 1979, 1985; Benbasat and Schroeder 1977; Chervany and Dickson 1974; Taylor 1975), the confidence in the decision made (Chervany and Dickson 1974; Taylor 1975), and the number of reports requested (Benbasat and Dexter 1979; Benbasat and Schroeder 1977). DeSanctis and Gallupe (1987) sug- gested member participation in decision making as a measure of decision effective- ness in group decision making. In a study which sought to measure the success of user- developed applications, Rivard and Huff (1984) included increased user productivity in their measure of success. DeBrabander and Thiers (1984) used efficiency of task accomplishment (time required to find a correct answer) as the dependent variable in their laboratory
  • 61. March 1992 71 DeLone • McLean TABLE 4 Empirical Measures of User Satisfaction Author(s) Description of Study Type Description of Measure(s) Alavi and Henderson (1981) Baitey and Pearson (1983) Baroudi, Olson, and Ives(1986) Barti and Huff(l985) Work force and production scheduling DSS; one university; 45 graduate students Overall I/S; 8 organizations. 32 managers Overall I/S; 200 firms. 200 production managers DSS; 9 organizations, 42 decision makers
  • 62. Lab Overall satisfaction witb DSS Field User satisfaction (39-item instrument) Field User information satisfaction Field User information satisfaction (modified Bailey & Pearson instrument) Bruwer(1984) Cats-Baril and Huber (1987) DeSanctis (1986) Doll and Ahmed (1985) Edmundson and JefFery (1984) Ginzberg (1981a) Ginzberg(l981b) Hogue(1987) Ives. Olson, and Baroudi(1983) Jenkins, Naumann, and Wetherbe (1984) King and Epstein (1983)
  • 63. Langle, Leitheiser, and Naumann (1984) Lehman, Van Wetering. and Vogel (1986) Lucas(1981) Overall I/S; one organization, 114 managers DSS; one university, 101 students Human resources 1/S; 171 human resource system professionals Specificl/S; 55 firms, 154 managers Accounting software package; 12 organizations On-line portfolio management system; U.S. bank. 29 portfolio managers Overall I/S; 35 I/S users DSS; 18 organizations Overall I/S; 200 firms, 200 production managers
  • 64. A specific I/S; 23 corporations, 72 systems development managers Overall I/S; 2 firms. 76 managers Overall 1/S; 78 oi^nizations, I/S development managers Business graphics; 200 organizations, DP managers Inventory ordering system; one university, 100 executives Field Lab Field Field Field Field Field Field Field Field
  • 65. Field Field Field U b User satisfaction Satisfaction with a DSS (multi-item scale) (1) Top management satisfaction (2) Personal management satisfaction User satisfaction (11 -item scale) User satisfaction (1 question) Overall satisfaction Overall satisfaction User satisfaction (1 question) User satisfaction (Bailey & Pearson instrument) User satisfaction (25-item instrument) User satisfaction (1 item: scale 0 to 100) User satisfaction (1 question)
  • 66. (1) Software satisfaction (2) Hardware satisfaction (1) Enjoyment (2) Satisfaction 72 Information Systems Research 3 : 1 Information Systems Success TABLE 4 Author(s) Description of Study Type Description of Measure(s) Mahmood (1987) Mahmood and Becker (1985-1986) Mahmood and Medewitz (1985) McKeen(l983) Nelson and Cheney (1987) Olson and Ives (1981) Olson and Ives (1982) Raymond (1985)
  • 67. Raymond (1987) Rivard and Huff (1984) Rushinek and Rushinek (1985) Rushinek and Rushinek (1986) Sanders and Courtney (1985) Sanders, Courtney, and Uy(I984) Taylor and Wang(1987) Specificl/S; 61 I/S managers Field Overall satisfaction Overall I/S; 59 firms. 118 Field User satisfaction managers DSS; 48 graduate students Lab User satisfaction (multi-Item scale) Application systems; 5 organizations Overall I/S; 100 top/middle managers Field Satisfaction with the development project (Powers and Dickson instrument)
  • 68. Field User satisfaction (Bailey & Pearson instrument) Overall I/S; 23 manufacturing Field Information dissatisfaction difference firms. 83 users between information needed and amount ofinformation received Overall I/S; 23 manufacturing Field Information satisfaction, difference firms, 83 users between information needed and information received Overall I/S; 464 small manufacturing firms Overall I/S; 464 small-firm finance managers User-developed applications; 10 large companies Accounting and billing system; 4448 users Overall I/S; 4448 users Financial DSS; 124 organizations Field Controller satisfaction (modified Bailey & Pearson instrument) Field User satisfaction (modified Bailey &
  • 69. Pearson instrument) Field User complaints regarding Information Center services Field Overall user satisfaction Field Overall user satisfaction Field (1) Overall satisfaction (2) Decision-making satisfaction Interactive Financial Planning Field (I) Decision-making satisfaction System (IFPS); 124 (2) Overall satisfaction oi^nizations, 373 users DBMS with multiple dialogue Lab User satisfaction with interface modes; one university, 93 students experiment. Finally, Sanders and Courtney (1985) adopted the speed of decision analysis resulting from DSS as one item in their DSS success measurement in- strument. Mason (1978) has suggested that one method of measuring I/S impact is to deter- mine whether the output ofthe system causes the receiver (i.e., the decision maker) to change his or her behavior. Ein-Dor, Segev, and Steinfeld (1981) asked decision makers: "Did use of PERT [a specific information system] ever lead to a change in a
  • 70. March 1992 73 DeLone • McLean decision or to a new decision?" Judd, Paddock, and Wetherbe (1981) measured whether a budget exception reporting system resulted in managers' taking investiga- tive action. Another approach to the measurement ofthe impact of an information system is to ask user managers to estimate the value of the information system. Cerullo (1980) asked managers to rank the value of their computer-based MIS on a scale of one to ten. Ronen and Falk (1973) asked participants to rank the value ofinformation received in an experimental decision context. Using success items developed by Schultz and Slevin (1975), King and Rodriguez (1978. 1981) asked users of their "Strategic Issue Competitive Information System" to rate the worth of that 1/S. Other researchers have gone a step further by asking respondents to place a dollar value on the information received. Gallagher (1974) asked managers about the maxi- mum amount they would be willing to pay for a particular report. Lucas (1978) reported using willingness to pay for an information system as one of his success
  • 71. measures. Keen (1981) incorporated willingness to pay development costs for im- proved DSS capability in his proposed "Value Analysis" for justification ofa DSS. In an experiment involving MBA students, Hilton and Swieringa (1982) measured what participants were willing to pay for specific information which they felt would lead to higher decision payoffs. Earlier. Garrity (1963) used MIS expenditures as a percent- age of annual capital expenditures to estimate the value ofthe MIS effort. Table 5, with 39 entries, contains the largest number of empirical studies. This in itself is a healthy sign, for it represents an attempt to move beyond the earlier inward- looking measures to those which offer the potential to gauge the contribution of information systems to the success ofthe enterprise. Also worth noting is the predomi- nance of laboratory studies. Whereas most ofthe entries in the preceding tables have been field experiments, 24 ofthe 39 studies reported here have used controlled labora- tory experiments as a setting for measuring the impact ofinformation on individuals. The increased experimental rigor which laboratory studies offer, and the extent to which they have been utilized at least in this success category, is an encouraging sign for the maturing of the field. Organizational Impact: The Effect of Information on Organizational Performance
  • 72. In a survey by Dickson, Leitheiser. Wetherbe, and Nechis (1984). 54 information systems professionals ranked the measurement ofinformation system effectiveness as the fifth most important I/S issue for the 1980s. In a recent update of that study by Brancheau and Wetherbe (1987). I/S professionals ranked measurement ofinforma- tion system effectiveness as the ninth most important I/S issue. Measures of individ- ual performance and, to a greater extent, organization performance are of consider- able importance to I/S practitioners. On the other hand. MIS academic researchers have tended to avoid performance measures (except in laboratory studies) because of the difficulty of isolating the effect ofthe I/S effort from other effects which influence organizational performance. As discussed in the previous section, the effect of an information system on individ- ual decision performance has been studied primarily in laboratory experiments using students and computer simulations. Many of these experiments were conducted at the University of Minnesota (Dickson, Chervany. and Senn 1977). Among these "Minnesota Experiments" were some that studied the effects of different information 74 Information Systems Research 3 : 1
  • 73. Information Systems Success formats and presentation modes on decision performance as measured in terms of lower production, inventory, or purchasing costs. King and Rodriguez (1978. 1981) measured decision performance by evaluating participant tests responses to various hypothesized strategic problems. In another laboratory study. Lucas and Nielsen (1980) measured participant perfor- mance (and thus, indirectly, organizational performance) in terms of profits in a logistics management game. In a later experiment, Lucas (1981) investigated the efiect of computer graphics on decisions involving inventory ordering. Finally, Remus (1984) used the costs of various scheduling decisions to evaluate the effects of graphical versus tabular displays. Field studies and case studies which have dealt with the influence ofinformation systems have chosen various organizational performance measures for their depen- dent variable. In their study, Chervany, Dickson, and Kozar (1972) chose cost reduc- tions as their dependent variable. Emery (1971. p. 6) has observed that: "Benefits from an information system can come from a variety of sources. An important one is the reduction in operating costs of activities external to the information processing system."
  • 74. Several researchers have suggested that the success of the MIS department is re- flected in the extent to which the computer is applied to critical or major problem areas ofthe firm (Garrity 1963; Couger and Wergin 1974; Ein- Dor and Segev 1978; Rockart 1979; Senn and Gibson 1981). In Garrity's early article (1963), company I/S operations were ranked partly on the basis o f t h e range and scope of its computer applications. In a later McKinsey study (1968), the authors used the range of "mean- ingful, functional computer applications" to distinguish between more or less success- ful MIS departments. In a similar vein, Vanlommel and DeBrabander (1975) used a weighted summation ofthe number of computer applications as a measure of MIS success in small firms. Finally, Cerullo (1980) ranked MIS success on the basis o f a firms's ability to computerize high complexity applications. In a survey of several large companies, Rivard and Huff (1984) interviewed data processing executives and asked them to assess the cost reductions and company profits realized from specific user-developed application programs. Lucas (1973) and Hamilton and Chervany (1981) suggested that company revenues can also be im- proved by computer-based information systems. In a study o f a clothing manufac- turer, Lucas (1975) used total dollar bookings as his measure of organizational perfor- mance. Chismar and Kriebel (1985) proposed measuring the relative efficiency ofthe
  • 75. information systems effort by applying Data Envelopment Analysis to measure the relationship of corporate outcomes such as total sales and return on investment to I/S inputs. More comprehensive studies ofthe effect of computers on an organization include both revenue and cost issues, within a cost/benefit analysis (Emery 1971). McFadden (1977) developed and demonstrated a detailed computer cost/benefit analysis using a mail order business as an example. In a paper entitled "What is the Value of Invest- ment in Information Systems?," Matlin (1979) presented a detailed reporting system for the measurement ofthe value and costs associated with an information system. Cost/benefit analyses are often found lacking due to the difficulty of quantifying "intangible benefits." Building on Keen's Value Analysis approach (1981). Money, Tromp. and Wegner (1988) proposed a methodology for identifying and quantifying March 1992 75 DeLone • McLean Authors) Aldag and Power (1986)
  • 76. Belardo, Kanvan, and Wallace (1982) Benbasat and Dexter (1985) Benbasat and Dexter (1986) Benbasat. Dexter, and Masulis (1981) Bergeron (1986a) Cats-Baril and Huber (1987) Crawford (1982) DeBrabanderand Thiers(l984) DeSanctis and Jarvenpaa (1985) Dickson, DeSanctis, and McBride (1986) Drury(1982) Ein-Dor, Segev,
  • 77. and Steinfeld (1981) TABLE 5 Empiricat Measures of Individual Impact Description of Study DSS; 88 business students Emergency management DSS; 10 emergency dispatchers Financiai; 65 business students Financial; 65 business students Pricing; one university, 50 students and factuly DP chargeback system; 54 organizations, 263 user managers DSS; one university, 101 students Electronic mail;
  • 78. computer vendor organization Specialized DSS; one university, 91 two- person teams Tables vs. graphs; 75 MBA students Graphics system; 840 undei^raduate students Chargeback system; 173 organizations, senior DP managers PERT; one R & D organization. 24 managers Type Lab U b Lab U b Lab Field
  • 79. Lab Case Lab U b U b Field Field Description of Measure(s) (1) User confidence (2) Quality of decision analysis (1) Efficient decisions (2) Time to arrive at a decision Time taken to complete a task Time taken to complete a task Time to make pricing decisions Extent to which users analyze charges and investigate
  • 80. budget variances (1) Quality of career plans (2) Number of objectives and • alternatives generated Improved personal productivity, hrs/wk/ manager (1) Time efficiency of task accomplishment (2) User adherence to plan Decision quality, forecast accuracy (1) Interpretation accuracy (2) Decision quality (1) Computer awareness (2) Cost awareness Change in decision behavior 76 Information Systems Research 3 : 1 Information Systems Success TABLE 5 ( a w / y ) Author(s) Description of Study Type Description of Measure(s)
  • 81. Fuerst and Cheney(1982) Goslar, Green, and Hughes (1986) Goul, Shane, and Tonge(1986) Green and Hughes (1986) DSS; 8 oil companies, 64 users DSS: 19 organizations, 43 sales and marketing personnel Knowledge-based DSS; one university, 52 students DSS; 63 city managers Field Value in assisting decision making Lab (I) Number of altematives
  • 82. considered (2) Time to decision (3) Confidence in decision (4) Ability to identify solutions L ^ Ability to identify strategic opoortunities or problems Lab (I) Time to decision (2) Number of alternatives considered (3) Amount of data considered Grudmtski(I98l) Gueutal. Surprenant, and Bubeck (1984) Hilton and Swieringa (1982) Hughes (1987) Judd. Paddock, and Wetherbe (1981)
  • 83. Kaspar(l985) King and Rodriguez (1981) Lee, MacLachlan. and Wallace (1986) Lucas(1981) Planning and control system: 65 business students Computer-aided design system; 69 students General !/S; one university, 56 MBA students DSS generator; 63 managers Budget exception reporting system; 116 MBA students DSS; 40 graduate students Strategic system; one
  • 84. university. 45 managers Performance I/S; 45 naarketing students Inventory ordering system; one university, 100 executives Lab Lab Lab Field Lab U b U b U b U b Precision of decision maker's forecast (I) Task performance (2) Confidence in performance
  • 85. Dollar value ofinformation (1) Time to reach decision (2) Number of alternatives considered Management takes investigative action Ability to forecast firm performance (1) Worth of information system (2) Quality of policy decisions (I) Accuracy of information interpretation (2) Time to solve problem User understanding of inventory problem March 1992 77 DeLone • Mcl^an Author(s) Lucas and Palley
  • 86. (1987) Luzi and Mackenzie (1982) Meador, Guyote, and Keen (1984) Millman and Hanwick (1987) Rivard and Huff (1984) Rivard and Huff (1985) Sanders and Courtney (1985) Snitkin and King (1986) Srinivasan (1985) Vogel. Lehman, and Dickson (1986) Watson and Driver (1983)
  • 87. Zmud (1983) Zmud. Blocher. and Moffie (1983) TABLE Description of Study Overall 1/S; 3 manufacturing firms, 87 plant managers Performance information system; one university, 200 business students DSS; 18 firms, 73 users Office I/S; 75 middle managers User-developed applications; 10 large companies User-developed I/S; 10 firms 272 users Financial DSS; 124 oi^nizations
  • 88. Personal DSS; 31 users Computer-based modeling systems; 29 firms Graphical Presentation System; 174 undergraduate students Graphical presentation of information; 29 undergraduate business students External information channels; 49 software development managers Invoicing system; 51 internal auditors 5 (cont'd) Type Field U b
  • 89. Field Field Field Field Field Field Field U b U b Field U b Description of Measure(s) (1) Power of I/S department (2) Influence of I/S department (1) Time to solve problem (2) Accuracy of problem solution (3) Ffficiency of effort
  • 90. (1) Effectiveness in supporting decisions (2) Time savings Personal effectiveness User productivity Productivity improvement Decision-making efficiency and effectiveness Effectiveness of personal DSS (1) Problem identification (2) Generation of alternatives Change in commitment of time and money (1) Immediate recall of information (2) Delayed recall of information Recognition and use of modern software practices (1) Decision accuracy (2) Decision confidence 78 Information Systems Research 3 ; 1
  • 91. Information Systems Success intangible benefits. The proposed methodology then applied a statistical test to deter- mine whether "significant" value can be attached to a decision support system. With the corporate "bottom line" in mind, several MIS frameworks have proposed that MIS effectiveness be determined by its contribution to company profits (Cher- vany, Dickson. and Kozar 1972; Lucas 1973; Hamilton and Chervany 1981), but few empirical studies have attempted to measure actual profit contribution. Ferguson and Jones (1969) based their evaluation of success on more profitable job schedules which resulted from decision-maker use ofthe information system. Ein-Dor, Segev, and Steinfeld (1981) attempted to measure contribution to profit by asking users ofa PERT system what savings were realized from use of PERT and what costs were incurred by using PERT. Another measure of organizational performance which might be appropriate for measuring the contribution of MIS is return on investment. Both Garrity (1963) and the McKinsey study (1968) reported using return on investment calculations to as- sess the success of corporate MIS efforts. Jenster (1987) included nonfinancial mea- sures of organizational impact in a field study of 124
  • 92. organizations. He included productivity, innovations, and product quality among his measures of I/S success. In a study of 53 firms, Perry (1983) measured the extent to which an office information system contributed to meeting organizational goals. Strassmann, in his book Information Payoffil9S5), presented a particularly com- prehensive view of the role of information systems with regards to performance, looking at it from the perspective ofthe individual, the organization, the top execu- tive, and society. His measure of performance was a specially constructed "Return on Management" (ROM) metric. In nonprofit organizations, specifically government agencies, Danziger (i 977) pro- posed using productivity gains as the measure ofinformation systems impact on the organization. He explained that productivity gains occur when the "functional out- put ofthe government is increased at the same or increased quality with the same or reduced resources inputs" (p. 213). In a presentation of several empirical studies conducted by the University of California, Irvine, Danziger included five productiv- ity measures: staff reduction, cost reduction, increased work volume, new informa- tion, and increased effectiveness in serving the public. The success of informati on systems in creating competitive advantage has prompted researchers to study I/S impacts not only on firm
  • 93. performance but also on industry structure (Clemons and Kimbrough 1986). Bakos (1987) reviewed the litera- ture on the impacts of information technology on firm and industry-level perfor- mance from the perspective of organization theory and industrial economics. At the firm level, he suggested measures of changes in organizational structure and of im- provements in process efficiency using Data Envelopment Analysis (Chismar and Kriebel 1985) as well as other financial measures. At the industry level, he found impact measures (e.g., economies of scale, scope, and market concentration) harder to identify in any readily quantifiable fashion and suggested that further work is needed. Johnston and Vitale (1988) have proposed a modified cost/benefit analysis ap- proach to measure the effects of interorganizational systems. Traditional cost/benefit analysis is applied to identify quantifiable benefits such as cost reductions, fee reve- nues, and increased product sales. Once the quantifiable costs and benefits have been identified and compared, Johnston and Vitale suggest that top management use March 1992 79 DeLone • McLean
  • 94. judgment to assess the value ofthe benefits which are more difficult to quantify such as reduction of overhead, increases in customer switching costs, barriers to new firm entry, and product differentiation. Table 6 is the last ofthe six tables summarizing the I/S success measures identified in this paper. Somewhat surprisingly, 20 empirical studies were found, with 13 using field-based measures (as opposed to the laboratory experiments characterizing the individual impacts) to get at the real-world effects of the impact of information systems on organizational performance. However, this is only a beginning; and it is in this area, "assessing the business value of information systems," where much work needs to be done. Discussion In reviewing the various approaches that I/S researcher s have taken in measuring MIS success, the following observations emerge. 1. As these research studies .show, the I/S researcher has a broad list of individual dependent variables from which to choose. It is apparent that there is no consensus on the measure of information systems success. Just as there are many steps in the production and dissemination ofinforma- tion, so too are there many variables which can be used as measures of "I/S success."
  • 95. In Table 7, all ofthe variables identified in each ofthe six success categories discussed in the preceding sections are listed. These include success variables which have been suggested but never used empirically as well as those that have actually been used in experiments. In reviewing these variables, no single measure is intrinsically better than another; so the choice ofa success variable is often a function ofthe objective ofthe study, the organizational context, the aspect ofthe information system which is addressed by the study, the independent variables under investigation, the research method, and the level of analysis, i.e., individual, organization, or society (Markus and Robey 1988). However, this proliferation of measures has been overdone. Some consolida- tion is needed. 2. Progress toward an MIS cumidative tradition dictates a significant reduction in the number of different dependent variable measures so that research results can be compared. One of the major purposes of this paper is the attempt to reduce the myriad of variables shown in Table 7 to a more manageable taxomony. However, within each of these major success categories, a number of variables still exist. The existence of so many different success measures makes it difficult to compare the results of similar
  • 96. studies and to build a cumulative body of empirical knowledge. There are, however, examples of researchers who have adopted measurement instruments developed in earlier studies. Ives, Olson, and Baroudi (1983) have tested the validity and reliability ofthe user-satisfaction questionnaire developed by Bailey and Pearson (1983) and used that instrument in an empirical study of user involvement (Baroudi, Olson and Ives 1986). Raymond (1985. 1987) used a subset ofthe Bailey and Pearson user-satisfac- tion instrument to study MIS success in small manufacturing firms. Similarly, Mah- mood and Becker (1986) and Nelson and Cheney (1987) have used the Bailey and Pearson instrument in empirical studies. In another vein, McKeen (1983) adopted 80 Information Systems Research 3 : 1 Information Systems Success the Powers and Dickson (1973) satisfaction scale to measure the success of I/S devel- opment strategies. King and Rodriguez (1978, 1981), Robey (1979), Sanders (1984). and Sanders and Courtney (1985) have adopted parts ofa measurement instrument which Schultz and Slevin (1975) developed to measure user attitudes and
  • 97. perceptions about the value of operations research models. Munro and Davis (1977) and Zmud (1978) utilized Gallaghef s questionnaire items (1974) to measure the perceived value of an informa- tion system. Finally, Blaylock and Rees (1984) used Larcker and Lessig's 40 informa- tion items (1980) to measure perceived information usefulness. These are encouraging trends. More MIS researchers should seek out success mea- sures that have been developed, validated, and applied in previous empirical re- search. 3. Not enough MIS field study research attempts to measure the influence ofthe MIS effort on organizational performance. Attempts to measure MIS impact on overall organizatio nal performance are not often undertaken because ofthe difficulty of isolating the contribution ofthe infor- mation systems function from other contributors to organizational performance. Nevertheless, this connection is of great interest to information system practitioners and to top corporate management. MIS organizational performance measurement deserves further development and testing. Cost/benefit schemes such as those presented by Emery (1971), McFadden (1977), and MatHn (1979) offer promising avenues for further study. The University of Cali- fornia, Irvine, research on the impact ofinformation systems on
  • 98. government activity (Danziger 1987) suggests useful impact measures for public as well as private organi- zations. Lucas (1975) included organizational performance in his descriptive model and then operationalized this variable by including changes in sale revenues as an explicit variable in his field study ofa clothing manufacturer. Garrity (1963) and the McKinsey & Company study (1968) reported on early attempts to identify MIS returns on investment. McLean (1989), however, pointed out the difficulties with these approaches, while at the same time attempting to define a framework for such analyses. Strassmann (1985) has developed his "Return on Management" metric as a way to assess the overall impact ofinformation systems on companies. These research efforts represent promising beginnings in measuring MIS impact on performance. 4. The six success categories and the many specific I/S measures within each of these categories clearly indicaie that MIS success is a multidimensional construct and that it should he measured as such. Vanlomme! and DeBrabander (1975) early pointed out that the success ofa com- puter-based information system is not a homoge neous concept and therefore the attempt should not be made to capture it by a simple measure. Ein-Dor and Segev
  • 99. (1978) admitted that their selection of MIS use as their dependent variable may not be ideal. They stated that "A better measure of MIS success would probably be some weighted average for the criteria mentioned above" (i.e.. use, profitability, applica- tion to major problems, performance, resulting quality decision, and user satis- faction). In reviewing the empirical studies cited in Tables 1 through 6. it is clear that most of them have attempted to measure I/S success in only one or possibly two success March 1992 81 DeLone • McLean TABLE 6 Measures of Organizational Impact Author(s) Description of Study Type Description of Measure(s) Benbasat and IDexter (1985) Benbasat and Dexter (1986) Benbasat. Dexter, and Masulis(198l) Bender(1986)
  • 100. CronandSobol(1983) Edelman(I981) Ein-Dor. Segev. and Steinfeld (1981) Griese and Kurpicz (1985) Jenster (1987) Kaspar and Cerveny (1985) Lincoln (1986) Lucas(I981) Miller and Doyle (1987) Miilman and Hartwick (1987) Perry (1983) Remus (1984) Financial; 65 business students Financial; 65 business students
  • 101. Pricing; one university. 50 students and faculty Overall I/S; 132 life insurance companies Overall I/S; 138 small to medium-sized wbolesalers Industrial relations; one firm, 14 operating units PERT; one R & D organization 24 managers Overall I/S; 69 firms I/S which monitors critical success factors; 124 organizations End user systems; 96 MBA students Specific 1/S applications; 20 organizations. 167 applications Inventory ordering system: one university, 100 executives Overall I/S; 21 financial firms, 276 user managers
  • 102. Office I/S; 75 middle managers Office 1/S; 53 firms Production scheduling system; one university, 53 junior business students Lab Profit performance Lab Profit performance Lab Profit Field Ratio of total general expense to total premium income Field (I) Pretax return on assets (2) Return on net worth (3) Pretax profits (% of sales) (4) Average 5-year sales growth Field Overall manager productivity (cost of information per employee) Field Profitability Field Number of computer applications Field (I) Economic performance (2) Marketing achievements (3) Productivity in production (4) Innovations
  • 103. (5) Product and management quality Lab (1) Return on assets (2) Market share (3) Stock price Field (I) Internal rate of return (2) Cost-benefit ratio L^b Inventory ordering costs Field Overall cost-effectiveness of I/S Field Organizational effectiveness Field I/S contribution to meeting goals Lab Production scheduling costs 82 Information Systems Research 3 : Information Systems Success TABLE 6 (com "rf) Author(s) Description of Study Type Description of Measure(s) Rivard andHuff(1984) Turner (1982) Vasarhelyi(l981) Yap and Walsham