The document discusses the ethical concerns and implications of data science, particularly related to algorithmic bias and its impact across various sectors such as law enforcement, education, and finance. It highlights initiatives like IEEE P7003 focusing on accountability in algorithm design to mitigate unjust impacts on users and outlines broader social issues like privacy and bias. Additionally, the document emphasizes the importance of ethical standards in technology development and advocates for responsible practices in data science.
Introduction to the ethics of data science, emphasizing personal views and disclaimers regarding institutional affiliations.
Discussion on predictive models and broader social trends affecting their application, highlighting issues like lack of transparency and practitioner training.
Overview of significant impact areas for data science, including concerns in credit, marketing, and healthcare analytics.
Exploration of using data science ethically to solve social issues, referencing IEEE's initiative for value-sensitive design in tech.
Introduction to various IEEE P7000 Working Groups focused on ethical considerations in technology and system design.
Overview of the IEEE P7003 standard aimed at ensuring accountability in algorithm design to prevent bias.
Discussion on broader social issues impacting IT professionals and public disillusionment with algorithms and data practices.
Exploration of various ethical dilemmas related to sustainability, bias, and data privacy, and their implications.
Concerns regarding user privacy, government surveillance, and ethical decision-making in algorithmic contexts.
Issues caused by technological advances in medical decisions, emphasizing the undermining of specialists' roles.
Critique of facial recognition technology in law enforcement, highlighting bias and transparency issues in algorithmic tools.
Issues concerning risk in data science projects, including unintentional misuse and ethical lapses.
Discusses the need for security and reliability frameworks in AI implementations across various industries.
Exploration of transparency and interpretability challenges in machine learning, alongside a call for accountability.
Insights into risks faced by organizations due to data science practices and the importance of ethical frameworks.
Discussion on the connection between ethics in data practices and how professional standards exist in computing.
Overview of ethical challenges in data science life cycles and the framework for analytics as code.
Importance of addressing ethical abuse scenarios in agile development and the need for diverse stakeholder inclusion.
Introduction to various IEEE societies and their roles in addressing ethical implications of technology.
Roadmap for addressing algorithmic bias, promoting responsible data science practices among stakeholders.
Final thoughts on the importance of ethical practices in data science and a call to action for responsible behavior.
Ethics of DataScience
Issues Surfaced by IEEE P7003 WG*
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07
2.
Disclaimers
My viewsonly
Not view of my employer
Not view of IEEE or IEEE SA
Not view of the IEEE P7003 WG
P7003 standards work still early stage
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
3.
“Minority Report” (2002)
The “PreCogs” have landed
Proprietary predictive models already deployed in several
states for
Law enforcement
Child welfare
“Pockets of poverty” identification
Educational / teacher assessment
Credit: Philip K. Dick (1956)
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
4.
Storm Clouds
Broadersocial trends impacting application contexts
ML weak transparency
Some analytics practitioners may have weak statistical background
Algorithms may not have faced usability or reproducibility tests
Computation on encrypted data is still primitive and/or slow
Anonymization practices may not take big data variety into account
Lightweight practitioner training
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
5.
Where Data ScienceDoes Matter
A smattering of impact areas
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
6.
Use Case Clusters
Credit worthiness (Offers for balance transfers not extended to minority groups)
Sales & marketing analytics (Some profitable prospects will be ignored)
Autonomous vehicles (Fatality, injury accidents)
Drug trials (Some patients will be rejected, even if potentially life-saving)
Recommendation engines (Users not given opportunity to fine-tune – You, Netflix)
Maintenance forecasting (Forecasts may not include loss of life risks)
Operations Management (Cause cost overrun while managing AWS cost / performance)
Cybersecurity (False positives could suggest you are an insider threat)
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
7.
Solving Poverty throughData Science
It’s Magic!
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
https://www.marketplace.org/shows/marketplace-morning-report 2018-07-30
8.
IEEE Initiative:
Value-Sensitive Design
“.. . to refine ethical system requirements in systems and software life cycles.”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
9.
IEEE P7000 WG:Marquis Group Charter
“Scope: The standard establishes a process model by which engineers and technologists can address
ethical consideration throughout the various stages of system initiation, analysis and design.
Expected process requirements include management and engineering view of new IT product
development, computer ethics and IT system design, value-sensitive design, and stakeholder
involvement in ethical IT system design. . .. The purpose of this standard is to enable the pragmatic
application of this type of Value-Based System Design methodology which demonstrates that
conceptual analysis of values and an extensive feasibility analysis can help to refine ethical system
requirements in systems and software life cycles.”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
10.
Related IEEE P70nnGroups
IEEE P7000 Ethical Systems Design
IEEE P7001 Transparency of Autonomous Systems
IEEE P7002 Data Privacy Process
IEEE P7003 Algorithmic Bias Considerations
IEEE P7004 Standard for Child and Student Data Governance
IEEE P7005 Standard for Transparent Employer Data Governance
IEEE P7006 Standard for Personal AI Agent
IEEE P7007 Ontological Standard for Ethically Driven Robotics and Automation Systems
IEEE P7008 -Standard for Ethically Driven Nudging for Robotic, Intelligent and Autonomous Systems
IEEE P7009 Standard for Fail-Safe Design of Autonomous and Semi-Autonomous Systems
IEEE P7010 Wellbeing Metrics Standard for Ethical Artificial Intelligence and Autonomous Systems
IEEE P7011 SSIE Standard for Trustworthiness of News Media
IEEE P7012 SSIE Machine Readable Personal Privacy Terms
IEEE P7013 Facial Analysis
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
11.
IEEE P7003 AlgorithmicBias Considerations
“This standard is designed to provide individuals or organizations creating algorithms, largely in
regards to autonomous or intelligent systems, certification oriented methodologies to provide
clearly articulated accountability and clarity around how algorithms are targeting, assessing and
influencing the users and stakeholders of said algorithm. Certification under this standard will
allow algorithm creators to communicate to users, and regulatory authorities, that up-to-date
best practices were used in the design, testing and evaluation of the algorithm to avoid
unjustified differential impact on users.”
IEEE PAR https://standards.ieee.org/develop/project/7003.html
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
12.
Who is IEEESA?
Why care what it does?
• Affordable, volunteer-driven, int’l
• IEEE SA members voting rights
• Collaboration with ISO, NIST
• Key standards include ethernet
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
13.
Broad Social Issues
Cross-cuttingConcerns for all Technologists
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
14.
Professional Pride, PublicDisillusionment
Broader acceptance within IT & Evidence-based Practices
Growth of data science inside many professions (R, Python)
Extraordinary explosion of OSS tooling
Big Data, ML, Real Time
Watson, AlphaGo, Alexa “AI” (Gee Whiz factor)
Public Perspective
“2017 was the year we fell out of love with algorithms.”
Cambridge Analytica, Equifax
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
15.
“When we fellout of love with algorithms.”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
16.
Ethical issues Alreadyin Play
Sustainability
Environment
Climate Change (*data center power consumption)
Bias concerns in gender, race, free speech
Social media technology responsibility
As propaganda platforms
Excessive use of cell phones by children: ADHD?
Weakened critical thinking, F2F social skills (Sherry Turkle Reclaiming Conversation 2015)
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
17.
Some Related ISOStandards
ISO 26000 “ISO 26000 provides guidance on how businesses and organizations can operate in a socially responsible
way. This means acting in an ethical and transparent way that contributes to the health and welfare of society”
ISO 27500:2016 is intended for executive board members and policy makers of all types of organizations (whether
large or small) in the private, public and non-profit sectors. It describes the values and beliefs that make an
organization human-centred, the significant business benefits that can be achieved, and explains the risks for the
organization of not being human-centered
ISO 37101:2016 establishes requirements for a management system for sustainable development in communities,
including cities, using a holistic approach, with a view to ensuring consistency with the sustainable development
policy of communities.
ISO/IEC JTC 1/SC 39 Standardization related to the intersection of resource efficiency and IT which supports
environmentally and economically viable development, application, operation and management aspects.
ISO/IEC 31010:2009 – Risk Management – Risk Assessment Techniques
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
18.
What do youknow about these
standards?
To date, a bit hidden among the blockchain and TensorFlow conversations.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
19.
Algorithmic Bias inData Science
*Including finance
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
20.
Linkage to Privacy,Surveillance, Distrust
Algorithms are bad because they:
Use data without our knowledge
Are based on incorrect or misleading knowledge about us
Are not accountable to individual citizens
Are used by governments to spy on citizens
Support drone warfare
Are built by specialists who do what they are told without asking questions
Represent a trend to automate jobs out of existence
Are built by big companies with no public accountability
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
21.
“Bias is theNew Breach”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“Researchers from MIT and Stanford University
tested three commercially released facial-analysis
programs from major technology companies and
will present findings that the software contains
clear skin-type and gender biases. Facial
recognition programs are good at recognizing
white males but fail embarrassingly with females
especially the darker the skin tone. The news
broke last week but will be presented in full at the
upcoming Conference on Fairness, Accountability,
and Transparency.“
https://www.cio.com/article/3256272/artificial-intelligence/in-the-ai-revolution-bias-is-the-new-
breach-how-cios-must-manage-risk.html
22.
Decisions vs. DecisionSupport:
Application Areas
Human-Computer Interactions in Decision-making
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
23.
Undermining Specialists*
Mark Underwood@knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“The threat the electronic health records
and machine learning post for physicians’
clinical judgment – and their well-being.” – NYT
2018-05-16
“’Food poisoning’ was diagnosed because
the strangulated hernia in the groin was
overlooked, or patients were sent to the
catheterization lab for chest pain because
no one saw the shingles rash on the left
chest.”
*Or adversely changing specialist behavior.
24.
“Rote Decision-Making”
Mark Underwood@knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“The authors, both emergency room physicians at
Brigham and Women’s Hospital in Boston, do a fine job
of sorting through most of the serious problems in
American medicine today, including the costs, over-
testing, overprescribing, overlitigation and general
depersonalization. All are caused at least in part, they
argue, by the increasing use of algorithms in medical
care.” -NYT 2018-04-01
25.
Facial Recognition forLaw Enforcement
“AMZ touts its Rekognition facial recognition system as ‘simple and easy to
use,’ encouraging customers to ‘detect, analyze, and compare faces for a
wide variety of user verification, people counting, and public safety use
cases.’ And yet, in a study released Thursday by the American Civil Liberties
Union, the technology managed to confuse photos of 28 members of
Congress with publicly available mug shots. Given that Amazon actively
markets Rekognition to law enforcement agencies across the US, that’s
simply not good enough. The ACLU study also illustrated the racial bias that
plagues facial recognition today. ‘Nearly 40 percent of Rekognition’s false
matches in our test were of people of color, even though they make up only
20 percent of Congress,’ wrote ACLU attorney Jacob Snow. ‘People of color
are already disproportionately harmed by police practices, and it’s easy to
see how Rekognition could exacerbate that.’“ -Wired 2018-07-26
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
26.
“Family” Impacts
Mark Underwood@knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“Charges of faulty forecasts have accompanied the
emergence of predictive analytics into public policy.
And when it comes to criminal justice, where
analytics are now entrenched as a tool for judges
and parole boards, even larger complaints have
arisen about the secrecy surrounding the workings
of the algorithms themselves — most of which are
developed, marketed and closely guarded by private
firms. That’s a chief objection lodged against two
Florida companies: Eckerd Connects, a nonprofit,
and its for-profit partner, MindShare Technology.” –
NYT “Can an algorithm tell when kids are in danger?” 2018-01-02
27.
Lawsuit over TeacherEvaluation Algorithm
Value-added measures for teacher evaluation, called the Education Value-
Added Assessment System, or EVAAS, in Houston, is a statistical method
that uses a student’s performance on prior standardized tests to predict
academic growth in the current year. This methodology—derided as
deeply flawed, unfair and incomprehensible—was used to make decisions
about teacher evaluation, bonuses and termination. It uses a secret
computer program based on an inexplicable algorithm (above).
In May 2014, seven Houston teachers and the Houston Federation of
Teachers brought an unprecedented federal lawsuit to end the policy,
saying it reduced education to a test score, didn’t help improve teaching or
learning, and ruined teachers’ careers when they were incorrectly
terminated. Neither HISD nor its contractor allowed teachers access to the
data or computer algorithms so that they could test or challenge the
legitimacy of the scores, creating a ‘black box.’” http://kbros.co/2EvxjU9
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
28.
Mark Underwood @knowlengr| Views my own | Creative Commons | *Thru 2018-07 | v1.6
Wells Fargo Credit Denial “Glitch”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-08 | v1.4
CNN: “Hundreds of people
had their homes foreclosed
on after software used by
Wells Fargo incorrectly denied
them mortgage
modifications.” 2018-08-05
https://money.cnn.com/2018/08/04/news/companies/wells-fargo-mortgage-modification/index.html
29.
. . .And not easy to “fix”
Risk mitigation for data science implementations is relatively immature.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
30.
Unintended Use Casesor Ethical Lapse?
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
• Algorithm corrected for color bias, but can
now be used for profiling
• “Red Teaming” or “Abuse User Stories” can
help
• Unintended use cases call for a safety vs. a
pure “assurance” framework
31.
“Lite” AI Security/ReliabilityFrameworks
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
https://motherboard.vice.com/en_us/article/bjbxbz/researchers-tricked-ai-into-doing-free-computations-it-wasnt-trained-to-do
“Google researchers demonstrated that a
neural network could be tricked into
performing free computations for an
attacker. They worry that this could one
day be used to turn our smartphones into
botnets by exposing them to images.”
32.
Machine Learning:
Expanding Impact
Impactacross all industries may be difficult to assess.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
33.
The “Cambrian Explosion”
MarkUnderwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
-- Ian Goodfellow
Staff Research Scientist
Google Brain
ACM Webinar 2018-07-24
34.
Transparency & Interpretability
AsSeen by Ian Goodfellow
Findings from adversarial machine learning
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
35.
Why Ethical IssuesBecome Secondary
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
For many applications, ML benefits will
prove irresistible.
Goodfellow considers reliability needs as a
useful example.
36.
Challenges of Interpretability
MarkUnderwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“Adversarial ML literature
suggests that ML models are
very easy to fool and even
linear models work in counter-
intuitive ways.” (Selvaraju et al, 2016)
• Reproducability
• Training sets including results of
other analytics (e.g., FICO)
• Provenance (think IoT)
• Opaque statistical issues
37.
Transparency General Challenges
Some data, algorithms are intellectual property
Some training data includes PII
Predictive analytical models are often “point in time”
“Transparent” according to whose definition?
Should algorithms have “opt-in?” Can they?
Training set big data variety reidentification risks
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
38.
Explainability / Interpretability
MarkUnderwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“[We need to] find ways of making techniques like
deep learning more understandable to their creators
and accountable to their users. Otherwise it will be
hard to predict when failures might occur—and it’s
inevitable they will. That’s one reason Nvidia’s car is
still experimental.”
39.
Enterprise Level Risk
Impact on reputation
Litigation
Unintentionally reveal sources, methods, data / interrupted data streams (e.g, web)
Loss of consumer confidence, impact on public safety
Misapplication of internally developed models
Financial losses from data science #fail
“. . . as long as our training is in the form of someone lecturing about the basics of gender or
racial bias in society, that training is not likely to be effective”.
Dr. Hanie Sedghi, Research Scientist, Google Brain
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
40.
Risk Management
and Lessonsfrom Others
Aligning with enterprise risk appetite, resilience
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
41.
Insights from MoreMature Settings
AI Analytics for distributed military coalitions
“. . . Research has recently started to address such concerns and
prominent directions include explainable AI [4], quantification of
input influence in machine learning algorithms [5], ethics
embedding in decision support systems [6], “interruptability” for
machine learning systems [7], and data transparency [8]. “
“. . . devices that manage themselves and generate their own
management policies, discussing the similarities between such
systems and Skynet.”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
S. Calo, D. Verma, E. Bertino, J. Ingham, and G. Cirincione, "How to prevent skynet
from forming (a perspective from Policy-Based autonomic device management),"
in 2018 IEEE 38th International Conference on Distributed Computing Systems
(ICDCS), Jul. 2018, pp. 1369-1376. [Online]. Available:
http://dx.doi.org/10.1109/ICDCS.2018.00137
42.
Some Fairness Approaches(Sedghi, 2018)
1) Post processing in terms of calibration of our model. What this means is that, we calibrate
classifiers parameters such that it has the same acceptance ratio for all subgroups of sensitive
features, e.g. race, sex, etc.
2) Data resampling to remove skewed sample. But, for many reasons, collecting more data is not
very easy and sometimes causes problems for individuals.
3) Causal reasoning: We capture different paths in a causal graph that can lead to the same
observational data. This basically means to model possible factors such that sex, race and other
sensitive features to make sure their impact is captured and does not directly affect the result
variable.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
43.
Decision Support forBias Detection
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
“Things like transparency, intelligibility, and explanation are new enough
to the field that few of us have sufficient experience to know everything
we should look for and all the ways that bias might lurk in our models,”
says Rich Caruna, a senior researcher at Microsoft who is working on
the bias-detection dashboard.”
Technology Review, Will Knight 2018-05-25
44.
“Fairness Flow”:
But willyou share your ethics guidance?
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
https://www.cnet.com/news/facebook-starts-building-ai-with-an-ethical-compass/
“Bin Yu, a professor at UC Berkeley, says
the tools from Facebook and Microsoft
seem like a step in the right direction,
but may not be enough. She suggests
that big companies should have outside
experts audit their algorithms in order
to prove they are not biased. ‘Someone
else has to investigate Facebook's
algorithms—they can't be a secret to
everyone,” Yu says.’”
-Technology Review 2018-05-25
45.
Professional Level Risk
Skills obtained at considerable cost/effort will be deprecated
Alternative approaches will be preferred
Good, suitable approaches will be discouraged
-- in order to avoid negative result approaches
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
46.
Codes of Ethics
IEEE& ACM
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
47.
Joint ACM-IEEE SoftwareEngr Code
Assignment: Find the ethical trouble spots.
https://www.computer.org/web/education/code-of-ethics
1. PUBLIC - Software engineers shall act consistently with the public interest.
2. CLIENT AND EMPLOYER - Software engineers shall act in a manner that is in the best interests of their client and
employer consistent with the public interest.
3. PRODUCT - Software engineers shall ensure that their products and related modifications meet the highest
professional standards possible.
4. JUDGMENT - Software engineers shall maintain integrity and independence in their professional judgment.
5. MANAGEMENT - Software engineering managers and leaders shall subscribe to and promote an ethical approach
to the management of software development and maintenance.
6. PROFESSION - Software engineers shall advance the integrity and reputation of the profession consistent with the
public interest.
7. COLLEAGUES - Software engineers shall be fair to and supportive of their colleagues.
8. SELF - Software engineers shall participate in lifelong learning regarding the practice of their profession and shall
promote an ethical approach to the practice of the profession.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
48.
Transparency & ProfessionalEthics
What connection to IEEE /ACM professional ethics?
ACM: “The entire computing profession benefits when the ethical decision making process is accountable to
and transparent to all stakeholders. Open discussions about ethical issues promotes this accountability and
transparency.”
ACM “A computing professional should be transparent and provide full disclosure of all pertinent system
limitations and potential problems. Making deliberately false or misleading claims, fabricating or falsifying data,
and other dishonest conduct are violations of the Code.”
ACM “Computing professionals should establish transparent policies and procedures that allow individuals to
give informed consent to automatic data collection, review their personal data, correct inaccuracies, and, where
appropriate, remove data.”
ACM “Organizational procedures and attitudes oriented toward quality, transparency, and the welfare of society
reduce harm to the public and raise awareness of the influence of technology in our lives. Therefore, leaders
should encourage full participation of all computing professionals in meeting social responsibilities and
discourage tendencies to do otherwise.”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
49.
State of ComputingProfession Ethics
@ACM_Ethics
Not part of professional training
Enterprise ethics training is not technical enough
Engineers are not challenged with realistic moral dilemmas
Overlay of employment, profit, team loyalty as potential crosswinds
Project Management practices dilute responsibility
Requirements Engineering is more art than engineering
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
50.
SDLC for DataScience
Life cycle of analytics | Analytics as Code
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
51.
Analytics as Code
Requirements Engineering
Traceability and provenance (training data, code, constraints)
Audit & Forensics
Ethical tradeoffs tagged like PII (e.g., sustainability tradeoff for an oil company)
Test automation (“left-shifted” test engineering, “Test-Driven Development”
Dependency management
Supporting infrastructure analogous to other application spaces
Dashboard integration
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
52.
Continuously Deployed Algorithms
Impactsof DevOps (IEEE P2675)
Test Harness construction
Workflow for retest, refinement, reproducibility
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
53.
Training Set Scrutiny
Training sets can get ethics off on the wrong foot
Set responsibility for ethical ownership - e.g., similar to Institutional
Review Boards (IRB)
Be prepared to defend incomplete, tentative, or experimental projects
“Data” is part of data science “code”
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
54.
Recognize Ethical, AbuseScenarios
For agile SDLC: User stories should include all stakeholders
Often supply chain, OSS, internal customers are omitted from
scenarios
Add “abuse” or “adversarial” user stories / security stories (Secure Scrum
parlance)
May require an external sensibility to discover & recite
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
55.
Other Relevant IEEESocieties
IEEE Groups
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
56.
IEEE Society onSocial Implications
of Technology
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
57.
IEEE Product SafetyEngineering Society
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
• “Do no harm.” – It’s not
so easy.
• Do you know a system is
safe before it’s been fully
scaled up -- & possibly
federated?
• What constitutes “a
reasonable explanation”?
58.
IEEE Reliability Society
MarkUnderwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
See free reliability analytics
toolkit
https://kbros.co/2rugRij
Roadmaps: The AdversarialUser Story
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
Image via Robert C. Seacord
writing about software security
61.
Algorithmic Bias RiskManagement
1. Recognize, socialize groups protected by statute (e.g.,
Equal Credit Opportunity Act)
2. Creatively consider other affected subpopulations
Sight impaired – other disabilities
Children, elderly
Unusual household settings (elder care, multi-family
housing)
Part time and workers
Novice vs. Experienced users
What counterfactuals are simply not being measured?
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
62.
Roadmap Attributes
Engageall stakeholders: consumers, co-workers, shareholders, specialists,
professional CoI’s
Embrace or consult “traditional” AI
Tie to enterprise risk appetite
Align with professional Codes of Ethics
“User Stories” should include adversarial ethical challenges
Traceable responsibility for decision-making (e.g., tradeoff analysis)
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
63.
Suggested Area ofStudy: Ethical Intent
Rao, Anand; Georgeff, Michael. 1995. Belief-Desire-Intention Agents: From Theory to
Practice. Australian Artificial Intelligence Institute, Melbourne, Australia. In ICMAS-95.
Casali, A., Godo, L. and Sierra, C., 2011. A graded BDI agent model to represent and reason about
preferences. Artificial Intelligence, 175(7-8), pp.1468-1478.
Xiaowei Wang, Nicola Guarino, Giancarlo Guizzardi, and John Mylopoulos. 2014. Towards an Ontology
of Software: a Requirements Engineering Perspective. FOIS 2014.
https://www.academia.edu/8744511/Towards_an_Ontology_of_Software_A_Requirements_Engineering_
Perspective. [local]
Guarino, N., 2006. Ontology and Terminology. In How can formal ontology help concept modelling
and terminology. Powerpoint in Workshop «Terminology, Concept Modelling and Ontology. Which
approach for which problem.
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
64.
A Final Rationale
MarkUnderwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6
65.
Mark Underwood @knowlengr| Views my own | Creative Commons | *Thru 2018-07 | v1.6
• Co-Chair NIST Big Data Public WG Security & Privacy subgroup https://bigdatawg.nist.gov/
• Chair Ontology / Taxonomy subgroup for IEEE P7000. Occasional participant in IEEE Standards
WGs P7007, P7003, P7002, P7004, P7010
• IEEE Standard P1915.1 Standard for Software Defined Networking and Network Function
Virtualization Security (member)
• IEEE Standard P2675 WG Security for DevOps (member)
• Current: Finance, large enterprise: supply chain risk, complex playbooks, many InfoSec tools,
workflow automation, big data logging; risks include fraud and regulatory #fail
• Authored chapter “Big Data Complex Event Processing for Internet of Things Provenance:
Benefits for Audit, Forensics, and Safety” in Cyber-Assurance for IoT (Wiley, 2017)
https://kbros.co/2GNVHBv
• @knowlengr dark@computer.org knowlengr.com https://linkedin.com/in/knowlengr
About Me
66.
This deck isreleased under
Creative Commons
Attribution-Share Alike.
See also: Stakeholders in Ethical Systems Design
Mark Underwood @knowlengr | Views my own | Creative Commons | *Thru 2018-07 | v1.6