The	Ethics	of	Machine	Learning	Algorithms		 	
	
Introduction	
The	advent	of	the	‘Information	Age’	gave	birth	to	novel	technologies	capable	of	
dealing	 with	 large	 data	 sets	 and	 capable	 of	 inferring	 useful	 information	 from	
large	data	sets.	One	such	technology	is	Machine	Learning.	Machine	Learning	is	
‘‘any	methodology	and	set	of	techniques	that	can	employ	data	to	come	up	with	
novel	 patterns	 and	 knowledge,	 and	 generate	 models	 that	 can	 be	 used	 for	
effective	predictions	about	the	data’’	(Van	Otterlo,	2013).		
Machine	 Learning	 has	 been	 used	 extensively	 to	 draw	 significant	 insights	 from	
unwieldy	data	sets,	and	have	been	proposed	to	be	the	“future”	of	algorithms	and	
analytics	 (Tutt,	 2016).	 More	 and	 more	 responsibilities	 are	 being	 delegated	 to	
machine	 learning	 algorithms	 (Mittelstadt	 et	 al,	 2016).	 For	 instance,	 Machine	
Learning	 algorithms	 are	 employed	 to	 recognise	 human	 bias	 or	 inaccurate	
information,	as	displayed	by	Wikipedia’s	Objective	Revision	Evaluation	Service,	
to	 act	 as	 recommendation	 systems	 for	 customers	 advising	 them	 on	 their	 next	
purchase	based	on	their	purchase	history	(Vries,	2010)	and	to	facilitate	in	the	
understanding	 of	 behavioural	 data	 emerging	 from	 the	 ‘Internet	 of	 Things’	
(Portmess	and	Tower,	2014).	This	list	is	by	no	means	exhaustive.			
Given	 their	 wide	 applications,	 it	 is	 plausible	 to	 assume	 that	 Machine	 Learning	
algorithms	are	an	integral	set	of	technologies	to	industrial	and	academic	settings.	
(Citron,	2007)	argues	that	developers	of	algorithms	have	a	tendency	to	accept	
that	 automated	 processes,	 as	 those	 encountered	 during	 the	 application	 of	
Machine	 Learning	 algorithms,	 are	 true	 and	 precise,	 overcoming	 the	 hurdle	 of	
human	bias	(Naik	and	Bhide,	2014).	As	controversial	as	this	statement	is,	it	does	
highlight	 the	 dependency	 of	 Machine	 Learning	 algorithms	 in	 industrial	 and	
academic	settings.		
However,	these	technologies	come	at	an	ethical	cost.	Given	that	algorithms	are	
unavoidably	 influenced	 by	 personal	 opinion,	 since	 the	 parameters	 are	 finely	
tuned	 by	 the	 developer	 (Wiener,	 1988),	 then	 it	 comes	 as	 no	 surprise	 that	
designing	 and	 developing	 algorithms	 does	 not	 guarantee	 that	 ethical	
considerations	will	not	be	violated.		For	example,	profiling	algorithms	have	to	
discriminate	 to	 achieve	 their	 purpose,	 but	 sometimes	 concomitantly	 are	
derogatory	(Barocas	and	Selbst,	2015).
The	 publically	 familiar	 definition	 of	 algorithms	 is	 broad.	 An	 algorithm	 can	
essentially	be	any	set	of	instructions	provided	to	achieve	a	desired	outcome.	For	
the	purpose	of	this	essay,	this	definition	is	too	broad	and	should	be	restricted	
appropriately.	Therefore,	this	essay	will	adopt	Hill’s	definition	of	an	algorithm	as	
a	 “mathematical	 construct	 with	 a	 finite,	 abstract,	 effective,	 compound	 control	
structure,	 imperatively	 given,	 accomplishing	 a	 given	 purpose	 under	 given	
provisions”	(Hill,	2015).		This	definition	takes	into	account	the	implementation	
of	an	algorithm	into	a	technology,	and	the	application	of	that	technology	aimed	at	
fulfilling	 a	 predefined	 purpose.	 However,	 the	 discussion	 regarding	 ethics	
implicated	 in	 the	 application	 of	 technologies	 based	 on	 Machine	 Learning	
algorithms	 will	 be	 limited.	 Moreover,	 this	 essay	 will	 only	 focus	 on	 algorithms	
whose	decision-making	is	difficult	to	explain	from	a	human	perspective	due	to	
the	 algorithm’s	 opaque	 nature	 (how	 a	 new	 input	 will	 be	 handled),	 and	
algorithms	 whose	 actions	 are	 difficult	 to	 predict.	 Algorithms	 implemented	 for	
mundane	tasks,	such	as	manufacturing,	are	not	considered	here.	The	essay	will	
discuss	 several	 broad	 ethical	 concerns	 regarding	 the	 use	 of	 algorithms,	 each	
placed	under	a	distinct	subcategory.		
	
Algorithmic	Bias		
Algorithms	 inescapably	 contain	 some	 element	 of	 bias,	 although	 some	 would	
argue	 that	 their	 deployment	 is	 actually	 an	 approach	 to	 overcome	 human	 bias	
(Naik	 and	 Bhide,	 2014).	 One	 example	 of	 bias,	 classed	 as	 ‘technical	 bias’,	 is	
observed	when	the	listing	of	companies	for	a	search	result	follows	alphabetical	
order,	potentially	resulting	in	more	business	for	companies	that	appear	earlier	in	
the	alphabet	(Friedman	and	Nissenbaum,	1996).	
The	reason	why	algorithms	can	be	bias	is	because	the	values	of	the	developer	are	
institutionalised	into	the	algorithm,	and	because	algorithms	are	designed	with	an	
intended	purpose	in	mind	(Macnish,	2012).	During	the	course	of	the	algorithm’s	
development,	there	is	not	always	an	objectively	correct	choice,	but	rather	there	
is	a	myriad	of	alternatives	that	could	still	achieve	the	same	desired	objective.		
The	bias	can	arise	from	many	situations	(Friedman	and	Nissenbaum,	1996).	For	
example,	 bias	 can	 arise	 during	 supervised	 learning	 tasks.	 During	 supervised	
learning,	the	Machine	Learning	algorithm	learns	from	human-tagged	data	inputs.
The	 labelling	 of	 these	 inputs	 can	 inadvertently	 be	 bias	 (Diakopoulos,	 2015).		
Furthermore,	the	interpretation	of	the	outputs	of	the	algorithm	can	also	be	bias	
(Hildebrandt,	2011).	
Not	 many	 improvements	 can	 be	 made	 for	 bias	 emerging	 from	 technical	
constraints	 (i.e.	 technical	 bias).	 However,	 a	 collective	 effort	 could	 reduce	 bias	
emerging	from	other	sources,	such	as	social	bias	that	become	manifested	in	the	
algorithms.	 Raymond	 (2014)	 proposes	 that	 human	 monitoring	 may	 be	 an	
effective	approach	to	reduce	some	bias.			
	
Unfair	Outcomes	Based	on	Discrimination		
One	 broad	 category	 of	 ethical	 considerations	 is	 discrimination	 resulting	 from	
profiling	(Vries,	2010).	Classification	algorithms	are	often	employed	in	industry	
and	academia	for	various	purposes.	These	classification	algorithms	classify	new	
inputs	 based	 on	 the	 characteristics	 of	 the	 training	 data.	 The	 description	 of	
classification	 algorithms	 already	 hints	 at	 some	 ethical	 issues	 that	 can	 be	
encountered	during	the	employment	of	these	algorithms.	Practical	examples	of	
discrimination	include	the	delivery	of	online	advertisements	based	on	perceived	
ethnicity	 (Sweeny,	 2013),	 and	 personalised	 pricing	 (Danna	 and	 Gandy,	 2002).	
Another	 example	 is	 an	 infamous	 ‘Data	 Science	 gone	 wrong’	 story	 pertinent	 to	
classification	algorithms,	which	is	of	Target’s	unfortunate	labelling	of	a	13-year-
old	 girl	 as	 pregnant,	 before	 her	 father	 was	 even	 aware	 of	 her	 pregnancy	
(Golgowksi,	 2012).	 It	 is	 important	 to	 note	 though,	 that	 the	 last	 example	 has	
ethical	 concerns	 reaching	 beyond	 discrimination.	 The	 ethics	 of	 classification	
algorithms	 are	 even	 more	 extensive	 in	 the	 context	 of	 classification	 algorithms	
recruited	 for	 medical	 diagnoses	 (Cohen	 et	 al,	 2014).	 Such	 models	 aid	 in	 the	
diagnosis	 and	 also	 recommend	 viable	 treatment	 plans	 to	 physicians	 (Mazoue,	
1990).	
Schermer	 (2011)	 argues	 that	 the	 act	 of	 discriminatory	 treatment	 itself	 is	 not	
ethically	challenging,	but	that	the	resulting	effects	from	discriminatory	treatment	
are	 ethically	 challenging.	 However,	 (Mittelstadt	 et	 al,	 2016)	 suggests	 that	
Schermer	is	‘muddling’	up	the	concepts	of	bias	and	discrimination,	incorrectly	
using	discrimination	to	describe	bias.
There	are	multiple	diverse	reasons	to	consider	discriminatory	effects	as	adverse,	
implying	 that	 they	 should	 be	 prevented	 if	 possible,	 or	 at	 least	 minimised	 to	
almost	extinction.	It	has	been	postulated	that	discrimination	can	contribute	to	
self-fulfilling	 prophecies	 and	 that	 it	 can	 exaggerate	 stigmatisation	 (Macnish,	
2012).	 Furthermore,	 discrimination	 may	 reinforce	 advantages	 and	
disadvantages	already	existing	in	society,	exacerbating	social	discrepancy,	which	
violates	 ethical	 and	 legal	 principles	 of	 fair	 treatment	 (Newell	 and	 Marabelli,	
2015).		
Fortunately,	 it	 may	 be	 possible	 to	 develop	 algorithms	 that	 do	 not	 take	 into	
account	 sensitive	 attributes	 or	 features,	 which	 could	 potentially	 contribute	 to	
discriminatory	outcomes	(Barocas	and	Selbst,	2015).	For	instance,	(Kamiran	and	
Calders,	2010)	have	developed	algorithms	that	do	not	take	into	account	gender	
or	ethnicity,	without	a	loss	in	performance.	However,	this	is	strongly	subject	to	
context;	 some	 classification	 tasks	 may	 require	 the	 inclusion	 of	 sensitive	
attributes	 in	 order	 to	 perform	 to	 an	 adequate	 level.	 Thus,	 the	 question	 then	
becomes,	what	classification	tasks	can	be	delegated	to	classification	algorithms	
in	the	first	place?	Moreover,	using	proxies	for	sensitive	attributes	may	not	be	an	
effective	solution	(Romei	and	Ruggieri,	2014).		
	
Inappropriate	Actions	Based	on	Inconclusive	Data	
Correlations	 are	 commonly	 used	 to	 complement	 inductive	 knowledge	 in	
decision-making	 tasks.	 Correlations	 based	 on	 a	 large	 volume	 of	 data	 are	
extrapolated	 to	 direct	 actions,	 without	 the	 requirement	 to	 establish	 a	 causal	
relationship	first	(Hildebrandt,	2011).	This	can	result	in	unjustified	actions	based	
on	 inconclusive	 evidence.	 Mittelstadt	 et	 al	 (2016)	 coin	 this	 as	 ‘actionable	
insights’.	To	exacerbate	this	issue	of	actionable	insights,	(Lazer	et	al,	2014)	argue	
that	 correlations	 identified	 in	 large	 proprietary	 data	 sets	 are	 seldom	
reproducible	 due	 to	 their	 commercial	 sensitivity,	 and	 thus	 often	 present	 the	
opportunity	of	being	fabricated.	Ananny	(2016)	postulates	that	correlations	are	
doubly	uncertain;	not	only	can	they	result	in	actionable	insights,	but	acting	upon	
on	 correlations	 concerns	 individuals,	 while	 the	 knowledge	 concerns	 the	
population	from	which	the	conclusion	was	drawn	from	(Illari	and	Russo,	2014).
Furthermore,	 the	 sheer	 volume	 of	 ‘Big	 Data’	 does	 not	 imply	 that	 fundamental	
theoretical	aspects	of	measurement,	reliability	and	face	validity	can	be	neglected,	
and	 that	 solely	 large	 quantities	 of	 data	 do	 not	 serve	 as	 an	 appropriate	
replacement	 for	 improper	 methodology	 (Boyd	 and	 Crawford,	 2012).	 This	
concern	is	not	specific	to	correlations	either,	as	it	is	applicable	to	all	domains	
where	‘Big	Data’	is	employed	to	draw	conclusions.	
One	possible	solution	to	this	problem	is	to	train	data	analysts	not	to	draw	hasty	
conclusions,	 and	 to	 ensure	 that	 data	 quality	 is	 of	 the	 highest	 form.	 Whether	
actions	based	on	conclusive	evidence	are	deemed	ethically	acceptable	is	not	a	
discussion	that	is	pertinent	to	the	discussion	of	actionable	insights.				
	
Accountability		
Another	major	ethical	concern	of	Machine	Learning	algorithms,	as	suggested	by	
(Mittelstadt	et	al,	2016),	is	traceability.	The	authors	coined	this	term	to	describe	
the	difficulty	of	assigning	responsibility	or	culpability	to	an	agent	when	ethical	
concerns,	 such	 as	 any	 form	 of	 harm	 tout	 court,	 are	 violated	 during	 the	
application	of	Machine	Learning	algorithms.	Firstly,	it	is	important	to	note	that	
identifying	whether	some	Machine	Learning	algorithms	suffer	from	some	ethical	
concerns	due	to	human	subjectivity	is	difficult	to	decipher.		
Typically,	it	is	the	designers	and	users	of	the	algorithm	that	are	held	accountable	
if	 adversity	 is	 encountered	 (Kraemer	 et	 al,	 2011).	 However,	 (Matthias,	 2004)	
argues	 that	 blame	 can	 only	 be	 equipped	 when	 the	 designer	 or	 user	 has	 some	
element	 of	 control	 over	 the	 algorithm,	 and	 is	 also	 ill	 intentioned.	 Often,	 the	
developer	 or	 user	 does	 not	 have	 as	 much	 control	 over	 the	 algorithm	 as	
conventionally	believed,	creating	an	‘accountability	gap’,	where	it	is	not	clear	to	
determine	who	is	culpable	for	the	violation	of	ethical	considerations	(Cardona,	
2008).	In	essence,	shifting	blame	upon	the	developer	may	only	be	applicable	to	
cases	 where	 the	 operational	 rules	 of	 the	 algorithm	 are	 hand-written,	 and	 not	
rules	that	the	algorithm	learns	independently	(i.e.	Machine	Learning	algorithms).		
The	opposing	view	suggests	that	algorithms,	which	act	independently	and	result	
in	 ethically	 unjust	 outcomes,	 should	 be	 held	 accountable	 (Wiltshire,	 2015).	
However,	this	may	provide	an	opportunity	for	developers	to	assign	blame	to	the	
algorithm	(Crnkovic	and	Curuklu,	2011).	Furthermore,	the	capacity	of	algorithms
to	 make	 moral	 and	 ethically	 justifiable	 decisions	 is	 still	 disputable	 (Anderson,	
2008).		
Nonetheless,	 it	 can	 be	 argued	 that	 developers	 and	 users	 should	 monitor	 the	
workings	 of	 the	 algorithm	 for	 any	 ethical	 violations,	 as	 done	 so	 by	 (Fole	 and	
Ruddick,	 2004).	 Allen	 et	 al.	 (2006)	 insinuate	 that	 the	 void	 in	 ‘machine	 ethics’	
needs	 to	 be	 filled,	 so	 that	 ethical	 violations	 can	 be	 avoided.	 There	 is	 much	
research	 currently	 focused	 on	 developing	 algorithms	 with	 a	 concrete	
understanding	of	moral	and	ethical	responsibilities	(Wiegel	and	Berg,	2009),	so	
that	 neither	 extreme	 of	 assigning	 blame	 (solely	 blaming	 the	 developer	 or	 the	
algorithm)	has	to	be	adopted.			
	
Opacity	of	Algorithms		
The	 opacity	 of	 Machine	 Learning	 algorithms	 is	 another	 major	 concern	 in	 the	
Literature.	Supposedly,	algorithms	that	constitute	mechanisms	that	are	difficult	
to	 explain,	 or	 algorithms	 whose	 behaviour	 is	 difficult	 to	 predict,	 are	 also	
concomitantly	difficult	to	control,	regulate	and	correct	(Tutt,	2016).	Introna	and	
Nissenbaum	 (2000)	 is	 among	 pioneering	 publications	 discussing	 this	 topic	 of	
transparency.	 The	 authors	 suggested	 that	 search	 engines	 filter	 information	
based	on	market	conditions,	with	a	slight	preference	towards	powerful	actors.	In	
the	paper,	the	authors	discuss	corrective	mechanisms	intended	to	improve	the	
‘fairness’	of	search	engines.		
Anyhow,	 the	 accessibility	 of	 information	 elucidating	 the	 functionality	 of	
algorithms	 is	 often	 restricted	 because	 of	 national	 security	 (Leese,	 2014),	
competitive	advantage	(Glenn	and	Monteith,	2014),	or	privacy	in	general.	Thus,	
transparency	is	not	always	desired,	especially	not	at	the	expense	of	invasion	of	
privacy.	On	the	contrary,	it	can	be	argued	that	the	privacy	of	an	algorithm	is	not	
as	important,	except	for	cases	concerning	competitive	advantage,	and	that	only	
the	privacy	of	the	data	itself	should	be	regarded	as	important.	The	argument	of	
exposing	 the	 algorithm’s	 architecture	 can	 also	 be	 swayed	 in	 the	 opposite	
direction,	that	is,	informing	the	public	about	the	architecture	of	algorithms	may	
prompt	some	users	to	abusively	manipulate	the	algorithm	for	tailored	use,	while	
also	placing	non-tech-savvy	users	at	a	disadvantage	(Granka,	2010).
Intelligibility	 also	 contributes	 to	 the	 opacity	 of	 algorithms	 (Burrell,	 2016).		
Machine	 Learning	 algorithms	 are	 commonly	 described	 as	 ‘black	 boxes’,	 which	
are	fed	input	to	produce	an	output,	with	the	mechanisms	of	the	algorithm	being	
obscured.	For	many	applications	of	Machine	Learning	algorithms,	the	developer	
cannot	articulate	the	exact	procedure	needed	to	achieve	a	particular	goal,	hence	
the	requirement	for	these	algorithms	in	the	first	place	(Datta	et	al,	2016).	Even	
algorithms	whose	operation	can	be	decomposed	into	a	set	of	hand-written	rules	
have	the	capacity	to	be	inscrutable	(Kitchin,	2016).	To	complicate	matters	even	
more,	 algorithms	 are	 often	 designed	 by	 groups	 of	 engineers,	 rendering	 a	
comprehensive	 understanding	 of	 the	 algorithms	 extremely	 difficult,	 if	 not	
practically	infeasible	(Sandvig	et	al,	2014).	It	is	also	important	to	note	that	the	
intelligibility	 of	 algorithms	 can	 be	 limited	 by	 the	 questioner’s	 capacity	 to	
understand	 it.	 The	 problem	 with	 this	 is	 that	 full	 legitimate	 consent	 cannot	 be	
granted	if	the	mechanisms	of	the	algorithm	are	not	fully	understood	(Schermer,	
2011).	 To	 make	 the	 algorithms	 more	 accessible,	 the	 operational	 logic	 of	 the	
algorithm	 could	 be	 explained	 in	 simplified	 terms.	 However,	 (Kitchin,	 2016)	
suggests	that	this	may	only	be	effective	for	simple	algorithms,	and	may	not	be	a	
plausible	 solution	 for	 more	 complex	 algorithms,	 which	 could	 then	 still	
potentially	escalate	to	a	lack	of	trust	for	both	the	algorithm,	and	the	data	analyst	
(Rubel	and	Jones,	2014).	It	is	important	to	recognise	that	trust	can	be	placed	in	
the	algorithm	independent	of	the	data	analyst,	but	whether	this	is	appropriate	is	
still	controversial,	as	in	the	case	of	autonomous	weapons	(Swiatek,	2012).		
(Zarsky,	2013)	suggests	that	transparency	should	be	disclosed	to	trained	third	
parties,	seen	as	regulators,	acting	in	the	interest	of	the	public,	as	opposed	to	the	
questioners	themselves	to	prevent	any	loss	of	trust.	This	sense	of	trust	is	pivotal	
(Cohen	 et	 al,	 2014)	 and	 can	 alleviate	 concerns	 with	 the	 opacity	 of	 algorithms	
based	on	personal	information	(Mazoue,	1990).		
	
Transformative	Effects		
The	 results	 obtained	 from	 Machine	 Learning	 algorithms	 can	 change	 the	
perceptions	of	how	humans	perceive	the	world.	These	transformative	effects	can	
lead	to	ethically	challenging	concepts.
One	 example	 of	 such	 ethical	 concern	 is	 observed	 during	 the	 employment	 of	
personalisation	 algorithms.	 These	 algorithms	 obscure	 content	 encountered	 by	
the	user	that	is	believed	to	be	unaligned	to	the	user’s	belief,	or	be	contradictory	
to	the	user’s	belief	(Barnet,	2009).	Ananny	(2016)	suggests	that	algorithms	need	
not	be	explicitly	coercing	to	influence	perception	and	decision-making	through	
filtering.	 Personalisation	 algorithms	 can	 be	 seen	 as	 supporting	 the	 user	 by	
filtering	out	relevant	information	from	the	masses	of	information	presented,	or	
they	can	be	seen	as	controlling	by	presenting	the	user	with	only	information	the	
personalisation	algorithms	deems	relevant	(Bozdag,	2013).	This	could	result	in	
ethical	 violations	 when	 the	 choice	 of	 what	 data	 is	 presented	 prioritises	 the	
interests	 of	 a	 third-party	 over	 the	 interests	 of	 the	 user	 (Applin	 and	 Fischer,	
2015),	 as	 commonly	 demonstrated	 with	 online	 consumers	 (Coll,	 2013).	 The	
users	may	then	be	prompted	to	make	decisions	not	purely	based	on	their	own	
preference,	but	rather,	choices	will	be	based	on	inherently	subjective	restrictions	
of	information	load	(Johnson,	2013).		
The	 diversity	 of	 information	 presented	 is	 paramount	 in	 exercising	 autonomy	
(Van	 Den	 Hoven	 and	 Rooksby,	 2008).	 Thus,	 any	 inappropriate	 restriction	 of	
information	diversity	can	disrupt	the	autonomy	of	the	user.	Furthermore,	a	lack	
of	information	diversity	may	handicap	the	exploration	of	new	ideas	and	options	
(Raymond,	 2014).	 To	 counteract	 this	 ethical	 concern,	 (Lewis	 and	 Westlund,	
2015)	propose	that	personalisation	algorithms	should	be	taught	to	be	aware	of	
this	ethical	consideration,	of	not	being	too	persuasive	at	the	expense	of	limiting	
the	decisional	autonomy	of	the	user.		
However,	it’s	not	only	personalisation	algorithms	that	can	have	a	transformative	
impact.	 The	 general	 utilisation	 of	 algorithms	 is	 also	 transforming	 the	
conventional	 notions	 of	 informational	 privacy	 (Mittelstadt	 et	 al,	 2016).	 For	
instance,	 (Schermer,	 2011)	 confronts	 identifiability	 in	 profiling	 algorithms	 by	
arguing	that	profiling	algorithms	aim	to	designate	data	inputs	into	meaningful	
assemblies,	 for	 which	 individual	 identity	 is	 truly	 trivial.	 Yet,	 (Van	 Wel	 and	
Royakkers,	 2004)	 rebut	 by	 suggesting	 that	 profiling	 is	 a	 form	 of	 de-
personalisation	by	assigning	judgements	based	on	group	characteristics,	and	not	
individual	 merit.	 However,	 it	 must	 be	 noted	 that	 this	 rebuttal	 addresses	 an
ethical	 concern	 not	 related	 to	 informational	 privacy,	 thus	 its	 relevance	 in	 this	
context	is	negligible.		
What	is	classified	as	‘identifiable’	still	remains	rather	controversial.	Data	subjects	
who	 are	 anonymised	 prior	 to	 processing	 are	 not	 considered	 ‘identifiable’	
(European	 Commission,	 2012),	 thus	 anonymised	 data	 is	 not	 considered	 as	 a	
breach	 of	 privacy.	 Furthermore,	 algorithmic	 techniques	 that	 do	 not	 require	
access	to	personal	information	may	alleviate	the	risks	of	privacy	breaches	(Fule	
and	Roddick,	2000).							
	
Conclusion	
This	essay	has	summarised	some	ethical	concerns	encountered	in	the	literature,	
but	 is	 by	 no	 means	 exhaustive.	 More	 ethical	 challenges	 will	 emerge	 as	 the	
application	of	algorithms	becomes	more	extensive.	(Tutt,	2016)	suggests	that	the	
challenges	 regarding	 the	 ethics	 of	 Machine	 Learning	 algorithms	 will	 become	
increasingly	 difficult	 as	 the	 complexity	 of	 the	 algorithms	 increases.	 The	
increasing	utilisation	of	Machine	Learning	algorithms,	or	algorithms	in	general,	
will	 undoubtedly	 have	 an	 impact	 on	 policies	 regarding	 the	 use	 of	 such	
technologies.	These	policies	should	be	designed	in	a	way	that	does	not	stifle	the	
predictive	 power	 of	 the	 algorithms,	 but	 rather,	 algorithms	 should	 be	 designed	
with	ethical	considerations	in	mind.		
Already	existing	ethical	issues	also	demand	further	research	(Mittelstadt	et	al,	
2016).	For	instance,	it	is	still	unclear	whether	all	the	ethical	concerns	discussed	
here	are	applicable	to	all	types	of	data	sets;	are	these	ethical	considerations	only	
pertinent	when	data	is	collected	on	humans,	as	opposed	to	objects	such	as	cars?		
Theoretically,	 ethical	 concerns	 may	 be	 difficult,	 if	 not	 impossible,	 to	 avoid.	
However,	it	may	be	possible	to	minimise	ethical	concerns	by	training	developers	
and	users	to	be	aware	of	the	ethical	issues	arising	from	the	application	of	the	
algorithms	in	the	specified	context.	Furthermore,	establishing	governing	bodies	
or	 a	 framework	 of	 regulatory	 policies	 such	 as	 the	 EU	 General	 Data	 Protection	
Regulation	 (2016/679),	 can	 facilitate	 in	 the	 fair	 regulation	 of	 algorithmic	
applications.
References		
	
• Allen	C,	Wallach	W,	Smit	I.	(2006)	Why	machine	ethics?	IEEE	Intelligent	
Systems.	Vol.	21:	pages	12-17.		
• Ananny	M.	(2016)	Towards	an	ethics	of	algorithms	convening,	observation,	
probability	and	timeliness.	Science,	Technology	and	Human	Values.	Vol.	
41:	pages	93-117.		
• Anderson	SL.	(2008)	Asimov’s	‘Three	Laws	of	Robotics’	and	machine	
metaethics.	AI	and	Society.	Vol.	22:	pages	477-493.		
• Applin	SA,	Fischer	MD.	(2015)	New	technologies	and	mixed-use	
convergence:	How	humans	and	algorithms	are	adapting	to	each	other.	IEEE	
international	symposium	on	technology	and	society,	Dublin:	pages	1-6.		
• Barnet	BA.	(2009)	Idiomedia:	The	rise	of	personalized,	aggregated	content.	
Continuum.	Vol.	23:	pages	93-99.		
• Barocas	S,	Selbst	AD.	(2015)	Big	data’s	disparate	impact.	California	Law	
Review.	Vol.	104.3:	pages	671-732.	
• Boyd	D,	Crawford	K.	(2012)	Critical	Questions	for	Big	Data.	Information,	
Communication	and	Society.	Vol.	15:	pages	662-679.		
• Bozdag	E.	(2013)	Bias	in	algorithmic	filtering	and	personalization.	Ethics	
and	Information	Technology.	Vol.	15:	pages	209-227.		
• Burrell	J.	(2016)	How	the	machine	‘thinks’:	Understanding	opacity	in	
machine	learning	algorithms.	Big	Data	and	Society.	Vol.	3:	pages	1-12.		
• Cardona	B.	(2008)	‘Healthy	ageing’	policies	and	anti-ageing	ideologies	and	
practices:	On	the	exercise	of	responsibility.	Medicine,	Health	Care	and	
Philosophy.	Vol.	11:	pages	475-483.		
• Citron	DK.	(2007)	Technological	Due	Process.	Washington	University	Law	
Review.	Vol.	85:	pages	1249-1313.		
• Cohen	IG,	Amarasingham	R,	Shah	A,	Xie	B,	Lo	B.	(2014)	The	legal	and	
ethical	conerns	that	arise	from	using	complex	predicitive	analytics	in	health	
care.	Health	Affairs.	Vol.	33:	pages	1139-1147.		
• Coll	S.	(2013)	Consumption	as	biopower:	Governing	bodies	with	loyalty	
cards.	Journal	of	Consumer	Culture.	Vol.	13:	pages	201-220.		
• Crnkovic	CD,	Curuklu	B.	(2011)	Robots:	ethical	by	design.	Ethics	and	
Information	Technology.	Vol.	14:	pages	61-71.		
• Danna	A,	Gandy	OH.	(2002)	All	that	glitters	is	not	gold:	Digging	beneath	
the	surface	of	data	mining.	Journal	of	Business	Ethics.	Vol.	40:	pages	373-
386.		
• Datta	A,	Sen	S,	Zick	Y.	(2016)	Algorithmic	transparency	via	quantitative	
input	influence.	Proceedings	of	37th	IEEE	symposium	on	Security	and	
Privacy:	pages	598-617.			
• Diakopoulos	N.	(2015)	Algorithmic	accountability:	Journalistic	
investigation	of	computation	power	structures.	Digital	Journalism.	Vol.	3:	
pages	398-415.		
• European	Commission.	(2012)	Regulation	of	the	European	Parliament	and	
of	the	Council	on	the	Protection	of	Individuals	with	regard	to	the	processing	
of	personal	data	and	on	the	free	movement	of	such	data	(General	Data	
Protection	Regulation).	Brussels:	European	Commission:	pages	1-119.
• Friedman	B,	Niseenbaum	H.	(1996)	Bias	in	computer	systems.	ACM	
transactions	on	Information	Systems.	Vol.	14:	pages	330-347.		
• Fule	P,	Roddick	JF.	(2004)	Detecting	privacy	and	ethical	sensitivity	in	data	
mining	results.	Proceedings	of	the	27th	Australian	Conference	on	
Computer	Science.	Vol.	26:	pages	159-166.	
• Glenn	T,	Monteith	S.	(2014)	New	measures	of	mental	state	and	behavior	
based	on	data	collected	from	sensors,	smart-phones,	and	the	internet.	
Current	Psychiatry	Reports.	Vol.	16:	pages	1-10.		
• Golgowski	N.	(2012)	How	Target	knows	when	its	shoppers	are	pregnant	–	
and	figured	out	a	teen	was	pregnant	before	her	father	did.	Daily	Mail,	18	
February	2012.	Link:	http://www.dailymail.co.uk/news/article-
2102859/How-Target-knows-shoppers-pregnant--figured-teen-father-
did.html.	
• Granka	LA.	(2010)	The	politics	of	search:	A	decade	retrospective.	The	
Information	Society.	Vol.	26:	pages	364-374.	
• Hildebrandt	M.	(2011)	Who	needs	stories	if	you	can	get	the	data?	ISPs	in	
the	era	of	big	number	crunching.	Philosophy	and	Technology.	Vol.	24:	
pages	371-390.		
• Hill	RK.	(2015)	What	an	algorithm	is.	Philosophy	and	Technology.	Vol.	29:	
pages	35-59.		
• Illari	PM,	Russo	F.	(2014)	Causality:	Philosophical	Theory	Meets	Scientific	
Practice.	Oxford:	Oxford	University	Press.		
• Introna	LD,	Nissenbaum	H.	(2000)	Shaping	the	Web:	Why	the	politics	of	
search	engines	matters.	The	Information	Society.	Vol.	16:	pages	169-185.		
• Johnson	JA.	(2013)	Ethics	of	data	mining	and	predictive	analytics	in	higher	
education.	Association	for	Institutional	Research	Annual	Forum.	
• Kamiran	F,	Calders	T.	(2010)	Classification	with	no	discrimination	by	
preferential	sampling.	Proceedings	of	the	19th	Machine	Learning	
Conference.	Belgium	and	the	Netherlands:	pages	1-6.		
• Kitchin	R.	(2016)	Thinking	critically	about	and	researching	algorithms.	
Information,	Communication	and	Society.	Vol.	20:	pages	14-29.		
• Kraemer	F,	Van	Overveld	K,	Peterson	M.	(2011)	Is	there	an	ethics	of	
algorithms?	Ethics	and	Information	Technology.	Vol.	13:	pages	251-260.		
Last	Accessed:	19th	April	2017.		
• Lazer	D,	Kennedy	R,	King	G,	Vespignani	A.	(2014)	The	parable	of	Google	
flu:	Traps	in	big	data	analysis.	Science.	Vol.	343:	pages	1203-1205.		
• Leese	M.	(2014)	The	new	profiling:	Algorithms,	black	boxes,	and	the	failure	
of	anti-discriminatory	safeguards	in	the	European	Union.	Security	
Dialogue.	Vol.	45:	pages	494-511.	
• Lewis	SC,	Westlund	O.	(2015)	Big	data	and	journalism.	Digital	Journalism.	
Vol.	3:	pages	447-466.		
• Macnish	K.	(2012)	Unblinking	eyes:	The	ethics	of	automating	surveillance.	
Ethics	and	Information	Technology.	Vol.	14:	pages	151-167.		
• Matthias	A.	(2004)	The	responsibility	gap:	Ascribing	responsibility	for	the	
actions	of	learning	automata.	Ethics	and	Information	Technology.	Vol.	6:	
pages	175-183.		
• Mazoue	JG.	(1990)	Diagnosis	without	doctors.	Journal	of	Medicine	and	
Philosophy.	Vol.	15:	559-579.
• Mittelstadt	BD,	Allo	P,	Taddeo	M,	Wachter	S,	Floridi	L.	(2016)	The	ethics	of	
algorithms:	Mapping	the	debate.	Big	Data	and	Society.	Vol.	3:	pages	1-21.		
• Naik	G,	Bhide	SS.	(2014)	Will	the	future	of	knowledge	work	automation	
transform	personalised	medicine?	Applied	and	Translation	Genomics,	
Inaugural	Issue.	Vol.	3:	pages	50-53.		
• Newell	S,	Marabelli	M.	(2015)	Strategic	opportunities	(and	challenges)	of	
algorithmic	decision-making:	A	call	for	action	on	the	long-term	societal	
effects	of	‘datafication’.	The	Journal	of	Strategic	Information	Systems:	Vol.	
24:	pages	3-14.	
• Portmess	L,	Tower	S.	(2014)	Data	barns,	ambient	intelligence	and	cloud	
computing:	The	tacit	epistemology	and	linguistic	representation	of	Big	
Data.	Ethics	and	Information	Technology.	Vol.	17:	pages	1-9.		
• Raymond	A.	(2014)	The	dilemma	of	private	justice	systems:	Big	Data	
sources,	the	cloud	and	predictive	analytics.	Northwestern	Journal	of	
International	Law	and	Business,	Forthcoming:	pages	1-44.		
• Romei	A,	Ruggieri	S.	(2014)	A	multidisciplinary	survey	on	discrimination	
analysis.	The	Knowledge	Engineering	Review.	Vol.	29:	pages	582-638.		
• Rubel	A,	Jones	KML.	(2014)	Student	privacy	in	learning	analytics:	An	
information	ethics	perspective.	The	Information	Society.	Vol.	32:	143-159.		
• Sandvig	C,	Hamilton	K,	Karahalio	K,	Langbort	C.	(2014)	Auditing	
algorithms:	research	methods	for	detecting	discrimination	on	internet	
platforms.	Data	and	Discrimination:	Converting	Critical	Concerns	into	
Productive	Inquiry:	pages	1-23.	
• Schermer	BW.	(2011)	The	limits	of	privacy	in	automated	profiling	and	data	
mining.	Computer	Law	and	Security	Review.	Vol.	27:	pages	45-52.		
• Sweeney	L.	(2013)	Discrimination	in	online	ad	delivery.	Queue.	Vol.	11:	
10:10-	10:29.		
• Swiatek	M.	(2012)	Intending	to	err:	the	ethical	challenge	of	lethal,	
autonomous	systems.	Ethics	and	Information	Technology.	Vol.	14:	pages	
241-254.		
• Tutt	A.	(2016)	An	FDA	for	Algorithms.	Administrative	Law	Review.	Vol.	67:	
pages	1-26.		
• Van	Den	Hoven	J,	Rooksby	E.	(2008)	Distributive	justice	and	the	value	of	
information:	A	(broadly)	Rawlsian	approach.	Information	Technology	and	
Moral	Philosophy:	pages	376-396.	Cambridge	University	Press.		
• Van	Otterlo	M.	(2013)	A	machine	learning	view	of	profiling.	Privacy,	Due	
process	and	the	Computational-Turn	Philosophers	of	Law	meet	
Philosophers	of	Technology:	pages	41-64.		
• Vries	K.	(2010)	Identity,	profiling	algorithms	and	a	world	of	ambient	
intelligence.	Ethics	and	Information	Technology.	Vol.	12:	pages	71-85.		
• Wiegel	V,	Van	Den	Berg	J.	(2009)	Combining	moral	theory,	modal	logic	and	
mas	to	create	well-behaving	artificial	agents.	International	Journal	of	
Social	Robotics.	Vol.1:	pages	233-242.		
• Wiener	N.	(1988)	The	Human	use	of	Human	beings:	Cybernetics	and	
Society.	Da	Capo	Press.		
• Wiltshire	TJ.	(2015)	A	prospective	framework	for	the	design	of	ideal	
artificial	moral	agents:	Insights	from	the	science	of	heroism	in	humans.	
Minds	and	Machines.	Vol.	25:	pages	57-71.
• Zarsky	T.	(2013)	Transparent	Predictions.	University	of	Illinois	Law	
Review.	Vol.	2013:	pages	1503-1570.

The Ethics of Machine Learning Algorithms

  • 1.
    The Ethics of Machine Learning Algorithms Introduction The advent of the ‘Information Age’ gave birth to novel technologies capable of dealing with large data sets and capable of inferring useful information from large data sets. One such technology is Machine Learning. Machine Learning is ‘‘any methodology and set of techniques that can employ data to come up with novel patterns and knowledge, and generate models that can be used for effective predictions about the data’’ (Van Otterlo, 2013). Machine Learning has been used extensively to draw significant insights from unwieldy data sets, and have been proposed to be the “future” of algorithms and analytics (Tutt, 2016). More and more responsibilities are being delegated to machine learning algorithms (Mittelstadt et al, 2016). For instance, Machine Learning algorithms are employed to recognise human bias or inaccurate information, as displayed by Wikipedia’s Objective Revision Evaluation Service, to act as recommendation systems for customers advising them on their next purchase based on their purchase history (Vries, 2010) and to facilitate in the understanding of behavioural data emerging from the ‘Internet of Things’ (Portmess and Tower, 2014). This list is by no means exhaustive. Given their wide applications, it is plausible to assume that Machine Learning algorithms are an integral set of technologies to industrial and academic settings. (Citron, 2007) argues that developers of algorithms have a tendency to accept that automated processes, as those encountered during the application of Machine Learning algorithms, are true and precise, overcoming the hurdle of human bias (Naik and Bhide, 2014). As controversial as this statement is, it does highlight the dependency of Machine Learning algorithms in industrial and academic settings. However, these technologies come at an ethical cost. Given that algorithms are unavoidably influenced by personal opinion, since the parameters are finely tuned by the developer (Wiener, 1988), then it comes as no surprise that designing and developing algorithms does not guarantee that ethical considerations will not be violated. For example, profiling algorithms have to discriminate to achieve their purpose, but sometimes concomitantly are derogatory (Barocas and Selbst, 2015).
  • 2.
    The publically familiar definition of algorithms is broad. An algorithm can essentially be any set of instructions provided to achieve a desired outcome. For the purpose of this essay, this definition is too broad and should be restricted appropriately. Therefore, this essay will adopt Hill’s definition of an algorithm as a “mathematical construct with a finite, abstract, effective, compound control structure, imperatively given, accomplishing a given purpose under given provisions” (Hill, 2015). This definition takes into account the implementation of an algorithm into a technology, and the application of that technology aimed at fulfilling a predefined purpose. However, the discussion regarding ethics implicated in the application of technologies based on Machine Learning algorithms will be limited. Moreover, this essay will only focus on algorithms whose decision-making is difficult to explain from a human perspective due to the algorithm’s opaque nature (how a new input will be handled), and algorithms whose actions are difficult to predict. Algorithms implemented for mundane tasks, such as manufacturing, are not considered here. The essay will discuss several broad ethical concerns regarding the use of algorithms, each placed under a distinct subcategory. Algorithmic Bias Algorithms inescapably contain some element of bias, although some would argue that their deployment is actually an approach to overcome human bias (Naik and Bhide, 2014). One example of bias, classed as ‘technical bias’, is observed when the listing of companies for a search result follows alphabetical order, potentially resulting in more business for companies that appear earlier in the alphabet (Friedman and Nissenbaum, 1996). The reason why algorithms can be bias is because the values of the developer are institutionalised into the algorithm, and because algorithms are designed with an intended purpose in mind (Macnish, 2012). During the course of the algorithm’s development, there is not always an objectively correct choice, but rather there is a myriad of alternatives that could still achieve the same desired objective. The bias can arise from many situations (Friedman and Nissenbaum, 1996). For example, bias can arise during supervised learning tasks. During supervised learning, the Machine Learning algorithm learns from human-tagged data inputs.
  • 3.
    The labelling of these inputs can inadvertently be bias (Diakopoulos, 2015). Furthermore, the interpretation of the outputs of the algorithm can also be bias (Hildebrandt, 2011). Not many improvements can be made for bias emerging from technical constraints (i.e. technical bias). However, a collective effort could reduce bias emerging from other sources, such as social bias that become manifested in the algorithms. Raymond (2014) proposes that human monitoring may be an effective approach to reduce some bias. Unfair Outcomes Based on Discrimination One broad category of ethical considerations is discrimination resulting from profiling (Vries, 2010). Classification algorithms are often employed in industry and academia for various purposes. These classification algorithms classify new inputs based on the characteristics of the training data. The description of classification algorithms already hints at some ethical issues that can be encountered during the employment of these algorithms. Practical examples of discrimination include the delivery of online advertisements based on perceived ethnicity (Sweeny, 2013), and personalised pricing (Danna and Gandy, 2002). Another example is an infamous ‘Data Science gone wrong’ story pertinent to classification algorithms, which is of Target’s unfortunate labelling of a 13-year- old girl as pregnant, before her father was even aware of her pregnancy (Golgowksi, 2012). It is important to note though, that the last example has ethical concerns reaching beyond discrimination. The ethics of classification algorithms are even more extensive in the context of classification algorithms recruited for medical diagnoses (Cohen et al, 2014). Such models aid in the diagnosis and also recommend viable treatment plans to physicians (Mazoue, 1990). Schermer (2011) argues that the act of discriminatory treatment itself is not ethically challenging, but that the resulting effects from discriminatory treatment are ethically challenging. However, (Mittelstadt et al, 2016) suggests that Schermer is ‘muddling’ up the concepts of bias and discrimination, incorrectly using discrimination to describe bias.
  • 4.
    There are multiple diverse reasons to consider discriminatory effects as adverse, implying that they should be prevented if possible, or at least minimised to almost extinction. It has been postulated that discrimination can contribute to self-fulfilling prophecies and that it can exaggerate stigmatisation (Macnish, 2012). Furthermore, discrimination may reinforce advantages and disadvantages already existing in society, exacerbating social discrepancy, which violates ethical and legal principles of fair treatment (Newell and Marabelli, 2015). Fortunately, it may be possible to develop algorithms that do not take into account sensitive attributes or features, which could potentially contribute to discriminatory outcomes (Barocas and Selbst, 2015). For instance, (Kamiran and Calders, 2010) have developed algorithms that do not take into account gender or ethnicity, without a loss in performance. However, this is strongly subject to context; some classification tasks may require the inclusion of sensitive attributes in order to perform to an adequate level. Thus, the question then becomes, what classification tasks can be delegated to classification algorithms in the first place? Moreover, using proxies for sensitive attributes may not be an effective solution (Romei and Ruggieri, 2014). Inappropriate Actions Based on Inconclusive Data Correlations are commonly used to complement inductive knowledge in decision-making tasks. Correlations based on a large volume of data are extrapolated to direct actions, without the requirement to establish a causal relationship first (Hildebrandt, 2011). This can result in unjustified actions based on inconclusive evidence. Mittelstadt et al (2016) coin this as ‘actionable insights’. To exacerbate this issue of actionable insights, (Lazer et al, 2014) argue that correlations identified in large proprietary data sets are seldom reproducible due to their commercial sensitivity, and thus often present the opportunity of being fabricated. Ananny (2016) postulates that correlations are doubly uncertain; not only can they result in actionable insights, but acting upon on correlations concerns individuals, while the knowledge concerns the population from which the conclusion was drawn from (Illari and Russo, 2014).
  • 5.
    Furthermore, the sheer volume of ‘Big Data’ does not imply that fundamental theoretical aspects of measurement, reliability and face validity can be neglected, and that solely large quantities of data do not serve as an appropriate replacement for improper methodology (Boyd and Crawford, 2012). This concern is not specific to correlations either, as it is applicable to all domains where ‘Big Data’ is employed to draw conclusions. One possible solution to this problem is to train data analysts not to draw hasty conclusions, and to ensure that data quality is of the highest form. Whether actions based on conclusive evidence are deemed ethically acceptable is not a discussion that is pertinent to the discussion of actionable insights. Accountability Another major ethical concern of Machine Learning algorithms, as suggested by (Mittelstadt et al, 2016), is traceability. The authors coined this term to describe the difficulty of assigning responsibility or culpability to an agent when ethical concerns, such as any form of harm tout court, are violated during the application of Machine Learning algorithms. Firstly, it is important to note that identifying whether some Machine Learning algorithms suffer from some ethical concerns due to human subjectivity is difficult to decipher. Typically, it is the designers and users of the algorithm that are held accountable if adversity is encountered (Kraemer et al, 2011). However, (Matthias, 2004) argues that blame can only be equipped when the designer or user has some element of control over the algorithm, and is also ill intentioned. Often, the developer or user does not have as much control over the algorithm as conventionally believed, creating an ‘accountability gap’, where it is not clear to determine who is culpable for the violation of ethical considerations (Cardona, 2008). In essence, shifting blame upon the developer may only be applicable to cases where the operational rules of the algorithm are hand-written, and not rules that the algorithm learns independently (i.e. Machine Learning algorithms). The opposing view suggests that algorithms, which act independently and result in ethically unjust outcomes, should be held accountable (Wiltshire, 2015). However, this may provide an opportunity for developers to assign blame to the algorithm (Crnkovic and Curuklu, 2011). Furthermore, the capacity of algorithms
  • 6.
    to make moral and ethically justifiable decisions is still disputable (Anderson, 2008). Nonetheless, it can be argued that developers and users should monitor the workings of the algorithm for any ethical violations, as done so by (Fole and Ruddick, 2004). Allen et al. (2006) insinuate that the void in ‘machine ethics’ needs to be filled, so that ethical violations can be avoided. There is much research currently focused on developing algorithms with a concrete understanding of moral and ethical responsibilities (Wiegel and Berg, 2009), so that neither extreme of assigning blame (solely blaming the developer or the algorithm) has to be adopted. Opacity of Algorithms The opacity of Machine Learning algorithms is another major concern in the Literature. Supposedly, algorithms that constitute mechanisms that are difficult to explain, or algorithms whose behaviour is difficult to predict, are also concomitantly difficult to control, regulate and correct (Tutt, 2016). Introna and Nissenbaum (2000) is among pioneering publications discussing this topic of transparency. The authors suggested that search engines filter information based on market conditions, with a slight preference towards powerful actors. In the paper, the authors discuss corrective mechanisms intended to improve the ‘fairness’ of search engines. Anyhow, the accessibility of information elucidating the functionality of algorithms is often restricted because of national security (Leese, 2014), competitive advantage (Glenn and Monteith, 2014), or privacy in general. Thus, transparency is not always desired, especially not at the expense of invasion of privacy. On the contrary, it can be argued that the privacy of an algorithm is not as important, except for cases concerning competitive advantage, and that only the privacy of the data itself should be regarded as important. The argument of exposing the algorithm’s architecture can also be swayed in the opposite direction, that is, informing the public about the architecture of algorithms may prompt some users to abusively manipulate the algorithm for tailored use, while also placing non-tech-savvy users at a disadvantage (Granka, 2010).
  • 7.
    Intelligibility also contributes to the opacity of algorithms (Burrell, 2016). Machine Learning algorithms are commonly described as ‘black boxes’, which are fed input to produce an output, with the mechanisms of the algorithm being obscured. For many applications of Machine Learning algorithms, the developer cannot articulate the exact procedure needed to achieve a particular goal, hence the requirement for these algorithms in the first place (Datta et al, 2016). Even algorithms whose operation can be decomposed into a set of hand-written rules have the capacity to be inscrutable (Kitchin, 2016). To complicate matters even more, algorithms are often designed by groups of engineers, rendering a comprehensive understanding of the algorithms extremely difficult, if not practically infeasible (Sandvig et al, 2014). It is also important to note that the intelligibility of algorithms can be limited by the questioner’s capacity to understand it. The problem with this is that full legitimate consent cannot be granted if the mechanisms of the algorithm are not fully understood (Schermer, 2011). To make the algorithms more accessible, the operational logic of the algorithm could be explained in simplified terms. However, (Kitchin, 2016) suggests that this may only be effective for simple algorithms, and may not be a plausible solution for more complex algorithms, which could then still potentially escalate to a lack of trust for both the algorithm, and the data analyst (Rubel and Jones, 2014). It is important to recognise that trust can be placed in the algorithm independent of the data analyst, but whether this is appropriate is still controversial, as in the case of autonomous weapons (Swiatek, 2012). (Zarsky, 2013) suggests that transparency should be disclosed to trained third parties, seen as regulators, acting in the interest of the public, as opposed to the questioners themselves to prevent any loss of trust. This sense of trust is pivotal (Cohen et al, 2014) and can alleviate concerns with the opacity of algorithms based on personal information (Mazoue, 1990). Transformative Effects The results obtained from Machine Learning algorithms can change the perceptions of how humans perceive the world. These transformative effects can lead to ethically challenging concepts.
  • 8.
    One example of such ethical concern is observed during the employment of personalisation algorithms. These algorithms obscure content encountered by the user that is believed to be unaligned to the user’s belief, or be contradictory to the user’s belief (Barnet, 2009). Ananny (2016) suggests that algorithms need not be explicitly coercing to influence perception and decision-making through filtering. Personalisation algorithms can be seen as supporting the user by filtering out relevant information from the masses of information presented, or they can be seen as controlling by presenting the user with only information the personalisation algorithms deems relevant (Bozdag, 2013). This could result in ethical violations when the choice of what data is presented prioritises the interests of a third-party over the interests of the user (Applin and Fischer, 2015), as commonly demonstrated with online consumers (Coll, 2013). The users may then be prompted to make decisions not purely based on their own preference, but rather, choices will be based on inherently subjective restrictions of information load (Johnson, 2013). The diversity of information presented is paramount in exercising autonomy (Van Den Hoven and Rooksby, 2008). Thus, any inappropriate restriction of information diversity can disrupt the autonomy of the user. Furthermore, a lack of information diversity may handicap the exploration of new ideas and options (Raymond, 2014). To counteract this ethical concern, (Lewis and Westlund, 2015) propose that personalisation algorithms should be taught to be aware of this ethical consideration, of not being too persuasive at the expense of limiting the decisional autonomy of the user. However, it’s not only personalisation algorithms that can have a transformative impact. The general utilisation of algorithms is also transforming the conventional notions of informational privacy (Mittelstadt et al, 2016). For instance, (Schermer, 2011) confronts identifiability in profiling algorithms by arguing that profiling algorithms aim to designate data inputs into meaningful assemblies, for which individual identity is truly trivial. Yet, (Van Wel and Royakkers, 2004) rebut by suggesting that profiling is a form of de- personalisation by assigning judgements based on group characteristics, and not individual merit. However, it must be noted that this rebuttal addresses an
  • 9.
    ethical concern not related to informational privacy, thus its relevance in this context is negligible. What is classified as ‘identifiable’ still remains rather controversial. Data subjects who are anonymised prior to processing are not considered ‘identifiable’ (European Commission, 2012), thus anonymised data is not considered as a breach of privacy. Furthermore, algorithmic techniques that do not require access to personal information may alleviate the risks of privacy breaches (Fule and Roddick, 2000). Conclusion This essay has summarised some ethical concerns encountered in the literature, but is by no means exhaustive. More ethical challenges will emerge as the application of algorithms becomes more extensive. (Tutt, 2016) suggests that the challenges regarding the ethics of Machine Learning algorithms will become increasingly difficult as the complexity of the algorithms increases. The increasing utilisation of Machine Learning algorithms, or algorithms in general, will undoubtedly have an impact on policies regarding the use of such technologies. These policies should be designed in a way that does not stifle the predictive power of the algorithms, but rather, algorithms should be designed with ethical considerations in mind. Already existing ethical issues also demand further research (Mittelstadt et al, 2016). For instance, it is still unclear whether all the ethical concerns discussed here are applicable to all types of data sets; are these ethical considerations only pertinent when data is collected on humans, as opposed to objects such as cars? Theoretically, ethical concerns may be difficult, if not impossible, to avoid. However, it may be possible to minimise ethical concerns by training developers and users to be aware of the ethical issues arising from the application of the algorithms in the specified context. Furthermore, establishing governing bodies or a framework of regulatory policies such as the EU General Data Protection Regulation (2016/679), can facilitate in the fair regulation of algorithmic applications.
  • 10.
    References • Allen C, Wallach W, Smit I. (2006) Why machine ethics? IEEE Intelligent Systems. Vol. 21: pages 12-17. • Ananny M. (2016) Towards an ethics of algorithms convening, observation, probability and timeliness. Science, Technology and Human Values. Vol. 41: pages 93-117. •Anderson SL. (2008) Asimov’s ‘Three Laws of Robotics’ and machine metaethics. AI and Society. Vol. 22: pages 477-493. • Applin SA, Fischer MD. (2015) New technologies and mixed-use convergence: How humans and algorithms are adapting to each other. IEEE international symposium on technology and society, Dublin: pages 1-6. • Barnet BA. (2009) Idiomedia: The rise of personalized, aggregated content. Continuum. Vol. 23: pages 93-99. • Barocas S, Selbst AD. (2015) Big data’s disparate impact. California Law Review. Vol. 104.3: pages 671-732. • Boyd D, Crawford K. (2012) Critical Questions for Big Data. Information, Communication and Society. Vol. 15: pages 662-679. • Bozdag E. (2013) Bias in algorithmic filtering and personalization. Ethics and Information Technology. Vol. 15: pages 209-227. • Burrell J. (2016) How the machine ‘thinks’: Understanding opacity in machine learning algorithms. Big Data and Society. Vol. 3: pages 1-12. • Cardona B. (2008) ‘Healthy ageing’ policies and anti-ageing ideologies and practices: On the exercise of responsibility. Medicine, Health Care and Philosophy. Vol. 11: pages 475-483. • Citron DK. (2007) Technological Due Process. Washington University Law Review. Vol. 85: pages 1249-1313. • Cohen IG, Amarasingham R, Shah A, Xie B, Lo B. (2014) The legal and ethical conerns that arise from using complex predicitive analytics in health care. Health Affairs. Vol. 33: pages 1139-1147. • Coll S. (2013) Consumption as biopower: Governing bodies with loyalty cards. Journal of Consumer Culture. Vol. 13: pages 201-220. • Crnkovic CD, Curuklu B. (2011) Robots: ethical by design. Ethics and Information Technology. Vol. 14: pages 61-71. • Danna A, Gandy OH. (2002) All that glitters is not gold: Digging beneath the surface of data mining. Journal of Business Ethics. Vol. 40: pages 373- 386. • Datta A, Sen S, Zick Y. (2016) Algorithmic transparency via quantitative input influence. Proceedings of 37th IEEE symposium on Security and Privacy: pages 598-617. • Diakopoulos N. (2015) Algorithmic accountability: Journalistic investigation of computation power structures. Digital Journalism. Vol. 3: pages 398-415. • European Commission. (2012) Regulation of the European Parliament and of the Council on the Protection of Individuals with regard to the processing of personal data and on the free movement of such data (General Data Protection Regulation). Brussels: European Commission: pages 1-119.
  • 11.
    • Friedman B, Niseenbaum H. (1996) Bias in computer systems. ACM transactions on Information Systems. Vol. 14: pages 330-347. • Fule P, Roddick JF. (2004) Detecting privacy and ethical sensitivity in data mining results. Proceedings of the 27th Australian Conference on Computer Science. Vol. 26: pages 159-166. •Glenn T, Monteith S. (2014) New measures of mental state and behavior based on data collected from sensors, smart-phones, and the internet. Current Psychiatry Reports. Vol. 16: pages 1-10. • Golgowski N. (2012) How Target knows when its shoppers are pregnant – and figured out a teen was pregnant before her father did. Daily Mail, 18 February 2012. Link: http://www.dailymail.co.uk/news/article- 2102859/How-Target-knows-shoppers-pregnant--figured-teen-father- did.html. • Granka LA. (2010) The politics of search: A decade retrospective. The Information Society. Vol. 26: pages 364-374. • Hildebrandt M. (2011) Who needs stories if you can get the data? ISPs in the era of big number crunching. Philosophy and Technology. Vol. 24: pages 371-390. • Hill RK. (2015) What an algorithm is. Philosophy and Technology. Vol. 29: pages 35-59. • Illari PM, Russo F. (2014) Causality: Philosophical Theory Meets Scientific Practice. Oxford: Oxford University Press. • Introna LD, Nissenbaum H. (2000) Shaping the Web: Why the politics of search engines matters. The Information Society. Vol. 16: pages 169-185. • Johnson JA. (2013) Ethics of data mining and predictive analytics in higher education. Association for Institutional Research Annual Forum. • Kamiran F, Calders T. (2010) Classification with no discrimination by preferential sampling. Proceedings of the 19th Machine Learning Conference. Belgium and the Netherlands: pages 1-6. • Kitchin R. (2016) Thinking critically about and researching algorithms. Information, Communication and Society. Vol. 20: pages 14-29. • Kraemer F, Van Overveld K, Peterson M. (2011) Is there an ethics of algorithms? Ethics and Information Technology. Vol. 13: pages 251-260. Last Accessed: 19th April 2017. • Lazer D, Kennedy R, King G, Vespignani A. (2014) The parable of Google flu: Traps in big data analysis. Science. Vol. 343: pages 1203-1205. • Leese M. (2014) The new profiling: Algorithms, black boxes, and the failure of anti-discriminatory safeguards in the European Union. Security Dialogue. Vol. 45: pages 494-511. • Lewis SC, Westlund O. (2015) Big data and journalism. Digital Journalism. Vol. 3: pages 447-466. • Macnish K. (2012) Unblinking eyes: The ethics of automating surveillance. Ethics and Information Technology. Vol. 14: pages 151-167. • Matthias A. (2004) The responsibility gap: Ascribing responsibility for the actions of learning automata. Ethics and Information Technology. Vol. 6: pages 175-183. • Mazoue JG. (1990) Diagnosis without doctors. Journal of Medicine and Philosophy. Vol. 15: 559-579.
  • 12.
    • Mittelstadt BD, Allo P, Taddeo M, Wachter S, Floridi L. (2016) The ethics of algorithms: Mapping the debate. Big Data and Society. Vol. 3: pages 1-21. • Naik G, Bhide SS. (2014) Will the future of knowledge work automation transform personalised medicine? Applied and Translation Genomics, Inaugural Issue. Vol. 3: pages 50-53. •Newell S, Marabelli M. (2015) Strategic opportunities (and challenges) of algorithmic decision-making: A call for action on the long-term societal effects of ‘datafication’. The Journal of Strategic Information Systems: Vol. 24: pages 3-14. • Portmess L, Tower S. (2014) Data barns, ambient intelligence and cloud computing: The tacit epistemology and linguistic representation of Big Data. Ethics and Information Technology. Vol. 17: pages 1-9. • Raymond A. (2014) The dilemma of private justice systems: Big Data sources, the cloud and predictive analytics. Northwestern Journal of International Law and Business, Forthcoming: pages 1-44. • Romei A, Ruggieri S. (2014) A multidisciplinary survey on discrimination analysis. The Knowledge Engineering Review. Vol. 29: pages 582-638. • Rubel A, Jones KML. (2014) Student privacy in learning analytics: An information ethics perspective. The Information Society. Vol. 32: 143-159. • Sandvig C, Hamilton K, Karahalio K, Langbort C. (2014) Auditing algorithms: research methods for detecting discrimination on internet platforms. Data and Discrimination: Converting Critical Concerns into Productive Inquiry: pages 1-23. • Schermer BW. (2011) The limits of privacy in automated profiling and data mining. Computer Law and Security Review. Vol. 27: pages 45-52. • Sweeney L. (2013) Discrimination in online ad delivery. Queue. Vol. 11: 10:10- 10:29. • Swiatek M. (2012) Intending to err: the ethical challenge of lethal, autonomous systems. Ethics and Information Technology. Vol. 14: pages 241-254. • Tutt A. (2016) An FDA for Algorithms. Administrative Law Review. Vol. 67: pages 1-26. • Van Den Hoven J, Rooksby E. (2008) Distributive justice and the value of information: A (broadly) Rawlsian approach. Information Technology and Moral Philosophy: pages 376-396. Cambridge University Press. • Van Otterlo M. (2013) A machine learning view of profiling. Privacy, Due process and the Computational-Turn Philosophers of Law meet Philosophers of Technology: pages 41-64. • Vries K. (2010) Identity, profiling algorithms and a world of ambient intelligence. Ethics and Information Technology. Vol. 12: pages 71-85. • Wiegel V, Van Den Berg J. (2009) Combining moral theory, modal logic and mas to create well-behaving artificial agents. International Journal of Social Robotics. Vol.1: pages 233-242. • Wiener N. (1988) The Human use of Human beings: Cybernetics and Society. Da Capo Press. • Wiltshire TJ. (2015) A prospective framework for the design of ideal artificial moral agents: Insights from the science of heroism in humans. Minds and Machines. Vol. 25: pages 57-71.
  • 13.