SlideShare a Scribd company logo
1 of 100
SHEILA Project Results
Dragan Gašević
@dgasevic
SHEILA Conference
5th June 2018
Brussels, Belgium
http://sheilaproject.eu/
Inclusive adoption process
Inclusive adoption process
http://sheilaproject.eu/
Inclusive adoption process
Macfadyen, L., Dawson, S., Pardo, A., Gašević, D., (2014). The learning analytics imperative and the sociotechnical challenge: Policy for complex systems. Research
& Practice in Assessment, 9(Winter 2014), 17-28.
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Adoption challenge
Leadership for strategic
implementation & monitoring
Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the
Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
Adoption challenge
Equal engagement with
different stakeholders
Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the
Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
Adoption challenge
Training to cultivate data literacy
among primary stakeholders
Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the
Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
Adoption challenge
Policies for learning analytics practice
Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the
Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Institutional survey &
interviews
Yi-Shan Tsai
University of Edinburgh
yi-shan.tsai@ed.ac.uk
@yi_shan_tsai
What is the state of the art?
What are the drivers?
What are the challenges?
Survey
• 22 countries, 46 institutions
• November 2016
NO P LA NS
IN P RE P A RA TION
IMP LE ME NT ED 2 13
15
16
The adoption of LA
Institution-wide Small scale N/A
Interviews
• 16 countries, 51 HEIs, 64 interviews, 78 participants
• August 2016 - January 2017
N O P L A N S
I N P R E P A R A T I O N
I M P L E M E N T E D
9 7 5
12
18
The adoption of learning analytics (interviews)
Institution-wide Partial/ Pilots Data exploration/cleaning
Motivations to adopt learning analytics
• To improve student learning performance – 40 (87%)
• To improve student satisfaction – 33 (72%)
• To improve teaching excellence – 33 (72 %)
• To improve student retention– 26 (57 %)
• To explore what learning analytics can do for our
institution/ staff/ students – 25 (54 %)
46 institutions
Motivations to adopt learning analytics
• To improve student learning performance – 40 (87%)
• To improve student satisfaction – 33 (72%)
• To improve teaching excellence – 33 (72 %)
• To improve student retention– 26 (57 %)
• To explore what learning analytics can do for our
institution/ staff/ students – 25 (54 %)
46 institutions
Motivations to adopt learning analytics
• To improve student learning performance – 40 (87%)
• To improve student satisfaction – 33 (72%)
• To improve teaching excellence – 33 (72 %)
• To improve student retention– 26 (57 %)
• To explore what learning analytics can do for our
institution/ staff/ students – 25 (54 %)
46 institutions
Why learning analytics?
LA
Learner
driver
Teaching
driver
Institutional
driver
Self-regulation
Learning support
Performance
Why learning analytics?
LA
Learner
driver
Teaching
driver
Institutional
driver
Self-regulation
Learning support
Performance
“People are thinking about learning analytics as a way
to try and personalise education and enhance
education. And actually make our education more
inclusive both by understanding how different students
engage with different bits of educational processes, but
also about through developing curricula to make them
more flexible and inclusive as a standard.”
“I think what we would be looking at is how do we
evolve the way we teach to provide better learning
outcomes for the students, greater mastery of the
subject.”
“We’re trying to understand better the curriculum that
needs to be offered for the students in our region.
And…I think importantly how our pedagogical model
fits that and deliver the best experience for our
students.”
Barriers to the success of learning analytics
• Analytics expertise – 34 (76%)
• A data-driven culture at the institution – 30 (67%)
• Teaching staff/tutor buy-in – 29 (64%)
• The affordances of current learning analytics technology – 29 (64%)
Ethical and privacy concerns
access transparency anonymity
Analytical capability challenge
• More than half of the institutions with 0-3 years experience indicated
noticeable gaps among different stakeholders regarding perceptions
and understanding of learning analytics.
Success
• We have achieved the goals that we set for learning analytics.
Success
• Most institutions did not have confirmed success.
• Half of the institutions with 0-3 years experience claimed “gaining
experience” as part of their success.
Implications
• Interests were high but experiences were premature.
• There was strong motivation in increasing institutional performance
by improving teaching quality.
• Key barriers were around skills, institutional culture, technology,
ethics and privacy.
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Student Survey Results
Alex Wainwright
University of Liverpool
a.wainwright@Liverpool.ac.uk
http://sheilaproject.eu/
Background
• 12 Items Survey
• Two Subscales:
• Ethical and Privacy Expectations
• Service Expectations
• 6 Distributions:
• Edinburgh (N = 884)
• Liverpool (N = 191)
• Tallinn (N = 161)
• Madrid (N = 543)
• Netherlands (N = 1247)
• Blanchardstown (N = 237)
http://sheilaproject.eu/
Ideal Expectation Scale Predicted Expectation Scale
Alternative Purpose Consent to Collect Identifiable Data Keep Data Secure Third Party Alternative Purpose Consent to Collect Identifiable Data Keep Data Secure Third Party
1
2
3
4
5
6
7
Item
Average
Location
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Ethical and Privacy Expectations http://sheilaproject.eu/
Keep Data Secure – Predicted Expectation Scale
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree
Response
10
20
30
40
50
Percentage
http://sheilaproject.eu/
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree
Response
10
20
30
Percentage
Consent to Collect – Predicted Expectation Scale http://sheilaproject.eu/
Ideal Expectation Scale Predicted Expectation Scale
ObligationtoAct
IntegrateintoFeedback
SkillDevelopment
RegularlyUpdate
CompleteProfile
StudentDecisionMaking
CourseGoals
ObligationtoAct
IntegrateintoFeedback
SkillDevelopment
RegularlyUpdate
CompleteProfile
StudentDecisionMaking
CourseGoals
1
2
3
4
5
6
7
Average
Location
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Service Expectations http://sheilaproject.eu/
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree
Response
10
20
30
Percentage
Course Goals – Predicted Expectation Scale http://sheilaproject.eu/
Blanchardstown
Edinburgh
Liverpool
Madrid
Open University of the Netherlands
Tallinn
Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree
Response
10
20
Percentage
Obligation to Act – Predicted Expectation Scale http://sheilaproject.eu/
Summary
• Beliefs towards learning analytics are not consistent.
• Emphasis on data security and improving learning.
http://sheilaproject.eu/
Student focus groups
Pedro Manuel Moreno Marcos
Department of Telematic Engineering
Universidad Carlos III de Madrid
http://sheilaproject.eu/
Goals
• Interest and expectations
• Awareness
• Concerns
Background
• 18 focus groups
• 4 partners’ institutions
• 74 students
• Interviews: Around 1h
http://sheilaproject.eu/
Interests and expectations
• Improve the quality of teaching
• Better student-teacher feedback
• Better academic resources and academic tools to improve learning
• Personalized support
• Recommendation of learning resources
• Feedback from a system, via a dashboard
• Provide an overview of the tasks to be done in a semester → improve
curriculum design
http://sheilaproject.eu/
Awareness
• Students do not know what LA is, but they recognise its importance if
it can solve students’ problems
• Students are not generally aware of the data collected → Transparency
• Students have not checked the conditions they have accepted about
data
http://sheilaproject.eu/
Concerns
http://sheilaproject.eu/
Surveillance Anonymization
Purpose of
data
Kind of data
Consent and
access
Security
Provision of
opt-outs
Stereotypes
and biases
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Staff Survey
Maren Scheffel
Open Universiteit Nederland
Welten Institute
@m_a_s_c
With regards to learning analytics …
… what do academic staff ideally expect to happen?
… what do academic staff predict to happen in reality?
Goal of the survey
4 academic institutions
University of Edinburgh Carlos III Madrid
n = 81 n = 26
Open Universiteit University of Tallinn
n = 54 n = 49
from spring to fall 2017
16 items, some examples
The university will provide me with guidance on how to access
LA about my students
The LA service will show how a student’s learning progress
compares to their learning goals/the course objectives
The teaching staff will have an obligation to act if the analytics
show that a student is at-risk of failing, underperforming, or
that they could improve their learning
University of Edinburgh:
• Ideal: LA will collect and present data that is accurate (M = 5.91) Q9
• Predicted: Providing guidance to access LA about students (M = 5.05) Q1
Carlos III de Madrid:
• Ideal: LA presented in a format that is understandable and easy to read (M
= 6.31) Q11
• Predicted: LA will present students with a complete profile of their learning
across every course (M = 5.27) Q12
Highest expectation values
Highest expectation values
Open Universiteit Nederland:
• Ideal: LA will collect and present data that is accurate (M = 6.60) Q9
• Predicted: Able to access data about students’ progress in a course that I
am teaching (M = 5.17) Q4
University of Tallinn:
• Ideal: Able to access data about students’ progress in a course that I am
teaching (M = 6.04) Q4
• Predicted: Able to access data about students’ progress in a course that I
am teaching (M = 5.49) Q4
Lowest expectation values
University of Edinburgh:
• Ideal: Teaching staff will have an obligation to act if students are found to be at-
risk of failing or under performing (M = 3.65) Q14
• Predicted: Teaching staff will be competent in incorporating analytics into the
feedback and support they provide to students (M = 3.49) Q13
Carlos III de Madrid:
• Ideal: Teaching staff will have an obligation to act if students are found to be at-
risk of failing or under performing (M = 4.42) Q14
• Predicted: Teaching staff will have an obligation to act if students are found to
be at-risk of failing or under performing (M = 3.77) Q14
Lowest expectation values
Open Universiteit Nederland:
• Ideal: Teaching staff will have an obligation to act if students are found to be at-
risk of failing or under performing (M = 4.44) Q14
• Predicted: Feedback from analytics will be used to promote students’ academic
and professional skill development for future employability (M = 3.24) Q15
University of Tallinn:
• Ideal: Teaching staff will have an obligation to act if students are found to be at-
risk of failing or under performing (M = 4.80) Q14
• Predicted: Q14 (M = 3.82)
Staff focus groups
Kairit Tammets
Centre for educational technology
School of Digital Technologies
Tallinn University
Goal
To better understand the viewpoints of academic staff on:
• Learning analytics opportunities in the HEIs from the
perspective of students, teachers and programs;
• Concerns related with adapting of learning analytics;
• Needed steps to adopt learning analytics at the HEIs
Study participants
• University of Edinburgh: 5 focus groups, 18 teaching staff
• Universidad Carlos III de Madrid: 4 focus groups, 16
teaching staff
• Open Universiteit Nederland: 2 focus groups, 5 teaching
staff
• Tallinn University: 5 focus groups, 20 teaching staff
Results: Expectations & LA opportunities
STUDENT
LEVEL
TEACHER
LEVEL
PROGRAM
LEVEL
Take responsibility for their
learning and enhancing their
SRL- skills
Assess the degree of success to
prevent students from begin
worried or optimistic about
their performance
Method to identify student’s
weaknesses and know where
students are with their progress
Understand how students
engage with learning content
Improve of the design and
provision of learning materials,
courses, curriculum and support
to students
Understand how program is
working (strengths and
bottlenecks)
Improve educational quality
(e.g. content level)
Results: Meaningful data
Results: Meaningful data
Results: Meaningful data
Results: Meaningful data
Results: Meaningful data
Results: concerns – student level
https://www.pinterest.com/pin/432486370448743887/
Results: Concerns –
student level
https://www.pinterest.com/pin/432486370448743887/
Results: concerns –
student level
https://www.pinterest.com/pin/432486370448743887/
Results: concerns – teacher level
https://www.pinterest.com/pin/432486370448743887/
Results: concerns –
teacher level
http://create-learning.com
https://www.pinterest.com/pin/432486370448743887/
Results: concerns –
teacher level
http://create-learning.com
https://www.pinterest.com/pin/432486370448743887/
Http://memegenerator.net
Results: concerns – program level
• Interpretation of learning:
• Was the right data collected?
• Were the accurate algorithms developed ?
• Was an appropriate message given for the students?
• Connecting LA to real learning – is this meaningful picture of
learning what is happening in online environments?
What we should consider?
• LA should be just one component of many for collecting
feedback and enhancing decision-making
• Involve stakeholders:
• Academic staff to in developing and setting up of LA
• Pedagogy experts involved to ensure data makes sense to
improve learning
• Provide training, communication!
What we should consider?
•Design of the tools that are:
•Easy to use
•Providing visualizations of data
•Not requiring mathematical/statistical skills
•Not taking a lot of time
•Considering ethical and privacy aspects
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Group Concept Mapping
Prof. Dr. Hendrik Drachsler
Open University Netherlands
DIPF / University of Frankfurt, Germany
• innovations in way network is delivered
• (investigate) corporate/structural alignment
• assist in the development of non-traditional partnerships (Rehab with the
Medicine Community)
• expand investigation and knowledge of PSN'S/PSO's
• continue STHCS sponsored forums on public health issues (medicine
managed care forum)
• inventory assets of all participating agencies (providers, Venn Diagrams)
• access additional funds for telemedicine expansion
• better utilization of current technological bridge
• continued support by STHCS to member facilities
• expand and encourage utilization of interface programs to strengthen the
viability and to improve the health care delivery system (ie teleconference)
• discussion with CCHN
...organize the
issues...
Work
quickly and
effectively
under
pressure
49
Organize the
work when
directions are
not specific.
39
Decide how to
manage
multiple tasks.
20 Manage resources effectively.
4
2. Sort
3. Rate
1. Brainstorm
27 March 2014@HDrachsler 82 / 31
Group Concept Mapping (GCM) Study
Onderwerp via >Beeld >Koptekst en
voettekst
Pagina 83
27 March 2014@HDrachsler 83 / 31
Group Concept Mapping
An essential feature of a higher education institution’s
learning analytics policy should be …
Online sorting
@HDrachsler 27 March 2014 84 / 31
Group Concept Mapping
Online rating
@HDrachsler 27 March 2014 85 / 31
Group Concept Mapping
Participants
Participants
Point Map
1 2
3
4 5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39 40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79 80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
9899
Cluster Replay Map
Cluster Replay Map
Cluster Replay Map
Cluster Map
1. privacy & transparency
2. roles & responsibilities
(of all stakeholders)
3. objectives of LA
(learner and teacher support)
4. risks & challenges
5. data management
6. research & data analysis
Rating Map – Importance
1. privacy & transparency
2. roles & responsibilities
(of all stakeholders)
3. objectives of LA
(learner and teacher support)
4. risks & challenges
5. data management
6. research & data analysis
Cluster Legend
Layer Value
1 5.08 to 5.27
2 5.27 to 5.46
3 5.46 to 5.65
4 5.65 to 5.84
5 5.84 to 6.03
Rating Map – Ease
1. privacy & transparency
2. roles & responsibilities
(of all stakeholders)
3. objectives of LA
(learner and teacher support)
4. risks & challenges
5. data management
6. research & data analysis
Cluster Legend
Layer Value
1 3.79 to 4.12
2 4.12 to 4.45
3 4.45 to 4.78
4 4.78 to 5.11
5 5.11 to 5.44
Rating Ladder Graph
importance ease
privacy & transparency
privacy & transparency
risks & challenges
risks & challenges
roles & responsibilities (of all stakeholders)
roles & responsibilities (of all stakeholders)
objectives of LA (learner and teacher support)
objectives of LA (learner and teacher support)
data management
data management
research & data analysis
research & data analysis
3.79 3.79
6.03 6.03
r = 0.66
Yi-Shan Tsai, Pedro Manuel
Moreno-Marcos, Kairit Tammets,
Kaire Kollom, and Dragan
Gašević. 2018. SHEILA policy
framework: informing
institutional strategies and
policy processes of learning
analytics. In Proceedings of the
8th International Conference on
Learning Analytics and
Knowledge (LAK '18). ACM, New
York, NY, USA, 320-329. DOI:
https://doi.org/10.1145/317035
8.3170367
Go Zone – Roles & Responsibilities
5
38
62
11
19
22
33
39 48
70
91
25
28
37
40
55
61
66
27
47 49
6.08
4.72
3.12
ease
3.83 5.48 6.59
importance
r = 0.26
55. being clear about the purpose of learning analytics
61. a clear articulation of responsibilities when it comes to the use of institutional data
Yi-Shan Tsai, Pedro Manuel
Moreno-Marcos, Kairit Tammets,
Kaire Kollom, and Dragan
Gašević. 2018. SHEILA policy
framework: informing
institutional strategies and
policy processes of learning
analytics. In Proceedings of the
8th International Conference on
Learning Analytics and
Knowledge (LAK '18). ACM, New
York, NY, USA, 320-329. DOI:
https://doi.org/10.1145/317035
8.3170367
@hdrachsler drachsler@dipf.de
99
Many thanks for your attention!
Questions now or later:
Slides: http://bit.ly/TrustedLA
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
SHEILA framework
SHEILA policy framework
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees
Methodology
Literature
- Policy
- Adoption
Academic staff
- Survey
- Focus groups
Students
- Survey
- Focus groups
Senior managers
- Survey
- Interviews
Experts
- Group concept
mapping
Policy
framework
Institutional
policy/strategy
Other stakeh.
- Workshops
- Committees

More Related Content

What's hot

Collaborating with Teachers to Make Great Referrals: Identifying treatment me...
Collaborating with Teachers to Make Great Referrals: Identifying treatment me...Collaborating with Teachers to Make Great Referrals: Identifying treatment me...
Collaborating with Teachers to Make Great Referrals: Identifying treatment me...Bilinguistics
 
Strategies for Enhancing Learning and Teaching focussed continuing profession...
Strategies for Enhancing Learning and Teaching focussed continuing profession...Strategies for Enhancing Learning and Teaching focussed continuing profession...
Strategies for Enhancing Learning and Teaching focussed continuing profession...Learning Enhancement and Development (LEaD)
 
Teaching Excellence Series: The Educator Perspective
Teaching Excellence Series: The Educator PerspectiveTeaching Excellence Series: The Educator Perspective
Teaching Excellence Series: The Educator PerspectiveNewcastle Educators
 
Development of educational tools that enable large-scale ethical empirical re...
Development of educational tools that enable large-scale ethical empirical re...Development of educational tools that enable large-scale ethical empirical re...
Development of educational tools that enable large-scale ethical empirical re...Hassan Khosravi
 
LAK15 panel - European Perspectives
 LAK15 panel - European Perspectives LAK15 panel - European Perspectives
LAK15 panel - European PerspectivesLACE Project
 
Open Education Research : Overview, Benefits and Challenges
Open Education Research : Overview, Benefits and Challenges Open Education Research : Overview, Benefits and Challenges
Open Education Research : Overview, Benefits and Challenges Robert Farrow
 
Hugh Mannerings & Jonathan Neves
Hugh Mannerings & Jonathan NevesHugh Mannerings & Jonathan Neves
Hugh Mannerings & Jonathan NevesLucia Garcia
 
Louise Naylor
Louise NaylorLouise Naylor
Louise NaylorSEDA
 
Successful implementation of the EPQ
Successful implementation of the EPQSuccessful implementation of the EPQ
Successful implementation of the EPQCranleigh School
 
Jekaterina Rogaten
Jekaterina RogatenJekaterina Rogaten
Jekaterina RogatenLucia Garcia
 
researchED Haninge 10 March, 2018
researchED Haninge 10 March, 2018researchED Haninge 10 March, 2018
researchED Haninge 10 March, 2018Gary Jones
 
Scaling up learning analytics
Scaling up learning analyticsScaling up learning analytics
Scaling up learning analyticsRebecca Ferguson
 
The truth about data: discovering what learners really want
The truth about data: discovering what learners really wantThe truth about data: discovering what learners really want
The truth about data: discovering what learners really wantLearningandTeaching
 
Elevator speech for rti
Elevator speech for rtiElevator speech for rti
Elevator speech for rtiKeith Eades
 
Academic integrity: Shining a light on cheating and ghostwriting
Academic integrity: Shining a light on cheating and ghostwritingAcademic integrity: Shining a light on cheating and ghostwriting
Academic integrity: Shining a light on cheating and ghostwritingLearningandTeaching
 

What's hot (20)

Collaborating with Teachers to Make Great Referrals: Identifying treatment me...
Collaborating with Teachers to Make Great Referrals: Identifying treatment me...Collaborating with Teachers to Make Great Referrals: Identifying treatment me...
Collaborating with Teachers to Make Great Referrals: Identifying treatment me...
 
Strategies for Enhancing Learning and Teaching focussed continuing profession...
Strategies for Enhancing Learning and Teaching focussed continuing profession...Strategies for Enhancing Learning and Teaching focussed continuing profession...
Strategies for Enhancing Learning and Teaching focussed continuing profession...
 
Teaching Excellence Series: The Educator Perspective
Teaching Excellence Series: The Educator PerspectiveTeaching Excellence Series: The Educator Perspective
Teaching Excellence Series: The Educator Perspective
 
Peter Chatterton
Peter ChattertonPeter Chatterton
Peter Chatterton
 
Development of educational tools that enable large-scale ethical empirical re...
Development of educational tools that enable large-scale ethical empirical re...Development of educational tools that enable large-scale ethical empirical re...
Development of educational tools that enable large-scale ethical empirical re...
 
LAK15 panel - European Perspectives
 LAK15 panel - European Perspectives LAK15 panel - European Perspectives
LAK15 panel - European Perspectives
 
Open Education Research : Overview, Benefits and Challenges
Open Education Research : Overview, Benefits and Challenges Open Education Research : Overview, Benefits and Challenges
Open Education Research : Overview, Benefits and Challenges
 
Hugh Mannerings & Jonathan Neves
Hugh Mannerings & Jonathan NevesHugh Mannerings & Jonathan Neves
Hugh Mannerings & Jonathan Neves
 
Louise Naylor
Louise NaylorLouise Naylor
Louise Naylor
 
Bob Gilworth
Bob GilworthBob Gilworth
Bob Gilworth
 
HEA Conference 2017 - Strategics for Enhancing L&T CPD
HEA Conference 2017 - Strategics for Enhancing L&T CPD HEA Conference 2017 - Strategics for Enhancing L&T CPD
HEA Conference 2017 - Strategics for Enhancing L&T CPD
 
Sharon Smith
Sharon SmithSharon Smith
Sharon Smith
 
Successful implementation of the EPQ
Successful implementation of the EPQSuccessful implementation of the EPQ
Successful implementation of the EPQ
 
CHEPSAA final networking meeting: activities and outputs
CHEPSAA final networking meeting: activities and outputsCHEPSAA final networking meeting: activities and outputs
CHEPSAA final networking meeting: activities and outputs
 
Jekaterina Rogaten
Jekaterina RogatenJekaterina Rogaten
Jekaterina Rogaten
 
researchED Haninge 10 March, 2018
researchED Haninge 10 March, 2018researchED Haninge 10 March, 2018
researchED Haninge 10 March, 2018
 
Scaling up learning analytics
Scaling up learning analyticsScaling up learning analytics
Scaling up learning analytics
 
The truth about data: discovering what learners really want
The truth about data: discovering what learners really wantThe truth about data: discovering what learners really want
The truth about data: discovering what learners really want
 
Elevator speech for rti
Elevator speech for rtiElevator speech for rti
Elevator speech for rti
 
Academic integrity: Shining a light on cheating and ghostwriting
Academic integrity: Shining a light on cheating and ghostwritingAcademic integrity: Shining a light on cheating and ghostwriting
Academic integrity: Shining a light on cheating and ghostwriting
 

Similar to SHEILA Results – Conference 5 June 2018

SHEILA-CRLI seminar
SHEILA-CRLI seminarSHEILA-CRLI seminar
SHEILA-CRLI seminarYi-Shan Tsai
 
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107Yi-Shan Tsai
 
ACTION RESEARCH
ACTION RESEARCHACTION RESEARCH
ACTION RESEARCHParvathy V
 
FerrellWalker.pptx
FerrellWalker.pptxFerrellWalker.pptx
FerrellWalker.pptxJisc
 
A collaborative and sustainable approach to institutional curriculum development
A collaborative and sustainable approach to institutional curriculum developmentA collaborative and sustainable approach to institutional curriculum development
A collaborative and sustainable approach to institutional curriculum developmentLearningandTeaching
 
A three tier model to promote the institutional adoption of learning analytics
A three tier model to promote the institutional adoption of learning analyticsA three tier model to promote the institutional adoption of learning analytics
A three tier model to promote the institutional adoption of learning analyticsUniversity of Newcastle, NSW.
 
Learning analytics research informed institutional practice
Learning analytics research informed institutional practiceLearning analytics research informed institutional practice
Learning analytics research informed institutional practiceYi-Shan Tsai
 
Leading with pedagogy Sam Harris, University of Queensland
Leading with pedagogy Sam Harris, University of QueenslandLeading with pedagogy Sam Harris, University of Queensland
Leading with pedagogy Sam Harris, University of QueenslandePortfolios Australia
 
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...ePortfolios Australia
 
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014FACT2 Learning Analytics Task Group Phase 2 report - CIT2014
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014Greg Ketcham
 
ABLE - NTU Danish visit February 2018
ABLE - NTU Danish visit February 2018ABLE - NTU Danish visit February 2018
ABLE - NTU Danish visit February 2018Ed Foster
 
E12+Analytics+Workshop+ppt.pptx
E12+Analytics+Workshop+ppt.pptxE12+Analytics+Workshop+ppt.pptx
E12+Analytics+Workshop+ppt.pptxchatbot9
 
ABLE - the NTU Student Dashboard - University of Derby
ABLE - the NTU Student Dashboard - University of DerbyABLE - the NTU Student Dashboard - University of Derby
ABLE - the NTU Student Dashboard - University of DerbyEd Foster
 
ABLE - NTU Eunis case study - Nov 2017
ABLE - NTU Eunis case study - Nov 2017ABLE - NTU Eunis case study - Nov 2017
ABLE - NTU Eunis case study - Nov 2017Ed Foster
 
HEA Framework for Employability
HEA Framework for EmployabilityHEA Framework for Employability
HEA Framework for EmployabilityMoira Wright
 
Leadership in eLearning: Developing the Next Generation of Leaders
Leadership in eLearning: Developing the Next Generation of LeadersLeadership in eLearning: Developing the Next Generation of Leaders
Leadership in eLearning: Developing the Next Generation of LeadersMark Brown
 
Administration and supervision
Administration and supervisionAdministration and supervision
Administration and supervisionKongz Campiseño
 
Designing Systemic Learning Analytics at the Open University
Designing Systemic Learning Analytics at the Open UniversityDesigning Systemic Learning Analytics at the Open University
Designing Systemic Learning Analytics at the Open UniversitySimon Buckingham Shum
 
Krakow presentation speak_appsmngm_final
Krakow presentation speak_appsmngm_finalKrakow presentation speak_appsmngm_final
Krakow presentation speak_appsmngm_finalSpeakApps Project
 
Learning Analytics: connecting analytics, theory and the organisation Nov 14
Learning Analytics: connecting analytics, theory and the organisation Nov 14Learning Analytics: connecting analytics, theory and the organisation Nov 14
Learning Analytics: connecting analytics, theory and the organisation Nov 14siwelsh
 

Similar to SHEILA Results – Conference 5 June 2018 (20)

SHEILA-CRLI seminar
SHEILA-CRLI seminarSHEILA-CRLI seminar
SHEILA-CRLI seminar
 
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107
Supporting Higher Education to Integrate Learning Analytics_EUNIS20171107
 
ACTION RESEARCH
ACTION RESEARCHACTION RESEARCH
ACTION RESEARCH
 
FerrellWalker.pptx
FerrellWalker.pptxFerrellWalker.pptx
FerrellWalker.pptx
 
A collaborative and sustainable approach to institutional curriculum development
A collaborative and sustainable approach to institutional curriculum developmentA collaborative and sustainable approach to institutional curriculum development
A collaborative and sustainable approach to institutional curriculum development
 
A three tier model to promote the institutional adoption of learning analytics
A three tier model to promote the institutional adoption of learning analyticsA three tier model to promote the institutional adoption of learning analytics
A three tier model to promote the institutional adoption of learning analytics
 
Learning analytics research informed institutional practice
Learning analytics research informed institutional practiceLearning analytics research informed institutional practice
Learning analytics research informed institutional practice
 
Leading with pedagogy Sam Harris, University of Queensland
Leading with pedagogy Sam Harris, University of QueenslandLeading with pedagogy Sam Harris, University of Queensland
Leading with pedagogy Sam Harris, University of Queensland
 
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...
Collaborative, Program-wide Alignment of Assessments and ePortfolios to Build...
 
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014FACT2 Learning Analytics Task Group Phase 2 report - CIT2014
FACT2 Learning Analytics Task Group Phase 2 report - CIT2014
 
ABLE - NTU Danish visit February 2018
ABLE - NTU Danish visit February 2018ABLE - NTU Danish visit February 2018
ABLE - NTU Danish visit February 2018
 
E12+Analytics+Workshop+ppt.pptx
E12+Analytics+Workshop+ppt.pptxE12+Analytics+Workshop+ppt.pptx
E12+Analytics+Workshop+ppt.pptx
 
ABLE - the NTU Student Dashboard - University of Derby
ABLE - the NTU Student Dashboard - University of DerbyABLE - the NTU Student Dashboard - University of Derby
ABLE - the NTU Student Dashboard - University of Derby
 
ABLE - NTU Eunis case study - Nov 2017
ABLE - NTU Eunis case study - Nov 2017ABLE - NTU Eunis case study - Nov 2017
ABLE - NTU Eunis case study - Nov 2017
 
HEA Framework for Employability
HEA Framework for EmployabilityHEA Framework for Employability
HEA Framework for Employability
 
Leadership in eLearning: Developing the Next Generation of Leaders
Leadership in eLearning: Developing the Next Generation of LeadersLeadership in eLearning: Developing the Next Generation of Leaders
Leadership in eLearning: Developing the Next Generation of Leaders
 
Administration and supervision
Administration and supervisionAdministration and supervision
Administration and supervision
 
Designing Systemic Learning Analytics at the Open University
Designing Systemic Learning Analytics at the Open UniversityDesigning Systemic Learning Analytics at the Open University
Designing Systemic Learning Analytics at the Open University
 
Krakow presentation speak_appsmngm_final
Krakow presentation speak_appsmngm_finalKrakow presentation speak_appsmngm_final
Krakow presentation speak_appsmngm_final
 
Learning Analytics: connecting analytics, theory and the organisation Nov 14
Learning Analytics: connecting analytics, theory and the organisation Nov 14Learning Analytics: connecting analytics, theory and the organisation Nov 14
Learning Analytics: connecting analytics, theory and the organisation Nov 14
 

More from LACE Project

SHEILA LAK18 Presentation
SHEILA LAK18 PresentationSHEILA LAK18 Presentation
SHEILA LAK18 PresentationLACE Project
 
SHEILA Project – LAK18 Workshop Slides
SHEILA Project – LAK18 Workshop SlidesSHEILA Project – LAK18 Workshop Slides
SHEILA Project – LAK18 Workshop SlidesLACE Project
 
SHEILA project LAK17 workshop slides
SHEILA project LAK17 workshop slidesSHEILA project LAK17 workshop slides
SHEILA project LAK17 workshop slidesLACE Project
 
SHEILA Project - Workshop Slides Online Educa Berlin 2016
SHEILA Project - Workshop Slides Online Educa Berlin 2016SHEILA Project - Workshop Slides Online Educa Berlin 2016
SHEILA Project - Workshop Slides Online Educa Berlin 2016LACE Project
 
Overview of the LAEP learning analytics project
Overview of the LAEP learning analytics projectOverview of the LAEP learning analytics project
Overview of the LAEP learning analytics projectLACE Project
 
EU-funded learning analytics projects
EU-funded learning analytics projectsEU-funded learning analytics projects
EU-funded learning analytics projectsLACE Project
 
Expert presentations: LAEP / LACE Amsterdam workshop
Expert presentations: LAEP / LACE Amsterdam workshopExpert presentations: LAEP / LACE Amsterdam workshop
Expert presentations: LAEP / LACE Amsterdam workshopLACE Project
 
Research in to Practice: Building and implementing learning analytics at Tribal
Research in to Practice: Building and implementing learning analytics at TribalResearch in to Practice: Building and implementing learning analytics at Tribal
Research in to Practice: Building and implementing learning analytics at TribalLACE Project
 
Leveraging the full value of learning analytics: The power of learning design
Leveraging the full value of learning analytics: The power of learning designLeveraging the full value of learning analytics: The power of learning design
Leveraging the full value of learning analytics: The power of learning designLACE Project
 
Learning analytics LACE SoLAR Flare 2015
Learning analytics LACE SoLAR Flare 2015Learning analytics LACE SoLAR Flare 2015
Learning analytics LACE SoLAR Flare 2015LACE Project
 
Scalable Learning Analytics and Interoperability – an assessment of potential...
Scalable Learning Analytics and Interoperability – an assessment of potential...Scalable Learning Analytics and Interoperability – an assessment of potential...
Scalable Learning Analytics and Interoperability – an assessment of potential...LACE Project
 
Assessment Analytics - EUNIS 2015 E-Learning Task Force Workshop
Assessment Analytics - EUNIS 2015 E-Learning Task Force WorkshopAssessment Analytics - EUNIS 2015 E-Learning Task Force Workshop
Assessment Analytics - EUNIS 2015 E-Learning Task Force WorkshopLACE Project
 
LACE Spring Briefing - WatchMe project overview
LACE Spring Briefing - WatchMe project overviewLACE Spring Briefing - WatchMe project overview
LACE Spring Briefing - WatchMe project overviewLACE Project
 
LACE Spring Briefing - Lea's Box Project overview
LACE Spring Briefing - Lea's Box Project overviewLACE Spring Briefing - Lea's Box Project overview
LACE Spring Briefing - Lea's Box Project overviewLACE Project
 
LACE Spring Briefing - PELARS project overview
LACE Spring Briefing - PELARS project overview LACE Spring Briefing - PELARS project overview
LACE Spring Briefing - PELARS project overview LACE Project
 
LACE Spring Briefing - the LACE project
LACE Spring Briefing - the LACE project LACE Spring Briefing - the LACE project
LACE Spring Briefing - the LACE project LACE Project
 
LACE Spring Briefing - Learning analytics are more than measurement
LACE Spring Briefing - Learning analytics are more than measurementLACE Spring Briefing - Learning analytics are more than measurement
LACE Spring Briefing - Learning analytics are more than measurementLACE Project
 
Themes in Learning Analytics - A Critical View
Themes in Learning Analytics - A Critical ViewThemes in Learning Analytics - A Critical View
Themes in Learning Analytics - A Critical ViewLACE Project
 
European Perspectives on Learning Analytics: LAK15 LACE panel
European Perspectives on Learning Analytics: LAK15 LACE panelEuropean Perspectives on Learning Analytics: LAK15 LACE panel
European Perspectives on Learning Analytics: LAK15 LACE panelLACE Project
 
Open Learning Analytics Network - Summit Europe 2014
Open Learning Analytics Network - Summit Europe 2014Open Learning Analytics Network - Summit Europe 2014
Open Learning Analytics Network - Summit Europe 2014LACE Project
 

More from LACE Project (20)

SHEILA LAK18 Presentation
SHEILA LAK18 PresentationSHEILA LAK18 Presentation
SHEILA LAK18 Presentation
 
SHEILA Project – LAK18 Workshop Slides
SHEILA Project – LAK18 Workshop SlidesSHEILA Project – LAK18 Workshop Slides
SHEILA Project – LAK18 Workshop Slides
 
SHEILA project LAK17 workshop slides
SHEILA project LAK17 workshop slidesSHEILA project LAK17 workshop slides
SHEILA project LAK17 workshop slides
 
SHEILA Project - Workshop Slides Online Educa Berlin 2016
SHEILA Project - Workshop Slides Online Educa Berlin 2016SHEILA Project - Workshop Slides Online Educa Berlin 2016
SHEILA Project - Workshop Slides Online Educa Berlin 2016
 
Overview of the LAEP learning analytics project
Overview of the LAEP learning analytics projectOverview of the LAEP learning analytics project
Overview of the LAEP learning analytics project
 
EU-funded learning analytics projects
EU-funded learning analytics projectsEU-funded learning analytics projects
EU-funded learning analytics projects
 
Expert presentations: LAEP / LACE Amsterdam workshop
Expert presentations: LAEP / LACE Amsterdam workshopExpert presentations: LAEP / LACE Amsterdam workshop
Expert presentations: LAEP / LACE Amsterdam workshop
 
Research in to Practice: Building and implementing learning analytics at Tribal
Research in to Practice: Building and implementing learning analytics at TribalResearch in to Practice: Building and implementing learning analytics at Tribal
Research in to Practice: Building and implementing learning analytics at Tribal
 
Leveraging the full value of learning analytics: The power of learning design
Leveraging the full value of learning analytics: The power of learning designLeveraging the full value of learning analytics: The power of learning design
Leveraging the full value of learning analytics: The power of learning design
 
Learning analytics LACE SoLAR Flare 2015
Learning analytics LACE SoLAR Flare 2015Learning analytics LACE SoLAR Flare 2015
Learning analytics LACE SoLAR Flare 2015
 
Scalable Learning Analytics and Interoperability – an assessment of potential...
Scalable Learning Analytics and Interoperability – an assessment of potential...Scalable Learning Analytics and Interoperability – an assessment of potential...
Scalable Learning Analytics and Interoperability – an assessment of potential...
 
Assessment Analytics - EUNIS 2015 E-Learning Task Force Workshop
Assessment Analytics - EUNIS 2015 E-Learning Task Force WorkshopAssessment Analytics - EUNIS 2015 E-Learning Task Force Workshop
Assessment Analytics - EUNIS 2015 E-Learning Task Force Workshop
 
LACE Spring Briefing - WatchMe project overview
LACE Spring Briefing - WatchMe project overviewLACE Spring Briefing - WatchMe project overview
LACE Spring Briefing - WatchMe project overview
 
LACE Spring Briefing - Lea's Box Project overview
LACE Spring Briefing - Lea's Box Project overviewLACE Spring Briefing - Lea's Box Project overview
LACE Spring Briefing - Lea's Box Project overview
 
LACE Spring Briefing - PELARS project overview
LACE Spring Briefing - PELARS project overview LACE Spring Briefing - PELARS project overview
LACE Spring Briefing - PELARS project overview
 
LACE Spring Briefing - the LACE project
LACE Spring Briefing - the LACE project LACE Spring Briefing - the LACE project
LACE Spring Briefing - the LACE project
 
LACE Spring Briefing - Learning analytics are more than measurement
LACE Spring Briefing - Learning analytics are more than measurementLACE Spring Briefing - Learning analytics are more than measurement
LACE Spring Briefing - Learning analytics are more than measurement
 
Themes in Learning Analytics - A Critical View
Themes in Learning Analytics - A Critical ViewThemes in Learning Analytics - A Critical View
Themes in Learning Analytics - A Critical View
 
European Perspectives on Learning Analytics: LAK15 LACE panel
European Perspectives on Learning Analytics: LAK15 LACE panelEuropean Perspectives on Learning Analytics: LAK15 LACE panel
European Perspectives on Learning Analytics: LAK15 LACE panel
 
Open Learning Analytics Network - Summit Europe 2014
Open Learning Analytics Network - Summit Europe 2014Open Learning Analytics Network - Summit Europe 2014
Open Learning Analytics Network - Summit Europe 2014
 

Recently uploaded

Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3JemimahLaneBuaron
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfchloefrazer622
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Disha Kariya
 
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...PsychoTech Services
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 

Recently uploaded (20)

Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdf
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 

SHEILA Results – Conference 5 June 2018

  • 1. SHEILA Project Results Dragan Gašević @dgasevic SHEILA Conference 5th June 2018 Brussels, Belgium http://sheilaproject.eu/
  • 4. Inclusive adoption process Macfadyen, L., Dawson, S., Pardo, A., Gašević, D., (2014). The learning analytics imperative and the sociotechnical challenge: Policy for complex systems. Research & Practice in Assessment, 9(Winter 2014), 17-28.
  • 5. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 6. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 7. Adoption challenge Leadership for strategic implementation & monitoring Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
  • 8. Adoption challenge Equal engagement with different stakeholders Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
  • 9. Adoption challenge Training to cultivate data literacy among primary stakeholders Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
  • 10. Adoption challenge Policies for learning analytics practice Tsai, Y. S., & Gasevic, D. (2017). Learning analytics in higher education – challenges and policies: a review of eight learning analytics policies. In Proceedings of the Seventh International Learning Analytics & Knowledge Conference (pp. 233-242).
  • 11. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 12. Institutional survey & interviews Yi-Shan Tsai University of Edinburgh yi-shan.tsai@ed.ac.uk @yi_shan_tsai
  • 13. What is the state of the art? What are the drivers? What are the challenges?
  • 14. Survey • 22 countries, 46 institutions • November 2016 NO P LA NS IN P RE P A RA TION IMP LE ME NT ED 2 13 15 16 The adoption of LA Institution-wide Small scale N/A
  • 15. Interviews • 16 countries, 51 HEIs, 64 interviews, 78 participants • August 2016 - January 2017 N O P L A N S I N P R E P A R A T I O N I M P L E M E N T E D 9 7 5 12 18 The adoption of learning analytics (interviews) Institution-wide Partial/ Pilots Data exploration/cleaning
  • 16. Motivations to adopt learning analytics • To improve student learning performance – 40 (87%) • To improve student satisfaction – 33 (72%) • To improve teaching excellence – 33 (72 %) • To improve student retention– 26 (57 %) • To explore what learning analytics can do for our institution/ staff/ students – 25 (54 %) 46 institutions
  • 17. Motivations to adopt learning analytics • To improve student learning performance – 40 (87%) • To improve student satisfaction – 33 (72%) • To improve teaching excellence – 33 (72 %) • To improve student retention– 26 (57 %) • To explore what learning analytics can do for our institution/ staff/ students – 25 (54 %) 46 institutions
  • 18. Motivations to adopt learning analytics • To improve student learning performance – 40 (87%) • To improve student satisfaction – 33 (72%) • To improve teaching excellence – 33 (72 %) • To improve student retention– 26 (57 %) • To explore what learning analytics can do for our institution/ staff/ students – 25 (54 %) 46 institutions
  • 21. “People are thinking about learning analytics as a way to try and personalise education and enhance education. And actually make our education more inclusive both by understanding how different students engage with different bits of educational processes, but also about through developing curricula to make them more flexible and inclusive as a standard.”
  • 22. “I think what we would be looking at is how do we evolve the way we teach to provide better learning outcomes for the students, greater mastery of the subject.”
  • 23. “We’re trying to understand better the curriculum that needs to be offered for the students in our region. And…I think importantly how our pedagogical model fits that and deliver the best experience for our students.”
  • 24. Barriers to the success of learning analytics • Analytics expertise – 34 (76%) • A data-driven culture at the institution – 30 (67%) • Teaching staff/tutor buy-in – 29 (64%) • The affordances of current learning analytics technology – 29 (64%)
  • 25. Ethical and privacy concerns access transparency anonymity
  • 26. Analytical capability challenge • More than half of the institutions with 0-3 years experience indicated noticeable gaps among different stakeholders regarding perceptions and understanding of learning analytics.
  • 27. Success • We have achieved the goals that we set for learning analytics.
  • 28. Success • Most institutions did not have confirmed success. • Half of the institutions with 0-3 years experience claimed “gaining experience” as part of their success.
  • 29. Implications • Interests were high but experiences were premature. • There was strong motivation in increasing institutional performance by improving teaching quality. • Key barriers were around skills, institutional culture, technology, ethics and privacy.
  • 30. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 31. Student Survey Results Alex Wainwright University of Liverpool a.wainwright@Liverpool.ac.uk http://sheilaproject.eu/
  • 32. Background • 12 Items Survey • Two Subscales: • Ethical and Privacy Expectations • Service Expectations • 6 Distributions: • Edinburgh (N = 884) • Liverpool (N = 191) • Tallinn (N = 161) • Madrid (N = 543) • Netherlands (N = 1247) • Blanchardstown (N = 237) http://sheilaproject.eu/
  • 33. Ideal Expectation Scale Predicted Expectation Scale Alternative Purpose Consent to Collect Identifiable Data Keep Data Secure Third Party Alternative Purpose Consent to Collect Identifiable Data Keep Data Secure Third Party 1 2 3 4 5 6 7 Item Average Location Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Ethical and Privacy Expectations http://sheilaproject.eu/
  • 34. Keep Data Secure – Predicted Expectation Scale Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree Response 10 20 30 40 50 Percentage http://sheilaproject.eu/
  • 35. Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree Response 10 20 30 Percentage Consent to Collect – Predicted Expectation Scale http://sheilaproject.eu/
  • 36. Ideal Expectation Scale Predicted Expectation Scale ObligationtoAct IntegrateintoFeedback SkillDevelopment RegularlyUpdate CompleteProfile StudentDecisionMaking CourseGoals ObligationtoAct IntegrateintoFeedback SkillDevelopment RegularlyUpdate CompleteProfile StudentDecisionMaking CourseGoals 1 2 3 4 5 6 7 Average Location Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Service Expectations http://sheilaproject.eu/
  • 37. Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree Response 10 20 30 Percentage Course Goals – Predicted Expectation Scale http://sheilaproject.eu/
  • 38. Blanchardstown Edinburgh Liverpool Madrid Open University of the Netherlands Tallinn Strongly Disagree Disagree Somewhat Disagree Neither Agree nor Disagree Somewhat Agree Agree Strongly Agree Response 10 20 Percentage Obligation to Act – Predicted Expectation Scale http://sheilaproject.eu/
  • 39. Summary • Beliefs towards learning analytics are not consistent. • Emphasis on data security and improving learning. http://sheilaproject.eu/
  • 40. Student focus groups Pedro Manuel Moreno Marcos Department of Telematic Engineering Universidad Carlos III de Madrid http://sheilaproject.eu/
  • 41. Goals • Interest and expectations • Awareness • Concerns
  • 42. Background • 18 focus groups • 4 partners’ institutions • 74 students • Interviews: Around 1h http://sheilaproject.eu/
  • 43. Interests and expectations • Improve the quality of teaching • Better student-teacher feedback • Better academic resources and academic tools to improve learning • Personalized support • Recommendation of learning resources • Feedback from a system, via a dashboard • Provide an overview of the tasks to be done in a semester → improve curriculum design http://sheilaproject.eu/
  • 44. Awareness • Students do not know what LA is, but they recognise its importance if it can solve students’ problems • Students are not generally aware of the data collected → Transparency • Students have not checked the conditions they have accepted about data http://sheilaproject.eu/
  • 45. Concerns http://sheilaproject.eu/ Surveillance Anonymization Purpose of data Kind of data Consent and access Security Provision of opt-outs Stereotypes and biases
  • 46. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 47. Staff Survey Maren Scheffel Open Universiteit Nederland Welten Institute @m_a_s_c
  • 48. With regards to learning analytics … … what do academic staff ideally expect to happen? … what do academic staff predict to happen in reality? Goal of the survey
  • 49. 4 academic institutions University of Edinburgh Carlos III Madrid n = 81 n = 26 Open Universiteit University of Tallinn n = 54 n = 49 from spring to fall 2017
  • 50. 16 items, some examples The university will provide me with guidance on how to access LA about my students The LA service will show how a student’s learning progress compares to their learning goals/the course objectives The teaching staff will have an obligation to act if the analytics show that a student is at-risk of failing, underperforming, or that they could improve their learning
  • 51.
  • 52.
  • 53. University of Edinburgh: • Ideal: LA will collect and present data that is accurate (M = 5.91) Q9 • Predicted: Providing guidance to access LA about students (M = 5.05) Q1 Carlos III de Madrid: • Ideal: LA presented in a format that is understandable and easy to read (M = 6.31) Q11 • Predicted: LA will present students with a complete profile of their learning across every course (M = 5.27) Q12 Highest expectation values
  • 54. Highest expectation values Open Universiteit Nederland: • Ideal: LA will collect and present data that is accurate (M = 6.60) Q9 • Predicted: Able to access data about students’ progress in a course that I am teaching (M = 5.17) Q4 University of Tallinn: • Ideal: Able to access data about students’ progress in a course that I am teaching (M = 6.04) Q4 • Predicted: Able to access data about students’ progress in a course that I am teaching (M = 5.49) Q4
  • 55.
  • 56. Lowest expectation values University of Edinburgh: • Ideal: Teaching staff will have an obligation to act if students are found to be at- risk of failing or under performing (M = 3.65) Q14 • Predicted: Teaching staff will be competent in incorporating analytics into the feedback and support they provide to students (M = 3.49) Q13 Carlos III de Madrid: • Ideal: Teaching staff will have an obligation to act if students are found to be at- risk of failing or under performing (M = 4.42) Q14 • Predicted: Teaching staff will have an obligation to act if students are found to be at-risk of failing or under performing (M = 3.77) Q14
  • 57. Lowest expectation values Open Universiteit Nederland: • Ideal: Teaching staff will have an obligation to act if students are found to be at- risk of failing or under performing (M = 4.44) Q14 • Predicted: Feedback from analytics will be used to promote students’ academic and professional skill development for future employability (M = 3.24) Q15 University of Tallinn: • Ideal: Teaching staff will have an obligation to act if students are found to be at- risk of failing or under performing (M = 4.80) Q14 • Predicted: Q14 (M = 3.82)
  • 58. Staff focus groups Kairit Tammets Centre for educational technology School of Digital Technologies Tallinn University
  • 59. Goal To better understand the viewpoints of academic staff on: • Learning analytics opportunities in the HEIs from the perspective of students, teachers and programs; • Concerns related with adapting of learning analytics; • Needed steps to adopt learning analytics at the HEIs
  • 60. Study participants • University of Edinburgh: 5 focus groups, 18 teaching staff • Universidad Carlos III de Madrid: 4 focus groups, 16 teaching staff • Open Universiteit Nederland: 2 focus groups, 5 teaching staff • Tallinn University: 5 focus groups, 20 teaching staff
  • 61. Results: Expectations & LA opportunities STUDENT LEVEL TEACHER LEVEL PROGRAM LEVEL Take responsibility for their learning and enhancing their SRL- skills Assess the degree of success to prevent students from begin worried or optimistic about their performance Method to identify student’s weaknesses and know where students are with their progress Understand how students engage with learning content Improve of the design and provision of learning materials, courses, curriculum and support to students Understand how program is working (strengths and bottlenecks) Improve educational quality (e.g. content level)
  • 67. Results: concerns – student level https://www.pinterest.com/pin/432486370448743887/
  • 68. Results: Concerns – student level https://www.pinterest.com/pin/432486370448743887/
  • 69. Results: concerns – student level https://www.pinterest.com/pin/432486370448743887/
  • 70. Results: concerns – teacher level https://www.pinterest.com/pin/432486370448743887/
  • 71. Results: concerns – teacher level http://create-learning.com https://www.pinterest.com/pin/432486370448743887/
  • 72. Results: concerns – teacher level http://create-learning.com https://www.pinterest.com/pin/432486370448743887/ Http://memegenerator.net
  • 73. Results: concerns – program level • Interpretation of learning: • Was the right data collected? • Were the accurate algorithms developed ? • Was an appropriate message given for the students? • Connecting LA to real learning – is this meaningful picture of learning what is happening in online environments?
  • 74. What we should consider? • LA should be just one component of many for collecting feedback and enhancing decision-making • Involve stakeholders: • Academic staff to in developing and setting up of LA • Pedagogy experts involved to ensure data makes sense to improve learning • Provide training, communication!
  • 75. What we should consider? •Design of the tools that are: •Easy to use •Providing visualizations of data •Not requiring mathematical/statistical skills •Not taking a lot of time •Considering ethical and privacy aspects
  • 76. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 77. Group Concept Mapping Prof. Dr. Hendrik Drachsler Open University Netherlands DIPF / University of Frankfurt, Germany
  • 78. • innovations in way network is delivered • (investigate) corporate/structural alignment • assist in the development of non-traditional partnerships (Rehab with the Medicine Community) • expand investigation and knowledge of PSN'S/PSO's • continue STHCS sponsored forums on public health issues (medicine managed care forum) • inventory assets of all participating agencies (providers, Venn Diagrams) • access additional funds for telemedicine expansion • better utilization of current technological bridge • continued support by STHCS to member facilities • expand and encourage utilization of interface programs to strengthen the viability and to improve the health care delivery system (ie teleconference) • discussion with CCHN ...organize the issues... Work quickly and effectively under pressure 49 Organize the work when directions are not specific. 39 Decide how to manage multiple tasks. 20 Manage resources effectively. 4 2. Sort 3. Rate 1. Brainstorm 27 March 2014@HDrachsler 82 / 31 Group Concept Mapping (GCM) Study
  • 79. Onderwerp via >Beeld >Koptekst en voettekst Pagina 83 27 March 2014@HDrachsler 83 / 31 Group Concept Mapping An essential feature of a higher education institution’s learning analytics policy should be …
  • 80. Online sorting @HDrachsler 27 March 2014 84 / 31 Group Concept Mapping
  • 81. Online rating @HDrachsler 27 March 2014 85 / 31 Group Concept Mapping
  • 84. Point Map 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 9899
  • 88. Cluster Map 1. privacy & transparency 2. roles & responsibilities (of all stakeholders) 3. objectives of LA (learner and teacher support) 4. risks & challenges 5. data management 6. research & data analysis
  • 89. Rating Map – Importance 1. privacy & transparency 2. roles & responsibilities (of all stakeholders) 3. objectives of LA (learner and teacher support) 4. risks & challenges 5. data management 6. research & data analysis Cluster Legend Layer Value 1 5.08 to 5.27 2 5.27 to 5.46 3 5.46 to 5.65 4 5.65 to 5.84 5 5.84 to 6.03
  • 90. Rating Map – Ease 1. privacy & transparency 2. roles & responsibilities (of all stakeholders) 3. objectives of LA (learner and teacher support) 4. risks & challenges 5. data management 6. research & data analysis Cluster Legend Layer Value 1 3.79 to 4.12 2 4.12 to 4.45 3 4.45 to 4.78 4 4.78 to 5.11 5 5.11 to 5.44
  • 91. Rating Ladder Graph importance ease privacy & transparency privacy & transparency risks & challenges risks & challenges roles & responsibilities (of all stakeholders) roles & responsibilities (of all stakeholders) objectives of LA (learner and teacher support) objectives of LA (learner and teacher support) data management data management research & data analysis research & data analysis 3.79 3.79 6.03 6.03 r = 0.66
  • 92. Yi-Shan Tsai, Pedro Manuel Moreno-Marcos, Kairit Tammets, Kaire Kollom, and Dragan Gašević. 2018. SHEILA policy framework: informing institutional strategies and policy processes of learning analytics. In Proceedings of the 8th International Conference on Learning Analytics and Knowledge (LAK '18). ACM, New York, NY, USA, 320-329. DOI: https://doi.org/10.1145/317035 8.3170367
  • 93. Go Zone – Roles & Responsibilities 5 38 62 11 19 22 33 39 48 70 91 25 28 37 40 55 61 66 27 47 49 6.08 4.72 3.12 ease 3.83 5.48 6.59 importance r = 0.26 55. being clear about the purpose of learning analytics 61. a clear articulation of responsibilities when it comes to the use of institutional data
  • 94. Yi-Shan Tsai, Pedro Manuel Moreno-Marcos, Kairit Tammets, Kaire Kollom, and Dragan Gašević. 2018. SHEILA policy framework: informing institutional strategies and policy processes of learning analytics. In Proceedings of the 8th International Conference on Learning Analytics and Knowledge (LAK '18). ACM, New York, NY, USA, 320-329. DOI: https://doi.org/10.1145/317035 8.3170367
  • 95. @hdrachsler drachsler@dipf.de 99 Many thanks for your attention! Questions now or later: Slides: http://bit.ly/TrustedLA
  • 96. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 99. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees
  • 100. Methodology Literature - Policy - Adoption Academic staff - Survey - Focus groups Students - Survey - Focus groups Senior managers - Survey - Interviews Experts - Group concept mapping Policy framework Institutional policy/strategy Other stakeh. - Workshops - Committees

Editor's Notes

  1. With senior managers, we were
  2. EADTU (European Association of Distance Teaching Universities) EUA (European University Association) HeLF (Heads of e-Learning Forum) EUNIS (European University Information Systems) SNOLA (Spanish Network of Learning Analytics) eMadrid
  3. 22 countries: Austria, Bulgaria, Cyprus , Czech Republic, Denmark, Estonia, Finland, Germany, Hungary, Ireland, Italy, Lithuania, Netherlands, Norway, Portugal, Romania, Serbia, Slovakia, Spain, Switzerland, Turkey, UK Interview + survey: 26 countries
  4. A survey question (multiple choices) provided 11 options for motivations specific to learning and teaching.
  5. All related to institutional performance: league ranking, satisfaction survey, teaching excellence framework But also dependent on teaching quality
  6. Early stage - exploration
  7. Most institutions seem to have incorporated all levels of goals into their planning or implementation of LA Enhance self-regulation skills: provide data-based information to guide students Improve learning support: curriculum, feedback, personalized support, pastoral care, timely support Increase institutional performance: retention rate, student satisfaction, league ranking
  8. Particularly teaching-level (1/1 – S; 18/19 – M; 6/7- L) – UoE data
  9. 13 options Moderately-sized, large, critical
  10. Three most mentioned issues regarding ethics and privacy
  11. 15 institutions that have implemented LA.
  12. UoE data
  13. Interest is strong Institutions were exploring what LA can do Using LA to enhance teaching so as to increase institutional performance is the biggest motivation among managers Barriers – skills, culture, technology, ethics and privacy
  14. Hi – I’m Alex Wainwright… and I’m going to give an overview of the student survey results… this is going to cover response rates and some general insights obtained…
  15. The student survey is composed of 12 items… and responses are made on two scales that correspond to a desired service… and what students expect in reality… so they reflect two levels of expectation… Through the development and validation process we have identified two subscales… these refers to ethical and privacy expectations… such as whether students expect to provide consent for the collection of their educational data… And the other subscales refers to service expectations… so this covers things such as whether students expect to receive updates on how their learning progress compares to a set goal…. As you can see… we have distributed the instrument at six different higher education institutions… with the highest response rate being at the open university of the netherlands… All distributions have shown the scales to be valid and to also show excellent measurement quality….
  16. Firstly… I am going to go over the ethical and privacy expectation items… On this figure you can see the average responses to these items by expectation scale and location…. The x axis provides an indication of what the items refer to… So we have beliefs about providing consent when data is used for an alternative purpose… or whether consent should be sought before distributing data to third party companies What can be seen is that students ideal expectations are generally higher than predicted expectations – this is anticipated as it is a desired level of service… Across both scales… however… we can see that the expectation that all collected data remain secure receives the highest average response… whereas… the expectation to provide consent before educational data is collected and analysed receives the lowest average response across these five items… and whilst students agree with this belief… it verges on indifference on the predicted expectation scale for the Spanish student sample…. It may be that students are open to universities collecting and analysing educational data… particularly as it is used for attendance purposes, for example… Whereas… they have stronger beliefs toward universities abiding by data handling policies that will ensure that all data remains secure…
  17. We can also look at these two particular ethical and privacy expectation items in more detail… This figure shows the percentage of students responding in a certain way to the data security expectation… with darker colours reflecting a higher percentage of students responding that way… And what is show is that… between 60 to 80% of students across all universities either agreed or strongly agreed with the expectation that universities will ensure data is kept secure…
  18. For the consent to collect expectation… this figure shows that there is more variation in the responses… For those students from Edinburgh, Liverpool, the Netherlands, and Blanchardstown… the largest response of around 30% is for strongly agree to this belief… Whereas… the largest percentage of responses for Madrid and Tallinn… which was around 25%... Was for somewhat agree…
  19. Looking at the service expectation items… we can that the average responses tend to be similar across locations… Of particular note… the obligation to act is the item with the lowest response on average… with students in Madrid, the Netherlands, and Tallinn generally showing indifference to this belief on the predicted expectation scale... The higher average responses… on the other hand… seem to be around aspects of self-regulated learning such as students expecting to receive a complete profile of their learning…. Making their own decisions on the analytics that they receive… and knowing how their progress compares to a set learning goal….
  20. Looking into what are the highest and lowest average response items… we can also understand differences within each sample… For knowing how progress compares to a set learning goal… between 20 to 35% of students across each sample agreed with this expectation… with around 4% disagreeing….
  21. As for the obligation to act… the highest response rates are variable… Around 20% of students in the Tallinn and Madrid samples somewhat disagreed with this expectation… For the Dutch students 24% expressed indifference to this belief… whereas in Liverpool and Blanchardstown around 28% showed agreement…
  22. The output from the student survey shows that the expectations of students towards learning analytics are not consistent across each sample… with students generally showing variations in what they want from such services… On the other hand… we can generally see that students expect a learning analytics service that emphasises data security… and provides tools that support learning as opposed to those that emphasise early interventions
  23. Engagement: what do they view, do with the content and relationship between participation and grades Academic data: Pre-requisite subjects to understand the background Previous studies to know if additional materials are needed Demographic: Previous studies Employment Health: Visually impaired – to adjust content Educational needs – to adjust content mental issues – to be aware
  24. Academic data: Pre-requisite subjects to understand the background Previous studies to know if additional materials are needed Demographic: Previous studies Employment Health: Visually impaired – to adjust content Educational needs – to adjust content mental issues – to be aware
  25. Demographic: Previous studies Employment Health: Visually impaired – to adjust content Educational needs – to adjust content mental issues – to be aware
  26. Health: Visually impaired – to adjust content Educational needs – to adjust content mental issues – to be aware
  27. Qualitative data to get addition to LA data students’ perceptions and understandings about teaching and learning processes
  28. Staff is worried that we will put a lot of efforts and resources for developing LA services – but maybe students do not want such service and will never benefit it as they should
  29. Privacy and autonomy – staff emphasized that services should accept the ethical regulations and behaviour in all the different levels
  30. Staff was worried that profiling of the students as e.g. low-performing might end up with the lost of motivation and anxiety.
  31. The main concern on teacher level was related with the time constraints: staff was not very convinced now that LA may help them, rather they felt it takes more time from them, especially if they do have to monitor each individual student.
  32. Performance judgement- will my professionalism will be evaluated based on LA data Fulfilling expectations
  33. Staff was wondering: shall I be objective?
  34. How can I be objective