Advertisement
Advertisement

More Related Content

Advertisement

Data fusion and mining in SPHERE

  1. Data fusion and mining in SPHERE Challenges & Opportunities for Machine Learning Tom Diethe Intelligent Systems Laboratory University of Bristol
  2. irc-sphere.ac.uk 21st Century Healthcare Challenges UK: 1.4 million aged > 85, by 2035 -> 3.6 million Japan: will have the oldest population in human history by 2050 (52 yrs) China: a retired population larger than Europe Ageing populations living with long term health conditions: obesity, diabetes, depression, heart disease, dementia … Technological solution to fill the gap between expectations and reality of healthcare
  3. Environmental Temperature, light level, humidity, air quality Water & electricity consumption Video emotion, gate, activity, interaction Wearables activity, sleep, etc. Contextual information medical history, demographics Feedback medical practitioner, users
  4. Use Cases • Clinician with information need • Hip injury example – how is their gait, are they walking better? • Causal relations – are there any common patterns of behaviour that lead to health issues • Information back to the user – better health-enhancing choices • Early warning system • Disease progress/treatment effectiveness • Many more …
  5. 5
  6. • 100 home deployment underway
  7. What do we want to learn? 7
  8. Prediction and Modelling Who is in the house What are they doing When are activities happening Where are these happening Why does this matter? 8
  9. Is this Big Data? • Sensors • Heterogeneous • Noisy/intermittent • Different spatial/temporal resolutions • Velocity ✔ Variety ✔ volume ? 9
  10. What’s Important? • Quantification of uncertainty • Transparent models • Online Learning: models must adapt to changing habits • How to incorporate medical history • Daily/Weekly/Seasonal patterns • Personalisation 10
  11. Model-Based Machine Learning • Model uncertainty using probabilities • frequency • belief • Model contains variables and factors 1. Build a model 2. Incorporate observations 3. Perform inference over “latent” variables 11
  12. What is a Model? 12 A simulator programme bool A = random.NextDouble() > 0; bool B = random.NextDouble() > 0; bool C = A & B; C A B & • A set of assumptions 1. A coin has an equal chance of landing on heads or tails 2. Coin tosses are independent
  13. Rules of Probability 13 SUM RULE PRODUCT RULE
  14. Bayes’ Rule 14 POSTERIOR PRIOR LIKELIHOOD NORMALISER
  15. • Designed for the data-scientist • Case study based www.mbmlbook.com
  16. Experiments • SPHERE House: • Scripted Experiments • Medium Term Stays (1-7 days) • Long Stays (1-4 weeks) • Full Deployments 16
  17. 17
  18. SPHERE House Script <<<<Downstairs>>>> Living Room Enter the room and close the door behind you Stand facing the mirror and jump twice. Turn light on Go to window and open and close the curtains Take off shoes Artificial activities … repeat 5 times with 3 seconds between each stand to bend bend to stand stand to kneel kneel to stand stand to sit sit to lie (back) lie (back) to lie (side) (on sofa) lie (side) to lie (back) lie (back) to sit sit to stand cough Turn light off 18
  19. SPHERE Challenge • Task: predict posture and ambulation labels given the sensor data • Accelerometer, RGB-D and environmental data 19
  20. SPHERE Challenge • Data was collected from a script in the SPHERE house • 10 participant, ~20-30 minutes per script • Even split between training and testing data • Test data split into short 10-30s sequences https://www.drivendata.org/competitions/42/senior-data-science- safe-aging-with-sphere bit.ly/sphere-challenge 20
  21. Targets • Each sequence was annotated at least twice • Not all annotators will agree all of the time • Start/end time of annotations may not be aligned • Actual label assigned to a time interval may not agree • Task: predict mean annotation on a per-second basis — the targets are probabilistic • Also provided localisation annotations (in training only) 21
  22. 22
  23. 23
  24. 24
  25. Challenge Participation • ~ 80 teams • > 100 participants • ~ 400 total registrants • > 770 individual submissions 25
  26. 26
  27. Challenges 1 & 2 Shifting Sands Humans are costly 27
  28. Goal: Activity Recognition in Smart Homes Deployment context differs from learning context (home/resident) TRANSFER LEARNING Labels costly and time-consuming to acquire ACTIVE LEARNING 28
  29. 29 ONLINE ACTIVE + TRANSFER + TRANSFER
  30. Method • Extension of the “Bayes Point Machine” • Additional layer of hierarchy: • model “shared” and “individual” weights • can smoothly evolve from generic to personalised predictions • Implemented using Infer.NET • http://research.microsoft.com/en- us/um/cambridge/projects/infernet/ 30
  31. irc-sphere.ac.uk Accelerometer Data • Source: 30 subjects, Smartphone, 50Hz, Video annotations • Target: 14 subjects, MotionNode, 100Hz, Observer annotations • Classes: Walking upstairs vs. Walking downstairs • Features: 48 features based on the ‘body’ acceleration signal 31 https://archive.ics.uci.edu/ml/datasets/Human+Activity+Recognition+ Using+Smartphones http://sipi.usc.edu/HAD/
  32. 32 Online Learning Active Transfer Learning VOI method fails to out-perform US
  33. irc-sphere.ac.uk • Bayesian framework very appealing • Transfer boosts initial accuracy to 70% • Active Learning -> ~5 instances to personalise Diethe, T., Twomey, N. and Flach, P., 2016. Active transfer learning for activity recognition. ESANN Summary
  34. Challenge 3: Where did I put my sensor? 34
  35. Unsupervised learning of sensor topologies • signal processing and information-theoretic techniques • learn an adjacency matrix • enables us to determine combinations of sensors useful for classification • Experiments using CASAS data: http://ailab.wsu.edu/casas/datasets/ 35
  36. irc-sphere.ac.uk Modelling of CASAS datasets • Experiments based on dataset 11 (Kyoto Daily Life 2010) • Existing methods: • Naïve Bayes, HMM, CRF (segmented data) Krishnan & Cook 2012 • SVM, Decision trees (streaming data) Cook 2012 • SOTA: ~80-90% accuracy in controlled environments, some transfer learning • Two approaches: • Undirected models: Further work using Linear Chain CRFs • Directed models: Online Bayesian classifiers
  37. Dataset: CASAS twor2009 37
  38. 38
  39. Results • 5-10% boost in classification performance • Can help • when transferring to new sensor configuration • disambiguating multiple residents Twomey, N., Diethe, T., Craddock, I. and Flach, P., 2016. Unsupervised learning of sensor topologies for improving activity recognition in smart environments. Neurocomputing. 39
  40. Challenge 4 What to compute, and when? 40
  41. HyperStream • Software for streaming data • High-level interfaces • Complex interlinked workflows • Online and offline execution modes https://github.com/IRC-SPHERE/HyperStream • Diethe, T., Twomey, N., Kull, M., Sokol, K., Song, H., Tonkin, E., & Flach, P.. (2017). IRC-SPHERE/HyperStream: First public pre-release version. Zenodo. http://doi.org/10.5281/zenodo.242227 • General purpose tool • Domain-independent • “Compute-on-request”
  42. Challenge 5 Tick-tock-tick-tock 42
  43. Circular Statistics 43
  44. Further Challenges Opportunities! 44
  45. Opportunities • True house-to-house transfer learning • How well does this all work with multiple residents • What happens when houses or people change/move? • Complex activities • Sleep • Real medical applications 45
  46. Summing up … 46
  47. 47
  48. Resources • SPHERE Code: • https://github.com/IRC-SPHERE • SPHERE Challenge Dataset • http://irc-sphere.ac.uk/sphere-challenge/home • http://bit.ly/sphere-challenge • Infer.NET • http://research.microsoft.com/en-us/um/cambridge/projects/infernet/ • MBML Book • http://mbmlbook.com/ 48

Editor's Notes

  1. Build a model, which if we’re taking the Bayesian approach is a joint distribution over the relevant variables. This can be represented as a graph. Incorporate observations. Run inference, which again in the Bayesian approach means computing the distributions over the desired variables. In real-time applications we iterate over the 2nd & 3rd steps, and extend the model as required.
  2. ----- Meeting Notes (09/09/2014 14:37) ----- - Coming to the end of year 1, working on setting things up. Names of the team members. - We're hiring. Visits. Collaboration. Who wants this? Nobody wants it for themselves, but everyone wants it for someone else Cut out 1 slide Cliniciian with information need Hip injury example – how it their gait, are they walking better? Causal relations – what causes illness Information back to the user – better health-enhancing choices Early warning system Disease progress/treatment effectiveness Embed video New “House” graphic to replace existing & WP slide Differing temporal/spatial resolutions
Advertisement