Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

Shipping Machine Learning to production

164 views

Published on

It has been a long time since the Turing test laid down the first foundations of the philosophy of Artificial Intelligence (AI). AI however has only recently reached the mainstream market, with applications in the industrial and consumer markets. This talk will take a look at a few specific techniques and architectures that help reduce the time and complexity of iterations when using Machine Learning algorithms to deliver value to end users using Big Data. As the cost of storing data today is rather cheap, companies usually store a huge volume. Consequently, the cost of training Machine Learning models to serve clients and the public in general is high because it takes time and uses up a lot of resources due to the high amount of data. To avoid this, companies need a distributed environment and speedup techniques to reduce the time and cost of training – for example, a distributed processing framework such as Spark. But once the model has been validated, the requirements to bring it to production are different. The resources required to provide predictions to the user do not depend on the model itself, as it has already been trained, but on the business part of the company and the number of users that need to receive information. A more appropriate architecture for this purpose may be more similar to a microservice than to a distributed processing framework. Thus different solutions should be used for both stages: while Apache Spark should be used for the former, a microservice should be used for the latter. This would allow companies to easily and rapidly iterate over models to keep them updated. This is indeed another challenge: As new data enters continuously, how can the model in production be updated at the same time, including both the historic and new data gathered? Model training uses a lot of resources and is a time-consuming operation due to the large amount of data to be processed. This training process cannot be launched from scratch every time the model has to be updated with new data, so different solutions are needed – and this goes for all situations in which a same trained model can be used. For example, training a model to recognize cars in images could also then be used as a basis for recognizing other types of vehicles.
By Raúl de la Fuente and Mateo Álvarez

Published in: Technology
  • Be the first to comment

Shipping Machine Learning to production

  1. 1. 0 Shipping ML to production: From Data Scientist to End User Mateo �lvarez Ra�l de la Fuente @neurozetta
  2. 2. Experience thepresentation xlic.es/v/2C8064
  3. 3. Shipping faster
  4. 4. ML cycle
  5. 5. QUESTION ML cycle
  6. 6. HYPOTHESISQUESTION ML cycle
  7. 7. HYPOTHESIS DEPLOY ML MODEL QUESTION ML cycle
  8. 8. HYPOTHESIS DEPLOY ML MODEL QUESTION ML cycle DATA
  9. 9. Which phase makes the most to ship faster? a) Form a HYPOTHESIS b) TRAIN a model c) EVOLVE and iterate a model d) DEPLOY a model
  10. 10. Inspired by Human Learning Learn from examples Retain learned knowledge Reusable in other tasks
  11. 11. HYPOTHESIS DEPLOY ML MODEL QUESTION ML cycle DATA + KNOWLEDGE
  12. 12. Knowledge Source Domain Target Domain
  13. 13. Knowledge is not static! Source Domain Target Domain
  14. 14. Size dataset Compute price Hyperparameter set up Limitations
  15. 15. Transfer Learning Data Knowledge
  16. 16. Knowledge
  17. 17. Knowledge Similar Domain / Differents Tasks
  18. 18. Knowledge Similar Domain / Differents Tasks Differents Domains / Same Task
  19. 19. Target Data Fine Tuning Multitasking Domain Adaptation Zero-shot Labelled Transfer DL Methods Knowledge Labelled Source Data Unlabelled
  20. 20. I KNOW KUNG FU
  21. 21. WAX ON WAX OFF
  22. 22. no transfer transfer Impact performance training
  23. 23. no transfer transfer Impact Shipping fasterperformance training
  24. 24. HYPOTHESIS DEPLOY ML MODEL QUESTION ML cycle
  25. 25. Deploy Prediction, please! Backend Spark Scikit Learn Tensor Flow
  26. 26. Deploy Prediction, please! Backend Microservice (Combust-ML) Spark Scikit Learn Tensor Flow
  27. 27. Deploy DCOS Stratio Intelligence Model Repository Model Microservice Prediction, please!
  28. 28. Thx! Those who don't remember the past are condemned to repeat it Ra�l de la Fuente @neurozetta Mateo Alvarez

×