Starting your AI/ML project right (May 2020)Julien SIMON
In this talk, we’ll see how you can put your AI/ML project on the right track from the get-go. Applying common sense and proven best practices, we’ll discuss skills, tools, methods, and more. We’ll also look at several real-life projects built by AWS customers in different industries and startups.
Optimize your Machine Learning Workloads on AWS (July 2019)Julien SIMON
Talk at Floor 28, Tel Aviv.
Infrastructure, tips to speed up training, hyperparameter optimization, model compilation, Amazon SageMaker Neo, cost optimization, Amazon Elastic Inference
Starting your AI/ML project right (May 2020)Julien SIMON
In this talk, we’ll see how you can put your AI/ML project on the right track from the get-go. Applying common sense and proven best practices, we’ll discuss skills, tools, methods, and more. We’ll also look at several real-life projects built by AWS customers in different industries and startups.
Optimize your Machine Learning Workloads on AWS (July 2019)Julien SIMON
Talk at Floor 28, Tel Aviv.
Infrastructure, tips to speed up training, hyperparameter optimization, model compilation, Amazon SageMaker Neo, cost optimization, Amazon Elastic Inference
End to End Model Development to Deployment using SageMakerAmazon Web Services
End to End Model Development to Deployment Using SageMaker
In this session we would be developing a model for image classification model (a convolutional neural network, or CNN). We would start off with some theory about CNNs, explore how they learn an image and then proceed towards hands-on lab. We would be using Amazon SageMaker to develop the model in Python, train the model and then to finally create an endpoint and run inference against it. We would be using a custom Conda Kernel for this exercise and would be looking at leveraging SageMaker features like LifeCycle Configurations to help us prepare the notebook before launch. Finally we would be deploying the model in production and run inference against it. We would also be able to monitor various parameters for endpoint performance such as endpoint’s CPU/Memory and Model inference performance metrics.
Level: 200-300
Using Amazon SageMaker to build, train, & deploy your ML ModelsAmazon Web Services
Machine Learning Workshops at the San Francisco Loft
Build, Train, and Deploy ML Models Using SageMaker
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Level: 200-300
Speaker: Martin Schade - R&D Engineer, AWS Solutions Architecture
Amazon SageMaker is a fully-managed platform that lets developers and data scientists build and scale machine learning solutions. First, we'll show you how SageMaker Ground Truth helps you label large training datasets. Then, using Jupyter notebooks, we'll show you how to build, train and deploy models using built-in algorithms and frameworks (TensorFlow, Apache MXNet, etc). Finally, we'll show you how to use 3rd-party models from the AWS marketplace.
Working with Amazon SageMaker Algorithms for Faster Model TrainingAmazon Web Services
Amazon SageMaker is a fully-managed service that enables developers and data scientists to quickly and easily build, train, and deploy machine learning (ML) models, at any scale. Amazon SageMaker provides high-performance, machine learning algorithms optimized for speed, scale, and accuracy, to perform training on petabyte-scale data sets. This webinar will introduce you to the collection of distributed streaming ML algorithms that come with Amazon SageMaker. You will learn about the difference between streaming and batch ML algorithms, and how SageMaker has been architected to run these algorithms at scale. We will demo Neural Topic Modeling of text documents using a sample SageMaker Notebook, which will be made available to attendees.
Using Amazon SageMaker to Build, Train, and Deploy Your ML ModelsAmazon Web Services
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Using Amazon SageMaker to build, train, and deploy your ML ModelsAmazon Web Services
by Gitansh Chadha, Solutions Architect AWS
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Build Deep Learning Applications with TensorFlow & SageMakerAmazon Web Services
Build Deep Learning Applications with TensorFlow and SageMaker
Deep learning continues to push the state of the art in domains such as computer vision, natural language understanding and recommendation engines. In this workshop, you’ll learn how to get started with the TensorFlow deep learning framework using Amazon SageMaker, a platform to easily build, train and deploy models at scale. You’ll learn how to build a model using TensorFlow by setting up a Jupyter notebook to get started with image and object recognition. You’ll also learn how to quickly train and deploy a model through Amazon SageMaker.
Level: 200-300
Speakers:
Martin Schade - R&D Engineer, AWS Solutions Architecture
Steve Sedlmeyer - Sr. Solutions Architect, World Wide Public Sector, AWS
Costruisci modelli di Machine Learning con Amazon SageMaker AutopilotAmazon Web Services
Amazon SageMaker AutoPilot è una funzionalità di Amazon SageMaker che crea automaticamente il miglior modello di apprendimento automatico per il tuo set di dati. Con SageMaker Autopilot, si fornisce un set di dati tabellare e si seleziona la variabile target da prevedere, che può essere numerica o categorica. SageMaker Autopilot esplorerà automaticamente diverse soluzioni per trovare il modello migliore. È quindi possibile distribuire direttamente il modello in produzione con un solo clic o esplorare le soluzioni consigliate con Amazon SageMaker Studio per migliorare ulteriormente la qualità del modello. In questo webinar approfondiremo questa capacità, con dimostrazioni pratiche su come utilizzare il servizio.
Build, Train, and Deploy Machine Learning for the Enterprise with Amazon Sage...Amazon Web Services
Machine learning (ML) is rapidly being adopted by enterprises, enabling them to be nimble and align technical solutions to solve real-world business problems. ML use cases include diagnosis and research in healthcare, financial fraud detection, natural language processing (NLU), and accurate statistics in sports. Amazon SageMaker is a fully managed platform that enables developers to build, train, and deploy enterprise-scale ML models quickly and easily. In this workshop, we build an ML model using Amazon SageMaker’s built-in algorithms and frameworks. We train the model to achieve a high level of accurate predictions, then we deploy the model in production to achieve best results. Gain an understanding of how Amazon SageMaker removes the complexity and barriers to use and deploy ML models.
End to End Model Development to Deployment using SageMakerAmazon Web Services
End to End Model Development to Deployment Using SageMaker
In this session we would be developing a model for image classification model (a convolutional neural network, or CNN). We would start off with some theory about CNNs, explore how they learn an image and then proceed towards hands-on lab. We would be using Amazon SageMaker to develop the model in Python, train the model and then to finally create an endpoint and run inference against it. We would be using a custom Conda Kernel for this exercise and would be looking at leveraging SageMaker features like LifeCycle Configurations to help us prepare the notebook before launch. Finally we would be deploying the model in production and run inference against it. We would also be able to monitor various parameters for endpoint performance such as endpoint’s CPU/Memory and Model inference performance metrics.
Level: 200-300
Using Amazon SageMaker to build, train, & deploy your ML ModelsAmazon Web Services
Machine Learning Workshops at the San Francisco Loft
Build, Train, and Deploy ML Models Using SageMaker
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Level: 200-300
Speaker: Martin Schade - R&D Engineer, AWS Solutions Architecture
Amazon SageMaker is a fully-managed platform that lets developers and data scientists build and scale machine learning solutions. First, we'll show you how SageMaker Ground Truth helps you label large training datasets. Then, using Jupyter notebooks, we'll show you how to build, train and deploy models using built-in algorithms and frameworks (TensorFlow, Apache MXNet, etc). Finally, we'll show you how to use 3rd-party models from the AWS marketplace.
Working with Amazon SageMaker Algorithms for Faster Model TrainingAmazon Web Services
Amazon SageMaker is a fully-managed service that enables developers and data scientists to quickly and easily build, train, and deploy machine learning (ML) models, at any scale. Amazon SageMaker provides high-performance, machine learning algorithms optimized for speed, scale, and accuracy, to perform training on petabyte-scale data sets. This webinar will introduce you to the collection of distributed streaming ML algorithms that come with Amazon SageMaker. You will learn about the difference between streaming and batch ML algorithms, and how SageMaker has been architected to run these algorithms at scale. We will demo Neural Topic Modeling of text documents using a sample SageMaker Notebook, which will be made available to attendees.
Using Amazon SageMaker to Build, Train, and Deploy Your ML ModelsAmazon Web Services
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Using Amazon SageMaker to build, train, and deploy your ML ModelsAmazon Web Services
by Gitansh Chadha, Solutions Architect AWS
Amazon SageMaker is a fully-managed service that enables data scientists and developers to quickly and easily build, train, and deploy machine learning models, at scale. This session will introduce you the features of Amazon SageMaker, including a one-click training environment, highly-optimized machine learning algorithms with built-in model tuning, and deployment without engineering effort. With zero-setup required, Amazon SageMaker significantly decreases your training time and overall cost of building production machine learning systems.
Build Deep Learning Applications with TensorFlow & SageMakerAmazon Web Services
Build Deep Learning Applications with TensorFlow and SageMaker
Deep learning continues to push the state of the art in domains such as computer vision, natural language understanding and recommendation engines. In this workshop, you’ll learn how to get started with the TensorFlow deep learning framework using Amazon SageMaker, a platform to easily build, train and deploy models at scale. You’ll learn how to build a model using TensorFlow by setting up a Jupyter notebook to get started with image and object recognition. You’ll also learn how to quickly train and deploy a model through Amazon SageMaker.
Level: 200-300
Speakers:
Martin Schade - R&D Engineer, AWS Solutions Architecture
Steve Sedlmeyer - Sr. Solutions Architect, World Wide Public Sector, AWS
Costruisci modelli di Machine Learning con Amazon SageMaker AutopilotAmazon Web Services
Amazon SageMaker AutoPilot è una funzionalità di Amazon SageMaker che crea automaticamente il miglior modello di apprendimento automatico per il tuo set di dati. Con SageMaker Autopilot, si fornisce un set di dati tabellare e si seleziona la variabile target da prevedere, che può essere numerica o categorica. SageMaker Autopilot esplorerà automaticamente diverse soluzioni per trovare il modello migliore. È quindi possibile distribuire direttamente il modello in produzione con un solo clic o esplorare le soluzioni consigliate con Amazon SageMaker Studio per migliorare ulteriormente la qualità del modello. In questo webinar approfondiremo questa capacità, con dimostrazioni pratiche su come utilizzare il servizio.
Build, Train, and Deploy Machine Learning for the Enterprise with Amazon Sage...Amazon Web Services
Machine learning (ML) is rapidly being adopted by enterprises, enabling them to be nimble and align technical solutions to solve real-world business problems. ML use cases include diagnosis and research in healthcare, financial fraud detection, natural language processing (NLU), and accurate statistics in sports. Amazon SageMaker is a fully managed platform that enables developers to build, train, and deploy enterprise-scale ML models quickly and easily. In this workshop, we build an ML model using Amazon SageMaker’s built-in algorithms and frameworks. We train the model to achieve a high level of accurate predictions, then we deploy the model in production to achieve best results. Gain an understanding of how Amazon SageMaker removes the complexity and barriers to use and deploy ML models.
In this talk, Jayesh will walk us through AWS ML and AI offerings and how they are better solutions for most of the industries out there looking to integrate ML features in their product. Also includes demo's of services.
Amazon SageMaker is a fully managed Machine Learning service which facilitates seamless adoption of #MachineLearning across various industries! Jayesh is walking us through details of SageMaker with demo in this talk!
Building Machine Learning inference pipelines at scale | AWS Summit Tel Aviv ...Amazon Web Services
Real-life Machine Learning (ML) workloads typically require more than training and predicting: data often needs to be pre-processed and post-processed, sometimes in multiple steps. Thus, developers and data scientists have to train and deploy not just a single algorithm, but a sequence of algorithms that will collaborate in delivering predictions from raw data. In this session, we’ll first show you how to use Apache Spark MLlib to build ML pipelines, and we’ll discuss scaling options when datasets grow huge. We’ll then show how to how implement inference pipelines on Amazon SageMaker, using Apache Spark, Scikit-learn, as well as ML algorithms implemented by Amazon.
Building Machine Learning inference pipelines at scale | AWS Summit Tel Aviv ...AWS Summits
Real-life Machine Learning (ML) workloads typically require more than training and predicting: data often needs to be pre-processed and post-processed, sometimes in multiple steps. Thus, developers and data scientists have to train and deploy not just a single algorithm, but a sequence of algorithms that will collaborate in delivering predictions from raw data. In this session, we’ll first show you how to use Apache Spark MLlib to build ML pipelines, and we’ll discuss scaling options when datasets grow huge. We’ll then show how to how implement inference pipelines on Amazon SageMaker, using Apache Spark, Scikit-learn, as well as ML algorithms implemented by Amazon.
Amazon SageMaker and Chainer: Tips & Tricks (AIM329-R1) - AWS re:Invent 2018Amazon Web Services
In this session, learn how to use Chainer, an open-source deep learning framework written in Python, in the Amazon SageMaker machine learning platform.
Building State-of-the-Art Computer Vision Models Using MXNet and Gluon (AIM36...Amazon Web Services
Implementing computer vision (CV) models just got simpler and faster. In this chalk talk, learn how to implement CV models using MXNet and the Gluon CV Toolkit, which provides implementations of state-of-the-art deep learning algorithms in computer vision to help engineers, researchers, and students quickly prototype products, validate new ideas, and learn computer vision.
Build, Train, and Deploy ML Models Quickly and Easily with Amazon SageMaker, ...Amazon Web Services
Amazon SageMaker is a fully managed platform that enables developers and data scientists to quickly and easily build, train, and deploy machine learning models at any scale. Amazon SageMaker takes away the heavy lifting of machine learning, thus removing the typical barriers associated with machine learning. In this session, we'll dive deep into the technical details of each of the modules of Amazon SageMaker to showcase the capabilities of the platform. We also discuss the practical deployments of Amazon SageMaker through real-world customer examples.
Building Machine Learning models with Apache Spark and Amazon SageMaker | AWS...Amazon Web Services
Amazon SageMaker is a fully-managed platform that enables developers and data scientists to quickly and easily build, train, and deploy machine learning models at any scale. In this session, we'll show you how to combine it with Apache Spark to build efficient Machine Learning pipeline.
Building Machine Learning Inference Pipelines at Scale (July 2019)Julien SIMON
Talk at OSCON, Portland, 18/07/2019
Real-life Machine Learning applications require more than a single model. Data may need pre-processing: normalization, feature engineering, dimensionality reduction, etc. Predictions may need post-processing: filtering, sorting, combining, etc.
Our goal: build scalable ML pipelines with open source (Spark, Scikit-learn, XGBoost) and managed services (Amazon EMR, AWS Glue, Amazon SageMaker)
An introduction to computer vision with Hugging FaceJulien SIMON
In this code-level talk, Julien will show you how to quickly build and deploy computer vision applications based on Transformer models. Along the way, you'll learn about the portfolio of open source and commercial Hugging Face solutions, and how they can help you deliver high-quality solutions faster than ever before.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
3. Agenda
1. Welcome & housekeeping
2. An introduction to Automatic Model Tuning (AMT) and AutoML
3. Labs
4. Wrap-up and clean-up
What you’ll learn today
• How to use AMT to find optimal model hyperparameters
• How to use AMT to explore deep learning architectures
• How to use Amazon SageMaker Autopilot to find the optimal algorithm, data preprocessing steps and hyper
parameters
4. Our team today
• Antje
• Chris
• Srikanth
• Wei
• Marc
• Michael E
• Matt
• Mike
• Guillaume
• Michael M
• Frank
• Shashank
• John
• Abhi
• Navjot
• Bo
• Boaz
• Mohamed
5. Housekeeping
• Please be a good neighbor ☺
• Turn off network backups and any network-hogging app
• Switch your phones to silent mode
• Help the people around you if you can
• Don’t stay blocked. Ask questions!
7. Hyperparameters
Neural Networks
Number of layers
Hidden layer width
Learning rate
Embedding
dimensions
Dropout
…
XGBoost
Tree depth
Max leaf nodes
Gamma
Eta
Lambda
Alpha
…
8. Tactics to find the optimal set of hyperparameters
• Manual Search: ”I know what I’m doing”
• Grid Search: “X marks the spot”
Typically training hundreds of models
Slow and expensive
• Random Search: “Spray and pray”
« Random Search for Hyper-Parameter Optimization », Bergstra & Bengio, 2012
Works better and faster than Grid Search
But… but… but… it’s random!
• Hyperparameter Optimization: use ML to predict hyperparameters
Training fewer models
Gaussian Process Regression and Bayesian Optimization
https://docs.aws.amazon.com/en_pv/sagemaker/latest/dg/automatic-model-tuning-how-it-works.html
9. Setting hyperparameters in Amazon SageMaker
• Built-in algorithms
• Python parameters for the relevant estimator (KMeans, LinearLearner, etc.)
• Built-in frameworks
• hyperparameters parameter for the relevant estimator (TensorFlow, MXNet, etc.)
• This must be a Python dictionary
tf_estimator = TensorFlow(…, hyperparameters={'epochs’: 1, ‘lr’: ‘0.01’})
• Your code must be able to accept them as command-line arguments (script mode)
• Bring your own container
• hyperparameters parameter for Estimator
• This must be Python dictionary
• It’s copied inside the container: /opt/ml/input/config/hyperparameters.json
10. Automatic Model Tuning in Amazon SageMaker
1. Define an Estimator the normal way
2. Define the metric to tune on
• Pre-defined metrics for built-in algorithms and frameworks
• Or anything present in the training log, provided that you pass a regular expression for it
3. Define parameter ranges to explore
• Type: categorical (avoid if possible), integer, continuous (aka floating point)
• Range
• Scaling: linear (default), logarithmic, reverse logarithmic
4. Create an HyperparameterTuner
• Estimator, metric, parameters, total number of jobs, number of jobs in parallel
• Strategy: bayesian (default), or random search
5. Launch the tuning job with fit()
11. Workflow
Training JobHyperparameter
Tuning Job
Tuning strategy
Objective
metrics
Training Job
Training Job
Training Job
Clients
(console, notebook, IDEs, CLI)
model name
model1
model2
…
objective
metric
0.8
0.75
…
eta
0.07
0.09
…
max_depth
6
5
…
…
12. Automatic Model Tuning in Amazon SageMaker
• You can view ongoing tuning jobs in the AWS console
• List of training jobs
• Best training job
• You can also query their status with the SageMaker SDK
• Calling deploy() on the HyperparameterTuner deploys the best job
• The best job so far if the tuning job has not yet completed
13. Tips
• Use the bayesian strategy for better, faster, cheaper results
• Most customers use random search as a baseline, to check that bayesian performs better
• Don’t run too many jobs in parallel
• This gives the bayesian strategy fewer opportunities to predict
• Instance limits!
• Don’t run too many jobs
• Bayesian typically requires 10x fewer jobs than random
• Cost!
14. Resources on Automatic Model Tuning
Documentation
https://docs.aws.amazon.com/sagemaker/latest/dg/automatic-model-tuning.html
https://sagemaker.readthedocs.io/en/stable/tuner.html
Notebooks
https://github.com/awslabs/amazon-sagemaker-examples/tree/master/hyperparameter_tuning
Blog posts
https://aws.amazon.com/blogs/aws/sagemaker-automatic-model-tuning/
https://aws.amazon.com/blogs/machine-learning/amazon-sagemaker-automatic-model-tuning-produces-better-models-faster/
https://aws.amazon.com/blogs/machine-learning/amazon-sagemaker-automatic-model-tuning-now-supports-early-stopping-of-
training-jobs/
https://aws.amazon.com/blogs/machine-learning/amazon-sagemaker-automatic-model-tuning-becomes-more-efficient-with-warm-
start-of-hyperparameter-tuning-jobs/
https://aws.amazon.com/blogs/machine-learning/amazon-sagemaker-automatic-model-tuning-now-supports-random-search-and-
hyperparameter-scaling/
16. AutoML
• AutoML aims at automating the process of building a model
• Problem identification: looking at the data set, what class of problem are we trying to solve?
• Algorithm selection: which algorithm is best suited to solve the problem?
• Data preprocessing: how should data be prepared for best results?
• Hyperparameter tuning: what is the optimal set of training parameters?
• Black box vs. white box
• Black box: the best model only
→ Hard to understand the model, impossible to reproduce it manually
• White box: the best model, other candidates, full source code for preprocessing and training
→ See how the model was built, and keep tweaking for extra performance
17. AutoML with Amazon SageMaker Autopilot
• SageMaker Autopilot covers all steps
• Problem identification: looking at the data set, what class of problem are we trying to solve?
• Algorithm selection: which algorithm is best suited to solve the problem?
• Data preprocessing: how should data be prepared for best results?
• Hyperparameter tuning: what is the optimal set of training parameters?
• Autopilot is white box AutoML
• You can understand how the model was built, and you can keep tweaking
• Supported algorithms at launch:
Linear Learner, Factorization Machines, KNN, XGBoost
18. AutoML with Amazon SageMaker Autopilot
1. Upload the unprocessed dataset to S3
2. Configure the AutoML job
• Location of dataset
• Completion criteria
3. Launch the job
4. View the list of candidates and the autogenerated notebook
5. Deploy the best candidate to a real-time endpoint, or use batch
transform
20. Labs
1. Use AMT to find optimal model hyperparameters for XGBoost
2. Use Autopilot to find the optimal algo, preprocessing steps and
hyper parameters
3. Use AMT to explore deep learning architectures on Keras
https://gitlab.com/juliensimon/aim361