This document discusses scaling machine learning models from initial development to production deployment for millions of users. It outlines several options for scaling models from a single instance to large distributed systems, including using Amazon EC2 instances with automation, Docker clusters on ECS/EKS, or the fully managed SageMaker service. SageMaker is recommended for ease of scaling training and inference with minimal infrastructure management required.
Building Machine Learning Inference Pipelines at Scale (July 2019)Julien SIMON
Talk at OSCON, Portland, 18/07/2019
Real-life Machine Learning applications require more than a single model. Data may need pre-processing: normalization, feature engineering, dimensionality reduction, etc. Predictions may need post-processing: filtering, sorting, combining, etc.
Our goal: build scalable ML pipelines with open source (Spark, Scikit-learn, XGBoost) and managed services (Amazon EMR, AWS Glue, Amazon SageMaker)
Managing Database Indexes: A Data-Driven Approach - Amadeus MagrabiAmadeus Magrabi
Talk at the Open Data Science Conference East 2020.
Abstract:
Database indexes can make or break the performance of a database. Efficient indexes need to be tailored to the specific queries that are sent to a database, but since query patterns can vary a lot and change over time, it is often a painful process to manually manage indexes. In this session, I will talk about our data-driven approach to automatically estimate optimal indexes from log data of our MongoDB databases. You will learn how we use Google Cloud Functions to stream log data from Stackdriver to Google BigQuery, how we use BigQuery to scale our data analysis, and how we use Python’s Jupyter Notebooks to visualize and monitor our results.
Building Machine Learning Inference Pipelines at Scale (July 2019)Julien SIMON
Talk at OSCON, Portland, 18/07/2019
Real-life Machine Learning applications require more than a single model. Data may need pre-processing: normalization, feature engineering, dimensionality reduction, etc. Predictions may need post-processing: filtering, sorting, combining, etc.
Our goal: build scalable ML pipelines with open source (Spark, Scikit-learn, XGBoost) and managed services (Amazon EMR, AWS Glue, Amazon SageMaker)
Managing Database Indexes: A Data-Driven Approach - Amadeus MagrabiAmadeus Magrabi
Talk at the Open Data Science Conference East 2020.
Abstract:
Database indexes can make or break the performance of a database. Efficient indexes need to be tailored to the specific queries that are sent to a database, but since query patterns can vary a lot and change over time, it is often a painful process to manually manage indexes. In this session, I will talk about our data-driven approach to automatically estimate optimal indexes from log data of our MongoDB databases. You will learn how we use Google Cloud Functions to stream log data from Stackdriver to Google BigQuery, how we use BigQuery to scale our data analysis, and how we use Python’s Jupyter Notebooks to visualize and monitor our results.
Amazon SageMaker is a fully-managed platform that lets developers and data scientists build and scale machine learning solutions. First, we'll show you how SageMaker Ground Truth helps you label large training datasets. Then, using Jupyter notebooks, we'll show you how to build, train and deploy models using built-in algorithms and frameworks (TensorFlow, Apache MXNet, etc). Finally, we'll show you how to use 3rd-party models from the AWS marketplace.
Integrating Deep Learning into your Enterprise
In this workshop we return to one of the popular Machine Learning Framework - scikit-learn. We scikit-learn's decision tree classifier to train the model. Decision Trees (DTs) are a non-parametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target variable by learning simple decision rules inferred from the data features. We follow the whole machine learning pipeline from algorithm selection, training and finally deployment of an endpoint. We would be working with the widely available Iris dataset and the endpoint would be predicting what species the sample belongs to from the Sepal width and length, Petal width and length. Through this workshop we would know all the internal details of how we use containers to train and deploy our machine learning workloads.
Level: 300-400
AWS Machine Learning Week SF: End to End Model Development Using SageMakerAmazon Web Services
AWS Machine Learning Week at the San Francisco Loft: End to End Model Development Using SageMaker
In this session we would be developing a model for image classification model (a convolutional neural network, or CNN). We would start off with some theory about CNNs, explore how they learn an image and then proceed towards hands-on lab. We would be using Amazon SageMaker to develop the model in Python, train the model and then to finally create an endpoint and run inference against it. We would be using a custom Conda Kernel for this exercise and would be looking at leveraging SageMaker features like LifeCycle Configurations to help us prepare the notebook before launch. Finally we would be deploying the model in production and run inference against it. We would also be able to monitor various parameters for endpoint performance such as endpoint’s CPU/Memory and Model inference performance metrics.
Presenter: Kris Skrinak
Amazon SageMaker is a fully managed Machine Learning service which facilitates seamless adoption of #MachineLearning across various industries! Jayesh is walking us through details of SageMaker with demo in this talk!
AWS ML and SparkML on EMR to Build Recommendation Engine Amazon Web Services
Machine Learning
A managed supervised learning environment to build different models, including Binary Classification / Multi-class classification / Regression ML. The demos will show a dataset of banking customers with demographics, predicting the likelihood of whether they are going to default using binary classification. Second one will be predicting a UK bike rental shop traffic using linear regression, and third one for predicting a rainforest soil type using multi-class classification.
Benefits: Managed and on-demand environment for supervised learning algorithm, available as batch processing or real-time API.
Spark ML Cluster
Running spark on AWS managed cluster, storing data on HDFS / S3 persistent storage, modules include MLib and Zeppelin (Web Notebook), to build a movie recommendation engine based on “Collaborative Filtering”. The dataset contains 10M ratings provided by grouplens from MovieLens website.
Benefits: Fully managed clusters, with HA, Scalability, Elasticity and Spot instance pricing
AWS re:Invent 2016: Transforming Industrial Processes with Deep Learning (MAC...Amazon Web Services
Deep learning has revolutionized computer vision by significantly increasing the accuracy of recognition systems. This session will discuss how the Amazon Fulfillment Technologies Computer Vision Research team has harnessed deep learning to identify inventory defects in Amazon’s warehouses. Beginning with a brief overview of how orders on Amazon.com are fulfilled, the talk will describe a combination of hardware and software that uses computer vision and deep learning that visually examine bins of Amazon inventory to locate possible mismatches between the physical inventory and inventory records. With the growth of deep learning, the emphasis of new system design shifts from clever algorithms to innovative ways to harness available data.
Costruisci modelli di Machine Learning con Amazon SageMaker AutopilotAmazon Web Services
Amazon SageMaker AutoPilot è una funzionalità di Amazon SageMaker che crea automaticamente il miglior modello di apprendimento automatico per il tuo set di dati. Con SageMaker Autopilot, si fornisce un set di dati tabellare e si seleziona la variabile target da prevedere, che può essere numerica o categorica. SageMaker Autopilot esplorerà automaticamente diverse soluzioni per trovare il modello migliore. È quindi possibile distribuire direttamente il modello in produzione con un solo clic o esplorare le soluzioni consigliate con Amazon SageMaker Studio per migliorare ulteriormente la qualità del modello. In questo webinar approfondiremo questa capacità, con dimostrazioni pratiche su come utilizzare il servizio.
Build Text Analytics Solutions with Amazon Comprehend and Amazon TranslateAmazon Web Services
by Pratap Ramamurthy, Partner Solutions Architect, AWS
Natural language holds a wealth of information like user sentiment and conversational intent. In this session, we'll demonstrate the capabilities of Amazon Comprehend, a natural language processing (NLP) service that uses machine learning to find insights and relationships in text. We'll show you how to build a VOC (Voice of the Customer) application and integrate it with other AWS services including AWS Lambda, Amazon S3, Amazon Athena, Amazon QuickSight, and Amazon Translate. We’ll also show you additional methods for NLP available through Amazon Sagemaker.
"Scaling ML from 0 to millions of users", Julien Simon, AWS Dev Day Kyiv 2019Provectus
AWS Dev Day Kyiv 2019
Track: Analytics & Machine Learning
Session: ""Scaling ML from 0 to millions of users""
Speaker: Julien Simon, Global AI & Machine Learning Evangelist at AWS
Level: 300
AWS Dev Day is a free, full-day technical event where new developers will learn about some of the hottest topics in cloud computing, and experienced developers can dive deep on newer AWS services.
Provectus has organized AWS Dev Day Kyiv in close collaboration with Amazon Web Services: 800+ participants, 18 sessions, 3 tracks, a really AWSome Day!
Now, together with Zeo Alliance, we're building and nurturing AWS User Group Ukraine — join us on Facebook to stay updated about cloud technologies and AWS services: https://www.facebook.com/groups/AWSUserGroupUkraine
Video: https://www.youtube.com/watch?v=N73u1mx9DqY
Julien Simon "Scaling ML from 0 to millions of users"Fwdays
In this session, I'll show you how to scale machine learning workloads using containers on AWS (Deep Learning AMI and containers, ECS, EKS, SageMaker). We'll discuss the pros and cons of these different services from a technical, operational and cost perspective. Of course, we'll run some demos :)
Amazon SageMaker is a fully-managed platform that lets developers and data scientists build and scale machine learning solutions. First, we'll show you how SageMaker Ground Truth helps you label large training datasets. Then, using Jupyter notebooks, we'll show you how to build, train and deploy models using built-in algorithms and frameworks (TensorFlow, Apache MXNet, etc). Finally, we'll show you how to use 3rd-party models from the AWS marketplace.
Integrating Deep Learning into your Enterprise
In this workshop we return to one of the popular Machine Learning Framework - scikit-learn. We scikit-learn's decision tree classifier to train the model. Decision Trees (DTs) are a non-parametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target variable by learning simple decision rules inferred from the data features. We follow the whole machine learning pipeline from algorithm selection, training and finally deployment of an endpoint. We would be working with the widely available Iris dataset and the endpoint would be predicting what species the sample belongs to from the Sepal width and length, Petal width and length. Through this workshop we would know all the internal details of how we use containers to train and deploy our machine learning workloads.
Level: 300-400
AWS Machine Learning Week SF: End to End Model Development Using SageMakerAmazon Web Services
AWS Machine Learning Week at the San Francisco Loft: End to End Model Development Using SageMaker
In this session we would be developing a model for image classification model (a convolutional neural network, or CNN). We would start off with some theory about CNNs, explore how they learn an image and then proceed towards hands-on lab. We would be using Amazon SageMaker to develop the model in Python, train the model and then to finally create an endpoint and run inference against it. We would be using a custom Conda Kernel for this exercise and would be looking at leveraging SageMaker features like LifeCycle Configurations to help us prepare the notebook before launch. Finally we would be deploying the model in production and run inference against it. We would also be able to monitor various parameters for endpoint performance such as endpoint’s CPU/Memory and Model inference performance metrics.
Presenter: Kris Skrinak
Amazon SageMaker is a fully managed Machine Learning service which facilitates seamless adoption of #MachineLearning across various industries! Jayesh is walking us through details of SageMaker with demo in this talk!
AWS ML and SparkML on EMR to Build Recommendation Engine Amazon Web Services
Machine Learning
A managed supervised learning environment to build different models, including Binary Classification / Multi-class classification / Regression ML. The demos will show a dataset of banking customers with demographics, predicting the likelihood of whether they are going to default using binary classification. Second one will be predicting a UK bike rental shop traffic using linear regression, and third one for predicting a rainforest soil type using multi-class classification.
Benefits: Managed and on-demand environment for supervised learning algorithm, available as batch processing or real-time API.
Spark ML Cluster
Running spark on AWS managed cluster, storing data on HDFS / S3 persistent storage, modules include MLib and Zeppelin (Web Notebook), to build a movie recommendation engine based on “Collaborative Filtering”. The dataset contains 10M ratings provided by grouplens from MovieLens website.
Benefits: Fully managed clusters, with HA, Scalability, Elasticity and Spot instance pricing
AWS re:Invent 2016: Transforming Industrial Processes with Deep Learning (MAC...Amazon Web Services
Deep learning has revolutionized computer vision by significantly increasing the accuracy of recognition systems. This session will discuss how the Amazon Fulfillment Technologies Computer Vision Research team has harnessed deep learning to identify inventory defects in Amazon’s warehouses. Beginning with a brief overview of how orders on Amazon.com are fulfilled, the talk will describe a combination of hardware and software that uses computer vision and deep learning that visually examine bins of Amazon inventory to locate possible mismatches between the physical inventory and inventory records. With the growth of deep learning, the emphasis of new system design shifts from clever algorithms to innovative ways to harness available data.
Costruisci modelli di Machine Learning con Amazon SageMaker AutopilotAmazon Web Services
Amazon SageMaker AutoPilot è una funzionalità di Amazon SageMaker che crea automaticamente il miglior modello di apprendimento automatico per il tuo set di dati. Con SageMaker Autopilot, si fornisce un set di dati tabellare e si seleziona la variabile target da prevedere, che può essere numerica o categorica. SageMaker Autopilot esplorerà automaticamente diverse soluzioni per trovare il modello migliore. È quindi possibile distribuire direttamente il modello in produzione con un solo clic o esplorare le soluzioni consigliate con Amazon SageMaker Studio per migliorare ulteriormente la qualità del modello. In questo webinar approfondiremo questa capacità, con dimostrazioni pratiche su come utilizzare il servizio.
Build Text Analytics Solutions with Amazon Comprehend and Amazon TranslateAmazon Web Services
by Pratap Ramamurthy, Partner Solutions Architect, AWS
Natural language holds a wealth of information like user sentiment and conversational intent. In this session, we'll demonstrate the capabilities of Amazon Comprehend, a natural language processing (NLP) service that uses machine learning to find insights and relationships in text. We'll show you how to build a VOC (Voice of the Customer) application and integrate it with other AWS services including AWS Lambda, Amazon S3, Amazon Athena, Amazon QuickSight, and Amazon Translate. We’ll also show you additional methods for NLP available through Amazon Sagemaker.
"Scaling ML from 0 to millions of users", Julien Simon, AWS Dev Day Kyiv 2019Provectus
AWS Dev Day Kyiv 2019
Track: Analytics & Machine Learning
Session: ""Scaling ML from 0 to millions of users""
Speaker: Julien Simon, Global AI & Machine Learning Evangelist at AWS
Level: 300
AWS Dev Day is a free, full-day technical event where new developers will learn about some of the hottest topics in cloud computing, and experienced developers can dive deep on newer AWS services.
Provectus has organized AWS Dev Day Kyiv in close collaboration with Amazon Web Services: 800+ participants, 18 sessions, 3 tracks, a really AWSome Day!
Now, together with Zeo Alliance, we're building and nurturing AWS User Group Ukraine — join us on Facebook to stay updated about cloud technologies and AWS services: https://www.facebook.com/groups/AWSUserGroupUkraine
Video: https://www.youtube.com/watch?v=N73u1mx9DqY
Julien Simon "Scaling ML from 0 to millions of users"Fwdays
In this session, I'll show you how to scale machine learning workloads using containers on AWS (Deep Learning AMI and containers, ECS, EKS, SageMaker). We'll discuss the pros and cons of these different services from a technical, operational and cost perspective. Of course, we'll run some demos :)
- Overview of a use case - Sentiment analysis
- Introduction - Using Jupyter Notebook & AWS SageMaker
- Setup New Project
- Setup and Run the Build CI/CD Pipeline
- Setup the Release Pipeline
- Test Build and Release Pipelines
- Testing the deployed solution
- Examining deployed model performance
Setting up custom machine learning environments on AWS - AIM204 - Chicago AWS...Amazon Web Services
Sometimes, you might need to set up your own deep learning environments for domain-specific performance optimization and integration with custom applications. AWS offers prepackaged, optimized Amazon Machine Images (AMIs) and Docker container images that make it easy to quickly deploy these custom environments by letting you skip the complicated process of building and optimizing your environments from scratch. In this session, you learn about how to use AWS Deep Learning AMIs and AWS Deep Learning Containers to create custom machine learning environments with TensorFlow and Apache MXNet frameworks.
DataTalks.Club - Building Scalable End-to-End Deep Learning Pipelines in the ...Rustem Feyzkhanov
One of the main issues with ML and DL deployment is finding the right way to train and operationalize the model within the company. Serverless approach for deep learning provides simple, scalable, affordable yet reliable architecture. The challenge of this approach is to keep in mind certain limitations in CPU, GPU and RAM, and organize training and inference of your model.
My presentation will show how to utilize services like Amazon SageMaker, AWS Batch, AWS Fargate, AWS Lambda and AWS Step Functions to organize deep learning workflows.
Automated machine learning (automated ML) automates feature engineering, algorithm and hyperparameter selection to find the best model for your data. The mission: Enable automated building of machine learning with the goal of accelerating, democratizing and scaling AI. This presentation covers some recent announcements of technologies related to Automated ML, and especially for Azure. The demonstrations focus on Python with Azure ML Service and Azure Databricks.
RightScale Webinar: Operationalize Your Enterprise AWS Usage Through an IT Ve...RightScale
We’ve all seen the trend everywhere around us: customers want self-service. It offers them the agility they need and gives businesses the ability to scale and lower their costs. With cloud deployments, enterprises can experience similar benefits through the use of a self-service portal where internal customers can provision their own resources while Central IT maintains control and visibility. This saves both time and money.
In this webinar, learn how to empower your internal customers to provision the necessary cloud resources when they need them but also ensure that what get receive is well within IT approved guidelines. Beyond simple convenience, this methodology permits you to operationalize your AWS cloud usage to easily roll cloud into an overall IT strategy.
Architects from Amazon Web Services (AWS) and RightScale, an Advanced Technology Partner, will provide an overview of the key business and technical considerations for operationalizing your AWS cloud usage. In the second half of the webinar, our technical experts will answer your questions. Priority will be given to pre-submitted questions.
To help illustrate the effectiveness of this approach, our architects will walk you through real-world examples and the overall impact on their organizations.
Key Topics:
1. Create an IT Vending machine with consistent and reproducible processes.
2. Enable your end users while maintaining visibility and control.
3. Use cost planning and forecasting to fine-tune and understand cloud spend.
4. Discover reporting and auditing tools to ensure compliance.
5. Avoid downtime through proven HA/DR architectures.
오토스케일링(Auto-scaling)은 AWS 클라우드를 통해 고확장성 서비스와 아키텍처를 구성하는 데 필요한 가장 중요한 요소 중 하나입니다. 이 강연에서는 효과적인 클라우드 인프라 구축을 위해 오토 스케일링을 활용하는 다양한 방법에 대해 자세히 소개해 드립니다.
오토 스케일링 그룹의 구성과 확장 계획에 따른 설정 방법, 오토 스케일링 라이프 사이클과 CloudWatch 및 알림을 이용한 관리 방법, 각종 오토스케일링 모범사례 등을 알아보실 수 있습니다.
AWS 201 Webinar Series - Rightsizing and Cost Optimizing your DeploymentAmazon Web Services
Leveraging the AWS Cloud can help you further lower your overall IT costs and avoid fixed, upfront IT investments. Learning how to right-size your environments can help you to go from capacity guessing to meeting QoE targets for your customers. The session will also cover best practices on how to Architect for Cost from real world customer use cases and ultimately how the AWS Cloud can help you increase revenue by focusing on Innovation and Return on Agility.
Key takeaways
- Replace up-front capital expenses with low variable costs
- Outsource undifferentiated IT tasks to useful services
- Evaluate the total Cost of (Non) Ownership
- Build Cost-aware architectures
- AWS features that help you reduce your spend
- Different purchasing options available with AWS
Who should attend
- Technical Users: Developers, engineers, system administrators and architects
- Decision Makers: IT Managers, directors and business leaders
Amazon ECS at Coursera: A unified execution framework while defending against...Brennan Saeta
In this talk, Frank Chen and Brennan Saeta discuss Coursera's use of Docker, and Amazon ECS. We discuss the implementation of our unified processing framework, and delve into the security challenges inherent in running un-trusted code.
(CMP406) Amazon ECS at Coursera: A general-purpose microserviceAmazon Web Services
"Coursera has helped millions of students learn computer science through MOOCs ranging from Introduction to Python, to state-of-the-art Functional-Reactive Programming in Scala. Our interactive educational experience relies upon an automated grading platform for programming assignments. But, because anyone can sign up for a course on Coursera for free, our systems must defend against arbitrary code execution.
Come learn how Coursera uses AWS services such as Amazon EC2 Container Service (ECS), and Amazon Virtual Private Cloud (VPC) to power a defense-in-depth strategy to secure our infrastructure against bad actors. We have modified the Amazon ECS Agent to support security layers including kernel privilege de-escalation, and enabling mandatory access control systems. Additionally, we post-process uploaded grading container images to defang binaries.
At the core of automated grading is a general-purpose near-line & batch scheduling and execution microservice built on top of the Amazon ECS APIs. We use this flexible system to power a variety of internal services across the company including data exports for instructors, course announcement emails, data reconciliation jobs, and more.
In this session, we detail aspects of our success from implementing Docker and Amazon ECS in production, providing ideas for your own scheduling, execution and hardening requirements."
AWS Summit 2013 | India - Running High Churn Development & Test Environments,...Amazon Web Services
The flexible and pay-as-you-go nature of AWS means that developers can spin up compute resources quickly and shut them down when not required. Learn about rapid deployment of applications to AWS as part of your development and testing cycle. Development and testing are a resource hungry function that requires numerous environments and the AWS Cloud allows you create these environments quickly. Hear about real-world examples of our existing customers that have benefited from using AWS for their development and testing.
AWS DeepLens Workshop: Building Computer Vision Applications - BDA201 - Atlan...Amazon Web Services
In this workshop, learn how to build and deploy computer vision models using the AWS DeepLens deep learning-enabled video camera. Learn to build a machine learning model from scratch using Amazon SageMaker, and get hands-on experience with AWS DeepLens by extending that model to build an end-to-end AI application using Amazon Rekognition. Attendees also learn about use cases built by the community which integrate other AWS services and extend the functionality of AWS DeepLens. Please note, you must have an AWS account to participate in this workshop. If setting up a new account, please do this at least 24 hours in advance of the workshop.
RightScale Webinar: January 13, 2011 – Watch this webinar for a look behind the scenes as we discuss ServerTemplates and how are they different from alternate approaches.
Similar to Scale Machine Learning from zero to millions of users (April 2020) (20)
An introduction to computer vision with Hugging FaceJulien SIMON
In this code-level talk, Julien will show you how to quickly build and deploy computer vision applications based on Transformer models. Along the way, you'll learn about the portfolio of open source and commercial Hugging Face solutions, and how they can help you deliver high-quality solutions faster than ever before.
Starting your AI/ML project right (May 2020)Julien SIMON
In this talk, we’ll see how you can put your AI/ML project on the right track from the get-go. Applying common sense and proven best practices, we’ll discuss skills, tools, methods, and more. We’ll also look at several real-life projects built by AWS customers in different industries and startups.
Optimize your Machine Learning Workloads on AWS (July 2019)Julien SIMON
Talk at Floor 28, Tel Aviv.
Infrastructure, tips to speed up training, hyperparameter optimization, model compilation, Amazon SageMaker Neo, cost optimization, Amazon Elastic Inference
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
2. Rationale
How to train ML models and deploy them in production, from
humble beginnings to world domination
Try to take reasonable and justified steps
Longer, more opinionated version: https://medium.com/@julsimon/scaling-machine-
learning-from-0-to-millions-of-users-part-1-a2d36a5e849
3.
4. And so itbegins
• You’ve trained a model on a local machine, using a popular open source library.
• You’ve measured the model’s accuracy, and things look good.
• Now you’d like to deploy it to check its actual behaviour, to run A/B tests, etc.
• You’ve embedded the model in your business application.
• You’ve deployed everything to a single virtual machine in the cloud.
• Everything works, you’re serving predictions, life is good!
5. Score card
Single EC2 instance
Infrastructure effort C’mon, it’s just one instance
ML setup effort pip install tensorflow
CI/CD integration Not needed
Build models DIY
Train models python train.py
Deploy models (at scale) python predict.py
Scale/HA inference Not needed
Optimize costs Not needed
Security Not needed
6.
7. A fewinstancesand models later…
• Life is not that good
• Too much manual work
• Time-consuming and error-prone
• Dependency hell
• No cost optimization
• Monolithic architecture
• Deployment hell
• Multiple apps can’t share the same model
• Apps and models scale differently
Use AWS-maintained tools
• Deep Learning Amazon Machine Images
• Deep Learning containers
Dockerize
Create a prediction service
• Model servers
• Bespoke API (Flask?)
8. AWS Deep LearningAMIs andContainers
Optimized environments on Amazon Linux or Ubuntu
Conda AMI
For developers who want pre-
installed pip packages of DL
frameworks in separate virtual
environments.
Base AMI
For developers who want a clean
slate to set up private DL engine
repositories or custom builds of DL
engines.
Containers
For developers who want pre-
installed containers for DL
frameworks (TensorFlow, PyTorch,
Apache MXNet)
9.
10.
11. Scaling alert!
• More customers, more team members, more models, woohoo!
• Scalability, high availability & security are now a thing
• Scaling up is a losing proposition.You need to scale out
• Only automation can save you:
IaC, CI/CD and all that good DevOps stuff
• What are your options?
12. Option 1:virtualmachines
• Definitely possible, but:
• Why? Seriously, I want to know.
• Operational and financial issues await if you don’t automate extensively
• Training
• Build on-demand clusters with CloudFormation,Terraform, etc.
• Distributed training is a pain to set up
• Prediction
• Automate deployement with CI/CD
• Scale with Auto Scaling, Load Balancers, etc.
• Spot, spot, spot
14. Option 2:Docker clusters
• This makes a lot of sense if you’re already deploying apps to Docker
• No change to the dev experience: same workflows, same CI/CD, etc.
• Deploy prediction services on the same infrastructure as business apps.
• Amazon ECS and Amazon EKS
• Lots of flexibility: mixed instance types (including GPUs), placement constraints, etc.
• Both come with AWS-maintainedAMIs that will save you time
• One cluster or many clusters ?
• Build on-demand development and test clusters with CloudFormation,Terraform, etc.
• Many customers find that running a large single production cluster works better
• Still instance-based and not fully-managed
• Not a hands-off operation: services / pods, service discovery, etc. are nice but you still have work to do
• And yes, this matters even if « someone else is taking care of clusters »
18. Model options on Amazon SageMaker
Training code
Factorization Machines
Linear Learner
Principal Component Analysis
K-Means Clustering
Image classification
And more
Built-in Algorithms (17)
No ML coding required
No infrastructure work required
Distributed training
Pipe mode
BringYour Own Container
Full control, run anything!
R, C++, etc.
No infrastructure work required
Built-in Frameworks
Bring your own code: script mode
Open source containers
No infrastructure work required
Distributed training
Pipe mode
19. TheAmazonSageMakerAPI
• Python SDK orchestrating all Amazon SageMaker activity
• High-level objects for algorithm selection, training, deploying,
automatic model tuning, etc.
https://github.com/aws/sagemaker-python-sdk
• Spark SDK (Python & Scala)
https://github.com/aws/sagemaker-spark/tree/master/sagemaker-spark-sdk
• AWS SDK
• For scripting and automation
• CLI : ‘aws sagemaker’
• Language SDKs: boto3, etc.
20. Training and deploying
tf_estimator = TensorFlow(entry_point='mnist_keras_tf.py',
role=role,
train_instance_count=1,
train_instance_type='ml.c5.2xlarge’,
framework_version='1.12',
py_version='py3',
script_mode=True,
hyperparameters={
'epochs': 10,
'learning-rate': 0.01})
tf_estimator.fit(data)
# HTTPS endpoint backed by a single instance
tf_endpoint = tf_estimator.deploy(initial_instance_count=1, instance_type=ml.t3.xlarge)
tf_endpoint.predict(…)
21. Training and deploying, atany scale
tf_estimator = TensorFlow(entry_point=’my_crazy_cnn.py',
role=role,
train_instance_count=8,
train_instance_type='ml.p3.16xlarge', # Total of 64 GPUs
framework_version='1.12',
py_version='py3',
script_mode=True,
hyperparameters={
'epochs': 200,
'learning-rate': 0.01})
tf_estimator.fit(data)
# HTTPS endpoint backed by 16 multi-AZ load-balanced instances
tf_endpoint = tf_estimator.deploy(initial_instance_count=16, instance_type=ml.p3.2xlarge)
tf_endpoint.predict(…)
22. Score card
EC2 ECS / EKS SageMaker
Infrastructure effort Maximal Some (Docker tools) None
ML setup effort Some (DLAMI) Some (DL containers) Minimal
CI/CD integration No change No change Some (SDK, Step Functions)
Build models DIY DIY 17 built-in algorithms
Train models (at scale) DIY DIY (Docker tools) SDK: 2 LOCs
Deploy models (at scale) DIY (model servers) DIY (Docker tools) SDK: 1 LOCs
Kubernetes support
Scale/HA inference DIY (Auto Scaling, LB) DIY (Services, pods, etc.) Built-in
Optimize costs DIY (Spot, RIs, automation) DIY (Spot, RIs, automation) On-demand/Spot training,
Auto Scaling for inference
Security DIY (IAM,VPC, KMS) DIY (IAM,VPC, KMS) API parameters
23. Score card
Flamewarin3,2,1…
EC2 ECS / EKS SageMaker
Infrastructure effort Maximal Some (Docker tools) None
ML setup effort Some (DLAMI) Some (DL containers) Minimal
CI/CD integration No change No change Some (SDK, Step Functions)
Build models DIY DIY 17 built-in algorithms
Train models (at scale) DIY DIY (Docker tools) 2 LOCs
Deploy models (at scale) DIY (model servers) DIY (Docker tools) SDK: 1 LOCs
Kubernetes support
Scale/HA inference DIY (Auto Scaling, LB) DIY (Services, pods, etc.) Built-in
Optimize costs DIY (Spot, RIs, automation) DIY (Spot, RIs, automation) Spot training,
Auto Scaling for inference
Security DIY (IAM,VPC, KMS) DIY (IAM,VPC, KMS) API parameters
Personal opinion Small scale only, unless you have
strong DevOps skills and enjoy
exercising them.
Reasonable choice if you’re a Docker
Docker shop, and if you’re able and
and willing to integrate with the
Docker/OSS ecosystem. If not, I’d
think twice: Docker isn’t an ML
platform.
Learn it in a few hours, forget
about servers, focus 100% on
ML, enjoy goodies like pipe
mode, distributed training, HPO,
HPO, inference pipelines and
more.
24. Conclusion
• Whatever works for you at this time is fine
• Don’t over-engineer, and don’t « plan for the future »
• Optimize for current business conditions, pay attention toTCO
• Models and data matter, not infrastructure
• When conditions change, move fast: smash and rebuild
• ... which is what cloud is all about!
• « 100% of our time spent on ML » shall be the whole of the Law
• Mix and match if it makes sense
• Train on SageMaker, deploy on ECS/EKS… or vice versa
• Write your own story!
AI Services:
AI Services are intentionally easy to use. They can be accessed via a simple API call.
We’ve pulled the best and most targeted capabilities into ready-made services--for example image recognition or transcription.
The focus here is really on enabling any developer—no ML skills required—to be able to develop AI applications using one of our services.
These API services, used in conjunction, create compelling solutions that really target business problems and use cases.
Customers can build these capabilities into their new and existing applications to reduce costs, increase speed, improve customer satisfaction and insight, and build ‘modern’ intelligent applications
What is your use case? What are the capabilities you might need? There’s an AI Service, or a pairing of services that will address the need.
AI Services descriptions for color:
Amazon Rekognition:
Rekognition makes it easy to add image and video analysis to your applications. You just provide an image or video to the Rekognition API, and the service can identify the objects, people, text, scenes, and activities, as well as detect any inappropriate content.
Amazon Rekognition also provides highly accurate facial analysis and facial recognition on images and video that you provide. You can detect, analyze, and compare faces for a wide variety of user verification, people counting, and public safety use cases.Rekognition is a simple and easy to use API that can quickly analyze any image or video file stored in Amazon S3. Amazon Rekognition is always learning from new data, and we are continually adding new labels and facial recognition features to the service.
More info: https://aws.amazon.com/rekognition/
Amazon Polly:
Amazon Polly is a service that turns text into lifelike speech, allowing you to create applications that talk, and build entirely new categories of speech-enabled products.
Polly is a text to speech service that uses advanced deep learning technologies to synthesize speech that sounds like a human voice.
With dozens of lifelike voices across a variety of languages, you can select the ideal voice and build speech-enabled applications that work in many different countries.
More info: https://aws.amazon.com/polly/
Amazon Transcribe:
Amazon Transcribe is an automatic speech recognition (ASR) service that makes it easy for developers to add speech-to-text capability to their applications.
Using the Amazon Transcribe API, you can analyze audio files stored in Amazon S3 and have the service return a text file of the transcribed speech.
Amazon Transcribe can be used for lots of common applications, including the transcription of customer service calls and generating subtitles on audio and video content.
The service can transcribe audio files stored in common formats, like WAV and MP3, with time stamps for every word so that you can easily locate the audio in the original source by searching for the text. Amazon Transcribe is continually learning and improving to keep pace with the evolution of language.
More info: https://aws.amazon.com/transcribe/
Amazon Translate:
Amazon Translate is a neural machine translation service that delivers fast, high-quality, and affordable language translation.
Neural machine translation is a form of language translation automation that uses deep learning models to deliver more accurate and more natural sounding translation than traditional statistical and rule-based translation algorithms.
Amazon Translate allows you to localize content - such as websites and applications - for international users, and to easily translate large volumes of text efficiently.
More info: https://aws.amazon.com/translate/
Amazon Comprehend:
Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to find insights and relationships in text.
The service identifies the language of the text; extracts key phrases, places, people, brands, or events; understands how positive or negative the text is; analyzes text using tokenization and parts of speech; and automatically organizes a collection of text files by topic.
Using these APIs, you can analyze text and apply the results in a wide range of applications including voice of customer analysis, intelligent document search, and content personalization for web applications.
More info: https://aws.amazon.com/comprehend
Amazon Lex:
Amazon Lex is a service for building conversational interfaces into any application using voice and text.
Amazon Lex provides the advanced deep learning functionalities of automatic speech recognition (ASR) for converting speech to text, and natural language understanding (NLU) to recognize the intent of the text, to enable you to build applications with highly engaging user experiences and lifelike conversational interactions.
With Amazon Lex, the same deep learning technologies that power Amazon Alexa are now available to any developer, enabling you to quickly and easily build sophisticated, natural language, conversational bots
More info: https://aws.amazon.com/lex
AI Services:
AI Services are intentionally easy to use. They can be accessed via a simple API call.
We’ve pulled the best and most targeted capabilities into ready-made services--for example image recognition or transcription.
The focus here is really on enabling any developer—no ML skills required—to be able to develop AI applications using one of our services.
These API services, used in conjunction, create compelling solutions that really target business problems and use cases.
Customers can build these capabilities into their new and existing applications to reduce costs, increase speed, improve customer satisfaction and insight, and build ‘modern’ intelligent applications
What is your use case? What are the capabilities you might need? There’s an AI Service, or a pairing of services that will address the need.
AI Services descriptions for color:
Amazon Rekognition:
Rekognition makes it easy to add image and video analysis to your applications. You just provide an image or video to the Rekognition API, and the service can identify the objects, people, text, scenes, and activities, as well as detect any inappropriate content.
Amazon Rekognition also provides highly accurate facial analysis and facial recognition on images and video that you provide. You can detect, analyze, and compare faces for a wide variety of user verification, people counting, and public safety use cases.Rekognition is a simple and easy to use API that can quickly analyze any image or video file stored in Amazon S3. Amazon Rekognition is always learning from new data, and we are continually adding new labels and facial recognition features to the service.
More info: https://aws.amazon.com/rekognition/
Amazon Polly:
Amazon Polly is a service that turns text into lifelike speech, allowing you to create applications that talk, and build entirely new categories of speech-enabled products.
Polly is a text to speech service that uses advanced deep learning technologies to synthesize speech that sounds like a human voice.
With dozens of lifelike voices across a variety of languages, you can select the ideal voice and build speech-enabled applications that work in many different countries.
More info: https://aws.amazon.com/polly/
Amazon Transcribe:
Amazon Transcribe is an automatic speech recognition (ASR) service that makes it easy for developers to add speech-to-text capability to their applications.
Using the Amazon Transcribe API, you can analyze audio files stored in Amazon S3 and have the service return a text file of the transcribed speech.
Amazon Transcribe can be used for lots of common applications, including the transcription of customer service calls and generating subtitles on audio and video content.
The service can transcribe audio files stored in common formats, like WAV and MP3, with time stamps for every word so that you can easily locate the audio in the original source by searching for the text. Amazon Transcribe is continually learning and improving to keep pace with the evolution of language.
More info: https://aws.amazon.com/transcribe/
Amazon Translate:
Amazon Translate is a neural machine translation service that delivers fast, high-quality, and affordable language translation.
Neural machine translation is a form of language translation automation that uses deep learning models to deliver more accurate and more natural sounding translation than traditional statistical and rule-based translation algorithms.
Amazon Translate allows you to localize content - such as websites and applications - for international users, and to easily translate large volumes of text efficiently.
More info: https://aws.amazon.com/translate/
Amazon Comprehend:
Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to find insights and relationships in text.
The service identifies the language of the text; extracts key phrases, places, people, brands, or events; understands how positive or negative the text is; analyzes text using tokenization and parts of speech; and automatically organizes a collection of text files by topic.
Using these APIs, you can analyze text and apply the results in a wide range of applications including voice of customer analysis, intelligent document search, and content personalization for web applications.
More info: https://aws.amazon.com/comprehend
Amazon Lex:
Amazon Lex is a service for building conversational interfaces into any application using voice and text.
Amazon Lex provides the advanced deep learning functionalities of automatic speech recognition (ASR) for converting speech to text, and natural language understanding (NLU) to recognize the intent of the text, to enable you to build applications with highly engaging user experiences and lifelike conversational interactions.
With Amazon Lex, the same deep learning technologies that power Amazon Alexa are now available to any developer, enabling you to quickly and easily build sophisticated, natural language, conversational bots
More info: https://aws.amazon.com/lex