This session was recorded in San Francisco on February 5th, 2019 and can be viewed here: https://youtu.be/cnU6sqd31JU
Developing meaningful AI applications requires complete data lifecycle management. Sourcing, harvesting, labelling and ensuring the conduit to consume data structures and repositories is critical for model accuracy....but, one of the least talked about subjects. Intel’s optimized technologies enable efficient delivery of complete data samples to develop (and deploy) meaningful outcomes. During this session, we’ll review the considerations and criticality of data lifecycle management for the AI production pipeline.
Bio: Meg brings more than 17 years of global product, engineering and solutions experience. She is presently a Solutions Architect with Intel Corporation specializing in Visual Compute and AAI (Analytics and AI) Architecture. She is passionate about the potential for technology to improve the quality of peoples’ lives and humanity on the whole.
Democratization - New Wave of Data Science (홍운표 상무, DataRobot) :: AWS Techfor...Amazon Web Services Korea
This document discusses the democratization of data science and machine learning using automated machine learning tools. It provides examples of how DataRobot has helped customers in various industries build predictive models faster and with less coding than traditional approaches. Specifically, it summarizes how DataRobot has helped customers in banking, insurance, retail, and other industries with use cases like predictive maintenance, sales forecasting, fraud detection, customer churn prediction, and insurance underwriting.
Top 10 ways BigInsights BigIntegrate and BigQuality will improve your lifeIBM Analytics
BigIntegrate and BigQuality offer 10 ways to improve an organization's ability to leverage Hadoop by providing cost-effective data integration and quality capabilities that eliminate hand coding, improve performance, ensure scalability and reliability, and increase productivity when working with Hadoop data.
Big Data & Analytics continues to redefine business. Data has transitioned from an underused asset to the lifeblood of the organisation, and a critical component of business intelligence, insight and strategy.
Big Data Scotland is the largest annual data analytics conference held in Scotland: it is supported by ScotlandIS and The Data Lab and free for delegates to attend. The conference is geared towards senior technologists and business leaders and aims to provide a unique forum for knowledge exchange, discussion and cross-pollination.
The programme will explore the evolution of data analytics; looking at key tools and techniques and how these can be applied to deliver practical insight and value. Presentations will span a wide array of topics from Data Wrangling and Visualisation to AI, Chatbots and Industry 4.0.
Key Topics
• Tools and techniques
• Corporate data culture, business processes, digital transformation
• Business intelligence, trends, decision making
• AI, Real-time Analytics, IoT, Industry 4.0, Robotics
• Security, regulation, privacy, consent, anonymization
• Data visualisation, interpretation and communication
• CRM and Personalisation
SAP provides analytics solutions to help customers run their businesses better by accessing all relevant information, defining plans to align performance goals, and responding instantly to changing conditions. Their solutions provide capabilities for data analysis, business intelligence applications, and collaboration. Customers in various industries and regions have been able to increase financial and operational performance through SAP's analytics offerings.
Pivotal Digital Transformation Forum: Becoming a Data Driven EnterpriseVMware Tanzu
Next Steps in Your Digital Transformation
This session brings together all the lessons learnt throughout the day and shares with you practical advice on how to get started with, or accelerate, your journey to become a digital business.
DataOps - Big Data and AI World London - March 2020 - Harvinder AtwalHarvinder Atwal
Title
DataOps, the secret weapon for delivering AI, data science, and business intelligence value at speed.
Synopsis
● According to recent research, just 7.3% of organisations say the state of their data and analytics is excellent, and only 22% of companies are currently seeing a significant return from data science expenditure.
● Poor returns on data & analytics investment are often the result of applying 20th-century thinking to 21st-century challenges and opportunities.
● Modern data science and analytics require secure, efficient processes to turn raw data from multiple sources and in numerous formats into useful inputs to a data product.
● Developing, orchestrating and iterating modern data pipelines is an extremely complex process requiring multiple technologies and skills.
● Other domains have to successfully overcome the challenge of delivering high-quality products at speed in complex environments. DataOps applies proven agile principles, lean thinking and DevOps practices to the development of data products.
● A DataOps approach aligns data producers, analytical data consumers, processes and technology with the rest of the organisation and its goals.
H2O.ai provides open source machine learning platforms and enterprise AI solutions that help companies implement artificial intelligence. It offers tools for data scientists to build models using Python and R and also provides support services to help customers successfully deploy models in production. H2O.ai aims to democratize AI and help companies become AI-driven by leveraging its experts, community knowledge, and world-class technology.
Democratization - New Wave of Data Science (홍운표 상무, DataRobot) :: AWS Techfor...Amazon Web Services Korea
This document discusses the democratization of data science and machine learning using automated machine learning tools. It provides examples of how DataRobot has helped customers in various industries build predictive models faster and with less coding than traditional approaches. Specifically, it summarizes how DataRobot has helped customers in banking, insurance, retail, and other industries with use cases like predictive maintenance, sales forecasting, fraud detection, customer churn prediction, and insurance underwriting.
Top 10 ways BigInsights BigIntegrate and BigQuality will improve your lifeIBM Analytics
BigIntegrate and BigQuality offer 10 ways to improve an organization's ability to leverage Hadoop by providing cost-effective data integration and quality capabilities that eliminate hand coding, improve performance, ensure scalability and reliability, and increase productivity when working with Hadoop data.
Big Data & Analytics continues to redefine business. Data has transitioned from an underused asset to the lifeblood of the organisation, and a critical component of business intelligence, insight and strategy.
Big Data Scotland is the largest annual data analytics conference held in Scotland: it is supported by ScotlandIS and The Data Lab and free for delegates to attend. The conference is geared towards senior technologists and business leaders and aims to provide a unique forum for knowledge exchange, discussion and cross-pollination.
The programme will explore the evolution of data analytics; looking at key tools and techniques and how these can be applied to deliver practical insight and value. Presentations will span a wide array of topics from Data Wrangling and Visualisation to AI, Chatbots and Industry 4.0.
Key Topics
• Tools and techniques
• Corporate data culture, business processes, digital transformation
• Business intelligence, trends, decision making
• AI, Real-time Analytics, IoT, Industry 4.0, Robotics
• Security, regulation, privacy, consent, anonymization
• Data visualisation, interpretation and communication
• CRM and Personalisation
SAP provides analytics solutions to help customers run their businesses better by accessing all relevant information, defining plans to align performance goals, and responding instantly to changing conditions. Their solutions provide capabilities for data analysis, business intelligence applications, and collaboration. Customers in various industries and regions have been able to increase financial and operational performance through SAP's analytics offerings.
Pivotal Digital Transformation Forum: Becoming a Data Driven EnterpriseVMware Tanzu
Next Steps in Your Digital Transformation
This session brings together all the lessons learnt throughout the day and shares with you practical advice on how to get started with, or accelerate, your journey to become a digital business.
DataOps - Big Data and AI World London - March 2020 - Harvinder AtwalHarvinder Atwal
Title
DataOps, the secret weapon for delivering AI, data science, and business intelligence value at speed.
Synopsis
● According to recent research, just 7.3% of organisations say the state of their data and analytics is excellent, and only 22% of companies are currently seeing a significant return from data science expenditure.
● Poor returns on data & analytics investment are often the result of applying 20th-century thinking to 21st-century challenges and opportunities.
● Modern data science and analytics require secure, efficient processes to turn raw data from multiple sources and in numerous formats into useful inputs to a data product.
● Developing, orchestrating and iterating modern data pipelines is an extremely complex process requiring multiple technologies and skills.
● Other domains have to successfully overcome the challenge of delivering high-quality products at speed in complex environments. DataOps applies proven agile principles, lean thinking and DevOps practices to the development of data products.
● A DataOps approach aligns data producers, analytical data consumers, processes and technology with the rest of the organisation and its goals.
H2O.ai provides open source machine learning platforms and enterprise AI solutions that help companies implement artificial intelligence. It offers tools for data scientists to build models using Python and R and also provides support services to help customers successfully deploy models in production. H2O.ai aims to democratize AI and help companies become AI-driven by leveraging its experts, community knowledge, and world-class technology.
Open source Apache Hadoop is a great framework for distributed processing of large data sets. But there’s a difference between “playing” with big data versus solving real problems. The reality is that Hadoop alone is not enough. In fact, almost every organization that plans to use Hadoop for production use quickly discovers that it lacks the required features for enterprise use. And, fewer still have the Hadoop specialists on hand to navigate through the complexity to build reliable, robust applications. As a result, many Hadoop projects never make it to production as executives say, “we just don’t have the skills.” In this session, we will discuss these enterprise capabilities and why they’re important: analytics, visualization, security, enterprise integration, developer/admin tools, and more. Additionally, we will share several real-world client examples who have found it necessary to use an enterprise-grade Hadoop platform to tackle some of the most interesting and challenging business problems.
Keynote by Mike Gualtieri, Forrester Research - Making AI Happen Without Gett...Sri Ambati
This session was recorded in San Francisco on February 5th, 2019 and can be viewed here: https://youtu.be/4a_Y0L7suBc
AI is real. Enterprises use it to automate decisions, hyper-personalize customer experiences, streamline operational processes, and much more. However, for most enterprise technology leaders, AI technologies and use cases are still far too mysterious. The field is moving fast. Enterprise leaders must forge a coherent, pragmatic AI strategy that is tied to business outcomes. In this session, guest speaker Forrester Research Vice President & Principal Analyst Mike Gualtieri will demystify enterprise AI, identify use cases most likely to succeed, and, most importantly, provide key advice to enterprise leaders that are charged with moving AI forward in their organization.
Bio: Mike's research focuses on software technologies, platforms, and practices that enable technology professionals to deliver digital transformations that lead to prescient digital experiences and breakthrough operational efficiency. His key technology coverage areas are AI, machine learning, deep learning, AI chips and systems, digital decisions, streaming analytics, prescriptive analytics, big data analytical platforms and tools (Hadoop/Spark/Flink; translytical databases), optimization, and emerging technologies that make software faster and smarter. Mike is also a leading expert on the intersection of business strategy, artificial intelligence, and innovation. Mike provides technology vendors with actionable, fine-tuned advisory sessions on strategy, messaging, competitive analysis, buyer-persona analysis, market trends, and product road maps for the areas he directly covers and adjacent areas that wish to launch into new markets or use new technologies. Mike is a recipient of the Forrester Courage Award for making bold calls that inspire leaders and guide great business and technology decisions.
Understanding Big Data Analytics - solutions for growing businesses - Rafał M...GetInData
Did you like it? Check out our blog to stay up to date: https://getindata.com/blog
Data Analytics became a central point in many Digital Transformation programs. Building a data-driven organisation requires a common understanding the foundations of data analytics on every level. This presentation will help you and your colleagues understand Big Data, Data Science, Machine Learning and Artificial Intelligence.
Watch our webinar about Big Data Analytics: https://youtu.be/jdfKHVWov6A
Speaker: Rafał Małanij
---
Getindata is a company founded in 2014 by ex-Spotify data engineers. From day one our focus has been on Big Data projects. We bring together a group of best and most experienced experts in Poland, working with cloud and open-source Big Data technologies to help companies build scalable data architectures and implement advanced analytics over large data sets.
Our experts have vast production experience in implementing Big Data projects for Polish as well as foreign companies including i.a. Spotify, Play, Truecaller, Kcell, Acast, Allegro, ING, Agora, Synerise, StepStone, iZettle and many others from the pharmaceutical, media, finance and FMCG industries.
https://getindata.com
This presentation was made on May 13, 2020 and the video recording of it can be viewed here: https://youtu.be/QAgYASr1SHA
Description:
Are AI and AutoML overhyped or the answer to our problems?
Beyond the hyperbole, what are AutoML and AI?
How are they helpful, and when are they not?
Why are they more relevant and valuable than ever?
Our world is changing rapidly, and that implies many organizations will need to adapt quickly. AI is unlocking new potential for every enterprise. Organizations are using AI and machine learning technology to inform business decisions, predict potential issues, and provide more efficient, customized customer experiences. The results can enable a competitive edge for the business. AI empowers data teams to scale and deliver trusted, production-ready models in an easier, faster, more cost-effective way than traditional machine learning approaches.
AI and AutoML are not magic but it can be transformative, find out how at this virtual meetup. Get practical tips and see AutoML in action with a real-world example. We’ll demonstrate how AutoML can augment your Data Scientists, supercharging your team and giving your organization the AI edge in record time.
Speakers' Bio:
James Orton: He has over a decade of experience in analytics and data science across a number of industries. He has managed data science teams and large scale projects, before more recently launching his own startup. His vision for AI and that of H2O.ai were so closely aligned, it was a fortuitous opportunity for James to join H2O.ai in the Australia and New Zealand region.
The Importance of DataOps in a Multi-Cloud WorldDATAVERSITY
There’s no denying that Cloud has evolved from being an outlying market disruptor to a mainstream method for delivering IT applications and services. In fact, it’s not uncommon to find that Enterprises use the services of more than one cloud at the same time. However, while a multi-cloud strategy offers many benefits, it also increases data management complexity and consequently reduces data availability. This webinar defines the meaning of DataOps and why it’s a crucial component for every multi-cloud approach.
This presentation was made on June 16, 2020.
A recording of the presentation can be viewed here: https://youtu.be/khjW1t0gtSA
AI is unlocking new potential for every enterprise. Organizations are using AI and machine learning technology to inform business decisions, predict potential issues, and provide more efficient, customized customer experiences. The results can enable a competitive edge for the business.
H2O.ai is a visionary leader in AI and machine learning and is on a mission to democratize AI for everyone. We believe that every company can become an AI company, not just the AI Superpowers. We are empowering companies with our leading AI and Machine Learning platforms, our expertise, experience and training to embark on their own AI journey to become AI companies themselves. All companies in all industries can participate in this AI Transformation.
Tune into this virtual meetup to learn how companies are transforming their business with the power of AI and where to start.
About Parul Pandey:
Parul is a Data Science Evangelist here at H2O.ai. She combines Data Science , evangelism and community in her work. Her emphasis is to spread the information about H2O and Driverless AI to as many people as possible, She is also an active writer and has contributed towards various national and international publications.
This document discusses semantic data management. It describes the goals of reducing the time data scientists spend collecting, cleaning and organizing data so they can focus more on analysis. It also aims to make data more accessible, understandable and usable for different stakeholders. Key challenges include heterogeneous data formats, models, semantics and quality. The document outlines research into semantic querying, processing knowledge graphs and mapping to help integrate, understand and apply enterprise data.
The AI Mindset: Bridging Industry and Academic PerspectivesSnapLogic
In this presentation, find out how Dr. Greg Benson brought ML into the SnapLogic platform and how to combine the strengths of industry practices and academic methodologies to achieve success with ML.
Introducción al Machine Learning AutomáticoSri Ambati
¿Cómo puede llevar el aprendizaje automático a las masas? Los proyectos de Machine Learning con la búsqueda de talento, el tiempo para construir e implementar modelos y confiar en los modelos que se construyen.
¿Cómo puede tener varios equipos en su organización para crear modelos de ML precisos sin ser expertos en ciencia de datos o aprendizaje automático?
¿Se pregunta sobre los diferentes sabores de AutoML?
H2O Driverless AI emplea las técnicas de científicos expertos en datos en una aplicación fácil de usar que ayuda a escalar sus esfuerzos de ciencia de datos. La inteligencia artificial Driverless permite a los científicos de datos trabajar en proyectos más rápido utilizando la automatización y la potencia de computación de vanguardia de las GPU para realizar tareas en minutos que solían tomar meses.
Con H2O Driverless AI, todos, incluyendo expertos y científicos de datos junior, científicos de dominio e ingenieros de datos pueden desarrollar modelos confiables de aprendizaje automático. Esta plataforma de aprendizaje automático de última generación ofrece una funcionalidad única y avanzada para la visualización de datos, la ingeniería de características, la interpretabilidad del modelo y la implementación de baja latencia.
H2O Driverless AI hace:
* Visualización automática de datos
* Ingeniería automática de funciones a nivel de Grandmaster
* Selección automática del modelo
* Ajuste y capacitación automáticos del modelo
* Paralelización automática utilizando múltiples CPU o GPU
* Ensamblaje automático del modelo
*automática del Interpretaciónaprendizaje automático (MLI)
* Generación automática de código de puntuación
¿Quieres probarlo tú mismo? Puede obtener una prueba gratuita aquí: H2O Driverless AI trial.
Venga a esta sesión y descubra cómo comenzar con el Aprendizaje automático automático con AI sin conductor H2O, y cree modelos potentes con solo unos pocos clics.
¡Te veo pronto!
Acerca de H2O.ai
H2O.ai es una empresa visionaria de software de código abierto de Silicon Valley que creó y reimaginó lo que es posible. Somos una empresa de fabricantes que trajeron al mercado nuevas plataformas y tecnologías para impulsar el movimiento de inteligencia artificial. Somos los creadores de, H2O, la principal plataforma de aprendizaje de ciencia de datos de fuente abierta y de aprendizaje automático utilizada por casi la mitad de Fortune 500 y en la que confían más de 14,000 organizaciones y cientos de miles de científicos de datos de todo el mundo.
This document discusses enterprise data science and machine learning. It begins by noting that data is now more plentiful and machine learning opportunities are everywhere. However, challenges remain around scaling data science work, making models production-ready, and meeting different team needs. The document then introduces Cloudera's Data Science Workbench for addressing these challenges. It claims the Workbench provides a secure, self-service environment allowing data scientists direct access to enterprise data and tools while meeting IT requirements. Examples are given of how it supports the full data science pipeline from exploration to production. In demos, it highlights features like connecting to Hadoop clusters securely and enabling collaboration. Overall, the document pitches Cloudera's Workbench as a solution
Near realtime AI deployment with huge data and super low latency - Levi Brack...Sri Ambati
Published on Nov 2, 2018
This talk was recorded in London on October 30th, 2018 and can be viewed here: https://youtu.be/erHt-1yBuUw
Session: Travelport is a leading travel commerce platform that has truly huge data and many complex needs in terms of processing, performance and latency. This talk will demonstrate how we were able to harness big data technologies, H2O and cloud integration to deploy AI at scale and at low latency. The talk to cover practical advice taken from our AI journey; you will learn the successful strategies and the pitfalls of near real-time retraining ML models with streaming data and using all opensource technologies.
Bio: As principal data scientist at Travelport, Levi Brackman leads a team of data scientists that are putting ML model into production. Prior to Travelport, Levi spent most of his career in the start-up world. He founded and led an organization that created innovative educational software applications and solutions used by high schools and youth organizations in the USA and Australia. Levi earned a PhD in the quantitative social sciences under the supervision of one the world's leading educational psychologists. He earned master’s degree from University College London and is author of a business book published in eight languages that was a bestseller in multiple countries. A native of North London (UK) Levi is married and has five children and now lives in Broomfield, Colorado.
TechKnow Fiesta 2021 - Powered by Amazon Web Service in September 2021TechknowFiesta
The Data Tech Labs understand the key to any organization's success story is employee engagement. Presenting TechKnow Fiesta 2021– Skill up to Scale Up with AWS
In this talk we will share the idea of developing self guiding application that would provide the most engaging user experience possible using crowd sourced knowledge on a mobile interface. We will discuss and share how historical usage data could be mined using machine learning to identify application usage patterns to generate probable next actions. #h2ony
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
Usually, DataOps means applying DevOps principles to existing data analytics projects. We accidentally reversed it, taking a DevOps initiative and catalyzing adoption of data-driven practices across our company.
What started as a practical initiative to bring better reliability and visibility to our software product had the unexpected effect of catalyzing a transformation that helped our organization become more data-driven across the company. What we learned in the process was how and why DevOps principles can naturally expand the role of a traditional operations team and bring wider culture change to the organization.
AI Foundations Course Module 1 - An AI Transformation JourneySri Ambati
The chances of successfully implementing AI strategies within an organization significantly improve when you can recognize where your organization is on the maturity scale. Over this course, you will learn the keys to unlocking value with AI which include asking the right questions about the problems you are solving and ensuring you have the right cross-section of talent, tools, and resources. By the end of this module, you should be able to recognize where your organization is on the AI transformation spectrum and identify some strategies that can get you to the next stage in your journey.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/PJgr2epM6qs
Speakers:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
Ingrid Burton (H2O.ai - CMO)
Krish Swamy + Balaji Gopalakrishnan, Wells Fargo - Building a World Class Dat...Sri Ambati
This session was recorded in San Francisco on February 5th, 2019 and can be viewed here: https://youtu.be/VAW2eDht7JA
Bio: Krish Swamy is an experienced professional with deep skills in applying analytics and BigData capabilities to challenging business problems and driving customer insights. Krish's analytic experience includes marketing and pricing, credit risk, digital analytics and most recently, big data analytics and data transformation. His key experiences lie in banking and financial services, the digital customer experience domain, with a background in management consulting. Other key skills include influencing organizational change towards a data and analytics driven culture, and building teams of analytics, statisticians and data scientists.
Bio: Balaji Gopalakrishnan has over 15 years experience in the Machine Learning and Data Science space. Balaji has led cross functional data science and engineering teams for developing cutting-edge machine learning and cognitive computing capabilities for insurance fraud and underwriting, telematics, multi-asset class risk, scheduling under uncertainty, and others. He is passionate about driving AI adoption in organizations and strongly believes in the power of cross functional collaboration for this purpose.
Introduction & Hands-on with H2O Driverless AISri Ambati
These slides were presented by Marios Michailids and John Spooner at Dive into H2O: London on June 17, 2019.
Marios's session can be found here: https://youtu.be/GMtgT-3hENY
John's session can be found here: https://youtu.be/5t2zw4bVfsw
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...Sri Ambati
In this session, you will learn about what you should do after you’ve taken an AI transformation baseline. Over the span of this session, we will discuss the next steps in moving toward AI readiness through alignment of talent and tools to drive successful adoption and continuous use within an organization.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/K1Cl3x3rd8g
Speaker:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
Defining a Practical Path to Artificial Intelligence Roman Chanclor
With the evolution of purpose built AI Infrastructures and the advancement of Graphics Processing Units (GPUs) that enable massively parallel, deep analysis in real-time; cognitive computing may be the norm in data centers in record time. But how?
Accelerating Real-Time Analytics Insights Through Hadoop Open Source EcosystemDataWorks Summit
This document discusses accelerating real-time analytics through the Hadoop open source ecosystem. It highlights Intel's contributions to open source projects like Apache Hadoop and Apache Spark to drive mainstream adoption of advanced analytics. Real-time analytics can provide insights using data as it arrives rather than after it is stored. The document explores use cases for real-time analytics in healthcare, social media, and security and how Intel is working to accelerate solutions in these domains using its data platform and open source technologies.
Open source Apache Hadoop is a great framework for distributed processing of large data sets. But there’s a difference between “playing” with big data versus solving real problems. The reality is that Hadoop alone is not enough. In fact, almost every organization that plans to use Hadoop for production use quickly discovers that it lacks the required features for enterprise use. And, fewer still have the Hadoop specialists on hand to navigate through the complexity to build reliable, robust applications. As a result, many Hadoop projects never make it to production as executives say, “we just don’t have the skills.” In this session, we will discuss these enterprise capabilities and why they’re important: analytics, visualization, security, enterprise integration, developer/admin tools, and more. Additionally, we will share several real-world client examples who have found it necessary to use an enterprise-grade Hadoop platform to tackle some of the most interesting and challenging business problems.
Keynote by Mike Gualtieri, Forrester Research - Making AI Happen Without Gett...Sri Ambati
This session was recorded in San Francisco on February 5th, 2019 and can be viewed here: https://youtu.be/4a_Y0L7suBc
AI is real. Enterprises use it to automate decisions, hyper-personalize customer experiences, streamline operational processes, and much more. However, for most enterprise technology leaders, AI technologies and use cases are still far too mysterious. The field is moving fast. Enterprise leaders must forge a coherent, pragmatic AI strategy that is tied to business outcomes. In this session, guest speaker Forrester Research Vice President & Principal Analyst Mike Gualtieri will demystify enterprise AI, identify use cases most likely to succeed, and, most importantly, provide key advice to enterprise leaders that are charged with moving AI forward in their organization.
Bio: Mike's research focuses on software technologies, platforms, and practices that enable technology professionals to deliver digital transformations that lead to prescient digital experiences and breakthrough operational efficiency. His key technology coverage areas are AI, machine learning, deep learning, AI chips and systems, digital decisions, streaming analytics, prescriptive analytics, big data analytical platforms and tools (Hadoop/Spark/Flink; translytical databases), optimization, and emerging technologies that make software faster and smarter. Mike is also a leading expert on the intersection of business strategy, artificial intelligence, and innovation. Mike provides technology vendors with actionable, fine-tuned advisory sessions on strategy, messaging, competitive analysis, buyer-persona analysis, market trends, and product road maps for the areas he directly covers and adjacent areas that wish to launch into new markets or use new technologies. Mike is a recipient of the Forrester Courage Award for making bold calls that inspire leaders and guide great business and technology decisions.
Understanding Big Data Analytics - solutions for growing businesses - Rafał M...GetInData
Did you like it? Check out our blog to stay up to date: https://getindata.com/blog
Data Analytics became a central point in many Digital Transformation programs. Building a data-driven organisation requires a common understanding the foundations of data analytics on every level. This presentation will help you and your colleagues understand Big Data, Data Science, Machine Learning and Artificial Intelligence.
Watch our webinar about Big Data Analytics: https://youtu.be/jdfKHVWov6A
Speaker: Rafał Małanij
---
Getindata is a company founded in 2014 by ex-Spotify data engineers. From day one our focus has been on Big Data projects. We bring together a group of best and most experienced experts in Poland, working with cloud and open-source Big Data technologies to help companies build scalable data architectures and implement advanced analytics over large data sets.
Our experts have vast production experience in implementing Big Data projects for Polish as well as foreign companies including i.a. Spotify, Play, Truecaller, Kcell, Acast, Allegro, ING, Agora, Synerise, StepStone, iZettle and many others from the pharmaceutical, media, finance and FMCG industries.
https://getindata.com
This presentation was made on May 13, 2020 and the video recording of it can be viewed here: https://youtu.be/QAgYASr1SHA
Description:
Are AI and AutoML overhyped or the answer to our problems?
Beyond the hyperbole, what are AutoML and AI?
How are they helpful, and when are they not?
Why are they more relevant and valuable than ever?
Our world is changing rapidly, and that implies many organizations will need to adapt quickly. AI is unlocking new potential for every enterprise. Organizations are using AI and machine learning technology to inform business decisions, predict potential issues, and provide more efficient, customized customer experiences. The results can enable a competitive edge for the business. AI empowers data teams to scale and deliver trusted, production-ready models in an easier, faster, more cost-effective way than traditional machine learning approaches.
AI and AutoML are not magic but it can be transformative, find out how at this virtual meetup. Get practical tips and see AutoML in action with a real-world example. We’ll demonstrate how AutoML can augment your Data Scientists, supercharging your team and giving your organization the AI edge in record time.
Speakers' Bio:
James Orton: He has over a decade of experience in analytics and data science across a number of industries. He has managed data science teams and large scale projects, before more recently launching his own startup. His vision for AI and that of H2O.ai were so closely aligned, it was a fortuitous opportunity for James to join H2O.ai in the Australia and New Zealand region.
The Importance of DataOps in a Multi-Cloud WorldDATAVERSITY
There’s no denying that Cloud has evolved from being an outlying market disruptor to a mainstream method for delivering IT applications and services. In fact, it’s not uncommon to find that Enterprises use the services of more than one cloud at the same time. However, while a multi-cloud strategy offers many benefits, it also increases data management complexity and consequently reduces data availability. This webinar defines the meaning of DataOps and why it’s a crucial component for every multi-cloud approach.
This presentation was made on June 16, 2020.
A recording of the presentation can be viewed here: https://youtu.be/khjW1t0gtSA
AI is unlocking new potential for every enterprise. Organizations are using AI and machine learning technology to inform business decisions, predict potential issues, and provide more efficient, customized customer experiences. The results can enable a competitive edge for the business.
H2O.ai is a visionary leader in AI and machine learning and is on a mission to democratize AI for everyone. We believe that every company can become an AI company, not just the AI Superpowers. We are empowering companies with our leading AI and Machine Learning platforms, our expertise, experience and training to embark on their own AI journey to become AI companies themselves. All companies in all industries can participate in this AI Transformation.
Tune into this virtual meetup to learn how companies are transforming their business with the power of AI and where to start.
About Parul Pandey:
Parul is a Data Science Evangelist here at H2O.ai. She combines Data Science , evangelism and community in her work. Her emphasis is to spread the information about H2O and Driverless AI to as many people as possible, She is also an active writer and has contributed towards various national and international publications.
This document discusses semantic data management. It describes the goals of reducing the time data scientists spend collecting, cleaning and organizing data so they can focus more on analysis. It also aims to make data more accessible, understandable and usable for different stakeholders. Key challenges include heterogeneous data formats, models, semantics and quality. The document outlines research into semantic querying, processing knowledge graphs and mapping to help integrate, understand and apply enterprise data.
The AI Mindset: Bridging Industry and Academic PerspectivesSnapLogic
In this presentation, find out how Dr. Greg Benson brought ML into the SnapLogic platform and how to combine the strengths of industry practices and academic methodologies to achieve success with ML.
Introducción al Machine Learning AutomáticoSri Ambati
¿Cómo puede llevar el aprendizaje automático a las masas? Los proyectos de Machine Learning con la búsqueda de talento, el tiempo para construir e implementar modelos y confiar en los modelos que se construyen.
¿Cómo puede tener varios equipos en su organización para crear modelos de ML precisos sin ser expertos en ciencia de datos o aprendizaje automático?
¿Se pregunta sobre los diferentes sabores de AutoML?
H2O Driverless AI emplea las técnicas de científicos expertos en datos en una aplicación fácil de usar que ayuda a escalar sus esfuerzos de ciencia de datos. La inteligencia artificial Driverless permite a los científicos de datos trabajar en proyectos más rápido utilizando la automatización y la potencia de computación de vanguardia de las GPU para realizar tareas en minutos que solían tomar meses.
Con H2O Driverless AI, todos, incluyendo expertos y científicos de datos junior, científicos de dominio e ingenieros de datos pueden desarrollar modelos confiables de aprendizaje automático. Esta plataforma de aprendizaje automático de última generación ofrece una funcionalidad única y avanzada para la visualización de datos, la ingeniería de características, la interpretabilidad del modelo y la implementación de baja latencia.
H2O Driverless AI hace:
* Visualización automática de datos
* Ingeniería automática de funciones a nivel de Grandmaster
* Selección automática del modelo
* Ajuste y capacitación automáticos del modelo
* Paralelización automática utilizando múltiples CPU o GPU
* Ensamblaje automático del modelo
*automática del Interpretaciónaprendizaje automático (MLI)
* Generación automática de código de puntuación
¿Quieres probarlo tú mismo? Puede obtener una prueba gratuita aquí: H2O Driverless AI trial.
Venga a esta sesión y descubra cómo comenzar con el Aprendizaje automático automático con AI sin conductor H2O, y cree modelos potentes con solo unos pocos clics.
¡Te veo pronto!
Acerca de H2O.ai
H2O.ai es una empresa visionaria de software de código abierto de Silicon Valley que creó y reimaginó lo que es posible. Somos una empresa de fabricantes que trajeron al mercado nuevas plataformas y tecnologías para impulsar el movimiento de inteligencia artificial. Somos los creadores de, H2O, la principal plataforma de aprendizaje de ciencia de datos de fuente abierta y de aprendizaje automático utilizada por casi la mitad de Fortune 500 y en la que confían más de 14,000 organizaciones y cientos de miles de científicos de datos de todo el mundo.
This document discusses enterprise data science and machine learning. It begins by noting that data is now more plentiful and machine learning opportunities are everywhere. However, challenges remain around scaling data science work, making models production-ready, and meeting different team needs. The document then introduces Cloudera's Data Science Workbench for addressing these challenges. It claims the Workbench provides a secure, self-service environment allowing data scientists direct access to enterprise data and tools while meeting IT requirements. Examples are given of how it supports the full data science pipeline from exploration to production. In demos, it highlights features like connecting to Hadoop clusters securely and enabling collaboration. Overall, the document pitches Cloudera's Workbench as a solution
Near realtime AI deployment with huge data and super low latency - Levi Brack...Sri Ambati
Published on Nov 2, 2018
This talk was recorded in London on October 30th, 2018 and can be viewed here: https://youtu.be/erHt-1yBuUw
Session: Travelport is a leading travel commerce platform that has truly huge data and many complex needs in terms of processing, performance and latency. This talk will demonstrate how we were able to harness big data technologies, H2O and cloud integration to deploy AI at scale and at low latency. The talk to cover practical advice taken from our AI journey; you will learn the successful strategies and the pitfalls of near real-time retraining ML models with streaming data and using all opensource technologies.
Bio: As principal data scientist at Travelport, Levi Brackman leads a team of data scientists that are putting ML model into production. Prior to Travelport, Levi spent most of his career in the start-up world. He founded and led an organization that created innovative educational software applications and solutions used by high schools and youth organizations in the USA and Australia. Levi earned a PhD in the quantitative social sciences under the supervision of one the world's leading educational psychologists. He earned master’s degree from University College London and is author of a business book published in eight languages that was a bestseller in multiple countries. A native of North London (UK) Levi is married and has five children and now lives in Broomfield, Colorado.
TechKnow Fiesta 2021 - Powered by Amazon Web Service in September 2021TechknowFiesta
The Data Tech Labs understand the key to any organization's success story is employee engagement. Presenting TechKnow Fiesta 2021– Skill up to Scale Up with AWS
In this talk we will share the idea of developing self guiding application that would provide the most engaging user experience possible using crowd sourced knowledge on a mobile interface. We will discuss and share how historical usage data could be mined using machine learning to identify application usage patterns to generate probable next actions. #h2ony
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
Usually, DataOps means applying DevOps principles to existing data analytics projects. We accidentally reversed it, taking a DevOps initiative and catalyzing adoption of data-driven practices across our company.
What started as a practical initiative to bring better reliability and visibility to our software product had the unexpected effect of catalyzing a transformation that helped our organization become more data-driven across the company. What we learned in the process was how and why DevOps principles can naturally expand the role of a traditional operations team and bring wider culture change to the organization.
AI Foundations Course Module 1 - An AI Transformation JourneySri Ambati
The chances of successfully implementing AI strategies within an organization significantly improve when you can recognize where your organization is on the maturity scale. Over this course, you will learn the keys to unlocking value with AI which include asking the right questions about the problems you are solving and ensuring you have the right cross-section of talent, tools, and resources. By the end of this module, you should be able to recognize where your organization is on the AI transformation spectrum and identify some strategies that can get you to the next stage in your journey.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/PJgr2epM6qs
Speakers:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
Ingrid Burton (H2O.ai - CMO)
Krish Swamy + Balaji Gopalakrishnan, Wells Fargo - Building a World Class Dat...Sri Ambati
This session was recorded in San Francisco on February 5th, 2019 and can be viewed here: https://youtu.be/VAW2eDht7JA
Bio: Krish Swamy is an experienced professional with deep skills in applying analytics and BigData capabilities to challenging business problems and driving customer insights. Krish's analytic experience includes marketing and pricing, credit risk, digital analytics and most recently, big data analytics and data transformation. His key experiences lie in banking and financial services, the digital customer experience domain, with a background in management consulting. Other key skills include influencing organizational change towards a data and analytics driven culture, and building teams of analytics, statisticians and data scientists.
Bio: Balaji Gopalakrishnan has over 15 years experience in the Machine Learning and Data Science space. Balaji has led cross functional data science and engineering teams for developing cutting-edge machine learning and cognitive computing capabilities for insurance fraud and underwriting, telematics, multi-asset class risk, scheduling under uncertainty, and others. He is passionate about driving AI adoption in organizations and strongly believes in the power of cross functional collaboration for this purpose.
Introduction & Hands-on with H2O Driverless AISri Ambati
These slides were presented by Marios Michailids and John Spooner at Dive into H2O: London on June 17, 2019.
Marios's session can be found here: https://youtu.be/GMtgT-3hENY
John's session can be found here: https://youtu.be/5t2zw4bVfsw
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...Sri Ambati
In this session, you will learn about what you should do after you’ve taken an AI transformation baseline. Over the span of this session, we will discuss the next steps in moving toward AI readiness through alignment of talent and tools to drive successful adoption and continuous use within an organization.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/K1Cl3x3rd8g
Speaker:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
Defining a Practical Path to Artificial Intelligence Roman Chanclor
With the evolution of purpose built AI Infrastructures and the advancement of Graphics Processing Units (GPUs) that enable massively parallel, deep analysis in real-time; cognitive computing may be the norm in data centers in record time. But how?
Accelerating Real-Time Analytics Insights Through Hadoop Open Source EcosystemDataWorks Summit
This document discusses accelerating real-time analytics through the Hadoop open source ecosystem. It highlights Intel's contributions to open source projects like Apache Hadoop and Apache Spark to drive mainstream adoption of advanced analytics. Real-time analytics can provide insights using data as it arrives rather than after it is stored. The document explores use cases for real-time analytics in healthcare, social media, and security and how Intel is working to accelerate solutions in these domains using its data platform and open source technologies.
This document describes building data science pipelines in Python using Luigi. It discusses the typical data science workflow, challenges with the current workflow approach, and how data science pipelines with Luigi can help address these challenges. Luigi is presented as a Python tool that allows defining data processing tasks, dependencies between tasks, scheduling, monitoring, and failure recovery for building reproducible and production-ready data science pipelines. An example problem of building a pipeline to predict player performance in a mobile game using Luigi is provided.
Data Science Pipelines in Python using LuigiShivam Bansal
This document describes building data science pipelines in Python using Luigi. It discusses the typical data science workflow, challenges with the current workflow approach, and how data science pipelines with Luigi can help address these challenges. Key features of Luigi that make it useful for data science pipelines are presented, including task templating, scheduling, monitoring, failure recovery, and enabling batch and parallel processing. The document concludes with a demonstration Luigi pipeline example to predict the performance score of mobile game users.
The document discusses Dell Technologies' artificial intelligence (AI) and data analytics solutions portfolio. It provides an overview of Dell's solutions for AI/machine learning, IoT/streaming data, augmented analytics/data warehousing, data lakes, and high-performance computing (HPC). The solutions leverage Dell infrastructure along with partner technologies and are designed to address various analytical use cases such as digital manufacturing, life sciences research, and retail loss prevention.
byteLAKE pioneers the use of AI and HPC to provide automated and data-driven solutions for its customers. Case study: https://www.intel.com/content/www/us/en/data-center/idc-bytelake-case-study.html
byteLAKE
Artificial Intelligence for Chemical Industry, Paper Industry and Manufacturing.
We build AI products and help design custom AI software.
About byteLAKE
byteLAKE is a software company that builds Artificial Intelligence products for the chemical industry, paper industry and manufacturing. byteLAKE's CFD Suite leverages AI to reduce CFD (Computational Fluid Dynamics) chemical mixing simulations’ time from hours to minutes. byteLAKE's Cognitive Services offer AI-assisted Visual Inspection and Big Data analytics for the paper industry, detecting and visually inspecting the so-called Water Line and complex tasks automation for manufacturing. The company also offers custom AI software development for real-time data analytics (image / video / sound / time-series). To learn more about byteLAKE’s innovations, go to www.byteLAKE.com.
2 pc enterprise summit cronin newfinal aug 18IntelAPAC
Intel discusses how the evolution of IoT and big data is driving business transformation. Intel provides leading technology from devices to the cloud to deliver end-to-end IoT and big data solutions. Intel is uniquely positioned through its technology, partnerships, and ecosystem to integrate physical systems with data and analytics from the edge to the cloud.
How Can AI and IoT Power the Chemical Industry?Xiaonan Wang
AI, IoT and Blockchain tech briefing to the industry to showcase our research at NUS.
by Dr. Xiaonan Wang
Assistant Professor
NUS Department of Chemical & Biomolecular Engineering
This document discusses future trends in big data. It notes that the amount of data produced grows enormously every year due to new technologies and devices. Big data provides businesses with better sources of analysis and insights. Key trends discussed include the growth of open source tools like Hadoop and Spark, increased use of machine learning and predictive analytics, edge computing and analytics to process IoT data more efficiently, integration of big data and cloud computing, use of big data for cybersecurity, and growing demand for data science jobs. The conclusion states that big data will significantly impact businesses and 15% of IT organizations will move services to the cloud by 2021.
Dell NVIDIA AI Roadshow - South Western OntarioBill Wong
- Artificial intelligence (AI) is mimicking human intelligence through machine algorithms like those used for chess and facial recognition. Machine learning (ML) is a subset of AI that uses algorithms to parse data, learn from data, and make predictions. Deep learning (DL) uses artificial neural networks to develop relationships in data and is used for applications like driverless cars and cybersecurity.
- AI technologies are enabling digital transformation and require infrastructure like edge computing, GPUs, FPGAs, deep learning accelerators, and specialized hardware to power applications of AI, ML, and DL. Dell Technologies provides platforms and solutions to accelerate AI workloads and support digital transformation.
This session was held by Vladimir Brenner, Partner Account Manager, Disruptors & AI, Intel AI at the Dive into H2O: London training on June 17, 2019.
Please find the recording here: https://youtu.be/60o3eyG5OLM
How Data Virtualization Puts Machine Learning into Production (APAC)Denodo
Watch full webinar here: https://bit.ly/3mJJ4w9
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spend most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Attend this session to learn how companies can use data virtualization to:
- Create a logical architecture to make all enterprise data available for advanced analytics exercise
- Accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- Integrate popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc
In this talk, Tong will start with the current landscape and typical use cases of Artificial Intelligence applications in the Telco domain. Then, she will introduce Intel’s strategy and products for Network AI, including our focus areas, our hardware portfolio, software stacks, roadmaps and some case studies.
Speaker: Tong Zhang, Principal Engineer and Chief Architect for AI and Analytics of the Network Platforms Group, Intel
Dell NVIDIA AI Powered Transformation in Financial Services WebinarBill Wong
Digital transformation through data analytics and AI can help financial services firms address business, technology, and labor challenges caused by COVID-19. Key trends include increased reliance on remote work and digital platforms, and the importance of data analytics for decision making. By 2025, 90% of new apps will use AI. The document discusses NVIDIA and Dell Technologies' partnership and strategies for providing infrastructure to support AI workloads through solutions like the DGX A100 system, which can support training, inference, and analytics on one platform through technologies like GPUs and MIG. This helps provide a more flexible and efficient infrastructure compared to traditional siloed approaches.
Ομιλία- Παρουσίαση: Ανδρέας Τσαγκάρης, VP & Chief Technology Officer, Performance Technologies
Τίτλος Παρουσίασης: “Big Data on Linux on Power Systems”
Top 5 AI and Deep Learning Stories - August 3, 2018NVIDIA
The document discusses the top 5 deep learning stories from August 3, 2018. It summarizes each story in 1-2 paragraphs. Story 1 is about Google making NVIDIA GPUs available on their cloud to accelerate AI projects. Story 2 describes NetApp's new AI data platform called Ontap AI that helps organizations manage their AI data. Story 3 discusses how machine learning is being used in healthcare to better monitor patients. Story 4 talks about how the Swiss Federal Railway is using deep learning with cameras and sensors to improve passenger safety. Story 5 is about an AI system that taught itself to solve a Rubik's Cube in 44 hours without human help.
IRJET- Search Improvement using Digital Thread in Data AnalyticsIRJET Journal
This document discusses the use of digital thread in data analytics to improve search and provide end-to-end visibility across product lifecycles. Digital thread is a communication system that connects manufacturing process elements and provides a complete view of each element throughout the lifecycle. It allows sharing of information across organizations and suppliers. Digital thread brings quality gains by managing large amounts of data and complex supply chains. It helps enterprises quickly redesign products and meet timelines while maintaining visibility of each component's journey. The document proposes using a Neo4j graph database hosted on AWS cloud to implement a digital thread that links product data. This would provide security, performance, and analytics benefits across the overall manufacturing process.
Miguel Angel Perdiguero - Head of BIG data & analytics Atos Iberia - semanain...COIICV
This document discusses Industry 4.0 and the digital transformation of industry. It describes key technological pillars like the Internet of Things, additive manufacturing, and data analytics. It provides examples of how these technologies can be applied through predictive maintenance and customized products. The document also introduces Atos Codex, an open industrial analytics platform that uses big data, high performance computing, and machine learning to deliver business insights and solutions.
AI for good: Scaling AI in science, healthcare, and more.Intel® Software
How do we scale AI to its full potential to enrich the lives of everyone on earth? Learn about AI hardware and software acceleration and how Intel AI technologies are being used to solve critical problems in high energy physics, cancer research, financial inclusion, and more. Get started on your AI Developer Journey @ software.intel.com/ai
Structuring Big Data results to create new information: Smart Data. These Smart Data can be used to advance knowledge and support decision-making processes.
A close cooperation between industry and science creates better conditions for cutting-edge research in Data Engineering/Smart Data.
Similar to Meg Mude, Intel - Data Engineering Lifecycle Optimized on Intel - H2O World San Francisco (20)
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
This document provides an overview of H2O.ai, an AI company that offers products and services to democratize AI. It mentions that H2O products are backed by 10% of the world's top data scientists from Kaggle and that H2O has customers in 7 of the top 10 banks, 4 of the top 10 insurance companies, and top manufacturing companies. It also provides details on H2O's founders, funding, customers, products, and vision to make AI accessible to more organizations.
Generative AI Masterclass - Model Risk Management.pptxSri Ambati
Here are some key points about benchmarking and evaluating generative AI models like large language models:
- Foundation models require large, diverse datasets to be trained on in order to learn broad language skills and knowledge. Fine-tuning can then improve performance on specific tasks.
- Popular benchmarks evaluate models on tasks involving things like commonsense reasoning, mathematics, science questions, generating truthful vs false responses, and more. This helps identify model capabilities and limitations.
- Custom benchmarks can also be designed using tools like Eval Studio to systematically test models on specific applications or scenarios. Both automated and human evaluations are important.
- Leaderboards like HELM aggregate benchmark results to compare how different models perform across a wide range of tests and metrics.
AI and the Future of Software Development: A Sneak Peek Sri Ambati
The document discusses developers including both in-house and outsourced developers. It also mentions dev shops and low/no code platforms. The main topics covered are developers, dev shops, and low-code platforms.
LLMOps: Match report from the top of the 5thSri Ambati
The document discusses LLMOps (Large Language Model Operations) compared to traditional MLOps. Some key points:
- LLMOps and MLOps face similar challenges across the development lifecycle, but LLMOps requires more GPU resources and integration is faster due to more models in each application. Evaluation is also less clear.
- The LLMOps field is around the 5th generation of models, with debates around proprietary vs open source models, and balancing privacy, cost and control.
- LLMOps platforms are emerging to provide solutions for tasks like prompting, embedding databases, evaluation, and governance, similar to how MLOps platforms have evolved.
Building, Evaluating, and Optimizing your RAG App for ProductionSri Ambati
The document discusses optimizing question answering systems called RAG (Retrieve-and-Generate) stacks. It outlines challenges with naive RAG approaches and proposes solutions like improved data representations, advanced retrieval techniques, and fine-tuning large language models. Table stakes optimizations include tuning chunk sizes, prompt engineering, and customizing LLMs. More advanced techniques involve small-to-big retrieval, multi-document agents, embedding fine-tuning, and LLM fine-tuning.
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Sri Ambati
Sandeep Singh, Head of Applied AI Computer Vision, Beans.ai
H2O Open Source GenAI World SF 2023
In the modern era of machine learning, leveraging both open-source and closed-source solutions has become paramount for achieving cutting-edge results. This talk delves into the intricacies of seamlessly integrating open-source Large Language Model (LLM) solutions like Vicuna, Falcon, and Llama with industry giants such as ChatGPT and Google's Palm. As the demand for fine-tuned and specialized datasets grows, it is imperative to understand the synergy between these tools. Attendees will gain insights into best practices for building and enriching datasets tailored for fine-tuning tasks, ensuring that their LLM projects are both robust and efficient. Through real-world examples and hands-on demonstrations, this talk will equip attendees with the knowledge to harness the power of both open and closed-source tools in a coherent and effective manner.
Patrick Hall, Professor, AI Risk Management, The George Washington University
H2O Open Source GenAI World SF 2023
Language models are incredible engineering breakthroughs but require auditing and risk management before productization. These systems raise concerns about toxicity, transparency and reproducibility, intellectual property licensing and ownership, disinformation and misinformation, supply chains, and more. How can your organization leverage these new tools without taking on undue or unknown risks? While language models and associated risk management are in their infancy, a small number of best practices in governance and risk are starting to emerge. If you have a language model use case in mind, want to understand your risks, and do something about them, this presentation is for you!
Dr. Alexy Khrabrov, Open Source Science Community Director, IBM
H2O Open Source GenAI World SF 2023
In this talk, Dr. Alexy Khrabrov, recently elected Chair of the new Generative AI Commons at Linux Foundation for AI & Data, outlines the OSS AI landscape, challenges, and opportunities. With new models and frameworks being unveiled weekly, one thing remains constant: community building and validation of all aspects of AI is key to reliable and responsible AI we can use for business and society needs. Industrial AI is one key area where such community validation can prove invaluable.
The document announces the launch of the H2O GenAI App Store, which provides a collection of applications that make it easier for average users to leverage large language models through custom interfaces for specific tasks like getting gardening advice or feedback on code. The app store is designed to accelerate the development of these GenAI apps using the H2O Wave platform and provides access to H2OGPTE for retrieval augmented generation and language model calls. Developers can also contribute their own apps through the GitHub repository listed.
Applied Gen AI for the Finance Vertical Sri Ambati
Megan Kurka, Vice President, Customer Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
Discover the transformative power of Applied Gen AI. Learn how the H2O team builds customized applications and workflows that integrate capabilities of Gen AI and AutoML specifically designed to address and enhance financial use cases. Explore real world examples, learn best practices, and witness firsthand how our innovative solutions are reshaping the landscape of finance technology.
This document discusses techniques for improving language models (LLMs) discussed in recent papers. It describes building blocks of LLMs like fine-tuning, foundation training, memory, and databases. Specific techniques covered include LIMA which uses 1,000 carefully curated examples, instruction backtranslation to generate question-answer pairs, fine-tuning models on API examples like Gorilla, and reducing false answers through techniques like not agreeing with incorrect user opinions. The goal is to discuss cutting edge tricks to build better LLMs.
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...Sri Ambati
Pascal Pfeiffer, Principal Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
This talk dives into the expansive ecosystem of Large Language Models (LLMs), offering practitioners an insightful guide to various relevant applications, from natural language understanding to creative content generation. While exploring use cases across different industries, it also honestly addresses the current limitations of LLMs and anticipates future advancements.
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...Sri Ambati
This document discusses using large language models (LLMs) for text classification tasks. It begins by describing how LLMs are commonly used for text generation and question answering. For classification, models are usually trained supervised on labeled data. The document then explores using LLMs for zero-shot classification without training, and techniques like fine-tuning LLMs on tasks to improve performance. It provides an example of fine-tuning an LLM on a financial sentiment dataset. The document concludes by describing H2O.ai's LLM Studio tool for fine-tuning and a few Kaggle competitions where LLMs achieved success in text classification.
1) Generative AI (GenAI) enables the creation of novel content by learning patterns in unstructured data rather than labeling outputs like traditional AI.
2) Both traditional and generative AI models lack transparency and may contain biases, but generative models can additionally hallucinate or leak private information.
3) To interpret generative models, researchers evaluate accuracy globally by checking for hallucinations or undesirable content, and locally by confirming the quality of individual responses.
Introducción al Aprendizaje Automatico con H2O-3 (1)Sri Ambati
En esta reunión virtual, damos una introducción a la plataforma de aprendizaje automático de código abierto número 1, H2O-3 y te mostramos cómo puedes usarla para desarrollar modelos para resolver diferentes casos de uso.
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...Sri Ambati
Numerai is an open, crowd-sourced hedge fund powered by predictions from data scientists around the world. In return, participants are rewarded with weekly payouts in crypto.
In this talk, Joe will give an overview of the Numerai tournament based on his own experience. He will then explain how he automates the time-consuming tasks such as testing different modelling strategies, scoring new datasets, submitting predictions to Numerai as well as monitoring model performance with H2O Driverless AI and R.
ML Model Deployment and Scoring on the Edge with Automatic ML & DFSri Ambati
Machine Learning Model Deployment and Scoring on the Edge with Automatic Machine Learning and Data Flow
YouTube Video URL: https://youtu.be/gB0bTH-L6DE
Deploying Machine Learning models to the edge can present significant ML/IoT challenges centered around the need for low latency and accurate scoring on minimal resource environments. H2O.ai's Driverless AI AutoML and Cloudera Data Flow work nicely together to solve this challenge. Driverless AI automates the building of accurate Machine Learning models, which are deployed as light footprint and low latency Java or C++ artifacts, also known as a MOJO (Model Optimized). And Cloudera Data Flow leverage Apache NiFi that offers an innovative data flow framework to host MOJOs to make predictions on data moving on the edge.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Project Management Semester Long Project - Acuityjpupo2018
Acuity is an innovative learning app designed to transform the way you engage with knowledge. Powered by AI technology, Acuity takes complex topics and distills them into concise, interactive summaries that are easy to read & understand. Whether you're exploring the depths of quantum mechanics or seeking insight into historical events, Acuity provides the key information you need without the burden of lengthy texts.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
OpenID AuthZEN Interop Read Out - AuthorizationDavid Brossard
During Identiverse 2024 and EIC 2024, members of the OpenID AuthZEN WG got together and demoed their authorization endpoints conforming to the AuthZEN API
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
Meg Mude, Intel - Data Engineering Lifecycle Optimized on Intel - H2O World San Francisco
1. Session Title
Name
Title
Company
Social Media (LinkedIn / Twitter)
#H2OWORLD
Data Engineering Lifecycle on IA
Meg Mude
Solutions Architecture, Intel
@megmyname
www.Linkedin.com/in/megmude
Software.intel.com
2.
3. Today is a big day... Announcing Blue Danube
https://www.prnewswire.com/news-releases/h2oai-teams-up-with-intel-to-drive-an-ai-transformation-in-the-enterprise-300789659.html
4. 46%
of Chief Information Officers (CIOs) have developed plans
to implement AI, but only
4%have implemented AI
so far.
According to a recent Gartner survey…
Ai adoption is nascent
Source: Gartner Says Nearly Half of CIOs Are Planning to Deploy Artificial Intelligence. February 2018 (https://www.gartner.com/newsroom/id/3856163)
5. Consider how the brain processes data, Endless, enormous quantities
of data…
and delivers useful insights.
6. 4 TButonomous vehicle
5 TBNECTED AIRPLANE
1 PBSmart Factory
1.5 GBrage internet user
750 pBoud video Provider
Daily By 2020
Source: Amalgamation of analyst data and Intel analysis.
Business
Insights
Operational
Insights
Security
Insights
The deluge of data
7. Artificial
I ntelligence
is the ability of machines to learn from
experience, without explicit programming, in
order to perform cognitive functions
associated with the human mind
Artificial Intelligence
Machine
learning
Algorithms whose performance
improve as they are exposed to
more data over time
Deep
learning
Subset of machine
learning in which
multi-layered neural
networks learn from
vast amounts of data
Analytics
8. Consumer Health Finance Retail
Governme
nt
Energy Transport Industrial Other
Smart Assistants
Chatbots
Search
Personalization
Augmented Reality
Robots
Enhanced Diagnostics
Drug
Discovery
Patient Care
Research
Sensory
Aids
Algorithmic Trading
Fraud Detection
Research
Personal Finance
Risk Mitigation
Support
Experience
Marketing
Merchandising
Loyalty
Supply Chain
Security
Defense
Data
Insights
Safety & Security
Resident Engagement
Smarter
Cities
Oil & Gas Exploration
Smart
Grid
Operational
Improvement
Conservation
Autonomous Cars
Automated Trucking
Aerospace
Shipping
Search & Rescue
Factory Automation
Predictive Maintenance
Precision Agriculture
Field Automation
Advertising
Education
Gaming
Professional & IT
Services
Telco/Media
Sports
Source: Intel forecast
AI will transform…everything
9. The AI lifecycle
2. Approach
Team breaks down the defined business
problem into workable steps to translate
the right data to achieve results
3. Expertise
A team of management sponsors,
data scientists, data engineers,
solution architects, and domain
experts identifies the right data and
works to translate the data to
achieve results
4. Philosophy
Team embraces fail-fast continuous
improvement practices to evaluate their success
in translating data to achieve results
5. Source Data
Team understands and obtains the
right data that explains the business
problem to achieve results
6. Infrastructure
Organization secures hardware and
software infrastructure that supports
data processing in a timely manner
7. Organization
Organization embraces data insights,
sponsors properly resourced teams, and
prioritizes analytic development work
1. Define the Challenge
10. hardwareMulti-purpose to purpose-built
AI compute from cloud to device
solutions Partner ecosystem to facilitate AI in
finance, health, retail, industrial & more
Intel
analytics
ecosystem
to get your
data ready
Data
Driving AI
forward
through R&D,
investments
and policy
Future
tools Software to accelerate development and
deployment of real solutions
Bring Your AI Vision to Life Using Our Extensive Portfolio
11. Edge
Device
ARTIFICIAL INTELLIGENCE
Platforms Finance Healthcare Energy Industrial Transport Retail Home More…
Data Center
TOOLKIT
S
App
Developers
libraries
Data
Scientists
foundatio
n
Library
Developers
*
*
*
*
FOR
* * * *
Hardware
IT System
Architects
Solution
s
Solution
Architects
AI Solutions Catalog
(Public & Internal)
DEEP LEARNING ACCELERATORS
Inference
DEEP LEARNING DEPLOYMENT
OpenVINO™ † Intel® Movidius™ SDK
Open Visual Inference & Neural Network Optimization toolkit
for inference deployment on CPU, processor graphics, FPGA
& VPU using TF, Caffe* & MXNet*
Optimized inference deployment
for all Intel® Movidius™ VPUs using
TensorFlow* & Caffe*
DEEP LEARNING FRAMEWORKS
Now optimized for CPU Optimizations in progress
TensorFlow* MXNet* Caffe* BigDL/Spark* Caffe2* PyTorch* PaddlePaddle*
DEEP LEARNING
Intel® Deep
Learning Studio‡
Open-source tool to compress deep
learning development cycle
MACHINE LEARNING LIBRARIES
Python R Distributed
•Scikit-learn
•Pandas
•NumPy
•Cart
•RandomF
orest
•e1071
•MlLib (on Spark)
•Mahout
ANALYTICS, MACHINE & DEEP LEARNING PRIMITIVES
Python DAAL MKL-DNN
Intel distribution
optimized for
machine learning
Intel® Data Analytics
Acceleration Library
(for machine learning)
Open-source deep neural
network functions for
CPU, processor graphics
DEEP LEARNING GRAPH COMPILER
Intel® nGraph™ Compiler (Alpha)
Open-sourced compiler for deep learning model computations
optimized for multiple devices (CPU, GPU, NNP) using multiple
frameworks (TF, MXNet, ONNX)
AI FOUNDATION
A
R
T
I
F
I
C
I
A
l
I
N
T
E
L
L
I
G
E
n
C
e NNP L-1000
* * * *
Ai.intel.com
† Formerly the Intel® Computer Vision SDK
*Other names and brands may be claimed as the property of others.
All products, computer systems, dates, and figures are preliminary based on current expectations, and are subject to change without notice.
12. • Typical HPC Cluster • Typical Cloud setup
One Example - Intel® Xeon®
Based Clusters
13. Customer Testimonials
“Thanks to Intel OpenVino toolkit, that
Learning Factory can now deliver the expected SLA of less
than 1 second for inferencing all 3 X-ray models that we
were targeting! I can’t believe, all the above happened in
almost a month since this effort got started! – Aruna
Narayanan, GE Healthcare AI DL Platform” – Aug 2018
(1)
“Taboola ended up sticking with Intel for reasons of speed and cost, said Ariel
Pisetzky, the company's vice president of information technology. Nvidia's
chip was far faster, but time spent shuffling data
back and forth to the chip negated the gains, Pisetzky
said. Second, Intel dispatched engineers to help Taboola tweak its computer
code so that the same servers could handle more than twice as many requests.
– Ariel Pisetzky, VP Taboola IT” July 2018 (3)
Intel and Philips achieved a speed improvement of 188 times for
the bone-age-prediction model, and a 38 times speed
improvement for the lung-segmentation model over the
baseline measurements. Vijayananda J., chief architect and fellow,
Data Science and AI at Philips HealthSuite Insights July 2018 (4)
“With PaddlePaddle now optimized for Intel
Xeon Scalable processors, developers and
data scientists can now use the same
hardware that powers the
world’s data centers and
clouds to advance their AI algorithms.”
– Jul 2018 (5)
The CERN team demonstrated that AI-based models have the potential to
act as orders-of-magnitude-faster replacements for
computationally expensive tasks in simulation, while maintaining a
remarkable level of accuracy. Dr. Federico Carminati, Gul Rukh Khattak,
and Dr. Sofia Vallecorsa at CERN, as well as Jean-Roch Vlimant at
Caltech. The work is part of a CERN openlab project in collaboration with
Intel Corporation, who partially funded the endeavor through the Intel
Parallel Computing Center (IPCC) program” – Aug 2018 (2)
“The collaboration team with representatives from Novartis and
Intel have shown more than 6X improvement in the time to
process a dataset of 10K images for training. Using the Broad
Bio-image Benchmark Collection* 021 (BBBC-021) dataset,
the team has achieved a total processing time of
31 minutes with over 99 percent accuracy.” May 2018
(6)
1) https://newsroom.intel.com/articles/solve-healthcare-intel-partners-demonstrate-real-uses-artificial-intelligence-healthcare/ 2) https://www.hpcwire.com/2018/08/14/cern-incorporates-ai-into-physics-based-simulations/ 3) https://www.reuters.com/article/us-nvidia-
intel/as-nvidia-expands-in-artificial-intelligence-intel-defends-turf-idUSKBN1L2051 4) https://venturebeat-com.cdn.ampproject.org/c/s/venturebeat.com/2018/08/14/intel-and-philips-use-xeon-chips-to-speed-up-ai-medical-scan-analysis/amp/
5) https://newsroom.intel.com/news/intel-ai-baidu-create-ai-camera-fpga-based-acceleration-xeon-scalable-optimizations-deep-learning/ 6) https://newsroom.intel.com/news/using-deep-neural-network-acceleration-image-analysis-drug-discovery/
(7) For more information, see http://aidc.gallery.video/detail/videos/china:-keynotes/video/5977039606001/large-scale-deep-learning-applications-at-baidu-and-open-source-ai-framework-paddlepaddle?autoStart=true
“Machine learning is a big part of our heritage. IT works
on GPUs today, but it also works on instances
powered by highly customized Intel
Xeon Processors” – Bratin Saha, VP & GM
Machine Learning Platforms, Amazon AI - Amazon
“Inference is one thing we do, but we do lots more. That’s
why flexibility is really essential” – Kim
Hazelwood, Head of AI Infrastructure Foundation,
Facebook
Public
Philips
“We rely heavily on Intel Xeon processors for
deep learning training and
inference workloads at Baidu”
– Dianhai Yu, Tech leader of Baidu PaddlePaddle
(7)
Baidu Customers
Internal Baidu
14. Intel® Confidential. For Internal Use ONLY
Intel works with customers across the entire AI lifecycle
TIME-
TO-
SOLUTI
ON
Opportunity Hypotheses Data Modeling Deployment Iteration Evaluation
15% 15% 23%
15% 15%
8% 8%
Experiment with
Topologies
Tune Hyper-
parameters
Share
ResultsLabel Data Load Data Augment Data
Support
Inference
Compute-intensiveLabor-intensive Labor-intensive
Proof
of
concept
Training
Source Data Scale & Deploy Inference Scale & Deploy inference within broader application
15%
15%
23%
15%
15%
8%
8%
Dev Cycle
…
Build,
Deploy
& Scale
AI customer example
The complete analytics pipeline
15. 15
Results
188X &
38x increase
Client: Philips, a worldwide
leader in healthcare
products for consumers,
patients, providers and
caregivers across the health
continuum.
Challenge: AI for medicalimagingischallenging
becausetheinformationis often high-resolutionand
multi-dimensional. Down-samplingimagesto lower
resolutionsdueto memory constraintscan cause
misdiagnoses.Philips’goalis to offerAI to its end
customers withoutsignificantlyincreasingthe cost of
the customers’systems,and withoutrequiring
modificationsto the hardwaredeployed in the field.
Solution: Philips and Intel tested two healthcare
use cases for deep learning inference, models:
one on X-rays of bones for bone-age-prediction
modeling, and the other on CT scans of lungs for
lung segmentation. The solution took advantage
of efficient multi-core processing Intel Xeon®
Scalable processors, along with the OpenVINO™
toolkit.
Intel® Distribution for
OpenVINO™
In inference performance over baseline (images
per second) for a 2S Intel® Xeon® Scalable 8168
processor
Bone age prediction model Lung segmentation model
*Other names and brands may be claimed as the property of others.
Configuration: 2-socket Intel® Xeon® Platinum 8168 processor, 2.70Ghz, HT OFF ,Total Memory 192 GB (2666 MHz), Ubuntu 18.04.1 LTS (GNU/Linux 4.15.0-29-generic x86_64*), BIOS: SE5C620.86B.0D.01.0010.072020182008, Intel
Math Kernel Library for Deep Neural Networks (Intel® MKL-DNN) v0.14. Source: https://ai.intel.com/ai/wp-content/uploads/sites/69/Intel-PhilipsAIHealthcare-CaseStudy-FinalV2-withquote.pdf
Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors.
Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary.
You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more
complete information visit http://www.intel.com/performance. Performance results are based on testing as of August 2018 and may not reflect all publicly available security updates.
See configuration disclosure for details. No product can be absolutely secure.
White paper: https://ai.intel.com/ai/wp-content/uploads/sites/69/Intel-PhilipsAIHealthcare-CaseStudy-FinalV2-withquote.pdf
Video: https://ai.intel.com/videos/philips/
Public
16. 16
LINKS:
Blog: Montefiore Health System Improves Patient Outcomes and Healthcare Efficiency with Semantic Data Lake and Artificial Intelligence Powered by Intel Technologies
Research brief: https://ai.intel.com/ai/wp-content/uploads/sites/69/montefiore-in-ai-case-study.pdf
Client: Montefiore, a
premier academic health
system in the Bronx, NY,
which has implemented a
Patient-centered Analytical
Machine Learning (PALM)
platform
Challenge: Risk stratification across a patient
population. For example, determining which
patients are at risk of respiratory failure, and
subsequent intubation (which significantly
diminishes the odds of a positive outcome).
A robust and scalable intelligent healthcare system
is needed, where models will need to be built on
data coming from a variety of sources (traditional
databases or in newer unstructured data stores),
while still complying with privacy regulations.
Solution: The PALM platform, which can tap into a
myriad of data stores, regardless of where the
information is located or how it is structured. PALM
is powered by Intel® Xeon® Scalable Gold
processors and Intel® Optane™ SSDs, and was first
deployed help identify patients at risk for
respiratory failure. This improved patient outcomes
and lowered costs, and is already starting to apply
PALM to a variety of other projects.
Intel does not control or audit third-party benchmark data or the web sites
referenced in this document. You should visit the referenced web site and
confirm whether referenced data are accurate.
Result
“With Intel’s solutions for AI, all of [these AI
capabilities] can occur on the same architecture
already in use for so many other traditional
enterprise activities, increasing efficiency and
improving time to value.”
Public
17. Intel® Confidential. For Internal Use ONLY
Sample End-to-End Solution
7
Complementary Public Cloud/Private Cloud
Sensors
logs
Messages
Smart
Machines
Transaction
logs
Source Data Sourcing and
Collection (Examples)
Storage Processing + Analysis
Kafka
Sqoop
Spark Streaming
Storm/Heron
Informatica,
DataDtage
Object Storage
RDS
In-memory
Cache
MPP DB
k/v storage
SAP HANA
Elastic
Search/Solr
Spark/BigDL
Impala/Presto
ML/deep
learning
Consumable, Visualized and
Syndicated Data / Information
Add Arcadia
Data
• Apache spark
and apache
• * Native
visualization
stacks
Post-processing
Ingest
19. AI
Is the
driving force
The path to deeper insight
Descriptive
Analytics
Diagnostic
Analytics
Predictive
Analytics
Prescriptive
Analytics
Cognitive
Analytics
Foresight
What Will Happen, When, and Why
Hindsight
What Happened?
Insight
What Happened and Why?
Forecast
How Should I Proceed?
Self-Learning
How Do I Proceed?
AI adoption is nascent – just coming into existence, but demonstrating enormous potential moving forward.
Forrester surveyed business and technology professionals and found that 58% of them were researching AI, but only 12% were using AI systems. This gap reflects growing interest in AI, but little actual use in practice. We expect enterprise interest in, and use of, AI to increase as software vendors roll out AI platforms and build AI capabilities into applications. Enterprises that plan to invest in AI expect to improve customer experiences, improve products and services, and disrupt their industry with new business models. AI technologies will increasingly be rapidly assimilated into analytics practices, giving business users unprecedented access to powerful insights that drive action. In 2018 and beyond, expect the flood gates to open even further, driven by the business’ voracious appetite for deeper contextual insights.
The business imperative for AI is firmly rooted in data. Data is the currency of the future. By 2020, we expect over 50 billion devices and 200 billion sensors to join the internet, and this hug explosion of smart & connected devices will lead to a ton of data being generated. Even just by 2020, as you can see in this slide, we’re talking about huge volumes of data. This data contains extremely valuable insights, in business, operations and security that we really want to extract. In order to extract that data, we need help, and analytics & AI are tools in our toolbag that will help us extract value from this treasure trove of data.
------ BACKUP INFO BELOW ------
The “People Devices” we’re all familiar with, PCs, tablets, phones, will remain an important part of the Intel and we will continue to invest and maximize returns in these businesses
Moving forward, these “people devices” are welcoming billions and billions of things to the internet
By 2020, 50B devices and 212B sensors will join the internet
At this point, 47% of total devices and connection will be Machine to Machine
Truly the rise of the machines…
These “things” will generate tremendous amounts of data
Consider this…
In 2020, it is expected that the average internet user will generate ~1.5 GB of traffic per day (Up from ~650MB in 2015)
Certainly a huge amount of data… until you consider the machines…
A Smart Hospital will generate 3,000 GB/day
Self-driving cars are generating over 4,000 GB/day… each
A connected plane will generate 5,000 gigabytes per day
A connected factory will generate 1 million gigabytes per day
This data, will need to be analyzed and interpreted in real time
Intel’s technology makes this possible, in order to unlock:
Operational insight – optimized efficiencies can lead to lower operational costs and higher quality
Business insishgt – understanding market needs/drivers can lead to more predictable outcomes and new opportunities
Security insight – recognizing behaviors and predicting vulnerabilities can lead to better protected IP and security planning
http://www.cisco.com/c/en/us/solutions/service-provider/vni-network-traffic-forecast/infographic.html
http://www.cisco.com/c/en/us/solutions/collateral/service-provider/global-cloud-index-gci/Cloud_Index_White_Paper.html
https://datafloq.com/read/self-driving-cars-create-2-petabytes-data-annually/172
http://www.cisco.com/c/en/us/solutions/collateral/service-provider/global-cloud-index-gci/Cloud_Index_White_Paper.html
http://www.cisco.com/c/en/us/solutions/collateral/service-provider/global-cloud-index-gci/Cloud_Index_White_Paper.html
The definition of “Artificial Intelligence” is continually evolving, but at its core, AI is about machines mimicking (and/or exceeding) cognitive functions associated with the human mind. In the universe of AI, which includes many different approaches, data-centric machine learning has emerged as a leader due to its increasing ability to tackle the three main AI sub-tasks: perception, planning/reasoning and control. Ultimately, AI is achieved through the fusion of multiple approaches to deliver ever more intelligent machines, and the nexus of AI developments in the near-future is centered on deep learning, with other approaches all playing important roles – depending on the dataset, problem, and unique requirements.
Which industries are the earliest adopters of AI? Generally, those segments with clear use cases, high purchasing power, and high rewards for making decisions quickly and/or more accurately will adopt AI fastest. Here are the segments that we believe will lead AI through 2020, ordered roughly by market opportunity (earliest at left).
Consumer
Smart Assistants – personal assistant that anticipates, optimizes, automates daily life (e.g. Amazon Alexa, Apple Siri, Google Assistant, Microsoft Cortana, Facebook Jarvis home automation, X.ai virtual assistant Amy)
Chatbots – 24/7/365 no waiting access to an informative or helpful agent (e.g. WeChat, Bank of America, Uber, Pizza Hut, Alaska Airlines, Amtrak, etc.)
Search – ability to more intelligently search more data types including image, video, context, etc (e.g. Improved Google search, Google Photos, ReSnap)
Personalization – ability to automatically adjust content/recommendations to suit individuals (e.g. Entefy, Netflix recommendation engine, Amazon personalized shopping recommendations)
Augmented Reality – overlay information on our field of view in real-time to identify interesting or undesirable things (e.g. Intel Project Alloy, Google Translate using smartphone camera)
Robots – personal robots that are able to perform household, yard, or other chores (e.g. Jibo robot for day-to-day functions, Roomba follow-ons)
Health (SME: Kristina Kermanshahche, Ketan Paranjape)
Enhanced Diagnosis – a tool for doctors to augment their own diagnosis with more data, experience, precision and accuracy (e.g. radiology image analysis, Journal of American Medicine Association paper on retina scan for diabetic retinopathy, skin lesion classification to recognize melanoma with 98% accuracy, medical history scraping, treatment outcome prediction)
Drug Discovery – computational drug discovery that intelligently hones in on the most promising treatments (e.g. speeding pharma drug development)
Patient Care – machines that aid with monitoring, treatment, and/or recovery of patients (e.g. visual patient monitoring, autonomous robotic surgery, friendly medication and/or physical therapy robots)
Research – instantly sifting through hundreds of new research papers and clinical trials that are published each day to make new connections (e.g. AI at University of North Carolina’s Lineberger Comprehensive Cancer Center)
Sensory Aids – filling in for various senses that are absent or challenged (e.g. visual aid, audio aid)
Finance (SME: Robert Geva)
Algorithmic Trading – augment rule-based algorithmic trading models and data sources using AI (e.g. Kensho analysis of myriad data to predict stock movement)
Fraud Detection – ability to identify fraudulent transactions and/or claims (e.g. USAA identifies insurance fraud)
Research – ability to intelligently assemble, parse, and extract meaning from troves of data that influence asset prices (e.g. Quid, FSI firm reducing time to insight for portfolio managers through smart knowledge management system)
Personal Finance – smarter recommendations, lower risk lending, greater efficiency (e.g. active portfolio recommendations, quickly parsing more data before issuing loan, automatic reading of check scans, etc.)
Risk Mitigation – detect risk factors and/or reduce the burden of regulation and minimize errors through automated compliance (e.g. IBM+Promontory Financial Group using natural language processing to detect excursions)
Retail (SME: Janet Kerby, Chris Hunt)
Support – bots providing shopping, ordering and support in lifelike interaction (e.g. My Starbucks Barista, KLM Dutch Airline customer support via social media, Nieman Marcus visual search, Pizza Hut order pizza via bot, Adobe Digital’s digital mirror that recommends clothes, intelligent phone menu routing based on NLP, ViSenze recommending similar items based on image, Adobe Digital’s digital mirror that recommends clothes)
Experience – deliver winning consumer experiences in-store (e.g. Amazon Go checkout-free grocery store, Macy’s mobile shopping assistant, Lowes Lowebots that roam stores answering simple questions and tracking inventory)
Marketing – precision marketing to consumers, promoting products and services how and where they want to hear (e.g. North Face “Expert Personal Shopper” on website)
Merchandising – better planning through accelerated and expanded insight into consumer buying patterns (e.g. Stitch Fix virtual styling, Skechers.com analyzing clicks in real-time to bring similar catalog items forward, Wal-mart pairing products that sell together, Cosabella evolutionary website tweaks)
Loyalty – transform the consumer experience through segmentation (e.g. Under Armour health app that constantly collects user data to deliver personalized fitness recommendations)
Supply Chain – optimize the supply chain and inventory management for efficiency and innovate new business models (e.g. OnProcess technology’s use of predictive analytics for inventory management)
Security – improve security of all consumer and business digital assets, such as real-time shoplifting/lifter detection, multi-factor identity verification, data breach detection (e.g. Mastercard pay with your face, Walmart facial recognition to catch shoplifters)
Government (SME: Harris Joyce)
Defense – drones, connected soldiers, defense strategy (e.g. military/surveillance drones, autonomous rescue vehicles, augmented connected soldier, real-time threat assessment and strategy recommendation)
Data Insights – analyze massive amounts of data to identify opportunities/inefficiencies in bureaucracy, cybersecurity threats and more, to ultimately implement better systems and policies (e.g. MIT AI that detects cyber security threats)
Crime Preventionusing AI to predict and help recover from disasters thanks to ability to quickly process large amounts of unstructured data and optimize limited resources (e.g. 1Concern, BlueLineGrid)
Safety & Security – crowd analytics, behavioral/sentiment analytics, social media analytics, face/vehicle recognition, online identity recognition, real-time video analytics, using AI to predict and help recover from disasters thanks to ability to quickly process large amounts of unstructured data and optimize limited resources (e.g. police analyzing social media to adjust police presence, license plate readers in police cars, 1Concern, BlueLineGrid)
Resident Engagement – new tools to facilitate citizen engagement like chatbots, at-risk citizen identification, (e.g. Amelia chatbot in North London Enfield council, North Carolina chatbot to help state employees with IT inquiries)
Smarter Cities – traffic/pedestrian management, lighting management, weather management, energy conservation, services analytics (e.g. San Francisco and Pittsburgh using sensors and AI to optimize traffic flow)
Energy (SME: Noe Garcia, Tonya Cosby)
Oil & Gas Exploration – automated geophysical feature detection (e.g. oil & gas producers using AI to augment traditional modeling & simulation)
Smart Grid – predictive and real-time intelligent generation, allocation, and storage of power to meet variable demand (e.g. GridSense, SoloGrid)
Operational Improvement – safety and efficiency improvements through predictive and/or insightful AI (e.g. GE Oil and Gas using predictive analytics and AI to predict and preempt potential operational problems)
Conservation – intelligent buildings, computing and appliances that reduce power consumption and are more efficient than producing another kWh of electricity (e.g. Google DeepMind datacenter energy reductions)
Transport (SME: Len Klebba)
Automated Cars – autonomous cars driving on the roadways (e.g. BMW, Google, Uber, many others)
Automated Trucking – autonomous trucks driving on the roadways (e.g. Daimler)
Aerospace – autonomous planes and other aerial vehicles (e.g. Boeing’s evolution of autopilot and drones)
Shipping – autonomous package delivery via drone or other vehicle (e.g. Amazon package delivery drone)
Search & Rescue – ability to deploy autonomous robot to search and rescue victims in potentially hazardous environments (e.g. war casualty extraction, miner rescue, firefighting, avalanche rescue)
Industrial (SME: Mary Bunzel, Esther Baldwin)
Factory Automation – highly-productive, efficient and safe factories with robots that can see, hear and adapt to their environment to produce goods with incredible quality and speed (e.g. assembly line)
Predictive Maintenance – ability to detect patterns that indicate the likelihood of an upcoming fault that would require maintenance (e.g. airline being able to adjust schedule to perform preventive maintenance before a failure)
Precision Agriculture – ability to deliver the precise amount of water, nutrients, sunlight, weed killer, etc to a particular crop or individual plant (e.g. farmer using visual weed search to zap only weeds with RoundUp, automated sorting of produce for market)
Field Automation – ability to automate heavy equipment beyond the factory walls (e.g. mining, excavation, construction, road repair)
Other
Advertising – interactive ads, adaptive ads, personalized ads, real-time ads (e.g. AdBrain, MetaMarkets, Proximic, RocketFuel)
Education – virtual mentors, foreign language instruction, automated study sheets, personalized assignments, cheating detection, deliberate practice, machine-to-machine instruction (e.g. Intelligent Tutor Systems, Content Technologies Inc, PR2 robot from Cornell)
Gaming – dynamic and interactive video game experiences (e.g. Xbox Kinect, Playstation Eye, Wii)
Professional & IT Services – sales, marketing, legal research, accounting/tax, assisted counseling, customized IT recommendations (e.g. Pinsent Masons law firm that emulates human decision-making, Salesforce use of AI)
Telco/Media – customized content/ads, network optimization, quality of service, mobile/home security (e.g. media company customizing tv show recommendations and ads, network operator ensuring efficient and high-quality delivery/repair, wireless company using multi-factor security)
Sports – intelligent analytics for injury prevention and betting (e.g. Kinduct injury prevention, Microsoft Cortana predicting football games)
Here is an even broader list of industries that will be impacted by AI: Advertising, Aerospace, Agriculture, Automotive, Building Automation, Business, Education, Fashion, Finance, Gaming, Government, Healthcare, IT, Investment, Legal, Life Sciences, Logistics, Manufacturing, Media & Entertainment, Oil/Gas/Mining, Real Estate, Retail, Sports & Fitness, Telecommunications, Transportation
Sources: Intel forecast (IDC, GII Research, Tractica, Technavio, Market Research Store, Allied Market Research, BCC Research)
Now, before we explore “what is AI”, it’s important to understand that implementing AI in your organization will be a journey. As we saw on the last slide, most businesses are at step 1 or 2 in this lifecycle, while the minority have gone full circle. Let’s step through it starting at the top and going clockwise…
The first step in any analytics or AI journey is to define the challenge you want to go solve, through brainstorming what challenges you’re facing across your organization, and prioritizing them based on business value and how much it will cost to solve them. If you think of a 2x2 chart with increasing business value on the y-axis and decreasing cost to solve on the x-axis, naturally the most impactful challenges to tackle first are those in the upper righthand quadrant. Once you’ve identified some high potential opportunities to investigate, the next step is to figure out which AI (or other) approach is best-suited to each problem, which we’ll explore in the next slide and next section. The next step is to assess whether or not you have the expertise required to implement the solution, and whether those people embrace a fail-fast continuous improvement philosophy, since AI projects typically involve a lot more uncertainty, trial & error, and exploration than more traditional and deterministic software development projects. Once the human element is in place, the next step is to source data and prepare it for analysis, as well as stand up whatever technology infrastructure is required to tackle the problem. Last, but certainly not least, you can do all the heavy-lifting to use data to solve business challenges, but if your organization isn’t ready to accept data-driven insights, then all that work may have been for naught. A classic example is the initial resistance to data analytics in sports, where general managers and scouts scoffed at the idea of computer algorithms outsmarting their years of experience and tribal knowledge. Bottom line, if think about all these steps in the AI lifecycle, you’ll stand a much better chance of realizing the business value that you set out to deliver in the first place through AI.
In the next few sections, we’ll unpack much of this AI lifecycle.
Intel’s commitment to AI is simple: help our customers bring their AI visions to life using our Extensive Portfolio.
The first step on your AI journey is getting your data ready. Intel and our partner ecosystem are ready to help you with one of multiple solutions to integrating, storing, processing and managing your data.
The complexity of bringing AI from model to reality takes a mix of hardware solutions, and our multi-architecture approach optimizes a variety of computing for different purposes, enabling application designers to choose what works best. They can use their existing multi-purpose CPU resources to begin their AI journey – including breakthrough deep learning through scaling on Xeon – and if/when it makes sense, choose from the broadest (and best) deep learning acceleration portfolio to maximize ROI for their unique requirements.
One level up this stack, software of course a critical requirement for AI, and we continue optimizing key open-source software like popular deep learning frameworks and working with our partners to bring tools to bear that reduce development time and overall time-to-solution. In addition, with the necessity of a multi-architecture approach to satisfy the demands of a wide variety of use cases, Intel is also in the process of developing tools to drive increasing harmony, reducing development and deployment complexity each step of the way.
Beyond technology and tools, we’re taking also taking a solutions-driven approach in building a strong partner ecosystem in order to scale AI and enrich the lives of every person on the planet. This will include ready-built solutions through Intel and our partner ecosystem for many segments and verticals, including healthcare, finance, retail, government, energy, transport, industrial & more.
Finally, Intel continues our push into the future by deepening our investments to push the forefront of AI computing into the next decade, including funding cutting-edge academic research, internal R&D, investments in leading innovators, and policy/ethics leadership.
In the next few slides we’ll unpack these pillars in more detail, or feel free to skip ahead to whichever section you’re most interested in.
Now that we’ve unpacked the Intel AI hardware portfolio, let’s build on top of that by looking at the important software and solutions stacks.
Software: Intel is investing in AI tools that get the most out of, and streamline development across, each hardware option in our portfolio – in order to ultimately accelerate total time-to-solution.
For application developers – those who deploy solutions using AI-based algorithms – Intel develops several tools to optimize performance and accelerate time-to-solution. For deep learning, the open-source OpenVINO™ (formerly the Intel® CV SDK and Deployment Toolkit) facilitates model deployment for inference, by converting & optimizing trained models for whichever hardware target is downstream, with support for TensorFlow, Caffe & MXNet on CPU, integrated GPU, VPU (Movidius Myriad 2 / Neural Compute Stick NCS) and FPGA. Similarly, the Intel® Movidius™ SDK supports inference deployment on TensorFlow & Caffe across the full range of VPUs. Intel is also in the process of developing the Intel® Deep Learning Studio (coming soon!) to help compress the end-to-end deep learning development cycle (including training).
For data scientists – those who create AI-based algorithms – Intel contributes to and optimizes a set of open-source libraries that are widely used for machine and deep learning. There are a number of such machine learning libraries that get the most out of Intel hardware today, spanning Python, R and Distributed. For deep learning, Intel aims to ensure that all the major DL frameworks and topologies run well on Intel hardware, and customers are of course free to choose whichever framework(s) best suit their needs. We’ve been directly optimizing the most popular AI frameworks first, based on market demand, and producing huge speedups (>100x!!!). Today, we have many optimized topologies available for TensorFlow, MXNet, Caffe and BigDL on Spark, and you can download & install the optimized version of these frameworks by clicking on the links in this slide. Going forward, we intend to enable even more frameworks in the future through the Intel® nGraph™ Compiler.
For library developers – those who develop and optimize API’s/libraries/frameworks to support new algorithms/topologies on the underlying hardware – Intel offers a host of foundational building blocks to get the most out of our hardware. Beginning on the left with the primitives category, the Data Analytics & Acceleration library (DAAL) and Intel Python distribution are important building blocks for machine learning. The ‘DNN’ (deep neural network) open source libraries contain CPU-optimized functions that are most relevant for, you guessed it, deep learning model development. On the right side of this row is a description of the Intel® nGraph™ Library (formerly the Nervana Graph), which takes the computational graph from each deep learning framework and creates an intermediate representation, which is executed by calling the math accelerator software libraries of each Intel hardware target. This compiler reduces the need for framework & model direct optimization for each hardware target using low-level software & math accelerator libraries. Today, it supports Xeon, GPU (CUDA) and the Crest family, with more hardware targets planned going forward.
Solutions: Many business don’t want to start their AI journey from scratch and/or don’t have AI expertise or desire to build a core competency in it, but would still like to harness its benefits as quickly and efficiently as possible. Enter the Intel AI builders program, which is a one-stop-shop to find Intel AI technology-based solutions, be it ready-to-develop platforms or customized solutions that address particular problems. For the more do-it-yourself (DIY) crowd, Intel also publishes case studies, reference solutions and reference blueprints through the builders program, that you can leverage to scope and implement your own AI solutions. For more information about both technical services and reference solutions, visit our builders site at builders.intel.com/ai
On the left is a typical HPC cluster showing underlying hardware to storage. On top is shown the OS, run time libraries, cluster and workload managers. The dark blue stack is AI workflow and brown stack is the HPC workflow. If the end user application is climate modeling, the HPC code will follow the brown stack, however if there is AI code as part of climate modeling for image recognition, it will follow the dark blue stack.
On the right is a cloud cluster showing an abstracted vide of running Virtual Machines and Containers. The difference when you are running containers is that you abstract away the OS – the container holds application code and all dependencies enabling much better portability.
Container – share host kernel; as far as libraries and run time environment – unique to each container. GCC/GlibC is unique to each container. So is HostOS/Kernel.
We have been working with several customers to implement AI. Here are some examples of customers we’ve worked with and their testimonials.
Discussion on what compute means
Let’s use an example to highlight the value that Intel brings to AI. The breakdown on this slide, including the proof-of-concept (POC) percentages, is from a real AI customer project focused on industrial defect detection. While other projects will differ in time breakdown, the steps are typically the same.
At the bottom, overall time-to-solution is the complete AI journey, including steps from opportunity assessment, to development & deployment, and ultimately evaluation of the end result.
In the middle, we zoom in on the develop & deploy portion of the overall solution, which includes sourcing data, proof-of-concept development, inference deployment, as well as integration into a broader application.
At the top, we zoom in further on the proof-of-concept development itself, where you can see that data preparation took a majority of the development time, followed by model training, and testing plus documentation.
The bottom line here is that while compute-intensive training (the slivers in yellow) dominate a lot of the discourse around deep learning, they’re a relatively small part of your overall time-to-solution, and Intel works with customers across the entire AI lifecycle to speedup overall time-to-solution. Thus, it’s important to think about how to spend your IT budget in a way that gets you to deployment fastest, rather than paying a premium to – now only marginally – accelerate one portion of your solution, when that acceleration may add data management and other headaches, then potentially sit idle collecting dust when it’s served its purpose.
- Demacation of Visual and Consume layers
the end2end big data solution includes data collection,data storage,data process&analyze and data visualize;each step includes a lot of different products and solutions; please select the proper solution according to your business requirements, we will further discuss the technical details of each solution in our 201 version.
Analytics is a constantly evolving science that companies can leverage for insight, innovation, and competitive advantage. Analytics has changed over the years and continues to advance through five stages of increasing scale & maturity: descriptive, diagnostic, predictive, prescriptive, and cognitive. AI is its own category, applied to all phases of the analytics pipeline (especially more advanced analytics), and a vital tool for reaching higher maturity & scale data analytics. AI is now a reality because of three key factors:
Data deluge: Our world of smart and connected devices has unleashed a data deluge, as the Internet of Things (IoT) joins apps in generating continuous streams of structured and unstructured data. The IoT will include a projected 200 billion smart-and-connected devices by 2020,8 and the data produced is expected to double every two years to total 40 zettabytes (40 trillion gigabytes) by 2020. These vast data stores are required to train many AI algorithms and are ripe to be mined for fresh insights.
Compute breakthrough: Paved by Moore’s Law, compute capability and architectural innovation have progressed to the point where we’ve crossed the threshold required to support the intense demands of machine intelligence. For example, the concept of deep learning through artificial neural networks has existed for at least 20 years, but not until the past few years have computing advancements enabled the practical application of these intensive algorithms, thanks to greater accuracy and speed.
Innovation surge: Of course, compute power and data are not enough on their own. The road to AI is also being driven by a surge of innovation that has pushed us over the tipping point from research to mainstream use. Each new AI algorithmic innovation and use case opens more eyes to the power of AI, leading more innovators to join the community and stimulating an ever-increasing demand for the technology. Neural network innovations in the 1990s renewed research into AI, but it was accuracy breakthroughs in both speech recognition and image recognition, in 2009 and 2012 respectively, that proved to be catalysts for today’s surge of innovation. In last year’s ImageNet Computer Vision contest, a neural network–based application even outperformed a human. As we progress, a plethora of unsolved AI challenges will continue to attract researchers and innovators around the world.
------------------ BACKUP ------------------
Descriptive identifies what happened in the past. It helped us understand, but it focuses on hindsight. In today’s competitive environment, hindsight is not a competitive position.
Diagnostic offers more insight into what happened, by describing why it happened. That’s more insight, but not helpful to identify what we should do as a company going forward in a fast-paced competitive world.
Predictive and Prescriptive analytics provides foresight, identifying potentials out of the many possibilities of forward pathways the business can go. It then points the business in the best directions to achieve desired outcomes. Use of many data sources and simulations help prescribe the best path forward. 40% of enterprises net-new investments in analytics will be in predictive + prescriptive analytics by 2020. (IDC, Big Data Forecast, November 2015)
Cognitive leverages what we’re learning and developing in machine and deep learning, artificial intelligence, and high performance data analytics to automate decisions using a human-like analysis.
“Traditional analytics” consists of Descriptive and Diagnostic analytics, while Artificial Intelligence plays an increasingly important role at the upper echelon of Diagnostic analytics and beyond, for Predictive, Prescriptive, and Cognitive analytics.
While there is much interest and focus on AI esp. deep learning, it is worth noting that both advanced analytics such as Predictive and the field of AI have been around for decades. The availability of open source frameworks and platforms (such as Hadoop) and advancements in analytics technologies along with a downward push on compute and storage prices have opened up the field of AI and advanced analytics to the mainstream.
Moreover, no single analytics tool or AI approach in itself is a silver bullet. For instance, not all workloads are well suited for deep learning. Successful advanced analytics and AI deployment are about layering the right analytic tools and approaches matched to the right workloads with proper scoping of projects.