An overview of why AI and Deep Learning are hot now? Overview f Machine Intelligence startups. What are the key ingredients for AI startup? How can AI startups compete with big tech companies and areas to focus on for differentiation?
An Introduction to Graph: Database, Analytics, and Cloud ServicesJean Ihm
Graph analysis employs powerful algorithms to explore and discover relationships in social network, IoT, big data, and complex transaction data. Learn how graph technologies are used in applications such as fraud detection for banking, customer 360, public safety, and manufacturing. This session will provide an overview and demos of graph technologies for Oracle Cloud Services, Oracle Database, NoSQL, Spark and Hadoop, including PGX analytics and PGQL property graph query language.
Presented at Analytics and Data Summit, March 20, 2018
This document discusses graph data science and Neo4j's capabilities. It describes how Neo4j can help simplify graph data science through its native graph database, graph data science library, and data visualization tool. Example use cases are also provided that demonstrate how Neo4j has helped companies with fraud detection, customer journey analysis, supply chain management, and patient outcomes.
Airbnb aims to democratize data within the company by building a graph database of all internal data resources connected by relationships. This graph is queried through a search interface to help employees explore, discover, and build trust in company data. Challenges include modeling complex data dependencies and proxy nodes, merging graph updates from different sources, and designing a data-dense interface simply. Future goals are to gamify content production, deliver recommendations, certify trusted content, and analyze the information network.
Slides from my talk at Big Data Conference 2018 in Vilnius
Doing data science today is far more difficult than it will be in the next 5-10 years. Sharing, collaborating on data science workflows in painful, pushing models into production is challenging.
Let’s explore what Azure provides to ease Data Scientists’ pains. What tools and services can we choose based on a problem definition, skillset or infrastructure requirements?
In this talk, you will learn about Azure Machine Learning Studio, Azure Databricks, Data Science Virtual Machines and Cognitive Services, with all the perks and limitations.
This introduction to graph databases is specifically designed for Enterprise Architects who need to map business requirements to architectural components like graph databases. It explains how and why graphs matter for Enterprise Architecture and reviews the architectural differences between relational and graph models.
Applying Noisy Knowledge Graphs to Real ProblemsDataWorks Summit
Knowledge graphs (KGs) have recently emerged as a powerful way to represent knowledge in multiple communities, including data mining, natural language processing and machine learning. Large-scale KGs like Wikidata and DBpedia are openly available, while in industry, the Google Knowledge Graph is a good example of proprietary knowledge that continues to fuel impressive advances in Google's semantic search capabilities. Yet, both crowdsourced and automatically constructed KGs suffer from noise, both during KG construction and during search and inference. In this talk, I will discuss how to build and use such knowledge graphs effectively, despite the noise and sparsity of labeled data, to solve real-world social problems such as providing insights in disaster situations, and helping law enforcement fight human trafficking. I will conclude by providing insight on the lessons learned, and the applicability of research techniques to industrial problems. The talk will be designed to appeal both to business and technical leaders.
The document discusses a presentation about connecting data and Neo4j. It covers data ecosystems and where different technologies fit, how Neo4j works as a graph database, and building graph-native organizations. It also discusses Neo4j's long term vision of connecting enterprise data and the state of data in 2018. Key points include how data structures have evolved from hierarchies to dynamic knowledge graphs and how different technologies like relational databases and Neo4j are suited for different types of queries and connected data problems.
AI, ML and Graph Algorithms: Real Life Use Cases with Neo4jIvan Zoratti
I gave this presentation at DataOps 19 in Barcelona.
You will find information about Neo4j and how to use it with Graph Algorithms for Machine Learning and Artificial Intelligence.
An Introduction to Graph: Database, Analytics, and Cloud ServicesJean Ihm
Graph analysis employs powerful algorithms to explore and discover relationships in social network, IoT, big data, and complex transaction data. Learn how graph technologies are used in applications such as fraud detection for banking, customer 360, public safety, and manufacturing. This session will provide an overview and demos of graph technologies for Oracle Cloud Services, Oracle Database, NoSQL, Spark and Hadoop, including PGX analytics and PGQL property graph query language.
Presented at Analytics and Data Summit, March 20, 2018
This document discusses graph data science and Neo4j's capabilities. It describes how Neo4j can help simplify graph data science through its native graph database, graph data science library, and data visualization tool. Example use cases are also provided that demonstrate how Neo4j has helped companies with fraud detection, customer journey analysis, supply chain management, and patient outcomes.
Airbnb aims to democratize data within the company by building a graph database of all internal data resources connected by relationships. This graph is queried through a search interface to help employees explore, discover, and build trust in company data. Challenges include modeling complex data dependencies and proxy nodes, merging graph updates from different sources, and designing a data-dense interface simply. Future goals are to gamify content production, deliver recommendations, certify trusted content, and analyze the information network.
Slides from my talk at Big Data Conference 2018 in Vilnius
Doing data science today is far more difficult than it will be in the next 5-10 years. Sharing, collaborating on data science workflows in painful, pushing models into production is challenging.
Let’s explore what Azure provides to ease Data Scientists’ pains. What tools and services can we choose based on a problem definition, skillset or infrastructure requirements?
In this talk, you will learn about Azure Machine Learning Studio, Azure Databricks, Data Science Virtual Machines and Cognitive Services, with all the perks and limitations.
This introduction to graph databases is specifically designed for Enterprise Architects who need to map business requirements to architectural components like graph databases. It explains how and why graphs matter for Enterprise Architecture and reviews the architectural differences between relational and graph models.
Applying Noisy Knowledge Graphs to Real ProblemsDataWorks Summit
Knowledge graphs (KGs) have recently emerged as a powerful way to represent knowledge in multiple communities, including data mining, natural language processing and machine learning. Large-scale KGs like Wikidata and DBpedia are openly available, while in industry, the Google Knowledge Graph is a good example of proprietary knowledge that continues to fuel impressive advances in Google's semantic search capabilities. Yet, both crowdsourced and automatically constructed KGs suffer from noise, both during KG construction and during search and inference. In this talk, I will discuss how to build and use such knowledge graphs effectively, despite the noise and sparsity of labeled data, to solve real-world social problems such as providing insights in disaster situations, and helping law enforcement fight human trafficking. I will conclude by providing insight on the lessons learned, and the applicability of research techniques to industrial problems. The talk will be designed to appeal both to business and technical leaders.
The document discusses a presentation about connecting data and Neo4j. It covers data ecosystems and where different technologies fit, how Neo4j works as a graph database, and building graph-native organizations. It also discusses Neo4j's long term vision of connecting enterprise data and the state of data in 2018. Key points include how data structures have evolved from hierarchies to dynamic knowledge graphs and how different technologies like relational databases and Neo4j are suited for different types of queries and connected data problems.
AI, ML and Graph Algorithms: Real Life Use Cases with Neo4jIvan Zoratti
I gave this presentation at DataOps 19 in Barcelona.
You will find information about Neo4j and how to use it with Graph Algorithms for Machine Learning and Artificial Intelligence.
Should a Graph Database Be in Your Next Data Warehouse Stack?Cambridge Semantics
This document discusses whether a graph database should be part of an organization's next data warehouse stack. It introduces Anzo, a managed data fabric that connects and blends enterprise data, and AnzoGraph, a data mart that can be embedded in applications or used with ETL and visualization tools. Graph databases are growing in popularity due to their ability to ask complex questions across complex data in a way that is not possible with SQL. Graph databases also support powerful graph algorithms and analysis not found in SQL. AnzoGraph is a graph database that provides standard analytics capabilities along with graph-specific features.
Democratizing Intelligence - Sri Ambati, CEO & Co-Founder, H2O.aiSri Ambati
Presented at #H2OWorld 2017 in Mountain View, CA.
Enjoy the video: https://youtu.be/ZrlJQqNaSMI.
Learn more about H2O.ai: https://www.h2o.ai/.
Follow @h2oai: https://www.twitter.com/h2oai.
The document discusses scaling tribal knowledge at Airbnb by building a graph database of the company's data resources. It describes collecting metadata on over 6,000 charts, dashboards, experiments and other data assets from various systems into a Neo4j graph database using Airflow. The graph is indexed in Elasticsearch for fast search. This allows employees to explore relationships between data and find relevant resources.
In this webinar we discuss the primary use cases for Graph Databases and explore the properties of Neo4j that make those use cases possible.
We cover the high-level steps of modeling, importing, and querying your data using Cypher and give an overview of the transition from RDBMS to Graph.
The year of the graph: do you really need a graph database? How do you choose...George Anadiotis
Graph databases have been around for more than 15 years, but it was AWS and Microsoft getting in the domain that attracted widespread interest. If they are into this, there must be a reason.
Everyone wants to know more, few can really keep up and provide answers. And as this hitherto niche domain is in the mainstream now, the dynamics are changing dramatically. Besides new entries, existing players keep evolving.
I’ve done the hard work of evaluating solutions, so you don’t have to. An overview of the domain and selection methodology, as presented in Big Data Spain 2018
Relational databases were conceived to digitize paper forms and automate well-structured business processes, and still have their uses. But, oftentimes with RDBMS, performance degrades with the increasing number and levels of data relationships and data size.
A graph database like Neo4j naturally stores, manages, analyzes, and uses data within the context of connections meaning Neo4j provides faster query performance and vastly improved flexibility in handling complex hierarchies than SQL.
This webinar explains why companies are shifting away from RDBMS towards graphs to unlock the business value in their data relationships.
This document discusses graphs and graph databases. It provides examples of graphs and compares SQL queries to Gremlin queries on graphs. It also discusses different types of graph databases for online transaction processing (OLTP) and online analytical processing (OLAP). The document then discusses how a social and data graph could help address the problem of data going dark in life sciences research by enabling collaboration, data sharing and discovery of relevant experts and data. It proposes using bi-clustering algorithms to identify relevant groups within the social and data graph to facilitate data and expert discovery.
Big Graph Analytics on Neo4j with Apache SparkKenny Bastani
In this talk I will introduce you to a Docker container that provides you an easy way to do distributed graph processing using Apache Spark GraphX and a Neo4j graph database. You'll learn how to analyze big data graphs that are exported from Neo4j and consequently updated from the results of a Spark GraphX analysis. The types of analysis I will be talking about are PageRank, connected components, triangle counting, and community detection.
Database technologies have evolved to be able to store big data, but are largely inflexible. For complex graph data models stored in a relational database there may be tedious transformations and shuffling around of data to perform large scale analysis.
Fast and scalable analysis of big data has become a critical competitive advantage for companies. There are open source tools like Apache Hadoop and Apache Spark that are providing opportunities for companies to solve these big data problems in a scalable way. Platforms like these have become the foundation of the big data analysis movement.
Speakers
The document discusses H2O.ai's Driverless AI product, which aims to automate and simplify the machine learning process. It provides an overview of H2O.ai as a company, their goals of operationalizing data science. Driverless AI uses techniques like automated feature engineering, model tuning and selection, and model ensembling to build accurate models fast. It also allows for interpreting and explaining machine learning models through features like model inspection and reason codes. A demo of Driverless AI predicting credit card default risk is shown to illustrate the system.
3. Relationships Matter: Using Connected Data for Better Machine LearningNeo4j
The document discusses how graph databases and graph data science can be used to enhance machine learning models by incorporating relationship data. It provides examples of how organizations are using Neo4j's graph data science platform to improve predictive models in areas like fraud detection, health outcomes, and supply chain reliability. The platform includes over 50 graph algorithms, graph-native machine learning workflows, and the ability to train, apply, and manage predictive models on graph data.
This webinar focuses on the particular use case of graph databases in Network & IT-Management. This webinar is designed for people who work with Network Management at telecom companies or professionals within industries that handle and rely on complex networks.
We’ll start with an overview of Neo4j and Graph-thinking within Networks, explaining how Neworks are naturally modelled as graphs. We’ll explain how graph databases vastly help mitigate some of the major challenges the Network and Security Managers face on daily basis — including intrusions and other cyber crimes, performance optimization, outage simulations, fraud prevention and more.
Introduction to the graph technologies landscapeLinkurious
Graph technologies allow modeling of complex relationships and connections through nodes and edges. There are three main layers of graph technologies: graph databases to store graph data, graph analysis frameworks to analyze large graphs, and graph visualization solutions to interact with graphs. Popular tools in each layer include Neo4j and Titan for databases, Giraph and GraphX for analysis, and Gephi and Cytoscape for visualization. Graph technologies are gaining more attention due to their ability to extract insights from connected data.
What is graph all about, and why should you care? Graphs come in many shapes and forms, and can be used for different applications: Graph Analytics, Graph AI, Knowledge Graphs, and Graph Databases.
Talk by George Anadiotis. Connected Data London Meetup June 29th 2020.
Up until the beginning of the 2010s, the world was mostly running on spreadsheets and relational databases. To a large extent, it still does. But the NoSQL wave of databases has largely succeeded in instilling the “best tool for the job” mindset.
After relational, key-value, document, and columnar, the latest link in this evolutionary proliferation of data structures is graph. Graph analytics, Graph AI, Knowledge Graphs and Graph Databases have been making waves, included in hype cycles for the last couple of years.
The Year of the Graph marked the beginning of it all before the Gartners of the world got in the game. The Year of the Graph is a term coined to convey the fact that the time has come for this technology to flourish.
The eponymous article that set the tone was published in January 2018 on ZDNet by domain expert George Anadiotis. George has been working with, and keeping an eye on, all things Graph since the early 2000s. He was one of the first to note the continuing rise of Graph Databases, and to bring this technology in front of a mainstream audience.
The Year of the Graph has been going strong since 2018. In August 2018, Gartner started including Graph in its hype cycles. Ever since, Graph has been riding the upward slope of the Hype Cycle.
The need for knowledge on these technologies is constantly growing. To respond to that need, the Year of the Graph newsletter was released in April 2018. In addition, a constant flow of graph-related news and resources is being shared on social media.
To help people make educated choices, the Year of the Graph Database Report was released. The report has been hailed as the most comprehensive of its kind in the market, consistently helping people choose the most appropriate solution for their use case since 2018.
The report, articles, news stream, and the newsletter have been reaching thousands of people, helping them understand and navigate this landscape. We’ll talk about the Year of the Graph, the different shapes, forms, and applications for graphs, the latest news and trends, and wrap up with an ask me anything session.
This document summarizes a live webinar about creating and querying a graph database of Olympic data. It describes loading data on athletes, countries, sports, events and medals from 1896-2012 into a Neo4j graph database. It then demonstrates several example queries of the Olympic graph, such as the number of sports per games, medals per country per sport, and athletes who medaled in multiple sports.
These webinar slides are an introduction to Neo4j and Graph Databases. They discuss the primary use cases for Graph Databases and the properties of Neo4j which make those use cases possible. They also cover the high-level steps of modeling, importing, and querying your data using Cypher and touch on RDBMS to Graph.
DataStax | Meaningful User Experience with Graph Data (Chris Lacava, Expero) ...DataStax
Congratulations, your data is up and running in a graph database! This is the first step of many to unlocking the potential in your data. It’s easy to get mired in the complexities of graph technology and forget that real users, mere mortals, will need to use this information to inform mission critical tasks. To get the value out of your graph investment, you’ll need to provide an experience that enables users to explore and visualize your graph data in meaningful ways. In this talk we’ll take a hands on approach to applying user-centered strategies and leveraging the latest UI tools to rapidly create great experiences with graph data. Topics will include:
Tailoring experiences to the intended audience and data
Interacting with complex data shouldn’t be complicated for users. The key is to understand your users and build a solution that targets them.
Zeroing in on user goals and creating a solution that targets them in a fast, lightweight and iterative way
Using live data and rapid prototyping to inform your navigation, visualization selections and overall design
Determining the the right visualization for the job
Just because you can display almost anything doesn’t mean you should. Choosing the right visualizations to achieve specific goals is a key factor in unlocking the usefulness of graph. We’ll demonstrate how to match the right visualization against user needs.
When is it appropriate to break out of the standard node view and visualize graph data in another context like a geospatial map?
How do I determine the dominant dimensions to filter a node chart?
What is the simplest, most efficient way to traverse time with large data sets?
Do I need to visually expose the graph at all?
Cutting through the clutter on choosing the right visualization tools
Once you’ve got your goals set, how do you make it happen? Will it perform at scale? We’ll demonstrate example use cases with sample graph data and some of these tools to highlight practical uses.
Cheryl Wiebe - Advanced Analytics in the Industrial WorldRehgan Avon
2018 Women in Analytics Conference
https://www.womeninanalytics.org/
Cheryl will talk about her consulting practice in Industrial Solutions, Analytic solutions for industrial IoT-enabled businesses, including connected factory, connected supply chain, smart mobility, connected assets. Her path to this practice has bounced between hands on systems development, IT strategy, business process reengineering, supply chain analytics, manufacturing quality analytics, and now Industrial IoT analytics. She spent time working in industry as a developer, as a management consultant, started and sold a company, before settling in to pursue this topic as a career analytics consultant. Cheryl will shed light on what's happening in industrial companies struggling to make the transition to digital, what that means, and what barriers they're challenged with. She'll touch on how/where artificial intelligence, deep learning, and machine learning technologies are being used most effectively in industrial companies, and what are the unique challenges they are facing. Reflecting on what's changed over the years, and her journey to witness this, Cheryl will pose what she considers important ideas to consider for women (and men) in pursuing an analytics career successfully and meaningfully.
This document provides an overview of Think Big Analytics, an analytics consulting firm. It discusses their services portfolio including data engineering, data science, analytics operations and managed services. It also highlights their global delivery model and successful projects with over 100 clients. The document then discusses their approach to artificial intelligence and deep learning, including applications across industries like banking, connected cars, and automated check processing. It emphasizes the need for a phased implementation approach to AI and challenges around technology, data, and deployment.
Should a Graph Database Be in Your Next Data Warehouse Stack?Cambridge Semantics
This document discusses whether a graph database should be part of an organization's next data warehouse stack. It introduces Anzo, a managed data fabric that connects and blends enterprise data, and AnzoGraph, a data mart that can be embedded in applications or used with ETL and visualization tools. Graph databases are growing in popularity due to their ability to ask complex questions across complex data in a way that is not possible with SQL. Graph databases also support powerful graph algorithms and analysis not found in SQL. AnzoGraph is a graph database that provides standard analytics capabilities along with graph-specific features.
Democratizing Intelligence - Sri Ambati, CEO & Co-Founder, H2O.aiSri Ambati
Presented at #H2OWorld 2017 in Mountain View, CA.
Enjoy the video: https://youtu.be/ZrlJQqNaSMI.
Learn more about H2O.ai: https://www.h2o.ai/.
Follow @h2oai: https://www.twitter.com/h2oai.
The document discusses scaling tribal knowledge at Airbnb by building a graph database of the company's data resources. It describes collecting metadata on over 6,000 charts, dashboards, experiments and other data assets from various systems into a Neo4j graph database using Airflow. The graph is indexed in Elasticsearch for fast search. This allows employees to explore relationships between data and find relevant resources.
In this webinar we discuss the primary use cases for Graph Databases and explore the properties of Neo4j that make those use cases possible.
We cover the high-level steps of modeling, importing, and querying your data using Cypher and give an overview of the transition from RDBMS to Graph.
The year of the graph: do you really need a graph database? How do you choose...George Anadiotis
Graph databases have been around for more than 15 years, but it was AWS and Microsoft getting in the domain that attracted widespread interest. If they are into this, there must be a reason.
Everyone wants to know more, few can really keep up and provide answers. And as this hitherto niche domain is in the mainstream now, the dynamics are changing dramatically. Besides new entries, existing players keep evolving.
I’ve done the hard work of evaluating solutions, so you don’t have to. An overview of the domain and selection methodology, as presented in Big Data Spain 2018
Relational databases were conceived to digitize paper forms and automate well-structured business processes, and still have their uses. But, oftentimes with RDBMS, performance degrades with the increasing number and levels of data relationships and data size.
A graph database like Neo4j naturally stores, manages, analyzes, and uses data within the context of connections meaning Neo4j provides faster query performance and vastly improved flexibility in handling complex hierarchies than SQL.
This webinar explains why companies are shifting away from RDBMS towards graphs to unlock the business value in their data relationships.
This document discusses graphs and graph databases. It provides examples of graphs and compares SQL queries to Gremlin queries on graphs. It also discusses different types of graph databases for online transaction processing (OLTP) and online analytical processing (OLAP). The document then discusses how a social and data graph could help address the problem of data going dark in life sciences research by enabling collaboration, data sharing and discovery of relevant experts and data. It proposes using bi-clustering algorithms to identify relevant groups within the social and data graph to facilitate data and expert discovery.
Big Graph Analytics on Neo4j with Apache SparkKenny Bastani
In this talk I will introduce you to a Docker container that provides you an easy way to do distributed graph processing using Apache Spark GraphX and a Neo4j graph database. You'll learn how to analyze big data graphs that are exported from Neo4j and consequently updated from the results of a Spark GraphX analysis. The types of analysis I will be talking about are PageRank, connected components, triangle counting, and community detection.
Database technologies have evolved to be able to store big data, but are largely inflexible. For complex graph data models stored in a relational database there may be tedious transformations and shuffling around of data to perform large scale analysis.
Fast and scalable analysis of big data has become a critical competitive advantage for companies. There are open source tools like Apache Hadoop and Apache Spark that are providing opportunities for companies to solve these big data problems in a scalable way. Platforms like these have become the foundation of the big data analysis movement.
Speakers
The document discusses H2O.ai's Driverless AI product, which aims to automate and simplify the machine learning process. It provides an overview of H2O.ai as a company, their goals of operationalizing data science. Driverless AI uses techniques like automated feature engineering, model tuning and selection, and model ensembling to build accurate models fast. It also allows for interpreting and explaining machine learning models through features like model inspection and reason codes. A demo of Driverless AI predicting credit card default risk is shown to illustrate the system.
3. Relationships Matter: Using Connected Data for Better Machine LearningNeo4j
The document discusses how graph databases and graph data science can be used to enhance machine learning models by incorporating relationship data. It provides examples of how organizations are using Neo4j's graph data science platform to improve predictive models in areas like fraud detection, health outcomes, and supply chain reliability. The platform includes over 50 graph algorithms, graph-native machine learning workflows, and the ability to train, apply, and manage predictive models on graph data.
This webinar focuses on the particular use case of graph databases in Network & IT-Management. This webinar is designed for people who work with Network Management at telecom companies or professionals within industries that handle and rely on complex networks.
We’ll start with an overview of Neo4j and Graph-thinking within Networks, explaining how Neworks are naturally modelled as graphs. We’ll explain how graph databases vastly help mitigate some of the major challenges the Network and Security Managers face on daily basis — including intrusions and other cyber crimes, performance optimization, outage simulations, fraud prevention and more.
Introduction to the graph technologies landscapeLinkurious
Graph technologies allow modeling of complex relationships and connections through nodes and edges. There are three main layers of graph technologies: graph databases to store graph data, graph analysis frameworks to analyze large graphs, and graph visualization solutions to interact with graphs. Popular tools in each layer include Neo4j and Titan for databases, Giraph and GraphX for analysis, and Gephi and Cytoscape for visualization. Graph technologies are gaining more attention due to their ability to extract insights from connected data.
What is graph all about, and why should you care? Graphs come in many shapes and forms, and can be used for different applications: Graph Analytics, Graph AI, Knowledge Graphs, and Graph Databases.
Talk by George Anadiotis. Connected Data London Meetup June 29th 2020.
Up until the beginning of the 2010s, the world was mostly running on spreadsheets and relational databases. To a large extent, it still does. But the NoSQL wave of databases has largely succeeded in instilling the “best tool for the job” mindset.
After relational, key-value, document, and columnar, the latest link in this evolutionary proliferation of data structures is graph. Graph analytics, Graph AI, Knowledge Graphs and Graph Databases have been making waves, included in hype cycles for the last couple of years.
The Year of the Graph marked the beginning of it all before the Gartners of the world got in the game. The Year of the Graph is a term coined to convey the fact that the time has come for this technology to flourish.
The eponymous article that set the tone was published in January 2018 on ZDNet by domain expert George Anadiotis. George has been working with, and keeping an eye on, all things Graph since the early 2000s. He was one of the first to note the continuing rise of Graph Databases, and to bring this technology in front of a mainstream audience.
The Year of the Graph has been going strong since 2018. In August 2018, Gartner started including Graph in its hype cycles. Ever since, Graph has been riding the upward slope of the Hype Cycle.
The need for knowledge on these technologies is constantly growing. To respond to that need, the Year of the Graph newsletter was released in April 2018. In addition, a constant flow of graph-related news and resources is being shared on social media.
To help people make educated choices, the Year of the Graph Database Report was released. The report has been hailed as the most comprehensive of its kind in the market, consistently helping people choose the most appropriate solution for their use case since 2018.
The report, articles, news stream, and the newsletter have been reaching thousands of people, helping them understand and navigate this landscape. We’ll talk about the Year of the Graph, the different shapes, forms, and applications for graphs, the latest news and trends, and wrap up with an ask me anything session.
This document summarizes a live webinar about creating and querying a graph database of Olympic data. It describes loading data on athletes, countries, sports, events and medals from 1896-2012 into a Neo4j graph database. It then demonstrates several example queries of the Olympic graph, such as the number of sports per games, medals per country per sport, and athletes who medaled in multiple sports.
These webinar slides are an introduction to Neo4j and Graph Databases. They discuss the primary use cases for Graph Databases and the properties of Neo4j which make those use cases possible. They also cover the high-level steps of modeling, importing, and querying your data using Cypher and touch on RDBMS to Graph.
DataStax | Meaningful User Experience with Graph Data (Chris Lacava, Expero) ...DataStax
Congratulations, your data is up and running in a graph database! This is the first step of many to unlocking the potential in your data. It’s easy to get mired in the complexities of graph technology and forget that real users, mere mortals, will need to use this information to inform mission critical tasks. To get the value out of your graph investment, you’ll need to provide an experience that enables users to explore and visualize your graph data in meaningful ways. In this talk we’ll take a hands on approach to applying user-centered strategies and leveraging the latest UI tools to rapidly create great experiences with graph data. Topics will include:
Tailoring experiences to the intended audience and data
Interacting with complex data shouldn’t be complicated for users. The key is to understand your users and build a solution that targets them.
Zeroing in on user goals and creating a solution that targets them in a fast, lightweight and iterative way
Using live data and rapid prototyping to inform your navigation, visualization selections and overall design
Determining the the right visualization for the job
Just because you can display almost anything doesn’t mean you should. Choosing the right visualizations to achieve specific goals is a key factor in unlocking the usefulness of graph. We’ll demonstrate how to match the right visualization against user needs.
When is it appropriate to break out of the standard node view and visualize graph data in another context like a geospatial map?
How do I determine the dominant dimensions to filter a node chart?
What is the simplest, most efficient way to traverse time with large data sets?
Do I need to visually expose the graph at all?
Cutting through the clutter on choosing the right visualization tools
Once you’ve got your goals set, how do you make it happen? Will it perform at scale? We’ll demonstrate example use cases with sample graph data and some of these tools to highlight practical uses.
Cheryl Wiebe - Advanced Analytics in the Industrial WorldRehgan Avon
2018 Women in Analytics Conference
https://www.womeninanalytics.org/
Cheryl will talk about her consulting practice in Industrial Solutions, Analytic solutions for industrial IoT-enabled businesses, including connected factory, connected supply chain, smart mobility, connected assets. Her path to this practice has bounced between hands on systems development, IT strategy, business process reengineering, supply chain analytics, manufacturing quality analytics, and now Industrial IoT analytics. She spent time working in industry as a developer, as a management consultant, started and sold a company, before settling in to pursue this topic as a career analytics consultant. Cheryl will shed light on what's happening in industrial companies struggling to make the transition to digital, what that means, and what barriers they're challenged with. She'll touch on how/where artificial intelligence, deep learning, and machine learning technologies are being used most effectively in industrial companies, and what are the unique challenges they are facing. Reflecting on what's changed over the years, and her journey to witness this, Cheryl will pose what she considers important ideas to consider for women (and men) in pursuing an analytics career successfully and meaningfully.
This document provides an overview of Think Big Analytics, an analytics consulting firm. It discusses their services portfolio including data engineering, data science, analytics operations and managed services. It also highlights their global delivery model and successful projects with over 100 clients. The document then discusses their approach to artificial intelligence and deep learning, including applications across industries like banking, connected cars, and automated check processing. It emphasizes the need for a phased implementation approach to AI and challenges around technology, data, and deployment.
CTO Radshow Hamburg17 - Keynote - The CxO responsibilities in Big Data and AI...Santiago Cabrera-Naranjo
When talking about how the future of Big Data will look like, this conversation often turns straight to Artificial Intelligence and Deep Learning. However, today data science is all too often a process where new insights and models get developed as a one-time effort or deployed to production on an ad-hoc basis i.e. they commonly require regular babysitting for monitoring and updating.
According to Gartner, the number of useless Data Lakes will be of 90% in 2018. Furthermore, only 15% of Big Data Products are mature enough to be deployed into Production - Who is responsible to make Big Data successful and Business relevant within an enterprise?
Webinar - The Agility Challenge - Powering Cloud Apps with Multi-Model & Mixe...DataStax
Building and managing cloud applications is not easy. Teams come face to face with these challenges: agility, manageability, performance, scalability, continuous availability and of course, security. Join us for “The Agility Challenge: Powering Cloud Applications with Multi-Model & Mixed Workloads” webinar where we will deep dive into challenges customers face with multiple data models such as graph, mixed workloads and how DataStax Enterprise can help.
Video: https://youtu.be/1tKDxkexzFE
Artificial intelligence is becoming a hot topic due to recent advances in hardware capabilities, neural networks research, and technology investments. Deep learning is driving this resurgence by using neural networks with multiple layers to interpret nonlinear relationships in high-dimensional data. Deep learning is delivering improved performance on complex problems and creating value with little domain knowledge required. The presentation provides examples of AI applications in industries like banking, automotive, and healthcare. It also outlines steps to get started with an AI pilot project and developing an AI strategy and roadmap.
Jupyter in the modern enterprise data and analytics ecosystem Gerald Rousselle
Gerald Rousselle presented on Jupyter in the modern enterprise analytical ecosystem. He discussed how Jupyter can help provide a unified access experience to manage increasing data complexity and enable collaboration. Jupyter is emerging as a technology to solve challenges around access, collaboration, and managing complexity. Rousselle showed how Jupyter is moving beyond data science into business analytics by extending its capabilities with tools like a SQL extension. Key takeaways were that Jupyter will be a central part of analytical ecosystems, help democratize access, and is more than just notebooks through its open source protocols.
This document discusses using LSTM neural networks for time series prediction. It provides examples of using LSTM to predict traffic times based on historical minute-by-minute traffic data and evaluating predictions against actual times. The document also discusses data preparation steps like feature engineering and dimensionality reduction needed before using LSTM on different types of time series data like text, images, or numerical values.
Building a Data Platform Strata SF 2019mark madsen
Building a data lake involves more than installing Hadoop or putting data into AWS. The goal in most organizations is to build multi-use data infrastructure that is not subject to past constraints. This tutorial covers design assumptions, design principles, and how to approach the architecture and planning for multi-use data infrastructure in IT.
[This is a new, changed version of the presentations of the same title from last year's Strata]
Designing the Next Generation Data LakeRobert Chong
This document contains a presentation by George Trujillo on designing the next generation data lake. It discusses how analytic platforms need to change to keep up with business demands. New technologies like cloud, object storage, and self-driving databases are allowing for more flexible and scalable data architectures. This is shifting analytics platforms from tightly coupled storage and compute to independent, elastic models. These changes will impact how organizations build projects, careers, and skills in the future by focusing more on innovation and delivering results faster.
Dr. Stefan Radtke gave a presentation on the journey to big data analytics. He discussed how analytics is affecting many industries and the evolution of analytic questions from descriptive to predictive to prescriptive. He emphasized the need to collect all potential data from both traditional and new sources. A strategic approach was presented that aligns business and IT goals, identifies strategic opportunities, prioritizes use cases, and recommends an analytics roadmap. Dell EMC offers various services to help customers with their big data and analytics initiatives and solutions.
Apache spark empowering the real time data driven enterprise - StreamAnalytix...Impetus Technologies
Apache Spark is one of the most popular Big Data frameworks today. It is fast becoming the de facto technology choice for stream processing, real-time analytics, data science and machine learning applications at scale. It has moved well beyond the early-adopter phase, is supported by a vibrant open source community and is enjoying accelerated adoption in enterprises.
Join our guest speaker from Forrester Research, VP & Principal Analyst, Mike Gualtieri and StreamAnalytix, Product Head, Anand Venugopal for a discussion on the trends and directions defining the growing importance of Apache Spark for stream processing, machine learning and other advanced data analytics applications.
DataOps: An Agile Method for Data-Driven OrganizationsEllen Friedman
DataOps expands DevOps philosophy to include data-heavy roles (data engineering & data science). DataOps uses better cross-functional collaboration for flexibility, fast time to value and an agile workflow for data-intensive applications including machine learning pipelines. (Strata Data San Jose March 2018)
Ext JS provides solutions for building data-intensive web applications. It addresses problems with displaying, analyzing, visualizing, and exploring large datasets. Features like buffered grids and pivot grids help optimize performance. Pre-built charts, the D3 adapter, and data binding allow custom visualization and rich user interaction. Ext JS robustly handles huge datasets and improves productivity over alternatives.
This document discusses a method called "data fingerprinting" to represent data as signatures that capture the underlying structure and semantics. It presents two case studies applying this method to question complexity analysis and image recognition with limited data. The method uses autoencoders trained on clustered data to extract and encode structural patterns, allowing data-hungry machine learning algorithms to be used for "small-data" applications. Evaluation results demonstrate it can accurately classify new data types not seen during training.
This document provides an overview of data science and its applications. It discusses:
1) Industries that are being disrupted by data science like telecom, banking, retail, and healthcare.
2) How companies like Amazon, Netflix, and Google were able to disrupt their industries through their ability to analyze patterns in data faster than competitors.
3) The factors driving more companies to adopt data science including competitive advantages, revenue growth, and cost optimization.
Saama Presents Is your Big Data Solution Ready for StreamingSaama
This document discusses how pharmaceutical companies can learn from other industries' use of IoT and streaming data. It outlines how edge computing works and considerations for clinical trials. Examples of IoT use cases in clinical trials are provided. The document also discusses ensuring a big data platform is ready to handle IoT and streaming data from various sources and devices. It emphasizes distributed architectures and cloud solutions.
The Maturity Model: Taking the Growing Pains Out of HadoopInside Analysis
The Briefing Room with Rick van der Lans and Think Big, a Teradata Company
Live Webcast on June 16, 2015
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=197f8106531874cc5c14081ca214eaff
Hadoop is arguably one of the most disruptive technologies of the last decade. Once lauded solely for its ability to transform the speed of batch processing, it has marched steadily forward and promulgated an array of performance-enhancing accessories, notably Spark and YARN. Hadoop has evolved into much more than a file system and batch processor, and it now promises to stand as the data management and analytics backbone for enterprises.
Register for this episode of The Briefing Room to learn from veteran Analyst Rick van der Lans, as he discusses the emerging roles of Hadoop within the analytics ecosystem. He’ll be briefed by Ron Bodkin of Think Big, a Teradata Company, who will explore Hadoop’s maturity spectrum, from typical entry use cases all the way up the value chain. He’ll show how enterprises that already use Hadoop in production are finding new ways to exploit its power and build creative, dynamic analytics environments.
Visit InsideAnalysis.com for more information.
1. The document discusses Teradata Vantage, a data analytics platform. It introduces Patrick Deglon, VP of Advanced Analytics at Teradata, and covers various topics around machine learning, analytics, and data management challenges.
2. An example is provided of an experiment run by eBay to test the effectiveness of marketing campaigns. Results showed increased purchases when marketing was used.
3. Teradata Vantage is positioned as a solution that can help organizations overcome challenges around data management, analytics, and machine learning by providing unified analytics capabilities across different workloads and data types.
Data science a practitioner's perspectiveAmir Ziai
The document provides an overview of data science from the perspective of a practitioner at ZEFR, an ad tech company. It discusses the history and growth of data science, common pitfalls, and the minimum skills required, including experience with SQL, NoSQL, machine learning frameworks, cloud computing, and software engineering best practices. It emphasizes the importance of understanding problems, communicating findings, and automating/scaling solutions given the petabyte-scale of data at ZEFR.
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...sameer shah
"Join us for STATATHON, a dynamic 2-day event dedicated to exploring statistical knowledge and its real-world applications. From theory to practice, participants engage in intensive learning sessions, workshops, and challenges, fostering a deeper understanding of statistical methodologies and their significance in various fields."
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
1. What you need to know to start an AI company?
Mo Patel, Practice Director
Artificial Intelligence & Machine Learning
Teradata Analytics
Tech Club Lunch & Learn
November 10, 2016
The Mack Institute
The Wharton School
University of Pennsylvania
Unemployment photo: Wikipedia – The Depression
Bias photo: http://psych.nyu.edu/freemanlab/research.htm
Security photo: Transcendence movie poster
Machine Human Photo: Shutterstock licensed by Teradata
Safety photo: Shutterstock via Google Image Search