This document provides an overview of Anzo Unstructured, a natural language processing (NLP) platform from Cambridge Semantics. It discusses the core capabilities of Anzo Unstructured, including intake of various file formats, extraction of entities and relationships, and semantic analysis. It also outlines example use cases in pharma and finance. The document demonstrates the configuration and visualization of Anzo Unstructured pipelines and annotations.
Transforming Data Management and Time to Insight with Anzo Smart Data Lake®Cambridge Semantics
This webinar is targeted to Federal Government CIOs and
staff that are researching enterprise data management and
mining tools to help them understand how Smart Data Lakes
enable a viable mechanism for addressing their top priorities.
Accelerating Insight - Smart Data Lake Customer Success StoriesCambridge Semantics
At Gartner Data & Analytics Summit 2017 Alok Prasad, President, was joined by Peter Horowitz of PricewaterhouseCoopers in presenting a session on how Cambridge Semantics' in-memory, massively parallel, semantic graph-based platform delivers an accelerating edge to data-driven organizations, while maintaining trust with security and governance.
Graph technology has truly burst onto the scene with diverse new products and services, proving that graph is relevant and that not all graph use cases are equal. Previously relegated to niche implementations and science projects, graph now finds itself deployed as the foundational technology for enterprise analytics solutions and enterprise Data Fabric strategies. It is no surprise that many are calling 2018 “The Year of the Graph”.
How can organizations give up the keys to data systems without creating data anarchy? The answer lies in Smart Data Lakes™. Learn how Smart Data Lakes are being used to design contextual data platforms for deeper insights and problem solving, responsibly and effectively introduce self-service independence from IT, put subject matter expertise to work overcoming volume and variety challenges and enable a backbone of collaboration and sharing to improve data and insights.
A data lake promises cheap storage and ubiquitous access for all of your enterprise data. However, most organizations are struggling to make sense of the data in the lake. How do you harmonize, add meaning, govern, secure and offer business self-service to your data lake? You build a Smart Data Lake.
Applying Data Engineering and Semantic Standards to Tame the "Perfect Storm" ...Cambridge Semantics
The financial industry is facing a perfect storm of disruptive drivers for data management. While regulators seek accuracy and transparency, institutions are struggling with fragmented data and IT infrastructures. The path forward is “data engineering” – applying consistent semantics with scalable infrastructure to harmonize data and enable traceable and dynamic analytics. In this webinar, we hear from industry practitioners and thought leaders on how this vision is being deployed and also see it in action.
From Data Lakes to the Data Fabric: Our Vision for Digital StrategyCambridge Semantics
In this presentation for Strata NY 2018, we share our vision for digital innovation as a shift to something powerful, expedient and future-proof. This is accomplished through the use of a 'Data Fabric'. Utilizing graph technology, this Data Fabric connects enterprise data in an overlay fashion that does not disrupt current investments for unprecedented access to data. This interconnected and reliable data can then be used to automate scalable AI and ML efforts to improve business outcomes.
Retail banks are moving beyond the data warehouse and data lake and are now implementing data fabric architectures to address data discovery and integration challenges.
These are the slides from our webinar "Modern Data Discovery and Integration in Retail Banking" in which we explore the role of the data discovery and integration layer in a data fabric with special focus on evolution from data warehouse to data fabric, semantics and graph data models in data fabric and example use cases in retail banks and B2C financial services.
Transforming Data Management and Time to Insight with Anzo Smart Data Lake®Cambridge Semantics
This webinar is targeted to Federal Government CIOs and
staff that are researching enterprise data management and
mining tools to help them understand how Smart Data Lakes
enable a viable mechanism for addressing their top priorities.
Accelerating Insight - Smart Data Lake Customer Success StoriesCambridge Semantics
At Gartner Data & Analytics Summit 2017 Alok Prasad, President, was joined by Peter Horowitz of PricewaterhouseCoopers in presenting a session on how Cambridge Semantics' in-memory, massively parallel, semantic graph-based platform delivers an accelerating edge to data-driven organizations, while maintaining trust with security and governance.
Graph technology has truly burst onto the scene with diverse new products and services, proving that graph is relevant and that not all graph use cases are equal. Previously relegated to niche implementations and science projects, graph now finds itself deployed as the foundational technology for enterprise analytics solutions and enterprise Data Fabric strategies. It is no surprise that many are calling 2018 “The Year of the Graph”.
How can organizations give up the keys to data systems without creating data anarchy? The answer lies in Smart Data Lakes™. Learn how Smart Data Lakes are being used to design contextual data platforms for deeper insights and problem solving, responsibly and effectively introduce self-service independence from IT, put subject matter expertise to work overcoming volume and variety challenges and enable a backbone of collaboration and sharing to improve data and insights.
A data lake promises cheap storage and ubiquitous access for all of your enterprise data. However, most organizations are struggling to make sense of the data in the lake. How do you harmonize, add meaning, govern, secure and offer business self-service to your data lake? You build a Smart Data Lake.
Applying Data Engineering and Semantic Standards to Tame the "Perfect Storm" ...Cambridge Semantics
The financial industry is facing a perfect storm of disruptive drivers for data management. While regulators seek accuracy and transparency, institutions are struggling with fragmented data and IT infrastructures. The path forward is “data engineering” – applying consistent semantics with scalable infrastructure to harmonize data and enable traceable and dynamic analytics. In this webinar, we hear from industry practitioners and thought leaders on how this vision is being deployed and also see it in action.
From Data Lakes to the Data Fabric: Our Vision for Digital StrategyCambridge Semantics
In this presentation for Strata NY 2018, we share our vision for digital innovation as a shift to something powerful, expedient and future-proof. This is accomplished through the use of a 'Data Fabric'. Utilizing graph technology, this Data Fabric connects enterprise data in an overlay fashion that does not disrupt current investments for unprecedented access to data. This interconnected and reliable data can then be used to automate scalable AI and ML efforts to improve business outcomes.
Retail banks are moving beyond the data warehouse and data lake and are now implementing data fabric architectures to address data discovery and integration challenges.
These are the slides from our webinar "Modern Data Discovery and Integration in Retail Banking" in which we explore the role of the data discovery and integration layer in a data fabric with special focus on evolution from data warehouse to data fabric, semantics and graph data models in data fabric and example use cases in retail banks and B2C financial services.
Accelerate Digital Transformation with an Enterprise Big Data FabricCambridge Semantics
In this webinar by Cambridge Semantics' VP of Solution Engineering, Ben Szekely, you will learn more about how the Enterprise Data Fabric prevails as the bedrock of enterprise digital strategy. Connected and highly available data is the new normal - powering analytics and AI. The data lake itself is commoditized, like raw compute or disk, and becomes an unseen part of the stack. Semantic graph technology is central to Data Fabric initiatives that meaningfully contribute to digital transformation.
We share our vision for digital innovation - a shift to something powerful, expedient and future-proof. The Data Fabric connects enterprise data for unprecedented access in an overlay fashion that does not disrupt current investments. Interconnected and reliable data drives business outcomes by automating scalable AI and ML efforts. Graph technology is the way forward to realize this future.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
Using Cloud Automation Technologies to Deliver an Enterprise Data FabricCambridge Semantics
The world of database management is changing. Cloud adoption is accelerating, offering a path for companies to increase their database capabilities while keeping costs in line. To help IT decision-makers survive and thrive in the cloud era, DBTA hosted this special roundtable webinar.
Building A Self Service Analytics Platform on HadoopCraig Warman
These slides were presented by Avinash Ramineni of Clairvoyant to the Atlanta Apache Spark User Group on Wednesday, March 22, 2017: https://www.meetup.com/Atlanta-Apache-Spark-User-Group/events/238109721/
Anzo Smart Data Lake 4.0 - a Data Lake Platform for the Enterprise Informatio...Cambridge Semantics
Only with a rich and interactive semantic layer can your data and analytics stack deliver true on-demand access to data, answers and insights - weaving data together from across the enterprise into an information fabric. In this webinar we introduce Anzo Smart Data Lake 4.0, which provides that rich and interactive semantic layer to your data.
In their webinar "Big Data Fabric 2.0 Drives Data Democratization" Ben Szekley, Cambridge Semantics’ SVP of Field Operations, and guest speaker, Forrester’s Noel Yuhanna, author of the Forrester report: “Big Data Fabric 2.0 Drives Data Democratization”, explored why data-driven businesses are making a big data fabric part of their data strategy to minimize data complexity, integrate siloed data, deliver real-time trusted insights, and to create new business opportunities. These are the slides from that webinar.
The most profitable insurance organizations will outperform competitors in key areas as personalized customer service, claims processing, subrogation recovery, fraud detection and product innovation. This requires thinking beyond the traditional data warehouse to the data fabric - an emerging data management architecture.
In this webinar Andy Sohn, Senior Advisor at NewVantage Partners, and Bob Parker, Senior Director for Insurance at Cambridge Semantics, explore the role of the data discovery and integration layer in an enterprise data fabric for the Insurance industry. These are their slides.
Should a Graph Database Be in Your Next Data Warehouse Stack?Cambridge Semantics
In this webinar, AnzoGraph’s graph database guru Barry Zane (former co-founder of Netezza) and data governance author Steve Sarsfield talk about how graph databases fit into the data warehouse modernization trend. They also explore how certain workloads can be better served with an analytical graph database and how today’s technology stacks offer new paradigms for deployment like the cloud, containers and graph analytics.
Sustainability Investment Research Using Cognitive AnalyticsCambridge Semantics
In this webinar Anthony J. Sarkis, Chief Strategy Officer at Parabole, and Steve Sarsfield, VP Product at Cambridge Semantics, explore how portfolio managers are using the recently developed Parabole/ AnzoGraph DB integration as their underlying infrastructure for conducting ML and cognitive analytics at scale to exploit data to identify potential risks and new opportunities.
Graph-driven Data Integration: Accelerating and Automating Data Delivery for ...Cambridge Semantics
In our webinar "A Data Fabric Market Update with Guest Speaker, VP, Principal Analyst Noel Yuhanna" Ben Szekely, Cambridge Semantics’ Co-founder and SVP of Field Operations, and guest speaker, Noel Yuhanna, VP and Principal Analyst at Forrester and author of the “The Forrester Wave™: Enterprise Data Fabric, Q2 2020”, discuss the state of the Data Fabric Market. These are Ben's slides from that webinar.
Knowledge Graph Discussion: Foundational Capability for Data Fabric, Data Int...Cambridge Semantics
Knowledge graphs are on the rise at businesses hungry for greater automation and intelligence with use cases spreading across industries, from fraud detection and chatbots, to risk analysis and recommendation engines. In this webinar we dive into key technical and business considerations, use cases and best practices in leveraging knowledge graphs for better knowledge management.
In this webinar, data analytics gurus Sathish Thyagarajan and Steve Sarsfield introduce AnzoGraph™, our graph OLAP database, demonstrate the different types of analyses you can perform with it and how it complements Neo4j, AWS Neptune and other OLTP systems. Finally, they’ll show how you can get it up and running on your laptop in about 5 minutes.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
AnzoGraph DB: Driving AI and Machine Insights with Knowledge Graphs in a Conn...Cambridge Semantics
Thomas Cook, director of sales, Cambridge Semantics, offers a primer on graph database technology and the rapid growth of knowledge graphs at Data Summit 2020 in his presentation titled "AnzoGraph DB: Driving AI and Machine Insights with Knowledge Graphs in a Connected World".
Semantic Graph Databases: The Evolution of Relational DatabasesCambridge Semantics
In this webinar, Barry Zane, our Vice President of Engineering, discusses the evolution of databases from Relational to Semantic Graph and the Anzo Graph Query Engine, the key element of scale in the Anzo Smart Data Lake. Based on elastic clustered, in-memory computing, the Anzo Graph Query Engine offers interactive ad hoc query and analytics on datasets with billions of triples. With this powerful layer over their data, end users can effect powerful analytic workflows in a self-service manner.
Freddie Mac makes homeownership and rental housing more accessible and affordable. Operating in the secondary mortgage market, we keep mortgage capital flowing by purchasing mortgage loans from lenders so they in turn can provide more loans to qualified borrowers. Our mission to provide liquidity, stability, and affordability to the U.S. housing market in all economic conditions extends to all communities from coast to coast.
We're using big data and advanced analytics to create powerful enhancements to better meet our customer’s needs: automated collateral evaluation, automated assessments for borrowers without credit scores, immediate certainty for collateral rep and warranty relief, and coming soon automated asset and income validation.
We’re building tools to help our customers cut costs and give them rep and warranty relief sooner in the loan manufacturing process.
We’ve designed Loan Advisor Suite with lenders to give our customers greater certainty, usability, reliability and efficiency. It's a simpler, better way to do business.
More Tools - Access powerful solutions for every stage of the loan production process.
More Loans - Increase output with automated data management and user-friendly controls.
Less Risk = Get alerted to loan issues and take action the moment they occur.
Hear the story of how ACE helped Freddie Mac reimagine the mortgage process and how HDP helped make it possible.
Speaker
Dennis Tally, Freddie Mac, Director
Automating Data Science over a Human Genomics Knowledge BaseVaticle
# Automating Data Science over a Human Genomics Knowledge Base
Radouane Oudrhiri, the CTO of Eagle Genomics, will talk about how Eagle Genomics is building a platform for automating data science over a human genomics knowledge base. Rad will dive into the architecture Eagle Genomics and also discuss how Grakn serves as the knowledge base foundation of the system. Rad also give a brief history of databases, semantic expressiveness and how Grakn fits in the big picture.
# Radouane Oudrhiri, CTO, Eagle Genomics
Radouane has an extensive experience in leading world-class software and data-intensive system developments in different industries from Telecom to Healthcare, Nuclear, Automotive, Financials. Radouane is Lean/Six Sigma Master Black Belt with speciality in high-tech, IT and Software engineering and he is recognised as the leader and early adaptor of Lean/Six Sigma and DFSS to IT and Software. He is a fellow of the Royal Statistical Society (RSS) and member of the ISO Technical Committee (TC69: Applications of Statistical methods) where he is co-author of the Lean & Six Sigma Standard (ISO 18404) as well as the new standard under development (Design for Six Sigma). He is also part of the newly formed international Group on Big Data (nominated by BSI as the UK representative/expert). Radouane has also been Chair of the working group on Measurement Systems for Automated Processes/Systems within the ISPE (International Society for Pharmaceutical Engineering).
A Dynamic Data Catalog for Autonomy and Self-ServiceDenodo
Watch Daves' presentation on-demand from Fast Data Strategy Virtual Summit here: https://buff.ly/2Kj7muc
Denodo’s new dynamic catalog is the new black. It combines the power of data delivery infrastructure with data catalog for contextual information and collective intelligence.
Attend this session to discover:
• What is unique about Dynamic Data Catalog?
• How it empowers a community of analysts and decisions makers?
• How it facilitates data curation and data stewardship in your organization?
Necessity of Data Lakes in the Financial Services SectorDataWorks Summit
With the emergence of regulations such as the General Data Protection Regulation from the European Union (effective May 2018), with fines up to 20m Euro, Data Lakes are emerging as the data architecture of choice amongst financial institutions. Banks are embarking on a journey to enable data scientists to unlock the value of the data silo'ed in many disparate data systems. By enabling self service data access and merging multiple streams of data by using data clustering, entity extraction, identity resolution and other techniques - we will show how banks have used Analytics to uncover business value without falling into the abyss of data swamps. The build out of the data lake requires the ingestion of data from multiple operational systems . By leveraging an automated Data Cataloging service, organizations are able to search, profile, discover, tag, track lineage and capture tribal knowledge delivered on the FICO Analytics Cloud enabling the data scientists to build innovative models, make automated decisions, track fraudulent usage, make intelligent marketing campaigns and improve the top line and bottom line for the financial institution.
Speaker:
Rohit Valia, Product Management and Strategy, Fico
PoT - probeer de mogelijkheden van datamining zelf uit 30-10-2014Daniel Westzaan
IBM Proof of Technology
Probeer de Mogelijkheden van Datamining zelf uit
30-10-2014 Amsterdam, IBM Client Center
Presentatie van Laila Fettah & Robin van Tilburg
Presentation given by Appistry's Vice President of Product Strategy, Sultan Meghi at the World Genome Data Analysis Summit. Meghi presented about the big data challenges facing labs as they strive to manage the flow of genetic data from sequencer to the clinic.
Accelerate Digital Transformation with an Enterprise Big Data FabricCambridge Semantics
In this webinar by Cambridge Semantics' VP of Solution Engineering, Ben Szekely, you will learn more about how the Enterprise Data Fabric prevails as the bedrock of enterprise digital strategy. Connected and highly available data is the new normal - powering analytics and AI. The data lake itself is commoditized, like raw compute or disk, and becomes an unseen part of the stack. Semantic graph technology is central to Data Fabric initiatives that meaningfully contribute to digital transformation.
We share our vision for digital innovation - a shift to something powerful, expedient and future-proof. The Data Fabric connects enterprise data for unprecedented access in an overlay fashion that does not disrupt current investments. Interconnected and reliable data drives business outcomes by automating scalable AI and ML efforts. Graph technology is the way forward to realize this future.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
Using Cloud Automation Technologies to Deliver an Enterprise Data FabricCambridge Semantics
The world of database management is changing. Cloud adoption is accelerating, offering a path for companies to increase their database capabilities while keeping costs in line. To help IT decision-makers survive and thrive in the cloud era, DBTA hosted this special roundtable webinar.
Building A Self Service Analytics Platform on HadoopCraig Warman
These slides were presented by Avinash Ramineni of Clairvoyant to the Atlanta Apache Spark User Group on Wednesday, March 22, 2017: https://www.meetup.com/Atlanta-Apache-Spark-User-Group/events/238109721/
Anzo Smart Data Lake 4.0 - a Data Lake Platform for the Enterprise Informatio...Cambridge Semantics
Only with a rich and interactive semantic layer can your data and analytics stack deliver true on-demand access to data, answers and insights - weaving data together from across the enterprise into an information fabric. In this webinar we introduce Anzo Smart Data Lake 4.0, which provides that rich and interactive semantic layer to your data.
In their webinar "Big Data Fabric 2.0 Drives Data Democratization" Ben Szekley, Cambridge Semantics’ SVP of Field Operations, and guest speaker, Forrester’s Noel Yuhanna, author of the Forrester report: “Big Data Fabric 2.0 Drives Data Democratization”, explored why data-driven businesses are making a big data fabric part of their data strategy to minimize data complexity, integrate siloed data, deliver real-time trusted insights, and to create new business opportunities. These are the slides from that webinar.
The most profitable insurance organizations will outperform competitors in key areas as personalized customer service, claims processing, subrogation recovery, fraud detection and product innovation. This requires thinking beyond the traditional data warehouse to the data fabric - an emerging data management architecture.
In this webinar Andy Sohn, Senior Advisor at NewVantage Partners, and Bob Parker, Senior Director for Insurance at Cambridge Semantics, explore the role of the data discovery and integration layer in an enterprise data fabric for the Insurance industry. These are their slides.
Should a Graph Database Be in Your Next Data Warehouse Stack?Cambridge Semantics
In this webinar, AnzoGraph’s graph database guru Barry Zane (former co-founder of Netezza) and data governance author Steve Sarsfield talk about how graph databases fit into the data warehouse modernization trend. They also explore how certain workloads can be better served with an analytical graph database and how today’s technology stacks offer new paradigms for deployment like the cloud, containers and graph analytics.
Sustainability Investment Research Using Cognitive AnalyticsCambridge Semantics
In this webinar Anthony J. Sarkis, Chief Strategy Officer at Parabole, and Steve Sarsfield, VP Product at Cambridge Semantics, explore how portfolio managers are using the recently developed Parabole/ AnzoGraph DB integration as their underlying infrastructure for conducting ML and cognitive analytics at scale to exploit data to identify potential risks and new opportunities.
Graph-driven Data Integration: Accelerating and Automating Data Delivery for ...Cambridge Semantics
In our webinar "A Data Fabric Market Update with Guest Speaker, VP, Principal Analyst Noel Yuhanna" Ben Szekely, Cambridge Semantics’ Co-founder and SVP of Field Operations, and guest speaker, Noel Yuhanna, VP and Principal Analyst at Forrester and author of the “The Forrester Wave™: Enterprise Data Fabric, Q2 2020”, discuss the state of the Data Fabric Market. These are Ben's slides from that webinar.
Knowledge Graph Discussion: Foundational Capability for Data Fabric, Data Int...Cambridge Semantics
Knowledge graphs are on the rise at businesses hungry for greater automation and intelligence with use cases spreading across industries, from fraud detection and chatbots, to risk analysis and recommendation engines. In this webinar we dive into key technical and business considerations, use cases and best practices in leveraging knowledge graphs for better knowledge management.
In this webinar, data analytics gurus Sathish Thyagarajan and Steve Sarsfield introduce AnzoGraph™, our graph OLAP database, demonstrate the different types of analyses you can perform with it and how it complements Neo4j, AWS Neptune and other OLTP systems. Finally, they’ll show how you can get it up and running on your laptop in about 5 minutes.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
AnzoGraph DB: Driving AI and Machine Insights with Knowledge Graphs in a Conn...Cambridge Semantics
Thomas Cook, director of sales, Cambridge Semantics, offers a primer on graph database technology and the rapid growth of knowledge graphs at Data Summit 2020 in his presentation titled "AnzoGraph DB: Driving AI and Machine Insights with Knowledge Graphs in a Connected World".
Semantic Graph Databases: The Evolution of Relational DatabasesCambridge Semantics
In this webinar, Barry Zane, our Vice President of Engineering, discusses the evolution of databases from Relational to Semantic Graph and the Anzo Graph Query Engine, the key element of scale in the Anzo Smart Data Lake. Based on elastic clustered, in-memory computing, the Anzo Graph Query Engine offers interactive ad hoc query and analytics on datasets with billions of triples. With this powerful layer over their data, end users can effect powerful analytic workflows in a self-service manner.
Freddie Mac makes homeownership and rental housing more accessible and affordable. Operating in the secondary mortgage market, we keep mortgage capital flowing by purchasing mortgage loans from lenders so they in turn can provide more loans to qualified borrowers. Our mission to provide liquidity, stability, and affordability to the U.S. housing market in all economic conditions extends to all communities from coast to coast.
We're using big data and advanced analytics to create powerful enhancements to better meet our customer’s needs: automated collateral evaluation, automated assessments for borrowers without credit scores, immediate certainty for collateral rep and warranty relief, and coming soon automated asset and income validation.
We’re building tools to help our customers cut costs and give them rep and warranty relief sooner in the loan manufacturing process.
We’ve designed Loan Advisor Suite with lenders to give our customers greater certainty, usability, reliability and efficiency. It's a simpler, better way to do business.
More Tools - Access powerful solutions for every stage of the loan production process.
More Loans - Increase output with automated data management and user-friendly controls.
Less Risk = Get alerted to loan issues and take action the moment they occur.
Hear the story of how ACE helped Freddie Mac reimagine the mortgage process and how HDP helped make it possible.
Speaker
Dennis Tally, Freddie Mac, Director
Automating Data Science over a Human Genomics Knowledge BaseVaticle
# Automating Data Science over a Human Genomics Knowledge Base
Radouane Oudrhiri, the CTO of Eagle Genomics, will talk about how Eagle Genomics is building a platform for automating data science over a human genomics knowledge base. Rad will dive into the architecture Eagle Genomics and also discuss how Grakn serves as the knowledge base foundation of the system. Rad also give a brief history of databases, semantic expressiveness and how Grakn fits in the big picture.
# Radouane Oudrhiri, CTO, Eagle Genomics
Radouane has an extensive experience in leading world-class software and data-intensive system developments in different industries from Telecom to Healthcare, Nuclear, Automotive, Financials. Radouane is Lean/Six Sigma Master Black Belt with speciality in high-tech, IT and Software engineering and he is recognised as the leader and early adaptor of Lean/Six Sigma and DFSS to IT and Software. He is a fellow of the Royal Statistical Society (RSS) and member of the ISO Technical Committee (TC69: Applications of Statistical methods) where he is co-author of the Lean & Six Sigma Standard (ISO 18404) as well as the new standard under development (Design for Six Sigma). He is also part of the newly formed international Group on Big Data (nominated by BSI as the UK representative/expert). Radouane has also been Chair of the working group on Measurement Systems for Automated Processes/Systems within the ISPE (International Society for Pharmaceutical Engineering).
A Dynamic Data Catalog for Autonomy and Self-ServiceDenodo
Watch Daves' presentation on-demand from Fast Data Strategy Virtual Summit here: https://buff.ly/2Kj7muc
Denodo’s new dynamic catalog is the new black. It combines the power of data delivery infrastructure with data catalog for contextual information and collective intelligence.
Attend this session to discover:
• What is unique about Dynamic Data Catalog?
• How it empowers a community of analysts and decisions makers?
• How it facilitates data curation and data stewardship in your organization?
Necessity of Data Lakes in the Financial Services SectorDataWorks Summit
With the emergence of regulations such as the General Data Protection Regulation from the European Union (effective May 2018), with fines up to 20m Euro, Data Lakes are emerging as the data architecture of choice amongst financial institutions. Banks are embarking on a journey to enable data scientists to unlock the value of the data silo'ed in many disparate data systems. By enabling self service data access and merging multiple streams of data by using data clustering, entity extraction, identity resolution and other techniques - we will show how banks have used Analytics to uncover business value without falling into the abyss of data swamps. The build out of the data lake requires the ingestion of data from multiple operational systems . By leveraging an automated Data Cataloging service, organizations are able to search, profile, discover, tag, track lineage and capture tribal knowledge delivered on the FICO Analytics Cloud enabling the data scientists to build innovative models, make automated decisions, track fraudulent usage, make intelligent marketing campaigns and improve the top line and bottom line for the financial institution.
Speaker:
Rohit Valia, Product Management and Strategy, Fico
PoT - probeer de mogelijkheden van datamining zelf uit 30-10-2014Daniel Westzaan
IBM Proof of Technology
Probeer de Mogelijkheden van Datamining zelf uit
30-10-2014 Amsterdam, IBM Client Center
Presentatie van Laila Fettah & Robin van Tilburg
Presentation given by Appistry's Vice President of Product Strategy, Sultan Meghi at the World Genome Data Analysis Summit. Meghi presented about the big data challenges facing labs as they strive to manage the flow of genetic data from sequencer to the clinic.
An invited talk for Lilly's Global IT Seminar Meeting In November 2016 on the subject of data, machine learning, AI, semantic web, text mining and spinach!
A Journey to Modern Apps with Containers, Microservices and Big DataEdward Hsu
2016-10-04 Reactive Summit - Mesosphere Keynote
Enterprises hear about the promise of application containers, but realizing meaningful business results from containers requires more than abandoning virtual machines. In order to implement containers correctly, businesses must consider the operational implications, as well as the new types of applications they want to build using microservices. In this session, Ed Hsu, Vice President of Enterprise DC/OS at Mesosphere, discusses how to capitalize on new opportunities that can accelerate your IT modernization initiatives.
We hear a lot about lambda architectures and how Cassandra and Spark can help us crunch our data both in batch and real-time. After a year in the trenches, I'll share how we at The Weather Company built a general purpose, weather-scale event processing pipeline to make sense of billions of events each day. If you want to avoid much of the pain learning how to get it right, this talk is for you.
This is a slide deck that was used for our 11/19/15 Nike Tech Talk to give a detailed overview of the SnappyData technology vision. The slides were presented by Jags Ramnarayan, Co-Founder & CTO of SnappyData
This TDWI EU 2012 presentation looks at the various options for implementing a data store for analytical purposes and shows that there's no 'one size fits all' solution available
Always On: Building Highly Available Applications on CassandraRobbie Strickland
Cassandra was built from the ground up to enable linearly scalable, always-on applications. But the path to high availability has many land mines that can mean failure for the inexperienced user. In this talk, I will offer practical advice on how to achieve 100% uptime on millions of transactions per second. I'll address all aspects of the topic, including deployment, configuration, application design, and operations.
NoLambda: Combining Streaming, Ad-Hoc, Machine Learning and Batch AnalysisHelena Edelson
Slides from my talk with Evan Chan at Strata San Jose: NoLambda: Combining Streaming, Ad-Hoc, Machine Learning and Batch Analysis. Streaming analytics architecture in big data for fast streaming, ad hoc and batch, with Kafka, Spark Streaming, Akka, Mesos, Cassandra and FiloDB. Simplifying to a unified architecture.
To date, Hadoop usage has focused primarily on offline analysis--making sense of web logs, parsing through loads of unstructured data in HDFS, etc. But what if you want to run map/reduce against your live data set without affecting online performance? Combining Hadoop with Cassandra's multi-datacenter replication capabilities makes this possible. If you're interested in getting value from your data without the hassle and latency of first moving it into Hadoop, this talk is for you. I'll show you how to connect all the parts, enabling you to write map/reduce jobs or run Pig queries against your live data. As a bonus I'll cover writing map/reduce in Scala, which is particularly well-suited for the task.
Fast and Simplified Streaming, Ad-Hoc and Batch Analytics with FiloDB and Spa...Helena Edelson
O'Reilly Webcast with Myself and Evan Chan on the new SNACK Stack (playoff of SMACK) with FIloDB: Scala, Spark Streaming, Akka, Cassandra, FiloDB and Kafka.
SnappyData, the Spark Database. A unified cluster for streaming, transactions...SnappyData
Apache Spark 2.0 offers many enhancements that make continuous analytics quite simple. In this talk, we will discuss many other things that you can do with your Apache Spark cluster. We explain how a deep integration of Apache Spark 2.0 and in-memory databases can bring you the best of both worlds! In particular, we discuss how to manage mutable data in Apache Spark, run consistent transactions at the same speed as state-the-art in-memory grids, build and use indexes for point lookups, and run 100x more analytics queries at in-memory speeds. No need to bridge multiple products or manage, tune multiple clusters. We explain how one can take regulation Apache Spark SQL OLAP workloads and speed them up by up to 20x using optimizations in SnappyData.
We then walk through several use-case examples, including IoT scenarios, where one has to ingest streams from many sources, cleanse it, manage the deluge by pre-aggregating and tracking metrics per minute, store all recent data in a in-memory store along with history in a data lake and permit interactive analytic queries at this constantly growing data. Rather than stitching together multiple clusters as proposed in Lambda, we walk through a design where everything is achieved in a single, horizontally scalable Apache Spark 2.0 cluster. A design that is simpler, a lot more efficient, and let’s you do everything from Machine Learning and Data Science to Transactions and Visual Analytics all in one single cluster.
Webinar: Lucidworks + Thomson Reuters for Improved Investment PerformanceLucidworks
Learn how the Lucidworks Fusion and Thomson Reuters Intelligent Tagging joint solution can help financial services professionals make better, faster investment decisions.
II-SDV 2017: Localizing International Content for Search, Data Mining and Ana...Dr. Haxel Consult
Advances in text mining, analytics and machine learning are transforming our applications and enabling ever more powerful applications, yet most applications and platforms are designed to deal with a single (normalized) language. Hence as our applications and platforms are increasingly required to ingest international content, the challenge becomes to find ways to normalize content to a single language without compromising quality. An extension of this question in terms of such applications is also how we define quality in this context and what, if any, bi-products a localization effort can produce that may enhance the usefulness of the application.
This talk will, using patent searching as an example use case, review the challenges and possible solution approaches for handling localization effectively and will show what current emerging technology offers, what to expect and what not to expect and provide an introductory practical guide to handling localization in the context of data mining and analytics.
Join Concept Searching and partner C/D/H for this thought-provoking webinar on what intelligent enterprise search should be.
Our solution is unique in the marketplace, and overcomes the limitations of other enterprise search engines. It was originally deployed as an enterprise search solution for engineers and support staff.
This webinar will focus on how one unified view of all unstructured, semi-structured, and structured data assets, including 2D and 3D images, can be integrated into the search interface, with previewers and navigational aids.
Both business and technical professionals will benefit from this session:
• Understand how the technology works, and how it can be set up with a platform and search engine of choice
• See how search returns results, and provides visual and navigational aids for all information retrieved
• Watch how to select an image based on color, size, or shape
• Learn how any business or artificial intelligence applications can benefit from the multi-term metadata created
• Find out why the search framework provides a responsive user interface for any tablet, PC or mobile device
Logo
Linguistic harmony in the Tower of Babel; how Amgen joins reference data from bench to bedside
The challenges associated with managing reference data create complexity and expense across all industries - in Life Sciences, Financial Services and Manufacturing, the problem is particularly acute.
Amgen – an American multinational biopharmaceutical company – tackles the reference data issue by creating a linked vocabulary that is leveraged across the Pharma pipeline to streamline processes and enable faster time to market.
Slides from Enterprise Search & Analytics Meetup @ Cisco Systems - http://www.meetup.com/Enterprise-Search-and-Analytics-Meetup/events/220742081/
Relevancy and Search Quality Analysis - By Mark David and Avi Rappoport
The Manifold Path to Search Quality
To achieve accurate search results, we must come to an understanding of the three pillars involved.
1. Understand your data
2. Understand your customers’ intent
3. Understand your search engine
The first path passes through Data Analysis and Text Processing.
The second passes through Query Processing, Log Analysis, and Result Presentation.
Everything learned from those explorations feeds into the final path of Relevancy Ranking.
Search quality is focused on end users finding what they want -- technical relevance is sometimes irrelevant! Working with the short head (very frequent queries) has the most return on investment for improving the search experience, tuning the results, for example, to emphasize recent documents or de-emphasize archive documents, near-duplicate detection, exposing diverse results in ambiguous situations, using synonyms, and guiding search via best bets and auto-suggest. Long-tail analysis can reveal user intent by detecting patterns, discovering related terms, and identifying the most fruitful results by aggregated behavior. all this feeds back into the regression testing, which provides reliable metrics to evaluate the changes.
By merging these insights, you can improve the quality of the search overall, in a scalable and maintainable fashion.
How To Implement Engineering Search Within Your Organization WebinarConcept Searching, Inc
What if you could not only search and discover, but also analyze, visualize and apply artificial intelligence to a normalized set of all structured and unstructured content within your organization, securely and in near real time?
Concept Searching, C/D/H, and Microsoft have partnered to bring you a global, cross-industry engineering search solution to deliver an unprecedented, unified view of all content, which your organization can rely on to grow and thrive.
Most organizations typically take twelve months or more to implement. We typically implement in two months – from scratch and hyper-agile.
This short How To webinar demonstrates our global cross-industry engineering search solution which delivers an unprecedented and unified view of all content within your organization.
• Fast – Find everything within your organization within three seconds
• Secure – Returns only results each person already has access to
• Normalized – based on a corporate governed nomenclature map
• Search – Concepts, compound terms, ranges, and normalized semantics
• Content – Both structured and unstructured
• Sources – From file shares, LOB systems, databases, websites
• Visualized – Full preview of all types of content for visual detection
• Organized – Refinement based on corporate taxonomy
• Analyzed – Applied artificial intelligence for predictive analytics
This Presentation presents the benefits of Data Science for those in retail broking practice. Employing Machine Learning techniques and text analytics, you not only get that competitive edge but also earn the customer's satisfaction and loyalty
This Presentation presents how Data science can bring manifold benefits to Retail Broking. Machine Learning & Text Analytics can impact your business in many positive ways- gives you that competitive edge and gains you customer satisfaction & loyalty
Cape Town - Bioschemas workshop before the Bioinformatics Education Summit.
Explains schema.org, Bioschemas, TeSS Case study, and the tools and implementation techniques adopters can use
Using Machine Teaching in Text Analysis: Case Study on Using Machine Teaching...Cambridge Semantics
At KDD 2020 Cambridge Semantics and Parabole.ai presented their joint paper 'Using Machine Teaching in Text Analysis: Case Study on Using Machine Teaching with Knowledge Graphs' by Thomas Cook, Rajib Saha, Aditya Narayanamoorthy and Sandip Bhaumik.
Fireside Chat with Bloor Research: State of the Graph Database Market 2020Cambridge Semantics
Sean Martin, CTO of Cambridge Semantics, Philip Howard, Research Director at Bloor Research and co-author of “Graph Database Market Update 2020”, and Steve Sarsfield, VP of Product at Cambridge Semantics, hold a fireside chat on the State of the Graph Database Market.
The Business Case for Semantic Web Ontology & Knowledge GraphCambridge Semantics
In this webinar Mark Wallace, Ontologist & Developer, Semantic Arts, and Thomas Cook, Director of Sales AnzoGraph DB, Cambridge Semantics, explore the benefits of building a Semantic Knowledge Graph with RDF*, wrapping up with an airline data demo that illustrates the value of schema, inference and reasoning in it.
In this webinar Thomas Cook, Sales Director, AnzoGraph DB, uses real-world flight data to discuss RDF and its newer property-graph-functionality iteration, RDF*, wrapping up with a pair of real-world demonstrations via Zeppelin notebooks.
In this webinar Thomas Cook, Sales Director, AnzoGraph DB, provides a history lesson on the origins of SPARQL, including its roots in the Semantic Web, and how linked open data is used to create Knowledge Graphs. Then, he dives into "What is RDF?", "What is a URI?" and "What is SPARQL?", wrapping up with a real-world demonstration via a Zeppelin notebook.
Healthcare and Life Sciences: Two Industries Separated by Common DataCambridge Semantics
Life Science and Healthcare industry leaders are finding success managing their disparate and unstructured data by implementing enterprise data fabrics. In this webinar you'll learn how leading organizations are using data fabrics to enable powerful and novel health sciences insights.
At Data-centric Architecture Forum 2020 Thomas Cook, our Sales Director of AnzoGraph DB, gave his presentation "Knowledge Graph for Machine Learning and Data Science". These are his slides.
When it comes to dealing with large, complex, and disparate data sets, traditional database technologies are unable to keep pace with the rich analytics necessary to power today’s data-driven applications. Graph analytics databases are becoming the underlying infrastructure for AI and machine learning. These databases allow users to ask complex questions across complex data, which is not always practical or even possible at scale using other approaches. They also enable faster insights against massive data sets when combined with pattern recognition, statistical analysis, and AI/ machine learning. And in the case of standards-based graph databases, they connect with popular visualization tools like Graphileon, allowing users to easily explore their data stores and quickly build compelling graph-based applications.
Pharma divisions, including translational research, medical affairs and patient safety are seeking to accelerate R&D with insights gained through analyzing results across multiple clinical trials. These efforts are hindered, however, by those results being spread across multiple disparate data sources. View these slides to learn more about how the Anzo platform provides a semantic layer to rapidly ingest, link, transform, and harmonize all your clinical data, then view the full webinar on demand.
Large Scale Graph Analytics with RDF and LPG Parallel ProcessingCambridge Semantics
Analytics that traverse large portions of large graphs have been problematic for both RDF and LPG graph engines. In this webinar Barry Zane, former co-founder of Netezza, Paraccel and SPARQL City and current VP of Engineering at Cambridge Semantics, discusses the native parallel-computing approach taken in AnzoGraph to yield interactive, scalable performance for RDF and LPG graphs.
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
7. What Solutions Benefit From Anzo?
• For aggregation of data from multiple, diverse data sources
• For integration of internal data with external data across the Web or
firewalls
• For solutions involving data sources, business rules, analytics and
actions that are not evident in advance
• For solutions that change often
• For analyzing diverse data sources with a diverse variety of access
control requirements with a need for full provenance and traceability
• For evolving solutions benefiting from ongoing involvement from
domain experts to update data models, data sources, and analytics as
needed
• For formal and informal day-to-day business activities that require
workflow, alerts, and automation
• For collecting & analyzing data that doesn’t currently have any system
of record (e.g. “shadow IT” systems)
17. Use Cases in Pharma
• PV & Safety Data Management - Automatic tagging of case reports with
customized curation workflow, text mining, and contextual search
• R&D Competitive Intelligence – Explore the competitive landscape for
Therapeutic Area, Indication, Target, Company, Compound, & Partners
• R&D Informatics– Understand and correlate your internal research and
how it may be related to any external developments or research
• Clinical Trial Site Selection and Optimization - Site selection, KOL search,
trial planning
• Scientific Affairs/Medical Science Liaisons - Track Key Opinion Leaders
(KOL) in literature and clinical trials & analyze feedback from medical
professionals and patients
• Information Landscape - Track and monitor data stewardship and usage
through the organization to drive more efficient usage.
• Commercial Analytics – Sales and Marketing, Rx Data, Text Analytics
18. Use Cases in Financial Services
• Compliance Policy & Procedure Management - Monitor structured and
unstructured data sources for relevant regulatory changes; have
collaborative workflows for policy & documentation development,
approval, and control; and establish targeted policy dissemination and
attestation workflows.
• Compliance Surveillance & Investigation– Legal and Compliance analysts
can create structures and views that provide analysis, rules, and alert
thresholds easily changed on-the-fly by investigators, who can then
comprehend and interact with the big data picture.
• Market and Customer Intelligence- Understand how clients and prospects
are thinking about your firm and competitors’ offerings
• Research - Automated analytics of news, chatter, IMs, secondary research
reports, emails, sentiment, etc. for research alerts, semantic search, and
relationship visualization, forming an integrated intelligence platform for
analysts, including Complex Event Processing.
• Information Landscape - Track and monitor data stewardship and usage
through the organization to drive more efficient usage.
• Commercial Analytics – Sales and Marketing, Tx Data, Text Analytics
Upper ontology vs lower ontology
Operational ontology
Main message: If your business faces challenges that show the characteristics on this slide, then you’d derive value from Anzo.
Additional details:
(for the first point): often these solutions combine “traditional” data sources (e.g. databases) with “non-traditional” data sources (e.g. spreadsheets or documents)
(for the second point): often these solutions involve some degree of back-and-forth collaboration between people inside a company and partners, suppliers, customers, etc.
(for the fourth point): Examples: the solution needs ad-hoc data, the solution needs rapidly evolving KPIs, the business is rapidly evolving due to mergers/acquisitions, etc.
At this point, can apply semantic services for reasoning, probabilistic reasoning, predictive analytics
Main message: If your business faces challenges that show the characteristics on this slide, then you’d derive value from Anzo.
Additional details:
(for the first point): often these solutions combine “traditional” data sources (e.g. databases) with “non-traditional” data sources (e.g. spreadsheets or documents)
(for the second point): often these solutions involve some degree of back-and-forth collaboration between people inside a company and partners, suppliers, customers, etc.
(for the fourth point): Examples: the solution needs ad-hoc data, the solution needs rapidly evolving KPIs, the business is rapidly evolving due to mergers/acquisitions, etc.
Main message: If your business faces challenges that show the characteristics on this slide, then you’d derive value from Anzo.
Additional details:
(for the first point): often these solutions combine “traditional” data sources (e.g. databases) with “non-traditional” data sources (e.g. spreadsheets or documents)
(for the second point): often these solutions involve some degree of back-and-forth collaboration between people inside a company and partners, suppliers, customers, etc.
(for the fourth point): Examples: the solution needs ad-hoc data, the solution needs rapidly evolving KPIs, the business is rapidly evolving due to mergers/acquisitions, etc.
Batch vs. incremental
- often not explicit that is different ontology
- especially when using same symbols
- seldom presented a full other ontology in practice
- updating bayesian priors from convincing stories
Suppose that instead of both annotators calling Corona a Product, one of them called it a Brand instead. Would we still be able to do this semantic correspondence linking and semantic overlay? Yes, with something called Rough Semantic Overlay.
Note that, because we chose to include a semantic correspondence linker in the pipeline, correspondences were found between Lexalytics and Calais regarding Company and regarding Product.
There was another correspondence picked up between Stocks found by our knowledgebase annotator and the company extractions as well, which is what will let us tie this all automatically to our structured data.
Semantic Overlay – Four Types: Direct Semantic Overlay; Meaning correspondence via shared class name; Rough Semantic Overlay; Meaning correspondence via compatible class names; Bridge Semantic Overlay
Bridging structured and unstructured derived classes; Contextual Correspondence; Leverages the contains-in-some-way relationship
Growing or evolving a semantic network by connecting, merging, or overlaying corresponding nodes of meaning from different sources
A given resource of class A can be treated the same for some purposes as some other resource of class B
A contextualization where the entities correspond in some way other than equivalence
Generic subsumption – a ‘contains’ or ‘part-of’ relationship; Type-specific relationships – e.g. third-party modifiers on sentiment
So given all this, we can start to ask some interesting questions. This path in the ontology here represents the question: What stocks from our structured dataset were found mentioned as acquirers in all our news and research reports, and what were the corresponding stocks that were the acquirees, also expressed in our original structured data. An analyst or compliance officer would be able to follow these paths when building a dashboard to ask such a question, of course without the need for any coding.
Another question they can ask is: What stocks/companies from my original structured dataset have a Corporate Interest in Products/brands from a different company? This is even though the CompanyProduct relationship came from Calais and the Corporate Interest relationship, a custom relationship, was based on Lexalytics entities, we are still able to follow the path, ask the question, build the dashboard, because of the semantic overlay.
Which leads us to notice that there was a trade, a buy, on STZ, Constellation Brands. This would not in and of itself be an issue, since only deal team members have knowledge of the deal.
Alignment of contexts
Often useful approach in data fusion and knowledge blending