Emcien overview v6 01282013


Published on

EMCIEN Scout social media, BIG Data analysis, pattern detection software

Published in: Social Media, Technology
  • Be the first to comment

  • Be the first to like this

No Downloads
Total Views
On Slideshare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide

Emcien overview v6 01282013

  1. 1. The Big Data Paradigm Shift: Insight Through Automation
  2. 2. In this white paper, you will learn about: • Big Data’s combinatorial explosion • Current and emerging technologies • Automation as the new way to leverage insight within Big Data • An algorithmic approach to the Big Data revolution 2 w w w. e m c i e n . c o m
  4. 4. ... a study of progress over a 15-year span on a benchmark production-planning task. Over that time, the speed of completing the calculations improved by a factor of 43 million. Of the total, a factor of roughly 1,000 was attributable to faster processor speeds. Yet a factor of 43,000 was due to improvements in the efficiency of software algorithms.” Martin Grotschel, a German scientist and mathematician White House Advisory Report DECEMBER 2010 4 w w w. e m c i e n . c o m
  5. 5. Data used to be scarce, and tiny bits of it were extremely valuable. Tiny bits of it are still extremely valuable, but now there is so much data that finding the valuable bits can be extremely difficult.” Executive Summary Big Data promises greater insight, competitive advantage, and the possibility of solving problems that have yet to be imagined. These insights will come from software applications that automate the analytics process. While infrastructure may be an important building block, this paper focuses on the algorithms that will deliver the insights organizations need. This paper does the following: • Proposes a paradigm shift away from analysts’ one-to-one relationship with data toward a relationship with algorithms. • Explains the combinatorial explosion that makes Big Data Analytics impossible for old data analytics tools and methodologies. • Examines current and emerging technologies. • Suggests that methods which accommodate Big Data users in the same way that smaller data sets are handled are missing the potential of Big Data. • Proposes that rather than search and crunch data, organizations need the ability to automate the process of analyzing, visualizing and ultimately leveraging the insight within their data. • Introduces an algorithmic approach that provides an efficient, sustainable, automated way to delve into Big Data, detect patterns, and discover insights hidden within that data. w w w. e m c i e n . c o m 5
  6. 6. The Big Data Paradigm Shift The Need for a Paradigm Shift Humankind has always possessed a love for data. We can do remarkable things with data and have built remarkable tools to collect, store, sift, sort, splice, dice, chart, report, predict, and visualize it. Data can change the way we perceive the world and how we interact with it. But the world is changing. Data used to be scarce, and tiny bits of it were extremely valuable. Tiny bits of it are still extremely valuable, but now there is so much data that finding the valuable bits can be extremely difficult. Big Data demands that organizations change the way they interact with data. In the past, analysts could stand by the data faucet and collect what was needed in a paper cup, but now data is the ocean in which they are floating. Paper cups are useless here. Why? Because the search for data is over. Data is everywhere. It creates noise. Now analysts are searching for the signal amidst the noise. They are looking for the important bits. And in this vast sea of information, that task is overwhelming. Current tools and methodologies are failing when it comes to finding the most critical information in a time-sensitive, cost-effective manner. The reason? Many of these emerging tools and technologies are trying to approach this challenge with new ways of doing the same old things. But a bigger cup is not what’s needed. That won’t solve it. What is needed is a completely new approach to Big Data Analytics. In this new approach, the only way to find the signal is to automate the process of data-to-insight conversion. And automation requires algorithms— fast, sophisticated, highly optimized algorithms. The volume of Big Data demands a change in the human relationship with data, from human to machine. The algorithms have to do the work, not the humans. In this brave new world, the machines and algorithms are the protagonists. The role of the analysts will be to select the best algorithms and approve the quality of results based on speed, quality and economics. 6 The Promise of Data and the Search for Insight Why is the world obsessed with data? Because the promise of data is insight. In the last few years, organizations have become exceptionally good at collecting data, and as the cost of storage has dropped, companies are now drowning in that “Big Data.” However, the business world has hit a wall where the amount of data available far exceeds the human capacity to process it. The amount of data also exceeds the capabilities of existing analytics and intelligence tools, which have served as mere data-shovels or pick axes in the search for the gold that is insight. Acquiring insight inevitably involves querying a database, or, most likely, several databases. Many analysts are mashing up data across data silos in an attempt to discover the connections between data points. For example, marketers are aggregating customer demographics, purchase data and social media data, while purchasers are aggregating supplier data with procurement and pricing data. This process produces a variety of data sets of different types and qualities. A simple query, for example, might ask for specific values within a subset of columns. So the real question becomes, “How many queries will it take to answer even one of these questions?” Consider how many queries one might make into even a small set of data, such as a table containing just ten columns: • If each column has 2 possible values, there are 59,048 possible queries. • If each column has 3 possible values, there are 1,048,575 possible queries. To think of it another way, a database with 100 columns and 6 choices per column yields more possible queries than there are atoms in the universe. w w w. e m c i e n . c o m
  7. 7. The Limitations of Search and Query-based Approaches Data becomes unwieldy due to the number of rows or the number of columns, or both. Data mash-ups, described previously, create a lot of columns. High volume transactional systems have lots of rows or records. However, having millions of records is not the problem. The depth of the data—the number of rows—merely impacts processing time in a linear fashion and can be reduced with fast or parallel computing. Thus, executing a query is simple enough. The problem, however, is the width of the data because the query explodes exponentially based on the number of columns. Number of Queries As a result, the real task of extracting insight from data is formulating the right queries. And manually laboring through thousands of queries to find the ones that deliver insight is not an efficient way to derive value from data. Therefore, when it comes to Big Data, the big challenge is knowing the right query. 30 Billion Exponential Explosion of Queries 10 Billion 1 Billion 1 Million 0 2 4 6 8 10 12 14 Number of Variables per Column w w w. e m c i e n . c o m 7
  8. 8. Landscape of Existing Methods and Tools Over 85% of all data is unstructured. 1 However, existing methods and tools are designed to analyze structured data. A high level categorization of analytics tools is critical to understanding the state of Big Data Analytics. Statistical Tool Kits The purpose of statistical analysis is to make inferences from samples of data, especially when data is scarce. In the era of Big Data, scarcity is not the problem. Traditional statistical methods have severe limitations in the realm of Big Data for the following reasons: • Statistical methods break down as dimensionality increases. • In unstructured data, dimensions are not well defined. • Attempts to define dimension for unstructured data result in millions of dimensions. Data Mining Data mining is a catchall phrase for a very broad category of methods. Essentially, it is a method for sifting through very large amounts of data in attempt to find useful information. It implies “digging through tons of data” to uncover patterns and relationships contained within the business activity and history. Data mining involves manually slicing and dicing the data until a pattern becomes obvious or by using software that analyzes the data automatically. The first limitation of data mining is that the data has to be put in a structured format first, such as a database. The second limitation is that most forms of data mining require that the analyst knows what to look for. For example, in classification and clustering analysis, the analyst is trying to find instances of known categories, such as people who have a high probability of defaulting on their mortgages. In anomaly detection, the analyst is looking for instances that do not match the known normal patterns or known suspicious patterns, such as people who pay cash for one-way plane tickets. 8 The overwhelming shortcoming of all these methods is that they are query-based and labor intensive.” Data Visualization Data visualization is the study of the visual representation of data, meaning information that has been abstracted in some schematic form, including attributes or variables for the units of information.2 Humans are better equipped to consume visual data than text. As we know, a picture is worth a thousand words. While visualization tools are interesting, they rely on human evaluation to extract insight and knowledge. The more severe limitation of visualization is that the visuals can only focus on two or three dimensions at the most before the amount of information is overwhelming. The most common limitation of visualization is that while it is a good test for small samples, it is not a sustainable method to gain insight into large volumes of higher dimensionality data. Consider a scenario in which there aren’t enough pixels on the screen to represent each item. An analyst can easily inspect a friendship network with 10-100 people, but not on billion. Business Intelligence & Analytics Business Intelligence (BI) is a catchall phrase for ad hoc reports created in a database. These are typically pre-canned reports based on metrics that users are comfortable reporting. “Analytics” includes any computation performed for reporting. Hence, BI tools are now called analytics. BI was created as a way to extract data from the database. While it continues to serve that purpose, it is time and labor intensive and is not intended to surface insights. w w w. e m c i e n . c o m
  9. 9. Limitations of Existing Tools The overwhelming shortcoming of all these methods is that they are query-based and labor-intensive. Big Data offers an infinite number of queries, causing all these methods to rely on analysts to produce questions. Any method that puts the burden on the user is a game-stopper. Although search remains the go-to information access interface, reliance on search needs to end. Search is not enough. A new type of information-processing focus is needed. The major shortcomings of the existing tools are as follows: • Because search helps you discover insights you already know about, it doesn’t help you discover things about which you’re completely unaware. • Query-based tools are time-consuming because search-based approaches require a virtually infinite number of queries. • Statistical methods are largely limited to numerical data; over 85% of data is unstructured. w w w. e m c i e n . c o m 9
  10. 10. Emergence of Big Data Tools Because Big Data includes data sets with sizes beyond the ability of commonly-used software tools to capture, curate, manage, and process within a tolerable elapsed time, new technologies are emerging to address the challenges brought on by these large quantities of data. These technologies can be categorized into two groups: Hadoop-based solutions and In-Memory based solutions. Hadoop and Hadoop-based Tools While Hadoop is not an analytics tool per se, it is often confused as being one. Apache Hadoop is an open-source software framework that supports data-intensive distributed applications. It supports the running of applications on large clusters of commodity hardware. Hadoop is used to break big tasks into smaller ones so that they can be run in parallel to gain speed and efficiency. This is great for a query on a large volume data set. The data set can be cut into smaller pieces, and the same query can be run on each smaller set. Hadoop aims to lower costs by storing data in chunks across many inexpensive servers and storage systems. The software can help speed up certain types of simple calculations by sending many queries to multiple machines at the same time. The technology has spawned a set of new start-ups, such as Hortonworks Inc. and Cloudera Inc., which help companies implement it. Hadoop helps companies store large amounts of data but doesn’t provide critical insights based on the naturally occurring connections within the data. The impulse to store lots of data because it is cheap to do so can lead to storing too much data, which can make answering simple questions more difficult. 10 IT professionals and analysts are asking the following questions: • Where is the insight? What is the data telling us? • How can I prove the return of investment? As a result, in-memory databases are gaining attention in an attempt to come closer to the goal of real-time business processing. 3,4 In-Memory-based Appliance Some of these approaches have been around for a long time in areas such as telecommunications or fields related to embedded databases. An example is SAP’s HANA (High Performance Analytical Appliance). This in-memory paradigm is now touted as the future database paradigm for Big Data. The primary limitation of in-memory is cost and size. There is a significant limit to the amount of data that can be held in memory. If you need to perform Big Data–style analysis and you want to see the bigger picture, in-memory is not enough. The cost is prohibitive and it is not sustainable. w w w. e m c i e n . c o m
  11. 11. The Need for a New Big Data Analytics Approach While these emerging technologies are attempting to address the challenge of Big Data, at the end of the day, they are heavy-handed and time-consuming because they lack automated intelligence for gaining insight. It’s time for an entirely new approach. This new approach demands a paradigm shift that focuses on the following: • A fundamental change in the role played by analysts from data-miners to insight-evaluators. • Fast and efficient algorithms that automatically convert data to insight for evaluation. • Continual improvement of these algorithms to keep up with the speed of data and critical need for timely insights. Old Paradigm: New Paradigm: Data Analyst Digs for Insights by Manually Querying a Database Algorithms Automatically Surface Insights to Evaluate Analysis Takes From Months to Years Automatic Insights in Seconds to Minutes Specialized Skills in Math and Computer Science Required Anyone (No Specialized Skills Required) Operational and Business Intelligence Immediate Insight and Perspective w w w. e m c i e n . c o m 11
  12. 12. Emergence of Algorithms as a New Class of Big Data Software Tools The size and speed of Big Data demands true automation, in which work is offloaded from human to machine. This automation happens with algorithms, which are designed for calculation, data processing, and automated reasoning. Algorithms are designed for tasks that are beyond human comprehension and require the speed of machines. This is the realm of Big Data. One of the most dramatic and game-changing examples of an algorithm was designed by Alan Turing to automatically decode German Navy messages at Bletchley Park during WWII. In this instance, the urgency was critical and demanded an automated approach to convert the data to intelligence. There were 158 million million million (158,000,000,000,000,000,000) possible ways that a message could be coded by the German Enigma machine. The decoder algorithm, Shark, worked and the results changed the course of the war. The Allies won because they had a competitive advantage. Bringing it to the present, the use of Big Data in the 2012 United States presidential election changed the face of political campaigns forever. Emcien’s approach to Big Data is to automate the process of data-to-insight in a timely and cost effective manner through sophisticated algorithms. The algorithms leverage advanced mathematics to solve complex problems of an unimaginable size, thereby pushing the frontier of innovation and competition. The following section details Emcien’s algorithmic approach to Big Data Analytics. 12 w w w. e m c i e n . c o m
  13. 13. Emcien’s Algorithmic Approach to Big Data Rather than search and crunch data, organizations need the ability to analyze, visualize and ultimately leverage the patterns and connections within their data. Emcien’s innovation is a suite of automatic pattern detection algorithms. These algorithms utilize a graph data model that captures the interconnectedness of the data elements and creates a very elegant representation of high volume data with unknown structure. These fast, sophisticated algorithms automatically detect patterns and self-organize what they find thereby providing immediate insight and perspective. Here is an outline of how the algorithm works: 1. Assesses the data in order to identify and measure connections between data points. 2. Converts the original high density/low value (structured, semi-structured or unstructured) data into a low density/high value graph. Structured If the data is structured: • Each row in the data table is considered an event. • Each cell in the row is converted to a node in the graph. • Cells that co-occur in a row (event) are connected by an arc. Unstructured If the data is unstructured: • Every word or data element is converted to a node. • Two words or data elements that occur simultaneously in an event are connected by an arc. • An event may be defined as a single document, message, email exchange, etc. 3. Builds the graph on non-Euclidean distances. This is important, as most of the data is unstructured and non-numeric. The distances and strengths are computed in non-Euclidean space. (For example – you may be closer to your family than your friends – but this is not Euclidean space!). 4. Computes millions of data points across the graphs to enable patterns to emerge. 5. Distills the noise, to allow the signal to emerge. This is made possible because the noise has patterns and the algorithms are designed to detect these patterns. 6. Enables the key topographical elements from the graph to emerge. The algorithm will then rank and focus these elements. 7. Categorizes these elements based on the application and output the insight. w w w. e m c i e n . c o m 13
  14. 14. Understanding the make-up of Graphs based on Connections The graph data model is very flexible and displays a distinct topography based on the density of connections. A cross sectional view of the graph data model will typically expose the following layers Layer Connectedness Description 1 Very Noisy Connections Typically the most highly prevalent data, this layer is composed of high volume interactions that may be mundane and blatantly obvious. 2 Highly Connected Nodes Lying just below the noise, this second layer is composed of the first signal that is interesting. This layer exhibits distinct patterns based on crowd behavior. 3 Weaker Connections The third layer is a weaker signal and displays the non-obvious connections. These relate to events that are less frequent and may be connected in non-obvious ways. 4 The Faint Signal Composed of very weak connections and interactions, this last layer is of interest for security and surveillance. In many cases, this layer only emerges when the data is very rich in entities, causing connections to emerge in very non-obvious ways. Advantages of Emcien’s Graph Data Model The graph data model exhibits a topography that signifies relationships and connectedness in a way that is not possible through any other method. Emcien’s algorithms have been designed to surface these patterns. Listed below are a few key attributes that help describe the characteristics of the algorithms. Attribute Advantage Software A critical distinction of Emcien’s graph data model is that it is software. Emcien’s software provides the computational engine with a data representation that lends itself to high-speed computing. As a result, the software runs on typical commodity computing environments. Algorithmic Layer Compact Representation of Data Data is big because the number of events can grow exponentially as the various entities are continually interacting. Noise Elimination 14 Although some products on the market model the graph into the database layer or hardware layer, they do not have an algorithmic layer, and, therefore, require the user to query the systems based on the old data-inquiring paradigm. Algorithms automate the data analysis process that is an absolute requirement for efficient Big Data analytics. The graph data model can be thought of in terms of layers, based on the connectedness of the data elements. The highly connected and noisy nodes are at the top layer, and the weak connections lay buried deep in the graphs. The noisy connections can be overwhelming and tend to render graph models as burdensome. Emcien utilizes a suite of patented algorithms to automatically distance the noise and detect critical patterns that relate to highly significant and relevant information. The graph representation is ideal for Big Data because it creates a very compact representation of the data. This is because the number of entities grows more slowly and reaches a natural steady state. In the graph data model these interactions translate to connection weights, allowing the graph model to encapsulate very big data in smaller structures. w w w. e m c i e n . c o m
  15. 15. Not Just Theory: Solving Real-World Problems The representation and visualization of complex networks as graphs helps surface critical, time-sensitive intelligence. One of the most important tasks in graph analysis is to identify closely connected network components comprising nodes that share similar properties. Detecting communities is of significant value in retail, healthcare, banking and intelligence work - verticals where loosely federated communities deliver insight and intelligence into the profile of a customer base or any other group being analyzed. How Can This Model Be Applied? Emcien’s Pattern Detection Engine: • Intelligence: Surfaces critical correlations between people that merit serious attention, determines key individuals in targeted social networks, and geo-locates persons of interest and their networks around the world – from gangs to terrorists. • Network Security: Auto-detects intrusion patterns and surfaces suspicious activity by providing immediate insight into highly linked variables. It then automatically identifies anomalies without the user having to query the data. For example, Emcien analyzes millions to billions of transactions to identify patterns in source and destinationIP addresses, ports, days, times and activity – to show you what you should be paying attention to. Emcien eliminates over 95% of the noise and identifies patterns that are “surprising” or that deviate from the norm. • Fraud Detection: Surfaces patterns in money laundering and fraud by identifying groups of customers, locations, or transaction types that occur together in banking transactions, • Customer Analytics: Surfaces insights on customer buying patterns, locations, demographics, loyalty, savings, lifestyle and insurance. • Healthcare Analytics: Analyzes massive volumes of clinical data on medications, allergies, medical claims, pharmacy therapies, lab results, medical records, clinician notes and more in order to surface patterns. • Performance and Operations Analytics: Analyzes raw information about performance and operations of every element of an organization which can be interpreted to increase profitability or improve customer service. In short, Emcien tackles one of the biggest challenges with Big Data, namely “What are the right questions to ask?” Emcien’s pattern-detection engine quickly discovers the value within massive data sets by making connections between disparate, seemingly unrelated bits of information and by finding the highest-ranked of these connections to focus on – which reveals time-sensitive, mission-critical insights. w w w. e m c i e n . c o m 15
  16. 16. Conclusion The Big Data Analytics revolution is underway. This revolution is a historic and game-changing expansion of the role that information plays in business, government and consumer realms. To harness the power of this data revolution, a paradigm shift is required. Organizations must be able to do more than query their Big Data stores; search is no longer enough. Up until now in the history of data analysis, the objective of queries was to find the signal in the noise. And it worked because we had clear-cut business questions and the size of the data was smaller, the data set was more complete, and we usually knew what we were looking for. We were playing in the realm of known knowns and known unknowns. In the new world of Big Data, it is now more important to know what to ignore. Because unless you know what to ignore, you’ll never get a chance to pay attention to what’s really important. Using algorithms to first ignore the noise and then find the insights is the way of the new world. Extracting insight from Big Data requires analytics methods that are fundamentally different from traditional querying, mining, and statistical analysis on small samples. Big Data is often noisy, dynamic, heterogeneous, unstructured, inter-related and untrustworthy. The combinatorial explosion requires new methods for finding insight in Big Data. The need for data sophistication is due to economics and time-criticality. As stated, manually laboring through thousands of queries to find the ones that deliver insight is not an efficient way to derive value from data. Emcien’s technology provides a “Command Center” for Big Data, automatically interpreting the data, discovering patterns, identifying complex and significant relationships, and surfacing the most relevant questions that lead to the insights analysts need to know. About Emcien Corp. Emcien’s automatic pattern-detection engine converts data to actionable insight that organizations can use immediately. Emcien breaks through time, cost and scale barriers that limit the ability to operationalize the value of data for mission-critical applications. Our patented algorithms recognize what’s important, defocus what’s not, evaluate all possible combinations and deliver the optimal results automatically. Emcien’s engine, fueled by several highly competitive NSF grants and years of research at Georgia Tech and MIT, is delivering unprecedented value to organizations across sectors that depend on immediate insight for success—banking, healthcare, insurance, retail, Intelligence and others. Visit emcien.com to learn more. 16 w w w. e m c i e n . c o m
  17. 17. Sources 1. Christopher C. Shilakes and Julie Tylman, “Enterprise Information Portals”, Merrill Lynch, 16 November, 1998. 2. Michael Friendly, “Milestones in the history of thematic cartography, statistical graphics, and data visualization,” 2008. 3. J. Vascellaro, “Hadoop Has Promise but Also Problems,” The Wall Street Journal, 23 February 2012. 4. R. Srinivasan, “Enterprise Hadoop: Five Issues With Hadoop That Need Addressing,” blog, 28 May 2012.