The document discusses how data science can help build better products. It explains that products are initially built to quickly test ideas through lightweight and imperfect means. Data science helps understand customer value and enables continuous learning through a process of analyzing data, making discoveries, and pivoting the product based on what is learned. This contrasts with the traditional approach where functionality is locked in place. The document advocates for an adaptive software environment that allows for rapid changes based on new insights. It provides tips for building successful data products through iterative improvements informed by data.
What Makes Healthcare Data Science so Hard & Interesting - Data Science Pop-u...Domino Data Lab
This is an introduction to the exciting world of healthcare data science though three real-world efforts we've delivered on in the past 3 years, and some of the open problems behind them. First, automated clinical coding will demonstrate the challenges in natural language understanding, variety of jargons and measurement stability. Second, patient risk prediction will illuminate the need for localized models, non-stationary models and explain-ability. Third, the master patient index covers some of the data quality, privacy and compliance constraints and the design choices they imply. We hope to convince that tackling these open problems is as intellectually exciting as it is socially important. Presented by David Talby, SVP Engineering at
Atigeo.
What is Your Data Worth? - Data Science Pop-up SeattleDomino Data Lab
With all the attention that big data’s place in the enterprise has been getting in the press as well as the coverage of a number of high-value data purchases and exchanges, people are beginning to wonder, “How much is my data worth?” Despite the volume of money being invested in data and data technology, methods for answering this question are severely lacking. In this talk, we will discuss relevant concepts to data’s valuation from related fields, the characteristics of data that make it unique from other economic goods, and practical considerations for how to begin thinking about the general value of data for the enterprise. Presented by Chloe Mawer, Data Scientist at SVDS.
First, we will explore the power of a compounding insight machine (as opposed to an ad hoc insight machine):
-Human time is focused on improving logic, rather than executing outcomes
-Less dependent on human biases or frailty
-Robust to and tested by a huge collection of scenarios
Second, we will explore the anatomy of such a machine:
-The roles you need to cast on your team and who to fill them with
-The key processes required for generating and capturing insight and, more importantly, for building upon those insights
-The technology required to enable this approach
Making Big Data Projects Successful - Data Science Pop-up SeattleDomino Data Lab
There are exciting and boring elements in any project, both of which must be addressed. This session will cover how to overcome the difficult but necessary problems that are essential to success. Presented by Aaron Cordova
CTO, co-founder at Koverse Inc.
Correctness in Data Science - Data Science Pop-up SeattleDomino Data Lab
Presented by: Benjamin S. Skrainka is a Principal Data Scientist and Lead Instructor at Galvanize, Inc. For several decades, he has built practical solutions to relevant problems using the best statistical and engineering tools. His expertise spans several problem domains, including sequencing DNA, estimating demand for differentiated products, measuring advertising efficacy, and forecasting for capacity planning. Ben earned an AB in Physics from Princeton University and a PhD in Economics from University College London.
Running Effective Controlled Experiments (aka A/B/n Tests) - Data Science Pop...Domino Data Lab
Microsoft's Analysis and Experimentation team has enabled many different feature teams to run controlled experiments in order to make feature ship/no-ship decisions. In this talk, I will review several case studies of experiments run by teams at Microsoft. I will highlight both the value that running controlled experiments has provided these teams as well as the challenges encountered in order to get trustworthy results from the controlled experiments. Presented by Brian Frasca, Partner Data Scientist Manager at
Microsoft.
What Makes Healthcare Data Science so Hard & Interesting - Data Science Pop-u...Domino Data Lab
This is an introduction to the exciting world of healthcare data science though three real-world efforts we've delivered on in the past 3 years, and some of the open problems behind them. First, automated clinical coding will demonstrate the challenges in natural language understanding, variety of jargons and measurement stability. Second, patient risk prediction will illuminate the need for localized models, non-stationary models and explain-ability. Third, the master patient index covers some of the data quality, privacy and compliance constraints and the design choices they imply. We hope to convince that tackling these open problems is as intellectually exciting as it is socially important. Presented by David Talby, SVP Engineering at
Atigeo.
What is Your Data Worth? - Data Science Pop-up SeattleDomino Data Lab
With all the attention that big data’s place in the enterprise has been getting in the press as well as the coverage of a number of high-value data purchases and exchanges, people are beginning to wonder, “How much is my data worth?” Despite the volume of money being invested in data and data technology, methods for answering this question are severely lacking. In this talk, we will discuss relevant concepts to data’s valuation from related fields, the characteristics of data that make it unique from other economic goods, and practical considerations for how to begin thinking about the general value of data for the enterprise. Presented by Chloe Mawer, Data Scientist at SVDS.
First, we will explore the power of a compounding insight machine (as opposed to an ad hoc insight machine):
-Human time is focused on improving logic, rather than executing outcomes
-Less dependent on human biases or frailty
-Robust to and tested by a huge collection of scenarios
Second, we will explore the anatomy of such a machine:
-The roles you need to cast on your team and who to fill them with
-The key processes required for generating and capturing insight and, more importantly, for building upon those insights
-The technology required to enable this approach
Making Big Data Projects Successful - Data Science Pop-up SeattleDomino Data Lab
There are exciting and boring elements in any project, both of which must be addressed. This session will cover how to overcome the difficult but necessary problems that are essential to success. Presented by Aaron Cordova
CTO, co-founder at Koverse Inc.
Correctness in Data Science - Data Science Pop-up SeattleDomino Data Lab
Presented by: Benjamin S. Skrainka is a Principal Data Scientist and Lead Instructor at Galvanize, Inc. For several decades, he has built practical solutions to relevant problems using the best statistical and engineering tools. His expertise spans several problem domains, including sequencing DNA, estimating demand for differentiated products, measuring advertising efficacy, and forecasting for capacity planning. Ben earned an AB in Physics from Princeton University and a PhD in Economics from University College London.
Running Effective Controlled Experiments (aka A/B/n Tests) - Data Science Pop...Domino Data Lab
Microsoft's Analysis and Experimentation team has enabled many different feature teams to run controlled experiments in order to make feature ship/no-ship decisions. In this talk, I will review several case studies of experiments run by teams at Microsoft. I will highlight both the value that running controlled experiments has provided these teams as well as the challenges encountered in order to get trustworthy results from the controlled experiments. Presented by Brian Frasca, Partner Data Scientist Manager at
Microsoft.
Keys to understanding when you are looking for a Data Scientist vs. Engineer,...Domino Data Lab
Knowing how to hire in this market is tough, (and) understanding what you are really looking for is key. This Lightning talk will cover some of the challenges in our current market, (as well as) tips to make the hiring process easier. Presented by Mary Kypreos
Recruiting Manager for the Open Source & Big Data Team
Greythorn.
Hiring for Data Scientists - Data Science Pop-up SeattleDomino Data Lab
Tales from the other side.
What you might be missing if you don't know what you are looking for. Presented by Amanda Casari, Senior Data Scientist at Concur.
What you till learn:
GOALS - What is the bar for data science teams
PITFALLS - What are common data science struggles
DIAGNOSES - Why so many of our efforts fail to deliver value
RECOMMENDATIONS - How to address these struggles with best practices
Presented by Mac Steele
Director of Product at Domino Data Lab
Moving Data Science from an Event to A Program: Considerations in Creating Su...Domino Data Lab
The exponential growth of Big Data and Analytics has outpaced the ability of organizations to govern their data appropriately. The ability to reuse the work done by data scientists work is becoming an economic necessity. The mix of data sources is changing from tradition transactional and ERP systems to include a mix of structured, semi-structured and unstructured data. Data Governance needs to adapt to these changes. This session discusses these data changes and proposed how to adapt current data governance processes. These include, how the concept of a stakeholder has changed and the need for expansion of communications and content management. We look at need to consolidate data from disparate systems and how it governed. Lastly we will investigate how context is emerging as an important factor in governance and how it can be leveraged to provide for accurate, reliable data reuse.
AI can give your organization the competitive advantage it needs, but the alarming truth is that only 1 in 10 data science projects ever make it into production. To be successful, organizations must not only correctly design and implement data science, but also raise the data, numerical, and technology literacy across the business.
Attend this webinar to learn what common pitfalls you need to avoid to keep your data science projects from failing. Then Data Scientist Gaby Lio will engage with the audience about project dos and don’ts and leave you with a checklist to ensure your projects success.
As a manager, what do you need to know in order for the data-science project you are leading to be successful?
This presentation looks into a data-science project lifecycle, points out common failures and gives some hints on how to avoid common pitfalls. Examples included.
The target audience is managerial - half technical.
Keynote - An overview on Big Data & Data Science - Dr Gregory Piatetsky-ShapiroData ScienceTech Institute
Data Science Tech Institute - Big Data and Data Science Conference around Dr Gregory Piatetsky-Shapiro.
Keynote - An overview on Big Data & Data Science Dr Gregory Piatetsky-Shapiro - KDnuggets.com Founder & Editor.
Paris May 23rd & Nice May 26th 2016 @ Data ScienceTech Institute (https://www.datasciencetech.institute/)
Dr Bonnie Cheuk IDC Future of Work Keynote: Workforce Transformation Human Ma...Bonnie Cheuk
Dr Bonnie Cheuk, AstraZeneca Digital Transformation & Global Capability Leader (Learning Culture and Learning Agility), delivered a keynote at IDC Future of Work Conference on 3 Mar 2020. She provoked the audience to go beyond the hype, and think deeper on how human and AI and data-driven Machine collaborate together.
These 3 questions were discussed:
1. How should human and machine collaborate? What skills are required?
2. Will machines replace (most) jobs?
3. Will there be new jobs to enable human-machine collaboration?
Drawing on Dervin's Sense-Making Methodology, Bonnie reminded us that human beings are not robotic machines. Human beings have feelings, experience, we are both scientists and artists, we are analytics and we are emotional.
Bonnie asked the audience how would you like to build a high performance team? Who do you want to put in the team? Do you want everyone to have the same strength, same skills? Or would you pick a team making up of players who can complement one another, and can bring out the best of one another. So in order to propose how human and machine should collaborate in the future of work, it is useful to first ask: what is the strength of human beings? What is the strength of the machine? We need to understanding how AI-driven machines learn vs how human beings learned, and play to one another's strength. And what is the strength of human? It is being human. Let the machine handle the deductive reasoning, the data-driven predictions, repetitive tasks. Let the humans do what we do well, adapting, navigating the unknown, use our human skills, promote collective sense making to make judgement, decisions. And free up the time to allow us to learn, create and innovate.
Bonnie highlighted that there are many unknowns as to how AI will be further developed, and there are ethical issues and risks that have to be addressed, and there are no precedents to follow. Collective human sense making is critical to bring out multiple perspectives from different stakeholders, to co-create AI-driven machines that human beings can trust, and to collectively address tricky ethical issues early on. Dervin’s Sense-Making Metaphor is introduced to facilitate two-way dialogue, to address power issues, and to explore common and divergent views to build common understanding of potential challenges, and co-create solutions to address them.
Idiots guide to setting up a data science teamAshish Bansal
Some nuggets of how I started the data science practice at Gale Partners on a budget. Presented at the Toronto Hadoop Users Group (THUG) in April, 2015.
With the launch of Digital India Program, the Indian Govt. aims to transform India into digitally empowered economy. The govt. seeks to integrate technology with everyday life of the citizen. This program intends to provide customized content through affordable internet access in local language across the country.
Keys to understanding when you are looking for a Data Scientist vs. Engineer,...Domino Data Lab
Knowing how to hire in this market is tough, (and) understanding what you are really looking for is key. This Lightning talk will cover some of the challenges in our current market, (as well as) tips to make the hiring process easier. Presented by Mary Kypreos
Recruiting Manager for the Open Source & Big Data Team
Greythorn.
Hiring for Data Scientists - Data Science Pop-up SeattleDomino Data Lab
Tales from the other side.
What you might be missing if you don't know what you are looking for. Presented by Amanda Casari, Senior Data Scientist at Concur.
What you till learn:
GOALS - What is the bar for data science teams
PITFALLS - What are common data science struggles
DIAGNOSES - Why so many of our efforts fail to deliver value
RECOMMENDATIONS - How to address these struggles with best practices
Presented by Mac Steele
Director of Product at Domino Data Lab
Moving Data Science from an Event to A Program: Considerations in Creating Su...Domino Data Lab
The exponential growth of Big Data and Analytics has outpaced the ability of organizations to govern their data appropriately. The ability to reuse the work done by data scientists work is becoming an economic necessity. The mix of data sources is changing from tradition transactional and ERP systems to include a mix of structured, semi-structured and unstructured data. Data Governance needs to adapt to these changes. This session discusses these data changes and proposed how to adapt current data governance processes. These include, how the concept of a stakeholder has changed and the need for expansion of communications and content management. We look at need to consolidate data from disparate systems and how it governed. Lastly we will investigate how context is emerging as an important factor in governance and how it can be leveraged to provide for accurate, reliable data reuse.
AI can give your organization the competitive advantage it needs, but the alarming truth is that only 1 in 10 data science projects ever make it into production. To be successful, organizations must not only correctly design and implement data science, but also raise the data, numerical, and technology literacy across the business.
Attend this webinar to learn what common pitfalls you need to avoid to keep your data science projects from failing. Then Data Scientist Gaby Lio will engage with the audience about project dos and don’ts and leave you with a checklist to ensure your projects success.
As a manager, what do you need to know in order for the data-science project you are leading to be successful?
This presentation looks into a data-science project lifecycle, points out common failures and gives some hints on how to avoid common pitfalls. Examples included.
The target audience is managerial - half technical.
Keynote - An overview on Big Data & Data Science - Dr Gregory Piatetsky-ShapiroData ScienceTech Institute
Data Science Tech Institute - Big Data and Data Science Conference around Dr Gregory Piatetsky-Shapiro.
Keynote - An overview on Big Data & Data Science Dr Gregory Piatetsky-Shapiro - KDnuggets.com Founder & Editor.
Paris May 23rd & Nice May 26th 2016 @ Data ScienceTech Institute (https://www.datasciencetech.institute/)
Dr Bonnie Cheuk IDC Future of Work Keynote: Workforce Transformation Human Ma...Bonnie Cheuk
Dr Bonnie Cheuk, AstraZeneca Digital Transformation & Global Capability Leader (Learning Culture and Learning Agility), delivered a keynote at IDC Future of Work Conference on 3 Mar 2020. She provoked the audience to go beyond the hype, and think deeper on how human and AI and data-driven Machine collaborate together.
These 3 questions were discussed:
1. How should human and machine collaborate? What skills are required?
2. Will machines replace (most) jobs?
3. Will there be new jobs to enable human-machine collaboration?
Drawing on Dervin's Sense-Making Methodology, Bonnie reminded us that human beings are not robotic machines. Human beings have feelings, experience, we are both scientists and artists, we are analytics and we are emotional.
Bonnie asked the audience how would you like to build a high performance team? Who do you want to put in the team? Do you want everyone to have the same strength, same skills? Or would you pick a team making up of players who can complement one another, and can bring out the best of one another. So in order to propose how human and machine should collaborate in the future of work, it is useful to first ask: what is the strength of human beings? What is the strength of the machine? We need to understanding how AI-driven machines learn vs how human beings learned, and play to one another's strength. And what is the strength of human? It is being human. Let the machine handle the deductive reasoning, the data-driven predictions, repetitive tasks. Let the humans do what we do well, adapting, navigating the unknown, use our human skills, promote collective sense making to make judgement, decisions. And free up the time to allow us to learn, create and innovate.
Bonnie highlighted that there are many unknowns as to how AI will be further developed, and there are ethical issues and risks that have to be addressed, and there are no precedents to follow. Collective human sense making is critical to bring out multiple perspectives from different stakeholders, to co-create AI-driven machines that human beings can trust, and to collectively address tricky ethical issues early on. Dervin’s Sense-Making Metaphor is introduced to facilitate two-way dialogue, to address power issues, and to explore common and divergent views to build common understanding of potential challenges, and co-create solutions to address them.
Idiots guide to setting up a data science teamAshish Bansal
Some nuggets of how I started the data science practice at Gale Partners on a budget. Presented at the Toronto Hadoop Users Group (THUG) in April, 2015.
With the launch of Digital India Program, the Indian Govt. aims to transform India into digitally empowered economy. The govt. seeks to integrate technology with everyday life of the citizen. This program intends to provide customized content through affordable internet access in local language across the country.
Myths and Mathemagical Superpowers of Data ScientistsDavid Pittman
Some people think data scientists are mythical beings, like unicorns, or they are some sort of nouveau fad that will quickly fade. Not true, says IBM big data evangelist James Kobielus. In this engaging presentation, with artwork created by Angela Tuminello, Kobielus debunks 10 myths about data scientists and their role in analytics and big data. You might also want to read the full blog by Kobielus that spawned this presentation: "Data Scientists: Myths and Mathemagical Superpowers" - http://ibm.co/PqF7Jn
For more information, visit http://www.ibmbigdatahub.com
How to Become a Data Scientist
SF Data Science Meetup, June 30, 2014
Video of this talk is available here: https://www.youtube.com/watch?v=c52IOlnPw08
More information at: http://www.zipfianacademy.com
Zipfian Academy @ Crowdflower
Tutorial on Deep learning and ApplicationsNhatHai Phan
In this presentation, I would like to review basis techniques, models, and applications in deep learning. Hope you find the slides are interesting. Further information about my research can be found at "https://sites.google.com/site/ihaiphan/."
NhatHai Phan
CIS Department,
University of Oregon, Eugene, OR
This talk is about how we applied deep learning techinques to achieve state-of-the-art results in various NLP tasks like sentiment analysis and aspect identification, and how we deployed these models at Flipkart
Data By The People, For The People
Daniel Tunkelang
Director, Data Science at LinkedIn
Invited Talk at the 21st ACM International Conference on Information and Knowledge Management (CIKM 2012)
LinkedIn has a unique data collection: the 175M+ members who use LinkedIn are also the content those same members access using our information retrieval products. LinkedIn members performed over 4 billion professionally-oriented searches in 2011, most of those to find and discover other people. Every LinkedIn search and recommendation is deeply personalized, reflecting the user's current employment, career history, and professional network. In this talk, I will describe some of the challenges and opportunities that arise from working with this unique corpus. I will discuss work we are doing in the areas of relevance, recommendation, and reputation, as well as the ecosystem we have developed to incent people to provide the high-quality semi-structured profiles that make LinkedIn so useful.
Bio:
Daniel Tunkelang leads the data science team at LinkedIn, which analyzes terabytes of data to produce products and insights that serve LinkedIn's members. Prior to LinkedIn, Daniel led a local search quality team at Google. Daniel was a founding employee of faceted search pioneer Endeca (recently acquired by Oracle), where he spent ten years as Chief Scientist. He has authored fourteen patents, written a textbook on faceted search, created the annual workshop on human-computer interaction and information retrieval (HCIR), and participated in the premier research conferences on information retrieval, knowledge management, databases, and data mining (SIGIR, CIKM, SIGMOD, SIAM Data Mining). Daniel holds a PhD in Computer Science from CMU, as well as BS and MS degrees from MIT.
How To Interview a Data Scientist
Daniel Tunkelang
Presented at the O'Reilly Strata 2013 Conference
Video: https://www.youtube.com/watch?v=gUTuESHKbXI
Interviewing data scientists is hard. The tech press sporadically publishes “best” interview questions that are cringe-worthy.
At LinkedIn, we put a heavy emphasis on the ability to think through the problems we work on. For example, if someone claims expertise in machine learning, we ask them to apply it to one of our recommendation problems. And, when we test coding and algorithmic problem solving, we do it with real problems that we’ve faced in the course of our day jobs. In general, we try as hard as possible to make the interview process representative of actual work.
In this session, I’ll offer general principles and concrete examples of how to interview data scientists. I’ll also touch on the challenges of sourcing and closing top candidates.
Presentation given by Dr. Diego Kuonen, CStat PStat CSci, on November 20, 2013, at the "IBM Developer Days 2013" in Zurich, Switzerland.
ABSTRACT
There is no question that big data has hit the business, government and scientific sectors. The demand for skills in data science is unprecedented in sectors where value, competitiveness and efficiency are driven by data. However, there is plenty of misleading hype around the terms big data and data science. This presentation gives a professional statistician's view on these terms and illustrates the connection between data science and statistics.
The presentation is also available at http://www.statoo.com/BigDataDataScience/.
Introduction to Mahout and Machine LearningVarad Meru
This presentation gives an introduction to Apache Mahout and Machine Learning. It presents some of the important Machine Learning algorithms implemented in Mahout. Machine Learning is a vast subject; this presentation is only a introductory guide to Mahout and does not go into lower-level implementation details.
Building Better Products: Selling Data and Decisions to your TeamHannah Flynn
Gathering support for a product feature or enhancement is a critical skill for Product Managers. Talking to customers, working with key stakeholders in the business and convincing development that a feature is necessary can be a daunting task. Join Product Management expert Cait Porte as she covers how to sell your ideas internally by leveraging data to drive decision making.
Creative Sky Blue Presentation Template
If you want to buy this presentation template, please visit http://punkl.com
Creating a presentation from scratch can be quite labour-intensive. Starting with a presentation template from Punkl is beneficial. It saves time, provides good visual design and means that you can primarily spend your time and attention on the content of your presentation.
Punkl Presentation Templates save you time, as they're a whole lot quicker than trying to design a deck from scratch. Also, starting with a template means that you can primarily spend your time and attention on the content of your presentation, while the visual style is already designed to be engaging.
Typically, the only elements that are changed while working with a presentation template are colors, typography, copy and any visual assets such as photos for example.
Creative Sky Blue Presentation Template
If you want to buy this presentation template, please visit http://punkl.com
Creating a presentation from scratch can be quite labour-intensive. Starting with a presentation template from Punkl is beneficial. It saves time, provides good visual design and means that you can primarily spend your time and attention on the content of your presentation.
Punkl Presentation Templates save you time, as they're a whole lot quicker than trying to design a deck from scratch. Also, starting with a template means that you can primarily spend your time and attention on the content of your presentation, while the visual style is already designed to be engaging.
Typically, the only elements that are changed while working with a presentation template are colors, typography, copy and any visual assets such as photos for example.
Power to the People: A Stack to Empower Every User to Make Data-Driven DecisionsLooker
Infectious Media runs on data. But, as an ad-tech company that records hundreds of thousands of web events per second, they have have to deal with data at a scale not seen by most companies. You can not make decisions with data when people need to write manual SQL only for queries take 10-20 minutes to return. Infectious Media made the switch to Google BigQuery and Looker and now every member of every team can get the data they need in seconds.
Infectious Media shares:
- Why they chose their current stack
- Why faster data means happier customers
- Advantages and practical implications of storing and processing that much data
Check out the recording at https://info.looker.com/h/i/308848878-power-to-the-people-a-stack-to-empower-every-user-to-make-data-driven-decisions
Interior Design & Home Decor Business Presentation Template
If you want to buy this presentation template, please visit http://punkl.com
Creating a presentation from scratch can be quite labour-intensive. Starting with a presentation template from Punkl is beneficial. It saves time, provides good visual design and means that you can primarily spend your time and attention on the content of your presentation.
Punkl Presentation Templates save you time, as they're a whole lot quicker than trying to design a deck from scratch. Also, starting with a template means that you can primarily spend your time and attention on the content of your presentation, while the visual style is already designed to be engaging.
Typically, the only elements that are changed while working with a presentation template are colors, typography, copy and any visual assets such as photos for example.
Technical Specification:
100 presentation slides in total.
50 unique and editable presentation slides design.
2 options of color themes variation.
16:9 HD widescreen slide format (1920 x 1080 pixels).
Image placeholder with slide master.
No need Adobe Photoshop or any other image editor.
Fully editable text.
Icons variation are included.
RGB color mode.
Drag and drop image to screen mockups.
Additional Note:
Photos in the preview are not included.
Vector Shape Illustrations are included.
Fonts used are not included, they could be dowloaded from the links on the Documentation File.
Prohibited Usage of Items:
Items Incorporated Into End-Products Cannot be Extracted or Used Separately
You cannot allow items incorporated into end-products to be extracted or used separately from the end-product, **including the vector illustrations and any other vector shapes**.
Enjoy and have a great day! :)
Usage:
Advertising Presentation, Architecture Presentation, Activity Presentation, Brand Guidelines Presentation, Business Presentation, Church Presentation, Cool Presentation, Creative Presentation, Education Presentation, Event Presentation, Finance Presentation, Keynote Presentation, Listing Presentation, Marketing Presentation, Medical Presentation, Pitch Deck Presentation, Product Presentation, Professional Presentation, Sales Presentation, Simple Presentation, Technology Presentation, Programs Presentation, Personal Presentation, Fashion Presentation, Promo Presentation, Shop Presentation, Boutique Presentation, Outlet Presentation, etc.
Interior Design & Home Decor Business Presentation Template
If you want to buy this presentation template, please visit http://punkl.com
Creating a presentation from scratch can be quite labour-intensive. Starting with a presentation template from Punkl is beneficial. It saves time, provides good visual design and means that you can primarily spend your time and attention on the content of your presentation.
Punkl Presentation Templates save you time, as they're a whole lot quicker than trying to design a deck from scratch. Also, starting with a template means that you can primarily spend your time and attention on the content of your presentation, while the visual style is already designed to be engaging.
Typically, the only elements that are changed while working with a presentation template are colors, typography, copy and any visual assets such as photos for example.
Technical Specification:
100 presentation slides in total.
50 unique and editable presentation slides design.
2 options of color themes variation.
16:9 HD widescreen slide format (1920 x 1080 pixels).
Image placeholder with slide master.
No need Adobe Photoshop or any other image editor.
Fully editable text.
Icons variation are included.
RGB color mode.
Drag and drop image to screen mockups.
Additional Note:
Photos in the preview are not included.
Vector Shape Illustrations are included.
Fonts used are not included, they could be dowloaded from the links on the Documentation File.
Prohibited Usage of Items:
Items Incorporated Into End-Products Cannot be Extracted or Used Separately
You cannot allow items incorporated into end-products to be extracted or used separately from the end-product, **including the vector illustrations and any other vector shapes**.
Enjoy and have a great day! :)
Usage:
Advertising Presentation, Architecture Presentation, Activity Presentation, Brand Guidelines Presentation, Business Presentation, Church Presentation, Cool Presentation, Creative Presentation, Education Presentation, Event Presentation, Finance Presentation, Keynote Presentation, Listing Presentation, Marketing Presentation, Medical Presentation, Pitch Deck Presentation, Product Presentation, Professional Presentation, Sales Presentation, Simple Presentation, Technology Presentation, Programs Presentation, Personal Presentation, Fashion Presentation, Promo Presentation, Shop Presentation, Boutique Presentation, Outlet Presentation, etc.
Gain a Holistic View of your Customer's JourneyPlatfora
Today, companies are capturing information about customers at every touchpoint, but the reality is that most companies are working with siloed marketing data because they’re using disparate tools to track online, offline, web, social, mobile, and advertising data.
In this presentation, Rod Fontecilla, VP of Application Modernization at Unisys, explains how his team uses Platfora to analyze, interact and understand data to drive customer success at Unisys.
Rod will highlight three specific Unisys use cases of Platfora, one of which involved a timely text survey sentiment analysis that produced insights enabling a course correction in favor of improved customer satisfaction.
Analytics thought-leader Thomas Davenport and leading industry experts discuss how—and why—organizations like yours use business analytics to empower more timely and precise decisions by bringing new insights into daily operations.
Whether you have an established website, are launching a new one or seeking ideas for a redesign, how do you know what works best for increasing conversions? Does user experience design interfere with your search engine marketing strategies? What design changes can you make now that will bring an instant conversions lift? This fun session provides tips, tools and action items to build a successful website that both search engines and your visitors will love.
Your Roadmap, Your Product Story & Datadriven Product ManagementProduct School
From this presentation you will find out more about becoming a Data-Driven Product Manager.
Get a FREE copy of our Product Book here: https://prdct.school/2BSES8J
Cryptocurrency Business Presentation Template
If you want to buy this presentation template, please visit http://punkl.com
Creating a presentation from scratch can be quite labour-intensive. Starting with a presentation template from Punkl is beneficial. It saves time, provides good visual design and means that you can primarily spend your time and attention on the content of your presentation.
Punkl Presentation Templates save you time, as they're a whole lot quicker than trying to design a deck from scratch. Also, starting with a template means that you can primarily spend your time and attention on the content of your presentation, while the visual style is already designed to be engaging.
Typically, the only elements that are changed while working with a presentation template are colors, typography, copy and any visual assets such as photos for example.
Technical Specification:
100 presentation slides in total.
50 unique and editable presentation slides design.
2 options of color themes variation.
16:9 HD widescreen slide format (1920 x 1080 pixels).
Image placeholder with slide master.
No need Adobe Photoshop or any other image editor.
Fully editable text.
Icons variation are included.
RGB color mode.
Drag and drop image to screen mockups.
Additional Note:
Photos in the preview are not included.
Vector Shape Illustrations are included.
Fonts used are not included, they could be dowloaded from the links on the Documentation File.
Prohibited Usage of Items:
Items Incorporated Into End-Products Cannot be Extracted or Used Separately
You cannot allow items incorporated into end-products to be extracted or used separately from the end-product, **including the vector illustrations and any other vector shapes**.
Enjoy and have a great day! :)
Usage:
Advertising Presentation, Architecture Presentation, Activity Presentation, Brand Guidelines Presentation, Business Presentation, Church Presentation, Cool Presentation, Creative Presentation, Education Presentation, Event Presentation, Finance Presentation, Keynote Presentation, Listing Presentation, Marketing Presentation, Medical Presentation, Pitch Deck Presentation, Product Presentation, Professional Presentation, Sales Presentation, Simple Presentation, Technology Presentation, Programs Presentation, Personal Presentation, Fashion Presentation, Promo Presentation, Shop Presentation, Boutique Presentation, Outlet Presentation, etc.
Similar to How Data Science Builds Better Products - Data Science Pop-up Seattle (20)
What's in your workflow? Bringing data science workflows to business analysis...Domino Data Lab
While business analysis rapidly grows more data-driven, the analyst community is slow to adapt the best practices of data science workflows. Many parallels exists between data science “top topics” (e.g. reproducibility) and business pain points, but these common needs are obscured by the different “languages” of these two communities. The opportunity cost is greatest in heavily regulated industries such as finance and insurance where documentation and compliance are paramount.
In this talk, we will review our experience transitioning Capital One business analysts from legacy systems to open-source workflows by developing user-friendly tools. We incentivized business analysts to adopt the data science mindset by curating open-source tools and developing code packages which simplify workflows and eliminate pain points.
Our internal R package, tidycf, reimagines cumbersome Excel cashflow statements as dataframes and uses RMarkdown templates and the RStudio IDE for an intuitive, user-friendly experience without the overhead of maintaining a custom GUI. We tackle challenges in documentation and communication while immersing new users in the R language.
We will share best practices and lessons learned from our experience designing tools for non-technical end-users, standardizing workflows based on the RStudio IDE’s infrastructure, and evangelizing data science methods.
The Proliferation of New Database Technologies and Implications for Data Scie...Domino Data Lab
In this talk, we’ll describe NoSQL (“not-only SQL”) and document-oriented databases and the value they provide for data science companies like Uptake. We will walk through the unique challenges such datastores pose for data science workflows. To make these challenges and lessons learned concrete, we’ll explore data science workflows through a discussion of the development efforts that led to “uptasticsearch”, an R package released by the Uptake Data Science team to reduce friction in interacting with a document store called Elasticsearch. The talk will conclude with a discussion of recent developments in NoSQL technologies and implications for data scientists.
Racial Bias in Policing: an analysis of Illinois traffic stops dataDomino Data Lab
Since 2004, Illinois has collected demographic information about traffic stops conducted by police in an effort to identify racial bias. This data has been used by groups such as the ACLU and the Stanford Open Policing Project to identify key markers that infer racial bias in policing. We have applied exploratory data analysis to investigate whether systemic racial bias may appear and to what extent. This talk will walk the audience through the insights gleaned from the exploration of this data along with the challenges posed and ongoing questions raised.
Data Quality Analytics: Understanding what is in your data, before using itDomino Data Lab
Analytics and data science are ever growing fields, as business decision makers continue to use data to drive decisions. The pinnacle of these fields are the models and their accuracy/fit,; what about the data? Is your data clean, and how do you know that? Our discussion will focus on best practices for data preprocessing for analytic uses. Beginning with essential distributional checks of a dataset to a propose method for automated data validation process during ETL for transactional data.
Supporting innovation in insurance with randomized experimentationDomino Data Lab
Recent technological advances, a dynamic competitive landscape, and an evolving regulatory environment have led to a period of rapid innovation for many insurance providers. Here, we’ll explore how data scientists may use randomized experiments to rigorously assess the causal impact of innovations on business outcomes. Particular emphasis will be placed on experimentation in “offline” channels, with some of the challenges and mitigation strategies highlighted.
Leveraging Data Science in the Automotive IndustryDomino Data Lab
Cars.com Inc. is a decision engine for car buyers and a growth engine for our partners. Data Science is the bread and butter of any decision engine and Cars is no different. In this talk, I will discuss how we quantify various parameters of a car and plan to make use of all the data in hand to put predictive models at various stages of a users’ automobile lifecycle. This talk will also cater to students looking to gain knowledge on how data science is utilized at scale while still following certain processes and leading the way for business and product partners.
Summertime Analytics: Predicting E. coli and West Nile VirusDomino Data Lab
Lake Michigan and outdoor recreation are enjoyable aspects of summers in Chicago, but it can come with risk of potential E. coli in Lake Michigan or West Nile Virus from mosquitos. This summer, the City of Chicago launched two new predictive analytics projects to forecasts the risks and to proactively limit these risks. Members of the research team, Gene Leynes and Nick Lucius discuss the projects and how they’re being used as part of city operations.
Today, more than ever before, maps are being used to bring data to life. In this presentation I will demonstrate how geoviz can make data science more tangible by providing an interactive canvas for spatial data. Gregory Brunner will shows several examples of how maps are being used enhance how we communicate data and how this applies across all scales, including spatial, temporal, and size of data.
Doing your first Kaggle (Python for Big Data sets)Domino Data Lab
You love python. You love Data Science. But the size of your data set keeps crashing your code. Is it time to bring in big data tools or simply code smarter? Lee is going to show you efficiency hacks, drawn from top Kaggle competitors, to get python to work on large data sets. Skip the hassle of creating a Big Data infrastructure. Let’s find out how far we can push our home laptop first.
Most of analytics modeling work today focuses on the production of single-purpose "artisanal" models for predictions. This approach to analytics is fragile with respect to model consistency, reorganization, and resource availability. This talk will argue that instead the focus of analytics modeling should be toward the production of analytics interchangeable parts, which can be combined in creative ways to produce a wide variety of analytics results. This "nuts and bolts" approach allows analytics groups to produce results in an agile way where the time between ask and answer is determined by the right combination of analytics, rather than the modeling.
How I Learned to Stop Worrying and Love Linked DataDomino Data Lab
In this presentation, Jon Loyens will share:
-Best practices for sharing context and knowledge about your data projects
-How linked data can augment your existing data science workflow and toolchain to accelerate your work
-How a social network can unlock power of Linked Data and data collaboration
-How Linked Data can help you easily combine private and Open Data for fun and profit
Although both disciplines are unique in their own ways, Software Engineering and Data Science make heavy use of programing languages to do their respective jobs. Data Science is a relatively new discipline and many of its practitioners have not previously been professional software engineers. There are a few techniques that Data Scientists can leverage from Software Engineering in order to make their tooling and environments, faster to design, more easily debugged and most importantly, clearer to read. This talk will be going over some practical tips that anyone can use to help better understand their code; give clarity around cloud environments, their uses and drawbacks and finally briefly touching on the Software Development Lifecycle.
Within marketing research, big data is often described as being “census” data for the population that it represents. The devil is in the details and when we take a closer look we can see that this isn’t the case. There are many situations that are not captured within the population that big data purports to be a census of. Big data isn’t even a census of itself since it’s not uncommon for records to be excluded either by accident during the collection process or by design in the cleaning processor. Unfortunately, our industry is so enamored with the size of big data that some users of data are willing to trade off precision for tonnage. Fortunately, if the shortcomings of big data are understood and corrected it can accurately represent the population that it measures in the correct proportion to the universe. We will discuss a method that Nielsen has developed called “Common Homes” that is designed to identify and correct the shortcomings of big data sets that represent media consumption.
Building Data Analytics pipelines in the cloud using serverless technologyDomino Data Lab
Big Data analytics is well known to uncover hidden insights that gives an organization an edge over the competition. But data does not need to be big in order to be useful. Smaller companies and startups may lack the volume of data that qualifies as big data, yet the variety of data can still yield a trove of insights that helps in driving the business strategies of a company. Startups may also lack the resources to fund an additional, seemingly expensive development project. The key is in simplicity, start small, simple and architect for scalability and performance. But how do you start? In this presentation, we share our experience in building a cost effective, AWS serverless data analytics platform that became an invaluable tool for sales, marketing and operational efficiencies.Serverless architectures simplify development work where servers and software are managed by a third party cloud provider. Developers can focus on just building the data wrangling and data analysis logic where critical aspects like scalability and high availability are guaranteed by the cloud provider. Besides, serverless services offer the pay as you go model, where you pay only based on the amount of resources you use. This turns out to be another attractive aspect where costs can be managed based on the usage. In this presentation we will focus on techniques and best practices to build a big data analytics platform using AWS serverless services like Lambda, DynamoDB, S3, Kinesis, Athena, QuickSight and Amazon ML. We will highlight the strengths of each of these services and what role each plays in the data analytics pipeline. We compare and contrast these services with some of the other popularly used big data technologies like Hadoop, Spark and Kafka. We also demonstrate the usage of these services to build intelligent components that detect anomalies, yield recommendations, simulate chat bots and generate predictive analytics.
Leveraging Open Source Automated Data Science ToolsDomino Data Lab
The data science process seeks to transform and empower organizations by finding and exploiting market inefficiencies and potentially hidden opportunities, but this is often an expensive, tedious process. However, many steps can be automated to provide a streamlined experience for data scientists. Eduardo Arino de la Rubia explores the tools being created by the open source community to free data scientists from tedium, enabling them to work on the high-value aspects of insight creation and impact validation.
The promise of the automated statistician is almost as old as statistics itself. From the creations of vast tables, which saved the labor of calculation, to modern tools which automatically mine datasets for correlations, there has been a considerable amount of advancement in this field. Eduardo compares and contrasts a number of open source tools, including TPOT and auto-sklearn for automated model generation and scikit-feature for feature generation and other aspects of the data science workflow, evaluates their results, and discusses their place in the modern data science workflow.
Along the way, Eduardo outlines the pitfalls of automated data science and applications of the “no free lunch” theorem and dives into alternate approaches, such as end-to-end deep learning, which seek to leverage massive-scale computing and architectures to handle automatic generation of features and advanced models.
The Role and Importance of Curiosity in Data ScienceDomino Data Lab
by Alfred Lee
Lead Data Scientist, White Ops
Is curiosity useful for more than serendipitous discovery? Can curiosity be taught? How do I foster curiosity in my team? Can someone be too curious? Questions!
by Jennifer Shin
Senior Principal Data Scientist, Nielsen
With more and more data being collected from consumers, finding a efficient solution to aligning data over time can become increasingly difficult and yet, even more necessary. Whether it's a change in the data collection process or an error in the system, working with big data requires tools that can account for real world complexities.
This talk with introduce the benefits and complexities of implementing a 'fuzzy' solution using the Levenshtein algorithm. Attendees will walk away with a high level understanding of fuzzy matching algorithms and learn how it can be effectively applied to solve real word business problem.
How to Effectively Combine Numerical Features and Categorical FeaturesDomino Data Lab
by Liangjie Hong
Head of Data Science, Etsy
Latent factor models and decision tree based models are widely used in tasks of prediction, ranking and recommendation. Latent factor models have the advantage of interpreting categorical features by a low-dimensional representation, while such an interpretation does not naturally fit numerical features. In contrast, decision tree based models enjoy the advantage of capturing the nonlinear interactions of numerical features, while their capability of handling categorical features is limited by the cardinality of those features. Since in real-world applications we usually have both abundant numerical features and categorical features with large cardinality (e.g. geolocations, IDs, tags etc.), we design a new model, called GB-CENT, which leverages latent factor embedding and tree components to achieve the merits of both while avoiding their demerits. With two real-world data sets, we demonstrate that GB-CENT can effectively (i.e. fast and accurately) achieve better accuracy than state-of-the-art matrix factorization, decision tree based models and their ensemble
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
StarCompliance is a leading firm specializing in the recovery of stolen cryptocurrency. Our comprehensive services are designed to assist individuals and organizations in navigating the complex process of fraud reporting, investigation, and fund recovery. We combine cutting-edge technology with expert legal support to provide a robust solution for victims of crypto theft.
Our Services Include:
Reporting to Tracking Authorities:
We immediately notify all relevant centralized exchanges (CEX), decentralized exchanges (DEX), and wallet providers about the stolen cryptocurrency. This ensures that the stolen assets are flagged as scam transactions, making it impossible for the thief to use them.
Assistance with Filing Police Reports:
We guide you through the process of filing a valid police report. Our support team provides detailed instructions on which police department to contact and helps you complete the necessary paperwork within the critical 72-hour window.
Launching the Refund Process:
Our team of experienced lawyers can initiate lawsuits on your behalf and represent you in various jurisdictions around the world. They work diligently to recover your stolen funds and ensure that justice is served.
At StarCompliance, we understand the urgency and stress involved in dealing with cryptocurrency theft. Our dedicated team works quickly and efficiently to provide you with the support and expertise needed to recover your assets. Trust us to be your partner in navigating the complexities of the crypto world and safeguarding your investments.
How Data Science Builds Better Products - Data Science Pop-up Seattle
1. #datapopupseattle
How Data Science Builds
Better Products
Sean McClure, Ph.D
Data Scientist, Senior Consultant, ThoughtWorks
WorldOfDataSci Thoughtworks
2. #datapopupseattle
UNSTRUCTURED
Data Science POP-UP in Seattle
www.dominodatalab.com
D
Produced by Domino Data Lab
Domino’s enterprise data science platform is used
by leading analytical organizations to increase
productivity, enable collaboration, and publish
models into production faster.
12. data science
the right decisions
+
understands strategyunderstands data
BETTER DISCOVERY
13. Count-controlled loops
Condition-controlled loops
Collection-controlled loops
Infinite loops
Restart loop
Generators
Early exit from loops
Loop variants and invariants
Loop system cross-references
Structured non-local control flow
Conditions
Exceptions
Loops
Flow
Control structures
If-then-(else)
Case and switch
Coroutines
Continuations
STANDARD SOFTWARE
What’s Wrong With the Usual Approach?
All functionality is locked in place
17. Learning algorithms
Model Validation
Model Performance
Data visualization
Operationalizing Models
Scientific computing libraries
Data cleansing
Data preparation
Probability and statistics
Loops
Flow
Control structures
If-then-(else)
Case and switch
Coroutines
Continuations
Count-controlled loops
Condition-controlled loops
Collection-controlled loops
Infinite loops
Restart loop
Generators
Early exit from loops
Loop variants and invariants
Loop system cross-references
Structured non-local control flow
Conditions
Exceptions
ADAPTIVE SOFTWARE
What is the New Approach?
unlocked
23. Successful Data Products
• establish early benchmarks
• understand true validation
• build sophistication via iteration
• provide APIs to model results
• get continuous exposure to domain experience
• design product experiments
Need to utilize technology choices that allow for
building data products successfully
24. Search Engine Marketing - Recommendation
• Increasing CTR?
• Decreasing CPC?
• Call volume trends
• Percentage of Good Call trends.
• Page Position
• Visits vs Cost Per Visit
• Impressions vs CTR graph.
• Breakdown of CVT types
• Click-to-call
• Daily Budget Spend
• Top 5 KWs vs Previous Good Cycle
• Budget distribution
• Impressions per publisher
• Revenue per publisher
• Page position per publisher
• Review for Negative KWs
• Review for Partner site issues
• Review for OAT
• Check Category page
• Impression Share
• Are the ads approved and running?
• Below 1st Page Bid KWs
• Quality Score
• Is it loading?
• Are all numbers replacing correctly?
• Out of Area Traffic
• High Spend – Low Revenue.
• Super Low CTRs
making decisions
25. Data Product
Hadoop Cluster
Databases
DB Data
Producer
Queue
Reporting Data
Operational Data
rl_op
rl_
keyword
rl_
report
HDFS
Flume
Data Core CPI Data Mart
Campaign
Creative
Publishers
Proxy Logs Call Logs
CPI
Admin
Console
Others
Others
Sqoop
CPI
Space
Raw
Nor
mali
zed
Core
Jobs
CPI
Jobs
Search Engine Marketing - Recommendation