Government agencies are collecting and producing data at an accelerating rate, and constituents want access to this data with decreasing latency. Meeting a digitally savvy polity's desire for data while ensuring that data is open, accessible, and interpretable by all comes with unique challenges. I'll share some of these while walking through how governments are building their own data products using open data as well as empowering civic hackers. I'll also walk through why data science at the government level is fundamentally different than data science in the private sector.
Put Alternative Data to Use in Capital Markets Cloudera, Inc.
Alternative data for capital markets, such as satellite imagery, logistics data, and social media feeds, has been getting a lot of attention recently. Like any trending topic, its uses and benefits can be hyped up a bit but if the right plumbing and creativity is in place, those benefits can be realized.
3 things to learn:
* Examples of alt data use cases, sources, and recent market trends
* Why a big data platform that facilitates self service and collaboration is critical in monetizing alternative data
* How alternative data can be applied to enhance current processes (Demo)
Modernizing Architecture for a Complete Data StrategyCloudera, Inc.
Data is the future of business. Either take advantage of it, or get surpassed by those who do.
In this webinar, Ovum's Tony Baer discusses the importance of building a modern data strategy that ensures your journey with Apache Hadoop and big data is a successful one. Together, we'll walk through how to build a plan for long-term success while realizing short-term gains, including:
How to pinpoint the business goals that matter most
How to assess your strengths and weaknesses to meet those goals
How to build a thoughtful approach that ensures your initiatives succeed
The Vortex of Change - Digital Transformation (Presented by Intel)Cloudera, Inc.
The vortex of change continues all around us – inside the company, with our customers and partners. A new norm is upon us. Business models are being turned upside down – the hunters now the hunted, global equalization – size is no longer a guarantee of success. The innovative survive and thrive…the nervous and slow go under...what does all this change means for you? Find out how does Intel’s strengths help our customers in this world of change.
Increase your ROI with Hadoop in Six Months - Presented by Dell, Cloudera and...Cloudera, Inc.
Are you struggling to validate the added costs of a Hadoop implementation? Are you struggling to manage your growing data?
The costs of implementing Hadoop may be more beneficial than you anticipate. Dell and Intel recently commissioned a study with Forrester Research to determine the Total Economic Impact of the Dell | Cloudera Apache Hadoop Solution, accelerated by Intel. The study determined customers can see a 6-month payback when implementing the Dell | Cloudera solution.
Join Dell, Intel and Cloudera, three big data market leaders, to understand how to begin a simplified and cost-effective big data journey and to hear case studies that demonstrate how users have benefited from the Dell | Cloudera Apache Hadoop Solution.
Cloudera Fast Forward Labs: Accelerate machine learningCloudera, Inc.
Machine learning and artificial intelligence can change the world. Diagnosing heart disease. Detecting fraud. Predicting insurance claims. Revolutionizing agriculture. In business, machine learning and artificial intelligence drive new sources of revenue and lower costs.
But executives struggle to define an investment strategy. Researchers introduce innovations in machine learning daily. Technical jargon is opaque. Vendor hype muddies the waters. Industry analysts cover the field, but only at a high level.
Cloudera Fast Forward Labs accelerates your machine learning journey. We deliver a unique blend of applied research and hands-on explanations that you can apply to your business today.
In this webinar you will:
Meet the Cloudera Fast Forward Labs team
Cut through machine learning hype
Explore recent examples of applied research
See exciting new ML techniques
Hear how machine learning is delivering real business value on multiple use cases
3 things to learn:
Explore recent examples of applied research
See exciting new ML techniques
Hear how machine learning is delivering real business value on multiple use cases
Get Started with Cloudera’s Cyber SolutionCloudera, Inc.
Cloudera empowers cybersecurity innovators to proactively secure the enterprise by accelerating threat detection, investigation, and response through machine learning and complete enterprise visibility. Cloudera’s cybersecurity solution, based on Apache Spot, enables anomaly detection, behavior analytics, and comprehensive access across all enterprise data using an open, scalable platform. But what’s the easiest way to get started?
Join Cloudera, StreamSets, and Arcadia Data as we show you first hand how we have made it easier to get your first use case up and running. During this session you will learn:
Signs you need Cloudera’s cybersecurity solution
How StreamSets can help increase enterprise visibility
Providing your security analyst the right context at the right time with modern visualizations
3 things to learn:
Signs you need Cloudera’s cybersecurity solution
How StreamSets can help increase enterprise visibility
Providing your security analyst the right context at the right time with modern visualizations
Becoming Data-Driven Through Cultural ChangeCloudera, Inc.
We've arrived at a crossroads. Big data is an initiative every business knows they should take on in order to evolve their business, but no one knows how to tackle the project.
This is the first in a series of webinars that describe how to break down the challenge into three major pieces: People, Process, and Technology. We'll discuss the industry trends around big data projects, the pitfalls with adopting a modern data strategy, and how to avoid them by building a culture of data-driven teams.
Put Alternative Data to Use in Capital Markets Cloudera, Inc.
Alternative data for capital markets, such as satellite imagery, logistics data, and social media feeds, has been getting a lot of attention recently. Like any trending topic, its uses and benefits can be hyped up a bit but if the right plumbing and creativity is in place, those benefits can be realized.
3 things to learn:
* Examples of alt data use cases, sources, and recent market trends
* Why a big data platform that facilitates self service and collaboration is critical in monetizing alternative data
* How alternative data can be applied to enhance current processes (Demo)
Modernizing Architecture for a Complete Data StrategyCloudera, Inc.
Data is the future of business. Either take advantage of it, or get surpassed by those who do.
In this webinar, Ovum's Tony Baer discusses the importance of building a modern data strategy that ensures your journey with Apache Hadoop and big data is a successful one. Together, we'll walk through how to build a plan for long-term success while realizing short-term gains, including:
How to pinpoint the business goals that matter most
How to assess your strengths and weaknesses to meet those goals
How to build a thoughtful approach that ensures your initiatives succeed
The Vortex of Change - Digital Transformation (Presented by Intel)Cloudera, Inc.
The vortex of change continues all around us – inside the company, with our customers and partners. A new norm is upon us. Business models are being turned upside down – the hunters now the hunted, global equalization – size is no longer a guarantee of success. The innovative survive and thrive…the nervous and slow go under...what does all this change means for you? Find out how does Intel’s strengths help our customers in this world of change.
Increase your ROI with Hadoop in Six Months - Presented by Dell, Cloudera and...Cloudera, Inc.
Are you struggling to validate the added costs of a Hadoop implementation? Are you struggling to manage your growing data?
The costs of implementing Hadoop may be more beneficial than you anticipate. Dell and Intel recently commissioned a study with Forrester Research to determine the Total Economic Impact of the Dell | Cloudera Apache Hadoop Solution, accelerated by Intel. The study determined customers can see a 6-month payback when implementing the Dell | Cloudera solution.
Join Dell, Intel and Cloudera, three big data market leaders, to understand how to begin a simplified and cost-effective big data journey and to hear case studies that demonstrate how users have benefited from the Dell | Cloudera Apache Hadoop Solution.
Cloudera Fast Forward Labs: Accelerate machine learningCloudera, Inc.
Machine learning and artificial intelligence can change the world. Diagnosing heart disease. Detecting fraud. Predicting insurance claims. Revolutionizing agriculture. In business, machine learning and artificial intelligence drive new sources of revenue and lower costs.
But executives struggle to define an investment strategy. Researchers introduce innovations in machine learning daily. Technical jargon is opaque. Vendor hype muddies the waters. Industry analysts cover the field, but only at a high level.
Cloudera Fast Forward Labs accelerates your machine learning journey. We deliver a unique blend of applied research and hands-on explanations that you can apply to your business today.
In this webinar you will:
Meet the Cloudera Fast Forward Labs team
Cut through machine learning hype
Explore recent examples of applied research
See exciting new ML techniques
Hear how machine learning is delivering real business value on multiple use cases
3 things to learn:
Explore recent examples of applied research
See exciting new ML techniques
Hear how machine learning is delivering real business value on multiple use cases
Get Started with Cloudera’s Cyber SolutionCloudera, Inc.
Cloudera empowers cybersecurity innovators to proactively secure the enterprise by accelerating threat detection, investigation, and response through machine learning and complete enterprise visibility. Cloudera’s cybersecurity solution, based on Apache Spot, enables anomaly detection, behavior analytics, and comprehensive access across all enterprise data using an open, scalable platform. But what’s the easiest way to get started?
Join Cloudera, StreamSets, and Arcadia Data as we show you first hand how we have made it easier to get your first use case up and running. During this session you will learn:
Signs you need Cloudera’s cybersecurity solution
How StreamSets can help increase enterprise visibility
Providing your security analyst the right context at the right time with modern visualizations
3 things to learn:
Signs you need Cloudera’s cybersecurity solution
How StreamSets can help increase enterprise visibility
Providing your security analyst the right context at the right time with modern visualizations
Becoming Data-Driven Through Cultural ChangeCloudera, Inc.
We've arrived at a crossroads. Big data is an initiative every business knows they should take on in order to evolve their business, but no one knows how to tackle the project.
This is the first in a series of webinars that describe how to break down the challenge into three major pieces: People, Process, and Technology. We'll discuss the industry trends around big data projects, the pitfalls with adopting a modern data strategy, and how to avoid them by building a culture of data-driven teams.
Webinar - Case Study: ProtectWise enhances network security with DataStax alw...DataStax
DataStax customer, ProtectWise, a leading enterprise security provider, is revolutionizing how network security is managed by using innovative technologies such as machine learning, gaming, and virtual reality powered by an always-on data platform from DataStax to effectively detect fraud and combat network attacks. Hear from ProtectWise Co-Founder & CTO, Gene Stevens, and learn how ProtectWise uses next-gen technologies, including the advanced data management solution from DataStax, to protect customer data.
View recording: https://youtu.be/E1I8sje9GfE
Explore all DataStax webinars: http://www.datastax.com/resources/webinars
Keynote from Big Data World Show Singapore, April 2015.
• How is data driving change?
• Where are the opportunities, across industries?
• What is required to gain value from data?
• How can you get started today?
The Big Picture: Real-time Data is Defining Intelligent OffersCloudera, Inc.
New research shows that 57% of the buying cycle is completed before a prospect even speaks to a company. Marketers already know this, Ninety-six percent (96%) of organizations believe that email personalization can improve email marketing performance. But where do we get this increasingly personal direction? The answer is likely in your customer data. In order to understand your customer needs contextualized in the moment they feel the need to act you will require a platform that can leverage real-time data. Apache Kudu is a Cloudera component that makes dealing with quickly changing data fast and easy. Companies are leveraging next generation data stores like Kudu to build data applications that deliver smart promotions, real-time offers, and personalized marketing. Join us as we discuss modern approaches to real-time application development and highlight key Cloudera use cases being powered by Cloudera’s operational database.
Optimizing Regulatory Compliance with Big DataCloudera, Inc.
3 Things to Learn:
-There are many challenges in the way financial firms deal with regulatory compliance today
-Some of these challenges are related to data management and can be solved by big data technologies
-Cloudera and its partners Trifacta and Qlik are offering a solution that can accelerate the time to obtain compliance reports by using automated workflows and fast analytics that work on top of Cloudera’s Enterprise Data Hub.
Rethink Analytics with an Enterprise Data HubCloudera, Inc.
Have you run into one or more of the following barriers or limitations with your existing data warehousing architecture:
> Increasingly high data storage and/or processing costs?
> Silos of data sources?
> Complexity of management and security?
> Lack of analytics agility?
IoT is reshaping the manufacturing and industrial processes, effectively changing the paradigm from one of repair and replace to more of predict and prevent. Using data streaming from connected equipment and machinery, organizations can now monitor the health of their assets and effectively predict when and how an asset might fail. However, without the right data management strategy and tools, investments in IoT can yield limited results. Join Cloudera and Tata Consultancy Services (TCS) for a joint webinar to learn more about how organizations are using advanced analytics and machine learning to drive IoT enabled predictive maintenance.
Webinar | Aligning GDPR Requirements with Today's Hybrid Cloud RealitiesDataStax
The European Union’s General Data Protection Regulation (GDPR) has sweeping effects on how enterprises manage their data. Without the right policies and safeguards in place, a tiny data mishap could end up turning into a catastrophic mistake. Join Datastax and our partner Thales eSecurity for a live webinar to learn how GDPR effects impact data management and the various ways enterprises can both comply and thrive in a hybrid cloud environment.
View recording: https://youtu.be/QZ48_qkK9PU
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
How Cloudera SDX can aid GDPR compliance 6.21.18Cloudera, Inc.
In this webinar, we will cover:
Technical capabilities required in your data platform including metadata classification on ingest, column-level lineage, fine-grained authorization, encryption, and more
How a shared data experience can facilitate the safe handling of metadata
Ways to enable your data platform for GDPR success
How to Power Innovation with Geo-Distributed Data Management in Hybrid CloudDataStax
Most enterprises understand the value of hybrid cloud. In fact, your enterprise is already working in a multi-cloud or hybrid cloud environment, whether you know it or not. View this SlideShare to gain a greater understanding of the requirements of a geo-distributed cloud database in hybrid and multi-cloud environments.
View recording: https://youtu.be/tHukS-p6lUI
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
Shortening the Sales Cycle with a Modern Data Warehouse 1.30.19Cloudera, Inc.
Join Cloudera as we outline how we use Cloudera technology to strengthen sales engagement, minimize marketing waste, and empower line of business leaders to drive successful outcomes.
Protecting health and life science organizations from breaches and ransomwareCloudera, Inc.
3 Things to Learn About:
* 1. Ransomware is a particular problem and currently the highest priority for healthcare organizations. Machine learning can use the structure of a malicious email to detect an attack even before the email is opened.
* 2. Big data architectures provide the machine-learning models with the volume and variety of data required to achieve complete visibility across the spectrum of IT activity—from packets to logs to alerts.
* 3. Intel and industry partners are currently running one-hour, complimentary, confidential benchmark engagements for HLS organizations that want to see how their security compares with the industry .
Webinar: How Active Everywhere Database Architecture Accelerates Hybrid Cloud...DataStax
In this webinar, we’ll discuss how an Active Everywhere database—a masterless architecture where multiple servers (or nodes) are grouped together in a cluster—provides a consistent data fabric between on-premises data centers and public clouds, enabling enterprises to effortlessly scale their hybrid cloud deployments and easily transition to the new hybrid cloud world, without changes to existing applications.
View recording: https://youtu.be/ob6tr-9YiF4
This session describes the roles and skill sets required when building a Data Science team, and starting a data science initiative, including how to develop Data Science capabilities, select suitable organizational models for Data Science teams, and understand the role of executive engagement for enhancing analytical maturity at an organization.
Objective 1: Understand the knowledge and skills needed for a Data Science team and how to acquire them.
After this session you will be able to:
Objective 2: Learn about the different organizational models for forming a Data Science team and how to choose the best for your organization.
Objective 3: Understand the importance of Executive support for Data Science initiatives and role it plays in their successful deployment.
Getting Started with Data Governance? Use Process Models!DATAVERSITY
In order to prepare for a data governance initiative, companies need to break down their organizational silos and address inefficiencis in business process. Any organization will use a number of different business processes for operations, sales, marketing, and other areas. Capturing the way a business works can be challenging, but data architects should work with business stakeholders to establish process models and ensure they are all speaking the same language to create a solid foundation for enterprise architecture and data governance. IDERA's Kim Brushaber will discuss the importance of business process modeling as the first step for data governance and demonstrate how ER/Studio Business Architect helps companies map and model the relationships between process, people, and data.
Webinar - Case Study: ProtectWise enhances network security with DataStax alw...DataStax
DataStax customer, ProtectWise, a leading enterprise security provider, is revolutionizing how network security is managed by using innovative technologies such as machine learning, gaming, and virtual reality powered by an always-on data platform from DataStax to effectively detect fraud and combat network attacks. Hear from ProtectWise Co-Founder & CTO, Gene Stevens, and learn how ProtectWise uses next-gen technologies, including the advanced data management solution from DataStax, to protect customer data.
View recording: https://youtu.be/E1I8sje9GfE
Explore all DataStax webinars: http://www.datastax.com/resources/webinars
Keynote from Big Data World Show Singapore, April 2015.
• How is data driving change?
• Where are the opportunities, across industries?
• What is required to gain value from data?
• How can you get started today?
The Big Picture: Real-time Data is Defining Intelligent OffersCloudera, Inc.
New research shows that 57% of the buying cycle is completed before a prospect even speaks to a company. Marketers already know this, Ninety-six percent (96%) of organizations believe that email personalization can improve email marketing performance. But where do we get this increasingly personal direction? The answer is likely in your customer data. In order to understand your customer needs contextualized in the moment they feel the need to act you will require a platform that can leverage real-time data. Apache Kudu is a Cloudera component that makes dealing with quickly changing data fast and easy. Companies are leveraging next generation data stores like Kudu to build data applications that deliver smart promotions, real-time offers, and personalized marketing. Join us as we discuss modern approaches to real-time application development and highlight key Cloudera use cases being powered by Cloudera’s operational database.
Optimizing Regulatory Compliance with Big DataCloudera, Inc.
3 Things to Learn:
-There are many challenges in the way financial firms deal with regulatory compliance today
-Some of these challenges are related to data management and can be solved by big data technologies
-Cloudera and its partners Trifacta and Qlik are offering a solution that can accelerate the time to obtain compliance reports by using automated workflows and fast analytics that work on top of Cloudera’s Enterprise Data Hub.
Rethink Analytics with an Enterprise Data HubCloudera, Inc.
Have you run into one or more of the following barriers or limitations with your existing data warehousing architecture:
> Increasingly high data storage and/or processing costs?
> Silos of data sources?
> Complexity of management and security?
> Lack of analytics agility?
IoT is reshaping the manufacturing and industrial processes, effectively changing the paradigm from one of repair and replace to more of predict and prevent. Using data streaming from connected equipment and machinery, organizations can now monitor the health of their assets and effectively predict when and how an asset might fail. However, without the right data management strategy and tools, investments in IoT can yield limited results. Join Cloudera and Tata Consultancy Services (TCS) for a joint webinar to learn more about how organizations are using advanced analytics and machine learning to drive IoT enabled predictive maintenance.
Webinar | Aligning GDPR Requirements with Today's Hybrid Cloud RealitiesDataStax
The European Union’s General Data Protection Regulation (GDPR) has sweeping effects on how enterprises manage their data. Without the right policies and safeguards in place, a tiny data mishap could end up turning into a catastrophic mistake. Join Datastax and our partner Thales eSecurity for a live webinar to learn how GDPR effects impact data management and the various ways enterprises can both comply and thrive in a hybrid cloud environment.
View recording: https://youtu.be/QZ48_qkK9PU
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
How Cloudera SDX can aid GDPR compliance 6.21.18Cloudera, Inc.
In this webinar, we will cover:
Technical capabilities required in your data platform including metadata classification on ingest, column-level lineage, fine-grained authorization, encryption, and more
How a shared data experience can facilitate the safe handling of metadata
Ways to enable your data platform for GDPR success
How to Power Innovation with Geo-Distributed Data Management in Hybrid CloudDataStax
Most enterprises understand the value of hybrid cloud. In fact, your enterprise is already working in a multi-cloud or hybrid cloud environment, whether you know it or not. View this SlideShare to gain a greater understanding of the requirements of a geo-distributed cloud database in hybrid and multi-cloud environments.
View recording: https://youtu.be/tHukS-p6lUI
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
Shortening the Sales Cycle with a Modern Data Warehouse 1.30.19Cloudera, Inc.
Join Cloudera as we outline how we use Cloudera technology to strengthen sales engagement, minimize marketing waste, and empower line of business leaders to drive successful outcomes.
Protecting health and life science organizations from breaches and ransomwareCloudera, Inc.
3 Things to Learn About:
* 1. Ransomware is a particular problem and currently the highest priority for healthcare organizations. Machine learning can use the structure of a malicious email to detect an attack even before the email is opened.
* 2. Big data architectures provide the machine-learning models with the volume and variety of data required to achieve complete visibility across the spectrum of IT activity—from packets to logs to alerts.
* 3. Intel and industry partners are currently running one-hour, complimentary, confidential benchmark engagements for HLS organizations that want to see how their security compares with the industry .
Webinar: How Active Everywhere Database Architecture Accelerates Hybrid Cloud...DataStax
In this webinar, we’ll discuss how an Active Everywhere database—a masterless architecture where multiple servers (or nodes) are grouped together in a cluster—provides a consistent data fabric between on-premises data centers and public clouds, enabling enterprises to effortlessly scale their hybrid cloud deployments and easily transition to the new hybrid cloud world, without changes to existing applications.
View recording: https://youtu.be/ob6tr-9YiF4
This session describes the roles and skill sets required when building a Data Science team, and starting a data science initiative, including how to develop Data Science capabilities, select suitable organizational models for Data Science teams, and understand the role of executive engagement for enhancing analytical maturity at an organization.
Objective 1: Understand the knowledge and skills needed for a Data Science team and how to acquire them.
After this session you will be able to:
Objective 2: Learn about the different organizational models for forming a Data Science team and how to choose the best for your organization.
Objective 3: Understand the importance of Executive support for Data Science initiatives and role it plays in their successful deployment.
Getting Started with Data Governance? Use Process Models!DATAVERSITY
In order to prepare for a data governance initiative, companies need to break down their organizational silos and address inefficiencis in business process. Any organization will use a number of different business processes for operations, sales, marketing, and other areas. Capturing the way a business works can be challenging, but data architects should work with business stakeholders to establish process models and ensure they are all speaking the same language to create a solid foundation for enterprise architecture and data governance. IDERA's Kim Brushaber will discuss the importance of business process modeling as the first step for data governance and demonstrate how ER/Studio Business Architect helps companies map and model the relationships between process, people, and data.
Geek Sync | Tackling Key GDPR Challenges with Data Modeling and GovernanceIDERA Software
You can watch the replay for this Geek Sync webcast in the IDERA Resource Center: http://ow.ly/tLtr50A5b4b
The General Data Protection Regulation (GDPR) is inevitable and goes live in the EU beginning May 25th 2018. It touches all technical and organizational measures as well as the design of internal systems and processes, and affects all companies around the world that have customers in the EU.
Join IDERA and Dr. Sultan Shiffa as he focuses on how data modeling, governance and collaboration help Executives, IT Managers, Architects, DBAs and Developers tackle the key challenges around data protection by design and by default, individual rights to access and erasure, valid consent, data protection roles and accountabilities, data breach notifications, and auditing the records of data processing activities. This session will also explore best practices and examples for how to master those challenges and assess the data protection impact. After this session, you can be prepared to become GDPR compliant ahead of the deadline and beyond.
Announcing the availability of the largest and most comprehensive repository of technology asset data on the planet. Current and constantly curated. Available on the software industry’s first open platform for this data. It’s about time.
Denodo DataFest 2017: Company Leadership from Data LeadershipDenodo
Watch the live session on-demand here: https://goo.gl/Sc6JNG
An increase in data leadership correlates to an increase in business success.
Every single item on a company mission statement relates to data at some level. It is from the position of data expertise that the mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data and projects that will deliver. After all, no matter what business you’re in, you’re in the business of information.
The data leader will anticipate the need -- the voracious need -- for data. If the need does not seem to exist, that is where to start. Commit to growing the data science at your organization. It's simply not enough to be responsive to urgent requests and be the data leader that companies need.
The speaker will share from experience some of the hallmarks of mature, leading data environments that leaders will be guiding their data environments towards in the next few years, with the goal of helping true data leadership emerge.
A Data Privacy & Security Year in Review: Top 10 Trends and PredictionsDelphix
Paying attention to data privacy and security is no longer optional. From a mega breach at Equifax to emerging regulations such as GDPR, data security is driving both today’s headlines and the IT initiatives of tomorrow. Join us for a fascinating discussion on how data privacy and security have evolved in 2017—and what to expect in 2018.
Data & Analytic Innovations: 5 lessons from our customersNick Smith
Nick Smith, managing director at Informatica Australia & New Zealand presented at Chief Data and Analytics Officer Forum in March 2017. In this presentation, he revealed the 5 key lessons learned by Informatica's customers on their quest for data-driven innovation.
Agenda:
1. Cyber Security - How it works, today!
2. Data Analytics, the What and the Why
3. The technical aspects
4. The pipeline
5. Opportunities - Gaps we're aiming for
6. Demo
This presentation, talks about how data analytics can play a significant role in the cyber security space and it also talks about various design challenges associated with datasets in cyber security and how they can be solved.
Are you drowning in data but lacking in insight? 80% of business leaders say data is critical in decision-making, yet 41% cite a lack of understanding of data because it is too complex or not accessible enough. You’ll learn how companies are using graph technology to leverage the relationships in their connected data to reveal new ways of solving their most pressing business problems and creating new business value for their enterprises. You’ll see real-world use cases that include fraud detection, AI/ML, supply chain management, real-time recommendations, Customer 360, network/IT operations and more.
Are you responsible for more than just AppSec? What do you do when you have more teams to support than security experts? How can you make security champions out of dissenters in the development team?
There just aren’t enough security experts to go around. You have to support the multitude of Agile and DevOps teams that are making production software changes anywhere from once a month to several times a day. The lack of resources coupled with the ever increasing responsibilities can make you feel like a rouge warrior in the battle against cybercrime. What’s a security professional to do? Whether you are a team of one or five, there aren’t enough hours in the day and even if there was more budget, good luck finding someone to fill that security role. What if I told you that through careful selection and good training it is possible to build your own army from the very people who own the development process?
What you will learn:
1. Who to recruit as security champions
2. How to train these champions in productive application security
3. How to measure success
4. How to build a scalable security program
5. What to expect from champions (responsibilities)
Creating an EDGE - Enterprise Data Governance ExperienceDATAVERSITY
Industry is challenged to ride the Big Data tidal wave. The average organization doesn’t use half of its structured data in decision-making, and less than one percent of unstructured data is analyzed or used at all, according to Harvard Business Review.
That’s a lot of untapped, ungoverned data assets and therefore unmitigated risks and missed opportunities. However, if data is accessible, reliable and actionable, it can drive serious results – from regulatory compliance (think GDPR) to topline revenue.
So, forget everything you know about data governance as it’s been practiced until now. It’s time to adopt a persona-based approach that joins IT and business functions to ensure organizational objectives are met with everyone – from executives on down – invested in and accountable for data use.
Creating such an end-to-end enterprise data governance experience makes data governance everyone’s business. Then they can manage data’s downsides while maximizing its upsides for optimal organizational performance.
Benchmarking Your GDPR Compliance: Will You Make the Grade? [TrustArc Webinar...TrustArc
Watch the webinar on-demand: https://info.trustarc.com/benchmarking-gdpr-compliance-webinar.html
Register now to watch this on-demand webinar to learn:
- How companies are approaching the GDPR
- Where they are prioritizing their effort
- How much they expect to spend
- These benchmarks can help you position your own program internally and build a case for further investment.
To register for upcoming/on-demand webinars visit: https://www.trustarc.com/events/webinar-schedule/
Who Will Make the Grade?
With less than one year to go before the GDPR is enforced across Europe, how has the industry responded to the GDPR requirements and how many companies will make the grade by May 2018? Recent TrustArc research conducted by Dimensional Research found that over 61% of companies have not even started their GDPR Compliance programs. Of those that had started - the three challenges cited most by the privacy professionals surveyed were difficult to maintain and update privacy programs (57%), lack of appropriate tools and technology (56%), and lack of internal resources (54%).
GrayMatter's partnership with Pentaho, spanning over a decade, reflects their deep expertise in business analytics. They have successfully served 100+ global companies, establishing best practices and a profound understanding of data-driven solutions.
Their performance dashboard and select clientele demonstrate their global reach and customer engagement metrics. GrayMatter's Top Projects Showcase provides concise insights into their Pentaho consulting, Big Data solutions, and stress performance testing capabilities.
At the heart of their services are Business Analytics, covering dashboards, reporting, ad-hoc analysis, visualization, and data mining. Data Integration Services ensure data quality, governance, ETL, master data management, and data warehousing, facilitating data accessibility and reliability.
In the realm of Big Data, GrayMatter's Pentaho proficiency empowers businesses to harness the potential of vast data sources. They also offer seamless migration services and support for embedding Pentaho within organizations, ensuring its seamless integration.
Their extensive project portfolio spans various industries, with notable case studies such as Stream Global, highlighting their remarkable scalability in implementing Enterprise Data Warehouses with Pentaho Data Integration.
In summary, GrayMatter's Pentaho Competency Centre is a testament to their dedication to data-driven excellence, making them a leading player in the business analytics arena. Their partnership with Pentaho isn't just about technology; it's about transforming businesses through data-driven insights that drive success.
Why Your Approach To Data Governance Needs a Major UpdateDelphix
Data is the fuel that powers today’s application-driven businesses. But efficiently governing this data is becoming harder as it grows in volume and proliferates across cloud and on-prem environments. Questions like “Where is my sensitive data located?,” “How do we protect it from breach?” and “How do I ensure the right people have access to it?” are becoming more difficult to answer.
5 Steps to Prepare for Digital Transformation & Real-Time Analytics DATUM LLC
"5 Steps to Prepare for Digital Transformation"
Speaker: David Woods
Abstract: The best time to plant a tree is 20 years ago, the second-best time is right now. In today's digital landscape, it is ideal to proactively plan and prepare for digitization by leveraging a staged approach. Completing the steps along the journey to real-time analytics will deliver tangible business outcomes that you can benefit from now. In this session, we describe the five ways to prepare your enterprise and why you should act now.
Similar to Digital Government: Data + Government Isn't Enough | Wrangle Conference 2017 (20)
Cloudera Data Impact Awards 2021 - Finalists Cloudera, Inc.
This annual program recognizes organizations who are moving swiftly towards the future and building innovative solutions by making what was impossible yesterday, possible today.
The winning organizations' implementations demonstrate outstanding achievements in fulfilling their mission, technical advancement, and overall impact.
The 2021 Data Impact Awards recognize organizations' achievements with the Cloudera Data Platform in seven categories:
Data Lifecycle Connection
Data for Enterprise AI
Cloud Innovation
Security & Governance Leadership
People First
Data for Good
Industry Transformation
2020 Cloudera Data Impact Awards FinalistsCloudera, Inc.
Cloudera is proud to present the 2020 Data Impact Awards Finalists. This annual program recognizes organizations running the Cloudera platform for the applications they've built and the impact their data projects have on their organizations, their industries, and the world. Nominations were evaluated by a panel of independent thought-leaders and expert industry analysts, who then selected the finalists and winners. Winners exemplify the most-cutting edge data projects and represent innovation and leadership in their respective industries.
Machine Learning with Limited Labeled Data 4/3/19Cloudera, Inc.
Cloudera Fast Forward Labs’ latest research report and prototype explore learning with limited labeled data. This capability relaxes the stringent labeled data requirement in supervised machine learning and opens up new product possibilities. It is industry invariant, addresses the labeling pain point and enables applications to be built faster and more efficiently.
Data Driven With the Cloudera Modern Data Warehouse 3.19.19Cloudera, Inc.
In this session, we will cover how to move beyond structured, curated reports based on known questions on known data, to an ad-hoc exploration of all data to optimize business processes and into the unknown questions on unknown data, where machine learning and statistically motivated predictive analytics are shaping business strategy.
Introducing Cloudera DataFlow (CDF) 2.13.19Cloudera, Inc.
Watch this webinar to understand how Hortonworks DataFlow (HDF) has evolved into the new Cloudera DataFlow (CDF). Learn about key capabilities that CDF delivers such as -
-Powerful data ingestion powered by Apache NiFi
-Edge data collection by Apache MiNiFi
-IoT-scale streaming data processing with Apache Kafka
-Enterprise services to offer unified security and governance from edge-to-enterprise
Introducing Cloudera Data Science Workbench for HDP 2.12.19Cloudera, Inc.
Cloudera’s Data Science Workbench (CDSW) is available for Hortonworks Data Platform (HDP) clusters for secure, collaborative data science at scale. During this webinar, we provide an introductory tour of CDSW and a demonstration of a machine learning workflow using CDSW on HDP.
Leveraging the cloud for analytics and machine learning 1.29.19Cloudera, Inc.
Learn how organizations are deriving unique customer insights, improving product and services efficiency, and reducing business risk with a modern big data architecture powered by Cloudera on Azure. In this webinar, you see how fast and easy it is to deploy a modern data management platform—in your cloud, on your terms.
Modernizing the Legacy Data Warehouse – What, Why, and How 1.23.19Cloudera, Inc.
Join us to learn about the challenges of legacy data warehousing, the goals of modern data warehousing, and the design patterns and frameworks that help to accelerate modernization efforts.
Leveraging the Cloud for Big Data Analytics 12.11.18Cloudera, Inc.
Learn how organizations are deriving unique customer insights, improving product and services efficiency, and reducing business risk with a modern big data architecture powered by Cloudera on AWS. In this webinar, you see how fast and easy it is to deploy a modern data management platform—in your cloud, on your terms.
Explore new trends and use cases in data warehousing including exploration and discovery, self-service ad-hoc analysis, predictive analytics and more ways to get deeper business insight. Modern Data Warehousing Fundamentals will show how to modernize your data warehouse architecture and infrastructure for benefits to both traditional analytics practitioners and data scientists and engineers.
Explore new trends and use cases in data warehousing including exploration and discovery, self-service ad-hoc analysis, predictive analytics and more ways to get deeper business insight. Modern Data Warehousing Fundamentals will show how to modernize your data warehouse architecture and infrastructure for benefits to both traditional analytics practitioners and data scientists and engineers.
Explore new trends and use cases in data warehousing including exploration and discovery, self-service ad-hoc analysis, predictive analytics and more ways to get deeper business insight. Modern Data Warehousing Fundamentals will show how to modernize your data warehouse architecture and infrastructure for benefits to both traditional analytics practitioners and data scientists and engineers.
Extending Cloudera SDX beyond the PlatformCloudera, Inc.
Cloudera SDX is by no means no restricted to just the platform; it extends well beyond. In this webinar, we show you how Bardess Group’s Zero2Hero solution leverages the shared data experience to coordinate Cloudera, Trifacta, and Qlik to deliver complete customer insight.
Federated Learning: ML with Privacy on the Edge 11.15.18Cloudera, Inc.
Join Cloudera Fast Forward Labs Research Engineer, Mike Lee Williams, to hear about their latest research report and prototype on Federated Learning. Learn more about what it is, when it’s applicable, how it works, and the current landscape of tools and libraries.
Analyst Webinar: Doing a 180 on Customer 360Cloudera, Inc.
451 Research Analyst Sheryl Kingstone, and Cloudera’s Steve Totman recently discussed how a growing number of organizations are replacing legacy Customer 360 systems with Customer Insights Platforms.
Build a modern platform for anti-money laundering 9.19.18Cloudera, Inc.
In this webinar, you will learn how Cloudera and BAH riskCanvas can help you build a modern AML platform that reduces false positive rates, investigation costs, technology sprawl, and regulatory risk.
Introducing the data science sandbox as a service 8.30.18Cloudera, Inc.
How can companies integrate data science into their businesses more effectively? Watch this recorded webinar and demonstration to hear more about operationalizing data science with Cloudera Data Science Workbench on Cazena’s fully-managed cloud platform.
In this webinar, we’ll show you how Cloudera SDX reduces the complexity in your data management environment and lets you deliver diverse analytics with consistent security, governance, and lifecycle management against a shared data catalog.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
We're SaaS business providing cloud-based solutions for data-driven government.
Data-as-a-service platform and cloud applications for government agencies
We make government data discoverable, reusable, and actionable
I’m a social scientist turned data scientist turned product manager, so I think a lot about:
how humans generate data and how that data gets encoded, and then how that encoded data gets turned into models.
Here I am at our annual employee summit – I was in the middle of talking to Dave Doyle, the City of Seattle open data program manager, who had just given closing remarks, and didn’t realize a group photo was forming. Or, as a coworker said to me after seeing this photo “man you must really love that laptop"
I own several of our backend services –- all of the way that our customers ingress data onto our platform – as well as a chunk of our machine learning infrastructure.
What do we build to enable open data?Open data portals –
- web interfacet to a cloud-based data-as-a-service platform that lets agencies provide open data to their communities with built-in APIs, search, etc. making the data discoverable and resuable.
- this is probably what most people think of when they think of open data
- lots of cities, counties, and states, some NGOs and federal agencies as well
Open data powers applications as well that help agencies plan and communicate with stakeholders:
- here’s the City of Seattle’s open budget
- Lets anyone explore the budget in nitty gritty detail, interactively
Becoming more popular – performance programs
- lets agencies provide transparency and accountability for the goals they've set for programs and initiatives
- think of it as a public dashboard with KPIs that anyone can check in on
- Budgets and performance dashboards are changes in the consumption experience for the raw data that open data portals host
- represent the overall maturation of open data, as we move from data for data's sake to solving specific problems with open data and putting open data in the path of government work, rather than a destination where data lives
- Even Steve Ballmer is getting in on the action
Side project to help Americans understand the flow of money in government
No LA Clippers salary cap explorer – but he’ll have 21 million or so freed up without Chris Paul in the upcoming season.
However, USA facts is having the same problems with engagement that I'll talk about later in this talk – at the national governor's association last week, he said after a big flurry of publicity, they're only getting about 4000 visitors a day.
Just putting data up for people to consume often doesn't produce a ton of engagement
What’s the state of open data in 2017?
- Broke the rules of giving a talk by discussing open data so far without really talking about what it is and what we mean by it
Open data as both an idea and a practice really picked up steam during the Obama Administration – but the concept has been around for a while.
So what do we mean?
Data can be open in a couple of ways –
machine readable, available programmatically – this means data in widely accepted formats like CSV, JSON, XML, not locked up in a PDF or stored in physical copies that require a FOIA to get at
It also means APIs that allow programmatic retrieval, and to enable developers to build applications with the data – if you've used a non-city-created transit application to find out when your bus/train is coming, you've benefited from this kind of open data.
permissively licensed for reuse
This is obviously a stickier issue – as licensing always is – and differs from agency to agency, but there is a generally agreed upon idea that open data should be available for reuse by anyone that wants to use it (at the very least non-commercially) but often this means for commercial reuse as well. Some examples of data reuse in applications include restaurant inspections in Yelp reviews or data used by Zillow for estimating housing value.
Sunlight Foundation, a non-profit dedicated to making governments accountable and transparent, has published guidelines for open data, many of which have been adopted by government agencies in creating their open data polices.
Over the past 10 years or so, we've seen a number of municipalities create policies or legislation mandating that any data that can be made open be made open. The implementation and wording differs from agency to agency, but a growing number.
Open Data Policies dot org – branched off from sunlight -- lists 101 current open data policies at the city, county, and state levels
Federal level: DATA Act: Digital Accountability and Transparency Act (2014) – how does the government spend its money?
This widespread adoption has produced real results for many agencies.
The city of Chicago saw their freedom of information act requests drop by 50% when they launched their open data portal. Chicago is a real leader in open data – constantly pushing us – and open sourcing a lot of their work for other governments to use. They recently relaunched their open data portal (data.cityofchicago.org) to great success.
The Dallas PD saw requests for data on officer-involved shootings drop to zero (dallasopendata.com) when they started releasing that data on their open data portal.
Fulfilling requests for records and freedom of information act requests are resource-intensive. They’re usually time-bound and failure to comply carries a penalty.
Beyond time saving and request-fulfilling efficiency, Where are we with open data? What are some of the success stories?Hopefully the data scientists in the crowd will allow me a bit of selection on the dependent variable for a moment.
NYC has one of the most active open data programs in the country, thanks to a combination of a very talented staff and a legislated primary goal for all city agencies to share open data.
The Department of Information Technology and Telecommunications (best acronym -- DoITT) and MODA (the Mayor’s Office of Data Analytics) manage a tremendously complicated program with great success – acting as central hubs for many of the city's agencies and their data. NYC also has stringent retention policies, so a lot of moving pieces to manage.
A favorite dataset of mine is the NYC tree census of all trees in NYC – whenever we're testing out geospatial features, we usually use that dataset as one of our testers.
- Probably the most common thing many people think of when they think of Open Data – citizens serving as independent watchdogs, finding inefficiencies, injustices, and just plain mistakes. “Transparency”
Ben Wellington is a quantitative analyst (who runs the popular blog I Quant NY) at Two Sigma – which if you’ve been to a recent PyData or SciPy conference will be familiar did this analysis using NYC open data.
Something worth over $33,000 in this picture – and it’s not the Mercedes.
Using NYC Open Data, found 84 tickets over a 4.5-month period -- $33,000 a year in fines, not including towing fees.
One block over, another hydrant generating 24,000 a year – over 55,000 a year on two blocks
Open data for the same reason you crowdsource things
Can’t think of all possible questions to ask, instead, rely on motivated individuals to ask them
Of course, this cost the city 55,000 a year!
This has a bit of a bizarre twist, as there’s some confusion over whether these cars were parked legally or not.
Using Google Maps – there is a protected bike lane between the cars and the fire hydrant.
New Orleans is a real leader in performance management – a city with a number of well-known and lesser-known problems that really embraced tying its decision-making to open data.
- NOLA combined data from the American Housing Survey, American Community Survey (both from the US Census) and FD admin data
Office of Performance & Analytics identified homes most at-risk for fire
Distributed 8,000 smoke detectors, reducing fire deaths & injuries
A family of 11 later that year escaped a house fire after a smoke alarm went off at 3am – that had been installed as part of this program
Also increased operational efficiency of FD by modeling where fires are most likely to occur
Blight is a well-known problem in New Orleans, which has experienced natural disasters and social crises over the past century – homes left abandoned, boarded up in a city that is always at risk of being reclaimed by the earth.
Blight is problematic for a number of reasons – crime, pest infestation, public safety, and that lot is blocking new development / revenue collection / depressing home values elsewhere.
- Blight affected up to ¼ of all residential addresses in NOLA after Katrina
Formed BlightSTAT – cross-department task force to reduce blight using data
Decreased blight by 30% especially impressive when benchmarked against peer cities where abandoment rates are climbing
cut response time from initial inspection to a hearing in half – over 3 months reduction
- Jackson, MS – like many cities and like New Orleans, plagued by aging infrastructure
- the kind of infrastructure that sees schoolbuses fall into sinkholes TWICE in three years – these are different buses on different streets.
- more than half of the city's bridges in dire need of repairs
Last year, one of these bridges targeted for repairs totally collapsed due to flooding – after closing, thankfully, but an example of the critical state of affairs.
Estimates to repair Jackson's infrastructure range from 750 million to 1 billion dollars – that's 5-6 times the annual revenue of the city all told.
City has one lever it can turn to generate revenue --
1 percent sales tax voted in, with revenue reserved for capital projects
Jackson knew that the only way to keep the public's trust was to be open and transparent about how this revenue was being used and how it was tracking to meet its goals
This slide, which you saw earlier, is actually part of Jackson's performance program – called JackStats
Of particular interest given the bus-sinkhole problem is Operation Orange Cone
By using 311data to more efficiently dispatch repair crews and identify problem areas,
Filled over 69,000 potholes in 2 years, a 60% increase in pothole-filling compared to before Operation Orange Cone.
Some of these complaints dated to 2010!
AND also decreased 311 calls
Now residents can track the progress of Operation Orange Cone in a number of places and see how the project is doing on time and on budget, as well as getting up-to-date information about what streets are planned for resurfacing.
Operation Orange Cone was supposed to be a two-week pilot program, but has run continuously for 2 years now.
From the agency’s perspective, you need a strategy for success.
Open data programs are most successful when they see high engagement with constituents and residents.
Engagement is key – the mayor/governor wants to see that these open data programs are providing value. One way they do that is via constituent engagement. Checking the open data box doesn’t drive eyeballs.
Need to figure out what problem you're solving and who you're solving it for. (Selfishly, I might say that open data programs need a good product manager but that’s probably another talk).
Just releasing open data for civic hackers only helps a small portion of the population.
When Jackson, MS needed to fix its roads to keep buses from falling into sinkholes, they made the data part of the planning process, the decisions that were made, and the reporting on the actions that resulted from those decisions.
Putting a CSV on an open data portal doesn’t drive engagement, and is at worst a good way to have stale, out-of-date data that no one trusts or relies on.
Each of the previous success stories began with some problem to be solved or goal to be accomplished – that’s what makes open data compelling. Data is only part of the story – what happens with the data is the magic of open data.
Data-informed decision-making (credit to Greg Reda at PyData Seattle) needs data.
For that data to be effective, that data needs to be up to date and authoritative. Government information workers need to be able to trust that data and integrate it into their workflows.
“Open data portals” are a red herring – it shouldn’t be where old CSVs go to retire – it should be where government workers look for data they need to do their job.
For every data scientist with a phd working at the NSA or the Census, there’s an analyst working at the department of public works in a city that has to work on 20 projects at once. They’re working with limited resources, limited time, and a public that often doesn’t care a whole lot about how busy they are.
Public service isn’t just a catchphrase – there’s real service here.
Your work is out there for the public to see. When your audience is everyone, you have to show your work. This means opening your data and explaining your models.
Sometimes your work is going to be featured in the local news, and there will often be press releases about it.
On the other hand, your work is going to be out there for everyone to see! I bet a lot of you don't get to talk about your work.
Over 30,000 emails published for anyone to read. Released every Friday. This used to be a FOIA request that the local news outlets would make every week, now they just release them automatically to save them the trouble. Complex workflow that is half-automated, half-manual (can't release PII, constituent information, etc.).
Government moves slowly – and upgrade cycles are no exception. Getting data from one agency to another is a real challenge.
Talking to an analyst at a major west coast city – people drive around different lots in the city to verify permitting status, then literally fill out forms by hand, which are then delivered to a central office for data entry at a later date. By the time they’re digitized, who knows what has changed?
Budgets run by department, but problems span departments – there’s no “department of homelessness”, but the problem requires action across agencies. Each agency has its own budget, and may use a different database or ERP system to track their data. It’s not as simple as a JOIN.
You can’t just use that new library you read about on Hacker News into production – IT departments vary wildly in their permissiveness, and changes to infrastructure are extremely difficult. CIOs or CDOs may make software decisions for entire departments.
You’re probably not going to be using Spark to build a predictive machine learning model that runs on a cluster and scales to millions of predictions a minute.
Some agencies allow users to submit data via forms – and it’s messy. Different languages, profanity, and then some. I was speaking to an open data program manager from a city in the south and he was telling me that when they were cleaning up this data, retention laws meant that they had to go in and indicate where data had been changed from its original form.
You think data scientists in the private sector complain about how hard it is to clean their data…
Accessibility is a compliance matter – data you release and websites you host it on have to be accessible to people with disabilities, and there’s a shifting landscape of standards.
You have no idea who will be accessing your data – it’s open! – and you can’t assume that they’ll be subject matter experts. Metadata, documentation are key – but of course they are time consuming.
On the other hand – your audience is everyone! You get to work on things that affect all sorts of people in very real ways. Your predictive models have real impact – like saving lives through fire detector placement.
For every data scientist with a phd working at the NSA or the Census, there’s an analyst working at the department of public works in a city that has to work on 20 projects at once. They’re working with limited resources, limited time, and a public that often doesn’t care a whole lot about how busy they are.
Public service isn’t just a catchphrase – there’s real service here.