Azavea develops software to analyze large geospatial datasets for social good. Their Big (Geo) Data Science work includes spatial forecasting of crime patterns in Philadelphia using statistical analysis of past incidents. They created automated maps and alerts for the police department to help predict crime hotspots and accelerate response times. Azavea also conducts open source research on high performance geoprocessing techniques to enable analysis of massive global datasets within seconds.
Crime Risk Forecasting and Predictive Analytics - Esri UCAzavea
Presentation at the 2011 Esri User Conference that included an overview of HunchLab features related to forecasting, specifically near repeat forecasts and load forecasts.
As we develop our crime analysis software, HunchLab, we are always on the look out for ways of examining and improving data quality as well as new academic research that shows promise to enhance crime analysis.
In this one-hour webinar, we first explain some of the ways we examine data quality when we utilize historic incident datasets for research and analysis and how you can use these techniques in your department. Then, we walk through a series of analytic techniques and practices that can help your department improve your crime analysis processes.
Crime Risk Forecasting: Near Repeat Pattern Analysis & Load ForecastingAzavea
http://www.azavea.com/hunchlab
This is a rather technical dive into the near repeat pattern analysis and load forecasting features that we've built into HunchLab. Both of these features are aimed at helping a law enforcement agency to better predict risk levels across their jurisdictions and allocate resources according. While no application of predictive analytics will be perfect, forecasting risk based on models of the past can help officers and analysts to anticipate the appropriate next steps.
Near repeat pattern analysis helps officers quantify the risk that arises from multiple incidents happening close to one another in space and time. What we are quantifying is how the fact that your neighbor's house is burgled raises your risk of a burglary in the coming days and weeks.
With load forecasting we are looking at cyclical temporal patterns in incidents. How does the time of year, time of day, and day of week change the levels of crime incidents that we should expect across a jurisdiction? By modeling these cyclical patterns we can project crime levels into the future, helping law enforcement agencies to allocate resources appropriately as well as better manage organizational accountability.
Crime Early Warning: Automated Data Mining of CAD and RMSAzavea
The genesis of HunchLab was the idea to mine law enforcement agencies' CAD and RMS databases to detect unusual levels of activity in particular areas and then send alerts to the appropriate police staff. While crime analysis tools often are aiming to display what has happened, the concept of a geographic early warning system, such as within HunchLab, tries to answer the question: "what is unusual that is happening?"
http://www.azavea.com/products/hunchlab/features/early-warning/
The Real-time Police Force: Publishing Analytic Information to the Field with...Azavea
http://www.azavea.com/products/hunchlab
Police agencies collect a wealth of data. Every call for services and every incident of crime is captured and logged (and often automatically geocoded to a point in space). Making sense of this wealth of data is critical to police agencies being led by intelligence and analysis and not simply putting cops out into the field haphazardly.
Most police forces have a process whereby this raw information is groomed into maps by a central crime analysis unit. Determining where hotspots are present and describing recent events is definitely useful, but how can we accelerate this process to adapt our analytic output in nearly real-time and then disseminate this information to the field?
The answer is by automating the flow of information. We see this feature as a core strength within our product, HunchLab. New information is automatically pulled into HunchLab through integration with police agencies computer aided dispatch (CAD) and records management systems (RMS). This new information is then immediately incorporated into analytic output. New incidents can trigger early warning alerts for spikes in activity or modify short-term risk assessment in a particular police district. But it's not just about consuming this information within HunchLab itself. The system provides secure access to analytic output via APIs that can be integrated into other back-end applications, further analytic tools, and even mobile applications.
http://www.azavea.com/products/hunchlab
Deep Learning for Public Safety in Chicago and San FranciscoSri Ambati
Presentation on Deep Learning for Public Safety using open data sets from the cities of San Francisco and Chicago.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
As per studies conducted by the University of California, it is observed that crime in any area follows the same pattern as that of earthquake aftershocks. It is difficult to predict an earthquake, but once it happens the aftershocks following it are quite predictable. Same is true for the crimes happening in a geographical area.
Crime Risk Forecasting and Predictive Analytics - Esri UCAzavea
Presentation at the 2011 Esri User Conference that included an overview of HunchLab features related to forecasting, specifically near repeat forecasts and load forecasts.
As we develop our crime analysis software, HunchLab, we are always on the look out for ways of examining and improving data quality as well as new academic research that shows promise to enhance crime analysis.
In this one-hour webinar, we first explain some of the ways we examine data quality when we utilize historic incident datasets for research and analysis and how you can use these techniques in your department. Then, we walk through a series of analytic techniques and practices that can help your department improve your crime analysis processes.
Crime Risk Forecasting: Near Repeat Pattern Analysis & Load ForecastingAzavea
http://www.azavea.com/hunchlab
This is a rather technical dive into the near repeat pattern analysis and load forecasting features that we've built into HunchLab. Both of these features are aimed at helping a law enforcement agency to better predict risk levels across their jurisdictions and allocate resources according. While no application of predictive analytics will be perfect, forecasting risk based on models of the past can help officers and analysts to anticipate the appropriate next steps.
Near repeat pattern analysis helps officers quantify the risk that arises from multiple incidents happening close to one another in space and time. What we are quantifying is how the fact that your neighbor's house is burgled raises your risk of a burglary in the coming days and weeks.
With load forecasting we are looking at cyclical temporal patterns in incidents. How does the time of year, time of day, and day of week change the levels of crime incidents that we should expect across a jurisdiction? By modeling these cyclical patterns we can project crime levels into the future, helping law enforcement agencies to allocate resources appropriately as well as better manage organizational accountability.
Crime Early Warning: Automated Data Mining of CAD and RMSAzavea
The genesis of HunchLab was the idea to mine law enforcement agencies' CAD and RMS databases to detect unusual levels of activity in particular areas and then send alerts to the appropriate police staff. While crime analysis tools often are aiming to display what has happened, the concept of a geographic early warning system, such as within HunchLab, tries to answer the question: "what is unusual that is happening?"
http://www.azavea.com/products/hunchlab/features/early-warning/
The Real-time Police Force: Publishing Analytic Information to the Field with...Azavea
http://www.azavea.com/products/hunchlab
Police agencies collect a wealth of data. Every call for services and every incident of crime is captured and logged (and often automatically geocoded to a point in space). Making sense of this wealth of data is critical to police agencies being led by intelligence and analysis and not simply putting cops out into the field haphazardly.
Most police forces have a process whereby this raw information is groomed into maps by a central crime analysis unit. Determining where hotspots are present and describing recent events is definitely useful, but how can we accelerate this process to adapt our analytic output in nearly real-time and then disseminate this information to the field?
The answer is by automating the flow of information. We see this feature as a core strength within our product, HunchLab. New information is automatically pulled into HunchLab through integration with police agencies computer aided dispatch (CAD) and records management systems (RMS). This new information is then immediately incorporated into analytic output. New incidents can trigger early warning alerts for spikes in activity or modify short-term risk assessment in a particular police district. But it's not just about consuming this information within HunchLab itself. The system provides secure access to analytic output via APIs that can be integrated into other back-end applications, further analytic tools, and even mobile applications.
http://www.azavea.com/products/hunchlab
Deep Learning for Public Safety in Chicago and San FranciscoSri Ambati
Presentation on Deep Learning for Public Safety using open data sets from the cities of San Francisco and Chicago.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
As per studies conducted by the University of California, it is observed that crime in any area follows the same pattern as that of earthquake aftershocks. It is difficult to predict an earthquake, but once it happens the aftershocks following it are quite predictable. Same is true for the crimes happening in a geographical area.
Our goal is to create a web application that would give insights to its user about the crime scenario and its various aspects in Chicago.
Our application will contain:
A search box/drop down list where user can select a district.
Geospatial analysis using ArcGIS maps and visualizations that are embedded into the web app which will be dynamically updated to show most interesting patterns or heat maps for that district.
Statistical analysis and visualizations on historical data to the user.
Prediction of the date when the next crime will happen and its probability.
Crime Analysis & Prediction System is a system to analyze & detect crime hotspots & predict crime.
It collects data from various data sources - crime data from OpenData sites, US census data, social media, traffic & weather data etc.
It leverages Microsoft's Azure Cloud and on premise technologies for back-end processing & desktop based visualization tools.
Crime analysis mapping, intrusion detection using data miningVenkat Projects
Crime analysis mapping, intrusion detection using data mining
Data Mining plays a key role in Crime Analysis. There are many different algorithms mentioned in previous research papers, among them are the virtual identifier, pruning strategy, support vector machines, and apriori algorithms. VID is to find relation between record and vid. The apriori algorithm helps the fuzzy association rules algorithm and it takes around six hundred seconds to detect a mail bomb attack. In this research paper, we identified Crime mapping analysis based on KNN (K – Nearest Neighbor) and ANN (Artificial Neural Network) algorithms to simplify this process. Crime Mapping is conducted and Funded by the Office of Community Oriented Policing Services (COPS). Evidence based research helps in analyzing the crimes. We calculate the crime rate based on the previous data using data mining techniques. Crime Analysis uses quantitative and qualitative data in combination with analytic techniques in resolving the cases. For public safety purposes, the crime mapping is an essential research area to concentrate on. We can identity the most frequently crime occurring zones with the help of data mining techniques. In Crime Analysis Mapping, we follow the following steps in order to reduce the crime rate: 1) Collect crime data 2) Group data 3) Clustering 4) Forecasting the data. Crime Analysis with crime mapping helps in understanding the concepts and practice of Crime Analysis in assisting police and helps in reduction and prevention of crimes and crime disorders.
Crime Mapping & Analysis – Georgia Tech
Crime analysis is a law enforcement function that involves systematic analysis for identifying and analyzing patterns and trends in crime and disorder. Information on patterns can help law enforcement agencies deploy resources in a more effective manner, and assist detectives in identifying and apprehending suspects.
PredPol: How Predictive Policing WorksPredPol, Inc
PredPol’s cloud-based predictive policing software enables law enforcement agencies to better prevent crime in their communities by generating predictions on the places and times that future crimes are most likely to occur.
PredPol’s technology has been helping law enforcement agencies to dramatically reduce crime in jurisdictions of all types and sizes, across the U.S. and overseas. Over the past year, Atlanta and Los Angeles have reduced specific crimes in targeted areas at rates ranging from nearly 20% to over 40%. Smaller jurisdictions, such as Norcross, Georgia, have seen nearly a 30% reduction in burglaries and robberies; in Alhambra, California, car burglaries have dropped 20% since the software technology was deployed.
Using advanced mathematics and computer learning, PredPol’s algorithms predict many types of crime, including property crimes, drug incidents, gang activity, and gun violence as well as traffic accidents.
Only three pieces of data are used to make predictions – type of crime, place of crime, and time of crime. No personal data is utilized in making these predictions.
Crime analysts and command staff using PredPol are 100% more effective than they are with traditional hotspot mapping at predicting where and when crimes are likely to occur. That means police have twice as many opportunities to deter and reduce crime.
Adoption of the R language has grown rapidly in the last few years, and is ranked as the number-one data science language in several surveys. This accelerating R adoption curve has been driven by the Big Data revolution, and the fact that so many data scientists — having learned R at university — are actively unlocking the secrets hidden in these new, vast data troves. In more than 6 years of writing for the Revolutions blog, I’ve discovered hundreds of applications of R in business, in government, and in the non-profit sector. Sometimes the use of R is obvious, and sometimes it takes a little bit of detective work to learn how R is operating behind the scenes. In this talk, I'll recount some of my favourite applications of R, and show how R is behind some amazing innovations in today’s world.
[Presented to the 7th China R Users Conference, Beijing, May 2014.]
Adoption of the R language has grown rapidly in the last few years, and is ranked as the number-one data science language in several surveys. This accelerating R adoption curve has been driven by the Big Data revolution, and the fact that so many data scientists — having learned R at university — are actively unlocking the secrets hidden in these new, vast data troves.
In more than 6 years of writing for the Revolutions blog, I’ve discovered hundreds of applications of R in business, in government, and in the non-profit sector. Sometimes the use of R is obvious, and sometimes it takes a little bit of detective work to learn how R is operating behind the scenes. In this talk, I’ll begin by presenting some recent statistics on the growth of R. Then I’ll recount some of my favourite applications of R, and show how R is behind some amazing innovations in today’s world.
Order Fulfillment Forecasting at John Deere: How R Facilitates Creativity and...Revolution Analytics
Statistical analysis has been known to be invaluable to any manufactory’s quality assurance for decades. Recently the value of valid statistical analysis has also been demonstrated to radically improve the ability of a company’s ability to weather extreme peaks and valley in customer demand. John Deere has been able to adjust to commodity spikes and housing downturns much better than its competitors have. This is in part due to the implementation of statistical analysis and the use of R software in the order fulfillment function of John Deere.
Using Data Mining Techniques to Analyze Crime PatternZakaria Zubi
Our proposed model will be able to extract crime patterns by using association rule mining and clustering to classify crime records on the basis of the values of crime attributes.
Video transmission over wireless networks is considered the most interesting application in our daily life nowadays. As
mobile data rates continue to increase and more people rely on wireless transmission, the amount of video transmitted over at least one
wireless hop will likely continue to increase. This kind of application needs large bandwidth, efficient routing protocols, and content
delivery methods to provide smooth video playback to the receivers. Current generation wireless networks are likely to operate on
internet technology combined with various access technologies. Achieving effective bandwidth aggregation in wireless environments
raises several challenges related to deployment, link heterogeneity, Network congestion, network fluctuation, and energy consumption.
In this work, an overview of technical challenges of over wireless networks is presented. A survey of wireless networks in recent video
transmission schemes is introduced. Demonstration results of few scenarios are showed.
Our goal is to create a web application that would give insights to its user about the crime scenario and its various aspects in Chicago.
Our application will contain:
A search box/drop down list where user can select a district.
Geospatial analysis using ArcGIS maps and visualizations that are embedded into the web app which will be dynamically updated to show most interesting patterns or heat maps for that district.
Statistical analysis and visualizations on historical data to the user.
Prediction of the date when the next crime will happen and its probability.
Crime Analysis & Prediction System is a system to analyze & detect crime hotspots & predict crime.
It collects data from various data sources - crime data from OpenData sites, US census data, social media, traffic & weather data etc.
It leverages Microsoft's Azure Cloud and on premise technologies for back-end processing & desktop based visualization tools.
Crime analysis mapping, intrusion detection using data miningVenkat Projects
Crime analysis mapping, intrusion detection using data mining
Data Mining plays a key role in Crime Analysis. There are many different algorithms mentioned in previous research papers, among them are the virtual identifier, pruning strategy, support vector machines, and apriori algorithms. VID is to find relation between record and vid. The apriori algorithm helps the fuzzy association rules algorithm and it takes around six hundred seconds to detect a mail bomb attack. In this research paper, we identified Crime mapping analysis based on KNN (K – Nearest Neighbor) and ANN (Artificial Neural Network) algorithms to simplify this process. Crime Mapping is conducted and Funded by the Office of Community Oriented Policing Services (COPS). Evidence based research helps in analyzing the crimes. We calculate the crime rate based on the previous data using data mining techniques. Crime Analysis uses quantitative and qualitative data in combination with analytic techniques in resolving the cases. For public safety purposes, the crime mapping is an essential research area to concentrate on. We can identity the most frequently crime occurring zones with the help of data mining techniques. In Crime Analysis Mapping, we follow the following steps in order to reduce the crime rate: 1) Collect crime data 2) Group data 3) Clustering 4) Forecasting the data. Crime Analysis with crime mapping helps in understanding the concepts and practice of Crime Analysis in assisting police and helps in reduction and prevention of crimes and crime disorders.
Crime Mapping & Analysis – Georgia Tech
Crime analysis is a law enforcement function that involves systematic analysis for identifying and analyzing patterns and trends in crime and disorder. Information on patterns can help law enforcement agencies deploy resources in a more effective manner, and assist detectives in identifying and apprehending suspects.
PredPol: How Predictive Policing WorksPredPol, Inc
PredPol’s cloud-based predictive policing software enables law enforcement agencies to better prevent crime in their communities by generating predictions on the places and times that future crimes are most likely to occur.
PredPol’s technology has been helping law enforcement agencies to dramatically reduce crime in jurisdictions of all types and sizes, across the U.S. and overseas. Over the past year, Atlanta and Los Angeles have reduced specific crimes in targeted areas at rates ranging from nearly 20% to over 40%. Smaller jurisdictions, such as Norcross, Georgia, have seen nearly a 30% reduction in burglaries and robberies; in Alhambra, California, car burglaries have dropped 20% since the software technology was deployed.
Using advanced mathematics and computer learning, PredPol’s algorithms predict many types of crime, including property crimes, drug incidents, gang activity, and gun violence as well as traffic accidents.
Only three pieces of data are used to make predictions – type of crime, place of crime, and time of crime. No personal data is utilized in making these predictions.
Crime analysts and command staff using PredPol are 100% more effective than they are with traditional hotspot mapping at predicting where and when crimes are likely to occur. That means police have twice as many opportunities to deter and reduce crime.
Adoption of the R language has grown rapidly in the last few years, and is ranked as the number-one data science language in several surveys. This accelerating R adoption curve has been driven by the Big Data revolution, and the fact that so many data scientists — having learned R at university — are actively unlocking the secrets hidden in these new, vast data troves. In more than 6 years of writing for the Revolutions blog, I’ve discovered hundreds of applications of R in business, in government, and in the non-profit sector. Sometimes the use of R is obvious, and sometimes it takes a little bit of detective work to learn how R is operating behind the scenes. In this talk, I'll recount some of my favourite applications of R, and show how R is behind some amazing innovations in today’s world.
[Presented to the 7th China R Users Conference, Beijing, May 2014.]
Adoption of the R language has grown rapidly in the last few years, and is ranked as the number-one data science language in several surveys. This accelerating R adoption curve has been driven by the Big Data revolution, and the fact that so many data scientists — having learned R at university — are actively unlocking the secrets hidden in these new, vast data troves.
In more than 6 years of writing for the Revolutions blog, I’ve discovered hundreds of applications of R in business, in government, and in the non-profit sector. Sometimes the use of R is obvious, and sometimes it takes a little bit of detective work to learn how R is operating behind the scenes. In this talk, I’ll begin by presenting some recent statistics on the growth of R. Then I’ll recount some of my favourite applications of R, and show how R is behind some amazing innovations in today’s world.
Order Fulfillment Forecasting at John Deere: How R Facilitates Creativity and...Revolution Analytics
Statistical analysis has been known to be invaluable to any manufactory’s quality assurance for decades. Recently the value of valid statistical analysis has also been demonstrated to radically improve the ability of a company’s ability to weather extreme peaks and valley in customer demand. John Deere has been able to adjust to commodity spikes and housing downturns much better than its competitors have. This is in part due to the implementation of statistical analysis and the use of R software in the order fulfillment function of John Deere.
Using Data Mining Techniques to Analyze Crime PatternZakaria Zubi
Our proposed model will be able to extract crime patterns by using association rule mining and clustering to classify crime records on the basis of the values of crime attributes.
Video transmission over wireless networks is considered the most interesting application in our daily life nowadays. As
mobile data rates continue to increase and more people rely on wireless transmission, the amount of video transmitted over at least one
wireless hop will likely continue to increase. This kind of application needs large bandwidth, efficient routing protocols, and content
delivery methods to provide smooth video playback to the receivers. Current generation wireless networks are likely to operate on
internet technology combined with various access technologies. Achieving effective bandwidth aggregation in wireless environments
raises several challenges related to deployment, link heterogeneity, Network congestion, network fluctuation, and energy consumption.
In this work, an overview of technical challenges of over wireless networks is presented. A survey of wireless networks in recent video
transmission schemes is introduced. Demonstration results of few scenarios are showed.
“Exploratory spatial analysis of illegal oil discharges detected off Canada’s Pacific Coast” Third International Workshop on "Geographical Analysis, Urban Modeling, Spatial Statistics"
The visibility estimation has an important impact in many economical and aesthetic fields, a mixed environment which contains madman objects like buildings with relief sol make a challenge for the visibility calculation. This paper presents a new method to solve this problem based on vector GIS data. The use of vector data gives the possibility to calculate the intervisibility, viewshed for mixed environment. The new method could identify the obstacles (relief, buildings identification) which block the visibility for a 3D environment points from observator, the intervisibility impact of a specific building could be calculated
An overview of traditional spatial analysis tools, an intro to hadoop and other tools for analyzing terabytes or more of data, and then a primer with examples on combining the two with data pulled from the Twitter streaming API. Given at the O'Reilly Where 2.0 conference in March 2010.
Big Social Data: The Spatial Turn in Big Data (Video available soon on YouTube)Rich Heimann
Big Social Data: The Spatial Turn in Big Data
By Richard Heimann & Abe Usher
University of Maryland Baltimore County Webinar Description:
The increased access to spatial data and overall improved application of spatial analytical methods present certain potential to social scientific research. This webinar is designed to focus on substantive social science research perspectives while exposing rewards involved in the application of geographic information systems (GIS), Big Data, and spatial analytics in their own research.
What is witnessed as the hype of Web 2.0 has worn off and the collaborative use of the Internet becomes a societal norm is an unprecedented explosion in the creation and analysis of geospatial data. Just as major governments are reducing their investments in location intelligence, individuals and non-government organizations are fueling a bonfire of innovation in the world of GIS data.
Traditional spatial analyses grew up in an era of sparse data and very weak computational power. Today, both of those circumstances are reversed and many of the old solutions are no longer suitable to answer todays questions.
"Big Social Data: The Spatial Turn in Big Data" reflects this change and combines two things which, until recently, engaged quite different groups of researchers and practitioners. Together, they require particular techniques and a sophisticated understanding of the special problems associated with spatial social data. Geographic Data Mining, or Geographic Knowledge Discovery, is not new, but is developing and changing rapidly as both more, and different, data becomes available, and people see new applications. The days of ‘Big Data’ require fresh thinking.
The webinar will highlight connections between spatial concepts and data availability. New emerging social media data will be promoted over traditional social science data, which better reflect some of the more recently developments in Big Data - most notably the socially critical exploration of such data.
In this study various techniques for exploratory spatial data analysis are reviewed : spatial autocorrelation, Moran's I statistic, hot spots analysis, spatial lag and spatial error models.
Vector Analysis at Undergraduate in Science (Math, Physics, Engineering) level. The presentation gives a general description of the subject.
Please send comments and suggestions to solo.hermelin@gmail.com, thanks. For more presentations, please visit my website at
http://www.solohermelin.com .
Forecasting Space-Time Events - Strata + Hadoop World 2015 San JoseAzavea
This presentation uses the speaker’s experience in building a crime forecasting package to outline some tools and techniques useful in modeling space-time event data. While the case study focuses on modeling crime, the techniques and tools presented are applicable to a broad selection of domains.
This presentation was given at Strata + Hadoop World 2015 in San Jose by Jeremy Heffner.
Cities are composed of complex systems with physical, cyber, and social components. Current works on extracting and understanding city events mainly rely on technology enabled infrastructure to observe and record events. In this work, we propose an approach to leverage citizen observations of various city systems and services such as traffic, public transport, water supply, weather, sewage, and public safety as a source of city events. We investigate the feasibility of using such textual streams for extracting city events from annotated text. We formalize the problem of annotating social streams such as microblogs as a sequence labeling problem. We present a novel training data creation process for training sequence labeling models. Our automatic training data creation process utilizes instance level domain knowledge (e.g., locations in a city, possible event terms). We compare this automated annotation process to a state-of-the-art tool that needs manually created training data and show that it has comparable performance in annotation tasks. An aggregation algorithm is then presented for event extraction from annotated text. We carry out a comprehensive evaluation of the event annotation and event extraction on a real-world dataset consisting of event reports and tweets collected over four months from San Francisco Bay Area. The evaluation results are promising and provide insights into the utility of social stream for extracting city events.
High Availability HPC ~ Microservice Architectures for Supercomputinginside-BigData.com
In this deck from the Stanford HPC Conference, Ryan Quick from Providentia Worldwide presents: High Availability HPC ~ Microservice Architectures for Supercomputing.
"Microservices power cloud-native applications to scale thousands of times larger than single deployments. We introduce the notion of microservices for traditional HPC workloads. We will describe microservices generally, highlighting some of the more popular and large-scale applications. Then we examine similarities between large-scale cloud configurations and HPC environments. Finally we propose a microservice application for solving a traditional HPC problem, illustrating improved time-to-market and workload resiliency."
Watch the video: https://insidehpc.com/2018/02/high-availability-hpc-microservice-architectures-supercomputing/
Learn more: http://www.providentiaworldwide.com/
and
http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
(BDT207) Real-Time Analytics In Service Of Self-Healing EcosystemsAmazon Web Services
Netflix strives to provide an amazing experience to each member. To accomplish this, Netflix needs to maintain very high availability across our systems. However, at a certain scale, humans can no longer scale their ability to monitor the status of all systems, making it critical for Netflix to build tools and platforms that can automatically monitor their production environments and make intelligent real-time operational decisions to remedy the problems they identify. In this session, we discuss how Netflix uses data mining and machine learning techniques to automate decisions in real-time with the goal of supporting operational availability, reliability, and consistency. We review how we got to the current states, the lessons we learned, and the future of real-time analytics at Netflix. While Netflix's scale is larger than most other companies, we believe the approaches and technologies we discuss are highly relevant to other production environments, and audience members should come away with actionable ideas that are implementable in, and benefit, most other environments.
Oplægget blev holdt ved InfinIT-arrangementet Big Data og data-intensive systemer i Danmark, der blev af holdt en 15. januar 2014. Læs mere om arrangementet her: http://infinit.dk/dk/arrangementer/tidligere_arrangementer/big_data_i_danmark.htm
Using New Tools to Analyze and Plan Your Urban Forest Azavea
Planting locations are often determined by organization goals, available funding, practical logistics that influence the number of trees you can plant and where you can plant them, and dozens of other factors. With the new toolkit from OpenTreeMap you can use existing sociodemographic and land-use data to make more informed planting decisions, and estimate the future environmental and economic benefits of your trees.
7 misconceptions about predictive policing webinarAzavea
Over the last few years, Predictive Policing has become more common in police departments around the world. With the rising interest in crime forecasting tools, important questions concerning ethics, privacy and fairness have been raised. We know that there are some misconceptions when it comes to the topic, and we want to dispel some of the common myths about Predictive Policing.
We invite you to join us as we walk through 7 Misconceptions of Predictive Policing. In this webinar, we aim to discuss some of the charged rhetoric and beliefs that surround the term. Also, we will highlight the some of the diverse crime modeling concepts that are used to make robust, predictions when forecasting crime.
OpenTreeMap is a platform that enables individuals and organizations to map and inventory their urban forest. This webinar provides an overview of OpenTreeMap's Green Infrastructure module and was given by Azavea on November 11, 2015. For more information on OpenTreeMap visit www.opentreemap.org or email us at opentreemap@azavea.com.
Growing Your Urban Forest: Using the OpenTreeMap Bulk UploaderAzavea
The "Growing Your Urban Forest: Using the OpenTreeMap Bulk Uploader" webinar was held on April 16, 2015. These slides provide an overview of that webinar.
November 12, 2014 Webinar: Hackers, Beer Geeks, and Arborly Love - Reaching o...Azavea
In this webinar based on our 2014 Partners in Community Forestry conference presentation, Andrew Thompson (OpenTreeMap), Erica Smith Fichman (TreePhilly), and Lee Mueller (Friends of Grand Rapids Parks) talked about three outreach events our organizations have done in urban forestry, and discussed tips and tricks your urban forestry group can use with your events and marketing to expand to new audiences. This webinar covered:
- A general framework for organizing events and campaigns geared toward exciting audiences and communities with little experience with urban forestry
- Pointers, tips, caveats, and potential downfalls to keep in mind to organize a successful event
- "Lessons learned" from three specific case studies organized by a government, nonprofit, and commercial company
This presentation covers the requirements to get started with HunchLab 2.0's predictive policing system. It starts discussing technical requirements (security, authentication) and then proceeds to discuss guidelines for configuring meaningful predictive models of crime. The presentation concludes with information about related geographic and temporal datasets that are useful in forecasting crime with recommendations on how to prioritize data sets to use in HunchLab.
Is it a Package or a Wrapper? Designing, Documenting, and Distributing a Pyth...Azavea
Andrew Thompson delivered this talk at the January 2014 joint meeting of the PhillyPUG Python User's Group and the GeoPhilly GIS Meetup group. Topics covered include Rest APIs, API wrappers, Python documentation tools, and Python module packaging practices and the Python Package Index.
Using Open Data and Citizen Science to Promote Citizen Engagement with Green ...Azavea
Presentation given at the Green Infrastructure and Water Management in Growing Metropolitan Areas conference on January 15, 2014 at the Patel College of Global Sustainability at the University of South Florida, Tampa, Florida.
This document presents a preview of the new version of HunchLab with a focus on geographic modeling. HunchLab 2.0 allows police departments to prioritize resource deployments by using predictive analytics that take into account many data sets and reflect the priorities of the police department. The webinar was recorded on September 25th, 2013.
Five Technology Trends Every Nonprofit Needs to KnowAzavea
Are you tired of hearing about big data, social media, web 2.0, and other buzzwords? This session will introduce five emerging technology trends that will fundamentally impact the independent sector. Join us and learn how to incorporate them into your current plans to better reach your donors, engage your constituents, and maximize your impact.
PhillyHistory.org - Tracking Metrics for a Digital ProjectAzavea
Presentation given at the Delaware Valley Archivists Group meeting on March 21, 2013. The slides provide an overview of how visitor statistics and user engagement are measured on PhillyHistory.org and how similar tracking may be done on other digital history projects.
Modeling Count-based Raster Data with ArcGIS and RAzavea
This presentation outlines the conceptual framework for building regression models of event counts where the unit of analysis is small. It explains how ArcGIS for Desktop can be used to build raster data sets that are modeled as generalized linear models within the open source R package.
This overview of OpenTreeMap, software for collaborative, geography enabled urban tree inventory, was given as part of the Alliance for Community Trees webcast training on January 17, 2013 - Tree Technology - Part V: Urban Forest Mapping. For more information, visit www.opentreemap.org.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
3. B Corporation
• Projects w/ Social Value
• Summer of Maps
• Pro Bono Program
• Donate share of profits
Research-Driven
• 10% Research Program
• Academic Collaborations
• Open Source
5. How Phila PD uses Maps
Customized Map Products
Weekly CompStat Meetings
Web Crime Analysis
6. INCT & PARS – main database sources
over 5,000 incidents daily, over 2 million annually
PARS
Complainant INCT
Verizon Daily download
911 District & Geocoding Routines
48 Desk
Incident Report
Completed by Officer District X
911 Operator
Police Officer Maps distributed
Through Intranet, District Y
Printing, CompStat
Radio
Dispatcher
CAD District Z
17. Crime Analysis – What has happened?
– Mapping (spatial / temporal densities)
– Trending
– Intelligence Dashboard
Early Warning – What is out of the ordinary?
– Statistical & Threshold-based Hunches (data mining)
– Alerting
Risk Forecasting – What is likely to happen next?
– Near Repeat Pattern
– Load Forecasting
23. Early Warning
• Geographic Early Warning System
– A system to alert staff of an unusual situation in a particular
location
– Ingests data sets to automatically “cook on” and only
involves staff when a statistically unusual situation is found
Geostatistical Engine
Operational
Operational
Database
Alerting
Operational
Database HunchLab
Database System
Databases
35. Contagious Crime?
• Near repeat pattern analysis
• “If one burglary occurs, how does the risk change nearby?”
36. What Do We Mean By Near Repeat?
• Repeat victimization
– Incident at the same location at a later time (likely related)
• Near repeat victimization
– Incident at a nearby location at a later time (likely related)
• Incident A (place, time) --> Incident B (place, time)
37. Near Repeat Pattern Analysis
• The goal:
– Quantify short term risk due to near-repeat victimization
• “If one burglary occurs, how does the risk of burglary for the
neighbors change?”
• What we know:
– Incident A (place, time) --> Incident B (place, time)
• Distance between A and B
• Timeframe between A and B
• What we need to know:
– What distances/timeframes are not simply random?
38. Near Repeat Pattern Analysis
• The process
– Observe the pattern in historic data
– Simulate the pattern in randomized historic data
– Compare the observed pattern to the simulated patterns
– Apply the non-random pattern to new incidents
• An example
– 180 days of burglaries in Division 6 of Philadelphia
43. Near Repeat Pattern Analysis
• How can you test your own data?
– Near Repeat Calculator
• http://www.temple.edu/cj/misc/nr/
• Papers
– Near-Repeat Patterns in Philadelphia Shootings (2008)
• One city block & two weeks after one shooting
– 33% increase in likelihood of a second event
Jerry Ratcliffe
Temple University
46. Improving CompStat
• Workload forecasting
• “Given the time of year, day of week, time of day and
general trend, what counts of crimes should I expect?”
47. What Do We Mean By Load Forecasting?
• Workload forecasting
• Generating aggregate crime counts for a future timeframe
using cyclical time series analysis
Measure cyclical patterns
+
Identify non-cyclical trend
Forecast expected count
bit.ly/gorrcrimeforecastingpaper
48. Load Forecasting
• Measure cyclical patterns
• Take historic incidents (for example: last five years)
• Generate multiplicative seasonal indices
– For each time cycle:
» time of year
» day of week
» time of day
– Count incidents within each time unit (for example: Monday)
– Calculate average per time unit if incidents were evenly
distributed
– Divide counts within each time unit by the calculated average to
generate multiplicative indices
» Index ~ 1 means at the average
» Index > 1 means above average
» Index < 1 means below average
53. Load Forecasting
• Identify non-cyclical trend
• Take recent daily counts (for example: last year daily counts)
• Remove cyclical trends by dividing by indices
• Run a trending function on the new counts
– Simple average
» Last X Days
– Smoothing function
» Exponential smoothing
» Holt’s linear exponential smoothing
54. Load Forecasting
• Forecast expected count
• Project trend into future timeframe
– Always flat
» Simple average
» Exponential smoothing
– Linear trend
» Holt’s linear exponential smoothing
• Multiple by seasonal indices to reseasonalize the data
57. How Do We Know It’s Accurate?
• Testing
• Generated forecasting techniques(examples)
– Commonly Used
» Average of last 30 days
» Average of last 365 days
» Last year’s count for the same time period
– Advanced Combinations
» Different cyclical indices (example: day of year vs. month of year)
» Different levels of geographic aggregation for indices
» Different trending functions
• Scoring methodologies (examples)
– Mean absolute percent error (with some enhancements)
– Mean percent error
– Mean squared error
• Run thousands of forecasts through testing framework
• Choose the right technique in the right situation
59. Research Topics
• Risk Forecasting
– Load forecasting enhancements
• Weather and special events
– Combining short and long term risk forecasts (Temple)
• Socioeconomic changes in neighborhoods
– Risk Terrain Modeling (Rutgers)
• Context of crime at the microplace
64. Robert’s Rules of Housing
Close to Center City somewhat important
Walk to Grocery Store vital
Nearby Restaurants very important
Library nice to have
Near a Park somewhat important
Biking / walking distance from our work very important
Biking distance to fencing somewhat important
65. Your factors might include…
Child Care
Local School Rankings
Farmer's Market
Car Share
Public Transit
73. Web is different from the Desktop
Lots of simultaneous users
Stateless environment
HTML+JS+CSS
Users are less skilled
Users are less patient
74. But wait … there’s a problem
10 – 60 second calculation time
Multiple simultaneous users …
… that are impatient
86. Optimizing one process sub-optimizes others
Complex to configure and maintain
Limited to one operation
No interpolation
No mixing
– cell sizes
– extents
– projections
etc.
87.
88. Broader set of functionality
Both raster and vector
Scala + Akka
Open source