The Federal Government has vast open data resources. This talk will present a few APIs: One from the Department of Labor that serves up the data on the goods and products made with forced and child labor, one from the US Census Bureau, and another from the Department of Commerce and tangles Income Inequality.
An intro into AI and how business leaders should use itLutz Finger
FOUR RULES: (1) AI is a tool not a business model. (2) Protect your data but use federated learning to share models. (3) Regulation should guide you & not stop you - use tools like LIME. (4) Think holistically and build a data culture of fair data usage.
Vowpal Wabbit is both an open-source machine learning toolkit and an active research platform. In this talk I introduce Vowpal Wabbit, discuss some of the design decisions, and the types of problems for which VW is (or is not) a good fit. The talk includes (live) demonstrations some of the latest features for recommendation, contextual bandit, and structured prediction problems.
Congratulations you privacy definitions have been successfully updatedAlberto Dávila
A little of history and update on how information technology and web content contributed to social media and big data, and how privacy settings has changed over the last decade.
An intro into AI and how business leaders should use itLutz Finger
FOUR RULES: (1) AI is a tool not a business model. (2) Protect your data but use federated learning to share models. (3) Regulation should guide you & not stop you - use tools like LIME. (4) Think holistically and build a data culture of fair data usage.
Vowpal Wabbit is both an open-source machine learning toolkit and an active research platform. In this talk I introduce Vowpal Wabbit, discuss some of the design decisions, and the types of problems for which VW is (or is not) a good fit. The talk includes (live) demonstrations some of the latest features for recommendation, contextual bandit, and structured prediction problems.
Congratulations you privacy definitions have been successfully updatedAlberto Dávila
A little of history and update on how information technology and web content contributed to social media and big data, and how privacy settings has changed over the last decade.
Нам кажется, что технологии меняют то, как мы учимся. Но такое заключение базируется лишь на наших ощущениях. Удается ли нам достичь лучших образовательных результатов за счет новых технологий? Конечно, можно попробовать разные новинки и протестировать результат. Но наш подход обратный - найти исследования, показывающие возможности для улучшения образовательного результата и применить технологии именно там, где они будут наиболее эффективны.
Coordinator: Ca‘ Foscari University of Venice, Department of Asian and North African Studies, Italy
Project Partners:
École Pratique des Hautes Études, Institut européen en sciences des religions,France
University of Salamanca, Research GRoup in InterAction and eLearning, Spain
University of Augsburg, Department for History Didactics, Germany
University of Southern Denmark, The Study of Religion, Denmark
OFXAM Italia Intercultura, Italy
Information literacy instruction session delivered to first year students. Topics covered included APA ciation, Google syntax and search strategies, library catalogue searching, and database interface features and functionality.
Find company reports in EBSCO Business Source CompleteJulie Anne Kent
Created for an Organizational / Human Resource course (BBUS) at Thompson River's University, this guide is embedded into the research guide in support of student learning outcomes. http://libguides.tru.ca/BBUS3810.
There are three things that matter in relation to a networked specific practice and media production. These three terms apply to the formal attributes of digital media, and so address the qualities that practice requires, and how we participate, use, and engage with networked media. There is no hierarchy amongst these three terms, and they may prove to be insufficient. The terms are porousness, granularity, and facets. The list does not include database, user, or interactivity, as these are not causes but consequences of this triumvirate of terms.
Government as a Platform: What We've Learned Since 2008 (pdf with notes)Tim O'Reilly
My talk at the UK Government Digital Service Sprint 15 event in London, February 2, 2015. I talk about my idea of government as a platform, and what I've learned since I first articulated the idea, with specific reference to what the GDS has taught me about the idea.
Нам кажется, что технологии меняют то, как мы учимся. Но такое заключение базируется лишь на наших ощущениях. Удается ли нам достичь лучших образовательных результатов за счет новых технологий? Конечно, можно попробовать разные новинки и протестировать результат. Но наш подход обратный - найти исследования, показывающие возможности для улучшения образовательного результата и применить технологии именно там, где они будут наиболее эффективны.
Coordinator: Ca‘ Foscari University of Venice, Department of Asian and North African Studies, Italy
Project Partners:
École Pratique des Hautes Études, Institut européen en sciences des religions,France
University of Salamanca, Research GRoup in InterAction and eLearning, Spain
University of Augsburg, Department for History Didactics, Germany
University of Southern Denmark, The Study of Religion, Denmark
OFXAM Italia Intercultura, Italy
Information literacy instruction session delivered to first year students. Topics covered included APA ciation, Google syntax and search strategies, library catalogue searching, and database interface features and functionality.
Find company reports in EBSCO Business Source CompleteJulie Anne Kent
Created for an Organizational / Human Resource course (BBUS) at Thompson River's University, this guide is embedded into the research guide in support of student learning outcomes. http://libguides.tru.ca/BBUS3810.
There are three things that matter in relation to a networked specific practice and media production. These three terms apply to the formal attributes of digital media, and so address the qualities that practice requires, and how we participate, use, and engage with networked media. There is no hierarchy amongst these three terms, and they may prove to be insufficient. The terms are porousness, granularity, and facets. The list does not include database, user, or interactivity, as these are not causes but consequences of this triumvirate of terms.
Government as a Platform: What We've Learned Since 2008 (pdf with notes)Tim O'Reilly
My talk at the UK Government Digital Service Sprint 15 event in London, February 2, 2015. I talk about my idea of government as a platform, and what I've learned since I first articulated the idea, with specific reference to what the GDS has taught me about the idea.
Platform Shift: How New Business Models Are Changing the Shape of IndustryMarshall Van Alstyne
Companies that can transform their traditional business models into network models will have a competitive advantage based on new insights into pricing, network effects, supply chains, and strategy. These principles show how dotcom companies like Airbnb, Amazon, Apple and Uber managed, in a relatively short time, to attract millions of clients worldwide. But they apply also to traditional product companies like Sony, shoe companies like Nike, and spice companies like McCormick. New business models help these companies extend existing transactions to new, associated products and services. Platforms beat products every time. This talk reveals the secret of Internet-driven platforms, why they happen, and what changes they imply.
GIS and Asset Management Moving to the Future : Symphony3
A 40 year roadmap for asset management
A presentation by Phillip Dooley of Symphony3. The presention was delivered at the Spatial Technology Summit, Melbourne, Brisbane and Sydney, May 2011.
The topics covers asset management and GIS in industries such as Local and State Government, Water Utilities and Private Industry. The implications of social media on Asset management are also addressed.
2. Economic Impact and Societal Considerations for Policy Decisions.Saurabh Mishra
This group reviewed resource allocation questions related to public and private investment including challenges for skill-reallocation, economic loss, including job loss, qualified labor force reduction, distributional challenges, income inequality, and opportunities for economic diversification.
Short talk on The Guardian and open/public data given by Chris Thorpe at the Gov2.0 Expo in Washington on the "Four perspectives of data.gov.uk" panel with Sir Tim Berners-Lee, John Sheridan and Dominic Campbell.
Open Data for Financial Innovations in the Developing WorldBiplav Srivastava
Financial innovations are key to getting developing countries like India their rightful role in the global economy. However, such innovations depend on data, advanced analytics and timely access to insights. In this talk, we will discuss open data collected by governments and people and how it can bootstrap innovations that matter. Open data helps not only overcome initial data bottleneck but also helps standardize solutions for scalability and wider adoption.
Fundamentals of Big Data in 2 minutes!!Simplify360
In today’s world where information is increasing every second, BIG DATA takes up a major role in transforming any business.
Learn the fundamentals of big data in just 2 minutes!
2014 Tech M&A Monthly - Mid-Year ReportCorum Group
This month, join us as we mark the halfway-point of this remarkable year for tech M&A, and look both back and ahead. We’ll look back at the notable deals, high valuations and key trends in all six market sand 26 subsectors that have made 2014 such a wild ride. Then, we’ll look ahead at the factors that could bring the ride to a sudden halt—either for the market as a whole, or for your company’s value in particular. Nothing lasts forever, so we’ll be examining the six events that could kill the market, as well as six events that, regardless of the overall environment, could kill your company’s value before you have a chance to realize that value.
Plus, special announcements of deals out of the UK, the US heartland and beyond, plus a look at the way Enmeshed Systems and the blurring lines between hardware and software are driving key deals and high valuations. Don’t miss this extended 45-minute presentation
The attention economy and the internetRoss Garrett
Today we’re going to take a look at how traditional or even de facto standards for integration aren’t always the best choice for web mobile and IoT applications. The standard I’m talking about is of course HTTP and ever popular REST APIs.
While I won’t be so bold as to disregard this integration pattern entirely, I do want us to take a critical look at how and where integration can be improved – by understanding the limitations of today’s app integration technologies and considering the business factors that impact success in the attention economy.
To view the full webinar visit: http://www.elasticpath.com/resources/webinar/commerce-content-perfected
The most innovative and profitable digital experiences require almost perfect choreography between your ecommerce and content management systems. In this webinar, we explain how you can achieve a perfect integration – and illustrate it with an insider’s look at how Elastic Path tackles the challenge with leading WCM solutions including Adobe® Experience Manager and OpenText® CEM.
Global Scientific Research as a Tool to Unlock and Engage Talent and Expand t...Tyrone Grandison
‘Science for social justice’ may only be achieved when politicians, decision-makers and science-policymakers set a considered and thoughtful agenda to utilize science, in reasoned and innovative ways, as a driving force for positive societal change to promote equity through innovation. However, to date, tangible results in many contexts have been mixed at best, especially in delivering a reliable mechanism for, or a path to, sustainable social equity and justice for all. As global inequality increases and much political decision-making remains myopic and contingent, the emotive and essential power of ‘science for social justice’ can be lost as scientists and decision-makers struggle to actualize meaningful change. We, as scientists, in collaboration with our decision-making peers, have a golden opportunity to correct this through clear and novel proposals for meaningful projects based on advanced research opportunities. In this regard, we contend that ‘science for social justice’ can only be fully realized if it is symbiotically connected to providing scientific opportunity, where no such opportunity previously existed. This inevitably foments and sustains prosperity, an essential factor for social justice to grow. Therefore, the goal must be to establish opportunity that serves as the bridge to prosperity. How can we accomplish this when most of the world relies on relatively few countries for new scientific advances and technologies?
During the COVID-19 Global Pandemic, there were multiple lessons provided to the world. In this talk, I set the stage for the discussion, highlight the issues we faced (and still face), I speak to an effort that contributed to help address one of those issues, then speak to future challenges and our responsibilities going forward.
Systemic Barriers in Technology: Striving for Equity and AccessTyrone Grandison
Technology is an integral part of our everyday lives through broad-band internet usage, protection of cyber-security security, or the usage of artificial intelligence (AI) to mimic human-operations. Historically, technology has perpetuated racial discrimination with biases in algorthims used in the health-care system, facial recognition in the criminal justice system, to Black and Latinx students lacking access to technological resources. This panel will discuss the historical context of racism in technology, current technology access issues in communities of color, as well as strategies and policies that dismantle systemic racism in technology.
Are There Ethical Limits to What Science Can Achieve or Should Pursue?Tyrone Grandison
As a computer scientist and data scientist, this is a nuanced question that requires prinicpled treatment. There are multiple factors that determine the answer to this prompt. The path that I take is going back to first principles and presenting a framework for evaluation.
The current model of invention needs to be augmented. We have to include user feedback more integrally and we need to invent to reduce unintended consequences.
We live in an amazing time. The only barrier to impact is execution. Every individual has the opportunity to take an idea from inception to invaluable and innovative solution in a matter of months. Every nation has the capacity, and the capability, to create a solid foundation for its citizens that has the potential to transform lives and sustain a thriving innovation ecosystem. This talk will examine the part that each of us must play in creating an innovation nation.
The mission of the IHME is to apply rigorous measurement and analysis to help policy makers make better decisions on a range of health policy issues. Like other organizations, the IHME have embraced containers and micro-services aggressively to better support hundreds of collaborating researchers.
In addition to containerized workloads, the IHME run a wide-variety of traditional analytic, simulation and high-performance computing workloads on an HPC cluster with 15,000 cores and 13PB of storage. Researchers increasingly need to combine both containerized and non-containerized elements into workflow pipelines, and a key challenge has been ensuring SLAs for various departments and avoiding duplicate infrastructure and unnecessary data movement and duplication. In collaboration with industry partners, IHME have deployed a unique solution based on Univa’s Navops technology that allows them to combine containerized and traditional analytic and high-performance application workloads on a single shared Kubernetes cluster, ensuring departmental SLAs and helping contain infrastructure costs.
In this talk Dr. Grandison will discuss IHME, their experience deploying containerized applications and how they went about using Kubernetes to support a variety of new containerized applications as well as a variety of traditional analytic applications.
It often goes unnoticed that the majority of innovations today stems from investments by government bodies to produce platforms, software and data for the greater societal good. The Internet, the Global Positioning System, voice-controlled software are all examples of these investments. The private industry has no business case for undertaking these efforts; as the business model and return on investment is often unknown. These well-known examples started as military projects in search of ethical commercial use cases. Private industry is often the biggest benefactors of the production of these systems. In this talk, I will speak about the cycle of open innovation, highlight a few examples, discuss what went and is wrong, and highlight course corrections. Specifically, the focus will be initiatives that were intentionally meant to be open , like weather data from NOAA, survey data from the Census Bureau, GDP (Gross Domestic Product) data from the Bureau of Economic Analysis, and public health data from the Institute of Health Metrics and Evaluation.
The Creative Economy within the United States of AmericaTyrone Grandison
The creative economy spans multiple industries and spaces. I show highlights of how the US Department of Commerce supports the American Creative Economy
The U.S. Department of Commerce collects, processes and disseminates data on a range of issues that impact our nation. Having a host of data and ensuring that this data is open and accessible to all are two separate issues. This session will cover the Commerce Data Usability Project (CDUP) - a community-driven public-private partnership to help data scientists, programmers and other users to access open knowledge from our open data.
Creating a Data-Driven Government: Big Data With PurposeTyrone Grandison
The U.S. Department of Commerce collects, processes and disseminates data on a range of issues that impact our nation. Whether it's data on the economy, the environment, or technology, data is critical in fulfilling the Department's mission of creating the conditions for economic growth and opportunity. It is this data that provides insight, drives innovation, and transforms our lives. The U.S. Department of Commerce has become known as "America's Data Agency" due to the tens of thousands of datasets including satellite imagery, material standards and demographic surveys.
But having a host of data and ensuring that this data is open and accessible to all are two separate issues. The latter, expanding open data access, is now a key pillar of the Commerce Department's mission. It was this focus on enhancing open data that led to the creation of the Commerce Data Service (CDS).
The mission at the Commerce Data Service is to enable more people to use big data from across the department in innovative ways and across multiple fields. In this talk, I will explore how we are using big data to create a data-driven government.
This talk is a keynote given at the Texas tech University's Big Data Symposium.
General tips to students at the Biomedical Data Science Mentoring Workshop at the Symposium of Health Informatics in Latin America and the Caribbean (SHILAC) 2015.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Neuro-symbolic is not enough, we need neuro-*semantic*
Building APIs in Government for Social Good
1. Building APIs in Government
for Social Good
Tyrone Grandison PhD
www.tyronegrandison.org @tyrgr
2. 2
Deputy Chief Data Officer,
US Department of Commerce (2015-16)
White House Presidential Innovation Fellow,
Department of Labor & US Census Bureau (2014-15)
My Time In Government
38. Sweat and Toil
•Monthly Data Users >
Web Traffic
•Three tools built using
this data.
38
So Far
CENSUS CitySDK
•Over 10 civic solutions
built using CitySDK
•Positive User Feedback
MIDAAS: Hack The Pay Gap Initiative
39. Sweat and Toil
•2016 Department of
Labor‘s Innovation Award.
MIDAAS
•2016 Nominee Fedscoop
Innovation Of The Year.
39
So Far
CENSUS CitySDK
•2016 Department of
Commerce Gold medal
•2016 Best Data API
Award, API:World
•2015 Fedscoop Innovation
Of The Year
By show of hands,
How many people here have worked in Government?
How many people think that there is a technical difference between delivering APIs in Government vs the Private Sector?
My mission here today to shed some light on the process.
If there is anything that you need to remember from this talk, it is that there are more similarities than differences.
I have spent the last two and a half years of my life working in government.
Before government, I worked in the private sector, academia and startup world as a developer (Python, Java, C, C++), manager, CTO, Consultant and Founder, which gives me a unique perspective.
This is the plan for today.
I am going to start with the reality of government, delve in three APIs that my teams and I have built, and finish with the lessons that we have learned.
While the private sector takes on elastic problems, like entertainment and marketing.
Government takes on hard problems like homelessness, health, safety, defense, trade, and social justice.
The problems are difficult optimization problems
And not just in one dimension.
But rather in multiple dimensions
And many of the dimensions are not anywhere technology or science related.
*** Laws, Systems, Personalities/Egos.
Developing APIs (or any tech) in Government is an issue in Organizational Change Management.
Where you have align Process, Policy and People in order to enable Technology development.
It is daunting, slow and normally takes a lot of effort.
However, there is hope on the horizon.
On May 9th 2013, the President created an Executive Order that stated that data produced by the US government should be machine-readable and open by default.
Executive Order -- Making Open and Machine Readable the New Default for Government Information
Data.gov, which was launched in May 2009, was used as the vehicle or mechanism to fuel compliance to M-13-13 and offer a wealth of new data sets and APIs to the community.
Additionally, the Administration has just released the initial guidelines on Federal Source Code.
You throw into the mix, an amazing set of agile and user-centered startup-like organizations within government working on accelerating the development of data products and services, and you see why there has been a steady stream of Social Good APIs coming from the Federal government over the last few years.
Now, look at three APIs, focused on solving social problems, that my teams and I have developed and deployed within the last 15 months.
The first is Sweat and Toil. The API details are in the top URL. The API fuels both Android and iPhone apps with the same name. The other two URLS contain the code. If you have feedback and or want to collaborate, please let us know.
The data itself is the information produced by the Bureau of International Labor Affairs (ILAB) on the countries that produce goods and products that use child and or forced labor.
ILAB’s mission is to use all available international channels to improve working conditions, raise living standards, protect workers' ability to exercise their rights, and address the workplace exploitation of children and other vulnerable populations.
Every year, ILAB has a congressional mandate to produce three reports on international child labor and forced labor.
Typically, the ILAB team spends nine months collecting information on all the countries in the world (except the US) to create over 1,000 pages of information that they then hand out in thick books; offering a PDF version on their website for increased accessibility.
Our idea was simple – liberate the data, make it available to the entire dev community and see what new and interesting things are done.
The first step in making any API stick within the Department was getting the buy-in from the Department’s leadership.
Carol Pier – the head of ILAB and Chris Lu and Tom Perez – the leaders of the Department of Labor.
Without their support, getting the necessary assistance from the different units within the Department would have been impossible and we would have spent years in discussion on a plan of action.
The second step was sitting with ILAB Program team (Tina Faulkner, Charita Castro, Chandra Ulca) and going through the process they use to create their thousand page deliverable.
In summary, there is a team of researchers that are tasked with manually gathering information from a set of countries in a given region. Each researcher records their findings in a single Word document per country. This document is passed around to peers and supervisors for vetting and editing until a final version is arrived at.
Once all the country profiles are complete, the Word documents are sent to a contractor to be converted to PDF documents. This PDF documents are merged and sent to a printing company that produces the books. The PDF documents are also used to populate the website content and are placed on the website itself.
Given the current process, the first point when the data is in a stable state is when the Word docs are sent to the contractor for PDF conversion.
So, this is the point where we should start the API creation process.
However, we just couldn’t use any technology that we wanted to build the API.
The Department of Labor’s Office of Public Affairs holds the responsibility for maintaining the Department’s APIs.
At the time, they have a v1 of their API and had embarked on a more full-featured v2 – which would handle API management for internal devs.
In order to ensure longevity, we had to use their platform, which is called Quarry – PHP, CodeIgniter – and currently open source.
The ILAB team worked with us to identify the user personas for their data – politicians, internal ILAB staff & executives, government officials, and the general public.
For each user group, the team helped us identify and prioritize their user stories.
We used this as our starting point for defining our endpoints.
The final step involved using all the prior information to create a simple RESTful API.
Our process involves taking over a 150 Word documents and 5 spreadsheets and converting them into a single JSON file (with accompanying XML and CSV files).
Because Quarry expects the data to be served from a database, we have to create a separate script to export structure and context to a MySQL DB. Because of unicode issues, I also had to export to a MSSQL.
It is MSSQL that currently drives the API.
The second API is “Census CitySDK”.
The first link is the URL for the project’s home; and the second is the project Github repo.
CitySDK is a software development kit that enables the easy and seamless integration of Federal and local data sources in order of help civic innovators quickly build solutions to their local problems.
Though, we start with the data from the US Census Bureau, the intention is to expand the number of Federal agencies included in each release.
The Census Bureau collects over 20,000 attributes on a representative sample of the 320 million people in American.
All the demographic studies that involve Americans and all solutions that include American geography use Census data.
Census has an API. However, most people prefer to perform bulk downloads rather than work with the Census API – because the simplest requests require multiple, non-intuitive steps.
To get to the specific issues that need to be fixed, we spoke to our users – the civic hackers.
We held a series of user discovery sessions and gathered the feedback from as many people as we could – legally.
We were constrained by the fact that the Leadership team needed us to help with increasing usage of the Census API.
The IT team behind Census API would use the feedback received from our engagements to chart their path forward.
For the Minimally Viable Product, we developed a thin layer that abstracted away the complexity of the Census API and allowed a JS developer to easily download Census attributes for any region and combine it with local datasets.
This was a decent start, but did not have the flexibility to be useful for developers that did not only want to visualize data.
So, we created a beta that can be used in any programming language. We implemented in using node.js and it enables richer data analysis scenarios.
The third API is MIDAAS, which stands for Making Income Data Accessible As a Service.
The first link is the website that contains all the information on the project.
And the second link is the Github repo for the API.
This project focuses on Income Inequality – how do we have an informed discussion around income and wealth & enable the developer community to start building system based on income data from the Census Bureau and the Bureau of Labor Statistics.
Income Inequality was defined by the President as one of the defining issues of our time.
The Department of Commerce has a Data Advisory Board called CDAC – The Commerce Data Advisory Committee - a group a high-level executives in the data space who echo’ed the President’s sentiment and provided the team with the initial user stories that we focused on.
Fortunately, the Census Bureau has a deep bench of experts who have working in the Income and Wealth space for over three decades.
Trudi Renwick, who leads the Bureau’s Survey of Income and Program Participation, was kind enough to validate all things that we did with the data.
Fortunately, we were finally exploring getting formal approval to use the cloud within the Department.
The timing was right and the Commerce Data Service – a startup with the Department – could be the vehicle for externalizing this initiative.
Both MVP and beta versions were built using AWS.
We took the ACS PUMS data, which is the most detailed (and most under-utilized) dataset produced by the US Census Bureau.
ACS stands for the American Community Survey and PUMS stands for the Public Microdata Use Sample.
We downloaded the full dataset, extracted the income and wealth dimensions and created an API specifically focused on accessing those dimensions.
For the MVP, we used Redshift, Lambda and API gateway. For the beta, we had to shift to an AWS stack that was FEDRAMP certified. So, we went to Postgres, S3 and EC2.
Let’s look at what we learned from all this.
It is important to have the support of leadership and the technology shop to both clear the way for these projects to happen and to ensure that these solutions persist.
It is extremely important that all projects are scoped for maximum awesomeness, which means that 1) they are focused, 2) they include user input, 3) that there must be a need for them, 4) that there should be access to both domain experts and end users, 5) there is a delivery path that fits naturally into an existing workflow.
Each of the APIs I presented used a different tech stack – because each had a different technical constraint.
Each solution was rigorously validated by the appropriate stakeholders because we need to make sure that we are appropriately using the data and that the data is high-quality.
Finally, we typically show our initial mockups or versions of what we are building in 2 to 4 weeks in order to start engaging our stakeholders.
Sweat and Toil: The monthly access stats for the data now triples the ILAB web traffic. We have taken the API to a few hackathons and a few teams have built interesting apps on the data.
CitySDK: In Minnesota, civil hackers built a CitySDK app that helps people with disabilities easily find a place to live or travel to in the state that satisfies their specific accessibility needs.
In Chicago, innovators created Purshable - a CitySDK mobile app helps reduce waste, increase grocer’s profit, and offer shoppers high quality food at a fraction of the cost.
In Washington DC, technologists developed HyperLocal - a CitySDK app that helps Food Truck operators without a lottery parking spot find customers by locating tweets from those who are hungry.
We have been fortunate enough to have caught the eye of a few organizations that have honored the team’s work.