Jaimi Dowdell presents "Data Journalism for Business Reporting" at the free business journalism workshop, "Be a Better Business Watchdog -- CAR for Business Journalists," hosted by the Donald W. Reynolds National Center for Business Journalists, The Seattle Times and the University of Washington.
The document discusses several examples of computer-assisted reporting (CAR) projects from various media outlets:
1. The Times (UK) obtained police misconduct claims data via FOI requests and found a total of £44M spent on claims and 76% settled out of court.
2. The Guardian (UK) obtained a gifts register from a government department via FOI and found weekly lavish gifts accepted from oil companies.
3. Stephen Grey plotted flight records from plane spotters to investigate extraordinary rendition flights by the CIA.
4. CBS News analyzed 50 states' suicide data and found veterans' suicide rates were much higher than non-veterans.
5. The Wall Street Journal
RPI Research in Linked Open Government SystemsJames Hendler
Open Government systems are changing the way that many governments around the world interact with their citizenry. Transparency and innovation are both enhanced by government openness and especially by government data sharing. Further, the linked data approach, using maturing semantic web technologies, has been shown to be very valuable in creating "mashups" in which government datasets can be combined in new and innovative ways, and turned into "live" infographics. In this talk, presented to the computer science department at PUC-RIO, we describe the research aspects of RPI's ongoing work in this area.
This document provides an overview of various online databases and resources that can be used for computer-assisted reporting, including press releases, book text searches, government transcripts, legal cases, census data, and more. It also discusses how to effectively search databases using Boolean operators and tips for online research such as verifying information and considering the credibility of sources.
The document provides an overview of various websites that can be used to find information on different topics. It discusses search engines like Google and Ask, subject lists like About.com and RefDesk.com, reference sources like Answers.com, government directories like USA.gov, and specialized information sites for consumers, health, and travel. The document aims to help users find relevant online resources through search engines and subject-organized directories.
This document contains statistics from March 18, 2015 on registered businesses and federal contract awards by socio-economic category. It shows the percentages of registered minorities, women, veterans, 8A, and HUBZone businesses, along with the number of contracts awarded and total dollars awarded to each group. Contact information is provided for questions.
Jaimi Dowdell, training director for the Investigative Reporters and Editors (IRE), and Mark Horvit, executive director of IRE, offer a guide to using data in business reporting for the free investigative workshop, "Accountability in Indian Country - Be a Better Business Watchdog," on July 18, 2013.
Presented by the Donald W. Reynolds National Center for Business Journalism, this workshop was part of the Native American Journalists Association's annual conference in Phoenix.
For more information about free training for business journalists, please visit businessjournalism.org.
For additional resources on using data to empower your coverage, please visit the training archive page at http://businessjournalism.org/2013/07/17/accountability-in-indian-country-be-a-better-business-watchdog-self-guided-training/.
Jaimi Dowdell presents "Data Journalism for Business Reporting" at the free business journalism workshop, "Be a Better Business Watchdog -- CAR for Business Journalists," hosted by the Donald W. Reynolds National Center for Business Journalists, The Seattle Times and the University of Washington.
The document discusses several examples of computer-assisted reporting (CAR) projects from various media outlets:
1. The Times (UK) obtained police misconduct claims data via FOI requests and found a total of £44M spent on claims and 76% settled out of court.
2. The Guardian (UK) obtained a gifts register from a government department via FOI and found weekly lavish gifts accepted from oil companies.
3. Stephen Grey plotted flight records from plane spotters to investigate extraordinary rendition flights by the CIA.
4. CBS News analyzed 50 states' suicide data and found veterans' suicide rates were much higher than non-veterans.
5. The Wall Street Journal
RPI Research in Linked Open Government SystemsJames Hendler
Open Government systems are changing the way that many governments around the world interact with their citizenry. Transparency and innovation are both enhanced by government openness and especially by government data sharing. Further, the linked data approach, using maturing semantic web technologies, has been shown to be very valuable in creating "mashups" in which government datasets can be combined in new and innovative ways, and turned into "live" infographics. In this talk, presented to the computer science department at PUC-RIO, we describe the research aspects of RPI's ongoing work in this area.
This document provides an overview of various online databases and resources that can be used for computer-assisted reporting, including press releases, book text searches, government transcripts, legal cases, census data, and more. It also discusses how to effectively search databases using Boolean operators and tips for online research such as verifying information and considering the credibility of sources.
The document provides an overview of various websites that can be used to find information on different topics. It discusses search engines like Google and Ask, subject lists like About.com and RefDesk.com, reference sources like Answers.com, government directories like USA.gov, and specialized information sites for consumers, health, and travel. The document aims to help users find relevant online resources through search engines and subject-organized directories.
This document contains statistics from March 18, 2015 on registered businesses and federal contract awards by socio-economic category. It shows the percentages of registered minorities, women, veterans, 8A, and HUBZone businesses, along with the number of contracts awarded and total dollars awarded to each group. Contact information is provided for questions.
Jaimi Dowdell, training director for the Investigative Reporters and Editors (IRE), and Mark Horvit, executive director of IRE, offer a guide to using data in business reporting for the free investigative workshop, "Accountability in Indian Country - Be a Better Business Watchdog," on July 18, 2013.
Presented by the Donald W. Reynolds National Center for Business Journalism, this workshop was part of the Native American Journalists Association's annual conference in Phoenix.
For more information about free training for business journalists, please visit businessjournalism.org.
For additional resources on using data to empower your coverage, please visit the training archive page at http://businessjournalism.org/2013/07/17/accountability-in-indian-country-be-a-better-business-watchdog-self-guided-training/.
Jaimi Dowdell presented computer-assisted reporting tips during the free Reynolds Center workshop, "Be a Better Business Watchdog -- CAR for Business Journalists," in Missoula, Mont. on Oct. 6, 2012.
For more information on free training for journalists, please visit businessjournalism.org.
This document provides tips for journalists on how to use data to find new story ideas, enhance reporting, and advance their careers. It discusses how data can be used to uncover trends, find compelling examples, and present statistics with context. The document then provides step-by-step instructions for basic spreadsheet skills and finding various types of publicly available data, including education, property, census, campaign finance, and government budget data. Journalists are advised to proceed carefully with data analysis and verify figures.
This document provides an overview of government documents librarianship. It discusses what government documents are, the types of information found in documents, and how to access document collections. It also outlines the day-to-day responsibilities of a government documents librarian, which include reference, instruction, web archiving, and collection development. Tips are provided for navigating document collections, such as determining the relevant agency and using catalog and database search tools. Popular document series and resources for new documents librarians are also highlighted.
This document provides information on finding government and legal resources. It outlines websites for accessing Florida state laws and statutes, federal laws and regulations, census and demographic data from the US Census Bureau, legal cases from LexisNexis Academic, international news and affairs from CIAO, and additional databases for issues, Latin America, statistical data, and broad information from ProQuest Central. Key websites are listed along with brief descriptions of search capabilities.
The document provides an overview of the resources available from the Kentucky Department for Libraries & Archives (KDLA) for accessing federal government information. KDLA serves as a selective federal depository library and provides access to government documents, reports, and databases in various formats. Key resources highlighted include the online catalog, statistical and census data, legislative information, and databases such as DARTS, HeinOnline, and the Homeland Security Digital Library. Contact information is provided for reference assistance.
Jaimi Dowdell presents "Data Journalism for Business Reporting" during the free, daylong workshop, "Be a Better Business Watchdog -- CAR for Business Journalists."
For more information about free training for business journalists, please visit businessjournalism.org.
State of Florida Neo4J Graph Briefing -Payments to Prescriptions AnalysisNeo4j
The document discusses using Neo4j to analyze connections between payments made by pharmaceutical companies to physicians and prescription patterns of opioid drugs by those physicians. It combines data from CMS Open Payments on pharmaceutical payments to doctors, Medicare Part D data on prescriptions written by doctors, and DEA data on pill distribution, to determine if payments are correlated with increased opioid prescriptions. The analysis focuses on payments and prescribing patterns from 2013 to 2015.
Locating scientific government information on the webShannon Lynch
This is a 2017 Powerpoint presentation given at the Department of Interior Library. The sources and information were correct at the time of presentation but have not been updated and should be double checked for current accuracy. Please feel free to contact the Department of Interior Library with any questions.
The document proposes a new service called Trellis that would provide concise summaries of local trial court judges' tentative rulings to help lawyers better understand and predict how those judges may rule. It notes current legal research tools focus on appellate courts but most cases are decided in local trial courts. Trellis would gather tentative rulings from individual counties and allow searching by judge, county, and topic. This would help lawyers determine a judge's temperament and tendencies. The plan is to start in California and New York, home to many lawyers and cases, then expand nationwide.
The presentation discussed strategies for obtaining public data and metadata through FOIA requests. It explained that data alone is not very useful and metadata is needed to understand the context and structure of the data. The presenters shared examples of FOIA requests they submitted to various agencies for data and metadata, and the challenges they faced obtaining usable information. They emphasized the importance of crafting focused yet broad requests, and being prepared to submit follow-up requests to get the necessary documentation to understand and work with the data.
National latina researchers network supercharge your search 2015 webinarMatthew Von Hendy
This document provides an overview of strategies and resources for conducting effective literature searches. It discusses using major search engines like Google while also being aware of limitations. The document recommends beginning with search engines, then searching major citation databases to find related works and who is citing sources. It suggests then searching subject-specific research databases. The document provides tips for effective searching and lists many free resources for full-text articles, images, citation management, legislation tracking, government reports, and grey literature. The presenter is available for questions.
This handout on data-driven enterprise reporting was prepared by Steve Doig, professor of journalism, specializing in data reporting, at the Walter Cronkite School of Journalism and Mass Communication at Arizona State University. It was distributed by Sarah Cohen, Knight Chair specializing in data journalism at Cronkite, at Phoenix NewsTrain on April 6-7, 2018. It includes why study data journalism; how to get started; how to find data, including selected data sets to get started; how to avoid errors; and where to get more training. It accompanies Cohen's presentation, "Data-driven enterprise." NewsTrain is a training initiative of Associated Press Media Editors (APME). More info; http://bit.ly/NewsTrain
Global Data Brokers is advertising an email list of multi-millionaires and billionaires for marketing purposes. The list contains over 1,800 billionaires worldwide and 540 in the US alone. It provides their contact information in an easy-to-use format. The company claims the list can help inventors, entrepreneurs, and businesses to connect with wealthy individuals to promote products, services, or investment opportunities. Global Data Brokers sources the information from public lists like the Forbes 500 and guarantees up-to-date accurate data on high-net-worth contacts.
Jason Parker gave a presentation on "Open Data Sources for Grants" to the Tennessee Chapter of the Grant Professionals Association on September 10, 2014. This presentation includes a wide variety of open data resources that grant writers can use to strengthen proposals.
The document provides tips and resources for conducting effective online searches and research. It recommends knowing your goals, available sources, and limits before searching. It also suggests using advanced search features like phrases, Boolean logic, and filtering by domain or file type. The document lists specific search engines, databases, and tools that can help locate public records, government data, social media information, and perform reverse lookups, people searches, and investigations into websites, emails, and backgrounds.
Infotrac provides access to the General Reference Center Gold database from Gale, which contains over 57 million articles from magazines, newspapers, encyclopedias and reference books. Users can search by basic, advanced, subject or publication browse. Search results can be filtered and include direct links to full text, images, audio and social sharing options. Infotrac allows searching across multiple Gale databases simultaneously and is a valuable resource for Michigan libraries.
This document provides an overview of why data journalism is useful, how to get started with it, finding relevant data sources, and learning more about data skills. Some key benefits of data journalism mentioned are discovering new stories, finding examples to illustrate reporting, and presenting statistics to show broader trends. The document recommends starting with accessible spreadsheets, learning skills gradually, and finding resources for help and inspiration. It also cautions that data needs verification and should be used to enhance traditional reporting.
The document outlines Marcy Phelps' presentation on cost-effective searching strategies at the Searchers Academy pre-conference workshop of Computers in Libraries 2013 on April 7, 2013. It lists sources for cost-effective searching, best business sources, and best local sources. Phelps' contact information is provided at the bottom.
This handout on data-driven enterprise reporting was prepared by Steve Doig, professor of journalism, specializing in data reporting, at the Walter Cronkite School of Journalism and Mass Communication at Arizona State University. It includes why study data journalism; how to get started; how to find data, including selected data sets to get started; how to avoid errors; and where to get more training. He compiled it for Seattle NewsTrain on Nov. 11, 2017. It accompanies a presentation of the same name. NewsTrain is a training initiative of Associated Press Media Editors (APME). More info; http://bit.ly/NewsTrain
Jaimi Dowdell presented computer-assisted reporting tips during the free Reynolds Center workshop, "Be a Better Business Watchdog -- CAR for Business Journalists," in Missoula, Mont. on Oct. 6, 2012.
For more information on free training for journalists, please visit businessjournalism.org.
This document provides tips for journalists on how to use data to find new story ideas, enhance reporting, and advance their careers. It discusses how data can be used to uncover trends, find compelling examples, and present statistics with context. The document then provides step-by-step instructions for basic spreadsheet skills and finding various types of publicly available data, including education, property, census, campaign finance, and government budget data. Journalists are advised to proceed carefully with data analysis and verify figures.
This document provides an overview of government documents librarianship. It discusses what government documents are, the types of information found in documents, and how to access document collections. It also outlines the day-to-day responsibilities of a government documents librarian, which include reference, instruction, web archiving, and collection development. Tips are provided for navigating document collections, such as determining the relevant agency and using catalog and database search tools. Popular document series and resources for new documents librarians are also highlighted.
This document provides information on finding government and legal resources. It outlines websites for accessing Florida state laws and statutes, federal laws and regulations, census and demographic data from the US Census Bureau, legal cases from LexisNexis Academic, international news and affairs from CIAO, and additional databases for issues, Latin America, statistical data, and broad information from ProQuest Central. Key websites are listed along with brief descriptions of search capabilities.
The document provides an overview of the resources available from the Kentucky Department for Libraries & Archives (KDLA) for accessing federal government information. KDLA serves as a selective federal depository library and provides access to government documents, reports, and databases in various formats. Key resources highlighted include the online catalog, statistical and census data, legislative information, and databases such as DARTS, HeinOnline, and the Homeland Security Digital Library. Contact information is provided for reference assistance.
Jaimi Dowdell presents "Data Journalism for Business Reporting" during the free, daylong workshop, "Be a Better Business Watchdog -- CAR for Business Journalists."
For more information about free training for business journalists, please visit businessjournalism.org.
State of Florida Neo4J Graph Briefing -Payments to Prescriptions AnalysisNeo4j
The document discusses using Neo4j to analyze connections between payments made by pharmaceutical companies to physicians and prescription patterns of opioid drugs by those physicians. It combines data from CMS Open Payments on pharmaceutical payments to doctors, Medicare Part D data on prescriptions written by doctors, and DEA data on pill distribution, to determine if payments are correlated with increased opioid prescriptions. The analysis focuses on payments and prescribing patterns from 2013 to 2015.
Locating scientific government information on the webShannon Lynch
This is a 2017 Powerpoint presentation given at the Department of Interior Library. The sources and information were correct at the time of presentation but have not been updated and should be double checked for current accuracy. Please feel free to contact the Department of Interior Library with any questions.
The document proposes a new service called Trellis that would provide concise summaries of local trial court judges' tentative rulings to help lawyers better understand and predict how those judges may rule. It notes current legal research tools focus on appellate courts but most cases are decided in local trial courts. Trellis would gather tentative rulings from individual counties and allow searching by judge, county, and topic. This would help lawyers determine a judge's temperament and tendencies. The plan is to start in California and New York, home to many lawyers and cases, then expand nationwide.
The presentation discussed strategies for obtaining public data and metadata through FOIA requests. It explained that data alone is not very useful and metadata is needed to understand the context and structure of the data. The presenters shared examples of FOIA requests they submitted to various agencies for data and metadata, and the challenges they faced obtaining usable information. They emphasized the importance of crafting focused yet broad requests, and being prepared to submit follow-up requests to get the necessary documentation to understand and work with the data.
National latina researchers network supercharge your search 2015 webinarMatthew Von Hendy
This document provides an overview of strategies and resources for conducting effective literature searches. It discusses using major search engines like Google while also being aware of limitations. The document recommends beginning with search engines, then searching major citation databases to find related works and who is citing sources. It suggests then searching subject-specific research databases. The document provides tips for effective searching and lists many free resources for full-text articles, images, citation management, legislation tracking, government reports, and grey literature. The presenter is available for questions.
This handout on data-driven enterprise reporting was prepared by Steve Doig, professor of journalism, specializing in data reporting, at the Walter Cronkite School of Journalism and Mass Communication at Arizona State University. It was distributed by Sarah Cohen, Knight Chair specializing in data journalism at Cronkite, at Phoenix NewsTrain on April 6-7, 2018. It includes why study data journalism; how to get started; how to find data, including selected data sets to get started; how to avoid errors; and where to get more training. It accompanies Cohen's presentation, "Data-driven enterprise." NewsTrain is a training initiative of Associated Press Media Editors (APME). More info; http://bit.ly/NewsTrain
Global Data Brokers is advertising an email list of multi-millionaires and billionaires for marketing purposes. The list contains over 1,800 billionaires worldwide and 540 in the US alone. It provides their contact information in an easy-to-use format. The company claims the list can help inventors, entrepreneurs, and businesses to connect with wealthy individuals to promote products, services, or investment opportunities. Global Data Brokers sources the information from public lists like the Forbes 500 and guarantees up-to-date accurate data on high-net-worth contacts.
Jason Parker gave a presentation on "Open Data Sources for Grants" to the Tennessee Chapter of the Grant Professionals Association on September 10, 2014. This presentation includes a wide variety of open data resources that grant writers can use to strengthen proposals.
The document provides tips and resources for conducting effective online searches and research. It recommends knowing your goals, available sources, and limits before searching. It also suggests using advanced search features like phrases, Boolean logic, and filtering by domain or file type. The document lists specific search engines, databases, and tools that can help locate public records, government data, social media information, and perform reverse lookups, people searches, and investigations into websites, emails, and backgrounds.
Infotrac provides access to the General Reference Center Gold database from Gale, which contains over 57 million articles from magazines, newspapers, encyclopedias and reference books. Users can search by basic, advanced, subject or publication browse. Search results can be filtered and include direct links to full text, images, audio and social sharing options. Infotrac allows searching across multiple Gale databases simultaneously and is a valuable resource for Michigan libraries.
This document provides an overview of why data journalism is useful, how to get started with it, finding relevant data sources, and learning more about data skills. Some key benefits of data journalism mentioned are discovering new stories, finding examples to illustrate reporting, and presenting statistics to show broader trends. The document recommends starting with accessible spreadsheets, learning skills gradually, and finding resources for help and inspiration. It also cautions that data needs verification and should be used to enhance traditional reporting.
The document outlines Marcy Phelps' presentation on cost-effective searching strategies at the Searchers Academy pre-conference workshop of Computers in Libraries 2013 on April 7, 2013. It lists sources for cost-effective searching, best business sources, and best local sources. Phelps' contact information is provided at the bottom.
This handout on data-driven enterprise reporting was prepared by Steve Doig, professor of journalism, specializing in data reporting, at the Walter Cronkite School of Journalism and Mass Communication at Arizona State University. It includes why study data journalism; how to get started; how to find data, including selected data sets to get started; how to avoid errors; and where to get more training. He compiled it for Seattle NewsTrain on Nov. 11, 2017. It accompanies a presentation of the same name. NewsTrain is a training initiative of Associated Press Media Editors (APME). More info; http://bit.ly/NewsTrain
Burt Hubbard is a data journalist who has worked with the Rocky Mountain News, Denver Post, Rocky Mountain PBS and 9News on investigative projects and documentaries. His numerous awards include two prestigious Best of The West awards, a national education award for investigative reporting, and Reporter of the Year in Colorado. Burt has taught computer-assisted reporting and internet research to graduate students for 20 years at the University of Colorado.
Mark Nichols prepared this presentation for Muncie NewsTrain on March 24, 2018. He is a data journalist on the national investigations team for USA Today. It discusses why to use data in reporting, where to find data, how to sort and filter data, and how to develop your own data-driven story off your beat. It is accompanied by a handout: Doing Data-Driven Reporting off Your Beat. NewsTrain is a training initiative of Associated Press Media Editors (APME). More info: http://bit.ly/NewsTrain
This document provides an overview of techniques for collecting, cleaning, and manipulating data for investigative reporting purposes. It discusses finding and obtaining data from various sources like government databases and agencies. It also covers integrity checks, data cleaning, and evaluating outside studies. Examples are given of reports that were produced using analyzed data from sources like Medicaid and EPA. The document emphasizes being thorough in understanding data and accounting for limitations and errors.
The document summarizes information about the United States Census Bureau, including its role in collecting population and economic data, how it conducts surveys and censuses, what data is collected, and how the data is used. It provides details on the decennial census process and products available from the Census Bureau.
Discovering and mapping your community needs - HealthLandscape
Presented at the 2013 Community Connections Pre-Application Workshops for The HealthPath Foundation of Ohio
This document discusses secondary data sources. It describes why secondary data is used, such as to provide context or validation for primary data. It outlines different types of official and unofficial secondary data sources across topics like agriculture, business, crime, education, and health. Specific sources mentioned include the US Census, American Community Survey, and TIGER database. Advantages of secondary data include low cost and timeliness, while disadvantages include lack of control over data collection methods.
This document discusses legal apps, websites, and resources available for legal research. It outlines various apps for tasks like trial presentation, productivity, and password management. It also summarizes free legal research sources like Casemaker, Fastcase, and websites through the Franklin County Law Library. These free sources provide primary law but lack edited secondary sources. The document also notes the existence of "hidden" legal information not found through basic searches, and how sites like Pacer provide access to large datasets of court records and filings. Overall, it encourages using free resources initially but also accessing the Franklin County Law Library for more comprehensive research needs.
FSU SLIS Week 14 Intro to Info Services: Health, Law and BusinessLorri Mon
This document provides an overview of resources for business, law, and health information. It discusses key databases for researching public and private companies, as well as non-profits. It also outlines the structure of the US legal system and describes primary and secondary legal sources. Additionally, it lists important medical databases and government health websites for researching clinical topics and finding medical statistics and guidelines.
This document provides an overview of strategies and resources for assisting patrons with data reference questions. It discusses John Snow's 1854 cholera map, examples of common data reference questions, and a five question framework for the data reference interview process. This includes determining the type of data needed, relevant variables and populations, geographic and time frame requirements, and data access and use conditions. Government agencies, data portals, and statistical resources are presented. Tips are offered for pairing with another librarian to practice sample questions using the interview framework.
This document provides guidance on researching companies and industries. It outlines a four step process: 1) choosing a topic and scope of research, 2) accessing relevant information sources effectively, 3) analyzing and evaluating sources, and 4) presenting findings. Key information sources discussed include company websites, annual reports, databases like Hoover's, Factiva, and Mergent, as well as government sites like EDGAR and SEDAR for filings and industry sources. Tips are provided on searching techniques, evaluating source reliability and bias, and analyzing business information.
This document provides tips for evaluating information found on the web. It discusses that the quality of websites varies greatly as the web is uncontrolled. It suggests evaluating websites based on factors like author credentials, date of publication, potential biases, accuracy and purpose. Recommended sources include subject directories, search engines like Google which can be refined, and .gov, .edu sites which are usually reliable. The document stresses the importance of fact checking online information against books and articles verified by experts.
Matt Wynn, developer and watchdog reporter at the Omaha World-Herald, offers tips on the following in this handout for Lincoln, Nebraska, NewsTrain on April 9, 2016:
--Why learn about data journalism?
--How to get started in data journalism
--Where to find data sets
--Examples of data sets for government, education, criminal justice, health, sports and other beats
--Where to learn more about data journalism
--Things he wishes someone had told him about data journalism
It accompanies his presentation, "Data-Driven Enterprise off Your Beat." NewsTrain is a training initiative of Associated Press Media Editors (APME). More info: http://bit.ly/NewsTrain
"Locating Statistics in Health, Education, and Criminal Justice" created for a Documents Interest Group of Oregon (DIGOR) workshop presented November 21, 2008 at George Fox University in Portland, OR. Edited for general use.
The document lists 18 online resources for business research, including websites that provide market data, company information from SEC filings, economic indicators, labor statistics, business environment profiles, industry reports, and information on human rights issues. It summarizes the type of information and coverage available on each site. The resources can be used to research industries, companies, economics, and other business topics.
The document discusses databases, including what they are, how they are organized, and common examples. It explains that a database is an organized collection of related information and a database management system allows users to create, edit, and retrieve data. Popular databases mentioned include Netflix, cell phone contacts, medical records, and school records. The document outlines the basic components of a database like entries, fields, and records. It also describes common operations with databases like sorting, searching, filtering and generating reports.
This handout accompanies a presentation, "Data-Driven Enterprise off Any Beat," by Manuel Torres, enterprise editor at The Times-Picayune | Nola.com. It details what data journalism can do for a journalist, how to get started with data journalism, how to find data and how to learn more about data journalism. It also lists links to many data sets by beat. NewsTrain is a training initiative of Associated Press Media Editors: http://bit.ly/NewsTrain
This document provides an overview of investigative reporting and the use of data in journalism. It discusses how investigative reporting can uncover new information, corruption, or injustice. It explains that investigative reporting often involves using data and records from public agencies to find stories. Some tips discussed include requesting data and records early, talking to real people, and keeping work organized. The document also provides guidance on ensuring the integrity of data through checks for missing information, duplicates, outliers and comparing to other sources. It notes that missing data or erroneous government databases can sometimes be stories in themselves.
Similar to A crash course in data for information graphics (20)
How journalists are using interactive tools to create data-driven maps on the web. Includes examples from a variety of media platforms and a survey of tools that beginners and experts can use.
This document discusses the history and uses of geographic information systems (GIS) for mapping news stories. It outlines how GIS was first used in the 1850s to map cholera outbreaks in London, then later to map Hurricane Andrew's damage in 1992 Miami. Modern applications include interactive maps, hyperlocal news maps, data exploration, and open-source options like QGIS. GIS allows layering of data, geocoding addresses, selecting data by location or attributes, and spatial analysis tools. While GIS programs require investment of money and time, there are also free and open-source options as well as online training resources.
We're building a news application about all the televised political ads that run in the Columbia-Jefferson City, Mo., local market. Here's the how and why behind the Political AdVault.
This document provides resources for accessing open records in Missouri. It outlines where to find state and local records retention schedules, state auditor reports, and state data portals. It also provides tips for identifying records custodians and additional tools like the Missouri Blue Book and Roster. Contact information is included for getting help with open records requests from organizations like the Missouri Sunshine Coalition.
Connecting to state data using OpenMissouri.orgDavid Herzog
This document summarizes the goals and activities of the OpenMissouri project, which aims to connect Missourians with public data from state and local government agencies. The project serves to inform citizens about available data, inspire its use, and educate about how to access and use public data. It catalogs over 250 datasets from 20 state agencies on its website. The site was built using open-source technologies and aims to expand its coverage to include more cities and counties as well as different data formats.
An overview of how geographic information systems work and how data journalists are using them to tell better stories in print and on the web. For the 2012 NICAR conference in St. Louis.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Building RAG with self-deployed Milvus vector database and Snowpark Container...Zilliz
This talk will give hands-on advice on building RAG applications with an open-source Milvus database deployed as a docker container. We will also introduce the integration of Milvus with Snowpark Container Services.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
2. Locating the data
Obtaining the data
Evaluating the data
Working with the data
Visualizing the data
3. “Database state of mind”
Data has to exist. Where?
Online
Offline
4. Government websites
Data.gov
U.S. Census Bureau
FDIC
Missouri Data Portal
Missouri Accountability Portal
5. U.S. agency FOIA pages
Drug Enforcement Administration
NGO sites
Right-to-Know Network
OpenMissouri.org
NICAR database library
ALA state agency databases wiki
6. Commercial services
Socrata
Infochimps
Geocommons
Foreclosure Radar
Oil Price Information Service
Search Systems
Junar
7. Academic data catalogs
ICPSR
Forms
Forms.gov
Web forms
▪ Columbia parade permits
8. Records retention schedules
Reports
State auditor
U.S. Government Accountability Office
U.S. Inspectors General
9. Google advanced search
Look for data files
Look for key words
Look only on government sites
10.
11. Data entry
In the field
At the office
Printouts/reports
Inspection forms
12. Download it
Write or request a scraper with ScraperWiki
Convert a PDF with
CometDocs
Zamzar
Just ask for it
13. U.S. Freedom of Information Act
Passed in 1966
Amended in 1996 to include electronic records
State open-records statutes
Missouri Sunshine Law
14. Get the roadmap!
Record layout
File layout
Data dictionary
Code sheet
Metadata
Data about the data
15. Look at it immediately when you get it
It is what you asked for/expected?
How many rows/records of data?
Is the file format OK?
16. Does it look too good to be true?
Beware of missing information
Who collected the information?
How? What are their methods?
Why?
What is their agenda?
Who supports them financially or otherwise?
21. Always keep original file
Never overwrite data columns
Tools
Spreadsheets
Database managers
Google Refine
Programming languages
22. Raw numbers, without context, rarely are
interesting.
Ask: Compared to what?
23. Raw (amount) change
New-Original
Percent change
Change/Original
Per capita rates
Per person
Per x people
24. Percent of total
Individual/Total
Ratio
Apples/oranges
Averages
Mean
Median
25. Be curious!
Cut out small slices
Spreadsheets for simple math and
comparisons
Spreadsheets for pivot tables
Database managers for more robust analysis
Always ask: Is this correct?