This document discusses using computational tools to analyze and provide access to large audiovisual collections. It describes a project called HiPSTAS that developed the open-source ARLO tool to analyze 68,000 hours of audio content using speech-to-text transcription, audio waveform analysis, and machine learning. The goal was to assess how these tools could help humanities scholars better access spoken word collections. Key challenges discussed include balancing accuracy with efficiency and ensuring tools are accessible and usable across disciplines.
Slides of my presentation on regional variation of Finnic runosongs at the conference "Expressions and Impressions: Aspects of Traditional Singing", 7.-9. Nov. 2018, Estonian Literary Museum, Tartu, Estonia.
After some introducting slides the results of stylometric and network analysis of texts in two Finnic folk song corpora are presented and commented: the division of material into regional groups and regional/stylistic peculiarities of these groups.
Dr. Pamela Burnard: Digital R&D in the Arts Annual ForumNesta
Dr Pamela Burnard, from the University of Cambridge, describes her participation and findings so far from the Sonic Pi: Live & Coding project that works alongside Cambridge Junction.
This presentation was delivered at the Digital R&D in the Arts Annual Forum at Vinopolis on 3 July 2014. For more information, visit: http://www.nesta.org.uk/event/digital-rd-arts-annual-forum
Slides of my presentation on regional variation of Finnic runosongs at the conference "Expressions and Impressions: Aspects of Traditional Singing", 7.-9. Nov. 2018, Estonian Literary Museum, Tartu, Estonia.
After some introducting slides the results of stylometric and network analysis of texts in two Finnic folk song corpora are presented and commented: the division of material into regional groups and regional/stylistic peculiarities of these groups.
Dr. Pamela Burnard: Digital R&D in the Arts Annual ForumNesta
Dr Pamela Burnard, from the University of Cambridge, describes her participation and findings so far from the Sonic Pi: Live & Coding project that works alongside Cambridge Junction.
This presentation was delivered at the Digital R&D in the Arts Annual Forum at Vinopolis on 3 July 2014. For more information, visit: http://www.nesta.org.uk/event/digital-rd-arts-annual-forum
Linked data for knowledge curation in humanities researchEnrico Daga
The identification and cataloguing of documentary evidence is an important part of empirical research in the humanities.
An increasing number of recent initiatives in the digital humanities have as a primary objective the curation of collections of digital artefacts augmented with fine-grained metadata, for example, mentioning the entities and their relations, often adopting the "Linked Data" paradigm. This talk is focused on exploring the potential of Linked Data to support humanities scholars in identifying, collecting, and curating documentary evidence. First, I will introduce the basic notions around Linked Data and place its emergence in the tradition of Knowledge Representation, an area of Artificial Intelligence (AI). Second, I will show how Linked Data and AI techniques have been successfully applied in the Listening Experience Database project to support the retrieval and curation of documentary evidence. Finally, I will conclude the presentation by discussing the potential (and challenges) of adopting a "knowledge extraction" paradigm to automate the identification and cataloguing of metadata about documentary evidence in texts.
Dot Porter (University of Pennsylvania)
'The Medieval Electronic Scholarly Alliance: a federated platform for discovery and research'.
Digital Classicist London & Institute of Classical Studies seminar 2013, Friday July 5th.
The Medieval Electronic Scholarly Alliance (MESA) is a federated international community of scholars, project, institutions, and organizations engaged in digital scholarship within the field of medieval studies. Funded by the Andrew W. Mellon Foundation, MESA seeks both to provide a community for those engaged in digital medieval studies and to meet emerging needs of this community, including making recommendations on technological and scholarly standards for electronic scholarship, the aggregation of data, and the ability to discover and repurpose this data.
This presentation will focus on the discovery aspect of MESA, and how it might serve the non-digital medievalist who may nevertheless be interested in finding and using digital resources. Starting with a history of medievalists and their interactions with digital technology as told through three data sets (the International Congress on Medieval Studies (first held in 1962), arts-humanities.net (a digital project database in the UK, sponsored by JISC and the Arts & Humanities Research Council), and two surveys, from 2002 and 2011, that looked specifically at medievalists' use of digital resources), I will draw out some potential issues that this history has for the current developers of digital resources for medievalists, and investigate how MESA might serve to address these issues.
Library Futures & the Importance of Understanding Communities of UsersChristine Madsen
In 2010 I finished a two year ethnographic study of that aimed at understanding how the digitization of rare texts is changing scholars’ work and in turn how it is changing their relationship with the library. I will present some highlights from the findings of that research and discuss more recent research to understand the future of libraries by understanding communities of users. In other words, what can we learn from Tibetan Buddhists, the Parakuyo Maasai, and the CTOs of our top technology companies about how to build the library of the future?
This presentation was the 2013 Breslauer Lecture at UCLA GSEIS.
presentation for a workshop on cataloging medieval manuscripts with Debra Cashion, Sheila Bair and Sue Steuer which was held at the Rare Book and Manuscript Section (RBMS) of the Association of College and Research Libraries (ACRL) in Minneapolis, MN on June 27, 2013.
New Mexico PBS and American Archive of Public Broadcasting staff present on collaborative grants with stations large and small to preserve programs and original materials contributed by 125 TV and radio stations, archives, and producers in New Mexico.
Access the recording by visiting PBS Hub and creating a free account at https://hub.pbs.org/posts/engage-your-community-to-celebrate-your-history?parentId=6881.
Join the American Archive of Public Broadcasting (AAPB) and Wikipedia's official United States affiliates for a virtual edit-a-thon to help strengthen the quality of the world’s largest online encyclopedia and improve the searchability of historic public radio and television collections in the AAPB.
Linked data for knowledge curation in humanities researchEnrico Daga
The identification and cataloguing of documentary evidence is an important part of empirical research in the humanities.
An increasing number of recent initiatives in the digital humanities have as a primary objective the curation of collections of digital artefacts augmented with fine-grained metadata, for example, mentioning the entities and their relations, often adopting the "Linked Data" paradigm. This talk is focused on exploring the potential of Linked Data to support humanities scholars in identifying, collecting, and curating documentary evidence. First, I will introduce the basic notions around Linked Data and place its emergence in the tradition of Knowledge Representation, an area of Artificial Intelligence (AI). Second, I will show how Linked Data and AI techniques have been successfully applied in the Listening Experience Database project to support the retrieval and curation of documentary evidence. Finally, I will conclude the presentation by discussing the potential (and challenges) of adopting a "knowledge extraction" paradigm to automate the identification and cataloguing of metadata about documentary evidence in texts.
Dot Porter (University of Pennsylvania)
'The Medieval Electronic Scholarly Alliance: a federated platform for discovery and research'.
Digital Classicist London & Institute of Classical Studies seminar 2013, Friday July 5th.
The Medieval Electronic Scholarly Alliance (MESA) is a federated international community of scholars, project, institutions, and organizations engaged in digital scholarship within the field of medieval studies. Funded by the Andrew W. Mellon Foundation, MESA seeks both to provide a community for those engaged in digital medieval studies and to meet emerging needs of this community, including making recommendations on technological and scholarly standards for electronic scholarship, the aggregation of data, and the ability to discover and repurpose this data.
This presentation will focus on the discovery aspect of MESA, and how it might serve the non-digital medievalist who may nevertheless be interested in finding and using digital resources. Starting with a history of medievalists and their interactions with digital technology as told through three data sets (the International Congress on Medieval Studies (first held in 1962), arts-humanities.net (a digital project database in the UK, sponsored by JISC and the Arts & Humanities Research Council), and two surveys, from 2002 and 2011, that looked specifically at medievalists' use of digital resources), I will draw out some potential issues that this history has for the current developers of digital resources for medievalists, and investigate how MESA might serve to address these issues.
Library Futures & the Importance of Understanding Communities of UsersChristine Madsen
In 2010 I finished a two year ethnographic study of that aimed at understanding how the digitization of rare texts is changing scholars’ work and in turn how it is changing their relationship with the library. I will present some highlights from the findings of that research and discuss more recent research to understand the future of libraries by understanding communities of users. In other words, what can we learn from Tibetan Buddhists, the Parakuyo Maasai, and the CTOs of our top technology companies about how to build the library of the future?
This presentation was the 2013 Breslauer Lecture at UCLA GSEIS.
presentation for a workshop on cataloging medieval manuscripts with Debra Cashion, Sheila Bair and Sue Steuer which was held at the Rare Book and Manuscript Section (RBMS) of the Association of College and Research Libraries (ACRL) in Minneapolis, MN on June 27, 2013.
New Mexico PBS and American Archive of Public Broadcasting staff present on collaborative grants with stations large and small to preserve programs and original materials contributed by 125 TV and radio stations, archives, and producers in New Mexico.
Access the recording by visiting PBS Hub and creating a free account at https://hub.pbs.org/posts/engage-your-community-to-celebrate-your-history?parentId=6881.
Join the American Archive of Public Broadcasting (AAPB) and Wikipedia's official United States affiliates for a virtual edit-a-thon to help strengthen the quality of the world’s largest online encyclopedia and improve the searchability of historic public radio and television collections in the AAPB.
Corrected transcripts improve the searchability of historic programs in the American Archive of Public Broadcasting and the staff needs YOUR help! This presentation includes an overview of the AAPB's mission, why FIX IT+ is important, and brief instructions on preferred editing conventions.
“Press Play on History” focuses on activities to engage students with primary sources in the AAPB's Protesting in America exhibit.
The American Archive of Public Broadcasting, a collaboration between the Library of Congress and public media producer GBH, streams primary and secondary public broadcasting content dating back more than 70 years. Over 130 organizations have contributed historic and culturally significant collections to the AAPB, providing educators with online, audiovisual materials for distance teaching and learning.
4,000 assets created by 230 different television and radio stations over a seventy-year period, stored on twenty-five different media formats, digitized and made accessible through a three-institution partnership during a pandemic. What could possibly go wrong? Four participants in this collaborative effort will discuss their contributions to the project, including innovative tools, evolving procedures, and collaborative strategies. They will also speak to the policies and tactics that have allowed the project to remain on track during the pandemic. Key topics include obtaining permissions from rights holders; creating and correcting speech-to-text transcripts, managing a remote workforce; conducting research during Covid-19 and curating an online exhibit; and digital asset management and quality control. Session sponsored by the News/TV/Docs committee.
Presented by:
Mary Lynn Miller, Brown Media Archives, University of Georgia
Kathleen Carter, Brown Media Archives, University of Georgia
Thomas May, Brown Media Archives, University of Georgia
Sally Smith, UNC School of Information and Library Science
Miranda Villesvik, GBH
Webinar hosted by the Boston Library Consortium and AAPB staff at WGBH. Presenters included Casey Davis Kaufman (AAPB, WGBH), Ryn Marchese (AAPB, WGBH), Ingrid Ockert (Princeton University), and Mark Williams (Dartmouth College).
Webinar hosted by American Archive of Public Broadcasting staff with presenters including Ryn Marchese (AAPB, WGBH), Kathryn Ostrofsky (Clark University), and Joshua Glick (Hendrix College).
Presentation about the CLIR-funded National Educational Television Collection Catalog Project by Sadie Roosa at the 2018 Digital Commonwealth conference.
Presentation given by Charles Hosale, Special Projects Assistant at WGBH/American Archive of Public Broadcasting; Leslie Bourgeois, Archivist at Louisiana Public Broadcasting; Ann Wilkens, Archivist at Wisconsin Public Television; and Rachel Curtis, AAPB Digital Conversion Specialist and Project Coordinator at the Library of Congress. The presentation was given at the 2017 Association of Moving Image Archivists conference in New Orleans.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
1. LET THE COMPUTER
DO THE WORK
Karen Cariani & Casey Davis
WGBH | Boston, MA, USA
2.
3.
4. the situation
■ 68,000 digitized television and radio programs
■ incomplete, inaccurate metadata records
■ limited staff resources
■ we need to know what we have in the collection
■ we have a responsibility to users to provide access to the collection
■ continued growth of the collection (content and sparse metadata)
9. The State of Recorded Sound
Preservation in the United States: A
National Legacy at Risk in the Digital Age
(2010)
Suggested that if scholars and students do not use sound archives,
cultural heritage institutions will be less inclined to preserve them.
Archives and libraries must collaborate with patrons and scholars to
understand how recordings are and might be used.
Scholars need to know what kinds of analysis are possible in an age
of large, freely available collections and advanced computational
analysis.
11. A vision
“ . . . the sound file would become . . . a text
for study, much like the visual document.
The acoustic experience of listening to the
poem would begin to compete with the
visual experience of reading the poem.”
Bernstein, Charles. Attack of the Difficult Poems: Essays and Inventions.
University Of Chicago Press, 2011, 114.
13. HiPSTAS team
• Tanya Clement, [PI] Assistant Professor, University of
Texas at Austin
• Loretta Auvil [Co-PI] Senior Project Coordinator at the
Illinois Informatics Institute (I3) at the University of
Illinois at Urbana-Champaign
• David Tcheng [Co-PI] Research Scientist at I3; ARLO
developer
• Tony Borries, Research Programmer working as a
consultant with I3; ARLO programmer
• David Enstrom, Biologist, University of Illinois at
Urbana-Champaign; consultant
14. Participants, Hipstas Institute, 2013-
2014
• 8 librarians and archivists
• 9 humanities scholars
• 3 advanced graduate students in humanities and
information science
15. Participating collections
• poetry from PennSound at the University of
Pennsylvania 30,000 audio files
• folklore at the Dolph the Briscoe Center for American
History at UT Austin, 57 feet of tapes (reels and
audiocassettes)
• storytelling traditions at the Native American Projects
(NAP) at the American Philosophical Society in
Philadelphia , 50 tribes, 3,000 hours
16. • Field recordings (200,000 recordings) American Folklife
Center, Library of Congress
• 30, 000 hours, Oral histories, Storycorps
• Speeches in the Southern Christian Leadership
Conference recordings, Emory University
• 700 recordings in the Elliston Poetry Collection at the
University of Cincinnati
• 36 interviews in the Dust, Drought and Dreams Gone
Dry: Oklahoma Women and the Dust Bowl (WDB) oral
history project out of the Oklahoma State Libraries
OTHER COLLECTIONS OF INTEREST TO PARTICIPANTS
17. To develop a virtual research environment in which users
can better access and analyze spoken word collections of
interest to humanists through:
1. an assessment of scholarly requirements for analyzing
sound
2. an assessment of technological infrastructures needed
to support discovery
3. preliminary tests that demonstrate the efficacy of using
such tools in humanities scholarship
4. A freely available, open-source, API-driven version for
general use
HIPSTAS: PRIMARY GOALS
18. ARLO (Adaptive Recognition with
Layered Optimization)
HZ, a unit
of
frequency
Time
a heat based color scheme.
White – hottest, most
intense
Yellow
Red
Green
Blue
Black – coolest, least
intense
Energy represented by
27. Takeaways:
■ What do scholars talk about when they talk about
sound?
• Language dynamics: tempo, pitch, tone/timbre,
volume, pace, laughter, silence, applause, moans,
screams, dialects, changing speakers, gender,
age, changing genres
• Environment: fan hums, car horns, chickens, train
whistles, bird calls, frogs mating
• Materiality: recording noises, needle drops,
feedback, the electronic grid, changing tracks
28. ■What do engineers talk about they talk about
audio?
• Resolution: Bit depth, Bit rate, sample rate
• Signal processing: Fast Fourier Transform (FFT)
and filter banks
• Dynamics: Damping ratios, gain, frequencies,
spectra, energy, and pitch energy
TAKEAWAYS
29. ■ What do computer scientists talk about when
they talk about ML?
• Features: What are we measuring?
• Ground Truth: What’s the answer? How do we
know when we’re accurate?
• Optimization: Accuracy vs. Efficiency – how do
you balance the accuracy of your results
against the computational resources you need
to achieve that level of accuracy?
Takeaways
30. Takeaways
• Literacy: How much do we need to know about the
technology of audio, of computational methods, and of
humanist inquiry to do new kinds of research in this area?
• Usability: What kinds of interfaces and tools facilitate AV
analysis in a diverse range of disciplines and communities?
Who gets access to these tools and for what kinds of
questions?
• Accuracy: Is good enough, good enough?
• Scalability: How much storage and processing power do
users need to conduct local and large-scale AV analyses? A
Laptop? A Supercomputer?
• Sustainability: What are local, national, and global scale
issues? How does this work fit back into the access
infrastructure already in place in archives, libraries,
classrooms? Is data enough to get us over the hump of our
limited means for discovery?
Perform machine learning with the instance based algorithm (with distance weighting power = 15 and threshold at .4 being optimal)using inductive biased optimization to come up with configuration; all machine learning algorithms have control parameters -- you need to try out different ones to find out what is optimal for your problem (classification threshold and distance weighting power)file-based cross validation was used to measure predictive performance; simulate the process for having the ground truth from other files and see how well you can predict it; applied the model to all the files (25 million examples);For each example in the file (an example is a 1/32nd second), we classify every example. To classify the time slice: compare it to all known slices and compute the distance between it and all known slices; computing the distance for 256 bands is 256 values (feature space has 256 dimensions in it); when we compute the distance, we are computing the distance between two points in a 256-dimensional space; this means for each dimension, compute the absolute value of the difference between each feature pair and then sum all the differences using the power of 1 (city block, taxi cab distance, hamming distance [straight lines] -- [Euclidean is 2; as the crow flies])http://taxicabgeometry.net/general/basics.htmlAfter computing the distance, convert the distance into a weight; the weight for every example is derived from its distance 1/distance(raised to a power) [which in our case is 15].[not using other predictions from near windows, previous or subsequent, to determine classes]Use the classes of all the examples and all of their weights to determine a “vote” for the current example; formula for the vote, single class probability: sum up (the actual class [0 or 1] times its weight) then divide by the sum of all weights; This creates a weighted prediction where some examples get more weight than others; [if you set it to 0 every slice would get the same weight]; ground truth is not taken out; [you will see perfect examples in the results;]