Introduction to Research project PoliMediaMartijn Kleppe
Presentation about our research project 'PoliMedia - Interlinking multimedia for the analysis of media coverage of political debates'. Presented at the PoliMedia symposium, 23 January 2013, Amsterdam, the Netherlands
The document discusses how the media product develops conventions from existing music videos. It analyzed videos from Pond, Tame Impala, and Temples that used techniques like cut shots, animation, and lighting effects to create dreamlike, stylized, and non-narrative experiences. The product aims to recreate this lack of narrative and interpretive style through landscape shots using different effects, while developing its own style. It keeps a basic studio performance setup but adds surrealism through lighting and post-production effects. Unlike average pop videos that advertise the artist, this media product enhances the listening experience through visually intriguing shots open to lyric interpretation.
Polimedia Syposium - Linking the data setsdamo_damr
Presentation describes the method for link discovery that aims to connect debate content (textual documents from the parliament) on a speech level with relevant articles that contain not just the mentions of speakers but also mentions of speakers in a context of topics or events that politicians tackled in their speech in parliament. Method uses semantic and information retrieval techniques to generate automatic queries that contain the context of the parliamentary speeches and to search newspaper, radio and video data sets for the connections between speeches and newspaper articles that are covering them.
Guest Lecture: Linked Open Data for the Humanities and Social SciencesLaura Hollink
The document discusses two projects, PoliMedia and Talk of Europe, that link government data to news data as linked open data. PoliMedia links speeches from the Dutch parliament between 1945-1995 to over 1.5 million newspaper articles, while Talk of Europe publishes the entire plenary debates of the European Parliament as linked open data consisting of over 14 million RDF statements about speeches between 1999-2014. Both projects model the data as structured events that can be queried to enable complex analysis across sources and time spans.
Talk of Europe – Linking European Parliament ProceedingsAstrid van Aggelen
This document summarizes the Talk of Europe project which links proceedings from the European Parliament from 1996 to the present. The primary goals are to represent the data in RDF format, publish it as linked open data, and promote applications that make use of the data. The project has enriched the data by linking it to databases about MEPs, their political backgrounds, and categorizing debates by topic. They are holding a creative camp to bring together developers and researchers to invent new uses for the dataset using natural language processing and other techniques.
Bringing parliamentary debates to the Semantic WebLaura Hollink
Presentation of the paper 'Bringing parliamentary debates to the Semantic Web' by Damir Juric, Laura Hollink and Geert-Jan Houben at the workshop on Detection, Representation, and Exploitation of Events in the Semantic Web (DeRiVE2012) in conjunction with the 11th International Semantic Web Conference 2012 in Boston, USA.
See also the homepage of the PoliMedia project: http://polimedia.nl/
Introduction to Research project PoliMediaMartijn Kleppe
Presentation about our research project 'PoliMedia - Interlinking multimedia for the analysis of media coverage of political debates'. Presented at the PoliMedia symposium, 23 January 2013, Amsterdam, the Netherlands
The document discusses how the media product develops conventions from existing music videos. It analyzed videos from Pond, Tame Impala, and Temples that used techniques like cut shots, animation, and lighting effects to create dreamlike, stylized, and non-narrative experiences. The product aims to recreate this lack of narrative and interpretive style through landscape shots using different effects, while developing its own style. It keeps a basic studio performance setup but adds surrealism through lighting and post-production effects. Unlike average pop videos that advertise the artist, this media product enhances the listening experience through visually intriguing shots open to lyric interpretation.
Polimedia Syposium - Linking the data setsdamo_damr
Presentation describes the method for link discovery that aims to connect debate content (textual documents from the parliament) on a speech level with relevant articles that contain not just the mentions of speakers but also mentions of speakers in a context of topics or events that politicians tackled in their speech in parliament. Method uses semantic and information retrieval techniques to generate automatic queries that contain the context of the parliamentary speeches and to search newspaper, radio and video data sets for the connections between speeches and newspaper articles that are covering them.
Guest Lecture: Linked Open Data for the Humanities and Social SciencesLaura Hollink
The document discusses two projects, PoliMedia and Talk of Europe, that link government data to news data as linked open data. PoliMedia links speeches from the Dutch parliament between 1945-1995 to over 1.5 million newspaper articles, while Talk of Europe publishes the entire plenary debates of the European Parliament as linked open data consisting of over 14 million RDF statements about speeches between 1999-2014. Both projects model the data as structured events that can be queried to enable complex analysis across sources and time spans.
Talk of Europe – Linking European Parliament ProceedingsAstrid van Aggelen
This document summarizes the Talk of Europe project which links proceedings from the European Parliament from 1996 to the present. The primary goals are to represent the data in RDF format, publish it as linked open data, and promote applications that make use of the data. The project has enriched the data by linking it to databases about MEPs, their political backgrounds, and categorizing debates by topic. They are holding a creative camp to bring together developers and researchers to invent new uses for the dataset using natural language processing and other techniques.
Bringing parliamentary debates to the Semantic WebLaura Hollink
Presentation of the paper 'Bringing parliamentary debates to the Semantic Web' by Damir Juric, Laura Hollink and Geert-Jan Houben at the workshop on Detection, Representation, and Exploitation of Events in the Semantic Web (DeRiVE2012) in conjunction with the 11th International Semantic Web Conference 2012 in Boston, USA.
See also the homepage of the PoliMedia project: http://polimedia.nl/
Talk of Europe: Linked data of the European ParliamentLaura Hollink
The document summarizes the Talk of Europe project, which publishes data from proceedings of the European Parliament as linked open data. It includes over 14 million triples about 30,000 speeches given over 15 years. The data is made available through a SPARQL endpoint and can be used to analyze topics discussed, differences between members and parties, and other insights. Creative camps are held for people to work with the data.
1) The document discusses using open datasets for research purposes. It describes several open datasets including PoliMedia, which covers Dutch parliamentary debates, and Talk of Europe, which covers debates in the European Parliament.
2) Some challenges discussed include finding datasets that match research questions and determining what makes a dataset truly open. Collaboration with computer scientists may be needed.
3) The goals of using open datasets are described as both answering existing research questions and finding new research questions. Examples of analyses that could be done using the described datasets are provided.
ICWE2013 - Discovering links between political debates and mediagjhouben
Discovering links between political debates and media
by Damir Juric, Laura Hollink, Geert-Jan Houben
TU Delft - WIS
at ICWE 2013, Aalborg, Denmark, July 2013
LIS 653 Knowledge Organization | Pratt Institute School of Information | Fall...PrattSILS
This document discusses challenges related to using Twitter data for research purposes. Twitter has restrictions on the distribution and download of tweet IDs and user IDs. Researchers are limited to hydrating up to 50,000 public tweets per day. Social media collections within web archives tend to be event-driven and limited in scope. The algorithms used by Twitter to generate sample sizes cannot be verified by researchers. Storage space and sufficient computing infrastructure are also challenges. The Library of Congress has archived over 170 billion tweets but has not yet provided full access due to technical limitations.
The document analyzes tweets containing the hashtag "#SpanishRevolution" from April-May 2013 during economic/housing crises in Spain. Key findings include:
1) "#SpanishRevolution" is strongly connected to hashtags like "#15M", "#StopDesahucios", representing opposition to governing parties and social/political issues at the time.
2) Related hashtags referred to important social actors ("#15M"), calls to action ("#NoLesVotes") or metaphorically represented anti-eviction movements with colors.
3) The main discourse among #SpanishRevolution tweets centered around "#15M", "#VAEO", "#Nolesvotes" and was focused/proposed surprisingly by
Using Topic Modeling to Study Everyday "Civic Talk" and Proto-political Engag...Tuukka Ylä-Anttila
We present a two-step topic modeling method of analysing political articulations in everyday proto-political "civic talk" on online social media and interpreting them in terms of cultural and political sociology.
Using Twitter as a Postgraduate ResearcherSimon Bishop
Second version of my talk. I tried to make it more focused and a better introduction. As ever, cute pictures need no explanation.
As for Up - try explaining the plot of it to someone who hasn't seen it... ridiculous, isn't it? There's no way to sell it that way, they have to see it. In the same way, to describe how Twitter works gives no indication as to its functionality. You have to play with it and learn by experience.
Slides from a practical workshop on gathering customer insights from social media using Social Network Analysis (SNA) with NodeXL and Twitter. SNA allows you to gain insight from thousands of tweets and messages on a range of topics for marketing research or academic use. NodeXL reports can be used for measuring and monitoring an organisation’s own performance as well as a competitors´ performance. At the highest level, a SNA approach allows social media managers to recognize what their audience looks like.
The document summarizes the findings of a study on the impact of digitized scholarly resources. It describes various quantitative and qualitative methods used in the study, including webometrics, analytics, log file analysis, interviews, focus groups, and surveys. The study analyzed five digitization projects and found they had positive impacts like improving research and enabling new types of quantitative analysis. Usage varied by project, with some seeing more impact through teaching resources while others saw more impact through computational analysis of materials.
The document provides information about Section A of the exam, which focuses on newspapers and online/social media. It will be worth 45 marks and include an in-depth study of both print newspapers and their online presence as well as social media feeds. Questions will focus on media industries, audiences, and language and representation. Students are expected to understand how different media construct representations and viewpoints, contextualize codes and conventions of broadsheet and tabloid newspapers both online and offline, and consider relevant academic theories.
This presentation explains how to use newspapers and online news in research. It highlights why students should use news but why they should be careful of issues such as bias. This presentation shows the sources that are available for searching newspaper archives, both at Goldsmiths and outside, as well as where to find video news online.
Tracking Social Media Participation: New Approaches to Studying User-Gener...Axel Bruns
PhD Seminar
Thursday 29 Oct., 9.30-11 a.m.
Seminar Room, Journalism & Media Research Centre, 1-3 Eurimbla St (corner High St), Randwick
The impact of user-generated content on a variety of media industries and practices is by now well understood from a conceptual perspective (e.g. Benkler 2006; Jenkins 2006; Bruns 2008). What remains less thoroughly explored is the possibility to utilise the affordances of Web 2.0 technologies themselves to generate large datasets that can be used to track and evaluate user participation practices in order to develop a solid evidence base for further research into social media, and further development of social media projects, technologies, and policies. This presentation outlines research possibilities across a number of social media spaces, and uses the example of a current research project studying the Australian political blogosphere to explore potential methodological approaches.
Researchers need to communicate their work more broadly to have greater societal impact. There are many communication channels available, including social media, blogs, and online forums. Some best practices are maintaining a policy blog like CAP Reform that provides commentary on agricultural issues, and running an online forum like Foodlog that facilitates discussion between citizens, researchers, and policymakers. As communication norms change, journals like EuroChoices may need to reconsider their business models and explore options like open access to improve accessibility and impact.
Twitter provides a selfie of envolving languageTERMCAT
Twitter provides a wealth of data that can be used to analyze language trends in real-time. The large volume of informal tweets sent daily help document the emergence and spread of new words and expressions. By following linguistic experts and word-related hashtags on Twitter, neologisms can be identified as they enter the language and their usage over time and in different locations can be tracked. Several organizations determine high-profile "Word of the Year" choices by analyzing which terms gained most attention over the past twelve months based on Twitter and other social media discourse.
- Alessandro Gallo, a sales manager at Springer, discusses the history of scholarly publishing and Springer's role as a global publisher. He outlines Springer's extensive online journal and book collections available on SpringerLink.
- Gallo describes Springer's open access models including Open Choice, which allows authors to pay a fee to make their articles openly accessible. He notes Springer's acquisition of BioMed Central, the largest open access publisher.
- The presentation concludes with an overview of upcoming semantic linking features on SpringerLink to improve search and navigation of content.
Twitter, Public Communication and the Media Ecology: The Case of the Queensla...Axel Bruns
Presented by Axel Bruns and Jean Burgess at the ATN-DAAD workshop The World According to Twitter, Brisbane, 27 June 2011.
Part of an ongoing collaboration between the Mapping Online Publics project (http://mappingonlinepublics.net/) at the ARC Centre of Excellence for Creative Industries and Innovation, QUT, Australia (http://cci.edu.au/), and the Nachwuchsforschergruppe Wissenschaft und Internet, Universität Düsseldorf, Germany (http://nfgwin.uni-duesseldorf.de/).
A History of Social Media Listening - Simon McDermott - AttentioInfluence People
Social media listening has grown directly with the rise of online social media platforms. Early social media included bulletin boards and Usenet groups in the 1970s-80s, with Google Groups archiving posts dating back to 1981. Blogs emerged in the late 1990s and grew exponentially. In the 2000s, companies emerged to perform social media listening for clients. As platforms like Facebook and Twitter grew in the late 2000s, the business applications of listening expanded to include influencer identification, campaign impact measurement, and reputation analysis. Today, most major brands perform social media listening, and ad hoc approaches are being replaced by continual, methodological listening and analytics to guide brand monitoring, marketing, and customer service. The future includes greater integration
Social Media and Architecture Journal ArchivesNoreen Whysel
This document outlines the social media and digital archiving activities of Architecture_MPS, a research organization that publishes an open-access journal. It describes Architecture_MPS' participation in social media platforms like Twitter, Facebook, and Pinterest to promote its journal articles, conferences, and other resources. The document also provides guidance to interns on social media best practices and analytics for measuring the effectiveness of Architecture_MPS' social media campaigns.
This document outlines the social media and digital archiving activities of Architecture_MPS, a research organization that publishes an open-access journal. It describes Architecture_MPS' participation in social media platforms like Twitter, Facebook, and Pinterest to promote its journal articles, conferences, and other resources. The document also provides guidance to interns on social media best practices and analytics to measure engagement. It discusses challenges in utilizing social media and opportunities to expand the audience and conversation around architecture, media, politics and society.
Creating and Analysing Linked Open Data for the EU ParliamentLaura Hollink
The document discusses creating linked open data from speeches and transcripts of the European Parliament. It describes the data structure for representing a speech as linked data, including properties like the speaker, date, title, text in multiple languages, and links to other speeches. It also discusses linking the data to external sources to identify speakers and add additional context about countries.
Enriching Linked Open Data with distributional semantics to study concept driftLaura Hollink
Presentation at the "Proximity in Information Retrieval" symposium on the occasion of the PhD thesis defense of Jeroen Vuurens
April 26, 2017, Delft University of Technology
More Related Content
Similar to Connecting political data to media data
Talk of Europe: Linked data of the European ParliamentLaura Hollink
The document summarizes the Talk of Europe project, which publishes data from proceedings of the European Parliament as linked open data. It includes over 14 million triples about 30,000 speeches given over 15 years. The data is made available through a SPARQL endpoint and can be used to analyze topics discussed, differences between members and parties, and other insights. Creative camps are held for people to work with the data.
1) The document discusses using open datasets for research purposes. It describes several open datasets including PoliMedia, which covers Dutch parliamentary debates, and Talk of Europe, which covers debates in the European Parliament.
2) Some challenges discussed include finding datasets that match research questions and determining what makes a dataset truly open. Collaboration with computer scientists may be needed.
3) The goals of using open datasets are described as both answering existing research questions and finding new research questions. Examples of analyses that could be done using the described datasets are provided.
ICWE2013 - Discovering links between political debates and mediagjhouben
Discovering links between political debates and media
by Damir Juric, Laura Hollink, Geert-Jan Houben
TU Delft - WIS
at ICWE 2013, Aalborg, Denmark, July 2013
LIS 653 Knowledge Organization | Pratt Institute School of Information | Fall...PrattSILS
This document discusses challenges related to using Twitter data for research purposes. Twitter has restrictions on the distribution and download of tweet IDs and user IDs. Researchers are limited to hydrating up to 50,000 public tweets per day. Social media collections within web archives tend to be event-driven and limited in scope. The algorithms used by Twitter to generate sample sizes cannot be verified by researchers. Storage space and sufficient computing infrastructure are also challenges. The Library of Congress has archived over 170 billion tweets but has not yet provided full access due to technical limitations.
The document analyzes tweets containing the hashtag "#SpanishRevolution" from April-May 2013 during economic/housing crises in Spain. Key findings include:
1) "#SpanishRevolution" is strongly connected to hashtags like "#15M", "#StopDesahucios", representing opposition to governing parties and social/political issues at the time.
2) Related hashtags referred to important social actors ("#15M"), calls to action ("#NoLesVotes") or metaphorically represented anti-eviction movements with colors.
3) The main discourse among #SpanishRevolution tweets centered around "#15M", "#VAEO", "#Nolesvotes" and was focused/proposed surprisingly by
Using Topic Modeling to Study Everyday "Civic Talk" and Proto-political Engag...Tuukka Ylä-Anttila
We present a two-step topic modeling method of analysing political articulations in everyday proto-political "civic talk" on online social media and interpreting them in terms of cultural and political sociology.
Using Twitter as a Postgraduate ResearcherSimon Bishop
Second version of my talk. I tried to make it more focused and a better introduction. As ever, cute pictures need no explanation.
As for Up - try explaining the plot of it to someone who hasn't seen it... ridiculous, isn't it? There's no way to sell it that way, they have to see it. In the same way, to describe how Twitter works gives no indication as to its functionality. You have to play with it and learn by experience.
Slides from a practical workshop on gathering customer insights from social media using Social Network Analysis (SNA) with NodeXL and Twitter. SNA allows you to gain insight from thousands of tweets and messages on a range of topics for marketing research or academic use. NodeXL reports can be used for measuring and monitoring an organisation’s own performance as well as a competitors´ performance. At the highest level, a SNA approach allows social media managers to recognize what their audience looks like.
The document summarizes the findings of a study on the impact of digitized scholarly resources. It describes various quantitative and qualitative methods used in the study, including webometrics, analytics, log file analysis, interviews, focus groups, and surveys. The study analyzed five digitization projects and found they had positive impacts like improving research and enabling new types of quantitative analysis. Usage varied by project, with some seeing more impact through teaching resources while others saw more impact through computational analysis of materials.
The document provides information about Section A of the exam, which focuses on newspapers and online/social media. It will be worth 45 marks and include an in-depth study of both print newspapers and their online presence as well as social media feeds. Questions will focus on media industries, audiences, and language and representation. Students are expected to understand how different media construct representations and viewpoints, contextualize codes and conventions of broadsheet and tabloid newspapers both online and offline, and consider relevant academic theories.
This presentation explains how to use newspapers and online news in research. It highlights why students should use news but why they should be careful of issues such as bias. This presentation shows the sources that are available for searching newspaper archives, both at Goldsmiths and outside, as well as where to find video news online.
Tracking Social Media Participation: New Approaches to Studying User-Gener...Axel Bruns
PhD Seminar
Thursday 29 Oct., 9.30-11 a.m.
Seminar Room, Journalism & Media Research Centre, 1-3 Eurimbla St (corner High St), Randwick
The impact of user-generated content on a variety of media industries and practices is by now well understood from a conceptual perspective (e.g. Benkler 2006; Jenkins 2006; Bruns 2008). What remains less thoroughly explored is the possibility to utilise the affordances of Web 2.0 technologies themselves to generate large datasets that can be used to track and evaluate user participation practices in order to develop a solid evidence base for further research into social media, and further development of social media projects, technologies, and policies. This presentation outlines research possibilities across a number of social media spaces, and uses the example of a current research project studying the Australian political blogosphere to explore potential methodological approaches.
Researchers need to communicate their work more broadly to have greater societal impact. There are many communication channels available, including social media, blogs, and online forums. Some best practices are maintaining a policy blog like CAP Reform that provides commentary on agricultural issues, and running an online forum like Foodlog that facilitates discussion between citizens, researchers, and policymakers. As communication norms change, journals like EuroChoices may need to reconsider their business models and explore options like open access to improve accessibility and impact.
Twitter provides a selfie of envolving languageTERMCAT
Twitter provides a wealth of data that can be used to analyze language trends in real-time. The large volume of informal tweets sent daily help document the emergence and spread of new words and expressions. By following linguistic experts and word-related hashtags on Twitter, neologisms can be identified as they enter the language and their usage over time and in different locations can be tracked. Several organizations determine high-profile "Word of the Year" choices by analyzing which terms gained most attention over the past twelve months based on Twitter and other social media discourse.
- Alessandro Gallo, a sales manager at Springer, discusses the history of scholarly publishing and Springer's role as a global publisher. He outlines Springer's extensive online journal and book collections available on SpringerLink.
- Gallo describes Springer's open access models including Open Choice, which allows authors to pay a fee to make their articles openly accessible. He notes Springer's acquisition of BioMed Central, the largest open access publisher.
- The presentation concludes with an overview of upcoming semantic linking features on SpringerLink to improve search and navigation of content.
Twitter, Public Communication and the Media Ecology: The Case of the Queensla...Axel Bruns
Presented by Axel Bruns and Jean Burgess at the ATN-DAAD workshop The World According to Twitter, Brisbane, 27 June 2011.
Part of an ongoing collaboration between the Mapping Online Publics project (http://mappingonlinepublics.net/) at the ARC Centre of Excellence for Creative Industries and Innovation, QUT, Australia (http://cci.edu.au/), and the Nachwuchsforschergruppe Wissenschaft und Internet, Universität Düsseldorf, Germany (http://nfgwin.uni-duesseldorf.de/).
A History of Social Media Listening - Simon McDermott - AttentioInfluence People
Social media listening has grown directly with the rise of online social media platforms. Early social media included bulletin boards and Usenet groups in the 1970s-80s, with Google Groups archiving posts dating back to 1981. Blogs emerged in the late 1990s and grew exponentially. In the 2000s, companies emerged to perform social media listening for clients. As platforms like Facebook and Twitter grew in the late 2000s, the business applications of listening expanded to include influencer identification, campaign impact measurement, and reputation analysis. Today, most major brands perform social media listening, and ad hoc approaches are being replaced by continual, methodological listening and analytics to guide brand monitoring, marketing, and customer service. The future includes greater integration
Social Media and Architecture Journal ArchivesNoreen Whysel
This document outlines the social media and digital archiving activities of Architecture_MPS, a research organization that publishes an open-access journal. It describes Architecture_MPS' participation in social media platforms like Twitter, Facebook, and Pinterest to promote its journal articles, conferences, and other resources. The document also provides guidance to interns on social media best practices and analytics for measuring the effectiveness of Architecture_MPS' social media campaigns.
This document outlines the social media and digital archiving activities of Architecture_MPS, a research organization that publishes an open-access journal. It describes Architecture_MPS' participation in social media platforms like Twitter, Facebook, and Pinterest to promote its journal articles, conferences, and other resources. The document also provides guidance to interns on social media best practices and analytics to measure engagement. It discusses challenges in utilizing social media and opportunities to expand the audience and conversation around architecture, media, politics and society.
Similar to Connecting political data to media data (20)
Creating and Analysing Linked Open Data for the EU ParliamentLaura Hollink
The document discusses creating linked open data from speeches and transcripts of the European Parliament. It describes the data structure for representing a speech as linked data, including properties like the speaker, date, title, text in multiple languages, and links to other speeches. It also discusses linking the data to external sources to identify speakers and add additional context about countries.
Enriching Linked Open Data with distributional semantics to study concept driftLaura Hollink
Presentation at the "Proximity in Information Retrieval" symposium on the occasion of the PhD thesis defense of Jeroen Vuurens
April 26, 2017, Delft University of Technology
Lecture at the advanced course on Data Science of the SIKS research school, May 20, 2016, Vught, The Netherlands.
Contents
-Why do we create Linked Open Data? Example questions from the Humanities and Social Sciences
-Introduction into Linked Open Data
-Lessons learned about the creation of Linked Open Data (link discovery, knowledge representation, evaluation).
-Accessing Linked Open Data
This document provides a step-by-step demo scenario for an online news application that allows users to explore topics in newspaper articles, how frequently topics are mentioned, which topics co-occur, and images selected for articles. The application allows side-by-side comparison of topics between two newspapers or topics. Users can select a newspaper, choose a topic, and select topics, categories or image concepts to explore, seeing which topics frequently appear with the selected topic. Charts also allow selecting articles to view. The document demonstrates comparing coverage of Bernie Sanders between the New York Times and Washington Post.
Presentation at Digital Humanities Benelux 2015, Antwerp, Belgium: The possibilities and challenges of using linked data for academic research: the case of the Talk of Europe project. linked data for academic research: the case of the Talk of Europe project. Laura Hollink, Martijn Kleppe, Max Kemman, Astrid van Aggelen, Willem Robert Van Hage.
WWW2013: Web Usage Mining with Semantic AnalysisLaura Hollink
Laura Hollink, Peter Mika and Roi Blanco. Web Usage Mining with Semantic Analysis. In proceedings of the International World Wide Web Conference, Rio de Janeiro, Brazil, May 2013.
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
1. Connecting political data to media data
Laura Hollink
VU University Amsterdam
Web & Media group
ASCoR Spring Colloquium ‘Big Data at the University of Amsterdam’
February 18, 2014
5. Questions we want to answer
• Which events have attracted
a lot of media attention?
• What are the differences
between different media?
E.g. in different newspapers,
or newspapers vs. radio
bulletins?
• Has the coverage changed
over time?
• How are the events visualized
(photos, layout of newspaper,
etc.).
6.
7. Transcriptions of all 9,294
meetings of the Dutch
parliament between
1945-1995, consisting of
1,208,903 speeches.
8. Transcriptions of all 9,294
meetings of the Dutch
parliament between
1945-1995, consisting of
1,208,903 speeches.
Archives of hundreds of
newspaper with tons of
newspaper issues or 10’s
of Millions of articles
between 1618-1995.
(We only use 1945-1995)
9. Transcriptions of all 9,294
meetings of the Dutch
parliament between
1945-1995, consisting of
1,208,903 speeches.
Roughly 1.8 Million news
bulletins between
1937-1984
(We only use 1945-1995)
Archives of hundreds of
newspaper with tons of
newspaper issues or 10’s
of Millions of articles
between 1618-1995.
(We only use 1945-1995)
11. Step 1: Translate the Dutch parliamentary debates
to the standard structured web format RDF
XML by
War in
Parliament
Project
Handelingen Verenigde
Vergadering...
Debate
PartOfDebate
DebateContext
rdf:type
rdf:type
rdf:type
1945-11-20
dc:date
Dutch
dc:language
nl.proc.sgd.d.
194519460000002
hasPart
nl.proc.sgd.d.
194519460000002.1
hasPart
nl.proc.sgd.d.
194519460000002.1.1
hasText
"De voorzitter
opent de
vergadering…"
dc:publisher
dc:id
http://statengeneraaldigitaal.nl/
dc:source
nl.proc.sgd.d.19720000002
hasSubsequentPartOfDebate
hasPart
dc:source
http://resolver.politicalmashup.nl/nl.proc.sgd.d.194519460000002
"Mijnheer de
Voorzitter, de
Commissie
van …"
member_of
_parliament
Speech
nl.proc.sgd.d.
194519460000002.2
hasSpokenText
hasRole
rdf:type
rdf:type
http://resolver.kb.nl/resolve?urn=sgd:mpeg21:19451946:0000002:pdf
Joannes Antonius James
Politician
foaf:firstName
Barge
foaf:lastName
nl.proc.sgd.d.
194519460000002.1.2
sem:hasActor
hasSpeaker
Speaker_0006
4
rdfs:label
Barge
dc:source
coveredIn
http://resolver.kb.nl/resolve?urn=ddd:011198136:mpeg21:a0525:ocr
hasSubsequentSpeech
http://resolver.politicalmashup.nl/nl.m.00064
hasParty
nl.proc.sgd.d.
194519460000002.1.3
Party
Katholieke Volkspartij
rdf:type
hasFullName
Party_kvp
hasAcronym
KVP
12. Modeling the debates as events
• An event has a date, a
location, actors, and
possibly sub-events.
• We build on the Simple
Event Model (SEM).
• links to the original sources
• reusing existing
vocabularies
Handelingen Verenigde
Vergadering...
Debate
dc:title
1945-11-20
rdf:type
dc:date
Dutch
dc:language
nl.proc.sgd.d.
194519460000002
dc:publisher
dc:id
http://statengeneraaldigitaal.nl/
dc:source
nl.proc.sgd.d.19720000002
dc:source
http://resolver.politicalmashup.nl/nl.proc.sgd.d.194519460000002
http://resolver.kb.nl/resolve?urn=sgd:mpeg21:19451946:0000002:pdf
14. "Mijnheer de
Voorzitter, de
Commissie
van …"
Speech
hasSpokenText
rdf:type
member_of
_parliament
Politician
Joannes Antonius James
hasRole
rdf:type
foaf:firstName
Barge
foaf:lastName
nl.proc.sgd.d.
194519460000002.1.2
sem:hasActor
coveredIn
hasSpeaker
Speaker_0006
4
rdfs:label
Barge
hasParty
Party
http://resolver.kb.nl/resolve?urn=ddd:011198136:mpeg21:a0525:ocr
Katholieke Volkspartij
rdf:type
hasFullName
Party_kvp
• the different roles and parties
that a speaker can have in his/
her career.
hasAcronym
KVP
15. Step 2: Linking speeches in the debate to the
newspaper articles that cover them
We created a linking method to deal with our two challenges:
1.How to link documents that are so different in nature?
2. Can we use the structure of the debates: people, chronologic
order of speeches, introductions to each new topic, etc?
Name of
speaker
Date of
debate
Search
newspaper
archive
Candidate
articles
Rank
candidate
articles
Debates
Detect
topics in
speeches
Topics
Create
queries
Detect
Named
Entities in
speeches
Named
Entities
Queries
Links
between
speeches
and articles
16. Step 2: Linking speeches in the debate to the
newspaper articles that cover them
Intuition 1: The name of the speaker should
appear in the article and the article should
be published within a week of the debate
Name of
speaker
Date of
debate
Search
newspaper
archive
Candidate
articles
Rank
candidate
articles
Debates
Detect
topics in
speeches
Topics
Create
queries
Detect
Named
Entities in
speeches
Named
Entities
Queries
Links
between
speeches
and articles
17. Step 2: Linking speeches in the debate to the
newspaper articles that cover them
Intuition 1: The name of the speaker should
appear in the article and the article should
be published within a week of the debate
Name of
speaker
Date of
debate
Search
newspaper
archive
Candidate
articles
Rank
candidate
articles
Debates
Detect
topics in
speeches
Topics
Create
queries
Detect
Named
Entities in
speeches
Named
Entities
Links
between
speeches
and articles
Queries
Intuition 2: the more the article and the
speech overlap in terms of topics and
named entities, the more they are related.
18. Evaluation: what do we use to rank the candidate
articles?
• Experiment on 150 <newspaper article, speech in debate> pairs, 2 raters, K
= 0.5
• Compare text of candidate articles to:
• Setting 1: Named Entities in speech
• Setting 2: Named Entities + Topics in speech
• Setting 3: Named Entities + Topics in speech and larger part-of-debate
Score
Setting 1 Setting 2 Setting 3
I don’t know
0.14
0.15
0.08
0 - unrelated
0.38
0.23
0.12
1- related
0.29
0.36
0.36
2- explicit mention of the debate 0.19
0.26
0.44
1+2
0.62
0.80
0.48
19. Results
• An open data set of Dutch parliamentary debates,
• with almost 3 Million
links between 450.000 speeches and URL’s of 1.5
Million news paper articles and radio bulletins at the National Library.
• accessible though a Web demonstrator and through a SPARQL endpoint.
27. SPARQL endpoint
• A service to query a knowledge
base using the SPARQL query
language.
“All speeches with more
than 60 associated news
items.”
SELECT ?speech ?no_newsitems {{
SELECT ?speech (COUNT(?news) AS ?no_news_items)
WHERE{
?speech <http://purl.org/linkedpolitics/nl/polivoc#coveredAt> ?news .
}
GROUP BY ?speech }
FILTER (?no_news_items > 60) }
28.
29.
30.
31.
32. Reflection: to what extend can we answer these
questions?
• Which events have attracted
a lot of media attention?
• What are the differences
between different media?
E.g. in different newspapers,
or newspapers vs. radio
bulletins?
• Has the coverage changed
over time?
• How are the events visualized
(photos, layout of newspaper,
etc.).
33. Future work
• More types of links
• From just “coveredIn” to “quotedIn”, “coveredIn”, “backgroundOf”
“talksAbout”
• More types of media
• More types of (political) events.
34. Project ‘Talk of Europe / Traveling Clarin Campus’
2014-2015
Funded by CLARIN-ERIC
From left to right: Max Kemman, Marnix van Berchum, Laura Hollink, Astrid van Aggelen, Steven Krauwer,
Henri Beunders. (Unfortunately, Martijn Kleppe and Johan Oomen were not present to join the group pic.)
35. Plans of ‘ToE/TTC’
1.Publish proceedings of the EU parliamentary debates in RDF
• hosted by DANS
2.Organize 3 workshops/hackathons/‘Traveling Clarin Campuses’ in which we
invite international partners to work with the data.
3.In collaboration with international partners:
• enrich with annotations, e.g. topics, structured data about people, parties,
etc.
• link to national datasets, e.g. media or national parliaments