Navidgator - Similarity Based Browsing for Image & Video DatabasesDamian Borth
Talk at the KI 2008: Advances in Artificial Intelligence in Kaiserslautern, Germany
Link to original publication: http://madm.dfki.de/publication&pubid=3834
Relevance Filtering meets Active Learning: Improving Web-based Concept Detec...Damian Borth
Talk at the ACM Int. Conference on Multimedia Information Retrieval (MIR) in Philadelphia, USA
Link to original publication: http://madm.dfki.de/publication&pubid=4535
This document describes the medical case of a woman who received a kidney transplant as a teenager and went on to have four pregnancies, despite being warned of the risks due to her transplant and history of urinary tract infections. Her kidney function declined rapidly after her fourth pregnancy, when it was discovered that her fourth child had kidney reflux. She was eventually placed back on dialysis and listed for another transplant due to complications of allograft nephropathy and recurrent infections.
Introduces "Slug" a web crawler (or "Scutter") designed for harvesting semantic web content. Implemented in Java using the Jena API, Slug provides a configurable, modular framework that allows a great degree of flexibility in configuring the retrieval, processing and storage of harvested content. The framework provides an RDF vocabulary for describing crawler configurations and collects metadata concerning crawling activity. Crawler metadata allows for reporting and analysis of crawling progress, as well as more efficient retrieval through the storage of HTTP caching data.
This document summarizes a study of CEO succession events among the largest 100 U.S. corporations between 2005-2015. The study analyzed executives who were passed over for the CEO role ("succession losers") and their subsequent careers. It found that 74% of passed over executives left their companies, with 30% eventually becoming CEOs elsewhere. However, companies led by succession losers saw average stock price declines of 13% over 3 years, compared to gains for companies whose CEO selections remained unchanged. The findings suggest that boards generally identify the most qualified CEO candidates, though differences between internal and external hires complicate comparisons.
The TIB|AV Portal : OSGeo conference videos as a resource for scientific res...Peter Löwe
The document discusses using the TIB|AV Portal to provide long-term preservation and access to open source geospatial (OSGeo) conference videos through assigning digital object identifiers (DOIs). It notes the portal currently hosts over 100 hours of OSGeo video content and is actively collecting more. Assigning DOIs to videos in the portal allows for citation, quotation of video segments, and integration into the linked open data framework to enable new ways of mining and analyzing video content. The goal is to better credit video producers, provide improved search capabilities for consumers, and ensure the scientific value of these resources is preserved over time.
Outlines an NLS pilot project with video sharing site YouTube and associated metadata issues. Part of the Cataloguing and Indexing Group in Scotland (CIGS) seminar "Toto, I've got a feeling we're not in Kansas anymore": metadata issues and Web2.0 services.
Navidgator - Similarity Based Browsing for Image & Video DatabasesDamian Borth
Talk at the KI 2008: Advances in Artificial Intelligence in Kaiserslautern, Germany
Link to original publication: http://madm.dfki.de/publication&pubid=3834
Relevance Filtering meets Active Learning: Improving Web-based Concept Detec...Damian Borth
Talk at the ACM Int. Conference on Multimedia Information Retrieval (MIR) in Philadelphia, USA
Link to original publication: http://madm.dfki.de/publication&pubid=4535
This document describes the medical case of a woman who received a kidney transplant as a teenager and went on to have four pregnancies, despite being warned of the risks due to her transplant and history of urinary tract infections. Her kidney function declined rapidly after her fourth pregnancy, when it was discovered that her fourth child had kidney reflux. She was eventually placed back on dialysis and listed for another transplant due to complications of allograft nephropathy and recurrent infections.
Introduces "Slug" a web crawler (or "Scutter") designed for harvesting semantic web content. Implemented in Java using the Jena API, Slug provides a configurable, modular framework that allows a great degree of flexibility in configuring the retrieval, processing and storage of harvested content. The framework provides an RDF vocabulary for describing crawler configurations and collects metadata concerning crawling activity. Crawler metadata allows for reporting and analysis of crawling progress, as well as more efficient retrieval through the storage of HTTP caching data.
This document summarizes a study of CEO succession events among the largest 100 U.S. corporations between 2005-2015. The study analyzed executives who were passed over for the CEO role ("succession losers") and their subsequent careers. It found that 74% of passed over executives left their companies, with 30% eventually becoming CEOs elsewhere. However, companies led by succession losers saw average stock price declines of 13% over 3 years, compared to gains for companies whose CEO selections remained unchanged. The findings suggest that boards generally identify the most qualified CEO candidates, though differences between internal and external hires complicate comparisons.
The TIB|AV Portal : OSGeo conference videos as a resource for scientific res...Peter Löwe
The document discusses using the TIB|AV Portal to provide long-term preservation and access to open source geospatial (OSGeo) conference videos through assigning digital object identifiers (DOIs). It notes the portal currently hosts over 100 hours of OSGeo video content and is actively collecting more. Assigning DOIs to videos in the portal allows for citation, quotation of video segments, and integration into the linked open data framework to enable new ways of mining and analyzing video content. The goal is to better credit video producers, provide improved search capabilities for consumers, and ensure the scientific value of these resources is preserved over time.
Outlines an NLS pilot project with video sharing site YouTube and associated metadata issues. Part of the Cataloguing and Indexing Group in Scotland (CIGS) seminar "Toto, I've got a feeling we're not in Kansas anymore": metadata issues and Web2.0 services.
This document describes tools and experiments for interactive video browsing and editing on personal computers. It proposes treating video like editable text by detecting shot boundaries and displaying a grid of key frames that can be manipulated similarly to a spreadsheet. The goal is to enable easy browsing, editing, and organization of home video archives on PCs in a more interactive way than traditional video editors. Shot detection and a flexible database are key enabling technologies. The document discusses using these tools to browse and abstract unedited home videos and publish edited or raw videos online or to discs.
The document discusses linking multimedia data on the web. It proposes applying linked data principles to multimedia content to allow fine-grained identification and description of multimedia fragments. This would enable capabilities like deep linking to specific points in videos, time-aligned annotations, and searching media using text. The document also outlines efforts by the W3C Media Fragments and Media Annotations working groups to develop standards for identifying multimedia fragments and annotating multimedia content.
This document discusses 3D technology reaching a tipping point in various industries such as television, movies, and healthcare. It provides examples of 3D technologies from 1844 to present day, including early 3D movies from the 1950s and current/future 3D movies. The document also discusses 3D television history, 3D education projects, 3D video games, 3D glasses technology including active and passive types, autostereoscopic displays, and the Nintendo 3DS. In addition, it addresses the concept of "3D Vision Syndrome" including symptoms, diagnosis, treatment, and a case example of a patient undergoing optometric vision therapy.
The 4th project meeting was held on December 15, 2009 in Munich. Davide Palmisano presented on building user profiles from social web data using a linked data approach. His presentation included collecting user activity data, reasoning over it to infer interests, and syndicating user profiles using OpenSocial. He demonstrated how profiles could be used to personalize news applications.
DISNEY DOES DATA: Data management implications of using animated video as tra...Louise Patterton
ABSTRACT The Network of Data and Information Curation Communities (NeDICC) created an animated video as study material for librarians attending RDM workshops. It was thought that the concept of a YouTube-uploaded animated video as training tool would provide a refreshing, unconventional and humorous mode of teaching concerning a subject area currently not understood my many librarians. Taking into account ease of use, affordability, customization options as well as proven track record, GoAnimate was chosen as video platform.
Data product implications for creators/curators of the video were numerous, and include intellectual property rights, sharing intent, format and restrictions, mobile device issues, preservation formats, conversion tools, and storage and backup locations. Speed of video technology evolvement, the benefits of a data management plan for the smallest of research outputs, the need for familiarization with ownership issues and IP rights when making use of web-based animation, and the omnipresent threat of format obsolescence as well as the possible future limitations of proprietary software, are some of the learnings gained.
Clipper project presentation at the Jisc Research Data Network meeting, Cambridge, 6th September 2016.
Clipper: A web annotation toolkit for research & practice with online audio visual media
Clipper is a web annotation toolkit created by a consortium including City of Glasgow College and The Open University to enable annotation and analysis of audiovisual media without copying large files. It allows users to create "virtual clips" from media sources and annotate them using text and share the clips via URIs. Clipper aims to make audiovisual media as easy to work with as text. It demonstrates potential benefits for research including open data, reproducibility, collaboration, and impact. The toolkit is built using MEAN stack technologies and aligns with emerging W3C annotation standards.
Programmers love Python because of how fast and easy it is to use. Python cuts development time in half with its simple to read syntax and easy compilation feature. Debugging your programs is a breeze in Python with its built in debugger. Using Python makes Programmers more productive and their programs ultimately better. Python is continued to be a favorite option for data scientists who use it for building and using Machine learning applications and other scientific computations.
Python runs on Windows, Linux/Unix, Mac OS and has been ported to Java and .NET virtual machines. Python is free to use, even for the commercial products, because of its OSI-approved open source license.
Python has evolved as the most preferred Language for Data Analytics and the increasing search trends on python also indicates that Python is the next "Big Thing" and a must for Professionals in the Data Analytics domain.
A growing working group is developing URIplay, a system for providing consistent identifiers and metadata for television and web content across devices and services. URIplay will use simple XML formats and APIs to describe shows, episodes, broadcasts and encodings, and allow different services and apps to reliably identify and link related content. The initial implementation is being built, and the group is finalizing the design and data model before opening it up for others to see, use and provide feedback on by the end of May.
Tape-basierter Object-Storage als S3 Speicherklasse und Cloud-Absicherungdata://disrupted®
This document discusses using tape-based storage as an S3 object storage class and for cloud backup. It describes how tape storage can provide scalable, high-throughput object storage with native S3 API access. Examples of using tape-based S3 storage include big data analytics, tiering from HDD storage, and backing up data from public cloud services like AWS S3 for added protection against risks like human errors and technology failures.
DISNEY DOES DATA: Data management implications of using animated video as tra...Louise Patterton
ABSTRACT The Network of Data and Information Curation Communities (NeDICC) created an animated video as study material for librarians attending RDM workshops. It was thought that the concept of a YouTube-uploaded animated video as training tool would provide a refreshing, unconventional and humorous mode of teaching concerning a subject area currently not understood my many librarians. Taking into account ease of use, affordability, customization options as well as proven track record, GoAnimate was chosen as video platform.
Data product implications for creators/curators of the video were numerous, and include intellectual property rights, sharing intent, format and restrictions, mobile device issues, preservation formats, conversion tools, and storage and backup locations. Speed of video technology evolvement, the benefits of a data management plan for the smallest of research outputs, the need for familiarization with ownership issues and IP rights when making use of web-based animation, and the omnipresent threat of format obsolescence as well as the possible future limitations of proprietary software, are some of the learnings gained.
There are different applications for getting the raw PCM audio on the client side. One can be client side recording, others can be for machine learning applications, call analysis, and automatic speech recognition.
In September, the W3C released the Media Stream Recording draft (https://www.w3.org/TR/mediastream-recording/).
In this session we will review the capabilities the draft provides and status of implementation in the different browsers.
Dr. Gerald Friedland's research focuses on advancing semantic analysis of multimedia content. His work aims to make multimedia data more "understandable" to computers by enabling capabilities like semantic navigation, search, automatic annotation, and inference over multimedia content. He has developed techniques for extracting objects from images and video using human interaction to provide context. His goal is to connect multimedia and semantic technologies to allow logical inference on perceptually encoded data and more natural interaction with computers.
This document discusses a metadata model for peer-to-peer media distribution developed as part of the P2P-Next project. The model defines core and optional metadata for describing media assets distributed via a peer-to-peer network. It also includes mappings to existing metadata standards. The model structures content into P2P-Next Items containing media files, metadata descriptors, and additional resources. An API based on MPEG-M allows creating and parsing the metadata descriptions.
Presentation given at the IPTC - Business Meets Technology Day 2010 held in Paris 9th March 2010.
An introduction to Semantic Web & Linked Data and its applicability to the News business.
The document outlines a mission to democratize data development through publicly available datasets. It presents the OSEMn framework for obtaining, scrubbing, exploring, modeling, and interpreting data. Examples are provided to illustrate analyzing over 1,300 NYC public datasets and providing log data analysis and recommender systems. Instructions are given on using the Mortar tool for exploring and analyzing public data sources.
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
This document describes tools and experiments for interactive video browsing and editing on personal computers. It proposes treating video like editable text by detecting shot boundaries and displaying a grid of key frames that can be manipulated similarly to a spreadsheet. The goal is to enable easy browsing, editing, and organization of home video archives on PCs in a more interactive way than traditional video editors. Shot detection and a flexible database are key enabling technologies. The document discusses using these tools to browse and abstract unedited home videos and publish edited or raw videos online or to discs.
The document discusses linking multimedia data on the web. It proposes applying linked data principles to multimedia content to allow fine-grained identification and description of multimedia fragments. This would enable capabilities like deep linking to specific points in videos, time-aligned annotations, and searching media using text. The document also outlines efforts by the W3C Media Fragments and Media Annotations working groups to develop standards for identifying multimedia fragments and annotating multimedia content.
This document discusses 3D technology reaching a tipping point in various industries such as television, movies, and healthcare. It provides examples of 3D technologies from 1844 to present day, including early 3D movies from the 1950s and current/future 3D movies. The document also discusses 3D television history, 3D education projects, 3D video games, 3D glasses technology including active and passive types, autostereoscopic displays, and the Nintendo 3DS. In addition, it addresses the concept of "3D Vision Syndrome" including symptoms, diagnosis, treatment, and a case example of a patient undergoing optometric vision therapy.
The 4th project meeting was held on December 15, 2009 in Munich. Davide Palmisano presented on building user profiles from social web data using a linked data approach. His presentation included collecting user activity data, reasoning over it to infer interests, and syndicating user profiles using OpenSocial. He demonstrated how profiles could be used to personalize news applications.
DISNEY DOES DATA: Data management implications of using animated video as tra...Louise Patterton
ABSTRACT The Network of Data and Information Curation Communities (NeDICC) created an animated video as study material for librarians attending RDM workshops. It was thought that the concept of a YouTube-uploaded animated video as training tool would provide a refreshing, unconventional and humorous mode of teaching concerning a subject area currently not understood my many librarians. Taking into account ease of use, affordability, customization options as well as proven track record, GoAnimate was chosen as video platform.
Data product implications for creators/curators of the video were numerous, and include intellectual property rights, sharing intent, format and restrictions, mobile device issues, preservation formats, conversion tools, and storage and backup locations. Speed of video technology evolvement, the benefits of a data management plan for the smallest of research outputs, the need for familiarization with ownership issues and IP rights when making use of web-based animation, and the omnipresent threat of format obsolescence as well as the possible future limitations of proprietary software, are some of the learnings gained.
Clipper project presentation at the Jisc Research Data Network meeting, Cambridge, 6th September 2016.
Clipper: A web annotation toolkit for research & practice with online audio visual media
Clipper is a web annotation toolkit created by a consortium including City of Glasgow College and The Open University to enable annotation and analysis of audiovisual media without copying large files. It allows users to create "virtual clips" from media sources and annotate them using text and share the clips via URIs. Clipper aims to make audiovisual media as easy to work with as text. It demonstrates potential benefits for research including open data, reproducibility, collaboration, and impact. The toolkit is built using MEAN stack technologies and aligns with emerging W3C annotation standards.
Programmers love Python because of how fast and easy it is to use. Python cuts development time in half with its simple to read syntax and easy compilation feature. Debugging your programs is a breeze in Python with its built in debugger. Using Python makes Programmers more productive and their programs ultimately better. Python is continued to be a favorite option for data scientists who use it for building and using Machine learning applications and other scientific computations.
Python runs on Windows, Linux/Unix, Mac OS and has been ported to Java and .NET virtual machines. Python is free to use, even for the commercial products, because of its OSI-approved open source license.
Python has evolved as the most preferred Language for Data Analytics and the increasing search trends on python also indicates that Python is the next "Big Thing" and a must for Professionals in the Data Analytics domain.
A growing working group is developing URIplay, a system for providing consistent identifiers and metadata for television and web content across devices and services. URIplay will use simple XML formats and APIs to describe shows, episodes, broadcasts and encodings, and allow different services and apps to reliably identify and link related content. The initial implementation is being built, and the group is finalizing the design and data model before opening it up for others to see, use and provide feedback on by the end of May.
Tape-basierter Object-Storage als S3 Speicherklasse und Cloud-Absicherungdata://disrupted®
This document discusses using tape-based storage as an S3 object storage class and for cloud backup. It describes how tape storage can provide scalable, high-throughput object storage with native S3 API access. Examples of using tape-based S3 storage include big data analytics, tiering from HDD storage, and backing up data from public cloud services like AWS S3 for added protection against risks like human errors and technology failures.
DISNEY DOES DATA: Data management implications of using animated video as tra...Louise Patterton
ABSTRACT The Network of Data and Information Curation Communities (NeDICC) created an animated video as study material for librarians attending RDM workshops. It was thought that the concept of a YouTube-uploaded animated video as training tool would provide a refreshing, unconventional and humorous mode of teaching concerning a subject area currently not understood my many librarians. Taking into account ease of use, affordability, customization options as well as proven track record, GoAnimate was chosen as video platform.
Data product implications for creators/curators of the video were numerous, and include intellectual property rights, sharing intent, format and restrictions, mobile device issues, preservation formats, conversion tools, and storage and backup locations. Speed of video technology evolvement, the benefits of a data management plan for the smallest of research outputs, the need for familiarization with ownership issues and IP rights when making use of web-based animation, and the omnipresent threat of format obsolescence as well as the possible future limitations of proprietary software, are some of the learnings gained.
There are different applications for getting the raw PCM audio on the client side. One can be client side recording, others can be for machine learning applications, call analysis, and automatic speech recognition.
In September, the W3C released the Media Stream Recording draft (https://www.w3.org/TR/mediastream-recording/).
In this session we will review the capabilities the draft provides and status of implementation in the different browsers.
Dr. Gerald Friedland's research focuses on advancing semantic analysis of multimedia content. His work aims to make multimedia data more "understandable" to computers by enabling capabilities like semantic navigation, search, automatic annotation, and inference over multimedia content. He has developed techniques for extracting objects from images and video using human interaction to provide context. His goal is to connect multimedia and semantic technologies to allow logical inference on perceptually encoded data and more natural interaction with computers.
This document discusses a metadata model for peer-to-peer media distribution developed as part of the P2P-Next project. The model defines core and optional metadata for describing media assets distributed via a peer-to-peer network. It also includes mappings to existing metadata standards. The model structures content into P2P-Next Items containing media files, metadata descriptors, and additional resources. An API based on MPEG-M allows creating and parsing the metadata descriptions.
Presentation given at the IPTC - Business Meets Technology Day 2010 held in Paris 9th March 2010.
An introduction to Semantic Web & Linked Data and its applicability to the News business.
The document outlines a mission to democratize data development through publicly available datasets. It presents the OSEMn framework for obtaining, scrubbing, exploring, modeling, and interpreting data. Examples are provided to illustrate analyzing over 1,300 NYC public datasets and providing log data analysis and recommender systems. Instructions are given on using the Mortar tool for exploring and analyzing public data sources.
Similar to TubeTagger – YouTube-based Concept Detection (20)
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
National Security Agency - NSA mobile device best practices
TubeTagger – YouTube-based Concept Detection
1. TubeTagger
— YouTube-based Concept Detection —
Adrian Ulges, Markus Koch
Damian Borth, Thomas M. Breuel
German Research Center for Artificial Intelligence (DFKI) &
University of Kaiserslautern
December 6 2009
D.Borth: : TubeTagger 1 December 6 2009
2. Outline
Motivation
TubeTagger System
TubeTagger Web Demo
Summary
D.Borth: : TubeTagger 2 December 6 2009
4. Data, data, data
”...TV, video on demand, Internet video, and P2P video will
account for over 91 percent of global consumer traffic by 2013...”
”Cisco Visual Networking Index”, 2009
D.Borth: : TubeTagger 3 December 6 2009
5. Data, data, data
”...TV, video on demand, Internet video, and P2P video will
account for over 91 percent of global consumer traffic by 2013...”
”Cisco Visual Networking Index”, 2009
”...more than 20 hours of video uploaded to YouTube every
minute, 1 billion views per day...”
, 2009
D.Borth: : TubeTagger 3 December 6 2009
6. Data, data, data
”...TV, video on demand, Internet video, and P2P video will
account for over 91 percent of global consumer traffic by 2013...”
”Cisco Visual Networking Index”, 2009
”...more than 20 hours of video uploaded to YouTube every
minute, 1 billion views per day...”
, 2009
Increasing importance of video live streams: Obama’s inauguration,
Michael Jackson’s memorial service, music video debuts...
, 2009
D.Borth: : TubeTagger 3 December 6 2009
9. Content-based Video Retrieval
How to search in large video databases?
b ah
ec
s ce
o cr
itriw
neve
Video Concept Detection [MediaMill, Columbia, IBM,...]
→ as key building block of CBVR
D.Borth: : TubeTagger 6 December 6 2009
10. Content-based Video Retrieval
How to search in large video databases?
b ah
ec
s ce
o cr
itriw
neve
S p rie
u evs d
Ma hn L ann
c ie e rig
Video Concept Detection [MediaMill, Columbia, IBM,...]
→ as key building block of CBVR
Supervised Machine Learning
→ need labeled training data
D.Borth: : TubeTagger 7 December 6 2009
11. Training Data
State-of-the-art Datasets
TRECVID
LSCOM
Acquisition
collaborative [Ayache07]
precise & high quality
time consuming effort
D.Borth: : TubeTagger 8 December 6 2009
12. Training Data
State-of-the-art Datasets
TRECVID
LSCOM
Acquisition
collaborative [Ayache07]
precise & high quality
time consuming effort
vocabulary size does not scale
→ from hundreds to thousands of concepts [Hauptmann07]
missing new emerging concepts of interest
→ e.g. ”Michael Jackson”, ”Obama”, ”iPod”1
1
top ranked searches 2009-Q3 by “Google Insights for Search” for web search, news search
and product search respectively
D.Borth: : TubeTagger 8 December 6 2009
14. Web-video as training data
Pros
available at large scale
high variability of data
enriched with
tags
comments
ratings
could allow automatic
concept learning
D.Borth: : TubeTagger 9 December 6 2009
15. Web-video as training data
Pros Cons
available at large scale weakly labeled
high variability of data incomplete / subjective
coarse
enriched with
tags focus of interest
comments domain change
ratings video portal as black box
could allow automatic
concept learning
D.Borth: : TubeTagger 9 December 6 2009
17. TubeTagger System Overview
Notation
t ∈ T := tags (≈concepts)
x ∈ Xdb := keyframe of a videos
P(t|x) := probability of tag t for keyframe x
D.Borth: : TubeTagger 10 December 6 2009
19. TubeTagger System Overview
Training Data Acquisition
use YouTube as primary
training data source
use tags as weak labels
pos. = tagged
neg. = all non-tagged
D.Borth: : TubeTagger 11 December 6 2009
20. TubeTagger System Overview
Training Data Acquisition Learning Pipelines
use YouTube as primary Visual Concept Learning
training data source Semantic Concept
use tags as weak labels Learning
pos. = tagged
neg. = all non-tagged
D.Borth: : TubeTagger 12 December 6 2009
26. Semantic Concept Learning
Query Formulation
concepts = textual queries
mapping of text queries to concept vocabulary
”computer” → Monitor, Windows-Desktop, iPhone
”funny” → Muppets, Cats, Commercials
learning of tag co-occurrences
D.Borth: : TubeTagger 14 December 6 2009
27. Semantic Concept Learning
Feature
bag-of-words features
t ∈ T → ht
q := hq
mapping to concepts as
weights
w (q, t) :=< ht , hq >
D.Borth: : TubeTagger 15 December 6 2009
28. Semantic Concept Learning
Feature Approach
bag-of-words features T5 = w (q, t)top5
t ∈ T → ht final fusion
q := hq P(q|x) =
mapping to concepts as
weights P w (q,t) P(t|x)
t∈T5 t∈T w (q,t)
5
w (q, t) :=< ht , hq >
D.Borth: : TubeTagger 15 December 6 2009
30. Experiments
Dataset
1200 hrs. (= 750k keyframes)
233 concepts
per concept:
150 videos for training
50 videos for testing ”soccer” ”traffic”
D.Borth: : TubeTagger 16 December 6 2009
31. Experiments
Dataset
1200 hrs. (= 750k keyframes)
233 concepts
per concept:
150 videos for training
50 videos for testing ”soccer” ”traffic”
Clips/Concept Evaluation
10 concepts
trained on N clips
tested on 200k keyframes
saturation at 100 − 150
clips/concepts for SURF+PAMIR
D.Borth: : TubeTagger 16 December 6 2009
33. Experiments
Feature & Classifier Evaluation
81 concepts feature
model SURF SIFT
video level testing (avg. fusion) SVMs 20.4 22.4
results: PAMIR 15.4 18.4
MAP: 22.4% MAXENT 14.1 15.5
MAP: 15.4% (6 times faster)
Performance Distribution
oi mi
ra
g s ce
ocr bs
u rs ua t
e t rn
a o eai - o
p rt nr m
o o
showing 78 representative concepts out of 233
D.Borth: : TubeTagger 17 December 6 2009
34. Insights in Web-based Concept Detection
Good Concepts
”boat/ship”, ”pyramids”
broad community of YouTube users
often ”interesting”, ”spectacular” ”boat ship”
Redundant Concepts
”drummer”, ”fencing”
series of clips, not sufficiently diverse
”drummer”
problem: generalization
Bad Concepts
”fence”, ”operation-room”
not regularly used as a tag
”fence”
D.Borth: : TubeTagger 18 December 6 2009
35. TubeTagger Web Demo
User Interface
1. selected a concept
2. or enter a query
3. switch between video or keyframe level
D.Borth: : TubeTagger 19 December 6 2009
36. TubeTagger Web Demo
Deep Tagging
frame-accurate concept detection beyond coarse tags
find concept (e.g. ”Christmas Tree”) within a video clip
D.Borth: : TubeTagger 20 December 6 2009
37. TubeTagger Web Demo
Text-based Search
matches queries to concepts
e.g. ”diving” → ”underwater”, ”shipwreck”, ”fish”, . . .
D.Borth: : TubeTagger 21 December 6 2009
38. Summary
TubeTagger - YouTube-based Concept Detection
utilize YouTube videos as training source for visual
concept detector learning
utilize YouTube tags for semantic model generation
web demo providing:
deep tagging
tag recommendation
text-based search
D.Borth: : TubeTagger 22 December 6 2009
39. questions?
project site: www.dfki.de/moonvid
web demo at: http://madm.dfki.de/demo/tubetagger/
D.Borth: : TubeTagger 23 December 6 2009