Search logs from user interactions with image archives can be analyzed and utilized in three ways:
1. To understand user search behavior and how professional users search differently than average users.
2. As training data to automatically annotate images with concepts using similar queries and clicked images, though reliability varies by concept.
3. As additional positive training samples to improve automated image classification systems, especially when combined with manually annotated samples.
We have built an online Movie Recommender System which is based on the analysis of users' ratings history to several movies and their demographic information. We used data from Movielens website. Collaborative filtering and matrix factorization techniques have been used for the implementation. The end result is a web application where a user is recommended with top 20 movies.
Codebase: http://goo.gl/nM7RMy
Demo Video: http://goo.gl/VgZ2uI
[RIIT 2017] Identifying Grey Sheep Users By The Distribution of User Similari...YONG ZHENG
Yong Zheng, Mayur Agnani, Mili Singh. “Identifying Grey Sheep Users By The Distribution of User Similarities In Collaborative Filtering”. Proceedings of The 6th ACM Conference on Research in Information Technology (RIIT), Rochester, NY, USA, October, 2017
This slide deck provides a survey of research in the field of coreference resolution. We survey 10 significant research papers and also provide a detailed description of the problem and also suggest future research directions.
Email Classification - Why Should it Matter to You?Sherpa Software
In this white paper, learn the basics of email classification, what it is, why it could assist your overall email management strategy and learn how to accomplish it.
Download Free Trial - http://bit.ly/vrIxKv
Get a Quick Quote - http://bit.ly/tw8pi3
Contact Us Now - http://bit.ly/sz9x5r
Exploiting User Comments for Audio-visual Content Indexing and Retrieval (ECI...Carsten Eickhoff
State-of-the-art content sharing platforms often require users to assign tags to pieces of media in order to make them easily retrievable. Since this task is sometimes perceived as tedious or boring, annotations can be sparse. Commenting on the other hand is a frequently used means of expressing user opinion towards shared media items. We propose the use of time series analyses in order to infer potential tags and indexing terms for audio-visual content from user comments. In this way, we mitigate the vocabulary gap between queries and document descriptors. Additionally, we show how large-scale encyclopedias such as Wikipedia can aid the task of tag prediction by serving as surrogates for high-coverage natural language vocabulary lists. Our evaluation is conducted on a corpus of several million real-world user comments from the popular video sharing platform YouTube, and demonstrates significant improvements in retrieval performance.
This work together with Wen Li and Arjen P. de Vries has been accepted for full oral presentation at the 35th European Conference on Information Retrieval (ECIR) in Moscow, Russia. The full version of the article is available at: http://link.springer.com/chapter/10.1007/978-3-642-36973-5_4
Better Contextual Suggestions by Applying Domain KnowledgeArjen de Vries
A talk summarizing the main lessons from the CWI participation in the 2014 TREC Contextual Suggestions track. If you want to suggest tourist locations, use tourist sources. If you want reproduceable research results, map these into Clueweb first.
We have built an online Movie Recommender System which is based on the analysis of users' ratings history to several movies and their demographic information. We used data from Movielens website. Collaborative filtering and matrix factorization techniques have been used for the implementation. The end result is a web application where a user is recommended with top 20 movies.
Codebase: http://goo.gl/nM7RMy
Demo Video: http://goo.gl/VgZ2uI
[RIIT 2017] Identifying Grey Sheep Users By The Distribution of User Similari...YONG ZHENG
Yong Zheng, Mayur Agnani, Mili Singh. “Identifying Grey Sheep Users By The Distribution of User Similarities In Collaborative Filtering”. Proceedings of The 6th ACM Conference on Research in Information Technology (RIIT), Rochester, NY, USA, October, 2017
This slide deck provides a survey of research in the field of coreference resolution. We survey 10 significant research papers and also provide a detailed description of the problem and also suggest future research directions.
Email Classification - Why Should it Matter to You?Sherpa Software
In this white paper, learn the basics of email classification, what it is, why it could assist your overall email management strategy and learn how to accomplish it.
Download Free Trial - http://bit.ly/vrIxKv
Get a Quick Quote - http://bit.ly/tw8pi3
Contact Us Now - http://bit.ly/sz9x5r
Exploiting User Comments for Audio-visual Content Indexing and Retrieval (ECI...Carsten Eickhoff
State-of-the-art content sharing platforms often require users to assign tags to pieces of media in order to make them easily retrievable. Since this task is sometimes perceived as tedious or boring, annotations can be sparse. Commenting on the other hand is a frequently used means of expressing user opinion towards shared media items. We propose the use of time series analyses in order to infer potential tags and indexing terms for audio-visual content from user comments. In this way, we mitigate the vocabulary gap between queries and document descriptors. Additionally, we show how large-scale encyclopedias such as Wikipedia can aid the task of tag prediction by serving as surrogates for high-coverage natural language vocabulary lists. Our evaluation is conducted on a corpus of several million real-world user comments from the popular video sharing platform YouTube, and demonstrates significant improvements in retrieval performance.
This work together with Wen Li and Arjen P. de Vries has been accepted for full oral presentation at the 35th European Conference on Information Retrieval (ECIR) in Moscow, Russia. The full version of the article is available at: http://link.springer.com/chapter/10.1007/978-3-642-36973-5_4
Better Contextual Suggestions by Applying Domain KnowledgeArjen de Vries
A talk summarizing the main lessons from the CWI participation in the 2014 TREC Contextual Suggestions track. If you want to suggest tourist locations, use tourist sources. If you want reproduceable research results, map these into Clueweb first.
Looking beyond plain text for document representation in the enterpriseArjen de Vries
In many real life scenarios, searching for information is not the user's end goal. In this presentation I look into the specific example of corporate strategy and business development in a university setting.
In today's academic institutions, strategic questions are those that relate to dependency on funding instruments, the public private partnerships that exist (and those that should be extended!), and the match between topic areas addressed by the research staff and those claimed important by policy makers. The professional search tasks encountered to answer questions in this domain are usually addressed by business intelligence (BI) tools, and not by search engines. However, professionals are known to be busy people inspired by their own research interests, and not particularly fond of keeping the
customer relationship management (CRM) or knowledge management systems up to date for the organisation's strategic interest. This then results in incomplete and inaccurate data.
Instead of requiring research staff (or their administrative support) to provide this management information, I will illustrate by example how the desired information usually exists already in the documents inherent to the academic work process. Information retrieval could thus play an important role in the computer systems that support the business analytics involved, and could significantly improve the coverage of entities of interest - i.e., to reduce the effort involved in achieving good recall in business analytics. The ranking functionality over the enterprise's (textual) content should however not be an isolated component. Our example setting integrates the information derived from research proposals, research publications and the financial systems, providing an excellent motivation for a more unified approach to structured and unstructured data.
Opening statement at the "Looking forward" panel at the 25 years of TREC celebration event, Nov 15th, 2016.
Webcast to appear within a week: https://www.nist.gov/news-events/events/2016/11/webcast-text-retrieval-conference
Models for Information Retrieval and RecommendationArjen de Vries
Online information services personalize the user experience by applying recommendation systems to identify the information that is most relevant to the user. The question how to estimate relevance has been the core concept in the field of information retrieval for many years. Not so surprisingly then, it turns out that the methods used in online recommendation systems are closely related to the models developed in the information retrieval area. In this lecture, I present a unified approach to information retrieval and collaborative filtering, and demonstrate how this let’s us turn a standard information retrieval system into a state-of-the-art recommendation system.
Huygens colloquium at Radboud University Science Faculty.
Effective web search engines (and the commercial success of a few internet giants) depend upon the data collected from the online seeking behaviour of huge numbers of users. Put differently, the high quality search results we accept for granted every day come at the price of reduced privacy.
A personal search engine would not only search the web, but also rich personal data including email, browsing history, documents read and contents of the user’s home directory. Results with so-called "slow search" indicate that the user experience can be improved significantly when the search engine gains access to additional data. However, will we be prepared to give up even more of our privacy, and eventually be prepared to give up control over all that personal information?
My proposal is to mitigate these concerns by developing a new architecture for web search, in which users control the trade-off between search result quality and the privacy risk inherent to sharing usage logs. Under this design, all data of the “personal search engine” (PSE) (web and usage data) resides in its owner’s personal digital infrastructure.
Two challenges need to be overcome to turn this into a viable alternative. Can we compensate for the loss of information about searches of large numbers of users? And, can we maintain an up-to-date index in a cost-effective manner? As a solution, I propose to organise personal search engines in a decentralised social network. This serves two goals: the index can be kept up-to-date collaboratively, and usage data may be traded with peers.
Query by Example of Speaker Audio Signals using Power Spectrum and MFCCsIJECEIAES
Search engine is the popular term for an information retrieval (IR) system. Typically, search engine can be based on full-text indexing. Changing the presentation from the text data to multimedia data types make an information retrieval process more complex such as a retrieval of image or sounds in large databases. This paper introduces the use of language and text independent speech as input queries in a large sound database by using Speaker identification algorithm. The method consists of 2 main processing first steps, we separate vocal and non-vocal identification after that vocal be used to speaker identification for audio query by speaker voice. For the speaker identification and audio query by process, we estimate the similarity of the example signal and the samples in the queried database by calculating the Euclidian distance between the Mel frequency cepstral coefficients (MFCC) and Energy spectrum of acoustic features. The simulations show that the good performance with a sustainable computational cost and obtained the average accuracy rate more than 90%.
Improving Semantic Search Using Query Log AnalysisStuart Wrigley
Despite the attention Semantic Search is continuously gaining, several challenges affecting tool performance and user experience remain unsolved. Among these are: matching user terms with the searchspace, adopting view-based interfaces in the Open Web as well as supporting users while building their queries. This paper proposes an approach to move a step forward towards tackling these challenges by creating models of usage of Linked Data concepts and properties extracted from semantic query logs as a source of collaborative knowledge. We use two sets of query logs from the USEWOD workshops to create our models and show the potential of using them in the mentioned areas.
The diversity and complexity of contents available on the web have dramatically increased in recent years. Multimedia content such as images, videos, maps, voice recordings has been published more often than before. Document genres have also been diversified, for instance, news, blogs, FAQs, wiki. These diversified information sources are often dealt with in a separated way. For example, in web search, users have to switch between search verticals to access different sources. Recently, there has been a growing interest in finding effective ways to aggregate these information sources so that to hide the complexity of the information spaces to users searching for relevant information. For example, so-called aggregated search investigated by the major search engine companies will provide search results from several sources in a single result page. Aggregation itself is not a new paradigm; for instance, aggregate operators are common in database technology.
This talk presents the challenges faced by the like of web search engines and digital libraries in providing the means to aggregate information from several and complex information spaces in a way that helps users in their information seeking tasks. It also discusses how other disciplines including databases, artificial intelligence, and cognitive science can be brought into building effective and efficient aggregated search systems.
Multimedia content based retrieval slideshare.pptgovintech1
information retrieval for text and multimedia content has become an important research area.
Content based retrieval in multimedia is a challenging problem since multimedia data needs detailed interpretation
from pixel values. In this presentation, an overview of the content based retrieval is presented along with
the different strategies in terms of syntactic and semantic indexing for retrieval. The matching techniques
used and learning methods employed are also analyzed.
Discovering User's Topics of Interest in Recommender SystemsGabriel Moreira
This talk introduces the main techniques of Recommender Systems and Topic Modeling.
Then, we present a case of how we've combined those techniques to build Smart Canvas (www.smartcanvas.com), a service that allows people to bring, create and curate content relevant to their organization, and also helps to tear down knowledge silos.
We present some of Smart Canvas features powered by its recommender system, such as:
- Highlight relevant content, explaining to the users which of his topics of interest have generated each recommendation.
- Associate tags to users’ profiles based on topics discovered from content they have contributed. These tags become searchable, allowing users to find experts or people with specific interests.
- Recommends people with similar interests, explaining which topics brings them together.
We give a deep dive into the design of our large-scale recommendation algorithms, giving special attention to our content-based approach that uses topic modeling techniques (like LDA and NMF) to discover people’s topics of interest from unstructured text, and social-based algorithms using a graph database connecting content, people and teams around topics.
Our typical data pipeline that includes the ingestion millions of user events (using Google PubSub and BigQuery), the batch processing of the models (with PySpark, MLib, and Scikit-learn), the online recommendations (with Google App Engine, Titan Graph Database and Elasticsearch), and the data-driven evaluation of UX and algorithms through A/B testing experimentation. We also touch topics about non-functional requirements of a software-as-a-service like scalability, performance, availability, reliability and multi-tenancy and how we addressed it in a robust architecture deployed on Google Cloud Platform.
Web Archives and the dream of the Personal Search EngineArjen de Vries
Keynote at the 4th Alexandria Workshop organised by Avishek Anand and Wolfgang Nejdl, L3S, Hannover (Germany). I argue that Web Archives should act as a pivot while revisiting the idea of decentralised search.
See also http://alexandria-project.eu/events/4th-int-alexandria-workshop-19-20-october-2017/
Lecture on Information Retrieval and Social Media, given to PhD students in the User-Centred Social Media Summer School, in Duisburg, September 19, 2017.
See also https://www.ucsm.info/events/118-new-frontiers-in-social-media-research-%E2%80%93-international-summer-school-2018
Recommender systems aim to predict the content that a user would like based on observations of the online behaviour of its users. Research in the Information Access group addresses different aspects of this problem, varying from how to measure recommendation results, how recommender systems relate to information retrieval models, and how to build effective recommender systems (note: last Friday, we won the ACM RecSys 2013 News Recommender Systems challenge). We would like to develop a general methodology to diagnose weaknesses and strengths of recommender systems. In this talk, I discuss the initial results of an analysis of the core component of collaborative filtering recommenders: the similarity metric used to find the most similar users (neighbours) that will provide the basis for the recommendation to be made. The purpose is to shed light on the question why certain user similarity metrics have been found to perform better than others. We have studied statistics computed over the distance distribution in the neighbourhood as well as properties of the nearest neighbour graph. The features identified correlate strongly with measured prediction performance - however, we have not yet discovered how to deploy this knowledge to actually improve recommendations made.
Social media sites (by some referred to as the web 2.0) allow their users to interact with each other, for example in collecting and sharing so-called user-generated content - these can be just bookmarks, but also blogs, images, and videos. Social media support co-creation: processes where customers (or users, if you prefer) do not just consume but play an active role in defining and shaping the end product. Famous examples include Six Degrees, LiveJournal, Digg, Epinions, Myspace, Flickr, YouTube, Linked-in, and Pinterest. Of course, today's internet giants Facebook and Twitter are key new developments. Finally, Wikipedia should not be overlooked - a major resource in many language technologies including information retrieval!
The second part of the lecture looks into the opportunities for information retrieval research. Social media platforms tend to provide access to user profiles, connections between users, the content these users publish or share, and how they react to each other's content through commenting and rating. Also, the large majority of social media platforms allow their users to categorize content by means of tags (or, in direct communication, through hash-tags), resulting in collaborative ways of information organization known as folksonomies. However, these social media also form a challenge for information retrieval research: the many platforms vary in functionalities, and we have only very little understanding of clearly desirable features like combining tag usage and ratings in content recommendation! A unifying approach based on random walks will be discussed to illustrate how we can answer some of these questions [1], but clearly the area has ample opportunity to leave your own marks.
In the final part of the lecture I will briefly touch upon an even wider range of opportunities, where data derived from social media form a key component to enable new research and insights. I will review a few important results from research centered on Wikipedia, facebook and twitter data, as well as a diverse range of new information sources including the geo- and temporal information derived from images and tweets, product reviews and comments on youtube videos, and how url shorteners may give a view on what is popular on the web.
[1] Maarten Clements, Arjen P. De Vries, and Marcel J. T. Reinders. 2010. The task-dependent effect of tags and ratings on social media access. ACM Trans. Inf. Syst. 28, 4, Article 21 (November 2010), 42 pages. http://doi.acm.org/10.1145/1852102.1852107
Recommendation and Information Retrieval: Two Sides of the Same Coin?Arjen de Vries
Status update on our current understanding of how collaborative filtering relates far more closely to information retrieval than usually thought. Includes work by Jun Wang and Alejandro Bellogín. This presentation has been given at the Siks PhD student course on computational intelligence, May 24th, 2013
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
How to Split Bills in the Odoo 17 POS ModuleCeline George
Bills have a main role in point of sale procedure. It will help to track sales, handling payments and giving receipts to customers. Bill splitting also has an important role in POS. For example, If some friends come together for dinner and if they want to divide the bill then it is possible by POS bill splitting. This slide will show how to split bills in odoo 17 POS.
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptxEduSkills OECD
Andreas Schleicher presents at the OECD webinar ‘Digital devices in schools: detrimental distraction or secret to success?’ on 27 May 2024. The presentation was based on findings from PISA 2022 results and the webinar helped launch the PISA in Focus ‘Managing screen time: How to protect and equip students against distraction’ https://www.oecd-ilibrary.org/education/managing-screen-time_7c225af4-en and the OECD Education Policy Perspective ‘Students, digital devices and success’ can be found here - https://oe.cd/il/5yV
This is a presentation by Dada Robert in a Your Skill Boost masterclass organised by the Excellence Foundation for South Sudan (EFSS) on Saturday, the 25th and Sunday, the 26th of May 2024.
He discussed the concept of quality improvement, emphasizing its applicability to various aspects of life, including personal, project, and program improvements. He defined quality as doing the right thing at the right time in the right way to achieve the best possible results and discussed the concept of the "gap" between what we know and what we do, and how this gap represents the areas we need to improve. He explained the scientific approach to quality improvement, which involves systematic performance analysis, testing and learning, and implementing change ideas. He also highlighted the importance of client focus and a team approach to quality improvement.
4. (C) 2008, The New York Times Company
Anchor tekst:
“continue reading”
5. Not much text to
get you here...
A fan’s hyves page:
Kyteman's HipHop Orchestra:
www.kyteman.com
Ticket Sales Luxor theatre:
May 22nd - Kyteman's hiphop Orchestra - www.kyteman.com
Kluun.nl:
De site van Kyteman
Blog Rockin’ Beats:
The 21-year-old Kyteman
(trumpet player, composer and
Producer Colin Benders),
has worked for 3 years on
his debute:
the Hermit sessions.
Jazzenzo:
... a performance by the popular
Kyteman’s Hiphop Orkest
6. ‘Co-creation’
Social Media:
Consumer becomes a co-creator
‘Data consumption’ traces
In essence: many new sources to
play the role of anchor text!
7. Tweets about blip.tv
E.g.: http://blip.tv/file/2168377
Amazing
Watching “World’s most realistic 3D city
models?”
Google Earth/Maps killer
Ludvig Emgard shows how maps/satellite pics
on web is done (learn Google and MS!)
and ~120 more Tweets
10. Types of feedback
Explicit user feedback
Images/videos marked as relevant/non-relevant
Selected keywords that are added to the query
Selected concepts that are added to the query
Implicit user feedback
Clicking on retrieved images/videos (click-through
data)
Bookmarking or sharing an image/video
Downloading/buying an image/video
11. Who interact with the data?
Interactive relevance feedback
Current user in current search
Personalisation
Current user in logged past searches
Context adaptation
Users similar to current user in logged past
searches
Collective knowledge
All users in logged past searches
12. Applications exploiting feedback
Given a query, rank all
images/videos based on past users
feedback
Given an image/video, rank all
images/videos based on past users
feedback
13. Applications exploiting feedback
Interactive relevance feedback
Modify query and re-rank, based on current
user's explicit feedback (and current ranking)
Blind relevance feedback
Modify query and re-rank, based on feedback
by past users and current ranking
14. Applications exploiting feedback
Query suggestion
Recommend keywords/concepts to support
users in interactive query modification
(refinement or expansion)
15.
16. ‘Police Sting’
Sting performs with The Police
‘Elton Diana’
Sting attends Versace memorial
service
‘Led Zeppelin’
Sting performs at Led Zeppelin concert
17. Exploiting User Logs
(FP6 Vitalas T4.2)
Aim
Understand the information-searching process
of professional users of a picture portal
Method
Building in collaboration with Belga an
increasingly large dataset that contains the
log of Belga's users' search interactions
Processing, analysing, and investigating the
use of this collective knowledge stored in
search logs in a variety of tasks
18. Search logs
Search logs in Vitalas
Searches performed by users through Belga's web
interface from 22/06/2007 to 12/10/2007 (101 days)
402,388 tuples <date,time,userid,action>
"SEARCH_PICTURES" (138,275) | "SHOW_PHOTO"
(192,168) | "DOWNLOAD_PICTURE" (38,070) |
"BROWSE_GALLERIES" (8,878) | "SHOW_GALLERY"
(24,352) | "CONNECT_IMAGE_FORUM" (645)
17,861 unique (‘lightly normalised’) queries
96,420 clicked images
Web image search (Craswell and Szummer,
2007):
Pruned graph has 1.1 million edges, 505,000 URLs and
202,000 queries
21. What could we learn?
Goals
What do users search for?
User context
How do professionals search image archives,
when compared to the average user?
Query modifications
How do users reformulate their queries within
a search session
28. Semantic analysis
Most studies investigate the search
logs at the syntactic (term-based)
level
Our idea: map the term occurrences
into linked open data (LOD)
29. Semantic Log Analysis
Method:
Map queries into linked data cloud, find 'abstract'
patterns, and re-use those for query suggestion, e.g.:
A and B play-soccer-in-team X
A is-spouse-of B
Advantages:
Reduces sparseness of the raw search log data
Provides higher level insights in the data
Right mix of statistics and semantics?
Overcomes the query drift problem of thesaurus-based
query expansion
39. Implications
Guide the selection of
ontologies/lexicons/etc. most suited
for your user population
Distinguish between successful and
unsuccessful queries when making
search suggestions
Improve session boundary detection
40. Finally… a ‘wild idea’
Image data is seldomly annoted
adequately
i.e., adequately to support search
Automatic image annotation or
‘concept detection’
Supervised machine learning
Requires labelled samples as training data; a
laborious and expensive task
41. FP6 Vitalas IP
Phase 1 – collect training data
Select ~500 concepten with collection owner
Manually select ~1000 positive and negative
examples for each concept
42. How to obtain training data?
Can we use click-through data
instead of manually labelled
samples?
Advantages:
Large quantities, no user intervention, collective
assessments
Disadvantages:
Noisy & sparse
Queries not based on strict visual criteria
43. Automatic Image Annotation
Research questions:
How to annotate images with concepts using
click-through data?
How reliable are click-through data based
annotations?
What is the effectiveness of these annotations
as training samples for concept classifiers?
44. Manual annotations
annotations per concept positive samples negative samples
MEAN 1020.02 89.44 930.57
MEDIAN 998 30 970
STDEV 164.64 132.84 186.21
46. 1. How to annotate? (1/4)
Use the queries for which images were clicked
Challenges:
Inherent noise: gap between queries/captions and concepts
queries describe the content+context of images to be retrieved
clicked images retrieved using their captions: content+context
concept-based annotations: based on visual content-only criteria
Sparsity: only cover part of the collection previously accessed
Mismatch between terms in concept descriptions and queries
47. How to annotate (2/4)
Basic ‘global’ method:
Given the keywords of a query Q
Find the query Q' in search logs that is most
textually similar to Q
Find the images I clicked for Q'
Find the queries Q'' for which these images
have been clicked
Rank the queries Q'' based on the number of
images clicked for them
48. How to annotate (3/4)
Exact: images clicked for queries exactly matching
the concept name
Example: 'traffic' -> 'traffic jam', 'E40', 'vacances', 'transport‘
Search log-based image representations:
Images represented by all queries for which they have been
clicked
Retrieval based on language models (smoothing, stemming)
Example: 'traffic' -> 'infrabel', 'deutsche bahn', 'traffic lights‘
Random walks over the click graph
Example: 'hurricane' -> 'dean', 'mexico', 'dean haiti', 'dean
mexico'
49. How to annotate (4/4)
Local method:
given the keywords of a query Q and its top
ranked images
Find the queries Q'' for which these images have
been clicked
Rank the queries Q'' based on the number of
images clicked for them
50. •Compare agreement of click-through-based annotations to manual ones,
examining the 111 VITALAS concepts with at least 10 images (for at least
one of the methods) in the overlap of clicked and manually annotated images
• Levels of agreement vary greatly across concepts
• 20% of concepts per method reach agreement of at least 0.8
What type of concepts can be reliably
annotated using clickthrough data?
• defined categories? not informative
activities, animals, events, graphics,
people,image_theme, objects,
setting/scene/site
Possible future research on types of concepts
• named entities?
• specific vs. broad?
•
2. Reliability
51. Train the classifiers for each of 25
concepts
positive samples:
images selected by each method
negative samples:
selected by random sampling the 100k set
exclude those already selected as positive samples
low-level visual features FW
:
texture description
integrated Weibull distribution extracted from overlapping image
regions
low-level textual features FT
:
a vocabulary of most frequent terms in captions is built for each
concept
compare each image caption is against each concept vocabulary
build a frequency-histogram for each concept
SVM classifiers with RBF kernel (and cross
3. Effectiveness (1/3)
52. 3. Effectiviness study (2/3)
•Experiment 1 (visual features):
–training: search-log based annotations
–test set for each concept: manual annotations (~1000 images)
–feasibility study: in most cases, AP considerably higher than the prior
3. Effectiveness (2/3)
53. •Experiments 2,3,4 (visual or textual features):
–Experiment 2 training: search-log based annotations
–Experiment 3 training: manual + search-log based annotations
–Experiment 4 training: manual annotations
–common test set: 56,605 images (subset of the 100,000 collection)
–contribution of search-log based annotations to training is positive
–particularly in combination with manual annotations
3. Effectiviness (3/3)
54. manually annotated positive samples search log based annotated positive samples
test set results
View results at:
http://olympus.ee.auth.gr/~diou/searchlogs/
Example: Soccer
57. Diversity from User Logs
Present different query variants'
clicked images in clustered view
Merge different query variants'
clicked images in a round robin
fashion into one list (CLEF)
66. ImageCLEF Findings
Many queries (>20%) without
clicked images
Corpus and available logs originated from
different time frame
67. Best results combine text search in
metadata with image click data for
topic title and each of the cluster
titles
Using query variants derived from
the logs increases recall with 50-
100%
However, also topic drift; reduced early
precision
ImageCLEF Findings
Editor's Notes
&lt;number&gt;
&lt;number&gt;
&lt;number&gt;
&lt;number&gt;
&lt;number&gt;
&lt;number&gt;
&lt;number&gt;
Explicit/implicit refers to whether a user&apos;s action translates into explicit/implicit evidence of relevance. Explicit evidence of relevance is when the user says a document is relevant. Implicit evidence of relevance is when the user may not say that something is relevant, but his actions/behaviour (i.e, clicking, looking at, etc.) indicate that to some degree the user finds it relevant. When a Belga user buys an image, he may not have said explicitly that it is relevant, but this action is very close to that.
Belga&apos;s logs currently include only type implicit evidence, whereas VITALAS logs will include both implicit and explicit.
Maybe “personalisation” should also become context adaptation so as to be consistent with IP3 and not get confused with the personalisation in WP5.
&lt;number&gt;
Explicit/implicit refers to whether a user&apos;s action translates into explicit/implicit evidence of relevance. Explicit evidence of relevance is when the user says a document is relevant. Implicit evidence of relevance is when the user may not say that something is relevant, but his actions/behaviour (i.e, clicking, looking at, etc.) indicate that to some degree the user finds it relevant. When a Belga user buys an image, he may not have said explicitly that it is relevant, but this action is very close to that.
Belga&apos;s logs currently include only type implicit evidence, whereas VITALAS logs will include both implicit and explicit.
Maybe “personalisation” should also become context adaptation so as to be consistent with IP3 and not get confused with the personalisation in WP5.