Usability and user satisfaction are of paramount importance
when designing interactive software solutions. Furthermore, the optimal
design can be dependent not only on the task but also on the type of
user. Evaluations can shed light on these issues; however, very few studies
have focused on assessing the usability of semantic search systems.
As semantic search becomes mainstream, there is growing need for standardised,
comprehensive evaluation frameworks. In this study, we assess
the usability and user satisfaction of dierent semantic search query input
approaches (natural language and view-based) from the perspective
of dierent user types (experts and casuals). Contrary to previous studies,
we found that casual users preferred the form-based query approach
whereas expert users found the graph-based to be the most intuitive.
Additionally, the controlled-language model oered the most support for
casual users but was perceived as restrictive by experts, thus limiting
their ability to express their information needs.
The document provides an overview and objectives of analyzing LibQUAL+® survey results. It discusses interpreting results internally by identifying areas of best and worst performance. It also covers benchmarking results externally against consortium data and longitudinally to assess the impact of changes. The document outlines quantitative analysis of survey scores and qualitative analysis of comments. It recommends using tools like LibQUAL+® Analytics and SPSS to further analyze results data.
Search Quality Evaluation: a Developer PerspectiveAndrea Gazzarini
Search quality evaluation is an ever-green topic every search engineer ordinarily struggles with. Improving the correctness and effectiveness of a search system requires a set of tools which help measuring the direction where the system is going.
The slides will focus on how a search quality evaluation tool can be seen under a practical developer perspective, how it could be used for producing a deliverable artifact and how it could be integrated within a continuous integration infrastructure.
Evaluating Semantic Search Systems to Identify Future Directions of ResearchStuart Wrigley
Recent work on searching the Semantic Web has yielded a wide range of approaches with respect to the style of input, the underlying search mechanisms and the manner in which results are presented. Each approach has an impact upon the quality of the information retrieved and the user's experience of the search process. This highlights the need for formalised and consistent evaluation to benchmark the coverage, applicability and usability of existing tools and provide indications of future directions for advancement of the state-of-the-art. In this paper, we describe a comprehensive evaluation methodology which addresses both the underlying performance and the subjective usability of a tool. We present the key outcomes of a recently completed international evaluation campaign which adopted this approach and thus identify a number of new requirements for semantic search tools from both the perspective of the underlying technology as well as the user experience.
Improving Semantic Search Using Query Log AnalysisStuart Wrigley
Despite the attention Semantic Search is continuously gaining, several challenges affecting tool performance and user experience remain unsolved. Among these are: matching user terms with the searchspace, adopting view-based interfaces in the Open Web as well as supporting users while building their queries. This paper proposes an approach to move a step forward towards tackling these challenges by creating models of usage of Linked Data concepts and properties extracted from semantic query logs as a source of collaborative knowledge. We use two sets of query logs from the USEWOD workshops to create our models and show the potential of using them in the mentioned areas.
2011 Search Query Rewrites - Synonyms & AcronymsBrian Johnson
July 27, 2011 Bay Area Search Presentation
Brian Johnson, Engineering Director, Query Services @ eBay
Query expansion is an important part of of the search recall for all search engines. In this talk I'll discuss some of the general trend driving Hadoop adoption within the Search Query Services team at eBay, and the types of algorithms/techniques we've moved to Hadoop at eBay. Over time we've moved from smaller, editorial data sets to large machine generated data sets mined from behavior log data, items/listings, catalogs, etc. One common workflow is to mine large candidate rewrites/expansions data sets from multiple data sources, use crowd sourced human judgment to classify a subset of the candidates (true positive, false positive), use machine learning techniques discard false positives, run automated validation on the final data set, and automatically push to production.
Ravi Jammalakadaka, Senior Applied Researcher, Query Services @ eBay
Ravi is a real engineer. Not a pointy haired manager like the previous speaker. Expect some real engineering:-) He'll be doing a literature review for acronym mining and discussing a real world implementation.
Title: Mining Acronyms From Raw Text
Abstract: Significant number of eBay products are known by their acronyms. eBay query expansion service expands user queries by their acronym equivalents to increase recall. The challenge is to mine acronyms from either seller ( ex. item descriptions, titles) or buyer ( ex. queries) data.
Ravi will present the state of the art algorithms from recent conferences that mine acronyms from raw text and present their limitations. He will present a new acronym mining algorithm that seeks to address the limitations identified with previous algorithms. He will present a machine learning classifier that seeks to remove the false positives generated from the acronym mining algorithm.
This is the deck for Science Advisory Board review of our recent progress in setting up a basic infrastructure -- hybrid system architecture to facilitate automatic question answering in Project Halo -- Vulcan's long-range strong AI effort to attack a key problem in the field of AI research.
Recommendations and User Understandingat StumbleUponDebora Donato
StumbleUpon provides personalized recommendations to help users discover new content across the web. They analyzed user data and conducted A/B tests to optimize recommendations for mobile users. They defined power users as those who regularly discover and interact with content. StumbleUpon also introduced lists, which over 45,000 users created in the first few months to organize content by topic. Data-driven techniques like topic modeling were used to recommend additional lists to users.
The document provides an overview of Harris Interactive's text analytics capabilities and services. It discusses how text analytics can [1] validate quantitative analysis by understanding what respondents mean, [2] provide more contextualized analysis driven less by preconceived categories, and [3] allow for more sophisticated classification. The document also covers how Harris Interactive's text analytics approach provides [1] lower costs for handling large volumes of text data, [2] reliable replication through natural language processing and domain training, and [3] systematic validation of themes.
The document provides an overview and objectives of analyzing LibQUAL+® survey results. It discusses interpreting results internally by identifying areas of best and worst performance. It also covers benchmarking results externally against consortium data and longitudinally to assess the impact of changes. The document outlines quantitative analysis of survey scores and qualitative analysis of comments. It recommends using tools like LibQUAL+® Analytics and SPSS to further analyze results data.
Search Quality Evaluation: a Developer PerspectiveAndrea Gazzarini
Search quality evaluation is an ever-green topic every search engineer ordinarily struggles with. Improving the correctness and effectiveness of a search system requires a set of tools which help measuring the direction where the system is going.
The slides will focus on how a search quality evaluation tool can be seen under a practical developer perspective, how it could be used for producing a deliverable artifact and how it could be integrated within a continuous integration infrastructure.
Evaluating Semantic Search Systems to Identify Future Directions of ResearchStuart Wrigley
Recent work on searching the Semantic Web has yielded a wide range of approaches with respect to the style of input, the underlying search mechanisms and the manner in which results are presented. Each approach has an impact upon the quality of the information retrieved and the user's experience of the search process. This highlights the need for formalised and consistent evaluation to benchmark the coverage, applicability and usability of existing tools and provide indications of future directions for advancement of the state-of-the-art. In this paper, we describe a comprehensive evaluation methodology which addresses both the underlying performance and the subjective usability of a tool. We present the key outcomes of a recently completed international evaluation campaign which adopted this approach and thus identify a number of new requirements for semantic search tools from both the perspective of the underlying technology as well as the user experience.
Improving Semantic Search Using Query Log AnalysisStuart Wrigley
Despite the attention Semantic Search is continuously gaining, several challenges affecting tool performance and user experience remain unsolved. Among these are: matching user terms with the searchspace, adopting view-based interfaces in the Open Web as well as supporting users while building their queries. This paper proposes an approach to move a step forward towards tackling these challenges by creating models of usage of Linked Data concepts and properties extracted from semantic query logs as a source of collaborative knowledge. We use two sets of query logs from the USEWOD workshops to create our models and show the potential of using them in the mentioned areas.
2011 Search Query Rewrites - Synonyms & AcronymsBrian Johnson
July 27, 2011 Bay Area Search Presentation
Brian Johnson, Engineering Director, Query Services @ eBay
Query expansion is an important part of of the search recall for all search engines. In this talk I'll discuss some of the general trend driving Hadoop adoption within the Search Query Services team at eBay, and the types of algorithms/techniques we've moved to Hadoop at eBay. Over time we've moved from smaller, editorial data sets to large machine generated data sets mined from behavior log data, items/listings, catalogs, etc. One common workflow is to mine large candidate rewrites/expansions data sets from multiple data sources, use crowd sourced human judgment to classify a subset of the candidates (true positive, false positive), use machine learning techniques discard false positives, run automated validation on the final data set, and automatically push to production.
Ravi Jammalakadaka, Senior Applied Researcher, Query Services @ eBay
Ravi is a real engineer. Not a pointy haired manager like the previous speaker. Expect some real engineering:-) He'll be doing a literature review for acronym mining and discussing a real world implementation.
Title: Mining Acronyms From Raw Text
Abstract: Significant number of eBay products are known by their acronyms. eBay query expansion service expands user queries by their acronym equivalents to increase recall. The challenge is to mine acronyms from either seller ( ex. item descriptions, titles) or buyer ( ex. queries) data.
Ravi will present the state of the art algorithms from recent conferences that mine acronyms from raw text and present their limitations. He will present a new acronym mining algorithm that seeks to address the limitations identified with previous algorithms. He will present a machine learning classifier that seeks to remove the false positives generated from the acronym mining algorithm.
This is the deck for Science Advisory Board review of our recent progress in setting up a basic infrastructure -- hybrid system architecture to facilitate automatic question answering in Project Halo -- Vulcan's long-range strong AI effort to attack a key problem in the field of AI research.
Recommendations and User Understandingat StumbleUponDebora Donato
StumbleUpon provides personalized recommendations to help users discover new content across the web. They analyzed user data and conducted A/B tests to optimize recommendations for mobile users. They defined power users as those who regularly discover and interact with content. StumbleUpon also introduced lists, which over 45,000 users created in the first few months to organize content by topic. Data-driven techniques like topic modeling were used to recommend additional lists to users.
The document provides an overview of Harris Interactive's text analytics capabilities and services. It discusses how text analytics can [1] validate quantitative analysis by understanding what respondents mean, [2] provide more contextualized analysis driven less by preconceived categories, and [3] allow for more sophisticated classification. The document also covers how Harris Interactive's text analytics approach provides [1] lower costs for handling large volumes of text data, [2] reliable replication through natural language processing and domain training, and [3] systematic validation of themes.
Building Large Arabic Multi-Domain Resources for Sentiment Analysis Hady Elsahar
This document discusses building large multi-domain resources for Arabic sentiment analysis. It describes the problem of existing resources being small, domain-specific and not publicly available. The authors built multi-domain datasets from online reviews totaling over 33,000 reviews across multiple domains. They also built multi-domain sentiment lexicons containing around 2,000 entries using machine learning techniques. Experiments were conducted on the datasets to evaluate their effectiveness for sentiment analysis and provide benchmarking. The results and resources are made publicly available.
This document discusses strategies for maximizing the effectiveness and efficiency of accessibility (A11Y) testing. It describes an experiment where a Fortune 50 client had all their pages, apps, and stages of development tested for A11Y by a single expert, which was less expensive and more accurate than previous approaches. However, the document warns against having QA testers conduct specialized tasks like mobile or screen reader testing without proper training. It proposes a new AVT (Accessibility Verification Testing) process that divides tasks between experts and QA based on whether they require conceptual A11Y knowledge or more concrete testing. The document also describes an experiment where an A11Y expert paired with an engineer to rapidly share their different strengths and perspectives to perform
Fusion 3.1 comes with exciting new features that will make your search more personal and better targeted. Join us for a webinar to learn more about Fusion's features, what's new in this release, and what's around the corner for Fusion.
1. The document discusses Optique 1.0, an ontology-based data access system that allows non-expert users to access data from multiple databases through a visualized query interface.
2. Optique 1.0 uses ontologies and declarative mappings to connect data sources, allowing queries to be formulated over ontologies and then pushed to and evaluated over the databases.
3. The system was tested on a real-world dataset from the oil and gas industry containing over 70 tables and 2.3 million triples extracted from databases into an ontology and triple store.
The penetration of mobile devices equipped with various embedded sensors also make it possible to capture the physical and virtual context of the user and surrounding environment. Further, the modeling of human behaviors based on those data becomes very important due to the increasing popularity of context-aware computing and people-centric applications, which utilize users' behavior pattern to improve the existing services or enable new services. In many natural settings, however, their broader applications are hindered by three main challenges: rarity of labels, uncertainty of activity granularities, and the difficulty of multi-dimensional sensor fusion.
This document discusses machine learning pipelines and introduces Evan Sparks' presentation on building image classification pipelines. It provides an overview of feature extraction techniques used in computer vision like normalization, patch extraction, convolution, rectification and pooling. These techniques are used to transform images into feature vectors that can be input to linear classifiers. The document encourages building simple, intermediate and advanced image classification pipelines using these techniques to qualitatively and quantitatively compare their effectiveness.
Wordnik's architecture is built around a large English word graph database and uses microservices and ephemeral Amazon EC2 storage. Key aspects include:
1) The system is built as independent microservices that communicate via REST APIs documented using Swagger specifications.
2) Databases for each microservice are kept small by design to facilitate operations like backups, replication, and index rebuilding.
3) Services are deployed across multiple Availability Zones and regions on ephemeral Amazon EC2 storage for high availability despite individual host failures.
Presentation on Secondary Indexes from the 9/11/12 HBase Contributor's Meetup. It discusses the current state of the discussion and some possible future directions.
The LDBC Social Network Benchmark Interactive Workload - SIGMOD 2015Ioan Toma
The document discusses the LDBC Social Network Benchmark for evaluating database and graph processing systems. It describes the benchmark's social network data generator which produces realistic data following power law distributions and correlations. It also outlines the benchmark's three workloads: interactive, business intelligence, and graph analytics. The focus is on the interactive workload, which includes complex read queries, simple read queries, and concurrent updates. It aims to identify choke points and measure the acceleration factor a system can sustain for the query mix while meeting a maximum query latency. Parameter curation is used to select query parameters that produce stable performance. The parallel query driver respects dependencies between queries to evaluate a system's ability to handle the workload concurrently.
Asyma E3 2012 - How to make reporting, budgeting & communications simple ...asyma
The document discusses tools to simplify reporting, budgeting, and communications tasks. It introduces Report Runner and Process Scheduler which allow automating financial report generation and scheduling. Information Manager is presented as a way to extend accounting data to include custom tables for key performance indicator reporting, budgeting, forecasting, and analytics. Finally, Info Explorer is highlighted as a tool that addresses issues with traditional reporting by allowing users to define report layouts and distribute scheduled data cubes.
This document provides an introduction to NVivo, a qualitative data analysis software. It describes how NVivo can be used to organize, analyze, and find insights in unstructured qualitative data like documents, interviews, and social media posts. The document outlines the basic NVivo workspace and functions for importing data sources, coding data, running queries, and visualizing results. It also provides guidance on setting up an NVivo project and includes some example tasks for getting started with the software.
Machine learning for IoT - unpacking the blackboxIvo Andreev
This document provides an overview of machine learning and how it can be applied to IoT scenarios. It discusses different machine learning algorithms like supervised and unsupervised learning. It also compares various machine learning platforms like Azure ML, BigML, Amazon ML, Google Prediction and IBM Watson ML. It provides guidance on choosing the right algorithm based on the data and diagnosing why machine learning models may fail. It also introduces neural networks and deep learning concepts. Finally, it demonstrates Azure ML capabilities through a predictive maintenance example.
K-SRL: Instance-based Learning for Semantic Role LabelingYunyao Li
This document summarizes a research paper on instance-based learning for semantic role labeling (SRL). It presents a simple but effective k-nearest neighbors approach using composite features that outperforms previous SRL systems on both in-domain and out-of-domain evaluation. The approach models global argument constraints and addresses SRL challenges like heavy-tailed label distributions and low-frequency exceptions through explicit representation of local biases in composite features.
Machine Learning Foundations for Professional ManagersAlbert Y. C. Chen
20180804@Taiwan AI Academy, Hsinchu
6 hour lecture for those new to machine learning, to grasps the concepts, advantages and limitations of various classical machine learning methods. More importantly, to learn the skills to break down large complicated AI projects into manageable pieces, where features and functionalities could be added incrementally and annotated data accumulated. Take home message: machine learning is always a delicate balance between model complexity M and number of data N so that the trained classifier generalizes well and does not overfit.
Relevance Improvements at Cengage - Ivan Provalovlucenerevolution
In the session we describe relevance improvements we have implemented in our Lucene-based search system for English and Chinese contents and the tests we have performed for Arabic and Spanish contents based on TREC data. We will also describe our relevance feedback web app for the end-users to rank results of various queries. The presentation will have information about the usage data we analyze to improve the relevance. We will also touch upon our OCR data indexing challenges for English and non-English content.
From Academic Papers To Production : A Learning To Rank StoryAlessandro Benedetti
This talk is about the journey to bring Learning To Rank (LTR from now on) to the e-commerce domain in a real world scenario, including all the pitfalls and disillutions involved.
LTR is a fantastic approach to solve complex ranking problems but industry domains are far from being the ideal world where those technologies were designed and experimented : open source software implementations are not working perfectly out of the box and require advanced tuning; industry training data is dirty, noisy and incomplete.
This talk will guide you through the different phases and technologies involved in a LTR project with a pragmatic approach.
Feature Engineering, Domain Modelling, Training Set Building, Model Training, Search Integration and Online Evaluation : each of them presents different challenges in the real world and must be carefully approached.
Facets and Pivoting for Flexible and Usable Linked Data ExplorationRoberto García
The success of Open Data initiatives has increased the amount of data available on the Web. Unfortunately, most of this data is only available in raw tabular form, what makes analysis and reuse quite difficult for non-experts. Linked Data principles allow for a more sophisticated approach by making explicit both the structure and semantics of the data. However, from the end-user viewpoint, they continue to be monolithic files completely opaque or difficult to explore by making tedious semantic queries. Our objective is to facilitate the user to grasp what kind of entities are in the dataset, how they are interrelated, which are their main properties and values, etc. Rhizomer is a tool for data publishing whose interface provides a set of components borrowed from Information Architecture (IA) that facilitate awareness of the dataset at hand. It automatically generates navigation menus and facets based on the kinds of things in the dataset and how they are described through metadata properties and values. Moreover, motivated by recent tests with end-users, it also provides the possibility to pivot among the faceted views created for each class of resources in the dataset.
The effects of preferred text formatting on performance and perceptual appealPaul Doncaster
Legal researchers were tested to see whether their preferred settings for reading documents online resulted in better perceived visual appeal and performance. Results showed that customized text formatting does not translate into improved reading performance, nor does it sustain its “preferred” status when compared to optimized alternatives.
This document discusses collaborative filtering and recommender systems. It begins with an overview of non-relational databases and graph databases. It then discusses collaborative filtering, including calculating similarity scores between users or items, predicting ratings for unseen items, and making recommendations. Specific methods discussed include Euclidean distance, Pearson correlation, and user-based filtering. The goal of collaborative filtering is to increase sales, market share, and targeted advertising by making personalized recommendations to users.
Towards Detecting Performance Anti-patterns Using Classification TechniquesJames Hill
This is the talk I gave on behalf of my Ph.D. student at the Machine Learning and Information Retrieval (MALIR) for Software Evolution (MALIR-SE) workshop at ASE 2013.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
More Related Content
Similar to Evaluating Semantic Search Query Approaches with Expert and Casual Users
Building Large Arabic Multi-Domain Resources for Sentiment Analysis Hady Elsahar
This document discusses building large multi-domain resources for Arabic sentiment analysis. It describes the problem of existing resources being small, domain-specific and not publicly available. The authors built multi-domain datasets from online reviews totaling over 33,000 reviews across multiple domains. They also built multi-domain sentiment lexicons containing around 2,000 entries using machine learning techniques. Experiments were conducted on the datasets to evaluate their effectiveness for sentiment analysis and provide benchmarking. The results and resources are made publicly available.
This document discusses strategies for maximizing the effectiveness and efficiency of accessibility (A11Y) testing. It describes an experiment where a Fortune 50 client had all their pages, apps, and stages of development tested for A11Y by a single expert, which was less expensive and more accurate than previous approaches. However, the document warns against having QA testers conduct specialized tasks like mobile or screen reader testing without proper training. It proposes a new AVT (Accessibility Verification Testing) process that divides tasks between experts and QA based on whether they require conceptual A11Y knowledge or more concrete testing. The document also describes an experiment where an A11Y expert paired with an engineer to rapidly share their different strengths and perspectives to perform
Fusion 3.1 comes with exciting new features that will make your search more personal and better targeted. Join us for a webinar to learn more about Fusion's features, what's new in this release, and what's around the corner for Fusion.
1. The document discusses Optique 1.0, an ontology-based data access system that allows non-expert users to access data from multiple databases through a visualized query interface.
2. Optique 1.0 uses ontologies and declarative mappings to connect data sources, allowing queries to be formulated over ontologies and then pushed to and evaluated over the databases.
3. The system was tested on a real-world dataset from the oil and gas industry containing over 70 tables and 2.3 million triples extracted from databases into an ontology and triple store.
The penetration of mobile devices equipped with various embedded sensors also make it possible to capture the physical and virtual context of the user and surrounding environment. Further, the modeling of human behaviors based on those data becomes very important due to the increasing popularity of context-aware computing and people-centric applications, which utilize users' behavior pattern to improve the existing services or enable new services. In many natural settings, however, their broader applications are hindered by three main challenges: rarity of labels, uncertainty of activity granularities, and the difficulty of multi-dimensional sensor fusion.
This document discusses machine learning pipelines and introduces Evan Sparks' presentation on building image classification pipelines. It provides an overview of feature extraction techniques used in computer vision like normalization, patch extraction, convolution, rectification and pooling. These techniques are used to transform images into feature vectors that can be input to linear classifiers. The document encourages building simple, intermediate and advanced image classification pipelines using these techniques to qualitatively and quantitatively compare their effectiveness.
Wordnik's architecture is built around a large English word graph database and uses microservices and ephemeral Amazon EC2 storage. Key aspects include:
1) The system is built as independent microservices that communicate via REST APIs documented using Swagger specifications.
2) Databases for each microservice are kept small by design to facilitate operations like backups, replication, and index rebuilding.
3) Services are deployed across multiple Availability Zones and regions on ephemeral Amazon EC2 storage for high availability despite individual host failures.
Presentation on Secondary Indexes from the 9/11/12 HBase Contributor's Meetup. It discusses the current state of the discussion and some possible future directions.
The LDBC Social Network Benchmark Interactive Workload - SIGMOD 2015Ioan Toma
The document discusses the LDBC Social Network Benchmark for evaluating database and graph processing systems. It describes the benchmark's social network data generator which produces realistic data following power law distributions and correlations. It also outlines the benchmark's three workloads: interactive, business intelligence, and graph analytics. The focus is on the interactive workload, which includes complex read queries, simple read queries, and concurrent updates. It aims to identify choke points and measure the acceleration factor a system can sustain for the query mix while meeting a maximum query latency. Parameter curation is used to select query parameters that produce stable performance. The parallel query driver respects dependencies between queries to evaluate a system's ability to handle the workload concurrently.
Asyma E3 2012 - How to make reporting, budgeting & communications simple ...asyma
The document discusses tools to simplify reporting, budgeting, and communications tasks. It introduces Report Runner and Process Scheduler which allow automating financial report generation and scheduling. Information Manager is presented as a way to extend accounting data to include custom tables for key performance indicator reporting, budgeting, forecasting, and analytics. Finally, Info Explorer is highlighted as a tool that addresses issues with traditional reporting by allowing users to define report layouts and distribute scheduled data cubes.
This document provides an introduction to NVivo, a qualitative data analysis software. It describes how NVivo can be used to organize, analyze, and find insights in unstructured qualitative data like documents, interviews, and social media posts. The document outlines the basic NVivo workspace and functions for importing data sources, coding data, running queries, and visualizing results. It also provides guidance on setting up an NVivo project and includes some example tasks for getting started with the software.
Machine learning for IoT - unpacking the blackboxIvo Andreev
This document provides an overview of machine learning and how it can be applied to IoT scenarios. It discusses different machine learning algorithms like supervised and unsupervised learning. It also compares various machine learning platforms like Azure ML, BigML, Amazon ML, Google Prediction and IBM Watson ML. It provides guidance on choosing the right algorithm based on the data and diagnosing why machine learning models may fail. It also introduces neural networks and deep learning concepts. Finally, it demonstrates Azure ML capabilities through a predictive maintenance example.
K-SRL: Instance-based Learning for Semantic Role LabelingYunyao Li
This document summarizes a research paper on instance-based learning for semantic role labeling (SRL). It presents a simple but effective k-nearest neighbors approach using composite features that outperforms previous SRL systems on both in-domain and out-of-domain evaluation. The approach models global argument constraints and addresses SRL challenges like heavy-tailed label distributions and low-frequency exceptions through explicit representation of local biases in composite features.
Machine Learning Foundations for Professional ManagersAlbert Y. C. Chen
20180804@Taiwan AI Academy, Hsinchu
6 hour lecture for those new to machine learning, to grasps the concepts, advantages and limitations of various classical machine learning methods. More importantly, to learn the skills to break down large complicated AI projects into manageable pieces, where features and functionalities could be added incrementally and annotated data accumulated. Take home message: machine learning is always a delicate balance between model complexity M and number of data N so that the trained classifier generalizes well and does not overfit.
Relevance Improvements at Cengage - Ivan Provalovlucenerevolution
In the session we describe relevance improvements we have implemented in our Lucene-based search system for English and Chinese contents and the tests we have performed for Arabic and Spanish contents based on TREC data. We will also describe our relevance feedback web app for the end-users to rank results of various queries. The presentation will have information about the usage data we analyze to improve the relevance. We will also touch upon our OCR data indexing challenges for English and non-English content.
From Academic Papers To Production : A Learning To Rank StoryAlessandro Benedetti
This talk is about the journey to bring Learning To Rank (LTR from now on) to the e-commerce domain in a real world scenario, including all the pitfalls and disillutions involved.
LTR is a fantastic approach to solve complex ranking problems but industry domains are far from being the ideal world where those technologies were designed and experimented : open source software implementations are not working perfectly out of the box and require advanced tuning; industry training data is dirty, noisy and incomplete.
This talk will guide you through the different phases and technologies involved in a LTR project with a pragmatic approach.
Feature Engineering, Domain Modelling, Training Set Building, Model Training, Search Integration and Online Evaluation : each of them presents different challenges in the real world and must be carefully approached.
Facets and Pivoting for Flexible and Usable Linked Data ExplorationRoberto García
The success of Open Data initiatives has increased the amount of data available on the Web. Unfortunately, most of this data is only available in raw tabular form, what makes analysis and reuse quite difficult for non-experts. Linked Data principles allow for a more sophisticated approach by making explicit both the structure and semantics of the data. However, from the end-user viewpoint, they continue to be monolithic files completely opaque or difficult to explore by making tedious semantic queries. Our objective is to facilitate the user to grasp what kind of entities are in the dataset, how they are interrelated, which are their main properties and values, etc. Rhizomer is a tool for data publishing whose interface provides a set of components borrowed from Information Architecture (IA) that facilitate awareness of the dataset at hand. It automatically generates navigation menus and facets based on the kinds of things in the dataset and how they are described through metadata properties and values. Moreover, motivated by recent tests with end-users, it also provides the possibility to pivot among the faceted views created for each class of resources in the dataset.
The effects of preferred text formatting on performance and perceptual appealPaul Doncaster
Legal researchers were tested to see whether their preferred settings for reading documents online resulted in better perceived visual appeal and performance. Results showed that customized text formatting does not translate into improved reading performance, nor does it sustain its “preferred” status when compared to optimized alternatives.
This document discusses collaborative filtering and recommender systems. It begins with an overview of non-relational databases and graph databases. It then discusses collaborative filtering, including calculating similarity scores between users or items, predicting ratings for unseen items, and making recommendations. Specific methods discussed include Euclidean distance, Pearson correlation, and user-based filtering. The goal of collaborative filtering is to increase sales, market share, and targeted advertising by making personalized recommendations to users.
Towards Detecting Performance Anti-patterns Using Classification TechniquesJames Hill
This is the talk I gave on behalf of my Ph.D. student at the Machine Learning and Information Retrieval (MALIR) for Software Evolution (MALIR-SE) workshop at ASE 2013.
Similar to Evaluating Semantic Search Query Approaches with Expert and Casual Users (20)
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
20240605 QFM017 Machine Intelligence Reading List May 2024
Evaluating Semantic Search Query Approaches with Expert and Casual Users
1. Evaluating Semantic Search Query
Approaches with Expert and Casual Users
Khadija Elbedweihy, Stuart N. Wrigley and Fabio Ciravegna
OAK Research Group,
Department of Computer Science,
University of Sheffield, UK
3. Motivation – Semantic Search
• Wikipedia states that Semantic Search “seeks to improve
search accuracy by understanding searcher intent and the
contextual meaning of terms as they appear in the
searchable dataspace, whether on the Web or within a
closed system, to generate more relevant results”.
• Covers broad category of applications in Semantic Web:
– Search engines (e.g., Swoogle, FalconS, Sindice)
– Closed-domain query interfaces (e.g., AquaLog, Querix)
– Open-domain query interfaces (e.g., PowerAqua)
4. Motivation - Evaluations
• Evaluation of software is critical.
• Large-scale evaluations foster research and development.
• Semantic search evaluations (SemSearch, TREC ELC,
QALD) focused on assessing retrieval performance.
Assessing usability of tools and user satisfaction is
important in Semantic Search.
5. Research Question
How do different types of users perceive the usability
of different query approaches?
• Method
- Assess usability and user satisfaction of:
* Free-NL, Controlled-NL, Form-based, Graph-based
- from the perspective of
* expert users and casual users
6. Query Approaches
Controlled-NL
Free-NL
Specific vocabulary
Natural language queries
Which state has river Submit
capital
What is the capital of Alabama? Submit lake
mountai
capital Alabama Submit n
a
any
Form-based Graph-based
Visualize the Visualize the
search space search space
7. Evaluation Design: Dataset
• Mooney Natural Language Learning Data
- simple and well-known domain (geography)
- used by other studies within the search community
- questions already available (877 NL questions)
• Geography Dataset:
– Concepts: State, City, Lake, Mountain, Capital, River, etc
– Properties: population of state, length of river, etc
– Relations linking concepts: State ‘hasCity’ City
8. Evaluation Design: Data Collected
• Objective data:
1) Input time
2) Number of attempts
3) Success rate
• Subjective data, collected using:
1) Questionnaires (e.g., System Usability Scale ‘SUS’)
2) Ranking of the tools (w.r.t: system, query approach,
results content, results presentation)
3) Observations
15. Results for expert users
• Visualizing the entire ontology supports query formulation
– Semantic Crystal: shows the entire ontology.
– Affective Graphs: shows selected concepts & relations.
16. Results for casual users
• Not showing ontology more complex for casual users:
– Semantic Crystal receiving higher scores.
– Affective Graphs perceived as complex and difficult to use
• 50% of the users found it to increase complexity and difficulty
18. Results for expert users
• Controlled-NL very restrictive for expert users (least-liked)
• Highest query input time
120
Best 1
0.9
100 0.8
Query Language Rank
0.7 Graph-based1
80
Input Time (Sec)
0.6 Graph-based2
60 0.5 Form-based
0.4 Controlled-NL
40
0.3 Free-NL
20 0.2
0.1
0 0
Worst
19. Results for casual users
• Controlled-NL provided most support for casual users.
• Users’ positive feedback for controlled-NL:
– allow only correct queries (50%)
– suggestions and guidance to formulate queries (40%)
Example: Although Ginseng is limited to specific vocabulary, I
knew that I will get answers once I can do the query because it
only allows the correct ones and thus I didn't keep trying a lot
of queries that I wasn't sure about.
21. Free-NL approach
+ simplest and most natural
- suffer from habitability problem.
• Feedback: “I have to guess the right words”
– Example: `run through’ with `river’ but not `traverse’.
• NLP-Reduce:
– lowest success rate: 20%
– highest number of attempts: 4.2
22. Negation
• Tell me which rivers do not traverse the state with the
capital Nashville?
1
0.9
0.8
0.7
Answer Found Rate
Graph-based1
0.6
Graph-based2
0.5
Form-based
0.4
Controlled-NL
0.3
Free-NL
0.2
0.1
0
Expert Users Casual Users
23. Negation
Tell me which states does the river Mississippi does not
traverse.
• “Closed world assumption (CWA): presumption that what
is not currently known to be true is false”.
<Mississippi, traverse, Louisiana>
• “Open world assumption (OWA): assumption that the
truth-value of a statement is independent of whether or
not it is known by any single observer or agent to be true”.
<Mississippi, not_traverse, Alabama>
25. Formal Query
• Benefit of showing formal query depends on user type.
• Formal query perceived by:
– Casual users: not understandable and confusing
– Expert users: increased confidence
Also, performing direct changes to the formal query
increased the expressiveness of the query language.
26. Results presentation
• Results presentation and format affected usability and user
satisfaction.
– Unless users are very familiar with the data, presenting URIs
alone is not very helpful.
– Example: A query for rivers returns one of the answers:
http://www.mooney.net/geo#tennesse2
27. Results Content
• Results should be augmented with associated information
to provide a `richer’ user experience.
• Users feedback:
– Maybe a `mouse over' function to show more
information.
– Perhaps related information with the results.
– Results very limited, would be good to have more
context.
29. Research Question & Approach
How do different types of users perceive the usability
of different query approaches?
- Assess usability and user satisfaction of:
* Free-NL, Controlled-NL, Form-based, Graph-based
- from the perspective of
* expert users and casual users
30. Conclusions
Expert Users Casual Users
• Graph-based most preferred • Form-based mid-point
- Intuitive - Allow more complex queries than
- Support complex queries NL.
- Easier than graph-based
• Controlled-NL least preferred
- Faster than graph-based
- Very restrictive.
- Limited expressiveness • Controlled-NL most supportive
• Prefer flexibility of free-NL - Only valid queries: Confidence
• Formal query provides confidence - Vocabulary suggestions: guidance
- Ability to change query increases • Formal Query not understandable
expressiveness. and confusing.
• Users want search results to be augmented with more
information to have a better understanding of the answers.
31. Recommendations
Cater to both expert and casual users:
• Hybridized query approach: Combine a view-based
approach (visualize search space) with a NL-input feature
(balance difficulty and speed) while including optional
suggestions for the NL input (provide guidance).
• Results Content: Augment results with ‘extra’ and ‘related’
information.
– extra information: for ‘State’: capital, area, population.
– related information: for ‘State’: rivers, lakes, mountains.
32. Limitations & Future work
• Limitation: Small size of the dataset.
• Assess learnability of different query approaches.
• Assess how interaction with the search tools affect the
information seeking process: usefulness.
– Use questions with an overall goal and compare users'
knowledge before and after the search task.