Google Goggles is a mobile visual search system that allows users to search for information by taking photos with their smartphone cameras. It uses computer vision techniques like interest point detection, feature extraction and indexing to match query images to images in large online databases. This kind of visual search is relevant because of the rise of powerful mobile devices with cameras and popular image-sharing apps. It enables new commercial opportunities for visual search and discovery on mobile phones.
This document discusses challenges with integrating elevation datasets including differences in vertical datums, resolutions, and times of capture. It presents results from a study to develop strategies for accurately transforming elevation data between datums. The strategies include using geoid and tide gauge data to transform between ellipsoidal and mean sea level datums. The document also describes a method for co-registering and blending multi-resolution elevation datasets to create a seamless digital elevation model while assessing accuracy using checkpoints.
Presentation at International Advanced School on Knowledge Co-creation and Service Innovation 2012, Japan Advanced Institute of Science and Technology, March 1
The document presents a zero-adjusted gamma model for estimating loss given default (LGD) on residential mortgage loans. It compares the performance of this model, which directly models the loss amount, to a traditional linear regression approach. The zero-adjusted gamma model is found to accommodate the non-linear relationships between loss amounts and predictor variables better than the linear model. It also estimates separate factors that predict the probability of loss and those that influence the loss amount. The zero-adjusted gamma model is shown to produce competitively predictive LGD estimates through validation testing.
Sparse feature analysis for detection of clustered microcalcifications in mam...Wesley De Neve
This document analyzes the use of sparse feature analysis for detecting clustered microcalcifications in mammogram images. It compares different feature types, combinations of features, and dictionary construction techniques for sparse representation based classification (SRC) of mammogram images. The experimental results show that texture features like Laws' texture features (LAW) are more effective than shape/morphology features. SRC using LAW features alone or combined with local binary patterns (LBP) achieved high performance. Larger dictionaries containing more atoms resulted in higher discriminative power for the SRC-based detection system.
This document summarizes a presentation on assessing quality of experience (QoE) for 3D television (3DTV) and beyond. It discusses testing methodologies for 3DTV image quality assessment (IQA), including both subjective and objective approaches. On the subjective side, it compares different rating scales that have been used to evaluate attributes like depth quality, visual comfort, and overall experience. It also addresses challenges in measuring long-term QoE factors like visual fatigue. On the objective side, it proposes initial approaches for 3D IQA metrics and recognizes the need to define new metrics that consider both visual quality and depth quality. Overall, the presentation examines moving from 2D visual quality evaluation to a multidimensional assessment of
Capping is one of the most complex phenomena in the pharmaceutical industry it is one of the mechanical defects in the tableting process in which catastrophic failure of the compact can occur. Understanding what influences tablet capping in terms of process variables, material properties, and density/stress distributions in tablets and developing specialized techniques to correlate these variables with mechanical failures are practical interests of the pharmaceutical industry. In this presentation, we describe a nondestructive ultrasonic device/methodology to predict the capping tendencies of tablet formulations based on their manufacturing performances.
1) The document summarizes a presentation given at the 2007 ACES Conference on using the Partial Element Equivalent Circuit (PEEC) technique to model frequency dependent phenomena like skin effect.
2) It describes using volume filaments and macro-basis functions to model skin effect in conductors with thickness between the skin depth. Broadband macromodels are generated from a frequency domain PEEC solver.
3) The presentation outlines modeling skin effect using an analytic solution of Maxwell's equations to derive impedance terms dependent on conductor thickness and frequency, and examines the asymptotic behavior of the model in the low- and high-frequency limits.
This document discusses challenges with integrating elevation datasets including differences in vertical datums, resolutions, and times of capture. It presents results from a study to develop strategies for accurately transforming elevation data between datums. The strategies include using geoid and tide gauge data to transform between ellipsoidal and mean sea level datums. The document also describes a method for co-registering and blending multi-resolution elevation datasets to create a seamless digital elevation model while assessing accuracy using checkpoints.
Presentation at International Advanced School on Knowledge Co-creation and Service Innovation 2012, Japan Advanced Institute of Science and Technology, March 1
The document presents a zero-adjusted gamma model for estimating loss given default (LGD) on residential mortgage loans. It compares the performance of this model, which directly models the loss amount, to a traditional linear regression approach. The zero-adjusted gamma model is found to accommodate the non-linear relationships between loss amounts and predictor variables better than the linear model. It also estimates separate factors that predict the probability of loss and those that influence the loss amount. The zero-adjusted gamma model is shown to produce competitively predictive LGD estimates through validation testing.
Sparse feature analysis for detection of clustered microcalcifications in mam...Wesley De Neve
This document analyzes the use of sparse feature analysis for detecting clustered microcalcifications in mammogram images. It compares different feature types, combinations of features, and dictionary construction techniques for sparse representation based classification (SRC) of mammogram images. The experimental results show that texture features like Laws' texture features (LAW) are more effective than shape/morphology features. SRC using LAW features alone or combined with local binary patterns (LBP) achieved high performance. Larger dictionaries containing more atoms resulted in higher discriminative power for the SRC-based detection system.
This document summarizes a presentation on assessing quality of experience (QoE) for 3D television (3DTV) and beyond. It discusses testing methodologies for 3DTV image quality assessment (IQA), including both subjective and objective approaches. On the subjective side, it compares different rating scales that have been used to evaluate attributes like depth quality, visual comfort, and overall experience. It also addresses challenges in measuring long-term QoE factors like visual fatigue. On the objective side, it proposes initial approaches for 3D IQA metrics and recognizes the need to define new metrics that consider both visual quality and depth quality. Overall, the presentation examines moving from 2D visual quality evaluation to a multidimensional assessment of
Capping is one of the most complex phenomena in the pharmaceutical industry it is one of the mechanical defects in the tableting process in which catastrophic failure of the compact can occur. Understanding what influences tablet capping in terms of process variables, material properties, and density/stress distributions in tablets and developing specialized techniques to correlate these variables with mechanical failures are practical interests of the pharmaceutical industry. In this presentation, we describe a nondestructive ultrasonic device/methodology to predict the capping tendencies of tablet formulations based on their manufacturing performances.
1) The document summarizes a presentation given at the 2007 ACES Conference on using the Partial Element Equivalent Circuit (PEEC) technique to model frequency dependent phenomena like skin effect.
2) It describes using volume filaments and macro-basis functions to model skin effect in conductors with thickness between the skin depth. Broadband macromodels are generated from a frequency domain PEEC solver.
3) The presentation outlines modeling skin effect using an analytic solution of Maxwell's equations to derive impedance terms dependent on conductor thickness and frequency, and examines the asymptotic behavior of the model in the low- and high-frequency limits.
Recent advances in visual information retrieval marques klu june 2010Oge Marques
The document summarizes key points from a 2010 presentation on visual information retrieval (VIR). It revisits conclusions from a 2000 paper on challenges facing content-based image retrieval (CBIR). While some predictions were accurate, like increased data sizes and interaction options, others were not, like solving image understanding. Significant progress was made on benchmarks and datasets but less on similarity metrics. Medical image retrieval poses new challenges to understand but offers opportunities if VIR methods can adapt to new domains.
This document discusses advances in image search and retrieval. It begins with an overview of visual information retrieval and its challenges, including the semantic gap between low-level visual features and high-level semantics. It then covers recent techniques like Google image search and similarity search. The document outlines core concepts like capturing similarity, large datasets, and user needs. It also revisits a 2000 paper on the challenges still facing the field, including the unsolved semantic gap and need for standardized evaluation benchmarks.
Using games to improve computer vision solutionsOge Marques
Dr. Oge Marques discusses using games to improve computer vision solutions. Specifically, Dr. Marques describes a two-player web-based guessing game called Ask'nSeek that helps solve the computer vision problems of object detection, labeling, and semantic scene segmentation. Ask'nSeek logs spatial relationships and labels from a small number of games per image to train machine learning models for these tasks.
Visual Information Retrieval: Advances, Challenges and OpportunitiesOge Marques
Visual Information Retrieval: Advances, Challenges and Opportunities discusses advances and challenges in visual information retrieval. Key points include:
- Visual information retrieval aims to find relevant images/videos based on visual and text queries, addressing the "semantic gap" between low-level features and high-level meanings.
- Advances include improved text-based, content-based, and mixed search methods, as well applications in medical image retrieval and mobile visual search.
- Ongoing challenges include capturing image similarity, addressing various representation gaps, understanding user intentions, and developing broad domain solutions.
A Comparative Study of Content Based Image Retrieval Trends and ApproachesCSCJournals
Content Based Image Retrieval (CBIR) is an important step in addressing image storage and management problems. Latest image technology improvements along with the Internet growth have led to a huge amount of digital multimedia during the recent decades. Various methods, algorithms and systems have been proposed to solve these problems. Such studies revealed the indexing and retrieval concepts, which have further evolved to Content-Based Image Retrieval. CBIR systems often analyze image content via the so-called low-level features for indexing and retrieval, such as color, texture and shape. In order to achieve significantly higher semantic performance, recent systems seek to combine low-level with high-level features that contain perceptual information for human. Purpose of this review is to identify the set of methods that have been used for CBR and also to discuss some of the key contributions in the current decade related to image retrieval and main challenges involved in the adaptation of existing image retrieval techniques to build useful systems that can handle real-world data. By making use of various CBIR approaches accurate, repeatable, quantitative data must be efficiently extracted in order to improve the retrieval accuracy of content-based image retrieval systems. In this paper, various approaches of CBIR and available algorithms are reviewed. Comparative results of various techniques are presented and their advantages, disadvantages and limitations are discussed.
Multimodal Analysis for Bridging Semantic Gap with Biologically Inspired Algo...techkrish
The amount and complexity of digital media being generated, stored, transmitted, analysed and accessed has exponentially increased as a result of advances in computer and Web technologies. Much of this information combines digital images, video, audio, graphics and textual data. Large-scale online video repositories enable users to creatively share material along a wide audience. Consequently, there is an increasing interest in associating media items with free-text annotations, ranging from simple titles and detailed descriptions of the video content. In an effort to reduce the complexity of the annotation task, this talk will outline some of the techniques developed for indexing large-scale multimedia repositories by exploiting multi-modality of information space. One such approach combines the use of semantic expansion and visual analysis for predicting user tags for online videos. The framework is designed to exploit visual features using biologically inspired algorithms and associated textual metadata, which is semantically, expanded using complementary textual resources. The experimental results indicate the usefulness of the proposed approach for analysing large-scale media items.
UNSUPERVISED VISUAL HASHING WITH SEMANTIC ASSISTANT FOR CONTENT-BASED IMAGE R...Nexgen Technology
This document proposes a novel unsupervised visual hashing approach called semantic-assisted visual hashing (SAVH) to improve content-based image retrieval. SAVH leverages rich semantics from auxiliary texts associated with images to boost visual hashing performance without requiring explicit semantic labels. It develops a unified framework to learn hash codes by preserving visual similarities between images, integrating semantic assistance from texts, and characterizing correlations between images and shared topics. Experimental results on several datasets show SAVH achieves superior performance over state-of-the-art techniques by effectively utilizing semantics from texts to assist visual hashing.
The Real Problem of Bridging the Multimedia “Semantic Gap” jrs21
WWW-2007 Panel Position:
- Since video search is visual, the semantic spaces should be defined visually as well
- Create large multimedia knowledge-base with exemplar content representing all semantic concepts relevant for search
- Allow semantics space to evolve from end-user perspectives (across sports, entertainment, news)
- Allow technology to focus on extracting the relevant semantics – truly providing the needed data-driven approach for bridging the multimedia semantic gap
Content-Based Image Retrieval (CBIR) systems employ colour as primary feature with texture and shape as secondary features. In this project a simple, image retrieval system will be implemented
The document discusses content-based image retrieval (CBIR) systems. It describes how CBIR systems use feature extraction to search large image databases based on visual content. The key components of CBIR systems are feature extraction, indexing, and system design. Feature extraction involves extracting information about images' colors, textures, shapes, and spatial locations. Effective features and indexing techniques are needed to make CBIR scalable for large image collections. Performance is evaluated based on how well systems return relevant images.
Literature Review on Content Based Image RetrievalUpekha Vandebona
This document summarizes a literature review on content-based image retrieval (CBIR). It discusses how CBIR uses computer vision techniques to automatically extract visual features from images for retrieval, unlike traditional concept-based methods that rely on metadata/text. The key visual features discussed are color, texture, and shape. A typical CBIR system architecture includes creating an image database, automatically extracting features, searching by example or semantics, and ranking results. Distance measures are used to compare image features and evaluate retrieval performance. Combining CBIR with concept-based techniques could improve image retrieval overall.
Visual search, also known as content-based image retrieval, allows users to search for images using either text queries, visual queries by uploading an example image, or visual queries by drawing an image. It has many applications including searching product catalogs, maps, photo archives, and for law enforcement. A visual search system typically uses low-level image descriptors for color, texture, shape and spatial layout to extract machine-understandable features from images. It then calculates similarity distances between images and indexes them to allow efficient searching. Performance is measured using precision and recall metrics. Existing visual search engines can still struggle with semantic gaps between low-level features and high-level human concepts.
This document outlines a presentation on content-based image retrieval (CBIR). It discusses the motivation for CBIR by describing limitations of text-based image retrieval, such as problems with image annotation, human perception, and queries that cannot be described with text. CBIR allows images to be retrieved based on automatically extracted visual features like color, texture, and histograms. A typical CBIR system extracts image features and then matches features to find visually similar images. Applications of CBIR include crime prevention, security, medical diagnosis, and intellectual property. The conclusion states that CBIR reduces computation time and increases user interaction compared to other methods.
Content Based Image and Video Retrieval AlgorithmAkshit Bum
The document describes content-based image and video retrieval (CBIR) algorithms. It discusses how CBIR works by extracting features from query images, indexing images, and retrieving similar images based on color, shape, and texture features. CBIR techniques include reverse image search, semantic retrieval using queries, and relevance feedback to refine searches based on user input about retrieved images. The document provides examples of CBIR applications in areas like crime prevention, military, web searching, and medical diagnosis.
The project aims at development of efficient segmentation method for the CBIR system. Mean-shift segmentation generates a list of potential objects which are meaningful and then these objects are clustered according to a predefined similarity measure. The method was tested on benchmark data and F-Score of .30 was achieved.
- Content-based image retrieval (CBIR) searches for images based on visual features like color, texture, and shape rather than keywords.
- CBIR systems extract features from images to create metadata and use those features to calculate visual similarity between images.
- Relevance feedback allows users to provide feedback on initial search results to help the system recalculate feature weights and improve subsequent results.
Content-based image retrieval (CBIR) uses visual image content to search large image databases according to user needs. CBIR systems represent images by extracting features related to color, shape, texture, and spatial layout. Features are extracted from regions of the image and compared to features of images in the database to find the most similar matches. CBIR has applications in medical imaging, fingerprints, photo collections, and more. Techniques include representing images with histograms of color and texture features extracted through transforms.
Vitamins are complex substances that regulate body processes and act as coenzymes in enzyme reactions. They are named alphabetically and classified as either fat-soluble or water-soluble. Fat-soluble vitamins like A, D, E and K can be stored in the body, while water-soluble vitamins like C and the B vitamins cannot be stored and must be replenished regularly through diet. Deficiencies in vitamins can cause a variety of symptoms and even permanent damage if left untreated. The document provides details on the functions, dietary sources, deficiency symptoms and risk factors for various vitamins.
This document discusses a project involving the development of a new type of aircraft called the WHITNEY S201. It mentions components like the MmLC2 AMM engine and provides technical specifications like a maximum takeoff weight of 41,000 pounds. The document also notes this project is still in development with further testing planned.
The document summarizes the key strategies and tactics used in Barack Obama's successful 2008 presidential campaign. It outlines how the campaign focused on themes of change, unity, reform, honesty and hope. It emphasizes expanding the electoral map, embracing new technologies, and expanding the electorate by appealing to youth and Latino voters. Advertising was tailored to different demographic groups and coordinated across paid, earned, and social media to turn interested voters into active supporters. The integrated approach helped Obama win the election.
Presented at DocTrain East 2007 by Joe Gollner, Stilo International -- This workshop will introduce participants to S1000D, a rapidly evolving standard that has gained growing level of adoption as a shared approach to addressing the wide range of requirements associated with planning, creating, managing, publishing and exchanging documentation for complex equipment systems. The workshop will provide guidelines for assessing the applicability of S1000D and an implementation framework for managing S1000D deployments. The following topics will receive specific attention:
* An overview of S1000D, its purpose and history
* A review of the S1000D schema framework
* A closer look at specific models
* The underlying identification and management schemes
* Recent changes and future directions
* Implementation examples
* Criteria for determining if S1000D is right for you
* Key considerations to keep in mind when implementing S1000D
* Comparing S1000D with other standards (e.g., DITA)
The document discusses a contact analysis of a deep groove ball bearing using ANSYS finite element software. A 3D parametric model of the bearing was created using APDL. Contact pairs between the inner/outer rings and balls were defined. Boundary conditions fixing the outer ring and applying a radial load to the inner ring were specified. A nonlinear contact analysis was performed and results for von Mises stress, strain, contact area shape, and penetration were found to be consistent with Hertzian contact theory values, validating the model and analysis. The contact analysis provides a scientific basis for optimizing bearing design under complex loads.
Recent advances in visual information retrieval marques klu june 2010Oge Marques
The document summarizes key points from a 2010 presentation on visual information retrieval (VIR). It revisits conclusions from a 2000 paper on challenges facing content-based image retrieval (CBIR). While some predictions were accurate, like increased data sizes and interaction options, others were not, like solving image understanding. Significant progress was made on benchmarks and datasets but less on similarity metrics. Medical image retrieval poses new challenges to understand but offers opportunities if VIR methods can adapt to new domains.
This document discusses advances in image search and retrieval. It begins with an overview of visual information retrieval and its challenges, including the semantic gap between low-level visual features and high-level semantics. It then covers recent techniques like Google image search and similarity search. The document outlines core concepts like capturing similarity, large datasets, and user needs. It also revisits a 2000 paper on the challenges still facing the field, including the unsolved semantic gap and need for standardized evaluation benchmarks.
Using games to improve computer vision solutionsOge Marques
Dr. Oge Marques discusses using games to improve computer vision solutions. Specifically, Dr. Marques describes a two-player web-based guessing game called Ask'nSeek that helps solve the computer vision problems of object detection, labeling, and semantic scene segmentation. Ask'nSeek logs spatial relationships and labels from a small number of games per image to train machine learning models for these tasks.
Visual Information Retrieval: Advances, Challenges and OpportunitiesOge Marques
Visual Information Retrieval: Advances, Challenges and Opportunities discusses advances and challenges in visual information retrieval. Key points include:
- Visual information retrieval aims to find relevant images/videos based on visual and text queries, addressing the "semantic gap" between low-level features and high-level meanings.
- Advances include improved text-based, content-based, and mixed search methods, as well applications in medical image retrieval and mobile visual search.
- Ongoing challenges include capturing image similarity, addressing various representation gaps, understanding user intentions, and developing broad domain solutions.
A Comparative Study of Content Based Image Retrieval Trends and ApproachesCSCJournals
Content Based Image Retrieval (CBIR) is an important step in addressing image storage and management problems. Latest image technology improvements along with the Internet growth have led to a huge amount of digital multimedia during the recent decades. Various methods, algorithms and systems have been proposed to solve these problems. Such studies revealed the indexing and retrieval concepts, which have further evolved to Content-Based Image Retrieval. CBIR systems often analyze image content via the so-called low-level features for indexing and retrieval, such as color, texture and shape. In order to achieve significantly higher semantic performance, recent systems seek to combine low-level with high-level features that contain perceptual information for human. Purpose of this review is to identify the set of methods that have been used for CBR and also to discuss some of the key contributions in the current decade related to image retrieval and main challenges involved in the adaptation of existing image retrieval techniques to build useful systems that can handle real-world data. By making use of various CBIR approaches accurate, repeatable, quantitative data must be efficiently extracted in order to improve the retrieval accuracy of content-based image retrieval systems. In this paper, various approaches of CBIR and available algorithms are reviewed. Comparative results of various techniques are presented and their advantages, disadvantages and limitations are discussed.
Multimodal Analysis for Bridging Semantic Gap with Biologically Inspired Algo...techkrish
The amount and complexity of digital media being generated, stored, transmitted, analysed and accessed has exponentially increased as a result of advances in computer and Web technologies. Much of this information combines digital images, video, audio, graphics and textual data. Large-scale online video repositories enable users to creatively share material along a wide audience. Consequently, there is an increasing interest in associating media items with free-text annotations, ranging from simple titles and detailed descriptions of the video content. In an effort to reduce the complexity of the annotation task, this talk will outline some of the techniques developed for indexing large-scale multimedia repositories by exploiting multi-modality of information space. One such approach combines the use of semantic expansion and visual analysis for predicting user tags for online videos. The framework is designed to exploit visual features using biologically inspired algorithms and associated textual metadata, which is semantically, expanded using complementary textual resources. The experimental results indicate the usefulness of the proposed approach for analysing large-scale media items.
UNSUPERVISED VISUAL HASHING WITH SEMANTIC ASSISTANT FOR CONTENT-BASED IMAGE R...Nexgen Technology
This document proposes a novel unsupervised visual hashing approach called semantic-assisted visual hashing (SAVH) to improve content-based image retrieval. SAVH leverages rich semantics from auxiliary texts associated with images to boost visual hashing performance without requiring explicit semantic labels. It develops a unified framework to learn hash codes by preserving visual similarities between images, integrating semantic assistance from texts, and characterizing correlations between images and shared topics. Experimental results on several datasets show SAVH achieves superior performance over state-of-the-art techniques by effectively utilizing semantics from texts to assist visual hashing.
The Real Problem of Bridging the Multimedia “Semantic Gap” jrs21
WWW-2007 Panel Position:
- Since video search is visual, the semantic spaces should be defined visually as well
- Create large multimedia knowledge-base with exemplar content representing all semantic concepts relevant for search
- Allow semantics space to evolve from end-user perspectives (across sports, entertainment, news)
- Allow technology to focus on extracting the relevant semantics – truly providing the needed data-driven approach for bridging the multimedia semantic gap
Content-Based Image Retrieval (CBIR) systems employ colour as primary feature with texture and shape as secondary features. In this project a simple, image retrieval system will be implemented
The document discusses content-based image retrieval (CBIR) systems. It describes how CBIR systems use feature extraction to search large image databases based on visual content. The key components of CBIR systems are feature extraction, indexing, and system design. Feature extraction involves extracting information about images' colors, textures, shapes, and spatial locations. Effective features and indexing techniques are needed to make CBIR scalable for large image collections. Performance is evaluated based on how well systems return relevant images.
Literature Review on Content Based Image RetrievalUpekha Vandebona
This document summarizes a literature review on content-based image retrieval (CBIR). It discusses how CBIR uses computer vision techniques to automatically extract visual features from images for retrieval, unlike traditional concept-based methods that rely on metadata/text. The key visual features discussed are color, texture, and shape. A typical CBIR system architecture includes creating an image database, automatically extracting features, searching by example or semantics, and ranking results. Distance measures are used to compare image features and evaluate retrieval performance. Combining CBIR with concept-based techniques could improve image retrieval overall.
Visual search, also known as content-based image retrieval, allows users to search for images using either text queries, visual queries by uploading an example image, or visual queries by drawing an image. It has many applications including searching product catalogs, maps, photo archives, and for law enforcement. A visual search system typically uses low-level image descriptors for color, texture, shape and spatial layout to extract machine-understandable features from images. It then calculates similarity distances between images and indexes them to allow efficient searching. Performance is measured using precision and recall metrics. Existing visual search engines can still struggle with semantic gaps between low-level features and high-level human concepts.
This document outlines a presentation on content-based image retrieval (CBIR). It discusses the motivation for CBIR by describing limitations of text-based image retrieval, such as problems with image annotation, human perception, and queries that cannot be described with text. CBIR allows images to be retrieved based on automatically extracted visual features like color, texture, and histograms. A typical CBIR system extracts image features and then matches features to find visually similar images. Applications of CBIR include crime prevention, security, medical diagnosis, and intellectual property. The conclusion states that CBIR reduces computation time and increases user interaction compared to other methods.
Content Based Image and Video Retrieval AlgorithmAkshit Bum
The document describes content-based image and video retrieval (CBIR) algorithms. It discusses how CBIR works by extracting features from query images, indexing images, and retrieving similar images based on color, shape, and texture features. CBIR techniques include reverse image search, semantic retrieval using queries, and relevance feedback to refine searches based on user input about retrieved images. The document provides examples of CBIR applications in areas like crime prevention, military, web searching, and medical diagnosis.
The project aims at development of efficient segmentation method for the CBIR system. Mean-shift segmentation generates a list of potential objects which are meaningful and then these objects are clustered according to a predefined similarity measure. The method was tested on benchmark data and F-Score of .30 was achieved.
- Content-based image retrieval (CBIR) searches for images based on visual features like color, texture, and shape rather than keywords.
- CBIR systems extract features from images to create metadata and use those features to calculate visual similarity between images.
- Relevance feedback allows users to provide feedback on initial search results to help the system recalculate feature weights and improve subsequent results.
Content-based image retrieval (CBIR) uses visual image content to search large image databases according to user needs. CBIR systems represent images by extracting features related to color, shape, texture, and spatial layout. Features are extracted from regions of the image and compared to features of images in the database to find the most similar matches. CBIR has applications in medical imaging, fingerprints, photo collections, and more. Techniques include representing images with histograms of color and texture features extracted through transforms.
Vitamins are complex substances that regulate body processes and act as coenzymes in enzyme reactions. They are named alphabetically and classified as either fat-soluble or water-soluble. Fat-soluble vitamins like A, D, E and K can be stored in the body, while water-soluble vitamins like C and the B vitamins cannot be stored and must be replenished regularly through diet. Deficiencies in vitamins can cause a variety of symptoms and even permanent damage if left untreated. The document provides details on the functions, dietary sources, deficiency symptoms and risk factors for various vitamins.
This document discusses a project involving the development of a new type of aircraft called the WHITNEY S201. It mentions components like the MmLC2 AMM engine and provides technical specifications like a maximum takeoff weight of 41,000 pounds. The document also notes this project is still in development with further testing planned.
The document summarizes the key strategies and tactics used in Barack Obama's successful 2008 presidential campaign. It outlines how the campaign focused on themes of change, unity, reform, honesty and hope. It emphasizes expanding the electoral map, embracing new technologies, and expanding the electorate by appealing to youth and Latino voters. Advertising was tailored to different demographic groups and coordinated across paid, earned, and social media to turn interested voters into active supporters. The integrated approach helped Obama win the election.
Presented at DocTrain East 2007 by Joe Gollner, Stilo International -- This workshop will introduce participants to S1000D, a rapidly evolving standard that has gained growing level of adoption as a shared approach to addressing the wide range of requirements associated with planning, creating, managing, publishing and exchanging documentation for complex equipment systems. The workshop will provide guidelines for assessing the applicability of S1000D and an implementation framework for managing S1000D deployments. The following topics will receive specific attention:
* An overview of S1000D, its purpose and history
* A review of the S1000D schema framework
* A closer look at specific models
* The underlying identification and management schemes
* Recent changes and future directions
* Implementation examples
* Criteria for determining if S1000D is right for you
* Key considerations to keep in mind when implementing S1000D
* Comparing S1000D with other standards (e.g., DITA)
The document discusses a contact analysis of a deep groove ball bearing using ANSYS finite element software. A 3D parametric model of the bearing was created using APDL. Contact pairs between the inner/outer rings and balls were defined. Boundary conditions fixing the outer ring and applying a radial load to the inner ring were specified. A nonlinear contact analysis was performed and results for von Mises stress, strain, contact area shape, and penetration were found to be consistent with Hertzian contact theory values, validating the model and analysis. The contact analysis provides a scientific basis for optimizing bearing design under complex loads.
OGC spet 2010 Meta-propagation of uncertainties within workflowsDidier, G. Leibovici
To begin with let us quote the QA4EO (Quality Assurance for Earth Observation)1:
“If the vision of GEOSS is to be achieved, Quality Indicators (QIs) should be ascribed to data and, in particular, to delivered information products, at each stage of the data processing chain - from collection and processing to delivery. A QI should provide sufficient information to allow all users to readily evaluate a product’s suitability for their particular application, i.e. its “fitness for purpose”. To ensure that this process is internationally harmonised and consistent, the QI needs to be based on a documented and quantifiable assessment of evidence demonstrating the level of traceability to internationally agreed (where possible SI) reference standards. Such standards may be manmade, natural or intrinsic in nature. The documented evidence should include a description of the processes used, together with an uncertainty budget (or other appropriate quality performance measure).The guidelines of QA4EO provide a template and guidance on how to achieve this in a harmonised and robust manner. “
For interoperability purposes, each data and process registered within EuroGEOSS possesses appropriate metadata elements. The metadata description and the semantics attached to each component of a workflow (datasets and processing services) allow updating/swapping of these components. With varying quality of the components of the workflow, the quality of the outputs of this workflow can become unreliable. With the knowledge of the level of uncertainty in each dataset involved and the sensitivity aspects of the processing steps it is possible to define the quality of a workflow and the level of uncertainty of the outputs by error propagation principles.
Reusing of a given model encapsulated in a scientific workflow implies running the workflow using either the same datasets but not necessarily coming from the same sources, or different datasets which have also not necessarily the required/desired scale specified by the workflow. From error propagation principles and the knowledge of the quality metadata of the components of the workflow, using datasets from different sources or at different scales can be assessed for the quality of the workflow. As part of the integrated modelling activity the latter assessment will help the modeller in choosing the appropriate datasets or in refining the workflow model for example by considering data assimilation, downscaling, multiple scale integration steps within the scientific model and its associated workflow. The workflow quality assessment will help also the modeller in swapping or refining the processing steps as well. Under these modelling activities, the workflow is then seen as the concrete support of a conceptual model, which evolves as the conceptual model does.
On top of quality descriptors existing in the ISO19157, the present document describes the requirements for uncertainty analysis within scientific workflows.
I Minds2009 Health Decision Support Prof Bart De Moor (Ibbt Esat Ku Leuven)imec.archive
This document discusses trends and opportunities in health decision support systems. It notes the exponential growth of data from technologies like genomics and imaging. This data tsunami creates opportunities for advanced decision support through integration of heterogeneous data sources. Multimodal imaging data and gene prioritization are examples given. The document also discusses building clinical decision support systems, policy decision support, and embedded decision support systems. It outlines several areas for further research and development like information security, population data mining, home health monitoring, and advanced signal processing.
This document outlines a master's project that aims to apply 2-Dimensional Digital Image Correlation (2D-DIC) to map bond strain and stress distribution in concrete pull-out specimens. Eleven concrete specimens with varying bar diameters and fiber contents were tested. 2D-DIC analysis was used to find displacement fields from images taken during testing, which were then used to calculate strain and stress distributions. Results showed good agreement between 2D-DIC displacements and measurements from LVDT sensors. Strain contours were mapped for two selected specimens.
The document discusses expressive gesture generation for the NAO robot. It aims to 1) generate communicative gestures integrated within an existing virtual agent platform and 2) focus on expressivity and synchronization of gestures with speech. The methodology includes building a gesture library from video, using a common framework to control virtual and physical agents, and specifying gestures symbolically to convey meaning while accounting for different embodiments.
Genmab reported financial results for the first nine months of 2011, with revenue of DKK 258 million and a net loss of DKK 553 million. Operating expenses decreased 22% year-over-year due to lower R&D costs. GSK sales of ofatumumab increased 45% year-over-year. Genmab updated 2011 guidance and objectives, including maximizing ofatumumab value, evaluating opportunities for zalutumumab, advancing daratumumab, and expanding the pipeline.
Lithography technology and trends for « Semiconductor frontier » held by Aman...Yole Developpement
Lithography technology and trends for « Semiconductor frontier »
Mask aligners are the fastest lithography technology
Stepper technology provides the best resolution
Key requirements for Advanced Packaging
LED manufacturers use small diameter wafers (2”, 3”, 4” or 6”) and transition more rapidly than traditional semiconductor’s industry to larger diameters
WAFER SIZE
Wafer bow can reach up to 50μm for 2” wafers and 100μm for 4”, inducing pattern distortion.
WAFER BOW
2”
4”
6”
LED manufacturers can use different substrates, mostly sapphire or SiCwafers, which are transparent with light-diffusing features such as rough or patterned surfaces. Also, they can use metal wafers for vertical structures, so there’s large material variability.
This document discusses methods for identifying and reporting on intellectual capital (IC). It outlines both manual and automatic content analysis approaches, noting their respective advantages and disadvantages. Manual coding allows for consideration of IC complexity but is time-consuming, while automatic methods save time but cannot account for nuances. The document also examines the dynamic and complex nature of IC. It proposes establishing an IC taxonomy and developing IC reporting that recognizes complexity, establishes an IC ecosystem, and is dynamic over time. The goal is to effectively locate, measure, and mobilize a company's IC.
D3 manufactures high-quality LED displays and digital signage systems. They offer indoor and outdoor displays in a variety of sizes with pixel pitches ranging from 6mm to 24mm. D3 displays have advanced features such as high image quality, reliable operation, energy efficiency, and easy maintenance. Their content management systems allow clients to display video, graphics, and other media on their LED signs. Major D3 clients include ABC Studios, Walgreens, JVC, Forever 21, Disney Stores, and various advertisers.
Sustainable growth in a sustained crisis - the business model as a tool to in...Kasper Roldsgaard
Research Festival 2012, April 19-20. Copenhagen, Denmark
Includes some examples of business model innovation and examples of companies that haven't innovated their business model.
This document discusses a study that used tagged magnetic resonance imaging (MRI) in the OsiriX platform to estimate myocardial torsion in the left ventricle. Tagged MRI was used to track motion over time by overlaying a grid pattern on the heart that deforms with cardiac motion. Researchers developed a plugin for OsiriX to segment the left ventricle boundaries and track motion between frames using harmonic phase flow imaging. They validated their method on data from 48 subjects, finding similar torsion trends but noisy results for some acquisitions. Future work will focus on quality indicators to improve reliability.
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey WaveletCSCJournals
Lip movement is an useful way to communicate with machines and it is extremely helpful in noisy environments. However, the recognition of lip motion is a difficult task since the region of interest (ROI) is nonlinear and noisy. In the proposed lip reading method we have used two stage feature extraction mechanism which is précised, discriminative and computation efficient. The first stage is to convert video frame data into 3 dimension space and the second stage trims down the raw information space by using 3 Dimension Discrete Wavelet Transform (DWT). These features are smaller in size to give rise a novel lip reading system. In addition to the novel feature extraction technique, we have also compared the performance of Back Propagation Neural Network (BPNN) and Support Vector Machine(SVM) classifier. CUAVE database and Tulips database are used for experimentation. Experimental results show that 3-D DWT feature mining is better than 2-D DWT. 3-D DWT with Dmey wavelet results are better than 3-D DWT Db4. Results of experimentation show that 3-D DWT-Dmey along with BNNN classifier outperforms SVM.
The document discusses the GS100 survey of global IT and business process outsourcing providers. It provides an overview of the 2010 survey results, including geographic coverage, industry sizes, revenue growth from 2008-2009, new customer contracts by size and sector, global delivery locations and headcounts, top employers, and common management practices. A model for service providers emphasizes customer maturity, global delivery maturity, services portfolio breadth, and management excellence.
Designing at 2x nanometers Some New Problems Appear & Some Old Ones Remainchiportal
Designing at the 2x nanometer scale presents new challenges. Some key challenges include increased complexity, higher power consumption, and difficulties with lithography at smaller scales. Potential solutions explored include non-planar transistor structures, double patterning lithography, and 3D stacking through silicon interposers. Tools are being enhanced to support these new device structures and integration approaches to continue scaling to smaller nodes.
This document describes a time-domain surface grinding model developed for dynamic simulation purposes. The model accounts for machine tool dynamics, grinding force modeling, and workpiece representation. An experimental campaign was used to calibrate and validate the model. Tests with introduced wheel unbalances showed good agreement between measured and simulated forces. Some evidence of a non-regenerative instability was also observed during single-pass grinding tests. Future work will involve analytical study of this instability and inclusion of wheel wear modeling.
This document discusses software tools for umbilical design, analysis, and modeling. UmbiliCAD is introduced as a cross-section design tool that allows early analysis of umbilical properties. Helica is then presented as a stress analysis tool that can be used with UmbiliCAD for more advanced analysis, including load sharing, stiffness properties, fatigue analysis, and bending behavior accounting for friction. The document demonstrates how UmbiliCAD and Helica can be used together for a bundled umbilical design workflow, from cross-section design to local and global analysis.
Similar to Image retrieval: challenges and opportunities (18)
Image Processing and Computer Vision in iOSOge Marques
- Image processing and computer vision applications are becoming more common on mobile devices like the iPhone and iPad. There are many opportunities to build successful apps that can improve how users work with images and videos.
- The talk provided an overview of developing image and computer vision apps for iOS, including recommended tools like Core Image and OpenCV. It also offered advice on focusing an app idea on solving a specific problem and being aware of competition and market timing.
- Mobile image processing and computer vision have a promising future, and there is a need for good solutions to specific problems in this area that developers can work on building.
Advances and Challenges in Visual Information Search and Retrieval (WVC 2012 ...Oge Marques
Part I – Concepts, challenges, and state of the art
Part II – Medical image retrieval
Part III – Mobile visual search
Part IV – Where is image search headed?
Mobile Visual Search (MVS) is a fascinating research field that has the potential to impact how visual data is organized, annotated, and retrieved using mobile devices. The document outlines opportunities in MVS, basic concepts, and technical aspects of MVS systems. It discusses the MVS pipeline including descriptor extraction, interest point detection, feature descriptor computation, feature indexing/matching, and geometric verification. Challenges of MVS like low latency, robust recognition, and handling broad/narrow domains are also covered. The Compressed Histogram of Gradients (CHoG) descriptor is presented as an example of a compact descriptor designed for MVS.
Image Processing and Computer Vision in iPhone and iPadOge Marques
This document provides an overview of image processing and computer vision applications for the iPhone and iPad. It discusses the growing market for mobile apps in this field and the technical capabilities of iPhone devices. The document outlines a mini-course on developing iPhone and iPad apps for image processing and computer vision. It covers fundamentals of iOS development like Xcode, Objective-C, classes and objects, and the model-view-controller design pattern. It also discusses OpenCV and examples of commercial apps and student projects.
Oge Marques (FAU) - invited talk at WISMA 2010 (Barcelona, May 2010)Oge Marques
- Image search and retrieval remains a challenging problem with many open issues even 10 years after it was deemed to be past its early years.
- While progress has been made in areas like datasets, benchmarks and interfaces, core problems around similarity, semantics, and bridging the semantic gap between low-level visual features and high-level concepts remain largely unsolved.
- Narrowing domains and combining content-based techniques with metadata and user involvement through tagging and feedback may provide more successful solutions going forward.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
ScyllaDB is making a major architecture shift. We’re moving from vNode replication to tablets – fragments of tables that are distributed independently, enabling dynamic data distribution and extreme elasticity. In this keynote, ScyllaDB co-founder and CTO Avi Kivity explains the reason for this shift, provides a look at the implementation and roadmap, and shares how this shift benefits ScyllaDB users.
Discover top-tier mobile app development services, offering innovative solutions for iOS and Android. Enhance your business with custom, user-friendly mobile applications.
"Scaling RAG Applications to serve millions of users", Kevin GoedeckeFwdays
How we managed to grow and scale a RAG application from zero to thousands of users in 7 months. Lessons from technical challenges around managing high load for LLMs, RAGs and Vector databases.
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor IvaniukFwdays
At this talk we will discuss DDoS protection tools and best practices, discuss network architectures and what AWS has to offer. Also, we will look into one of the largest DDoS attacks on Ukrainian infrastructure that happened in February 2022. We'll see, what techniques helped to keep the web resources available for Ukrainians and how AWS improved DDoS protection for all customers based on Ukraine experience
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Must Know Postgres Extension for DBA and Developer during MigrationMydbops
Mydbops Opensource Database Meetup 16
Topic: Must-Know PostgreSQL Extensions for Developers and DBAs During Migration
Speaker: Deepak Mahto, Founder of DataCloudGaze Consulting
Date & Time: 8th June | 10 AM - 1 PM IST
Venue: Bangalore International Centre, Bangalore
Abstract: Discover how PostgreSQL extensions can be your secret weapon! This talk explores how key extensions enhance database capabilities and streamline the migration process for users moving from other relational databases like Oracle.
Key Takeaways:
* Learn about crucial extensions like oracle_fdw, pgtt, and pg_audit that ease migration complexities.
* Gain valuable strategies for implementing these extensions in PostgreSQL to achieve license freedom.
* Discover how these key extensions can empower both developers and DBAs during the migration process.
* Don't miss this chance to gain practical knowledge from an industry expert and stay updated on the latest open-source database trends.
Mydbops Managed Services specializes in taking the pain out of database management while optimizing performance. Since 2015, we have been providing top-notch support and assistance for the top three open-source databases: MySQL, MongoDB, and PostgreSQL.
Our team offers a wide range of services, including assistance, support, consulting, 24/7 operations, and expertise in all relevant technologies. We help organizations improve their database's performance, scalability, efficiency, and availability.
Contact us: info@mydbops.com
Visit: https://www.mydbops.com/
Follow us on LinkedIn: https://in.linkedin.com/company/mydbops
For more details and updates, please follow up the below links.
Meetup Page : https://www.meetup.com/mydbops-databa...
Twitter: https://twitter.com/mydbopsofficial
Blogs: https://www.mydbops.com/blog/
Facebook(Meta): https://www.facebook.com/mydbops/
GraphRAG for LifeSciences Hands-On with the Clinical Knowledge Graph
Image retrieval: challenges and opportunities
1. Image retrieval:
challenges and opportunities
Oge Marques
Florida Atlantic University
Boca Raton, FL - USA
June
4,
2012
UTFPR
Curi3ba,
PR
-‐
Brazil
3. Google Goggles
• Mobile visual search (MVS) solution
– Android and iPhone
– Narrow-domain search and retrieval
h@p://www.google.com/mobile/goggles
Oge
Marques
4. Outline
• How does it work?
• Why is it relevant?
• What else is going on?
• Which challenges and opportunities lie ahead?
Oge
Marques
6. Fundamentals
• Google Goggles is (one of) the first – and maybe
the best-known – solution for MVS
• It is a contemporary example of content-based
image retrieval (CBIR)
• Its technical details (algorithms, etc.) are not
publicly available
• However…
Oge
Marques
7. MVS: Pipeline for image retrieval
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
9. MVS: descriptor extraction
• Interest point detection
• Feature descriptor computation
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
10. Interest point detection
• Numerous interest-point detectors have been proposed in
the literature:
– Harris Corners (Harris and Stephens 1988)
– Scale-Invariant Feature Transform (SIFT) Difference-of-Gaussian
(DoG) (Lowe 2004)
– Maximally Stable Extremal Regions (MSERs) (Matas et al. 2002)
– Hessian affine (Mikolajczyk et al. 2005)
– Features from Accelerated Segment Test (FAST) (Rosten and
Drummond 2006)
– Hessian blobs (Bay, Tuytelaars and Van Gool 2006)
• Different tradeoffs in repeatability and complexity
• See (Mikolajczyk and Schmid 2005) for a comparative
performance evaluation of local descriptors in a common
framework.
Girod
et
al.
IEEE
Signal
Processing
Magazine
2011
Oge
Marques
11. Feature descriptor computation
• After interest-point detection, we compute a
visual word descriptor on a normalized patch.
• Ideally, descriptors should be:
– robust to small distortions in scale, orientation, and
lighting conditions;
– discriminative, i.e., characteristic of an image or a small
set of images;
– compact, due to typical mobile computing constraints.
Girod
et
al.
IEEE
Signal
Processing
Magazine
2011
Oge
Marques
12. Feature descriptor computation
• Examples of feature descriptors in the literature:
– SIFT (Lowe 1999)
– Speeded Up Robust Feature (SURF) interest-point
detector (Bay et al. 2008)
– Gradient Location and Orientation Histogram (GLOH)
(Mikolajczyk and Schmid 2005)
– Compressed Histogram of Gradients (CHoG)
(Chandrasekhar et al. 2009, 2010)
• See (Winder, (Hua,) and Brown CVPR 2007, 2009) and
(Mikolajczyk and Schmid PAMI 2005) for comparative
performance evaluation of different descriptors.
Girod
et
al.
IEEE
Signal
Processing
Magazine
2011
Oge
Marques
13. Feature descriptor computation
• What about compactness?
– Option 1: Compress off-the-shelf descriptors.
• Result: poor rate-constrained image-retrieval
performance.
– Option 2: Design a descriptor with compression in
mind.
– Example: CHoG (Compressed Histogram of Gradients)
(Chandrasekhar et al. 2009, 2010)
Girod
et
al.
IEEE
Signal
Processing
Magazine
2011
Oge
Marques
14. CHoG: Compressed Histogram of Gradients
Gradients
Gradient distributions
Patch
for each bin
dx
dy
dx
dy
011101
Spatial
0100101
binning
01101
101101
Histogram
0100011
111001
compression
0010011
01100
1010100
CHoG
Descriptor
Bernd Girod: Mobile Visual Search
Chandrasekhar
et
al.
CVPR
09,10
Oge
Marques
15. CHoG: Compressed Histogram of Gradients
[3B2-9] mmu2011030086.3d 30/7/011 16:27 Page 92
• Performance evaluation
– Recall vs. bit rate
Industry and Standards
100
features, as they arrive.15 On
98 finds a result that has sufficien
ing score, it terminates the searc
96 ately sends the results back. T
optimization reduces system
Classification accuracy (%)
94
other factor of two.
92 Overall, the SPS system dem
using the described array of tec
90 bile visual-search systems can ac
ognition accuracy, scale to re
88
databases, and deliver search r
86 ceptable time.
84 Send feature (CHoG) Emerging MPEG standard
Send image (JPEG) As we have seen, key compo
82
Send feature (SIFT) gies for mobile visual search alr
80 we can choose among several p
100 101 102
tures to design such a system. W
Query size (Kbytes)
these options at the beginnin
Figure 7. Comparison of different schemes with regard to classification The architecture shown in Figur
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
est one to implement on a mobi
accuracy and query size. CHoG descriptor data is an order of magnitude
smaller compared to JPEG images or uncompressed SIFT descriptors. requires fast networks such as W
good performance. The archite
16. MVS: feature indexing and matching
• Goal: produce a data structure that can quickly return a short
list of the database candidates most likely to match the query
image.
– The short list may contain false positives as long as the correct match
is included.
– Slower pairwise comparisons can be subsequently performed on just
the short list of candidates rather than the entire database.
• Example of a technique: Vocabulary Tree (VT)-Based Retrieval
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
17. MVS: geometric verification
• Goal: use location information of features in
query and database images to confirm that the
feature matches are consistent with a change in
viewpoint between the two images.
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
18. ik2, c, ikNk 6 is sorted, it is more
utive ID differences 5 dk1 5 ik1,
es. is used to encode the inverted index.
2 ik1Nk 212 6 in place of the IDs. This
dex [58] can significantly reduce
cting recognition accuracy. First, [64] and recursive bottom-up complete (RBUC) code [65] have
been shown to be at least ten times faster in decoding than
MVS: geometric verification
AC, while achieving comparable compression gains as AC. The
carryover and RBUC codes attain these speedups by enforcing
ed in text retrieval [62]. Second, word-aligned memory accesses.
n be quantized to a few repre- Figure S6(a) compares the memory usage of the invert-
• Method: perform ed index with and without feature descriptorsRBUC evaluate
Max quantization. Third, the dis- pairwise matching of compression using the and
ces and visit counts are far from code. Index compression reduces memory usage from near-
geometricrate ly 10 GBof correspondences.
coding can be much more
consistency to 2 GB. This five times reduction leads to a sub-
• Techniques:
oding. Using the distributions of stantial speedup in server-side processing, as shown in
counts, each inverted list can be Figure S6(b). Without compression, the large inverted
c code (AC) [63]. The geometricindex causes swapping between main anddatabase image is usually
– Since keeping transform between the query and virtual memory estimated
very important for interactive regression down the retrieval engine. After compression,
using robust and slows techniques such as:
ions, a scheme that allows ultra- sample consensus (RANSAC) (Fischlermemory congestion
• Random memory swapping is avoided and and Bolles 1981)
red over AC. The carryover code delays no longer contribute to the query latency.
• Hough transform (Lowe 2004)
– The transformation is often represented by an affine mapping or a homography.
• Note: GV is computationally expensive, which is why it’s only used for a subset
of images selected during the feature-matching stage.
onsistency checks to rerank
tion and scale information of
[53] and [69] propose incor-
tion into the VT matching or
71], the authors investigate
stimation itself. Philbin et al.
atching features to propose
c transformation model and
hypotheses. Weak geometric
cally used to rerank a larger
ore a full GVt
al.
Iperformed on011
Girod
e is EEE
Mul3media
2 Oge
Marques
[FIG4] In the GV step, we match feature descriptors pairwise and
find feature correspondences that are consistent with a geometric
add a geometric reranking step
20. Relevance
• Explosive growth and increasing popularity of
mobile devices and apps
• (Finally!) a good use case for CBIR
• Many commercial opportunities
Oge
Marques
21. Mobile visual search: driving factors
• Age of mobile computing
h@p://60secondmarketer.com/blog/2011/10/18/more-‐mobile-‐phones-‐than-‐toothbrushes/
Oge
Marques
22. Mobile visual search: driving factors
• Why do I need a camera? I have a smartphone…
(22 Dec 2011)
h@p://www.cellular-‐news.com/story/52382.php
Oge
Marques
23. Mobile visual search: driving factors
• Powerful devices
1 GHz ARM
Cortex-A9
processor,
PowerVR
SGX543MP2,
Apple A5 chipset
h@p://www.apple.com/iphone/specs.html
h@p://www.gsmarena.com/apple_iphone_4s-‐4212.php
Oge
Marques
24. Mobile visual search: driving factors
• Powerful devices
h@p://europe.nokia.com/PRODUCT_METADATA_0/Products/Phones/8000-‐series/808/Nokia808PureView_Whitepaper.pdf
h@p://www.nokia.com/fr-‐fr/produits/mobiles/808/
Oge
Marques
25. Mobile visual search: driving factors
• Instagram:
– 50 million registered users (35 M in last four
months)
– 7 employees
– A (growing ecosystem) based on it!
• Search
• Send postcards
• Manage your photos
• Build a poster
• etc.
– Sold to Facebook (for $ 1 Billion !)
earlier this year
h@p://thenextweb.com/apps/2011/12/07/instagram-‐hits-‐15m-‐users-‐and-‐has-‐2-‐people-‐working-‐on-‐an-‐android-‐app-‐right-‐now/
h@p://www.nuwomb.com/instagram/
Oge
Marques
26. Search system, a low-latency interactive visual search system. base and is the key to very fast retr
Several sidebars in this article invite the interested reader to dig features they have in common wit
deeper into the underlying algorithms. of potentially similar images is sele
Finally, a geometric verificatio
Mobile visual search: driving factors
ROBUST MOBILE IMAGE RECOGNITION
Today, the most successful algorithms for content-based image
most similar matches in the datab
spatial pattern between features of
retrieval use an approach that is referred to as bag of features didate database image to ensure
(BoFs) or bag of words (BoWs). The BoW idea is borrowed from Example retrieval systems are pres
• A natural use case for CBIR with QBE (at last!)
text retrieval. To find a particular text document, such as a Web
page, it is sufficient to use a few well-chosen words. In the
For mobile visual search, ther
to provide the users with an int
– The example is right in front of the user!
database, the document itself can be likewise represented by a deployed systems typically transm
the server, which might require t
large databases, the inverted file in
memory swapping operations slow
ing stage. Further, the GV step
and thus increases the response t
the retrieval pipeline in the follow
the challenges of mobile visual se
Query Feature
Image Extraction
[FIG2] A Pipeline for image retrieva
from the query image. Feature mat
[FIG1] A snapshot of an outdoor mobile visual search system images in the database that have m
being used. The system augments the viewfinder with with the query image. The GV step
information about the objects it recognizes in the image taken feature locations that cannot be pl
with a camera phone. in viewing position.
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
27. MVS: commercial opportunities
• Example app (La Redoute by pixlinQ)
h@p://www.youtube.com/watch?v=qUZCFtc42Q4
Oge
Marques
29. Context
• Research: datasets and groups
• Standardization: MPEG CDVS efforts
• Commercial: main players (so far)
Oge
Marques
30. Datasets for MVS research
• Stanford Mobile Visual Search Data Set
(http://web.cs.wpi.edu/~claypool/mmsys-dataset/2011/stanford/)
– Key characteristics:
• rigid objects
• widely varying lighting conditions
• perspective distortion
• foreground and background clutter
• realistic ground-truth reference data
• query data collected from heterogeneous low and high-end
camera phones.
Chandrasekhar
et
al.
ACM
MMSys
2011
Oge
Marques
31. SMVS Data Set: categories and examples
• DVD covers
h@p://web.cs.wpi.edu/~claypool/mmsys-‐2011-‐dataset/stanford/mvs_images/dvd_covers.html
Oge
Marques
32. SMVS Data Set: categories and examples
• CD covers
h@p://web.cs.wpi.edu/~claypool/mmsys-‐2011-‐dataset/stanford/mvs_images/cd_covers.html
Oge
Marques
33. SMVS Data Set: categories and examples
• Museum paintings
h@p://web.cs.wpi.edu/~claypool/mmsys-‐2011-‐dataset/stanford/mvs_images/museum_pain3ngs.html
Oge
Marques
34. Other MVS data sets
ISO/IEC
JTC1/SC29/WG11/N12202
-‐
July
2011,
Torino,
IT
Oge
Marques
35. MPEG Compact Descriptors for Visual Search (CDVS)
• Objective
– Define a standard that enables efficient
implementation of visual search functionality on mobile
devices
• Scope
• bitstream of descriptors
• parts of descriptor extraction process (e.g. key-point
detection) needed to ensure interoperability
– Additional info:
• https://mailhost.tnt.uni-hannover.de/mailman/listinfo/cdvs
• http://mpeg.chiariglione.org/meetings/geneva11-1/geneva_ahg.htm (Ad hoc groups)
Bober,
Cordara,
and
Reznik
(2010)
Oge
Marques
36. MPEG CDVS
[3B2-9] mmu2011030086.3d 1/8/011 16:44 Page 93
• Summarized timeline
Table 1. Timeline for development of MPEG standard for visual search.
When Milestone Comments
March, 2011 Call for Proposals is published Registration deadline: 11 July 2011
Proposals due: 21 November 2011
December, 2011 Evaluation of proposals None
February, 2012 1st Working Draft First specification and test software model that can
be used for subsequent improvements.
July, 2012 Committee Draft Essentially complete and stabilized specification.
January, 2013 Draft International Standard Complete specification. Only minor editorial
changes are allowed after DIS.
July, 2013 Final Draft International Finalized specification, submitted for approval and
Standard publication as International standard.
that among several component technologies for existing standards, such as MPEG Query For-
image retrieval, such a standard should focus pri- mat, HTTP, XML, JPEG, and JPSearch.
marily on defining the format of descriptors and
Girod
et
al.
IEEE
Mul3media
2011
Oge
Marques
parts of their extraction process (such as interest Conclusions and outlook
point detectors) needed to ensure interoperabil- Recent years have witnessed remarkable
38. SnapTell
• One of the earliest (ca. 2008) MVS apps for iPhone
– Eventually acquired by Amazon (A9)
• Proprietary technique (“highly accurate and robust
algorithm for image matching: Accumulated Signed Gradient
(ASG)”).
h@p://www.snaptell.com/technology/index.htm
Oge
Marques
39. oMoby (and the IQ Engines API)
– iPhone app
h@p://omoby.com/pages/screenshots.php
Oge
Marques
40. oMoby (and the IQ Engines API)
• The IQ Engines API:
“vision as a service”
h@p://www.iqengines.com/applica3ons.php
Oge
Marques
41. Moodstocks: overview
• Offline image recognition thanks to a smart image
signatures synchronization
h@p://www.youtube.com/watch?v=tsxe23b12eU
Oge
Marques
43. MVS: technical challenges
• How to ensure low latency (and interactive
queries) under constraints such as:
– Network bandwidth
– Computational power
– Battery consumption
• How to achieve robust visual recognition in spite
of low-resolution cameras, varying lighting
conditions, etc.
• How to handle broad and narrow domains
Oge
Marques
44. Other technical challenges
• How to handle the (infamous) semantic gap
• Combination of text-based and visual queries
• Visualization of results
• Users' needs and intentions
Oge
Marques
45. The semantic gap
• The semantic gap is the lack of coincidence
between the information that one can extract
from the visual data and the interpretation that
the same data have for a user in a given situation.
• “The pivotal point in content-based retrieval is that the user
seeks semantic similarity, but the database can only provide
similarity by data processing. This is what we called the
semantic gap.” [Smeulders et al., 2000]
Oge
Marques
54. Challenge: users’ needs and intentions
• Users and developers have quite different views
• Cultural and contextual information should be
taken into account
• User intentions are hard to infer
– Privacy issues
– Users themselves don’t always know what they want
– Who misses the MS Office paper clip?
Oge
Marques
55. Concluding thoughts
(Mobile) visual search and retrieval is a fascinating
research field with many open challenges and
opportunities which have the potential to impact
the way we organize, annotate, and retrieve visual
data (images and videos).
Oge
Marques