Advances and Challenges in Visual Information Search and Retrieval (WVC 2012 ...Oge Marques
Part I – Concepts, challenges, and state of the art
Part II – Medical image retrieval
Part III – Mobile visual search
Part IV – Where is image search headed?
Advances and Challenges in Visual Information Search and Retrieval (WVC 2012 ...Oge Marques
Part I – Concepts, challenges, and state of the art
Part II – Medical image retrieval
Part III – Mobile visual search
Part IV – Where is image search headed?
Lecture 4: Social Web Personalization (2012)Lora Aroyo
This is the fourth lecture in the Social Web course at the VU University Amsterdam
Visit the website for more information: http://semanticweb.cs.vu.nl/socialweb2012/
Thanks to Fabian Abel for letting me adopt slides from his lectures
SSII2021 [SS2] Deepfake Generation and Detection – An Overview (ディープフェイクの生成と検出)SSII
SSII2021 [SS2] Deepfake Generation and Detection – An Overview (ディープフェイクの生成と検出)
6/10 (木) 14:30~15:00
講師:Huy H. Nguyen 氏(総合研究大学院大学/国立情報学研究所)
概要: Advances in machine learning and their interference with computer graphics allow us to easily generate high-quality images and videos. State-of-the-art manipulation methods enable the real-time manipulation of videos obtained from social networks. It is also possible to generate videos from a single portrait image. By combining these methods with speech synthesis, attackers can create a realistic video of some person saying something that they never said and distribute it on the internet. This results in loosing social trust, making confusion, and harming people’s reputation. Several countermeasures have been proposed to tackle this problem, from using hand-crafted features to using convolutional neural network. Some countermeasures use images as input and other leverage temporal information in videos. Their output could be binary (bona fide or fake) or muti-class (deepfake detection), or segmentation masks (manipulation localization). Since deepfake methods evolve rapidly, dealing with unseen ones is still a challenging problem. Some solutions have been proposed, however, this problem is not completely solved. In this talk, I will provide an overview on both deepfake generation and deepfake detection/localization. I will mainly focus on image and video domain and also introduce some audiovisual-based methods on both sides. Some open discussions and future directions are also included.
Using the Crowd to Understand and Adapt User InterfacesJeffrey Nichols
Keynote given at the Engineering Interactive Computing Systems (EICS) conference. June 25, 2013
Abstract: Engineering user interfaces has long implied careful design carried out using formal methods applied by human experts and automated systems. While these methods have advantages, especially for creating interfaces that have the flexibility to adapt to users and situations, they can also be time consuming, expensive, and there are relatively few experts able to apply them effectively. In particular, many engineering methods require the construction of one or more models, each of which can only be created through many hours of work by an expert. In this keynote, I will explore how social and human computation methods can be applied to reduce the barriers to achieving user interface flexibility and ultimately to using engineering methods. In a first example, I will illustrate how groups of users can work together to modify and improve user interfaces through end-user programming examples from the CoScripter and Highlight projects. I will then discuss some initial work on using a crowd of novice workers to create models of existing user interfaces. I hope this keynote will inspire the engineering community to consider alternate approaches that creatively combine formal methods with the power of crowds.
Slidedeck for the second class on Information Architecture: Part 2. This one examines the basics of how to create a set of wireframes and accessibility requirements for the Web.
The slidedeck for the fourth and final class of the Information Architecture course (Part 2) I teach at the University of Toronto's iSchool. This class covers:
- Creating a Web Style Guide
- Icons/Expression in Design
- Localization 101
- Change Management
- Creating a Functional Specification for Your CMS
Using games to improve computer vision solutionsOge Marques
Technical talk given at Alpen-Adria Universität (Klagenfurt, Austria), May 21, 2013.
Take-home message: "Solutions to many problems in computer vision can be improved using human computation, particularly through properly designed games (with a purpose)."
An introduction to GPUImage that is an excellent image processing framework for iOS. The architecture, threading model and how to leverage OpenGL ES are addressed in the slides.
Secure Multi-Party Negotiation: An Analysis for Electronic Payments in Mobile...IDES Editor
This paper is an attempt to base on auctions which
presents a frame work for the secure multi-party decision
protocols. In addition to the implementations which are very
light weighted, the main focus is on synchronizing security
features for avoiding agreements manipulations and reducing
the user traffic. Through this paper one can understand that
this different auction protocols on top of the frame work can
be collaborated using mobile devices. This paper present the
negotiation between auctioneer and the proffered and this
negotiation shows that multiparty security is far better than
the existing system.
What's a Core Image? An Image-Processing Framework on iOS and OS XFlatiron School
Flatiron students Steven Zhou and Heidi Hansen explain how core images work on iOS and OS X to help developers process images efficiently without dealing with low level interactions with GPU or CPU.

Want to know the untold secrets of imaging on iOS? This talks goes through performance considerations about a number of imaging APIs on iOS, including some examples of how we integrated them in our own apps. Image loading, processing, and display will be analysed and discussed to find best APIs for particular use cases.
Lecture 4: Social Web Personalization (2012)Lora Aroyo
This is the fourth lecture in the Social Web course at the VU University Amsterdam
Visit the website for more information: http://semanticweb.cs.vu.nl/socialweb2012/
Thanks to Fabian Abel for letting me adopt slides from his lectures
SSII2021 [SS2] Deepfake Generation and Detection – An Overview (ディープフェイクの生成と検出)SSII
SSII2021 [SS2] Deepfake Generation and Detection – An Overview (ディープフェイクの生成と検出)
6/10 (木) 14:30~15:00
講師:Huy H. Nguyen 氏(総合研究大学院大学/国立情報学研究所)
概要: Advances in machine learning and their interference with computer graphics allow us to easily generate high-quality images and videos. State-of-the-art manipulation methods enable the real-time manipulation of videos obtained from social networks. It is also possible to generate videos from a single portrait image. By combining these methods with speech synthesis, attackers can create a realistic video of some person saying something that they never said and distribute it on the internet. This results in loosing social trust, making confusion, and harming people’s reputation. Several countermeasures have been proposed to tackle this problem, from using hand-crafted features to using convolutional neural network. Some countermeasures use images as input and other leverage temporal information in videos. Their output could be binary (bona fide or fake) or muti-class (deepfake detection), or segmentation masks (manipulation localization). Since deepfake methods evolve rapidly, dealing with unseen ones is still a challenging problem. Some solutions have been proposed, however, this problem is not completely solved. In this talk, I will provide an overview on both deepfake generation and deepfake detection/localization. I will mainly focus on image and video domain and also introduce some audiovisual-based methods on both sides. Some open discussions and future directions are also included.
Using the Crowd to Understand and Adapt User InterfacesJeffrey Nichols
Keynote given at the Engineering Interactive Computing Systems (EICS) conference. June 25, 2013
Abstract: Engineering user interfaces has long implied careful design carried out using formal methods applied by human experts and automated systems. While these methods have advantages, especially for creating interfaces that have the flexibility to adapt to users and situations, they can also be time consuming, expensive, and there are relatively few experts able to apply them effectively. In particular, many engineering methods require the construction of one or more models, each of which can only be created through many hours of work by an expert. In this keynote, I will explore how social and human computation methods can be applied to reduce the barriers to achieving user interface flexibility and ultimately to using engineering methods. In a first example, I will illustrate how groups of users can work together to modify and improve user interfaces through end-user programming examples from the CoScripter and Highlight projects. I will then discuss some initial work on using a crowd of novice workers to create models of existing user interfaces. I hope this keynote will inspire the engineering community to consider alternate approaches that creatively combine formal methods with the power of crowds.
Slidedeck for the second class on Information Architecture: Part 2. This one examines the basics of how to create a set of wireframes and accessibility requirements for the Web.
The slidedeck for the fourth and final class of the Information Architecture course (Part 2) I teach at the University of Toronto's iSchool. This class covers:
- Creating a Web Style Guide
- Icons/Expression in Design
- Localization 101
- Change Management
- Creating a Functional Specification for Your CMS
Using games to improve computer vision solutionsOge Marques
Technical talk given at Alpen-Adria Universität (Klagenfurt, Austria), May 21, 2013.
Take-home message: "Solutions to many problems in computer vision can be improved using human computation, particularly through properly designed games (with a purpose)."
An introduction to GPUImage that is an excellent image processing framework for iOS. The architecture, threading model and how to leverage OpenGL ES are addressed in the slides.
Secure Multi-Party Negotiation: An Analysis for Electronic Payments in Mobile...IDES Editor
This paper is an attempt to base on auctions which
presents a frame work for the secure multi-party decision
protocols. In addition to the implementations which are very
light weighted, the main focus is on synchronizing security
features for avoiding agreements manipulations and reducing
the user traffic. Through this paper one can understand that
this different auction protocols on top of the frame work can
be collaborated using mobile devices. This paper present the
negotiation between auctioneer and the proffered and this
negotiation shows that multiparty security is far better than
the existing system.
What's a Core Image? An Image-Processing Framework on iOS and OS XFlatiron School
Flatiron students Steven Zhou and Heidi Hansen explain how core images work on iOS and OS X to help developers process images efficiently without dealing with low level interactions with GPU or CPU.

Want to know the untold secrets of imaging on iOS? This talks goes through performance considerations about a number of imaging APIs on iOS, including some examples of how we integrated them in our own apps. Image loading, processing, and display will be analysed and discussed to find best APIs for particular use cases.
Edge detection is the name for a set of mathematical methods which aim at identifying points in a digital image at which the image brightness changes sharply or, more formally, has discontinuities.
The President’s Speech in Cairo: A New Beginning - TurkishObama White House
President Obama’s speech in Cairo on America’s relationship with Muslim communities around the world. June 4th, 2009. http://www.whitehouse.gov/blog/newbeginning/
Award-winning preview of the new Graphic Novel adventure about a boy who discovers his biracial heritage in Africa.
He learns from the great warrior Shaka Zulu who somehow invented African feminism and inspired Gandhi's passive resistance.
Based on a true story. Winner of the 2009 Fuze Multimedia Storytelling contest.
For more visit http://www.whiteshaka.com
Created and storyboarded by Alan Brody
Graphics by Revo Yanson
Music: Yeah, Yeah by Draztik from the Zulu-language album Imbube
EA Draffan Assess 2010 Presentation on Accessibility & Assistive TechnologyNeil Milliken
EA Draffan's Assess 2010 Presentation on Accessibility & Assistive Technology
E.A. Draffan trained as a Speech and Language Therapist and worked in a District Hospital for several years before specialising in Dyslexia and Assistive Technology. She has since worked with disabled students in Further and Higher Education, and set up a regional Assistive Technology Centre at the University of Sussex. She is currently working at the University of Southampton on the JISC funded LexDis project as well as other projects related to the use of assistive technologies and accessibility of web 2.0 and e-learning environments. http:// www.lexdis.org
Working with Image basically based on python. It uses python libraries.With the help of this Software we can easily edit our image in just two minutes without using any extra software. It is very easy to use and also user friendly interface made him very effective for editing.
The iOS technical interview: get your dream job as an iOS developerJuan C Catalan
So you have been doing tutorials, sample projects, and watching videos on iOS development for a while. You are trying to publish an app in the App Store or maybe you got one already there. You dream of becoming a professional iOS developer.
Believe me, I was in the same situation six years ago. I started as an indie developer, self employed, and landed a few short contracts, then a six-month contract, and finally, one day, I got a job as a full-time professional iOS developer with a corporation. I have interviewed for a few companies and I have also interviewed come iOS candidates.
In this talk I will explain how to prepare yourself for the iOS technical interview. I will go thru the most usual questions, give my personal advice on how to succeed and pass the interview, and provide links to training material.
Neil Perlin - We're Going Mobile! Great! Are We Ready?LavaConConference
In this session attendees will learn:
Technical options for going mobile, including responsive design, converting traditional online help to an app, and creating a “true” app using RMAD (Rapid Mobile App Development) tools. The pros and cons of each approach and some of the tools available for creating each option.
Anticipated changes in content creation practices and workflows including the elimination of local formatting, adoption of a “mobile first” philosophy, rethinking the role of tables, and more.
How company issues like terminology standardization, strategic benefit, politics, and the development of metrics and standards can help or hinder a move to mobile.
Deep Learning is the area of machine learning and one of the most talked about trends in business and computer science today.
In this talk, I will give a review of Deep Learning explaining what it is, what kinds of tasks it can do today, and what it probably could do in the future.
Presentation to college students about creating infographics relating to course content. Explores the theory behind infographics, gives examples, profiles free tools for creating infographics, and touches on copyright and appropriate use of creative commons images.
Greens Technology is a leading Training and Placement company in Chennai. We are
known for our practical approach towards trainings that enable students to gain real-
time exposure on competitive technologies. Trainings are offered by employees from
MNCs to give a real corporate exposure.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
1. Image
Processing
and
Computer
Vision
in
iOS
Oge Marques, PhD
omarques@fau.edu
Uberaba, MG - Brazil
18 December 2013
2. Take-home message
Mobile image processing and computer vision
applications are coming of age.
There are many opportunities for building successful
apps that may improve the way we capture,
organize, edit, share, annotate, and retrieve images
and videos using iPhone and iPad.
3. Disclaimer #1
• I'm a teacher, researcher, graduate advisor, author,
…
• … not a developer
5. Disclaimer #3
• I'm an Apple fan!
– Since 2001…
• 4 iPod, 3 iPhone, 2 iPad, 2 iMac, 4
MacBook, and more (AirPort, AirPort
Express, Apple TV, etc.)
– Since 2010…
• Created and co-taught
iOS Programming classes at FAU
6.
7. In 2013…
• 1.4 billion people have a smartphone with
camera
• 350 million photos uploaded to Facebook every
day
• Instagram reaches 150
million users, with a total
of 16 billion photos
shared and 1 billion likes
each day
8. In 2013…
• Selfie was the Oxford Dictionary's new word of
the year
10. Background and Motivation
Two
sides
of
the
coin
• The maturity and
popularity of image
processing and computer
vision techniques and
algorithms
• The unprecedented
success of mobile
devices, particularly the
iPhone and the iPad
11. Motivation
• Rich capabilities of iPhone/iPad for image and
video processing
• Apple support for image and multimedia:
frameworks, libraries, etc.
• Third-party support for iPhone-based
development: open APIs, OpenCV, etc.
• Success stories and ever-growing market
12. Motivation
• Q: Why DIP and CV?
• A: Because they are still relevant and growing
fields whose techniques and can help solve many
problems.
• Q: Why iOS / mobile?
• A: Because some problems are better solved in
that context and some still need to be solved in a
away that is consistent with ergonomics (devices'
size etc.) and user needs (quick fix + filter before
sharing).
13. search. As an example, we then present the Stanford Product
Search system, a low-latency interactive visual search system.
Several sidebars in this article invite the interested reader to dig
deeper into the underlying algorithms.
each query feature vector with all t
base and is the key to very fast retr
features they have in common wit
of potentially similar images is sele
Finally, a geometric verificatio
most similar matches in the datab
spatial pattern between features of
didate database image to ensure
Example retrieval systems are pres
For mobile visual search, ther
to provide the users with an int
deployed systems typically transm
the server, which might require t
large databases, the inverted file in
memory swapping operations slow
ing stage. Further, the GV step
and thus increases the response t
the retrieval pipeline in the follow
the challenges of mobile visual se
Example: a natural use case for CBIR
ROBUST MOBILE IMAGE RECOGNITION
Today, the most successful algorithms for content-based image
retrieval use an approach that is referred to as bag of features
(BoFs) or bag of words (BoWs). The BoW idea is borrowed from
text retrieval. To find a particular text document, such as a Web
page, it is sufficient to use a few well-chosen words. In the
database, the document itself can be likewise represented by a
• Content-Based Image Retrieval (CBIR) using the
Query-By-Example (QBE) paradigm
– The example is right there, in front of the user!
Query
Image
[FIG1] A snapshot of an outdoor mobile visual search system
being used. The system augments the viewfinder with
information about the objects it recognizes in the image taken
with a camera phone.
Feature
Extraction
[FIG2] A Pipeline for image retrieva
from the query image. Feature mat
images in the database that have m
with the query image. The GV step
feature locations that cannot be pl
in viewing position.
15. iPhone photo apps
• 400+ photo- and video-related apps available in
iTunes store
– Entire sites for reviews, discussions, etc.
– Subcategories include:
• Camera enhancements
• Image editing and processing
• Image sharing
• Image printing, wireless transfer, etc.
20. Developing DIP/CV apps for iOS
• Checklist:
– Get a Mac running OS X
– Sign up to become a registered iOS developer
– Download / install xCode and latest version of iOS
SDK
– Download / install iOS simulator
– Learn Objective-C and the basics of iOS programming
– Get an iPhone, iPod Touch, or iPad (optional)
21. Developing DIP/CV apps for iOS
• Topics to study in greater depth:
– The main classes that you need to understand in order
to develop basic applications involving images, camera,
and photo library for the iPhone are:
• UIImageView
• UIImagePickerController
• UIImage
– Check out the documentation for the A/V foundation
22. Developing DIP/CV apps for iOS
• Topics to study in greater depth (cont'd):
– Learn about Core Image and its main classes:
• CIFilter: a mutable object that represents an effect. A filter
object has at least one input parameter and produces an
output image.
• CIImage: an immutable object that represents an image.
• CIContext: an object through which Core Image draws the
results produced by a filter.
23. Core Image
• Image processing and analysis technology designed to provide
near real-time processing for still and video images.
• Hides the details of low-level graphics processing by providing an
easy-to-use API.
• Brought into iOS since iOS 5 (Oct'11)
24. OpenCV
• OpenCV (Open Source Computer Vision) is
a library of programming functions for realtime computer vision.
• OpenCV is released under a BSD license; it is
free for both academic and commercial use.
• Goal: to provide a simple-to-use computer
vision infrastructure that helps people build
fairly sophisticated vision applications quickly.
• The library has 2000+ optimized algorithms.
– It is used around the world, has 2M downloads
and 40K people in the user group.
25. OpenCV
• 5 main components:
1. CV: basic image processing and higher-level computer
vision algorithms
2. ML: machine learning algorithms
3. HighGUI: I/O routines and functions for storing and
loading video and images
4. CXCore: basic data structures and content upon
which the three components above rely
5. CvAux: defunct areas + experimental algorithms; not
well-documented.
27. A bit of advice…
• Go beyond the DIP/CV and iOS boxes
– Learn about ergonomics, human factors, human psych,
HCI, UX
• Don't reinvent the wheel!
– Reuse code and ideas whenever possible
• Avoid the trap of building solutions looking for
problems
• Tackle ONE problem and solve it well!
• Beware of competition
• Beware of narrow windows of opportunity and
ephemeral success: timing is everything!
28. Learn more about it
• iOS Programming
– Apple online documentation
– Core Image
• OpenCV
– Official website
– Learning OpenCV book
– Instant OpenCV for iOS book
• Our work
– Slideshare (WVC 2011)
– Upcoming book (Springer Briefs, 2014)
29. Concluding thoughts
• Mobile image processing, image search, and computer
vision-based apps have a promising future.
• There is a great need for good solutions to specific
problems.
• I hope this talk has provided a good starting point
and many useful pointers.
• I look forward to working with some of you!
30. Let's get to work!
• Which computer vision or image processing app
would you like to build?
• Contact me with ideas: omarques@fau.edu