This document discusses deep learning techniques for person re-identification. It begins with an overview of supervised and unsupervised person re-identification. It then discusses the challenges of annotation cost and data size for re-ID. Next, it covers active learning approaches for person re-ID using human-in-the-loop feedback to incrementally train models. Finally, it discusses relationships between person re-ID and attribute learning, person detection, and multi-target multi-camera tracking.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/person-re-identification-and-tracking-at-the-edge-challenges-and-techniques-a-presentation-from-the-university-of-auckland/
Morteza Biglari-Abhari, Senior Lecturer at the University of Auckland, presents the “Person Re-Identification and Tracking at the Edge: Challenges and Techniques” tutorial at the May 2021 Embedded Vision Summit.
Numerous video analytics applications require understanding how people are moving through a space, including the ability to recognize when the same person has moved outside of the camera’s view and then back into the camera’s view, or when a person has passed from the view of one camera to the view of another. This capability is referred to as person re-identification and tracking. It’s an essential technique for applications such as surveillance for security, health and safety monitoring in healthcare and industrial facilities, intelligent transportation systems and smart cities. It can also assist in gathering business intelligence such as monitoring customer behavior in shopping environments. Person re-identification is challenging.
In this talk, Biglari-Abhari discusses the key challenges and current approaches for person re-identification and tracking, as well as his initial work on multi-camera systems and techniques to improve accuracy, especially fusing appearance and spatio-temporal models. He also briefly discusses privacy-preserving techniques, which are critical for some applications, as well as challenges for real-time processing at the edge.
Presentation for the Berlin Computer Vision Group, December 2020 on deep learning methods for image segmentation: Instance segmentation, semantic segmentation, and panoptic segmentation.
Object Detection using Deep Neural NetworksUsman Qayyum
Recent Talk at PI school covering following contents
Object Detection
Recent Architecture of Deep NN for Object Detection
Object Detection on Embedded Computers (or for edge computing)
SqueezeNet for embedded computing
TinySSD (object detection for edge computing)
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/person-re-identification-and-tracking-at-the-edge-challenges-and-techniques-a-presentation-from-the-university-of-auckland/
Morteza Biglari-Abhari, Senior Lecturer at the University of Auckland, presents the “Person Re-Identification and Tracking at the Edge: Challenges and Techniques” tutorial at the May 2021 Embedded Vision Summit.
Numerous video analytics applications require understanding how people are moving through a space, including the ability to recognize when the same person has moved outside of the camera’s view and then back into the camera’s view, or when a person has passed from the view of one camera to the view of another. This capability is referred to as person re-identification and tracking. It’s an essential technique for applications such as surveillance for security, health and safety monitoring in healthcare and industrial facilities, intelligent transportation systems and smart cities. It can also assist in gathering business intelligence such as monitoring customer behavior in shopping environments. Person re-identification is challenging.
In this talk, Biglari-Abhari discusses the key challenges and current approaches for person re-identification and tracking, as well as his initial work on multi-camera systems and techniques to improve accuracy, especially fusing appearance and spatio-temporal models. He also briefly discusses privacy-preserving techniques, which are critical for some applications, as well as challenges for real-time processing at the edge.
Presentation for the Berlin Computer Vision Group, December 2020 on deep learning methods for image segmentation: Instance segmentation, semantic segmentation, and panoptic segmentation.
Object Detection using Deep Neural NetworksUsman Qayyum
Recent Talk at PI school covering following contents
Object Detection
Recent Architecture of Deep NN for Object Detection
Object Detection on Embedded Computers (or for edge computing)
SqueezeNet for embedded computing
TinySSD (object detection for edge computing)
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
Image classification using convolutional neural networkKIRAN R
For separating the images from a large collection of images or from a large dataset this classifier can be used, Here deep neural network is used for training and classifying the images. The convolutional neural network is the most suitable algorithm for classifier images. This Classifier is a machine learning model, so the more you train it the more will be the accuracy.
발표자: 이인웅 (연세대 박사과정)
발표일: 2017.12.
개요:
영상에서 사람의 행동을 인식하는 방법은 크게 영상에서 직접적으로 행동 라벨을 추출하는 것과 자세 정보를 기반으로 행동 라벨을 추출하는 것으로 나뉠 수 있습니다.
본 발표는 행동 인식에 대한 전반적인 개요를 설명하고 그 중에서도 사람의 자세 정보를 기반으로 하는 행동 인식 기술에 초점을 두고 최근 ICCV 2017 학회에서 발표된 Temporal Sliding LSTM 네트워크를 활용한 행동 인식 기술을 중점적으로 설명합니다. 구체적으로, 스켈레톤 기반 행동 인식 이슈, 제안하는 방법과 실험 결과들이 소개되고 앞으로 나아갈 만한 새로운 연구 이슈들도 추가적으로 설명합니다.
Computer vision has received great attention over the last two decades.
This research field is important not only in security-related software but also in the advanced interface between people and computers, advanced control methods, and many other areas.
Scikit-Learn is a powerful machine learning library implemented in Python with numeric and scientific computing powerhouses Numpy, Scipy, and matplotlib for extremely fast analysis of small to medium sized data sets. It is open source, commercially usable and contains many modern machine learning algorithms for classification, regression, clustering, feature extraction, and optimization. For this reason Scikit-Learn is often the first tool in a Data Scientists toolkit for machine learning of incoming data sets.
The purpose of this one day course is to serve as an introduction to Machine Learning with Scikit-Learn. We will explore several clustering, classification, and regression algorithms for a variety of machine learning tasks and learn how to implement these tasks with our data using Scikit-Learn and Python. In particular, we will structure our machine learning models as though we were producing a data product, an actionable model that can be used in larger programs or algorithms; rather than as simply a research or investigation methodology.
This describes the supervised machine learning, supervised learning categorisation( regression and classification) and their types, applications of supervised machine learning, etc.
Real Time Object Detection with Audio Feedback using Yolo v3ijtsrd
In this paper, we propose a system that combines real time object detection using the YOLOv3 algorithm with audio feedback to assist visually impaired individuals in locating and identifying objects in their surroundings. The YOLOv3 algorithm is a state of the art object detection algorithm that has been used in numerous studies for various applications. Audio feedback has also been studied in previous research as a useful tool for assisting visually impaired individuals. Our proposed system builds on the effectiveness of both these technologies to provide a valuable tool for improving the independence and quality of life of visually impaired individuals. We present the architecture of our proposed system, which includes a YOLOv3 model for object detection and a text to speech engine for providing audio feedback. We also present the results of our experiments, which demonstrate the effectiveness of our system in detecting and identifying objects in real time. Our proposed system can be used in various settings, such as indoor and outdoor environments, and can assist visually impaired individuals in various activities such as the navigation and object identification. Dr. K. Nagi Reddy | K. Sreeja | M. Sreenivasulu Reddy | K. Sireesha | M. Triveni "Real Time Object Detection with Audio Feedback using Yolo_v3" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-7 | Issue-2 , April 2023, URL: https://www.ijtsrd.com.com/papers/ijtsrd55158.pdf Paper URL: https://www.ijtsrd.com.com/engineering/electronics-and-communication-engineering/55158/real-time-object-detection-with-audio-feedback-using-yolov3/dr-k-nagi-reddy
Image classification using convolutional neural networkKIRAN R
For separating the images from a large collection of images or from a large dataset this classifier can be used, Here deep neural network is used for training and classifying the images. The convolutional neural network is the most suitable algorithm for classifier images. This Classifier is a machine learning model, so the more you train it the more will be the accuracy.
발표자: 이인웅 (연세대 박사과정)
발표일: 2017.12.
개요:
영상에서 사람의 행동을 인식하는 방법은 크게 영상에서 직접적으로 행동 라벨을 추출하는 것과 자세 정보를 기반으로 행동 라벨을 추출하는 것으로 나뉠 수 있습니다.
본 발표는 행동 인식에 대한 전반적인 개요를 설명하고 그 중에서도 사람의 자세 정보를 기반으로 하는 행동 인식 기술에 초점을 두고 최근 ICCV 2017 학회에서 발표된 Temporal Sliding LSTM 네트워크를 활용한 행동 인식 기술을 중점적으로 설명합니다. 구체적으로, 스켈레톤 기반 행동 인식 이슈, 제안하는 방법과 실험 결과들이 소개되고 앞으로 나아갈 만한 새로운 연구 이슈들도 추가적으로 설명합니다.
Computer vision has received great attention over the last two decades.
This research field is important not only in security-related software but also in the advanced interface between people and computers, advanced control methods, and many other areas.
Scikit-Learn is a powerful machine learning library implemented in Python with numeric and scientific computing powerhouses Numpy, Scipy, and matplotlib for extremely fast analysis of small to medium sized data sets. It is open source, commercially usable and contains many modern machine learning algorithms for classification, regression, clustering, feature extraction, and optimization. For this reason Scikit-Learn is often the first tool in a Data Scientists toolkit for machine learning of incoming data sets.
The purpose of this one day course is to serve as an introduction to Machine Learning with Scikit-Learn. We will explore several clustering, classification, and regression algorithms for a variety of machine learning tasks and learn how to implement these tasks with our data using Scikit-Learn and Python. In particular, we will structure our machine learning models as though we were producing a data product, an actionable model that can be used in larger programs or algorithms; rather than as simply a research or investigation methodology.
This describes the supervised machine learning, supervised learning categorisation( regression and classification) and their types, applications of supervised machine learning, etc.
Real Time Object Detection with Audio Feedback using Yolo v3ijtsrd
In this paper, we propose a system that combines real time object detection using the YOLOv3 algorithm with audio feedback to assist visually impaired individuals in locating and identifying objects in their surroundings. The YOLOv3 algorithm is a state of the art object detection algorithm that has been used in numerous studies for various applications. Audio feedback has also been studied in previous research as a useful tool for assisting visually impaired individuals. Our proposed system builds on the effectiveness of both these technologies to provide a valuable tool for improving the independence and quality of life of visually impaired individuals. We present the architecture of our proposed system, which includes a YOLOv3 model for object detection and a text to speech engine for providing audio feedback. We also present the results of our experiments, which demonstrate the effectiveness of our system in detecting and identifying objects in real time. Our proposed system can be used in various settings, such as indoor and outdoor environments, and can assist visually impaired individuals in various activities such as the navigation and object identification. Dr. K. Nagi Reddy | K. Sreeja | M. Sreenivasulu Reddy | K. Sireesha | M. Triveni "Real Time Object Detection with Audio Feedback using Yolo_v3" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-7 | Issue-2 , April 2023, URL: https://www.ijtsrd.com.com/papers/ijtsrd55158.pdf Paper URL: https://www.ijtsrd.com.com/engineering/electronics-and-communication-engineering/55158/real-time-object-detection-with-audio-feedback-using-yolov3/dr-k-nagi-reddy
The slide is about pedestrian re-identification (re-ID) based on Deep Learning methods. I mainly review some prevalent methods and try to give some insights to the future work. Zhedong Zheng @2017.7.27
Face Association by Model Evolution: Learning people's face from weakly labelled web images. Just search the one in Goggle, eliminate irrelevant images automatically and train a classifier.
My CVPR paper...
Youtube: https://www.youtube.com/watch?v=9JeOHyQew6M
Martha Larson, Zhuoran Liu, Simon Brugman and Zhengyu Zhao, Pixel Privacy: Increasing Image Appeal while Blocking Automatic Inference of Sensitive Scene Information. Proc. of MediaEval 2018, 29-31 October 2018, Sophia Antipolis, France.
Abstract: We introduce a new privacy task focused on images that users share online. The task benchmarks image transformation algorithms that are capable of blocking the ability of automatic classifiers to infer sensitive information in images. At the same time, the image transformations should maintain the original value of the image to the user who is sharing it, either by leaving it not obviously changed, or by enhancing it to increase its visual appeal. This year, the focus is on a set of 60 scene categories, selected from the Places365-Standard data set, that can be considered privacy sensitive.
Presented by Martha Larson
Adversarial Multi Scale Features Learning for Person Re Identificationijtsrd
Person re identification Re ID is the task of matching a target person across different cameras, which has drawn extensive attention in computer vision and has become an essential component in the video surveillance system. Pried can be considered as a problem of image retrieval. Existing person re identification methods depend mostly on single scale appearance information. In this work, to address issues, we demonstrate the benefits of a deep model with Multi scale Feature Representation Learning MFRL using Convolutional Neural Networks CNN and Random Batch Feature Mask RBFM is proposed for pre id in this study. The RBFM is enlightened by the drop block and Batch Drop Block BDB dropout based approaches. However, great challenges are being faced in the pre id task. First, in different scenarios, appearance of the same pedestrian changes dramatically by reason of the body misalignment frequently, various background clutters, large variations of camera views and occlusion. Second, in a public space, different pedestrians wear the same or similar clothes. Therefore, the distinctions between different pedestrian images are subtle. These make the topic of pre id a huge challenge. The proposed methods are only performed in the training phase and discarded in the testing phase, thus, enhancing the effectiveness of the model. Our model achieves the state of the art on the popular benchmark datasets including Market 1501, duke mtmc re id and CUHK03. Besides, we conduct a set of ablation experiments to verify the effectiveness of the proposed methods. Mrs. D. Radhika | D. Harini | N. Kirujha | Dr. M. Duraipandiyan | M. Kavya "Adversarial Multi-Scale Features Learning for Person Re-Identification" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42562.pdf Paper URL: https://www.ijtsrd.comengineering/computer-engineering/42562/adversarial-multiscale-features-learning-for-person-reidentification/mrs-d-radhika
Global-local attention with triplet loss and label smoothed cross-entropy for...IAESIJAI
Person re-identification (Person Re-ID) is a research direction on tracking and identifying people in surveillance camera systems with non-overlapping camera perspectives. Despite much research on this topic, there are still some practical problems that Person Re-ID has not yet solved, in reality, human objects can easily be obscured by obstructions such as other people, trees, luggage, umbrellas, signs, cars, motorbikes. In this paper, we propose a multi-branch deep learning network architecture. In which one branch is for the representation of global features and two branches are for the representation of local features. Dividing the input image into small parts and changing the number of parts between the two branches helps the model to represent the features better. In addition, we add an attention module to the ResNet50 backbone that enhances important human characteristics and eliminates irrelevant information. To improve robustness, the model is trained by combining triplet loss and label smoothing cross-entropy loss (LSCE). Experiments are carried out on datasets Market1501, and duke multi-target multi-camera (DukeMTMC) datasets, our method achieved 96.04% rank-1, 88,11% mean average precision (mAP) on the Market1501 dataset, and 88.78% rank-1, 78,6% mAP on the DukeMTMC dataset. This method achieves performance better than some state-of-the-art methods.
Beyond Frontal Faces: Improving Person Recognition Using Multiple CuesJoão Gabriel Lima
We explore the task of recognizing peoples’ identities
in photo albums in an unconstrained setting. To facilitate
this, we introduce the new
People In Photo Albums (PIPA)
dataset, consisting of over 60000 instances of
2000 in-
dividuals collected from public Flickr photo albums. With
only about half of the person images containing a frontal
face, the recognition task is very challenging due to the
large variations in pose, clothing, camera viewpoint, image
resolution and illumination. We propose the Pose Invariant
PErson Recognition (PIPER) method, which accumulates
the cues of poselet-level person recognizers trained by deep
convolutional networks to discount for the pose variations,
combined with a face recognizer and a global recognizer.
Experiments on three different settings confirm that in our
unconstrained setup PIPER significantly improves on the
performance of DeepFace, which is one of the best face rec-
ognizers as measured on the LFW dataset.
Hardoon Image Ranking With Implicit Feedback From Eye MovementsKalle
In order to help users navigate an image search system, one could
provide explicit information on a small set of images as to which
of them are relevant or not to their task. These rankings are learned
in order to present a user with a new set of images that are relevant
to their task. Requiring such explicit information may not
be feasible in a number of cases, we consider the setting where
the user provides implicit feedback, eye movements, to assist when
performing such a task. This paper explores the idea of implicitly
incorporating eye movement features in an image ranking task
where only images are available during testing. Previous work had
demonstrated that combining eye movement and image features improved
on the retrieval accuracy when compared to using each of
the sources independently. Despite these encouraging results the
proposed approach is unrealistic as no eye movements will be presented
a-priori for new images (i.e. only after the ranked images are
presented would one be able to measure a user’s eye movements
on them). We propose a novel search methodology which combines
image features together with implicit feedback from users’
eye movements in a tensor ranking Support Vector Machine and
show that it is possible to extract the individual source-specific
weight vectors. Furthermore, we demonstrate that the decomposed
image weight vector is able to construct a new image-based semantic
space that outperforms the retrieval accuracy than when solely
using the image-features.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
3. Person re-identification (re-id) aims at matching people
across non-overlapping camera views distributed at distinct
locations.
Camera A Camera B
4. Presentation Outline
• Supervised Person Re-Identification
• Unsupervised Person Re-Identification
• Active Learning for Person Re-Identification
5. Supervised Person Re-Identification
• Training and testing data are from same domain
Lavi, B., Serj, M.F. and Ullah, I., 2018. Survey on deep learning techniques for person re-identification task
Contrastive loss Triplet lossClassification loss
10. Presentation Outline
• Supervised Person Re-Identification
• Unsupervised Person Re-Identification
• Active Learning for Person Re-Identification
11. Unsupervised Person Re-Identification
Training and testing data are from different domain
-> learned on the source domain and transfer the knowledge to
target domain (unsupervised domain adaptation)
Train: Market1501 Test: DUKE
12. Challenges:
• Source and target domains have unknown camera viewing conditions
• The identity/class between source and target domains are non-overlapping
therefore presents a more challenging open-set recognition problem
-> Transferring knowledge of the source domain to target domain in
attribute space
Wang J, Zhu X, Gong S, Li W. Transferable joint attribute-identity deep learning
for unsupervised person re-identification. CVPR,2018
Unsupervised Person Re-Identification
13. Wang J, Zhu X, Gong S, Li W. Transferable joint attribute-identity deep learning
for unsupervised person re-identification. CVPR,2018
Unsupervised Person Re-Identification
14. Unsupervised Target Domain Adaptation
Wang J, Zhu X, Gong S, Li W. Transferable joint attribute-identity deep learning
for unsupervised person re-identification. CVPR,2018
Unsupervised Person Re-Identification
15. Unsupervised Person Re-Identification
Image-to-image translation method: SPGAN
Deng et al., Image-image domain adaptation with preserved self-similarity
and domain-dissimilarity for person re-identification. CVPR 2018
preserved self-similarity and domain dissimilarity
18. Presentation Outline
• Supervised Person Re-Identification
• Unsupervised Person Re-Identification
• Active Learning for Person Re-Identification
19. Make AI work in the real world: Human-In-The-Loop
Human-in-the-Loop (HITL) explores human feedback
in an incremental learning cycle of the machine for
rapid model domain adaptation
20. Active learning is a special case of machine learning in which a learning algorithm
is able to interactively query the user (or some other information source) to obtain
the desired outputs at new data points.
21. There are three scenarios for Active learning :
1. Membership Query Synthesis: the learner generates/constructs an
instance (from some underlying natural distribution).
2. Stream-Based selective sampling, i.e, each sample is considered separately
in our case for label-querying or rejection. Similarly to online-learning, the
data is not saved, there are no assumptions on data distribution, and
therefore it is adaptive to change.
22. 3. Pool-Based sampling, i.e., sampled are chosen from a pool of
unlabeled data for the purpose of labeling
24. Concept
A user annotates few informative pedestrian pairs recommended
by an adaptive agent in a human-in-the-loop learning
mechanism
Re-ID
Model
Sample Selection
(agent)
annotator
Pairwise Data
human-in-the-loop
Agent
action!"
query
ancho
r
query for label
unlabeled gallery
pool
state
reward
Goal: Sample Informative Pair
Action: Select One Sample at Each Step
State: Reflect Sample Correlation
Reward: Uncertainty
Liu, Z.*, Wang, J *., Gong, S., Lu, H. and Tao, D. Deep Reinforcement
Active Learning for Human-in-the-Loop Person Re-Identification. ICCV,
2019,Oral
25. STATE
annotator
Re-ID Loss(Triplet)
REWARD
ACTION
Sample Selection Strategy
gallery pool
query
q
…
g
1
g
2
g
N
0 0.83 0.71 0.66 0.47 0.36
0.83 0 0.85 0 0.87 0
0.71 0.85 0 0 0 0
0.66 0 0 0 0 0
0.47 0.87 0 0 0 0.77
0.36 0 0 0 0.77 0
gKq
Methodology
Joint Reinforcement Active Learning in A Deep Network
false
CNN
!
Agent
"
Liu, Z.*, Wang, J *., Gong, S., Lu, H. and Tao, D. Deep Reinforcement
Active Learning for Human-in-the-Loop Person Re-Identification. ICCV,
2019,Oral
26. Action: Select One Sample at Each Step
State: We construct a sparse similarity graph among query and gallery samples and take
it as the state value (Reflect Sample Correlation)
1. Base CNN Network
2. A Deep Reinforced Active Learner - An Agent
Methodology
27. An example of state updating with different human feedback
Reward: we perform a similar hard triplet loss to measure the uncertainty of data.
30. Presentation Outline
Link Person Re-Identification with ….
• Attribute Learning
• Detection (Person Search )
• Tracking (Multi-target multi-camera tracking)
31. Attribute recognition usually denotes local structures of a person
Person Re-Identification and Attribute Learning
Ø How do human brain match person?
Long hair
bag
31
Attribute recognition usually denotes local structures of a person
Person Re-Iden3fica3on and A6ribute Learning
Ø How do human brain match person?
Long hair
bag
32. Attribute Recognition in in Surveillance
ØChallenges:
• Poor image quality
• Complex background clutter
• Uncontrolled viewing conditions
• Small number of labelled training
33. ØMain idea:
•Discover the interdependency and correlation among
attributes
•Explore visual context as an extra information source to
assist attribute recognition
ØContributions:
•A novel end-to-end encoder-decoder architecture capable
of jointly learning image level context and attribute level
sequential correlation
•Exploit more latent and richer higher-order dependency
among attributes
Wang, J, et al. "Attribute recognition by joint recurrent learning of
context and correlation." ICCV. 2017
Attribute Recognition in in Surveillance
34. Person Re-Identification and Attribute Recognition
Lin, Yutian, et al. "Improving person re-identification by attribute and
identity learning." Pattern Recognition (2019).
40. Detection
Xiao, Tong, et al. "Joint detection and identification feature learning for
person search." CVPR. 2017
Person Re-Identification and Detection
41. Liu, Hao, et al. "Neural person search machines." ICCV. 2017.
Person Re-Identification and Detection
42. Presentation Outline
Link Person Re-Identification with ….
• Attribute Learning
• Detection (Person Search )
• Tracking (Multi-target multi-camera tracking)
43. Multi-target multi-camera tracking
1st MTMCT and ReID workshop CVPR 2017
2nd MTMCT and ReID workshop CVPR 2019
Duke MTMC (Multi-Target, Multi-Camera) dataset
45. Conclusion
• Supervised Person Re-Identification
• Unsupervised Person Re-Identification
• Active Learning for Person Re-Identification
Link Person Re-Identification with ….
Ø Attribute Learning
Ø Detection (Person Search )
Ø Tracking (Multi-target multi-camera tracking)