SlideShare a Scribd company logo
Wen-Chih Lo¹, Ching-Ling Fan¹, Jean Lee¹, Chun-Ying Huang²,
Kuan-Ta Chen³, and Cheng-Hsin Hsu¹
¹National Tsing Hua University, HsinChu, Taiwan
²National Chiao Tung University, HsinChu, Taiwan
³Institute of Information Science Academia Sinica, Taipei,
Taiwan
1
ACM MMSys’17, Dataset Track, Taipei, Taiwan, June 22, 2017
 Introduction
 Basic statistics
 Dataset structure
 Content trace collection
 Sensor trace collection
 How to choose
 Sample applications
 Teaser
2
3
Wow,
it’s good!
 A 360° video is a view that every direction is recorded at the
same time
 With planar monitors is passive experiences
 Head-Mounted Displays (HMDs) offer more immersive
experiences
4
 VR/AR deliver a total $3.9 billion, including $2.7 billion
VR and $1.2 billion AR, revenue in 2016 [1]
5[1] After mixed year, mobile AR to drive $108 billion VR/AR market by 2021, Digi-capital, Jan 2017.
https://goo.gl/Blcv2f
 Latency
 Extremely high resolution
 The distortion while stitching or projecting videos/images
 Compress tremendous amount of video data in real-time
 Reduce the computational cost using tile-based viewing
method
6
Distortio
n
7
 We recruit 50 subjects, each of them is asked to watch ten
360°videos
 52% are male
 Most of them are in early twenties
 56% of them for the first time
8
56%
42%
2%
How often do you watch
360°video using HMDs?
Never Seldom Often
9
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Video traces
Image saliency map
Motion map
10
 We collect ten 360° videos from YouTube
 1 minute long, 4K resolution, and 30fps
11
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Content
 Convolutional Neural Networks (CNN)
 Based on a pre-trained VGG-16 networks
 Gray-scale image (from 0 to 255)
12
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
[1] M. Cornia, L. Baraldi, G. Serra, and R. Cucchiara. 2016. A Deep Multi-Level Network for
Saliency Prediction. In International Conference on Pattern Recognition (ICPR’16).
Content
 Relative motions
 Lucas-Kanade optical flow
 Black-and-white images (0 or 1)
13
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
[2] B. Lucas and T. Kanade. 1981. An iterative image registration technique with an application to
stereo vision. In Proc. of the International Joint Conference on Artificial Intelligence (IJCAI’7)
Content
 Collect sensor data from HMDs while viewers are watching
360° videos
 Frame Capturer: GamingAnywhere[1]
 Sensor Logger: OpenTrack[2]
14
[1] http://gaminganywhere.org/
[2] https://github.com/opentrack/opentrack
360° video
Sensor data
with timestamp250Hz
Video frame with timestamp
30Hz
Oculus
DK2
 Raw data
 Orientation data
 Tile data
15
 Raw sensor data from HMDs
 Timestamp with epoch time
 Position (x, y, and z)
 Orientation (yaw, pitch, and roll)
16
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Raw sensor data from HMDs
 Timestamp with epoch time
 Position (x, y, and z)
 Orientation (yaw, pitch, and roll)
17
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Raw sensor data from HMDs
 Timestamp with epoch time
 Position (x, y, and z)
 Orientation (yaw, pitch, and roll)
18
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
x
y
z
Sensor
 Raw sensor data from HMDs
 Timestamp with epoch time
 Position (x, y, and z)
 Orientation (yaw, pitch, and roll)
19
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
x
y
z
roll
pitchyaw
Sensor
 Align the sensor data and video frames
 Different viewers introduce different bias
 A 35-sec calibration procedure
20
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Align the sensor data and video frames
 Different viewers introduce different bias
 Design a calibration procedure
21
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Align the sensor data and video frames
 Different viewers introduce different bias
 Design a calibration procedure
22
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Align the sensor data and video frames
 Different viewers introduce different bias
 Design a calibration procedure
23
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Align the sensor data and video frames
 Different viewers introduce different bias
 Design a calibration procedure
24
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Field-of-View (FoV) are 100° x 100° circle
 We divide each frame into 192x192 tiles
 We number the tiles from upper-left to lower-right (from 0
to 199)
25
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
100°
100°
Sensor
 Field-of-View (FoV) is 100° x 100° circle
 We divide each frame into 192x192 tiles
 We number the tiles from upper-left to lower-right (from 0
to 199)
26
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
 Field-of-View (FoV) is 100° x 100° circle
 We divide each frame into 192x192 tiles
 We number the tiles from upper-left to lower-right (from 0
to 199)
27
Dataset
Content
Trace
Video Trace
Saliency
Map
Motion Map
Sensor
Trace
Raw
Orientation
Tile
Sensor
0
Head
moveme
nt
Eye
moveme
nt
Content
-driven
data
Open-
source
software
Applicati
on-driven
Lo et al. [1]
Rai et al. [2]
Corbillon et al.
[3]
Wu et al. [4]
28
[1] W. Lo, C. Fan, J. Lee, C. Huang, K. Chen, and C. Hsu. “360° Video Viewing Dataset in Head-Mounted Virtual
Reality.” In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017.
[2] Y. Rai, J. Gutiérrez, and P. Callet. “A Dataset of Head and Eye Movements for 360 Degree Images.” In Proc. of the
8th ACM on Multimedia Systems Conference (MMSys'17). 2017.
[3] X. Corbillon, F. Simone, and G. Simon. “360-Degree Video Head Movement Dataset.” In Proc. of the 8th ACM on
Multimedia Systems Conference (MMSys'17). 2017.
[4] C. Wu, Z. Tan, Z. Wang, and S. Yang. “A Dataset for Exploring User Behaviors in VR Spherical Video Streaming.”
In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017.
 Viewed tile predictions
 Bitrate allocation
α
β
θ
FoV
0°
1 1 1 1
1 1 1 1
1 1 1
0 0 0 0 0 1
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0
0 0
011 1 1 1
1 1 1 1
1 1 1
1 1
 NOSSDAV’17
 Tomorrow (6/23) 2:10pm - 3:10pm at 2nd Conference Room
 C. Fan, J. Lee, W. Lo, C. Huang, K. Chen, and C. Hsu,
“Fixation Prediction for 360˚ Video Streaming in Head-
Mounted Virtual Reality”
30
Wen-Chih Lo
wchih.lo@gmail.com
Dataset link address:
https://nmsl.cs.nthu.edu.tw/dropbox/3
60dataset.zip
31

More Related Content

Similar to 360° Video Viewing Dataset in Head-Mounted Virtual Reality

Understanding Users Behaviours in User-Centric Immersive Communications
Understanding Users Behaviours in User-Centric Immersive CommunicationsUnderstanding Users Behaviours in User-Centric Immersive Communications
Understanding Users Behaviours in User-Centric Immersive Communications
Förderverein Technische Fakultät
 
538 207-219
538 207-219538 207-219
538 207-219
idescitation
 
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
Universitat Politècnica de Catalunya
 
3d television
3d television3d television
3d television
Ishtdeep Hora
 
2013 09-01 enviroinfo presentation - final
2013 09-01 enviroinfo presentation - final2013 09-01 enviroinfo presentation - final
2013 09-01 enviroinfo presentation - final
Denis Havlik
 
TB-Survey-2020.pdf
TB-Survey-2020.pdfTB-Survey-2020.pdf
TB-Survey-2020.pdf
ssuser50a5ec
 
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Universitat Politècnica de Catalunya
 
최근의 공간정보 분야 동향과 시사점
최근의 공간정보 분야 동향과 시사점최근의 공간정보 분야 동향과 시사점
최근의 공간정보 분야 동향과 시사점
SANGHEE SHIN
 
Secure IoT Systems Monitor Framework using Probabilistic Image Encryption
Secure IoT Systems Monitor Framework using Probabilistic Image EncryptionSecure IoT Systems Monitor Framework using Probabilistic Image Encryption
Secure IoT Systems Monitor Framework using Probabilistic Image Encryption
IJAEMSJORNAL
 
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
Universitat Politècnica de Catalunya
 
Video Liveness Verification
Video Liveness VerificationVideo Liveness Verification
Video Liveness Verification
ijtsrd
 
DW2020 Data Models - FIWARE Platform
DW2020 Data Models - FIWARE PlatformDW2020 Data Models - FIWARE Platform
DW2020 Data Models - FIWARE Platform
Fernando Lopez Aguilar
 
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
Gianni Cristian Iannelli
 
GITA PNW 2015 Peter Batty
GITA PNW 2015 Peter BattyGITA PNW 2015 Peter Batty
GITA PNW 2015 Peter Batty
Peter Batty
 
Call for papers - 9th International Conference on Signal, Image Processing an...
Call for papers - 9th International Conference on Signal, Image Processing an...Call for papers - 9th International Conference on Signal, Image Processing an...
Call for papers - 9th International Conference on Signal, Image Processing an...
sipij
 
Technological advances in dental implant surgery
Technological advances in dental implant surgeryTechnological advances in dental implant surgery
Technological advances in dental implant surgery
Periowiki.com
 
Navigation-aware adaptive streaming strategies for omnidirectional video
Navigation-aware adaptive streaming strategies for omnidirectional videoNavigation-aware adaptive streaming strategies for omnidirectional video
Navigation-aware adaptive streaming strategies for omnidirectional video
Silvia Rossi
 
Hierarchical structure adaptive
Hierarchical structure adaptiveHierarchical structure adaptive
Hierarchical structure adaptive
NEERAJ BAGHEL
 
Augmented Reality And Virtual Reality
Augmented Reality And Virtual RealityAugmented Reality And Virtual Reality
Augmented Reality And Virtual Reality
Brandi Gonzales
 
COSC 426 Lect. 1 - Introduction to AR
COSC 426 Lect. 1 - Introduction to ARCOSC 426 Lect. 1 - Introduction to AR
COSC 426 Lect. 1 - Introduction to AR
Mark Billinghurst
 

Similar to 360° Video Viewing Dataset in Head-Mounted Virtual Reality (20)

Understanding Users Behaviours in User-Centric Immersive Communications
Understanding Users Behaviours in User-Centric Immersive CommunicationsUnderstanding Users Behaviours in User-Centric Immersive Communications
Understanding Users Behaviours in User-Centric Immersive Communications
 
538 207-219
538 207-219538 207-219
538 207-219
 
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
Deep Convnets for Video Processing (Master in Computer Vision Barcelona, 2016)
 
3d television
3d television3d television
3d television
 
2013 09-01 enviroinfo presentation - final
2013 09-01 enviroinfo presentation - final2013 09-01 enviroinfo presentation - final
2013 09-01 enviroinfo presentation - final
 
TB-Survey-2020.pdf
TB-Survey-2020.pdfTB-Survey-2020.pdf
TB-Survey-2020.pdf
 
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
 
최근의 공간정보 분야 동향과 시사점
최근의 공간정보 분야 동향과 시사점최근의 공간정보 분야 동향과 시사점
최근의 공간정보 분야 동향과 시사점
 
Secure IoT Systems Monitor Framework using Probabilistic Image Encryption
Secure IoT Systems Monitor Framework using Probabilistic Image EncryptionSecure IoT Systems Monitor Framework using Probabilistic Image Encryption
Secure IoT Systems Monitor Framework using Probabilistic Image Encryption
 
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
Deep Learning for Computer Vision (3/4): Video Analytics @ laSalle 2016
 
Video Liveness Verification
Video Liveness VerificationVideo Liveness Verification
Video Liveness Verification
 
DW2020 Data Models - FIWARE Platform
DW2020 Data Models - FIWARE PlatformDW2020 Data Models - FIWARE Platform
DW2020 Data Models - FIWARE Platform
 
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
DATA-CAPTURE-GEM-Userguide-Footprint-Homogenous-Zones-201401-V01
 
GITA PNW 2015 Peter Batty
GITA PNW 2015 Peter BattyGITA PNW 2015 Peter Batty
GITA PNW 2015 Peter Batty
 
Call for papers - 9th International Conference on Signal, Image Processing an...
Call for papers - 9th International Conference on Signal, Image Processing an...Call for papers - 9th International Conference on Signal, Image Processing an...
Call for papers - 9th International Conference on Signal, Image Processing an...
 
Technological advances in dental implant surgery
Technological advances in dental implant surgeryTechnological advances in dental implant surgery
Technological advances in dental implant surgery
 
Navigation-aware adaptive streaming strategies for omnidirectional video
Navigation-aware adaptive streaming strategies for omnidirectional videoNavigation-aware adaptive streaming strategies for omnidirectional video
Navigation-aware adaptive streaming strategies for omnidirectional video
 
Hierarchical structure adaptive
Hierarchical structure adaptiveHierarchical structure adaptive
Hierarchical structure adaptive
 
Augmented Reality And Virtual Reality
Augmented Reality And Virtual RealityAugmented Reality And Virtual Reality
Augmented Reality And Virtual Reality
 
COSC 426 Lect. 1 - Introduction to AR
COSC 426 Lect. 1 - Introduction to ARCOSC 426 Lect. 1 - Introduction to AR
COSC 426 Lect. 1 - Introduction to AR
 

Recently uploaded

怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
kekzed
 
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
OECD Directorate for Financial and Enterprise Affairs
 
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
OECD Directorate for Financial and Enterprise Affairs
 
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussionArtificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
OECD Directorate for Financial and Enterprise Affairs
 
Using-Presentation-Software-to-the-Fullf.pptx
Using-Presentation-Software-to-the-Fullf.pptxUsing-Presentation-Software-to-the-Fullf.pptx
Using-Presentation-Software-to-the-Fullf.pptx
kainatfatyma9
 
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
SkillCertProExams
 
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
gpww3sf4
 
Disaster Management project for holidays homework and other uses
Disaster Management project for holidays homework and other usesDisaster Management project for holidays homework and other uses
Disaster Management project for holidays homework and other uses
RIDHIMAGARG21
 
Competition and Regulation in Professions and Occupations – ROBSON – June 202...
Competition and Regulation in Professions and Occupations – ROBSON – June 202...Competition and Regulation in Professions and Occupations – ROBSON – June 202...
Competition and Regulation in Professions and Occupations – ROBSON – June 202...
OECD Directorate for Financial and Enterprise Affairs
 
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussion
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussionPro-competitive Industrial Policy – LANE – June 2024 OECD discussion
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussion
OECD Directorate for Financial and Enterprise Affairs
 
XP 2024 presentation: A New Look to Leadership
XP 2024 presentation: A New Look to LeadershipXP 2024 presentation: A New Look to Leadership
XP 2024 presentation: A New Look to Leadership
samililja
 
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
OECD Directorate for Financial and Enterprise Affairs
 
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
OECD Directorate for Financial and Enterprise Affairs
 
Proposal: The Ark Project and The BEEP Inc
Proposal: The Ark Project and The BEEP IncProposal: The Ark Project and The BEEP Inc
Proposal: The Ark Project and The BEEP Inc
Raheem Muhammad
 
IEEE CIS Webinar Sustainable futures.pdf
IEEE CIS Webinar Sustainable futures.pdfIEEE CIS Webinar Sustainable futures.pdf
IEEE CIS Webinar Sustainable futures.pdf
Claudio Gallicchio
 
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
OECD Directorate for Financial and Enterprise Affairs
 
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
OECD Directorate for Financial and Enterprise Affairs
 
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdfBRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
Robin Haunschild
 
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussion
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussionPro-competitive Industrial Policy – OECD – June 2024 OECD discussion
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussion
OECD Directorate for Financial and Enterprise Affairs
 
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdfWhy Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
Ben Linders
 

Recently uploaded (20)

怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
怎么办理(lincoln学位证书)英国林肯大学毕业证文凭学位证书原版一模一样
 
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
Competition and Regulation in Professions and Occupations – OECD – June 2024 ...
 
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
Artificial Intelligence, Data and Competition – SCHREPEL – June 2024 OECD dis...
 
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussionArtificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
Artificial Intelligence, Data and Competition – LIM – June 2024 OECD discussion
 
Using-Presentation-Software-to-the-Fullf.pptx
Using-Presentation-Software-to-the-Fullf.pptxUsing-Presentation-Software-to-the-Fullf.pptx
Using-Presentation-Software-to-the-Fullf.pptx
 
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
ServiceNow CIS-ITSM Exam Dumps & Questions [2024]
 
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
原版制作贝德福特大学毕业证(bedfordhire毕业证)硕士文凭原版一模一样
 
Disaster Management project for holidays homework and other uses
Disaster Management project for holidays homework and other usesDisaster Management project for holidays homework and other uses
Disaster Management project for holidays homework and other uses
 
Competition and Regulation in Professions and Occupations – ROBSON – June 202...
Competition and Regulation in Professions and Occupations – ROBSON – June 202...Competition and Regulation in Professions and Occupations – ROBSON – June 202...
Competition and Regulation in Professions and Occupations – ROBSON – June 202...
 
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussion
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussionPro-competitive Industrial Policy – LANE – June 2024 OECD discussion
Pro-competitive Industrial Policy – LANE – June 2024 OECD discussion
 
XP 2024 presentation: A New Look to Leadership
XP 2024 presentation: A New Look to LeadershipXP 2024 presentation: A New Look to Leadership
XP 2024 presentation: A New Look to Leadership
 
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
The Intersection between Competition and Data Privacy – CAPEL – June 2024 OEC...
 
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
The Intersection between Competition and Data Privacy – OECD – June 2024 OECD...
 
Proposal: The Ark Project and The BEEP Inc
Proposal: The Ark Project and The BEEP IncProposal: The Ark Project and The BEEP Inc
Proposal: The Ark Project and The BEEP Inc
 
IEEE CIS Webinar Sustainable futures.pdf
IEEE CIS Webinar Sustainable futures.pdfIEEE CIS Webinar Sustainable futures.pdf
IEEE CIS Webinar Sustainable futures.pdf
 
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
Artificial Intelligence, Data and Competition – ČORBA – June 2024 OECD discus...
 
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
The Intersection between Competition and Data Privacy – COLANGELO – June 2024...
 
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdfBRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
BRIC_2024_2024-06-06-11:30-haunschild_archival_version.pdf
 
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussion
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussionPro-competitive Industrial Policy – OECD – June 2024 OECD discussion
Pro-competitive Industrial Policy – OECD – June 2024 OECD discussion
 
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdfWhy Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
Why Psychological Safety Matters for Software Teams - ACE 2024 - Ben Linders.pdf
 

360° Video Viewing Dataset in Head-Mounted Virtual Reality

  • 1. Wen-Chih Lo¹, Ching-Ling Fan¹, Jean Lee¹, Chun-Ying Huang², Kuan-Ta Chen³, and Cheng-Hsin Hsu¹ ¹National Tsing Hua University, HsinChu, Taiwan ²National Chiao Tung University, HsinChu, Taiwan ³Institute of Information Science Academia Sinica, Taipei, Taiwan 1 ACM MMSys’17, Dataset Track, Taipei, Taiwan, June 22, 2017
  • 2.  Introduction  Basic statistics  Dataset structure  Content trace collection  Sensor trace collection  How to choose  Sample applications  Teaser 2
  • 4.  A 360° video is a view that every direction is recorded at the same time  With planar monitors is passive experiences  Head-Mounted Displays (HMDs) offer more immersive experiences 4
  • 5.  VR/AR deliver a total $3.9 billion, including $2.7 billion VR and $1.2 billion AR, revenue in 2016 [1] 5[1] After mixed year, mobile AR to drive $108 billion VR/AR market by 2021, Digi-capital, Jan 2017. https://goo.gl/Blcv2f
  • 6.  Latency  Extremely high resolution  The distortion while stitching or projecting videos/images  Compress tremendous amount of video data in real-time  Reduce the computational cost using tile-based viewing method 6 Distortio n
  • 7. 7
  • 8.  We recruit 50 subjects, each of them is asked to watch ten 360°videos  52% are male  Most of them are in early twenties  56% of them for the first time 8 56% 42% 2% How often do you watch 360°video using HMDs? Never Seldom Often
  • 10. Video traces Image saliency map Motion map 10
  • 11.  We collect ten 360° videos from YouTube  1 minute long, 4K resolution, and 30fps 11 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Content
  • 12.  Convolutional Neural Networks (CNN)  Based on a pre-trained VGG-16 networks  Gray-scale image (from 0 to 255) 12 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile [1] M. Cornia, L. Baraldi, G. Serra, and R. Cucchiara. 2016. A Deep Multi-Level Network for Saliency Prediction. In International Conference on Pattern Recognition (ICPR’16). Content
  • 13.  Relative motions  Lucas-Kanade optical flow  Black-and-white images (0 or 1) 13 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile [2] B. Lucas and T. Kanade. 1981. An iterative image registration technique with an application to stereo vision. In Proc. of the International Joint Conference on Artificial Intelligence (IJCAI’7) Content
  • 14.  Collect sensor data from HMDs while viewers are watching 360° videos  Frame Capturer: GamingAnywhere[1]  Sensor Logger: OpenTrack[2] 14 [1] http://gaminganywhere.org/ [2] https://github.com/opentrack/opentrack 360° video Sensor data with timestamp250Hz Video frame with timestamp 30Hz Oculus DK2
  • 15.  Raw data  Orientation data  Tile data 15
  • 16.  Raw sensor data from HMDs  Timestamp with epoch time  Position (x, y, and z)  Orientation (yaw, pitch, and roll) 16 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 17.  Raw sensor data from HMDs  Timestamp with epoch time  Position (x, y, and z)  Orientation (yaw, pitch, and roll) 17 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 18.  Raw sensor data from HMDs  Timestamp with epoch time  Position (x, y, and z)  Orientation (yaw, pitch, and roll) 18 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile x y z Sensor
  • 19.  Raw sensor data from HMDs  Timestamp with epoch time  Position (x, y, and z)  Orientation (yaw, pitch, and roll) 19 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile x y z roll pitchyaw Sensor
  • 20.  Align the sensor data and video frames  Different viewers introduce different bias  A 35-sec calibration procedure 20 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 21.  Align the sensor data and video frames  Different viewers introduce different bias  Design a calibration procedure 21 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 22.  Align the sensor data and video frames  Different viewers introduce different bias  Design a calibration procedure 22 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 23.  Align the sensor data and video frames  Different viewers introduce different bias  Design a calibration procedure 23 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 24.  Align the sensor data and video frames  Different viewers introduce different bias  Design a calibration procedure 24 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 25.  Field-of-View (FoV) are 100° x 100° circle  We divide each frame into 192x192 tiles  We number the tiles from upper-left to lower-right (from 0 to 199) 25 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile 100° 100° Sensor
  • 26.  Field-of-View (FoV) is 100° x 100° circle  We divide each frame into 192x192 tiles  We number the tiles from upper-left to lower-right (from 0 to 199) 26 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor
  • 27.  Field-of-View (FoV) is 100° x 100° circle  We divide each frame into 192x192 tiles  We number the tiles from upper-left to lower-right (from 0 to 199) 27 Dataset Content Trace Video Trace Saliency Map Motion Map Sensor Trace Raw Orientation Tile Sensor 0
  • 28. Head moveme nt Eye moveme nt Content -driven data Open- source software Applicati on-driven Lo et al. [1] Rai et al. [2] Corbillon et al. [3] Wu et al. [4] 28 [1] W. Lo, C. Fan, J. Lee, C. Huang, K. Chen, and C. Hsu. “360° Video Viewing Dataset in Head-Mounted Virtual Reality.” In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017. [2] Y. Rai, J. Gutiérrez, and P. Callet. “A Dataset of Head and Eye Movements for 360 Degree Images.” In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017. [3] X. Corbillon, F. Simone, and G. Simon. “360-Degree Video Head Movement Dataset.” In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017. [4] C. Wu, Z. Tan, Z. Wang, and S. Yang. “A Dataset for Exploring User Behaviors in VR Spherical Video Streaming.” In Proc. of the 8th ACM on Multimedia Systems Conference (MMSys'17). 2017.
  • 29.  Viewed tile predictions  Bitrate allocation α β θ FoV 0° 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 011 1 1 1 1 1 1 1 1 1 1 1 1
  • 30.  NOSSDAV’17  Tomorrow (6/23) 2:10pm - 3:10pm at 2nd Conference Room  C. Fan, J. Lee, W. Lo, C. Huang, K. Chen, and C. Hsu, “Fixation Prediction for 360˚ Video Streaming in Head- Mounted Virtual Reality” 30
  • 31. Wen-Chih Lo wchih.lo@gmail.com Dataset link address: https://nmsl.cs.nthu.edu.tw/dropbox/3 60dataset.zip 31

Editor's Notes

  1. Hello everyone, I am Wen-Chih Lo from NTHU. Today I am going to present a 360-degree video dataset. This work is joint with my collaborators from NTHU and Prof. Huang from NCTU and Prof. Chen from Academia Sinica.
  2. In this talk, I will briefly describe our dataset structure and the two major components in our 360-degree video viewing dataset. Besides, I will also show the basic statistics of our dataset and a sample application using our dataset.
  3. A 360-degree video is known as spherical video or immersive video. Watching these videos using traditional planar monitors gives viewers passive experience. Nowadays, HMDs are widely available. Using HMDs give viewers more immersive experience than previous one.
  4. In digi-capital report, they says AR/VR will keep growing up. In the future, watching 360-degree video using HMDs might be a common experience.
  5. However, it is challenging to deliver such a high resolution video over today’s network. For example... bla bla bla. More and more researchers, engineers are jumping into this topic. At that time, there are no public standard dataset can be used to evaluate their system performance and develop the new algorithms. But I guess that we don't need to worry about this problem. There are several dataset have been published here.
  6. At the beginning, I want to show you some basic statistics of our dataset. We recruit 50 subjects in our dataset. (description) Each of them is asked to watch ten videos. During the experiments, all of them are asked to stand when watching videos.
  7. Our dataset is unique, because it contains not only the sensor-driven data, but also the content-driven data.
  8. First, I want to show you the content-driven data, we called it content trace. The content trace means the impact of video content on viewer’s attentions. (click) For example, the video traces, the saliency map and motion map of 360-degree video. I will explain each of them later.
  9. Here, I want to show you the video traces we collected from YouTube. We divide the 10 videos into three groups, including fast-paced Natural Images (NI), slow-paced NI, and fast-paced Computer-Generated (CG) All of them are 1 minute 4k resolution videos.
  10. The other one is saliency map. We develop a deep neural network based on pre-trained VGG networks. We use this network to produce the saliency map. (demo video) A saliency map indicates the attraction levels of the video. It means which part of video that attracts viewer’s attention. As many previous presenters mentioned the saliency map. Which FoV should we stream to meet the viewer’s needs in the next moment.
  11. The last one is motion map. We analyze the optical flow of the video frames to produce the motion map using OpenCV library. (demo video) Motion map indicate the relative motions between the objects in video and the viewers.
  12. Second, I want to show you the sensor-drive data. It is the viewing orientation data from HMDs when a viewer watching videos. (click) In our testbed, we render the videos to Oculus DK2 and Oculus Video using Oculus SDK. (click) We use GamingAnywhere to be our Frame capturer. It records the video from HMDs. The rate of frame capturer is 30 Hz. (click) We modified and enhanced the opentrack to be our Sensor logger. Opentrack allows you to track user's head movements. It captures and timestamps the orientation data from HMDs. The rate of sensor logger is 250 Hz
  13. (click) The sensor trace contains 3 different data, the raw data from HMDs, the processed viewing orientation data, and the processed tile data.
  14. Here, I want to show you the raw sensor data from HMDs. There are 7 fields in the raw sensor data, including timestamp (epoch time), position (x, y, and z), and orientation (yaw, pitch ,and roll).
  15. Here, I want to show you the raw sensor data from HMDs. There are 7 fields in the raw sensor data, including timestamp (epoch time), position (x, y, and z), and orientation (yaw, pitch ,and roll).
  16. Here, I want to show you the raw sensor data from HMDs. There are 7 fields in the raw sensor data, including timestamp (epoch time), position (x, y, and z), and orientation (yaw, pitch ,and roll).
  17. Here, I want to show you the raw sensor data from HMDs. There are 7 fields in the raw sensor data, including timestamp (epoch time), position (x, y, and z), and orientation (yaw, pitch ,and roll).
  18. To simplify the usage of our dataset, we align the timestamps in the raw sensor data from HMD and video frames which captured from GA. However, in our pilot experiments, we find that different viewers tend to introduce different amount of bias. We then insert a 35-sec calibration video before each viewer starts watching 360° videos. This is not only the calibration procedure, but also help viewers to familiar how to watch the 360-degree video.
  19. There are 10 fields in the orientation data, Including index, position (x, y, and z), raw orientation (yaw, pitch ,and roll), and calibrated orientation data.
  20. There are 10 fields in the orientation data, Including index, position (x, y, and z), raw orientation (yaw, pitch ,and roll), and calibrated orientation data.
  21. There are 10 fields in the orientation data, Including index, position (x, y, and z), raw orientation (yaw, pitch ,and roll), and calibrated orientation data.
  22. There are 10 fields in the orientation data, Including index, position (x, y, and z), raw orientation (yaw, pitch ,and roll), and calibrated orientation data.
  23. We measure that the FoV of HMD is about 100°x100° circles. Therefore, we process the view orientation data, and generate viewed tile data to further simplify the usage of our dataset. There are 2 fields in the tile data, including index and the tiles which are watched by viewer.
  24. We measure that the FoV of HMD is about 100°x100° circles. Therefore, we process the view orientation data, and generate viewed tile data to further simplify the usage of our dataset. There are 2 fields in the tile data, including index and the tiles which are watched by viewer.
  25. We measure that the FoV of HMD is about 100°x100° circles. (click) Therefore, we process the view orientation data, and generate viewed tile data to further simplify the usage of our dataset. There are 2 fields in the tile data, including index and the tiles which are watched by viewer.
  26. Now, there are several dataset are available. Here is a table summarizing different dataset efforts. Each of them is unique and contains different features. For example, if you need to take content-driven data into consideration, you can choose the first dataset. If you need to take eye movement into consideration, you can choose the second one. I need to emphasize that this is preliminary, and if you think differently, please don't hesitate to let us know.
  27. Our dataset can be used in various 360-degree video applications. (click) For example, when a viewer with HMD rotates his/her head to watch some new tiles which have not been requested. (click) It may take several seconds to deliver these new tiles. To reduce the buffer time, (click) predict a viewer who watches a tile-based 360-degree video using HMDs is important. Our dataset can be used for developing and evaluating new algorithms for viewed tile predictions. Besides, our dataset also has video content with diverse characteristics, for example, it also can be used for bitrate allocation for 360° video streaming to HMDs.
  28. Here, I want to take a quick teaser of our prediction work. Tomorrow we will present our fixation prediction network using our dataset. If you are interested in this topic or how to use our dataset, Please feel free to join us tomorrow afternoon at the 2nd conference room. See you tomorrow.
  29. That is. Thank you. I am ready to take questions. I am not quiet sure. I guess….blablabla. Here is what I propose to do for deeper investigations. I will let you know my findings. I hope that we can stay in touch with each other. 1. We tiled the video in the equirectangular domain. 2. We provide original video sequences in our dataset, and they are protected with password to avoid copyright issues. 3. It is our individual judgment. 4. Saliency maps are computed based on equirectangular images using a classical image-based saliency mapping approach. There is no research that indicates with sufficient confidence that this can be done. If we tiled the video into multi-level sphere window, I guess it will reduce the distortion influence. 5. If we use the cubic map projection and process the 6 sides of the cubes separately using the same saliency mapping software. It seems to me that it won't achieve the same result.