This document provides information on conducting subjective video quality assessments. It discusses different subjective assessment methods like double stimulus impairment scale (DSIS) and single stimulus continuous quality evaluation (SSCQE). It describes test parameters like number of observers, viewing conditions, grading scales and how to present the results. Guidelines are provided for tasks like screening observers, conducting test sessions, introducing impairments and collecting opinion scores to evaluate video coding standards and compression artifacts.
Video Compression, Part 4 Section 1, Video Quality Assessment Dr. Mohieddin Moradi
This document provides an overview of video compression artifacts that can occur when video is compressed for streaming or storage. It discusses both spatial artifacts, such as blurring, blocking, ringing, and color bleeding, as well as temporal artifacts like flickering and mosquito noise. For each artifact, it describes the visual appearance and potential causes from factors like quantization during compression, motion compensation between frames, and chroma subsampling. The document aims to help understand how compression can degrade perceptual video quality and different types of artifacts that may be evaluated both objectively and subjectively.
This document provides information about various camera settings and technologies for capturing clear images, including:
1. Clear Scan helps eliminate banding caused when a camera's frame rate does not match a CRT display's refresh rate.
2. Slow Shutter extends the camera's exposure time to produce blur effects or allow more light in low-light scenes.
3. Super Sampling uses a 1080p camera to produce sharper 720p images by maintaining higher frequency response.
4. Detail correction adds a spike-shaped detail signal to make edges appear sharper without degrading resolution. Settings like detail level and H/V ratio control the amount and balance of detail correction.
5. Other topics covered
This document provides information about 4K lens specifications and performance. It discusses key optical parameters for 4K lenses such as sharpness, chromatic aberration, depth of field, and resolution. The document explains how 4K lenses are designed to minimize chromatic aberration and enhance modulation transfer function to improve image quality. It also describes the benefits of 4K lenses for wide color gamut and high dynamic range imaging applications. These benefits include reduced color fringing, flare, and black level for increased dynamic range. Examples are provided comparing image quality between 4K and HD lenses. The document concludes with information about Canon's cinema lens lineup and technologies.
This document provides an overview of high dynamic range (HDR) technology and workflows for HDR video production and mastering. It discusses HDR standards like SMPTE ST 2084 and ARIB STB-B67, camera log curves, luminance levels, and tools for setting up HDR monitoring including waveform monitors. Specific topics covered include HDR graticules, setting luminance levels for highlights and grey points, and using zebra patterns and zoom modes to evaluate highlight levels in HDR images.
The document describes the latest UHK-430 4K camera from Ikegami, highlighting its key technologies and innovations. It features newly developed 2/3-inch native 4K CMOS sensors, a next generation high speed video processor called AXII, and supports high dynamic range video formats. The camera employs a 3-chip design for improved resolution, sensitivity, and reduced aliasing compared to single-chip cameras. It allows for uncompressed 4K signal transmission and future-proof expandability through modular components.
This document outlines elements of high-quality image production, including spatial and temporal resolution, dynamic range, color gamut, bit depth, and coding. It discusses color gamut conversion, gamma correction, HDR and SDR mastering, tone mapping, and backwards compatibility. The document also covers HDR metadata standards and different distribution scenarios for HDR content.
This document provides an overview of high definition television (HDTV) standards and concepts such as color gamut, color bars test signals, colorimetry, chroma adjustment, and luminance adjustment. It discusses differences between standard definition (SDTV) and HDTV color bars, how wider color gamuts in HDTV allow for deeper colors, and how to use various elements of the color bars signal to properly adjust a display's color, brightness, contrast, and chroma. The document contains diagrams demonstrating color gamuts and examples of how objects appear within different gamuts.
This document provides an overview of color spaces and high dynamic range (HDR) technologies. It begins with definitions of color gamut and chromaticity coordinates. It then discusses several key color spaces including Rec.709, Rec.2020, DCI-P3, ACES, and S-Gamut3. It also covers HDR formats like PQ, HLG, and log encoding. The document aims to explain the essential aspects of different color spaces and HDR technologies used for digital cinema and television production.
Video Compression, Part 4 Section 1, Video Quality Assessment Dr. Mohieddin Moradi
This document provides an overview of video compression artifacts that can occur when video is compressed for streaming or storage. It discusses both spatial artifacts, such as blurring, blocking, ringing, and color bleeding, as well as temporal artifacts like flickering and mosquito noise. For each artifact, it describes the visual appearance and potential causes from factors like quantization during compression, motion compensation between frames, and chroma subsampling. The document aims to help understand how compression can degrade perceptual video quality and different types of artifacts that may be evaluated both objectively and subjectively.
This document provides information about various camera settings and technologies for capturing clear images, including:
1. Clear Scan helps eliminate banding caused when a camera's frame rate does not match a CRT display's refresh rate.
2. Slow Shutter extends the camera's exposure time to produce blur effects or allow more light in low-light scenes.
3. Super Sampling uses a 1080p camera to produce sharper 720p images by maintaining higher frequency response.
4. Detail correction adds a spike-shaped detail signal to make edges appear sharper without degrading resolution. Settings like detail level and H/V ratio control the amount and balance of detail correction.
5. Other topics covered
This document provides information about 4K lens specifications and performance. It discusses key optical parameters for 4K lenses such as sharpness, chromatic aberration, depth of field, and resolution. The document explains how 4K lenses are designed to minimize chromatic aberration and enhance modulation transfer function to improve image quality. It also describes the benefits of 4K lenses for wide color gamut and high dynamic range imaging applications. These benefits include reduced color fringing, flare, and black level for increased dynamic range. Examples are provided comparing image quality between 4K and HD lenses. The document concludes with information about Canon's cinema lens lineup and technologies.
This document provides an overview of high dynamic range (HDR) technology and workflows for HDR video production and mastering. It discusses HDR standards like SMPTE ST 2084 and ARIB STB-B67, camera log curves, luminance levels, and tools for setting up HDR monitoring including waveform monitors. Specific topics covered include HDR graticules, setting luminance levels for highlights and grey points, and using zebra patterns and zoom modes to evaluate highlight levels in HDR images.
The document describes the latest UHK-430 4K camera from Ikegami, highlighting its key technologies and innovations. It features newly developed 2/3-inch native 4K CMOS sensors, a next generation high speed video processor called AXII, and supports high dynamic range video formats. The camera employs a 3-chip design for improved resolution, sensitivity, and reduced aliasing compared to single-chip cameras. It allows for uncompressed 4K signal transmission and future-proof expandability through modular components.
This document outlines elements of high-quality image production, including spatial and temporal resolution, dynamic range, color gamut, bit depth, and coding. It discusses color gamut conversion, gamma correction, HDR and SDR mastering, tone mapping, and backwards compatibility. The document also covers HDR metadata standards and different distribution scenarios for HDR content.
This document provides an overview of high definition television (HDTV) standards and concepts such as color gamut, color bars test signals, colorimetry, chroma adjustment, and luminance adjustment. It discusses differences between standard definition (SDTV) and HDTV color bars, how wider color gamuts in HDTV allow for deeper colors, and how to use various elements of the color bars signal to properly adjust a display's color, brightness, contrast, and chroma. The document contains diagrams demonstrating color gamuts and examples of how objects appear within different gamuts.
This document provides an overview of color spaces and high dynamic range (HDR) technologies. It begins with definitions of color gamut and chromaticity coordinates. It then discusses several key color spaces including Rec.709, Rec.2020, DCI-P3, ACES, and S-Gamut3. It also covers HDR formats like PQ, HLG, and log encoding. The document aims to explain the essential aspects of different color spaces and HDR technologies used for digital cinema and television production.
The document discusses high dynamic range (HDR) video technology including:
- Different HDR formats such as SMPTE ST 2084 (PQ), ARIB STB-B67/ITU-R BT.2100 (HLG)
- Code value ranges for 10-bit and 12-bit RGB and color difference signals in narrow and full ranges
- Recommendations for using narrow versus full signal ranges for PQ and HLG
- Transcoding concepts when converting between PQ and HLG formats
- Considerations for including standard dynamic range (SDR) content in HDR programs
CCTV, or closed-circuit television, is a technology that uses cameras to transmit video images to a specific place for monitoring and security purposes. The document discusses various types of CCTV cameras like dome cameras, bullet cameras, and their applications in areas like industries, crime prevention, and traffic monitoring. It also covers CCTV system components, types of lenses, image sensors, features like WDR, and considerations for camera selection and system design. Digital video recorders (DVRs) are also discussed which record video from CCTV cameras.
This document provides an overview of video standards and concepts related to standard definition television (SDTV) and high definition television (HDTV). It begins with definitions of key terms like interlacing, progressive scanning, and frame rates. It then covers standards for monochrome signals, including signal timings, synchronization pulses, and blanking intervals. Digital SDTV standards like line counts, field structures, and ancillary data space are also summarized. The document concludes with discussions of spatial resolution, optimal viewing distances, and different aspect ratios used in television.
HDR, wide color gamut, and higher frame rates are new technologies that can improve image quality for ultra high definition televisions. They provide benefits like more vivid colors, deeper blacks, better shadow detail, and a more immersive viewing experience. However, supporting these new features requires significantly more data bandwidth compared to legacy standards. Future video standards will need to efficiently support higher resolutions, wider color, high dynamic range, and high frame rates to deliver next-generation picture quality while still allowing content to be economically distributed.
This document presents a system for 3D TV broadcasting and distribution that is compatible with existing 2D TV systems. The aim is to develop a system that can deliver stereoscopic image pairs to mobile and home users based on digital multimedia broadcasting (DMB) and digital video broadcasting - handheld (DVB-H) standards. It discusses encoding and transmitting left and right views of 3D video using these standards in a way that maintains backward compatibility with 2D receivers. The system is designed to support 3D content on mobile devices and televisions in both broadcast and high definition formats.
The document discusses high dynamic range (HDR) imaging technologies including:
- Standards for HDR encoding like SMPTE ST 2084 (PQ) and ARIB/ITU-R BT.2100 (HLG)
- Opto-electronic transfer functions (OETFs) and electro-optical transfer functions (EOTFs) used in HDR systems
- The human visual system's sensitivity to luminance levels and how this relates to quantization in HDR images
This document discusses key elements that contribute to high quality image production, including spatial resolution, frame rate, dynamic range, color gamut, bit depth, and compression artifacts. It examines these elements in the context of 4K and 8K broadcast cameras and their advantages over HD. Factors like wider viewing angles, increased perceived motion, and benefits for nature documentaries are cited as motivations for 8K. Technical details covered include lens flange back distance, flare, shading, chromatic aberration, and testing procedures. Overall quality is represented as a function of these various image quality factors.
Serial Digital Interface (SDI), From SD-SDI to 24G-SDI, Part 1Dr. Mohieddin Moradi
The document discusses standards for serial digital interface (SDI) video signals. It provides information on:
- Early SDI standards including SMPTE 259M for SD-SDI at 270Mbps and how they standardized a serial digital video connection.
- Video signal sampling structures and resolutions for SD, HD, and UHD formats.
- The development of higher data rate SDI standards up to 12G-SDI and 24G-SDI to support higher resolution video.
- Electrical parameters and cable distance limitations for different SDI data rates.
Serial Digital Interface (SDI), From SD-SDI to 24G-SDI, Part 2Dr. Mohieddin Moradi
This document discusses high definition video standards including SMPTE 274M, 292M, 372M and dual link SDI formats. It provides details on:
- The HD-SDI standards that define 1080p and 720p video formats and carriage through 1.5Gb/s serial digital interface.
- The timing reference signal codes used in HD-SDI to identify lines and perform error checking.
- How a 12-bit color depth can be achieved within the dual link standard by mapping the additional bits across both links.
- The benefits of 3Gb/s SDI and dual link formats for working at higher resolutions and color spaces prior to finishing.
Video and television systems work by presenting a sequence of images rapidly enough that the human eye perceives them as continuous motion. Different regions use different television standards that determine aspects like the number of lines, frames per second, and color systems. Video compression codecs like MPEG remove spatial and temporal redundancy to greatly reduce file sizes for storage and transmission while maintaining adequate quality.
Dr. Mohieddin Moradi provides an outline on high dynamic range (HDR) technology. The 3-page document covers various topics related to HDR including different HDR technologies, tone mapping, color representation, and HDR standards. It discusses concepts such as scene-referred vs display-referred conversions, and direct mapping vs tone mapping when converting between HDR and SDR formats. The document also examines potential side effects when mixing different conversion techniques in a production workflow.
This document discusses various standards and technologies related to video. It begins by explaining the different analog television color encoding systems used in different countries: PAL, SECAM, and NTSC. It then provides more details on each system, including lines, frames per second, and regions of use. The document also covers standards conversion between these analog systems, aspect ratios, composite and component video, and digital interfaces like HDMI and DVI.
What is In side a Television Broadcasting StationAnvar sha S
Industrial Training Report (TV Broadcast Station )Under My guidance .
For The Post Graduate Students in Electronics and Communication
Amal Jyothi College of Engineering
The document discusses various types of text, text compression methods, video formats and standards, and video recording systems. It describes unformatted, formatted and hypertext; Huffman coding and LZ/LZW compression; common text formats like TXT, DOC, RTF, PDF and PS. It also explains video signal formats, transmission standards including NTSC, PAL, EDTV, CIF and HDTV, and video recording systems such as VHS, camcorder, laserdisc, VCD and DVD.
The document provides an overview of key elements and trends in high-quality image production, including spatial resolution, temporal resolution, dynamic range, color gamut, quantization, and related technologies. It discusses technologies like HD, UHD, HDR and WCG and how they improve the total quality of experience. Images and charts are included to illustrate comparisons of technologies and results from industry surveys on trends and commercial projects.
A video signal carries visual information from a video source to a display. There are different types of analog video signals like composite, S-video, and component video. Digital video represents moving images using encoded digital data rather than a continuous analog signal. Digital video has advantages like no loss of quality during transmission and support for higher resolutions. Common digital video interfaces include DVI and HDMI, while TV standards are categorized as SDTV, EDTV, or HDTV based on resolution and frame rate.
H.120 was the first digital video coding standard developed in 1984. H.261 in the late 1980s was the first widespread success and established the modern structure for video compression that is still used today. MPEG-1 and MPEG-2/H.262 built upon H.261 with improvements like bidirectional prediction and half-pixel motion compensation. H.263 further enhanced compression performance and is now dominant for videoconferencing, adding features such as overlapped block motion compensation.
This document summarizes Juan Pedro López Velasco's thesis work on developing visual attention and perception models for assessing video quality. The work has two main objectives: 1) Predicting visual discomfort in 3D stereoscopic video by analyzing factors like motion, disparity, and parallax changes. 2) Improving 2D video quality metrics by applying visual attention models that weight regions of interest to better correspond to human perception. The work involves conducting subjective testing to determine important quality factors, developing computational models of visual attention, and incorporating these models into new objective metrics to provide more accurate quality assessment.
NY Radiation Safety/Quality Assurance Programqubyx
PerfectLum software can help hospitals comply with New York State's Primary Diagnostic Monitor quality assurance program. It performs all required tests, including DICOM GSDF verification and luminance tests, through a wizard interface. It saves reports to demonstrate compliance and automatically schedules future tests. PerfectLum's remote management allows centralized quality assurance across multiple hospital sites.
The document discusses high dynamic range (HDR) video technology including:
- Different HDR formats such as SMPTE ST 2084 (PQ), ARIB STB-B67/ITU-R BT.2100 (HLG)
- Code value ranges for 10-bit and 12-bit RGB and color difference signals in narrow and full ranges
- Recommendations for using narrow versus full signal ranges for PQ and HLG
- Transcoding concepts when converting between PQ and HLG formats
- Considerations for including standard dynamic range (SDR) content in HDR programs
CCTV, or closed-circuit television, is a technology that uses cameras to transmit video images to a specific place for monitoring and security purposes. The document discusses various types of CCTV cameras like dome cameras, bullet cameras, and their applications in areas like industries, crime prevention, and traffic monitoring. It also covers CCTV system components, types of lenses, image sensors, features like WDR, and considerations for camera selection and system design. Digital video recorders (DVRs) are also discussed which record video from CCTV cameras.
This document provides an overview of video standards and concepts related to standard definition television (SDTV) and high definition television (HDTV). It begins with definitions of key terms like interlacing, progressive scanning, and frame rates. It then covers standards for monochrome signals, including signal timings, synchronization pulses, and blanking intervals. Digital SDTV standards like line counts, field structures, and ancillary data space are also summarized. The document concludes with discussions of spatial resolution, optimal viewing distances, and different aspect ratios used in television.
HDR, wide color gamut, and higher frame rates are new technologies that can improve image quality for ultra high definition televisions. They provide benefits like more vivid colors, deeper blacks, better shadow detail, and a more immersive viewing experience. However, supporting these new features requires significantly more data bandwidth compared to legacy standards. Future video standards will need to efficiently support higher resolutions, wider color, high dynamic range, and high frame rates to deliver next-generation picture quality while still allowing content to be economically distributed.
This document presents a system for 3D TV broadcasting and distribution that is compatible with existing 2D TV systems. The aim is to develop a system that can deliver stereoscopic image pairs to mobile and home users based on digital multimedia broadcasting (DMB) and digital video broadcasting - handheld (DVB-H) standards. It discusses encoding and transmitting left and right views of 3D video using these standards in a way that maintains backward compatibility with 2D receivers. The system is designed to support 3D content on mobile devices and televisions in both broadcast and high definition formats.
The document discusses high dynamic range (HDR) imaging technologies including:
- Standards for HDR encoding like SMPTE ST 2084 (PQ) and ARIB/ITU-R BT.2100 (HLG)
- Opto-electronic transfer functions (OETFs) and electro-optical transfer functions (EOTFs) used in HDR systems
- The human visual system's sensitivity to luminance levels and how this relates to quantization in HDR images
This document discusses key elements that contribute to high quality image production, including spatial resolution, frame rate, dynamic range, color gamut, bit depth, and compression artifacts. It examines these elements in the context of 4K and 8K broadcast cameras and their advantages over HD. Factors like wider viewing angles, increased perceived motion, and benefits for nature documentaries are cited as motivations for 8K. Technical details covered include lens flange back distance, flare, shading, chromatic aberration, and testing procedures. Overall quality is represented as a function of these various image quality factors.
Serial Digital Interface (SDI), From SD-SDI to 24G-SDI, Part 1Dr. Mohieddin Moradi
The document discusses standards for serial digital interface (SDI) video signals. It provides information on:
- Early SDI standards including SMPTE 259M for SD-SDI at 270Mbps and how they standardized a serial digital video connection.
- Video signal sampling structures and resolutions for SD, HD, and UHD formats.
- The development of higher data rate SDI standards up to 12G-SDI and 24G-SDI to support higher resolution video.
- Electrical parameters and cable distance limitations for different SDI data rates.
Serial Digital Interface (SDI), From SD-SDI to 24G-SDI, Part 2Dr. Mohieddin Moradi
This document discusses high definition video standards including SMPTE 274M, 292M, 372M and dual link SDI formats. It provides details on:
- The HD-SDI standards that define 1080p and 720p video formats and carriage through 1.5Gb/s serial digital interface.
- The timing reference signal codes used in HD-SDI to identify lines and perform error checking.
- How a 12-bit color depth can be achieved within the dual link standard by mapping the additional bits across both links.
- The benefits of 3Gb/s SDI and dual link formats for working at higher resolutions and color spaces prior to finishing.
Video and television systems work by presenting a sequence of images rapidly enough that the human eye perceives them as continuous motion. Different regions use different television standards that determine aspects like the number of lines, frames per second, and color systems. Video compression codecs like MPEG remove spatial and temporal redundancy to greatly reduce file sizes for storage and transmission while maintaining adequate quality.
Dr. Mohieddin Moradi provides an outline on high dynamic range (HDR) technology. The 3-page document covers various topics related to HDR including different HDR technologies, tone mapping, color representation, and HDR standards. It discusses concepts such as scene-referred vs display-referred conversions, and direct mapping vs tone mapping when converting between HDR and SDR formats. The document also examines potential side effects when mixing different conversion techniques in a production workflow.
This document discusses various standards and technologies related to video. It begins by explaining the different analog television color encoding systems used in different countries: PAL, SECAM, and NTSC. It then provides more details on each system, including lines, frames per second, and regions of use. The document also covers standards conversion between these analog systems, aspect ratios, composite and component video, and digital interfaces like HDMI and DVI.
What is In side a Television Broadcasting StationAnvar sha S
Industrial Training Report (TV Broadcast Station )Under My guidance .
For The Post Graduate Students in Electronics and Communication
Amal Jyothi College of Engineering
The document discusses various types of text, text compression methods, video formats and standards, and video recording systems. It describes unformatted, formatted and hypertext; Huffman coding and LZ/LZW compression; common text formats like TXT, DOC, RTF, PDF and PS. It also explains video signal formats, transmission standards including NTSC, PAL, EDTV, CIF and HDTV, and video recording systems such as VHS, camcorder, laserdisc, VCD and DVD.
The document provides an overview of key elements and trends in high-quality image production, including spatial resolution, temporal resolution, dynamic range, color gamut, quantization, and related technologies. It discusses technologies like HD, UHD, HDR and WCG and how they improve the total quality of experience. Images and charts are included to illustrate comparisons of technologies and results from industry surveys on trends and commercial projects.
A video signal carries visual information from a video source to a display. There are different types of analog video signals like composite, S-video, and component video. Digital video represents moving images using encoded digital data rather than a continuous analog signal. Digital video has advantages like no loss of quality during transmission and support for higher resolutions. Common digital video interfaces include DVI and HDMI, while TV standards are categorized as SDTV, EDTV, or HDTV based on resolution and frame rate.
H.120 was the first digital video coding standard developed in 1984. H.261 in the late 1980s was the first widespread success and established the modern structure for video compression that is still used today. MPEG-1 and MPEG-2/H.262 built upon H.261 with improvements like bidirectional prediction and half-pixel motion compensation. H.263 further enhanced compression performance and is now dominant for videoconferencing, adding features such as overlapped block motion compensation.
This document summarizes Juan Pedro López Velasco's thesis work on developing visual attention and perception models for assessing video quality. The work has two main objectives: 1) Predicting visual discomfort in 3D stereoscopic video by analyzing factors like motion, disparity, and parallax changes. 2) Improving 2D video quality metrics by applying visual attention models that weight regions of interest to better correspond to human perception. The work involves conducting subjective testing to determine important quality factors, developing computational models of visual attention, and incorporating these models into new objective metrics to provide more accurate quality assessment.
NY Radiation Safety/Quality Assurance Programqubyx
PerfectLum software can help hospitals comply with New York State's Primary Diagnostic Monitor quality assurance program. It performs all required tests, including DICOM GSDF verification and luminance tests, through a wizard interface. It saves reports to demonstrate compliance and automatically schedules future tests. PerfectLum's remote management allows centralized quality assurance across multiple hospital sites.
Article on Remote Visual Inspection | Integrated NDE Solution.pdfintegratedndesolutio
With the use of cutting-edge GE borescope equipment, Integrated NDE Solution in Pune provides skilled remote visual inspection services for tight and challenging-to-reach cavities. For visual inspection of industrial gas turbines, diesel engines, steam turbines, airplane engines, and more, our borescopes are perfect. This approach ensures that parts fulfill industry norms, standards, or customer specifications while reducing the need for needless maintenance, particularly for big turbines. Our extensive NDT services offer the highest levels of safety and dependability through both laboratory and on-site examinations.
This document discusses key considerations for machine vision systems. It explains that vision systems are unique to each application and require custom engineering. It then discusses important elements like lighting, lenses, camera sensors, and integration with control systems. Proper lighting and component selection are essential for system success. The document also outlines common vision system types and imaging options like 2D and 3D.
Iec 62676 5 standardized spezification of image qualityHenry Rutjes
This document discusses IEC 62676-5, a standard for specifying image quality performance metrics for camera devices used in security applications. It outlines several methods defined in the standard for evaluating key image quality parameters such as resolution, optical/electronic characteristic function (OECF), dynamic range, minimum illumination, visible dynamic range, infrared illumination operating distance, image distortion, and image flare. The intention of the standard is to provide standardized procedures and definitions for measuring and reporting these parameters in a meaningful way to help users select the right camera for their application.
Yole Intel RealSense 3D camera module and STM IR laser 2015 teardown reverse ...Yole Developpement
INNOVATIVE 3D CAMERA FOR FACIAL ANALYSIS AND HAND/FINGER TRACKING, BASED ON RESONANT MICRO-MIRROR, IR LASER, VISIBLE AND NEAR INFRARED IMAGE SENSORS.
Intel RealSense is an intelligent 3D camera equipped with a system of three components: a conventional camera, a near infrared image sensor and an infrared laser projector. Infrared parts are used to calculate the distance between objects, but also to separate objects on different planes. They serve for facial recognition as well as gestures tracking.
The Intel 3D camera can scan the environment from 0.2m to 1.2m. The fixed-focal length camera will support up to 1080p @30FPS capture in RGB with a 77° FOV. Its lens has a built in IR cut filter. The 640x480 pixel VGA camera has a frame rate up to 60fps with a 90° FOV, moreover its lens has an IR Band Pass filter.
More information on that report at http://www.i-micronews.com/reports.html
The eye gaze system is a communication and control system that uses eye movements to allow interaction for people with physical disabilities. It works by using an infrared camera to track the pupil and correlate eye position with points on a computer screen. The system requires the user to have control of one eye and adequate vision. It must be calibrated to each individual user by having them follow a moving target on screen. Once calibrated, users can control software programs and devices with their eyes by fixating on icons or buttons. While expensive, the eye gaze system provides an important access method for those unable to use other input systems.
This interim report provides an overview of a project to develop a monitor calibrator kit. The kit aims to quickly calibrate CRT monitors used in broadcasting to ensure accurate color reproduction. It will generate test signals, measure the monitor's output using a handheld sensor, and guide the user to adjust monitor settings accordingly. So far, market research has been conducted, objectives defined, and initial prototyping has included circuit design and component testing. Future work includes further sensor testing and developing the signal generator. The project remains on schedule and within budget.
The eye gaze system is a communication and control system that uses eye tracking to allow people with physical disabilities to control a computer. It works by using an infrared camera to track the user's pupil and determine where they are looking on the screen. It then uses this eye gaze data to select icons, type, or control other devices. The system requires the user to first calibrate by looking at points on the screen. It has applications in communication, education, and research. While expensive, it provides an important access method for people unable to use other input devices.
This document discusses optimizing 360-degree video streaming to head-mounted virtual reality. It covers challenges like existing codecs only supporting 2D videos and 360 videos having wider views than conventional videos. Approaches proposed include fixation prediction to avoid streaming unwatched parts, QoE modeling designed for 360 videos to improve user experience, and an adaptive streaming platform to select and transmit tiles based on fixation prediction while allocating bitrates based on the QoE model. Part I discusses fixation prediction including using neural networks trained on viewing features. Part II covers QoE modeling, noting limitations of existing metrics and factors that affect QoE like content and bitrates. It constructs a logarithmic linear QoE model. Part III outlines an
A survey on Measurement of Objective Video Quality in Social Cloud using Mach...IRJET Journal
This document discusses using machine learning techniques for objective measurement of video quality in social cloud applications. It proposes a method using convolutional neural networks (CNNs) trained on a large dataset of videos evaluated by human observers to predict video quality. The CNN model extracts quality-related features from videos which are used to train a machine learning model. The model achieved high accuracy in predicting subjective quality outcomes for new videos. Machine learning provides an objective way to assess video quality that can benefit users and service providers by optimizing video streaming quality. The document reviews key concepts like video quality assessment, deep learning and CNNs, and the typical methodology used in machine learning-based video quality measurement.
qualification of instrument(UV & FTIR) BY Bhumi Suratiya, M.Pharm sem 2.pptxBhumiSuratiya
So, Here it's presentation on Qualification of instrument(UV & FTIR) BY Bhumi Suratiya, M.Pharm sem 2. Qualification of analytical instruments. it also includes design qualifications (DQ), Installation qualification(IQ)
Operational qualification(OQ)
Performance qualification(PQ).
Barco provides healthcare professionals with diagnostic displays and imaging solutions to support patient care. Their product catalog outlines diagnostic displays in various resolutions, sizes, and modalities from their Nio and Coronis lines. These displays feature technologies like LED backlights, sensors, and interfaces that ensure accurate and consistent images over time for improved diagnosis.
Proposal to supply, installation and testing of CCTV Camera Denis kisina
The document is a proposal from The Technology Doctor to supply, install, and commission a CCTV surveillance system. It includes an executive summary, introduction, technical proposal, management proposal, and commercial proposal. The key details are:
- The proposal offers a high performance managed surveillance service with 24/7/365 availability using commercial off-the-shelf equipment in a modular system.
- The technical solution includes CCTV cameras, network equipment, software, and backup power. It will be designed to integrate seamlessly with existing systems.
- Project management will follow PRINCE2 methodology to deliver the solution on time and on budget, with a dedicated project manager and quality assurance processes.
The document discusses quality attributes of embedded processors. It outlines operational quality attributes such as response, throughput, reliability, maintainability, security, and safety. Response measures how fast the system tracks inputs. Throughput deals with efficiency. Reliability measures dependability. Maintainability supports troubleshooting. Security ensures confidentiality, integrity, and availability of information. Safety analyzes potential damage. Non-operational attributes include testability, evolvability, portability, time-to-prototype, and per-unit cost and revenue. Testability and debug-ability facilitate testing. Evolvability allows adapting to new technologies. Portability provides system independence. Time-to-prototype and revenue influence bringing products to market.
Calibrating medical displays to DICOM standards is important for diagnostic accuracy and quality assurance. The human eye perceives light and color differently than digital displays. DICOM Part 14 specifies how pixel values should be displayed to match human vision. Regular calibration ensures displays accurately depict images over time as components like backlights degrade, and allows consistent viewing across clinical environments. Most LCD, LED, and OLED displays can be calibrated using software that generates lookup tables to map pixel values to luminance levels defined by DICOM standards.
The document describes an efector dualis vision sensor that can solve a variety of inspection and error-proofing applications throughout the manufacturing process. It has a compact CMOS image sensor and fast image capture and processing capabilities. The sensor includes integrated lighting, evaluation electronics, and an industrial housing. It can be easily configured via an Ethernet interface and Setup Wizard software. Example applications discussed include part verification, orientation detection, sortation, and detecting part presence or absence. The sensor provides a simpler and more cost-effective solution than traditional vision systems or sensor clusters.
The document proposes an electronic security installation for the Food and Drugs Authority that includes CCTV cameras, access control systems, and other physical security measures. It provides an overview of the assessment conducted, highlights risks around access control and CCTV, outlines priority areas for installation, and presents design criteria and equipment specifications. The proposal recommends a full security system from Afwest International Security, highlighting their experience, support, and some client examples.
Kingpak Corp's main production is CMOS image sensors. Over the past 5 years, they have built up their internal image sensor testing capabilities by:
1) Designing and implementing their own auto image test systems using self-designed test kits, software, and an auto handler.
2) Working with customers to define image specifications and test requirements.
3) Saving costs by developing their own test equipment rather than outsourcing, reducing the cost of test sets from NT$500K in 2006 to NT$50K in 2012.
Similar to Video Compression, Part 4 Section 2, Video Quality Assessment (20)
This document provides information about quality control testing of audiovisual content. It discusses various quality control tests that can be performed, including tests for analogue frame synchronization errors, black bars, constant colour frames, flashing video, macroblocking, video deinterlacing artifacts, and digital tape dropouts. Examples are provided for how each test can be configured and what results might look like. The goal of the quality control tests is to help broadcasters optimize their automated quality control systems and cope with increasing amounts of digital content.
The document outlines topics related to video over IP infrastructure and standards. It discusses IP technology trends, networking basics, video and audio over IP standards, SMPTE ST 2110, NMOS, infrastructure considerations, timing issues, clean switching methods, compression, broadcast controller/orchestration, and case studies for migrating broadcast facilities to IP. The document provides an overview and outline for presenting on designing, integrating, and managing IP-based broadcast facilities and production workflows.
This document outlines an educational course on audio and video over IP. The course covers IP networking fundamentals and standards including TCP/IP, OSI models, and SMPTE ST 2110. It also examines IP infrastructure, routing, timing issues, switching, compression techniques and case studies for broadcast facilities transitioning to IP. The document provides an in-depth outline of topics covered in each session, from IP basics to designing and integrating both hybrid and fully IP-based outside broadcast trucks. The goal is to educate on best practices for implementing audio and video over IP workflows and infrastructure.
This document provides definitions and explanations of various optical terminology related to light passing through a lens, including:
- Dispersion, refraction, diffraction, reflection, focal point, focal length, principal point, image circle, aperture ratio, numerical aperture, optical axis, and more. It discusses concepts such as entrance pupil, exit pupil, angular aperture, and how they relate to lens performance. The document also covers topics like vignetting, the cosine law, and flare. Overall, it serves as a comprehensive reference for understanding optical and photographic lens terminology.
This document discusses various optical and technical aspects of camera lenses, including:
1) It defines focal length as the distance between a lens and the point where light passing through converges, known as the focal point. Shorter focal lengths provide wide-angle views while longer focal lengths provide magnified close-up views.
2) F-number and f-stop are defined, with f-number indicating the maximum light a lens can admit and f-stop indicating light levels at smaller iris openings. Smaller f-numbers and f-stop numbers admit more light.
3) The relationship between aperture, focal length, and depth of field is explained. Smaller apertures provide deeper depth of field while
This document provides an overview of color video signals and color perception by the human visual system. It discusses:
1. The sensitivity of human cone cells to different wavelengths of light and how this determines color perception.
2. How color video signals like YUV, RGB, and composite video encode color and brightness information.
3. Standards for analog color television transmission including NTSC, PAL, and SECAM which differ in aspects like lines, frame rate, and color encoding.
This document provides an overview of analog and digital triax systems used for video transmission. It discusses key aspects of triax cables such as their ability to transmit multiple signals simultaneously through bundled cables. Both analog and digital triax systems are described, with analog transmitting component signals on different carrier frequencies and digital transmitting signals in digital format. The document also covers triax cable specifications, common connectors types used for broadcasting applications from different standards, fiber optic cable types including single mode and multi-mode, and common fiber connectors. Transmission distances and electrical properties of triax cables are discussed.
The document provides an overview of key concepts in high definition television (HDTV) including:
- Standards and definitions for SDTV and HDTV
- Interlacing and de-interlacing techniques
- Video scaling, edge enhancement, and frame rate conversion
- Signal quality issues in HDTV production and broadcast
- Cables and connectors used for HDTV production
The document contains diagrams and explanations of topics like color bars, genlocking, sampling, interlacing, field order, and 3D video sampling structures. It compares progressive and interlaced scanning and discusses concepts such as the Nyquist frequency, aliasing, and field dominance.
1. The document discusses color temperature and how different light sources emit different color spectrums that video cameras must account for through color balancing. Color temperature is used as a reference to adjust the camera's color balance to match the light source.
2. After color temperature conversion optically or electronically, white balance is then used to precisely match the light source color temperature by adjusting the camera's video amplifiers.
3. Other topics covered include polarizers, neutral density filters, and technical aspects of video such as gamma correction and clipping levels.
Hue refers to the dominant wavelength of light, which determines the color as perceived by the observer. Saturation refers to the purity of the hue, or the amount of white light mixed with it. Luminance refers to the brightness or intensity of the color.
The document discusses radiometry and photometry, which deal with measuring light across the electromagnetic spectrum and in the visible spectrum respectively. It defines terms like luminous flux, luminous intensity, illuminance, and luminance.
It also covers topics like additive and subtractive color mixing, primary and secondary colors, color spaces, and video signal formats like RGB, YUV, and YCbCr which are used to represent color images and video. Human cone sensitivity
This document provides an overview of sound and hearing, including:
1. It describes how the human ear works, from collecting sound waves through the outer ear and transmitting vibrations through the ossicles to the cochlea where hair cells detect different frequencies.
2. It discusses properties of sound like loudness, pitch, and timbre, and how they are perceived. Loudness depends on amplitude, pitch on frequency, and timbre on waveform complexity.
3. It explains characteristics of sound waves like wavelength, frequency, speed of sound, and the decibel scale used to measure sound intensity and pressure levels.
Video Compression, Part 3-Section 2, Some Standard Video CodecsDr. Mohieddin Moradi
This document discusses MPEG-2 Transport Streams and Packetized Elementary Streams. It describes how MPEG-2 Transport Streams use fixed length 188 byte packets containing compressed video, audio or data from one or more programs identified by Packet IDs. These packets can contain Packetized Elementary Stream packets which contain compressed elementary streams with timestamps for synchronization. The document also discusses how Transport Streams allow for synchronous multiplexing of multiple programs from independent time bases into a single stream.
Video Compression, Part 3-Section 1, Some Standard Video CodecsDr. Mohieddin Moradi
- ISO/IEC JTC 1/SC 29 and ITU-T are the main organizations that develop video coding standards through working groups like MPEG and VCEG.
- Early standards include H.261 for video telephony and conferencing, and MPEG-1 for DVD quality video.
- Later standards like H.264/AVC, HEVC, and future VVC provide increasingly higher compression through use of block transforms, motion compensation, and entropy coding in a hybrid video codec framework.
- Key organizations periodically collaborate through joint teams like JVT and JCT-VC to develop standards like AVC and HEVC.
This document discusses video compression techniques. It begins by outlining the history of video compression and describing the basic components of a generic video encoder/decoder system. It then covers specific compression methods including differential pulse code modulation, transform coding using discrete cosine transform, quantization, and entropy coding. The document also discusses techniques for reducing both spatial and temporal redundancy in video, such as prediction coding. It provides examples of how quantization is used to control quality and compression ratio in both lossy and lossless compression systems.
The document discusses video compression history and standards, including codecs such as H.261, H.262/MPEG-2, H.263, H.264/AVC, H.265/HEVC, and the roles of organizations like MPEG, VCEG, and ITU-T in developing video coding standards to ensure interoperability. It also covers video encoding and decoding principles, as well as common container formats and their applications in areas like broadcasting, streaming, and storage.
The document provides a history of the development of television technology from the late 1800s through the 1920s. Some key developments include:
- In 1873, experiments with selenium, which is light-sensitive and formed the basis for early televisions.
- In 1884, the Nipkow disk laid down many basic concepts like scanning and synchronization.
- In 1923, Vladimir Zworykin developed the Kinescope, which allowed television programs to be recorded on film.
- In 1924, John Logie Baird transmitted the first television image.
- In 1925, Vladimir Zworykin demonstrated 60-line television using a curved-line image structure typical of mechanical television at the time.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELijaia
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Comparative analysis between traditional aquaponics and reconstructed aquapon...bijceesjournal
The aquaponic system of planting is a method that does not require soil usage. It is a method that only needs water, fish, lava rocks (a substitute for soil), and plants. Aquaponic systems are sustainable and environmentally friendly. Its use not only helps to plant in small spaces but also helps reduce artificial chemical use and minimizes excess water use, as aquaponics consumes 90% less water than soil-based gardening. The study applied a descriptive and experimental design to assess and compare conventional and reconstructed aquaponic methods for reproducing tomatoes. The researchers created an observation checklist to determine the significant factors of the study. The study aims to determine the significant difference between traditional aquaponics and reconstructed aquaponics systems propagating tomatoes in terms of height, weight, girth, and number of fruits. The reconstructed aquaponics system’s higher growth yield results in a much more nourished crop than the traditional aquaponics system. It is superior in its number of fruits, height, weight, and girth measurement. Moreover, the reconstructed aquaponics system is proven to eliminate all the hindrances present in the traditional aquaponics system, which are overcrowding of fish, algae growth, pest problems, contaminated water, and dead fish.
VARIABLE FREQUENCY DRIVE. VFDs are widely used in industrial applications for...PIMR BHOPAL
Variable frequency drive .A Variable Frequency Drive (VFD) is an electronic device used to control the speed and torque of an electric motor by varying the frequency and voltage of its power supply. VFDs are widely used in industrial applications for motor control, providing significant energy savings and precise motor operation.
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...shadow0702a
This document serves as a comprehensive step-by-step guide on how to effectively use PyCharm for remote debugging of the Windows Subsystem for Linux (WSL) on a local Windows machine. It meticulously outlines several critical steps in the process, starting with the crucial task of enabling permissions, followed by the installation and configuration of WSL.
The guide then proceeds to explain how to set up the SSH service within the WSL environment, an integral part of the process. Alongside this, it also provides detailed instructions on how to modify the inbound rules of the Windows firewall to facilitate the process, ensuring that there are no connectivity issues that could potentially hinder the debugging process.
The document further emphasizes on the importance of checking the connection between the Windows and WSL environments, providing instructions on how to ensure that the connection is optimal and ready for remote debugging.
It also offers an in-depth guide on how to configure the WSL interpreter and files within the PyCharm environment. This is essential for ensuring that the debugging process is set up correctly and that the program can be run effectively within the WSL terminal.
Additionally, the document provides guidance on how to set up breakpoints for debugging, a fundamental aspect of the debugging process which allows the developer to stop the execution of their code at certain points and inspect their program at those stages.
Finally, the document concludes by providing a link to a reference blog. This blog offers additional information and guidance on configuring the remote Python interpreter in PyCharm, providing the reader with a well-rounded understanding of the process.
Applications of artificial Intelligence in Mechanical Engineering.pdfAtif Razi
Historically, mechanical engineering has relied heavily on human expertise and empirical methods to solve complex problems. With the introduction of computer-aided design (CAD) and finite element analysis (FEA), the field took its first steps towards digitization. These tools allowed engineers to simulate and analyze mechanical systems with greater accuracy and efficiency. However, the sheer volume of data generated by modern engineering systems and the increasing complexity of these systems have necessitated more advanced analytical tools, paving the way for AI.
AI offers the capability to process vast amounts of data, identify patterns, and make predictions with a level of speed and accuracy unattainable by traditional methods. This has profound implications for mechanical engineering, enabling more efficient design processes, predictive maintenance strategies, and optimized manufacturing operations. AI-driven tools can learn from historical data, adapt to new information, and continuously improve their performance, making them invaluable in tackling the multifaceted challenges of modern mechanical engineering.
Build the Next Generation of Apps with the Einstein 1 Platform.
Rejoignez Philippe Ozil pour une session de workshops qui vous guidera à travers les détails de la plateforme Einstein 1, l'importance des données pour la création d'applications d'intelligence artificielle et les différents outils et technologies que Salesforce propose pour vous apporter tous les bénéfices de l'IA.
Discover the latest insights on Data Driven Maintenance with our comprehensive webinar presentation. Learn about traditional maintenance challenges, the right approach to utilizing data, and the benefits of adopting a Data Driven Maintenance strategy. Explore real-world examples, industry best practices, and innovative solutions like FMECA and the D3M model. This presentation, led by expert Jules Oudmans, is essential for asset owners looking to optimize their maintenance processes and leverage digital technologies for improved efficiency and performance. Download now to stay ahead in the evolving maintenance landscape.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
2. Section I
− Perceptual Artifacts in Compressed Video
− Quality Assessment in Compressed Video
− Objective Assessment of Compressed Video
− Objective Assessment of Compressed Video, Codec Assessment for Production
Section II
− Subjective Assessment of Compressed Video
− Subjective Assessment of Compressed Video, Subjective Assessment by Expert Viewing
− Performance Comparison of Video Coding Standards: An Adaptive Streaming Perspective
− Subjective Assessment by Visualizer™ Test Pattern
2
Outline
4. 4
Upper anchor, Source material
Material to be evaluated
Lower anchor , Worst case
VRT-medialab: onderzoek en
innovatie
Subjective Assessment
EBU Triple Stimulus Continuous Evaluation Scale
(TSCES) quality assessment rig
5. In subjective evaluation, a panel of human observers are given the task to rate the quality of image or video
sequence.
− There are normally 15 to 25 subjects, but the minimum is 15
− Subjects should have normal HVS (age 18-30)
− They should NOT be familiar with Image degradations (but now it is changed!)
− The average values of the score is called: Mean Opinion Score (MOS)
− MOS is normally defined in the range of 1-5
Subjective Assessment
5
6. − Golden Standard
• Provides the best predictive value of actual subjective ratings
− Absolute Category Rating (ACR)
• Human subjects are shown two video sequences (original and processed) and are asked to assess the
overall quality of the processed sequence with respect to the original (reference) sequence
− Mean Opinion Score (MOS)
• Numerical measure of the human-judged overall quality
• On a scale of 1 (bad), 2 (poor), 3 (fair), 4 (good), 5 (excellent)
− DMOS (Differential Mean Opinion Scores)
• It is defined as the difference between the raw quality score of the reference and test images.
Mean Opinion Score (MOS)
6
7. Laboratory Viewing Environment
− The laboratory viewing environment is intended to provide critical conditions to check systems.
− The laboratory viewing environment parameters have been selected to define an environment slightly
more critical than the typical home viewing situations.
Home Viewing Environment
− The home viewing environment is intended to provide a means to evaluate quality at the consumer side
of the TV chain.
General Viewing Conditions
7
8. In a Laboratory Environment
The assessors’ viewing conditions should be arranged as follows:
− Room illumination: low
− Chromaticity of background: D65
− Peak luminance (it should be adjusted according to the room illumination): 70-250 𝒄𝒅/𝒎 𝟐
− Monitor contrast ratio: ≤ 0.02
− Ratio of luminance of background behind picture monitor to peak luminance of picture:≃0.15
General Viewing Conditions for Subjective Assessments
8
9. In Home Environment
The assessors’ viewing conditions should be arranged as follows:
− Environmental illuminance on the screen (incident light from the environment falling on the screen, should
be measured perpendicularly to the screen): 200 lux
− Peak luminance (it should be adjusted according to the room illumination): 70-500 𝒄𝒅/𝒎 𝟐
− Ratio of luminance of inactive screen to peak luminance monitor contrast ratio: ≤ 0.02
General Viewing Conditions for Subjective Assessments
9
10. Preferred Viewing Distance (PVD)
− The PVD is based upon viewers’ preferences which have been determined empirically.
− The PVD contains a number of data sets collected from available sources.
− This information may be referred to for designing a subjective assessment test.
Viewing Distance
10
11. Design Viewing Distance (DVD).
− The design viewing distance (DVD), or optimal viewing distance, for a digital system is the distance at
which two adjacent pixels subtend an angle of 1 arc-min at the viewer’s eye; and the optimal horizontal
viewing angle as the angle under which an image is seen at its optimal viewing distance.
Viewing Distance
11
12. − Using monitors with different characteristics will yield different subjective image qualities.
− It is therefore strongly recommended that characteristics of the monitors used should be checked
beforehand.
• Recommendation ITU-R BT.1886 – Reference electro-optical transfer function for Flat Panel Displays
(FPD) used in HDTV studio production
• Report ITU-R BT.2129 – User requirements for a Flat Panel Display (FPD) as a Master monitor in an HDTV
programme production environment may be referred to when professional FPD monitors are used for
subjective assessment.
• Report ITU-R BT.2390 provides information concerning laboratory and home displays and viewing
environments for the assessment of High Dynamic Range (HDR) images.
The Monitor
12
13. − Monitor processing such as image scaling, frame rate conversion, image enhancer, if implemented,
should be done in such a way as to avoid introducing artefacts.
− HDR processing should be appropriate to the HDR system being assessed or being used during the
assessment. For consumer environment or distribution assessments, this may include the use of the
appropriate static or dynamic metadata.
− Full details of such metadata should be included in the notes of the assessments in order that other
laboratories can accurately repeat the assessments.
Monitor Processing
13
14. − When using consumer displays for subject image assessments, it is important that all image-processing
options are disabled unless the impact of such image processing is the subject of the assessment(s).
− When accessing interlace images, the test report should indicate whether de-interlacer has been used or
not. It is preferable not to use de-interlacer if interlaced signals can be displayed without it.
Monitor Processing
14
15. − The resolution of professional monitors usually complies with the required standards for subjective
assessments in their luminance operating range.
− To check and report the maximum and minimum resolutions (center and corners of the screen) at the
used luminance value might be suggested.
− If consumer FPD TV displays are used for subjective assessments, it is strongly recommended to check and
report the maximum and minimum resolutions (center and corners of the screen) at the used luminance
value.
− At present the most practical system available to subjective assessments performers, in order to check
monitors or consumer TV sets resolutions, is the use of a swept test pattern electronically generated.
Monitor resolution
15
16. Contrast
− Contrast could be strongly influenced by the environment illuminance.
− Professional monitors seldom use technologies to improve their contrast in a high illuminance environment,
so it is possible they do not comply with the requested contrast standard if used in a high illuminance
environment.
− Consumer monitors typically use technologies to get a better contrast in a high illuminance environment.
Brightness
− When adjusting the LCD monitor brightness, it is preferable to use backlight intensity control rather than
using “signal level” scaling to retain the bit precision.
− In the case of other display technologies that do not use a backlight, the white level should be adjusted
by means other than signal level scaling.
Monitor Contrast and Brightness
16
17. − The display should not introduce motion artefacts that are introduced by specific display technologies.
− On the other hand, the motion effects included in the input signal should be represented on the display.
− When using consumer displays, it is vital that ALL motion processing options are disabled.
Monitor Motion Processing Options
17
18. Assessment problem Material used
Overall performance with average material General, “critical but not unduly so”
Capacity, critical applications (e.g. contribution, post-
processing, etc.)
Range, including very critical material for the application
tested
Performance of “adaptive” systems Material very critical for “adaptive” scheme used
Identify weaknesses and possible improvements Critical, attribute-specific material
Identify factors on which systems are seen to vary Wide range of very rich material
Conversion among different standards Critical for differences (e.g. field rate)
A Survey of Typical Assessment Problems and Test Materials Used to Address Them
ITU-R Test Sequences
− A number of organizations have developed test still images and sequences.
− Report ITU-R BT.2245 - HDTV and UHDTV including HDR-TV test materials for assessment of picture quality,
gives details of HDTV and UHDTV test material that can be used for subjective assessment.
18
19. − Observers may be expert or non-expert depending on the objectives of the assessment.
− An expert observer is an observer that has expertise in the image artefacts that may be introduced by the
system under test.
− A non-expert (“naive”) observer is an observer that has no expertise in the image artefacts that may be
introduced by the system under test.
− In any case, observers should not be, or have been, directly involved, i.e. enough to acquire specific and
detailed knowledge, in the development of the system under study.
Expert and Non-expert Observers
19
20. − At least 15 observers should be used.
− The number of assessors needed depends upon the sensitivity and reliability of the test procedure
adopted and upon the anticipated size of the effect sought.
− For studies with limited scope, e.g. of exploratory nature, fewer than 15 observers may be used. In this
case, the study should be identified as ‘informal’.
− The level of expertise in television image quality assessment of the observers should be reported.
Observers Quantity
20
21. − Prior to a session, the observers should be screened for (corrected-to-) normal visual acuity on the Snellen
or Landolt chart, and for normal colour vision using specially selected charts (Ishihara, for instance).
− A study of consistency between results at different testing laboratories has found that systematic
differences can occur between results obtained from different laboratories.
− A possible explanation for the differences between different laboratories is that there may be different skill
levels amongst different groups of assessors.
− Suggested data to be provided could include an occupation category (e.g. broadcast organization
employee, university student, office worker, ...), gender, and age range.
Observer Screening and Test Consistency
21
22. − Assessors should be carefully introduced to
• the method of assessment
• the types of impairment or quality factors likely to occur
• the grading scale
• the sequence and timing
− Training sequences demonstrating the range and the type of the impairments to be assessed should be
used with illustrating images other than those used in the test, but of comparable sensitivity.
− In the case of quality assessments, quality may be defined as to consist of specific perceptual attributes.
Instructions for the Assessment
22
24. − A session should not last more than half an hour.
− At the beginning of the first session, about five “dummy presentations” should be introduced to stabilize
the observers’ opinion.
− The data issued from these presentations must not be considered in the results of the test.
− If several sessions are necessary, about three dummy presentations are only necessary at the beginning of
the following session.
− A random order should be used for the presentations (for example, derived from Graeco-Latin squares);
but the test condition order should be arranged so that any effects on the grading of tiredness or
adaptation are balanced out from session to session.
− Some of the presentations can be repeated from session to session to check coherence.
The Test Session
24
25. − For each test parameter, the mean and 95% confidence interval of the statistical distribution of the
assessment grades must be given.
− The results must be given together with the following information:
• details of the test configuration;
• details of the test materials;
• type of image source and display monitors;
• number and type of assessors;
• reference systems used;
• the grand mean score for the experiment;
• original and adjusted mean scores and 95% confidence interval if one or more observers have been
eliminated according to the procedure given below.
Presentation of the Results
25
26. When the time-varying picture quality of the processed video without reference is evaluated by the subjects,
the SSCQE is preferred.
− Viewers are shown each video sequence one by one and asked them to vote.
− Each sequence is shown only once and it consists of 30 seconds to several minutes of duration.
− Since video sequences are long, they are segmented into 10s shots, and for each video segment an MOS
is calculated.
− During the presentation of sequence, viewers continuously provide their rating scores by using a slider.
− The variations in slider are observed after every 500 msec which provides the continuous scale reading.
− The rating scale is divided into five equal lengths as explained for DSCQS.
Single Stimulus Continuous Quality Evaluation (SSCQE)
26
27. − A typical assessment might call for an evaluation of either a new system, or the effect of a transmission
path impairment.
− The double-stimulus (EBU) method is cyclic in that the assessor is first presented with an unimpaired
reference, then with the same image impaired.
− Following this, he is asked to vote on the second, keeping in mind the first.
The Double-stimulus Impairment Scale (DSIS)
27
28. − In sessions, which last up to half an hour, the assessor is presented with a series of images or sequences in
random order and with random impairments covering all required combinations.
− The unimpaired image is included in the images or sequences to be assessed.
− At the end of the series of sessions, the mean score for each test condition and test image is calculated.
− The method uses the impairment scale, for which it is usually found that the stability of the results is greater
for small impairments than for large impairments.
− Although the method sometimes has been used with limited ranges of impairments, it is more properly
used with a full range of impairments.
The Double-stimulus Impairment Scale (DSIS)
28
29. − Assessors are presented with a series of test images or sequences. They are arranged in pairs such that the
first in the pair comes direct from the source, and the second is the same image via the system under test.
The Double-stimulus Impairment Scale (DSIS)
Grading scales
• Imperceptible
• Perceptible, But Not Annoying
• Slightly Annoying
• Annoying
• Very Annoying
29
Source Signal
or
System Under Test.
30. Presentation of the Test Material
− A test session comprises a number of presentations.
− There are two variants to the structure of presentations, I and II outlined below.
• Variant I: The reference image or sequence and the test image or sequence are presented only once.
• Variant II: The reference image or sequence and the test image or sequence are presented twice. It is
more time consuming, may be applied if the discrimination of very small impairments is required or
moving sequences are under test.
The Double-stimulus Impairment Scale (DSIS)
T1 = 10s Reference image
T2 = 3s Mid-grey produced by video level of around 200 mV
T3 = 10s Test condition
T4 = 5 to 11s Mid-grey
Experience suggests that extending the periods T1 and T3 beyond 10 s does not
improve the assessor’s ability to grade the image sequences.
30
31. The Introduction to the Assessments
− At the beginning of each session, an explanation is given to the observers about the type of assessment,
the grading scale, the sequence and timing (reference image, grey, test image, voting period).
− The range and type of the impairments to be assessed should be illustrated on images other than those
used in the tests, but of comparable sensitivity.
− It must not be implied that the worst quality seen necessarily corresponds to the lowest subjective grade.
− Observers should be asked to base their judgement on the overall impression given by the image, and to
express these judgements in terms of the wordings used to define the subjective scale.
− The observers should be asked to look at the image for the whole of the duration of T1 and T3.
− Voting should be permitted only during T4.
The Double-stimulus Impairment Scale (DSIS)
31
32. The Test Session
− The images and impairments should be presented in a pseudo-random sequence and, preferably in a
different sequence for each session.
− In any case, the same test image or sequences should never be presented on two successive occasions
with the same or different levels of impairment.
− The range of impairments should be chosen so that all grades are used by the majority of observers; a
grand mean score (averaged overall judgements made in the experiment) close to three should be
aimed at.
− A session should not last more than roughly half an hour, including the explanations and preliminaries; the
test sequence could begin with a few images indicative of the range of impairments; judgements of these
images would not be taken into account in the final results.
The Double-stimulus Impairment Scale (DSIS)
32
33. − Video sequences are presented to observers in pair and the first video will always be reference video
followed by the coded video sequence, and instead of quality, impairment is rated.
− For voting: there are two methods used:
1st method: the reference video is followed by coded video and viewers are asked to vote.
2nd method: both reference and coded video sequences are presented twice to the observers and then
asked them for vote as illustrated in figure.
Double Stimulus Impairment Scale (DSIS), Summary
33
34. − A typical assessment might call for evaluation of a new system or of the effects of transmission paths on
quality.
− The double-stimulus method is thought to be especially useful when it is not possible to provide test
stimulus test conditions that exhibit the full range of quality.
− The method is cyclic in that the assessor is asked to view a pair of images, each from the same source, but
one via the process under examination, and the other one directly from the source.
− He is asked to assess the quality of both.
− In sessions which last up to half an hour, the assessor is presented with a series of image pairs (internally
random) in random order, and with random impairments covering all required combinations.
− At the end of the sessions, the mean scores for each test condition and test image are calculated.
The Double-stimulus Continuous Quality-scale (DSCQS)
34
35. General arrangement for test system for DSCQS method
The Double-stimulus Continuous Quality-scale (DSCQS)
T1 = 10 s Reference image
T2 = 3 s Mid-grey produced by video level of around 200 mV
T3 = 10 s Test condition
T4 = 5 to 11 s Mid-grey
35
36. Presentation of the Test Material
Variant I (a single observer):
− The assessor is free to switch between the A and B signals until the assessor has the mental measure of
the quality associated with each signal.
− The assessor may typically choose to do this two or three times for periods of up to 10s.
Variant II (use a number of observers simultaneously):
− Prior to recording results, the pair of conditions is shown one or more times for an equal length of time
to allow the assessor to gain the mental measure of the qualities associated with them, then the pair is
shown again one or more times while the results are recorded.
− The number of repetitions depends on the length of the test sequences.
− For still images, a 3-4 s sequence and five repetitions (voting during the last two) may be appropriate.
− For moving images with time-varying artefacts, a 10s sequence with two repetitions (voting during the
second) may be appropriate.
The Double-stimulus Continuous Quality-scale (DSCQS)
36
37. Grading Scale
− The method requires the assessment of two versions of each test image.
− One of each pair of test images is unimpaired while the other presentation might or might
not contain an impairment.
− The unimpaired image is included to serve as a reference, but the observers are not told
which the reference image is.
− In the series of tests, the position of the reference image is changed in pseudo-random
fashion.
− The observers are simply asked to assess the overall image quality of each presentation by
inserting a mark on a vertical scale.
− The vertical scales are printed in pairs to accommodate the double presentation of each
test image.
− The scales provide a continuous rating system to avoid quantizing errors, but they are divided
into five equal lengths which correspond to the normal ITU-R five-point quality scale.
The Double-stimulus Continuous Quality-scale (DSCQS)
37
38. Analysis of the Results
− The pairs of assessments (reference and test) for each test condition are converted from
measurements of length on the score sheet to normalized scores in the range 0 to 100.
− Then, the differences between the assessment of the reference and the test condition are
calculated.
− Experience has shown that the scores obtained for different test sequences are dependent
on the criticality of the test material used.
− A more complete understanding of codec performance can be obtained by presenting
results for different test sequences separately, rather than only as aggregated averages
across all the test sequences used in the assessment.
− If results for individual test sequences are arranged in a rank order of test sequence
criticality on an abscissa it is possible to present a crude graphical description of the image
content failure characteristic of the system under test.
The Double-stimulus Continuous Quality-scale (DSCQS)
38
39. Interpretation of the Results
− When using this DSCQS method, it could be hazardous, and even wrong, to derive conclusions about
the quality of the conditions under test by associating numerical DSCQS values with adjectives coming
from other tests protocols (e.g. imperceptible, perceptible but not annoying, ... coming from the DSIS
method).
− It is noted that results obtained from the DSCQS method should not be treated as absolute scores but as
differences of scores between a reference condition and a test condition.
− Thus, it is erroneous to associate the scores with a single quality description term even with those which
come from the DSCQS protocol itself (e.g. excellent, good, fair, ...).
− In any test procedure it is important to decide acceptability criteria before the assessment is
commenced.
− This is especially important when using the DSCQS method because of the tendency for inexperienced
users to misunderstand the meaning of the quality scale values produced by the method.
The Double-stimulus Continuous Quality-scale (DSCQS)
39
40. − Each assessment consists of a pair of video sequences A then B followed by the same sequences A and B.
− The order of reference and coded video sequences are shuffled and is unknown to viewers.
− Viewers are asked to rate the quality of both video sequences.
− Each video sequence has typically 8 to 10 seconds duration.
− The voting pattern of DSCQS is illustrated in figure below.
Double Stimulus Continuos Quality Scale (DSCQS), Summary
40
− Viewers rate the quality on continuous scale which is divided into five equal
intervals:
– Excellent for (100-80),
– Good for (79-60),
– Fair for (59-40),
– Poor for (39-20)
– Bad for (19-0)
− The final score is rated from 1(Bad) to 5 (Excellent)
41. − For distribution applications, quality specifications can be expressed in terms of the subjective judgement
of observers.
− Such codecs can in theory therefore be assessed subjectively against these specifications.
− The quality of a codec designed for contribution applications however, could not in theory be specified in
terms of subjective performance parameters because its output is destined not for immediate viewing, but
for studio post-processing, storing and/or coding for further transmission.
− Because of the difficulty of defining this performance for a variety of post-processing operations, the
approach preferred has been to specify the performance of a chain of equipment, including a post-
processing function, which is thought to be representative of a practical contribution application.
Subjective Assessment for Distribution or Contribution Codec
41
42. − This chain might typically consist of a codec, followed by a studio post-processing function (or another
codec in the case of basic contribution quality assessment), followed by yet another codec before the
signal reaches the observer.
− Adoption of this strategy for the specification of codecs for contribution applications means that the
measurement procedures given in this Recommendation can also be used to assess them.
− The most completely reliable method of evaluating the ranking order of high-quality codecs is to assess
all the candidate systems at the same time under identical conditions.
− Tests made independently, where fine differences of quality are involved, should be used for guidance
rather than as indisputable evidence of superiority.
Subjective Assessment for Distribution or Contribution Codec
42
43. − Where a codec is being assessed for distribution applications, this quality refers to pictures decoded after
a single pass through a codec pair.
− For contribution codecs, basic quality may be assessed after several codecs in series, in order to simulate
a typical contribution application.
− Where the range of quality to be assessed is small, as will normally be the case for television codecs, the
testing methodology to be used is variant II of the double-stimulus continuous quality-scale.
− The original source sequence will be used as the reference condition. Further consideration is being given
to the duration of presentation sequences.
− In the recent tests on codecs for 4:2:2 component video, it was considered advantageous to modify the
presentation from that given in this Recommendation.
− Composite pictures were used as an additional reference to provide a lower quality level against which to
judge the codec performance.
Evaluations of Basic Picture Quality
43
44. − It is recommended that at least six picture sequences be used in the assessment, plus an additional one
to be used for training purposes prior to the start of the trial.
− The sequences should range between moderately critical and critical in the context of the bit-rate
reduction application being considered.
− Suppose the importance is stressed of testing digital codecs with picture sequences which are critical in
the context of television bit-rate reduction.
• It is therefore reasonable to ask:
• How critical a particular image sequence is for a particular bit-rate reduction task.
or
• Whether one sequence is more critical than another.
Evaluations of Basic Picture Quality
44
45. − A simple but not especially helpful answer is that “criticality” means very different things to different
codecs.
− For example:
• To an intrafield codec a still picture containing much detail could well be critical.
• To an interframe codec a still picture containing much detail would present no difficulty at all.
Critical Sequences to All Classes of Codec
− Moving Texture + Complex motion → These types of sequences are most useful to generate or identify.
• Complex motion may take the form of movements which are predictable to an observer but not to coding
algorithms, such as tortuous periodic motion.
Evaluations of Basic Picture Quality
45
46. Fixed Word-length Intra-field Codecs
− While it is possible and valid to assess these codecs on still images, the use of moving sequences is
recommended since coding noise processes are easier to observe and this is more realistic of television
applications.
− If still images are used in computer simulations of codecs, processing should be performed over the entire
assessment sequence in order to preserve temporal aspects of any source noise, for example.
− The scenes chosen should contain as many as possible of the following details:
• static and moving textured areas (some with coloured texture)
• static and moving objects with sharp high contrast edges at various orientation (some with colour)
• static plain mid-grey areas.
− At least one sequence in the ensemble should exhibit just perceptible source noise
− At least one sequence should be synthetic (i.e. computer generated) so that it is free from camera
imperfections such as scanning aperture and lag.
Evaluations of Basic Picture Quality
46
47. Fixed Word-length Interframe Codecs
− The test scenes chosen should all contain movement and as many as possible of the following details:
• moving textured areas (some coloured)
• objects with sharp, high contrast edges moving in a direction perpendicular to these edges and at
various orientations (some coloured)
− At least one sequence in the ensemble should exhibit just perceptible source noise
− At least one sequence should be synthetic.
Evaluations of Basic Picture Quality
47
48. Variable Word-length Intra-field Codecs
− It is recommended that these codecs be tested with moving image sequence material for the same
reasons as the fixed word-length codecs.
− It should be noted that by virtue of its variable word-length coding and associated buffer store, these
codecs can dynamically distribute coding bit-capacity throughout the image.
− Thus, for example, if half of a picture consists of a featureless sky which does not require many bits to
code, capacity is saved for the other parts of the picture which can therefore be reproduced with high
quality even if they are critical.
− The important conclusion from this is that if a picture sequence is to be critical for such a codec, the
content of every part of the screen should be detailed.
Evaluations of Basic Picture Quality
48
49. Variable Word-length Intra-field Codecs
− So if a picture sequence is to be critical for such a codec, the content of every part of the screen should
be detailed.
− It should be filled with:
• moving and static texture
• as much colour variation as possible
• objects with sharp, high contrast edges
− At least one sequence in the text ensemble should exhibit just perceptible source noise
− At least one sequence should be synthetic.
Evaluations of Basic Picture Quality
49
50. Variable Word-length Interframe Codecs
− This is the most sophisticated class of codec and the kind which requires the most demanding material to
stress it.
− Not only should every part of the scene be filled with detail as in the intra-field variable word-length case,
but this detail should also exhibit motion.
− Furthermore, since many codecs employ motion compensation methods, the motion throughout the
sequence should be complex.
Evaluations of Basic Picture Quality
50
51. Variable Word-length Interframe Codecs
− Examples of complex motion are:
• scenes employing simultaneous zooming and panning of a camera
• a scene which has as a background a textured or detailed curtain blowing in the wind
• a scene containing objects which are rotating in the three-dimensional world
• scenes containing detailed objects which accelerate across the screen.
− All scenes should contain substantial motion of objects with different velocities, textures and high contrast
edges as well as a varied colour content.
− At least one sequence in the test ensemble should exhibit just perceptible source noise
− At least one sequence should have complex computer-generated camera motion from a natural still
picture (so that it is free from noise and camera lag)
− At least one sequence should be entirely computer generated.
Evaluations of Basic Picture Quality
51
52. − This assessment is intended to permit judgement to be made on the suitability of a codec for contribution
applications with respect to a particular post-process e.g. colour matte, slow motion, electronic zoom.
− The minimum arrangement of equipment for such an assessment is a single pass through the codec under
test, followed by the post-process of interest, followed by the viewer.
− It may, however, be more representative of a contribution application to employ further codecs after the
post-process.
− The test methodology to be used is variant II of the double-stimulus continuous quality-scale method (use
a number of observers simultaneously).
− Here however the reference condition will be the source subjected to the same post-processing as the
decoded pictures.
− If inclusion of a lower quality reference is considered to be advantageous then it too should be subjected
to the same post-process.
Evaluations of Picture Quality after Downstream Processing
52
53. − Test sequences required for post-processing assessments are subject to exactly the same criticality criteria
as sequences for other digital applications.
− Because of the practical constraints of possibly having to assess a codec with several post-processes, the
number of test picture sequences used may be a minimum of three with an additional one available for
demonstration purposes.
− The nature of the sequences will be dependent upon the post-processing task being studied but should
range between moderately critical and critical in the context of television bit-rate reduction and for the
process under consideration.
− For slow motion assessment a display rate of 1/10th of the source rate may be suitable.
Evaluations of Picture Quality after Downstream Processing
53
54. − In subjective assessments of impairments in codec images, due to imperfections in the transmission or
emission channel, a minimum of five, but preferably more, bit-error ratios or selected
transmission/emission conditions should be chosen, approximately logarithmically spaced and
adequately sampling the range which gives rise to codec impairments from “imperceptible” to “very
annoying”.
− It is possible that codec assessments could be required at transmission bit error ratios which result in visible
transients so infrequent that they may not be expected to occur during a 10s test sequence period.
− The presentation timing suggested here is clearly not suitable for such tests.
Evaluations of Failure Characteristics
54
55. − If recordings of a codec output under fairly low bit error ratio conditions (resulting in a small number of
visible transients within a 10s period) are to be made for later editing into subjective assessment
presentations, care should be taken to ensure that the recording used is typical of the codec output
viewed over a longer time-span.
− Because of the need to explore codec performance over a range of transmission bit error ratios, practical
constraints suggest that three test image sequences with an additional demonstration sequence will
probably be adequate.
− Sequences should be of the order of 10s in duration but it should be noted that test viewers may prefer a
duration of 15-30 s.
− It should range between moderately critical and critical in the context of television bit-rate reduction.
− As the tests will span the full range of impairment, the double-stimulus impairment scale (DSIS) method is
appropriate and should be used.
Evaluations of Failure Characteristics
55
57. Display selection and set-up
− The display used should be a flat panel display featuring performances typical of professional applications
(e.g. broadcasting studios or vans)
− The display diagonal dimension may vary from 22’ (minimum) to 40’ (suggested), but it may extend to 50’
or higher, when image systems with a resolution of HDTV or higher are assessed.
− It is allowed to use a reduced portion of the active viewing area of a display; in this case the area around
the active part of the display should be set to mid-grey. In this condition of use it should not be allowed to
set the monitor to a resolution different from its native one.
− The display should allow a proper set-up and calibration for luminance and colour, using a professional
light-meter instrument. The calibration of the display should comply with the parameters specified in the
relevant Recommendation for the test being undertaken.
57
Subjective Assessment by Expert Viewing
58. Viewing distance
− The viewing distance at which the experts are seated should be chosen according to the resolution of the
screen, and to the height of the active part of the screen, according to the design viewing distance as
described in Recommendation ITU-R BT.2022 or shorter viewing distance, according to the requirements in
terms of critical viewing conditions.
Viewing conditions
− An expert viewing protocol (EVP) experiment should not necessarily be run in a test laboratory, but it is
important that the testing location is protected from audible and or visible disturbances (e.g. a quiet
office or meeting room may be used as well).
− Any direct or reflected source of light falling on the screen should be eliminated; other ambient light
should be low, maintained to the minimum level that can allow filling scoring sheets (if used).
− The number of experts seated in front of the monitor, may vary according to the screen size, in order to
guarantee the same image rendering and stimulus presentation for all the viewers.
58
Subjective Assessment by Expert Viewing
59. Viewers
− The viewers participating to an EVP experiment should be expert in the domain of study.
− Viewers should not necessarily be screened for visual acuity or colour blindness, since they should be
chosen among qualified persons.
− The minimum number of different viewers should be nine.
− To reach the minimum number of viewers, the same experiment may be conducted at the same location
repeating the test, or in more than one location. The scores from different locations participating to an
expert viewing session may be statistically processed together.
59
Subjective Assessment by Expert Viewing
60. The basic test cell
− A basic test cell (BTC) for each couple of coding conditions to be assessed.
− The source reference sequences (SRC) and the processed video sequences (PVSs) clips to consider in a
BTC should always be related to the same video sequence, in order that the experts may be able to
identify any improvement in visual quality provided by the compression algorithms under test.
60
Subjective Assessment by Expert Viewing
The BTC should be organised as follows:
− 0.5 seconds with the screen set to a mid-grey (mean value in the luminance scale);
− 10 seconds presentation of the reference uncompressed video clip;
− 0.5 seconds showing the message “A” (first video to assess) on a mid-grey background;
− 10 seconds presentation of an impaired version of the video clip;
− 0.5 seconds showing the message “B” (second video to assess) on a mid-grey background;
− 10 seconds presentation of an impaired version of the video clip;
− 5 seconds showing a message that asks the viewers to express their opinion.
The message “Vote” should be followed by a number that helps to get synchronised on the scoring sheet.
61. Scoring sheet and rating scale
− The presentation of the video clips should be arranged in such a way that the unimpaired reference (SRC)
is shown at first, followed by two impaired video sequences (PVS).
− The order of presentation of the PVS should be randomly changed for each BTC and the viewers should
not know the order of presentation.
61
Subjective Assessment by Expert Viewing
62. Test design and session creation
− A random order by the test designer, in such a way that the same video clip is not shown two consecutive
times as well as the same impaired clip.
− Any viewing session should begin with a “stabilization phase” including the “best”, the “worst” and two
“mid quality” BTC among those included in each test session.
− This will allow the viewers to have an immediate impression of the quality range, already at the beginning
the test session.
− If the viewing session is longer than 20 minutes, the test designer should split it into two (or more) separate
viewing sessions, each of them not exceeding 20 minutes. In this case, the “stabilization phase” should be
provided before each viewing session.
62
Subjective Assessment by Expert Viewing
63. Training
− Even if this procedure is foreseen for use with the participation of experts, a short (5-6 BTC) training viewing
session should preferably be organised prior to each experiment.
− The video material used in the training session may be the same that will be used during the actual
sessions, but the order of presentation should be different.
− The viewers should be trained on the use of the 11-grade scale by asking them to carefully look at the
video clips shown immediately after the message “A” and “B” on the screen, and check whether they
can see any difference to the video clip shown as first (the SRC).
63
Subjective Assessment by Expert Viewing
64. Data collection and processing
− The scores should be collected at the end of each session and logged on an electronic spreadsheet to
compute the MEAN values.
− A “post screening” of the viewers should desirably be performed, using a linear Pearson’s correlation.
− The “correlation” function should be applied considering all the scores of each subject in relation to the
mean opinion scores (MOS); a threshold may be set to define each viewer as “acceptable” or “rejected”
(Recommendation ITU-T P.913 suggests the use of a “reject” threshold value equal to 0.75).
Terms of use of the expert viewing protocol results
− The expert viewing protocol (EVP) may be used when time and resources do not allow running a formal
subjective assessment experiment.
− EVP requires less time than a formal subjective assessment and may be executed in an “informal”
environment. The only mandatory conditions are related to the ambient illumination and to the viewing
conditions (display, angle of observation and viewing distance) as described in the above paragraphs.
64
Subjective Assessment by Expert Viewing
65. Limitations of use of the EVP results
− Even if the EVP is demonstrating to be able to provide acceptable results with only nine viewers, the MOS
provided by an EVP experiment cannot be considered as a replacement of the results obtainable with a
formal subjective assessment experiment.
− The MOS data obtained using EVP might be used to make a preliminary ranking of the video processing
schemes under evaluation.
− Where retained convenient or necessary, an EVP experiment can be run in parallel in more locations,
assuming the viewing conditions, viewing distance and the test design are identical.
− If the number of expert viewers involved in the same EVP experiment, also if running the experiment in
different locations, is equal or higher than 15, the raw subjective data might be processed to obtain MOS,
standard deviation and confidence interval data, that may help to perform a more accurate ranking of
the cases under test. In this last case more accurate inferential statistical analysis may be performed, e.g.
T-Student test.
65
Subjective Assessment by Expert Viewing
66. 66
EVP in the Presence of a Large Number of Expert Assessors
67. 67
EVP in the Presence of a Large Number of Expert Assessors
68. − Observing the results it can be noted that:
• 15 and the 18 viewers graphs show an homogeneous slope from high quality to low quality MOS
values.
• 9 and the 12 viewers graphs show some “inversions” of ranking compared to the 18 viewers graph,
even if the variations of scores are rather limited in their extension.
− In conclusion, the EVP experiments here described show a very good performance of EVP protocol,
confirming that the use of the EVP protocol, even if it cannot be considered a full replacement of a formal
subjective experiment, might be considered an evaluation procedure stable and providing results very
close to those obtained when many more viewers are available and a formal subjective assessment is
done.
68
EVP in the Presence of a Large Number of Expert Assessors
70. Distribution Applications
− The quality specifications can be expressed in terms of the subjective judgement of observers.
− Such codecs can in theory therefore be assessed subjectively against these specifications.
Contribution Applications
− The quality could not in theory be specified in terms of subjective performance parameters because its
output is destined not for immediate viewing, but for
• studio post-processing
• storing
• and/or coding for further transmission.
70
Distribution Applications and Contribution Application Codecs Test
71. Contribution Applications
− Because of the difficulty of defining this performance for a variety of post-processing operations, the
approach preferred has been to specify the performance of a chain of equipment, including a post-
processing function, which is thought to be representative of a practical contribution application.
− This chain might typically consist of a codec, followed by a studio post-processing function (or another
codec in the case of basic contribution quality assessment), followed by yet another codec before the
signal reaches the observer.
71
Distribution Applications and Contribution Application Codecs Test
Encoder
Studio Post-processing
Function
EncoderDecoder
First
Generation
Second
Generation
Input
Video
Decoder
Representative of a practical
contribution application
72. Contribution Applications
− The most completely reliable method of evaluating the ranking order of high-quality codecs is to assess
all the candidate systems at the same time under identical conditions.
− Tests made independently, where fine differences of quality are involved, should be used for guidance
rather than as indisputable evidence of superiority.
− A useful subjective measure may be impairment determined as a function of the bit error ratio which
occurs in the contribution link between coder and decoder.
− At present there is insufficient experimental knowledge of true transmission error statistics to recommend
parameters for a model which accounts for error clustering or bursts. Until this information becomes
available Poisson-distributed errors may be used.
72
Distribution Applications and Contribution Application Codecs Test
73. − One article might claim that codec A is 15% better than codec B, while the next one might assert that
codec B is 10% better than codec A. Why?
− Because the testing methodology and content play a crucial role in the evaluation of video codecs.
The several factors that impact the assessment of video codecs:
• Encoder Implementation
• Encoder Settings
• Methodology
• Content
• Metrics
Factors in Codec Comparisons
73
74. − Video coding standards are instantiated in software or hardware with goals as varied as research,
broadcasting, or streaming.
− A ‘reference encoder’ is a software implementation used during the video standardization process and for
research, and as a reference by implementers.
− Generally, this is the first implementation of a standard and not used for production. Afterward, production
encoders developed by the open-source community or commercial entities come along.
− These are practical implementations deployed by most companies for their encoding needs and are
subject to stricter speed and resource constraints. Therefore, the performance of reference and
production encoders might be substantially different.
− Besides, the standard profile and specific version influence the observed performance, especially for a
new standard with still immature implementations.
− Netflix deploys production encoders tuned to achieve the highest subjective quality for streaming.
Encoder Implementation
74
75. − Encoding parameters such as the number of coding passes, parallelization tools, rate-control, visual
tuning, and others introduce a high degree of variability in the results.
− The selection of these encoding settings is mainly application-driven.
− Standardization bodies tend to use test conditions that let them compare one tool to another, often
maximizing a particular objective metric and reducing variability over different experiments.
− For example, rate-control and visual tunings are generally disabled, to focus on the effectiveness of core
coding tools.
− Netflix encoding recipes focus on achieving the best quality, enabling the available encoder tools that
boost visual appearance, and thus, giving less weight to indicators like speed or encoder footprint that
are crucial in other applications.
Encoding Settings
75
76. − Testing methodology in codec standardization establishes well-defined “common test conditions” to
assess new coding tools and to allow for reproducibility of the experiments.
− Common test conditions consist of a relatively small set of test sequences (single shots of 1 to 10 seconds)
that are encoded only at the input resolution with a fixed set of quality parameters.
− Quality (PSNR) and bitrate are gathered for each of these quality points and used to compute the
average bitrate savings, the so-called BD-rate (Bjontegaard Delta (BD) rate savings).
Methodology
76
77. − While the methodology in standards has been suitable for its intended purpose, other considerations
come into play in the adaptive streaming world.
− Notably, the option to offer multiple representations of the same video at different bitrates and resolutions
to match network bandwidth and client processing and display capabilities.
− The content, encoding, and display resolutions are not necessarily tied together. Removing this constraint
implies that quality can be optimized by encoding at different resolutions.
Methodology
77
78. − Per-resolution rate-quality curves cross, so there is a range of rates for which each encoding resolution
gives the best quality.
− The ‘convex hull’ is derived by selecting the optimal curve at each rate for the entire range.
− Then, the BD-rate difference is computed on the convex hulls, instead of using the single resolution curves.
− The flexibility in delivering bitstreams on the convex hull as opposed to just the single resolution ones leads
to remarkable quality improvements
Methodology
78
79. − The dynamic optimizer (DO) methodology generalizes this concept to sequences with multiple shots.
− DO operates on the convex hull of all the shots in a video to jointly optimize the overall rate-distortion by
finding the optimal compression path for an encode across qualities, resolutions, and shots.
− It showed a 25% in BD-rate savings for multiple-shot videos.
− Three characteristics that make DO particularly wellsuited for adaptive streaming and codec comparisons
are:
• It is codec agnostic since it can be applied in the same way to any encoder.
• It can use any metric to guide its optimization process.
• It eliminates the need for high-level rate control across shots in the encoder. Lower-level rate control, like
adaptive quantization within a frame, is still useful, because DO does not operate below the shot level.
− DO, as originally presented, is a non-real-time, computationally expensive algorithm.
− However, because of the exhaustive search, DO could be seen as the upper-bound performance for
high-level rate control algorithms.
Methodology
79
80. − For a fair comparison, the testing content should be balanced, covering a variety of distinct types of
video (natural vs. animation, slow vs. high motion, etc.) or reflect the kind of content for the application at
hand.
− The testing content should not have been used during the development of the codec.
− Netflix has produced and made public long video sequences with multiple shots, such as ‘El Fuente’ or
‘Chimera’, to extend the available videos for R&D and mitigate the problem of conflating training and
test content.
− Internally, we extensively evaluate algorithms using full titles from our catalog.
Content
80
81. − Traditionally, PSNR has been the metric of choice given its simplicity, and it
reasonably matches subjective opinion scores.
− Other metrics, such as VIF or SSIM, better correlate with the subjective
scores.
− Metrics have commonly been computed at the encoding resolution.
− Netflix highly relies on VMAF (Video Multi-Method Assessment Fusion)
throughout its video pipeline.
− VMAF is a perceptual video quality metric that models the human visual
system.
− It correlates better than PSNR with subjective opinion over a wide quality
range and content.
− VMAF enables reliable codec comparisons across the broad range of
bitrates and resolutions occurring in adaptive streaming.
Metrics
81
Approximate correspondence
between VMAF values and
subjective opinion
82. Two relevant aspects when employing metrics are the resolution at which they are computed and the
temporal averaging:
Scaled Metric:
− VMAF is not computed at the encoding resolution, but at the display resolution, which better emulates our
members viewing experience. This is not unique to VMAF, as PSNR and other metrics can be applied at any
desired resolution by appropriately scaling the video.
Temporal Average:
− Metrics are calculated on a per-frame basis.
− Normally, the arithmetic mean has been the method of choice to obtain the temporal average across the entire
sequence.
− We employ the harmonic mean, which gives more weight to outliers than the arithmetic mean.
− The rationale for using the harmonic mean is that if there are few frames that look really bad in a shot of a show
you are watching, then your experience is not that great, no matter how good the quality of the rest of the shot
is. The acronym for the harmonic VMAF is HVMAF.
Metrics
82
83. − Putting in practice the factors mentioned above, we show the results drawn from the two distinct codec
comparison approaches, the traditional and the adaptive streaming one.
− Three commonly used video coding standards are tested
• H.264/AVC by ITU.T and ISO/MPEG
• H.265/HEVC by ITU.T and ISO/MPEG
• VP9 by Google.
− For each standard, we use the reference encoder and a production encoder.
Codec Comparison Results by Netflix
83
84. Results with the traditional approach
− The traditional approach uses fixed QP (quality) encoding for a set of short sequences.
− Encoder settings are listed in the table below.
Codec Comparison Results by Netflix
84
85. Codec Comparison Results by Netflix
85
Results with the traditional approach
− Methodology: Five fixed quality encodings per sequence at the content resolution are generated.
− Content: 14 standard sequences from the MPEG Common Test Conditions set (mostly from JVET) and 14
from the Alliance for Open Media (AOM) set. All sequences are 1080p. These are short clips: about10
seconds for the MPEG set and 1 second for the AOM set. Mostly, they are single shot sequences.
− Metrics: BD-rate savings are computed using the Classic PSNR for the luma component.
• BD-rates are given in percentage with respect to x264.
• Positive numbers indicate an average increase of bitrate, while negative numbers indicate bitrate
reduction.
− Interestingly, using the MPEG set doesn’t seem to benefit HEVC encoders, or using the AOM set help VP9
encoders.
86. Codec Comparison Results by Netflix
86
Results from an adaptive streaming perspective
− It builds on top of the traditional approach, modifying some aspects in each of the factors:
− Encoder settings: Settings are changed to incorporate perceptual tunings. The rest of the settings remain
as previously defined.
87. Codec Comparison Results by Netflix
87
Results from an adaptive streaming perspective
− Methodology:
• Encoding is done at 10 different resolutions for each shot, from 1920x1080 down to 256x144.
• DO performs the overall encoding optimization using HVMAF (harmonic VMAF).
− Content:
• A set of 8 Netflix full titles (like Orange is the New Black, House of Cards or Bojack) is added to the other two
test sets.
• Netflix titles are 1080p, 30fps, 8 bits/component. They contain a wide variety of content in about 8 hours of
video material.
− Metrics:
• HVMAF is employed to assess these perceptually-tuned encodings.
• The metric is computed over the relevant quality range of the convex hull.
• HVMAF is computed after scaling the encodes to the display resolution (assumed to be 1080p), which also
matches the resolution of the source content.
88. Codec Comparison Results by Netflix
88
Results from an adaptive streaming perspective
− Additionally, we split results into two ranges to visualize performance at different qualities.
• The low range refers to HVMAF between 30 and 63
• The high range refers to HVMAF between 63 and 96, which correlates with high subjective quality.
− The highlighted rows in the HVMAF BD-rates table are the most relevant operation points for Netflix.
89. − Encoder, encoding settings, methodology, testing content, and metrics should be thoroughly described in
any codec comparison since they greatly influence results.
− A different selection of the testing conditions leads to different conclusions.
• For example, EVE-VP9 is about 1% worse than x265 in terms of PSNR for the traditional approach, but
about 12% better for the HVMAF high range case.
− Given the vast amount of video compressed and delivered by services like Netflix and that traditional and
adaptive streaming approaches do not necessarily converge to the same outcome, it would be
beneficial if the video coding community considered the adaptive streaming perspective in the
comparisons.
− For example, it is relatively easy to compute metrics on the convex hull or to add the HVMAF numbers to
the reported metrics.
Codec Comparison Results by Netflix
89
92. − It enables accurate evaluation and calibration of a wide range of digital video systems (to evaluate more than
20 key parameters of video quality from a single screen).
− The pattern is available as an uncompressed or compressed video sequence. In streaming form—as
HEVC/H.265, H.264/ MPEG-4 AVC and MPEG-2 video—it is suitable for testing both file-based and streaming
systems.
• Quantifies compression fidelity
• Identifies color matrix mismatch
• Determines bit depth and chroma subsampling
• Reveals skipped frames
• Quantifies lip sync errors
• Provides 18 additional tests
− HDR+WCG Version
• Reveals tone mapping and clipping
• Shows color gamut mismatch
• Indicates gamut mapping
Visualizer™ SDR/HDR+WCG Digital Video Test Pattern
92
93. − At the left, H and V linear frequency sweeps from zero to the Nyquist limit.
− At right, after conversion from 1920x1080 to 1280x720 and back.
− Note the aliasing at the top that should have been filtered away to gray.
− Note the strong Moiré pattern 2/3 the way up. 2/3 = 720/1080.
− Use pattern’s 12-unit grid scale to measure frequency ratios (8/12 = 720/1080).
Frequency Response
93
94. − Right: Some regions have missing detail.
− At each frequency, note the highest bit depth with visible signal.
− Left: Uncompressed shows visible signal from bottom to top in every frequency burst. (Contrast
exaggerated.)
Compression Fidelity
94
95. − Check skin tones, detail in hair and jewelry; check for highlight clipping, black clipping.
− The image is colormanaged for gamma and RGB primaries.
Reference Image
95
96. − Calibrated linear light stairstep from 100% (10,000 nits) down 6 decades in 24 steps,
with +/- 1/3 stop deltas. Check monitor tone mapping and clipping, and darkest
visible light level.
Jacob’s Ladder
96
97. − See User Guide Appendix for RGB values for each chip.
− Note six dim blocks in the fifth row. They are 1/200, 1/400,
− 1/800, 1/1600, 1/3200, 1/6400 of the linear light range.
ST-303M
97
98. − Right half: luma only.
− Left half: chroma only.
− Look for jumps or stutter indicating dropped or repeated frames.
− 1/4 way up, speed is 1 line per frame or field: look for jaggies there.
− Look for differences between luma and chroma resolution and
motion smoothness.
Lava Lamp
98
101. − 1002, 940, 932, 914 (107%, 100, 99, 97).
− Headroom is clipped if you can’t see the difference between 107 and 100.
− You should see a difference between 100 and 99 (adjust monitor contrast down).
White Pluge
101
Correct Blown out
102. Chroma Upsampling Back to 4:4:4
102
4:4:4 4:2:2 to 4:4:4
4:2:0p to 4:4:4 4:2:0i to 4:4:4
103. − ± 4, 2, 1%.
− Left image: monitor brightness too high.
− Adjust down until 1st and 2nd columns merge (middle image).
− All 3rd column chips are different than 2nd column.
− Right: monitor brightness too high, but column 1 (super black) has been clipped off upstream or in the
monitor.
Pluge
103
104. − Left: 10-bit shows no contour lines.
− Dithered 8-bit may look almost as smooth.
− Right: 8-bit video has rotating contour lines.
− Compression artifacts may also appear. (Contrast exaggerated.)
Bit Depth
104
105. − 12 color points around each gamut, with small chips desaturated by 5 delta-E.
− Reveals gamut mapping.
Gamut Bars
105
106. − Upper field means the odd line comes first. Even field means the even line comes first.
− The field dominance refers to the choice of which field of an interlaced video signal is chosen as the point
at which video edits or switches occur.
− NTSC and PAL are both lower-field (or even-field) dominant. This means they record all the even-
numbered lines of your image first, then the odd-numbered lines a fraction of a second later.
− Most HD formats are upper-field dominant, which means they record the odd-numbered lines first.
Field Dominance
106
Correct
Reversed
107. − Check skin tones, detail in hair and jewelry; check for highlight clipping, black clipping.
− The image is colormanaged for gamma and RGB primaries.
Lip Sync
107
Correct: Tick is heard when center mark flashes.
Incorrect: Audio is 3 frames/fields early
108. − Demonstrates incorrect processing of Chroma in interlaced 4:2:0 systems.
Chroma Motion Error
108
Correct
Incorrect
109. − When center-cropped to the indicated aspect ratio, the associated dashed lines will be the edge pixels
of the new image.
Border Marquee & Crop Lines
109
Correct: Top white dashed line is showing Top line is missing