In October 2017, ISO/IEC JCT1 SC29/WG11 MPEG and ITU-T SG16/Q6 VCEG have jointly published a Call for Proposals on Video Compression with Capability beyond HEVC and its current extensions. It is targeting at a new generation of video compression technology that has substantially higher compression capability than the existing HEVC standard. The responses to the call are evaluated in April 2018, forming the kick-off for a new standardization activity in the Joint Video Experts Team (JVET) of VCEG and MPEG, with a target of finalization by the end of the year 2020. Three categories of video are addressed: Standard dynamic range video (SDR), high dynamic range video (HDR), and 360° video. While SDR and HDR cover variants of conventional video to be displayed e.g. on a suitable TV screen at very high resolution (UHD), the 360° category targets at videos capturing a full-degree surround view of the scene. This enables an immersive video experience with the possibility to look around in the rendered scene, e.g. when viewed using a head-mounted display. This application triggers various technical challenges which need to be addressed in terms of compression, encoding, transport, and rendering. The talk summarizes the current state of the complete standardization project. Focussing on the SDR and 360° video categories, it highlights the development of selected coding tools compared to the state of the art. Representative examples of the new technological challenges as well as corresponding proposed solutions are presented.
This document provides an overview and comparison of the H.264 and HEVC video coding standards. It describes the key features and innovations that allow each standard to compress video more efficiently than previous standards. H.264 introduced features like adaptive block sizes, multi-frame prediction, quarter-pixel motion compensation and loop filtering that improved compression performance over prior standards. HEVC aims to further increase compression efficiency through innovations such as larger coding tree blocks, additional intra-prediction modes, and improved entropy coding. The document analyzes these standards to understand how their new coding tools enable significantly higher compression ratios and support for new applications like higher resolution video.
Video Compression, Part 3-Section 1, Some Standard Video CodecsDr. Mohieddin Moradi
- ISO/IEC JTC 1/SC 29 and ITU-T are the main organizations that develop video coding standards through working groups like MPEG and VCEG.
- Early standards include H.261 for video telephony and conferencing, and MPEG-1 for DVD quality video.
- Later standards like H.264/AVC, HEVC, and future VVC provide increasingly higher compression through use of block transforms, motion compensation, and entropy coding in a hybrid video codec framework.
- Key organizations periodically collaborate through joint teams like JVT and JCT-VC to develop standards like AVC and HEVC.
Introduction to H.264 Advanced Video CompressionIain Richardson
The document discusses H.264 advanced video compression. It provides an agenda that covers what H.264 is, how it works through prediction, transform and quantization techniques, its syntax, examples, and going deeper into its implementation. H.264 is widely used for video compression in broadcast digital TV, DVDs/Blu-Rays, IPTV, web video and mobile video. It works by predicting pixels from previous frames, applying transforms and quantization to remove redundant information, and using entropy coding techniques to further compress the data. The document provides resources to learn more about H.264 standards, implementations, and extensions.
Video coding is an essential component of video streaming, digital TV, video chat and many other technologies. This presentation, an invited lecture to the US Patent and Trade Mark Office, describes some of the key developments in the history of video coding.
Many of the components of present-day video codecs were originally developed before 1990. From 1990 onwards, developments in video coding were closely associated with industry standards such as MPEG-2, H.264 and H.265/HEVC.
The presentation covers:
- Basic concepts of video coding
- Fundamental inventions prior to 1990
- Industry standards from 1990 to 2014
- Video coding patents and patent pools.
The document discusses high dynamic range (HDR) video technology including:
- Different HDR formats such as SMPTE ST 2084 (PQ), ARIB STB-B67/ITU-R BT.2100 (HLG)
- Code value ranges for 10-bit and 12-bit RGB and color difference signals in narrow and full ranges
- Recommendations for using narrow versus full signal ranges for PQ and HLG
- Transcoding concepts when converting between PQ and HLG formats
- Considerations for including standard dynamic range (SDR) content in HDR programs
This document provides an overview and comparison of the H.264 and HEVC video coding standards. It describes the key features and innovations that allow each standard to compress video more efficiently than previous standards. H.264 introduced features like adaptive block sizes, multi-frame prediction, quarter-pixel motion compensation and loop filtering that improved compression performance over prior standards. HEVC aims to further increase compression efficiency through innovations such as larger coding tree blocks, additional intra-prediction modes, and improved entropy coding. The document analyzes these standards to understand how their new coding tools enable significantly higher compression ratios and support for new applications like higher resolution video.
Video Compression, Part 3-Section 1, Some Standard Video CodecsDr. Mohieddin Moradi
- ISO/IEC JTC 1/SC 29 and ITU-T are the main organizations that develop video coding standards through working groups like MPEG and VCEG.
- Early standards include H.261 for video telephony and conferencing, and MPEG-1 for DVD quality video.
- Later standards like H.264/AVC, HEVC, and future VVC provide increasingly higher compression through use of block transforms, motion compensation, and entropy coding in a hybrid video codec framework.
- Key organizations periodically collaborate through joint teams like JVT and JCT-VC to develop standards like AVC and HEVC.
Introduction to H.264 Advanced Video CompressionIain Richardson
The document discusses H.264 advanced video compression. It provides an agenda that covers what H.264 is, how it works through prediction, transform and quantization techniques, its syntax, examples, and going deeper into its implementation. H.264 is widely used for video compression in broadcast digital TV, DVDs/Blu-Rays, IPTV, web video and mobile video. It works by predicting pixels from previous frames, applying transforms and quantization to remove redundant information, and using entropy coding techniques to further compress the data. The document provides resources to learn more about H.264 standards, implementations, and extensions.
Video coding is an essential component of video streaming, digital TV, video chat and many other technologies. This presentation, an invited lecture to the US Patent and Trade Mark Office, describes some of the key developments in the history of video coding.
Many of the components of present-day video codecs were originally developed before 1990. From 1990 onwards, developments in video coding were closely associated with industry standards such as MPEG-2, H.264 and H.265/HEVC.
The presentation covers:
- Basic concepts of video coding
- Fundamental inventions prior to 1990
- Industry standards from 1990 to 2014
- Video coding patents and patent pools.
The document discusses high dynamic range (HDR) video technology including:
- Different HDR formats such as SMPTE ST 2084 (PQ), ARIB STB-B67/ITU-R BT.2100 (HLG)
- Code value ranges for 10-bit and 12-bit RGB and color difference signals in narrow and full ranges
- Recommendations for using narrow versus full signal ranges for PQ and HLG
- Transcoding concepts when converting between PQ and HLG formats
- Considerations for including standard dynamic range (SDR) content in HDR programs
This document provides information about quality control testing of audiovisual content. It discusses various quality control tests that can be performed, including tests for analogue frame synchronization errors, black bars, constant colour frames, flashing video, macroblocking, video deinterlacing artifacts, and digital tape dropouts. Examples are provided for how each test can be configured and what results might look like. The goal of the quality control tests is to help broadcasters optimize their automated quality control systems and cope with increasing amounts of digital content.
Dr. Mohieddin Moradi provides an outline on high dynamic range (HDR) technology. The 3-page document covers various topics related to HDR including different HDR technologies, tone mapping, color representation, and HDR standards. It discusses concepts such as scene-referred vs display-referred conversions, and direct mapping vs tone mapping when converting between HDR and SDR formats. The document also examines potential side effects when mixing different conversion techniques in a production workflow.
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. Some key points:
- HEVC was created as a new video compression standard to address the growing needs of higher resolution video content and more efficient compression compared to prior standards like H.264.
- It achieves 50% bitrate reduction over H.264 for the same visual quality or improved quality at the same bitrate.
- The standard uses a block-based coding structure with coding tree units and supports intra-frame and inter-frame coding with motion estimation/compensation.
- It introduces more intra-prediction modes and block sizes along with improved transforms, quantization, and entropy coding.
This document provides an overview of high dynamic range (HDR) technology and workflows for HDR video production and mastering. It discusses HDR standards like SMPTE ST 2084 and ARIB STB-B67, camera log curves, luminance levels, and tools for setting up HDR monitoring including waveform monitors. Specific topics covered include HDR graticules, setting luminance levels for highlights and grey points, and using zebra patterns and zoom modes to evaluate highlight levels in HDR images.
The document discusses video compression history and standards, including codecs such as H.261, H.262/MPEG-2, H.263, H.264/AVC, H.265/HEVC, and the roles of organizations like MPEG, VCEG, and ITU-T in developing video coding standards to ensure interoperability. It also covers video encoding and decoding principles, as well as common container formats and their applications in areas like broadcasting, streaming, and storage.
Versatile Video Coding: Compression Tools for UHD and 360° VideoMathias Wien
The document discusses the development of the Versatile Video Coding (VVC) standard. It describes how a call for proposals was issued to develop coding tools beyond HEVC. 46 proposals were submitted across standard dynamic range, high dynamic range, and 360-degree video categories. The proposals were evaluated through subjective testing and shown to provide over 40% bitrate reduction compared to HEVC and over 10% reduction compared to the Joint Exploration Model, with the best proposals demonstrating visual quality equal or better than HEVC at higher bitrates. Seven proposals were identified as significantly better than the Joint Exploration Model. This marked the starting point for developing the VVC standard based on the selected coding tools from the top-performing proposals
An Overview of High Efficiency Video Codec HEVC (H.265)Varun Ravi
The document provides an overview of the High Efficiency Video Coding (HEVC) H.265 standard. It discusses the need for improved video compression standards due to increasing video content and limited bandwidth. HEVC was developed to meet this need by providing around 50% better compression over its predecessor H.264 while still maintaining high video quality. The document describes the various techniques used in HEVC such as improved block partitioning, transform sizes, prediction modes, and entropy coding that help achieve its compression gains. Both hardware and software implementations of HEVC decoders and encoders are discussed.
MIPI DevCon 2021: Meeting the Needs of Next-Generation Displays with a High-P...MIPI Alliance
Presented by Alain Legault, Hardent Inc.; Joe Rodriguez, Rambus Inc.; and Justin Endo, Mixel, Inc.
Next-generation display applications have an insatiable appetite for bandwidth. Using a combination of VESA Display Stream Compression (DSC) and MIPI DSI-2℠ technology, designers can achieve display resolutions up to 8K without compromise to video quality, battery life or cost. This presentation discusses a fully integrated, off-the-shelf display IP subsystem solution, consisting of Mixel (MIPI C-PHY℠/D-PHY℠ combo), Rambus (MIPI DSI-2® controller) and Hardent (VESA DSC) IP, that can deliver this state-of-the-art performance in a power-efficient and compact footprint.
Video compression techniques exploit various types of redundancy in video signals to reduce the data required to represent them. Key techniques include intra-frame compression which uses spatial redundancy within frames via DCT, inter-frame compression which uses temporal redundancy between consecutive frames by encoding differences, and motion compensation which accounts for motion between frames. Popular video compression standards like MPEG use a combination of these techniques including I, P and B frames along with motion estimation to achieve much higher compression ratios than image compression alone.
This document provides definitions and explanations of various optical terminology related to light passing through a lens, including:
- Dispersion, refraction, diffraction, reflection, focal point, focal length, principal point, image circle, aperture ratio, numerical aperture, optical axis, and more. It discusses concepts such as entrance pupil, exit pupil, angular aperture, and how they relate to lens performance. The document also covers topics like vignetting, the cosine law, and flare. Overall, it serves as a comprehensive reference for understanding optical and photographic lens terminology.
Video Compression Standards - History & IntroductionChamp Yen
This document provides an overview of several video compression standards including MPEG-1/2, MPEG-4, H.264, and HEVC/H.265. It discusses the key concepts of video coding such as entropy coding, quantization, transformation, and intra- and inter-prediction. For each standard, it describes the main coding tools and improvements over previous standards, focusing on techniques for more efficient prediction and extraction of redundant spatial and temporal information while maintaining quality. The development of these standards has moved towards more fine-grained partitioning and new coding ideas and tools to reduce bitrates further.
Video Compression, Part 4 Section 1, Video Quality Assessment Dr. Mohieddin Moradi
This document provides an overview of video compression artifacts that can occur when video is compressed for streaming or storage. It discusses both spatial artifacts, such as blurring, blocking, ringing, and color bleeding, as well as temporal artifacts like flickering and mosquito noise. For each artifact, it describes the visual appearance and potential causes from factors like quantization during compression, motion compensation between frames, and chroma subsampling. The document aims to help understand how compression can degrade perceptual video quality and different types of artifacts that may be evaluated both objectively and subjectively.
HEVC/H.265 is a video compression standard that provides around 50% better compression over H.264/AVC for the same level of video quality. It was finalized in 2013 by the joint collaboration of MPEG and ITU-T. Key features of HEVC include support for higher resolutions like 4K and 8K, improved parallel processing abilities, increased coding efficiency through larger block sizes and an expanded set of prediction modes.
Machine learning approaches are being explored for video compression. Conservative approaches replace individual MPEG blocks with deep learning blocks, while disruptive end-to-end approaches aim to replace the entire MPEG chain. Optical flow networks can exploit temporal redundancy by estimating motion between frames. Fully neural network-based video compression models jointly learn motion estimation, motion compression, and residual compression in an end-to-end optimized framework. However, performance gains must be balanced against increased complexity, and neural network approaches are not yet mature enough to be included in video compression standards.
The document provides a history of the development of television technology from the late 1800s through the 1920s. Some key developments include:
- In 1873, experiments with selenium, which is light-sensitive and formed the basis for early televisions.
- In 1884, the Nipkow disk laid down many basic concepts like scanning and synchronization.
- In 1923, Vladimir Zworykin developed the Kinescope, which allowed television programs to be recorded on film.
- In 1924, John Logie Baird transmitted the first television image.
- In 1925, Vladimir Zworykin demonstrated 60-line television using a curved-line image structure typical of mechanical television at the time.
The document provides an introduction to video compression. It discusses key concepts such as lossy vs lossless compression, encoders, decoders, and codecs. It also covers techniques used in video compression like sampling, quantization, model-based transforms, the human visual system, color space transforms, block-based coding, and the discrete cosine transform. Video compression standards like MPEG compress video using techniques like motion estimation, motion compensation, and encoding frames individually.
The latest video compression standard, H.264 (also known as MPEG-4 Part 10/AVC for Advanced Video
Coding), is expected to become the video standard of choice in the coming years.
H.264 is an open, licensed standard that supports the most efficient video compression techniques available
today. Without compromising image quality, an H.264 encoder can reduce the size of a digital video file by
more than 80% compared with the Motion JPEG format and as much as 50% more than with the MPEG-4
Part 2 standard. This means that much less network bandwidth and storage space are required for a video
file. Or seen another way, much higher video quality can be achieved for a given bit rate.
Video Compression, Part 3-Section 2, Some Standard Video CodecsDr. Mohieddin Moradi
This document discusses MPEG-2 Transport Streams and Packetized Elementary Streams. It describes how MPEG-2 Transport Streams use fixed length 188 byte packets containing compressed video, audio or data from one or more programs identified by Packet IDs. These packets can contain Packetized Elementary Stream packets which contain compressed elementary streams with timestamps for synchronization. The document also discusses how Transport Streams allow for synchronous multiplexing of multiple programs from independent time bases into a single stream.
The document discusses video compression basics and MPEG-2 video compression. It explains that video frames contain redundant spatial and temporal data that can be compressed. MPEG-2 uses three frame types (I, P, B frames) and compresses frames using intra-frame and inter-frame encoding techniques like DCT, quantization, and entropy encoding to remove redundancy. The encoding process transforms raw video frames to compressed bitstreams for efficient storage and transmission.
This document provides information about quality control testing of audiovisual content. It discusses various quality control tests that can be performed, including tests for analogue frame synchronization errors, black bars, constant colour frames, flashing video, macroblocking, video deinterlacing artifacts, and digital tape dropouts. Examples are provided for how each test can be configured and what results might look like. The goal of the quality control tests is to help broadcasters optimize their automated quality control systems and cope with increasing amounts of digital content.
Dr. Mohieddin Moradi provides an outline on high dynamic range (HDR) technology. The 3-page document covers various topics related to HDR including different HDR technologies, tone mapping, color representation, and HDR standards. It discusses concepts such as scene-referred vs display-referred conversions, and direct mapping vs tone mapping when converting between HDR and SDR formats. The document also examines potential side effects when mixing different conversion techniques in a production workflow.
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. Some key points:
- HEVC was created as a new video compression standard to address the growing needs of higher resolution video content and more efficient compression compared to prior standards like H.264.
- It achieves 50% bitrate reduction over H.264 for the same visual quality or improved quality at the same bitrate.
- The standard uses a block-based coding structure with coding tree units and supports intra-frame and inter-frame coding with motion estimation/compensation.
- It introduces more intra-prediction modes and block sizes along with improved transforms, quantization, and entropy coding.
This document provides an overview of high dynamic range (HDR) technology and workflows for HDR video production and mastering. It discusses HDR standards like SMPTE ST 2084 and ARIB STB-B67, camera log curves, luminance levels, and tools for setting up HDR monitoring including waveform monitors. Specific topics covered include HDR graticules, setting luminance levels for highlights and grey points, and using zebra patterns and zoom modes to evaluate highlight levels in HDR images.
The document discusses video compression history and standards, including codecs such as H.261, H.262/MPEG-2, H.263, H.264/AVC, H.265/HEVC, and the roles of organizations like MPEG, VCEG, and ITU-T in developing video coding standards to ensure interoperability. It also covers video encoding and decoding principles, as well as common container formats and their applications in areas like broadcasting, streaming, and storage.
Versatile Video Coding: Compression Tools for UHD and 360° VideoMathias Wien
The document discusses the development of the Versatile Video Coding (VVC) standard. It describes how a call for proposals was issued to develop coding tools beyond HEVC. 46 proposals were submitted across standard dynamic range, high dynamic range, and 360-degree video categories. The proposals were evaluated through subjective testing and shown to provide over 40% bitrate reduction compared to HEVC and over 10% reduction compared to the Joint Exploration Model, with the best proposals demonstrating visual quality equal or better than HEVC at higher bitrates. Seven proposals were identified as significantly better than the Joint Exploration Model. This marked the starting point for developing the VVC standard based on the selected coding tools from the top-performing proposals
An Overview of High Efficiency Video Codec HEVC (H.265)Varun Ravi
The document provides an overview of the High Efficiency Video Coding (HEVC) H.265 standard. It discusses the need for improved video compression standards due to increasing video content and limited bandwidth. HEVC was developed to meet this need by providing around 50% better compression over its predecessor H.264 while still maintaining high video quality. The document describes the various techniques used in HEVC such as improved block partitioning, transform sizes, prediction modes, and entropy coding that help achieve its compression gains. Both hardware and software implementations of HEVC decoders and encoders are discussed.
MIPI DevCon 2021: Meeting the Needs of Next-Generation Displays with a High-P...MIPI Alliance
Presented by Alain Legault, Hardent Inc.; Joe Rodriguez, Rambus Inc.; and Justin Endo, Mixel, Inc.
Next-generation display applications have an insatiable appetite for bandwidth. Using a combination of VESA Display Stream Compression (DSC) and MIPI DSI-2℠ technology, designers can achieve display resolutions up to 8K without compromise to video quality, battery life or cost. This presentation discusses a fully integrated, off-the-shelf display IP subsystem solution, consisting of Mixel (MIPI C-PHY℠/D-PHY℠ combo), Rambus (MIPI DSI-2® controller) and Hardent (VESA DSC) IP, that can deliver this state-of-the-art performance in a power-efficient and compact footprint.
Video compression techniques exploit various types of redundancy in video signals to reduce the data required to represent them. Key techniques include intra-frame compression which uses spatial redundancy within frames via DCT, inter-frame compression which uses temporal redundancy between consecutive frames by encoding differences, and motion compensation which accounts for motion between frames. Popular video compression standards like MPEG use a combination of these techniques including I, P and B frames along with motion estimation to achieve much higher compression ratios than image compression alone.
This document provides definitions and explanations of various optical terminology related to light passing through a lens, including:
- Dispersion, refraction, diffraction, reflection, focal point, focal length, principal point, image circle, aperture ratio, numerical aperture, optical axis, and more. It discusses concepts such as entrance pupil, exit pupil, angular aperture, and how they relate to lens performance. The document also covers topics like vignetting, the cosine law, and flare. Overall, it serves as a comprehensive reference for understanding optical and photographic lens terminology.
Video Compression Standards - History & IntroductionChamp Yen
This document provides an overview of several video compression standards including MPEG-1/2, MPEG-4, H.264, and HEVC/H.265. It discusses the key concepts of video coding such as entropy coding, quantization, transformation, and intra- and inter-prediction. For each standard, it describes the main coding tools and improvements over previous standards, focusing on techniques for more efficient prediction and extraction of redundant spatial and temporal information while maintaining quality. The development of these standards has moved towards more fine-grained partitioning and new coding ideas and tools to reduce bitrates further.
Video Compression, Part 4 Section 1, Video Quality Assessment Dr. Mohieddin Moradi
This document provides an overview of video compression artifacts that can occur when video is compressed for streaming or storage. It discusses both spatial artifacts, such as blurring, blocking, ringing, and color bleeding, as well as temporal artifacts like flickering and mosquito noise. For each artifact, it describes the visual appearance and potential causes from factors like quantization during compression, motion compensation between frames, and chroma subsampling. The document aims to help understand how compression can degrade perceptual video quality and different types of artifacts that may be evaluated both objectively and subjectively.
HEVC/H.265 is a video compression standard that provides around 50% better compression over H.264/AVC for the same level of video quality. It was finalized in 2013 by the joint collaboration of MPEG and ITU-T. Key features of HEVC include support for higher resolutions like 4K and 8K, improved parallel processing abilities, increased coding efficiency through larger block sizes and an expanded set of prediction modes.
Machine learning approaches are being explored for video compression. Conservative approaches replace individual MPEG blocks with deep learning blocks, while disruptive end-to-end approaches aim to replace the entire MPEG chain. Optical flow networks can exploit temporal redundancy by estimating motion between frames. Fully neural network-based video compression models jointly learn motion estimation, motion compression, and residual compression in an end-to-end optimized framework. However, performance gains must be balanced against increased complexity, and neural network approaches are not yet mature enough to be included in video compression standards.
The document provides a history of the development of television technology from the late 1800s through the 1920s. Some key developments include:
- In 1873, experiments with selenium, which is light-sensitive and formed the basis for early televisions.
- In 1884, the Nipkow disk laid down many basic concepts like scanning and synchronization.
- In 1923, Vladimir Zworykin developed the Kinescope, which allowed television programs to be recorded on film.
- In 1924, John Logie Baird transmitted the first television image.
- In 1925, Vladimir Zworykin demonstrated 60-line television using a curved-line image structure typical of mechanical television at the time.
The document provides an introduction to video compression. It discusses key concepts such as lossy vs lossless compression, encoders, decoders, and codecs. It also covers techniques used in video compression like sampling, quantization, model-based transforms, the human visual system, color space transforms, block-based coding, and the discrete cosine transform. Video compression standards like MPEG compress video using techniques like motion estimation, motion compensation, and encoding frames individually.
The latest video compression standard, H.264 (also known as MPEG-4 Part 10/AVC for Advanced Video
Coding), is expected to become the video standard of choice in the coming years.
H.264 is an open, licensed standard that supports the most efficient video compression techniques available
today. Without compromising image quality, an H.264 encoder can reduce the size of a digital video file by
more than 80% compared with the Motion JPEG format and as much as 50% more than with the MPEG-4
Part 2 standard. This means that much less network bandwidth and storage space are required for a video
file. Or seen another way, much higher video quality can be achieved for a given bit rate.
Video Compression, Part 3-Section 2, Some Standard Video CodecsDr. Mohieddin Moradi
This document discusses MPEG-2 Transport Streams and Packetized Elementary Streams. It describes how MPEG-2 Transport Streams use fixed length 188 byte packets containing compressed video, audio or data from one or more programs identified by Packet IDs. These packets can contain Packetized Elementary Stream packets which contain compressed elementary streams with timestamps for synchronization. The document also discusses how Transport Streams allow for synchronous multiplexing of multiple programs from independent time bases into a single stream.
The document discusses video compression basics and MPEG-2 video compression. It explains that video frames contain redundant spatial and temporal data that can be compressed. MPEG-2 uses three frame types (I, P, B frames) and compresses frames using intra-frame and inter-frame encoding techniques like DCT, quantization, and entropy encoding to remove redundancy. The encoding process transforms raw video frames to compressed bitstreams for efficient storage and transmission.
Trends and Recent Developments in Video Coding StandardizationMathias Wien
This document summarizes a tutorial on trends and recent developments in video coding standardization. It discusses the history of video coding standards organizations and the standards they have developed. These include MPEG-1, H.261, H.262, H.264, H.265 and the upcoming H.266 Versatile Video Coding standard. The document outlines the tutorial, which will cover topics like video resolutions, current compression techniques, VVC, and future trends in areas like multi-camera coding.
The document summarizes an upcoming webinar on new developments in MPEG standards. It will discuss Versatile Video Coding (VVC), MPEG-H 3D Audio Baseline Profile, video-based point cloud compression (V-PCC), and MPEG Immersive Video (MIV). The webinar will provide overviews of each standard and their applications, as well as results from recent verification tests that evaluated subjective quality and performance. Speakers will include leaders from MPEG working groups and the Joint Video Experts Team.
1) The document discusses video compression and streaming technologies, including standards like H.264 and challenges of streaming over heterogeneous networks.
2) It outlines objectives to develop versatile encoder and decoder architectures, efficient compression algorithms, and new concepts for adaptive streaming over IP networks.
3) Key outcomes included advanced encoder and decoder architectures, improved video processing algorithms, an end-to-end H.264 streaming system, and a scalable video coding scheme.
The document discusses standardization in multimedia technologies. It describes how standardization is a strategic process that requires early collaboration between researchers, industry, and policymakers. It then provides details on some key international and European standardization organizations and processes. Examples of standardized multimedia technologies like JPEG, MPEG, and H.264 are also summarized. The document concludes by discussing intellectual property issues in standardization and the importance of having a strategic, multi-lateral approach to participation in standardization.
Tutorial High Efficiency Video Coding Coding - Tools and Specification.pdfssuserc5a4dd
This document provides an outline for a tutorial on High Efficiency Video Coding (HEVC). It discusses the motivation for developing a new video coding standard to support higher resolutions and bandwidth efficiency. It describes the formation of the Joint Collaborative Team on Video Coding (JCT-VC) by MPEG and VCEG to develop the HEVC specification. It also gives an overview of the hybrid coding scheme used in HEVC and other video coding standards, including prediction, transform coding of residuals, and entropy coding.
The document discusses dynamics and trends in mobile video, including:
1) Mobile devices have become the dominant platform for media interactions, with over 1/3 of daily interactions occurring on smartphones.
2) Advancements in mobile devices like higher resolution displays and better cameras, as well as increased availability of digital content, have led to mobile video becoming the dominant source of mobile data traffic growth.
3) However, the need for higher resolution video like 4K and 8K will require data capacities to increase 64-fold in the next decade, posing challenges unless video compression efficiency improves beyond current standards.
The H.264/AVC Advanced Video Coding Standard: Overview and ...Videoguy
This document provides an overview of the H.264/AVC video coding standard and its Fidelity Range Extensions (FRExt). It discusses how H.264/AVC was developed jointly by ISO/IEC MPEG and ITU-T VCEG to improve coding efficiency over prior standards. The FRExt amendment adds support for higher chroma sampling, bit depths, and other capabilities for demanding professional applications. Initial industry feedback indicates rapid adoption of the High Profile added in FRExt.
The document discusses challenges and recommendations for delivering 4K and virtual reality content. It begins by defining key terms like resolution, codecs, formats, and encoding for 4K and VR. It then examines use cases for delivering 360-degree VR videos and live 4K experiences. Challenges include high bandwidth requirements, ensuring low latency and quality for VR, and achieving resilient ingest for live streams. The document recommends techniques like tiled encoding for VR and using Akamai's media services for live 4K delivery to address these challenges. It emphasizes that new technologies are emerging in over-the-top media before traditional broadcast.
The document discusses the H.265/HEVC video coding standard. It provides an overview of HEVC version 1.0 and its extensions, including its coding efficiency compared to prior standards. Studies show HEVC achieves 50% bitrate reduction over H.264/AVC for the same subjective quality. For low delay applications, HEVC requires 48-73% fewer bits than VP9 or H.264/AVC encoders. While reference encoders are slow, real-time encoders have approached the coding efficiency of the reference within 1.5 years. Future extensions include higher color depths, multiview, and scalable coding.
R&S. UH. Practical aspects of the implementationSergii Pedorenko
The document discusses Sky's implementation of a 24/7 UHD television service. It provides timelines for the project starting in March 2016 and going live in October 2016. It describes the equipment used for ingest, production, playout, distribution and monitoring of UHD content. This includes servers, storage, encoding and monitoring solutions from Rohde & Schwarz.
MPEG Immersive Media
By Thomas, Director, Technical Standards at Qualcomm
at 2nd ITU-T Mini-Workshop on Immersive Live Experience (ILE) in 19 January 2017
This document provides an overview of MPEG Immersive Video (MIV) including:
- MIV enables 6 degrees of freedom immersive video playback through compression of multi-view or multi-plane video, geometry, and texture.
- It specifies a bitstream format that leverages existing 2D video codecs for storage and distribution of immersive video over networks.
- The MIV test model includes an encoder, conforming decoder, and renderer to experiment with different MIV configurations.
Video services are evolving from traditional two-dimensional video to virtual reality and holograms, which offer six degrees of freedom to users, enabling them to freely move around in a scene and change focus as desired. However, this increase in freedom translates into stringent requirements in terms of ultra-high bandwidth (in the order of Gigabits per second) and minimal latency (in the order of milliseconds). To realize such immersive services, the network transport, as well as the video representation and encoding, have to be fundamentally enhanced. The purpose of this tutorial article is to provide an elaborate introduction to the creation, streaming, and evaluation of immersive video. Moreover, it aims to provide lessons learned and to point at promising research paths to enable truly interactive immersive video applications toward holography.
The document describes a real-time JVT SD encoder demo that will be shown at the 4th JVT meeting in Klagenfurt, Austria from July 22-26, 2002. The demo uses an MPEG-4 Part 10/H.264 compliant real-time video encoder that can encode standard definition video at 30 frames per second near 1 Mbps with DVD quality. The document invites attendees to see the demo and provides contact information. It also provides diagrams of the demo system and describes the hardware and software functions and components as well as future plans.
The document summarizes improvements made to the Video Conferencing (VIC) tool to support high definition video for access grids. Key points:
1) The updated VIC tool leverages existing open source resources like FFmpeg to incorporate modern video codecs like MPEG-4 and H.264, allowing for higher quality video streaming in HD resolutions.
2) It adds features for error resilience, efficient color conversion, scaling viewing windows, and full-screen snapshots not previously supported by VIC.
3) Performance tests show the MPEG-4 codec can achieve television quality at 1Mbps for 720x480 video streams, using less bandwidth than older codecs like H.261.
The document summarizes improvements made to the Video Conferencing (VIC) tool to support high definition video for access grids. Key points:
1) The updated VIC tool leverages existing open source resources like FFmpeg to incorporate modern video codecs like MPEG-4 and H.264, allowing for higher quality video streaming in HD resolutions.
2) It adds features for error resilience, efficient color conversion, scaling viewing windows, and full-screen snapshots not previously supported by VIC.
3) Performance tests show the MPEG-4 codec can support 720p video at 25 frames/sec using around 1Mbps bandwidth while maintaining good quality.
Similar to Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360° Video (20)
A recent direction in Business Process Management studied methodologies to control the execution of Business Processes under several sources of uncertainty in order to always get to the end by satisfying all constraints. Current approaches encode business processes into temporal constraint networks or timed game automata in order to exploit their related strategy synthesis algorithms. However, the proposed encodings can only synthesize single-strategies and fail to handle loops. To overcome these limits I will discuss a recent approach based on supervisory control. The approach considers structured business processes with resources, parallel and mutually exclusive branches, loops, and uncertainty. I will discuss an encoding into finite state automata and prove that their concurrent behavior models exactly all possible executions of the process. After that, I will introduce tentative commitment constraints as a new class of constraints restricting the executions of a process. Finally, I will discuss a tree decomposition of the process that plays a central role in modular supervisory control.
In his ignite talk „The Digital Transformation of Education: A Hyper-Disruptive Era through Blockchain and Generative AI,“ Dr. Alexander Pfeiffer delves into the intricate challenges and potential benefits associated with integrating blockchain technologies and generative AI into the educational landscape. He scrutinizes consensus algorithms and explores sustainable methods of operating blockchain systems, while also examining how smart contracts and transactions can be tailored to meet the specific needs of the educational sector. Alexander underscores the importance of establishing secure digital identities and ensuring robust data protection, while simultaneously casting a critical eye on potential risks and vulnerabilities. The topic of digital identities, facilitated through tokenization, forms a bridge between storing data using blockchain-based databases and the increasingly urgent need for content verification of AI-generated material.
Alexander explores the profound alterations occurring in teaching methodologies, assignment creation, and evaluation processes, shedding light on the hyper-disruptive impact these changes are having on both research and practical applications in education. The production of textual content by educators and students is analyzed with a focus on ensuring clear traceability of content sources and editors, and its proper citation, a critical aspect in the responsible use of AI. In addition to generative text and graphics, AI plays a crucial role in future learning and assignment practices, particularly through adaptive game-based learning and assessment. Alexander will provide a brief glimpse into his game „Gallery-Defender,“ a prototype demonstrating how AI and blockchain can be effectively implemented in serious gaming scenarios.
Furthermore, he emphasizes the imperative for ongoing education and professional development for educational personnel, advocating for a proactive stance in addressing the (legal) challenges associated with AI-generated images and text. This ignite talk aims to provide a balanced and critically reflective perspective on hyper-disruptive technologies, setting the stage for further discourse and exploration in the subsequent discussion.
The simulation of melee combat is central to many contemporary and traditional strategic games and simulations. In order to elevate this element of play from mere exercises of stats-comparison and dice rolling to a meaningful experience of play, strategy games rely on a rich plethora of cultural motives as deciding factors of their mechanic design. On the example of Samurai-themed skirmishing games, my talk elaborates on the impact that (popular) culture and other inspirations have on gaming experiences. It provides concrete examples from Japanese history, its traditional cinema, and postmodern Western reflections of Japanese cultural practices. Based on these insights, it compares four tabletop strategy games, muses on which phenomena they have adapted in their mechanics, and asks why or why not they may succeed in capturing a cultural essence via their rules.
Ultimately, this comparative approach shall serve to decipher the interplay of dice mechanics and aesthetic properties as the longing for a dramatic ideal in tabletop gaming and encourage participants to reflect on the idea in a subsequent, shared gaming experience.
How does a development team expand on an already existing game?
We will look at the two community driven and committee led expansions to the abandoned Tabletop game 'GuildBall' and explore the stages of development that the game went through. The art and lore driven approach employed will show us how rough sketches and concept ideas become a fully fledged ruleset and ultimately miniatures that can be put on the table. We will also explore pitfalls in rules design like over complicating abilities, the lack of streamlining across the game or simply creating expansions who break the game instead of the mold.
The document discusses Ben Calvert-Lee's work developing miniatures for tabletop games. It begins with an introduction to Ben's background and current role as a freelance lead sculptor. It then outlines the typical development pipeline for miniatures, from initial concepts and artwork to production. The document also discusses different miniature production methods. A case study details Ben's process for developing the Tengu faction for a game, including exploring species archetypes and incorporating unexpected developments into the designs.
In recent years, we have experienced an exponential growth in the amount of data generated by IoT devices. Data have to be processed strict low latency constraints, that cannot be addressed by conventional computing paradigm and architectures. On top of this, if we consider that we recently hit the limit codified by the Moore’s law, satisfying low-latency requirements of modern applications will become even more challenging in the future. In this talk, we discuss challenges and possibilities of heterogeneous distributed systems in the Post-Moore era.
In the modern world, we are permanently using, leveraging, interacting with, and relying upon systems of ever higher sophistication, ranging from our cars, recommender systems in eCommerce, and networks when we go online, to integrated circuits when using our PCs and smartphones, security-critical software when accessing our bank accounts, and spreadsheets for financial planning and decision making. The complexity of these systems coupled with our high dependency on them implies both a non-negligible likelihood of system failures, and a high potential that such failures have significant negative effects on our everyday life. For that reason, it is a vital requirement to keep the harm of emerging failures to a minimum, which means minimizing the system downtime as well as the cost of system repair. This is where model-based diagnosis comes into play.
Model-based diagnosis is a principled, domain-independent approach that can be generally applied to troubleshoot systems of a wide variety of types, including all the ones mentioned above. It exploits and orchestrates techniques for knowledge representation, automated reasoning, heuristic problem solving, intelligent search, learning, stochastics, statistics, decision making under uncertainty, as well as combinatorics and set theory to detect, localize, and fix faults in abnormally behaving systems.
In this talk, we will give an introduction to the topic of model-based diagnosis, point out the major challenges in the field, and discuss a selection of approaches from our research addressing these challenges. For instance, we will present methods for the optimization of the time and memory performance of diagnosis systems, show efficient techniques for a semi-automatic debugging by interacting with a user or expert, and demonstrate how our algorithms can be effectively leveraged in important application domains such as scheduling or the Semantic Web.
Function-as-a-Service (FaaS) is the latest paradigm of cloud computing in which developers deploy their codes as serverless functions, while the entire underlying platform and infrastructure is completely managed by cloud providers. Each cloud provider offers a huge set of cloud services and many libraries to simplify development and deployment, but only inside their clouds, often in a single cloud region. With such „help“ of cloud providers, users are locked to use resources and services of the selected cloud provider, which are often limited. Moreover, such heterogeneous and distributed environment of multiple cloud regions and providers challenge scientists to engineer cloud applications, often in a form of serverless workflows. In this talk, I will present our design principle „code once, run everywhere, with everything“. In particular, I will present challenges and our approaches and techniques how to program, model, orchestrate, and run distributed serverless workflow applications in federated FaaS.
This document summarizes a presentation on machine learning and fluid network planes. It begins with an agenda and introduction to fluid network planes and instances. It then discusses the role of machine learning in fluid network planes, including applications such as optimization, virtual network embedding problems, run-time operations, and intent-based closed-loop automation. Recent research is presented on machine learning-based YouTube QoE estimation using real 4G/5G network traces to predict video quality and inform control actions. Results are shown comparing 4G and 5G networks in terms of radio parameters, stalling events, handovers, and video resolutions under different mobility conditions.
The dynamics of networks enables the function of a variety of systems we rely on every day, from gene regulation and metabolism in the cell to the distribution of electric power and communication of information. Understanding, steering and predicting the function of interacting nonlinear dynamical systems, in particular if they are externally driven out of equilibrium, relies on obtaining and evaluating suitable models, posing at least two major challenges. First, how can we extract key structural system features of networks if only time series data provide information about the dynamics of (some) units? Second, how can we characterize nonlinear responses of nonlinear multi-dimensional systems externally driven by fluctuations, and consequently, predict tipping points at which normal operational states may be lost? Here we report recent progress on nonlinear response theory extended to predict tipping points and on model-free inference of network structural features from observed dynamics.
When it comes to integrating digital technologies into the classroom in higher education, many teachers face similar challenges. Nevertheless, it is difficult for teachers to share experiences because it is usually not possible to transfer successful teaching scenarios directly from one area to another, as subject-specific characteristics make it difficult to reuse them. To address this problem, instructional scenarios can be described as patterns that have been used previously in educational contexts. Patterns can capture proven teaching strategies and describe instructional scenarios in a consistent structure that can be reused. Because priorities for content, methods, and tools are different in each domain, a consensus-tested taxonomy was first developed with the goal of modeling a domain-independent database to collect digital instructional practices. In addition, this presentation will present preliminary insights into a data-driven approach to identifying effective instructional practices from interdisciplinary data as patterns. A web-based application will be developed for this that can both collect teaching/learning scenarios and individually extract scenarios from patterns for a learning platform.
The document discusses performance characterization across a computing continuum from the edge to the cloud. It evaluates the performance of video encoding and machine learning tasks on different devices. For video encoding, older single-board computers had significantly higher encoding times than other resources but provided lower data transfer times. For machine learning, training a convolutional neural network took much longer than a simpler model. Cloud and fog resources generally outperformed edge devices for more complex tasks. The document recommends offloading large or complex tasks to more powerful resources when possible.
East-west oriented photovoltaic power system is a new trend in orienting photovoltaic system. This lecture presents an evaluation of east–west oriented photovoltaic power system. A comparison between east–west oriented photovoltaic system and south oriented photovoltaic system in terms of cost of energy and technical requirement is conducted is presented in this lecture. In addition to that, the benefits of using east–west oriented photovoltaic system are discussed in this paper.
The document discusses using randomized recurrent neural networks and signature-based methods for machine learning in finance. It proposes splitting the input-output map of a dynamical system into a "reservoir" part and a linear "readout" part. The signature of the input signal provides a natural candidate for the reservoir, as it is point-separating and linear functions on the signature can approximate continuous functionals via the universal approximation theorem. The goal of the talk is to prove how dynamical systems can be approximated using randomized recurrent networks, with precise convergence rates, and to view randomized deep networks through this lens.
We live in a “digital” world, the separation between physical and virtual makes (almost) no sense anymore. Here, the Corona pandemic has also acted as an accelerator/magnifier demonstrating that the future of our digital society is here with all its possibilities, but also shortcomings.
In his talk, Hannes Werthner will briefly reflect on the history of computer science, and then discuss the need for an interdisciplinary response to these shortcomings. Such an answer is the Digital Humanism, which looks at this interplay of technology and humankind, it analyzes, and, most importantly, tries to influence the complex interplay of technology and humankind, for a better society and life. In the second part he will discuss this approach, and show what was achieved since its first workshop in 2019, and what lies ahead.
In the latest years, we have witnessed a growing number of media transmitted and stored on computers and mobile devices. For this reason, there is an actual need to employ smart compression algorithms to reduce the size of our media files. However, such techniques are often responsible for severe reduction of user perceived quality. In this talk we present several approaches we have developed to restore degraded images and videos to match their original quality, making use of Generative Adversarial Networks. The aim of the talk is to highlight the main features of our research work, including the advantages of our solution, the current challenges and the possible directions for future improvements.
Recommendation systems today are widely used across many applications such as in multimedia content platforms, social networks, and ecommerce, to provide suggestions to users that are most likely to fulfill their needs, thereby improving the user experience. Academic research, to date, largely focuses on the performance of recommendation models in terms of ranking quality or accuracy measures, which often don’t directly translate into improvements in the real-world. In this talk, we present some of the most interesting challenges that we face in the personalization efforts at Netflix. The goal of this talk is to sunshine challenging research problems in industrial recommendation systems and start a conversation about exciting areas of future research.
The document discusses the evolution to 5G networks and their benefits. It covers 5G principles like enhanced mobile broadband, massive machine-type communication, and ultra-reliable low-latency communications. Statistics are provided on 5G subscriptions, deployments, and expected growth in mobile data traffic. Use cases like smart cities, VR/AR, and autonomous vehicles are described. The presentation outlines Ericsson's 5G experience and global footprint.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Project Management Semester Long Project - Acuityjpupo2018
Acuity is an innovative learning app designed to transform the way you engage with knowledge. Powered by AI technology, Acuity takes complex topics and distills them into concise, interactive summaries that are easy to read & understand. Whether you're exploring the depths of quantum mechanics or seeking insight into historical events, Acuity provides the key information you need without the burden of lengthy texts.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360° Video
1. Versatile Video Coding – Video Compression beyond
HEVC: Coding Tools for SDR and 360°Video
AAU Klagenfurt, May 14th, 2018
Mathias Wien
Institut für Nachrichtentechnik
RWTH Aachen University
wien@ient.rwth-aachen.de
2. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
2
• Introduction
Standardization development and process
• Versatile Video Coding Development
Joint Call for Proposals Outcome
• Coding Tools
Versatile Video Coding Test Model
• Tools proposed by RWTH
Geometric Partitioning
360° Tools
• Summary and Outlook
Next steps
Outline
3. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
3
INTRODUCTION
4. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
4
Video coding standardization organisations
• ISO/IEC MPEG = “Moving Picture Experts Group”
(ISO/IEC JTC 1/SC 29/WG 11 = International Standardization Organization and International Electrotechnical
Commission, Joint Technical Committee 1, Subcommittee 29, Working Group 11)
• ITU-T VCEG = “Video Coding Experts Group”
(ITU-T SG16/Q6 = International Telecommunications Union – Telecommunications Standardization Sector
[United Nations Organization, formerly CCITT], Study Group 16, Working Party 3, Question 6)
• JVT = “Joint Video Team” collaborative team of MPEG & VCEG, responsible for developing Advanced Video
Coding (AVC) (discontinued in 2009), documents and software publicly available
• JCT-VC = “Joint Collaborative Team on Video Coding” team of MPEG & VCEG , responsible for
developing High Efficiency Video Coding (HEVC) (established January 2010), documents and software
publicly available
• JVET = “Joint Video Exploration Team” exploring potential for new technology beyond HEVC (established
Oct. 2015) – renamed to “Joint Video Experts Team” responsible for developing Versatile Video Coding
(VVC) from April 2018, documents and software publicly available
5. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
5
History of international video coding standardization
H.263/+/++
(1995-2000+)
MPEG-4
Visual
(1998-2001+)
MPEG-1
(1993)
ISO/IECITU-T
H.120
(1984-1988)
H.261
(1990+)
H.262 / 13818-2
(1994/95-1998+)
H.264 / 14496-10
AVC
(2003-2008+)
H.265 / 23008-2
HEVC
(2013-2016+)
Videotelephony
Computer
SD HD 4K UHD
(Advanced Video Coding) (High Efficiency
Video Coding)
(MPEG-2)
H.26x / 23090-3
VVC
(2020-...)
8K, 360, ...
(Versatile Video Coding)
6. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
6
The scope of video standardization
• Only Specifications of the Bitstream, Syntax, and Decoder are standardized:
• Permits optimization beyond the obvious
• Permits complexity reduction for implementability
• Provides no guarantees of quality
Pre-Processing Encoding
Source
Destination
Post-Processing
& Error Recovery
Decoding
Scope of Standard
7. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
7
Hybrid Coding Concept
Basis of every standard since H.261
8. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
8
Performance history of standard generations
0 100 200 300
28
30
32
34
36
38
40
bit rate (kbit/s)
PSNR
(dB)
Foreman
10 Hz, QCIF
100 frames
HEVC
AVC H.262/MPEG-2 H.261H.263 +
MPEG-4 Visual
JPEG
35
Bit-rate Reduction: 50%
9. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
9
• Video is continually increasing by resolution
HD existing, UHD (4Kx2K, 8Kx4K) appearing
Mobile services going towards HD/UHD
Stereo, multi-view, 360° video
• Devices available to record and display ultra-high resolutions
Becoming affordable for home and mobile consumers
• Video has multiple dimensions to grow the data rate
Frame resolution, Temporal resolution
Color resolution, bit depth
Multi-view
Visible distortion still an issue with existing networks
• Necessary video data rate grows faster than feasible network transport capacities
Better video compression (50% rate of current HEVC) needed, even after availability of 5G
Motivation for improved video compression
10. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
10
VVC DEVELOPMENT
11. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
11
• Exploration activity in the JVET
starting Oct/2015
Investigation of tools integrated on top of HEVC:
Joint Exploration Model (JEM) software
Larger block sizes / transforms, improved intra / inter
prediction tools, decoder-side derivation / refinement
methods, adaptive loop filters, …(*)
Objective gains >25% measurable on test set
But: Evaluation software only, significant increase of
encoder run times
Joint Call for Evidence (issued Mar/2017, evaluated Jul/2017):
Significant compression gains asserted
Joint Call for Proposals (issued Oct/2017, evaluated Apr/2018):
Kick-off for Versatile Video Coding (VVC)
Steps towards next generation standard – Versatile Video Coding (VVC)
Figure from: JVET AHG report: Tool evaluation (AHG1) [JVET-H0001](*) Details on JEM coding tools in VCAS lecture
12. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
12
• VVC should be applicable for many types of data
SDR and HDR up to extremly high resolutions
All kind of camera generated content
Computer generated content
Non-camera video modalities e.g. medical data
360°, lightfield, depth, and volumetric video
• VVC should support flexible random and localized access
Low delay, random access, trick modes
Error resilience, video buffer, system layer interface
Possible support for scalability and multi-view
Steps towards next generation standard – Versatile Video Coding (VVC)
13. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
13
• Document JVET-H1002
• Test categories
Standard dynamic range (SDR): 5 UHD and 5 HD sequences
High dynamic range (HDR): 3 HLG and 5 PQ sequences
360° video (360): 5 sequences in ERP format
• Constraint sets
Constraint set 1 (C1): Random access configuration
Max 1.1s random access intervals, structural delay max 16 pictures
Constraint set 2 (C2): Low delay configuration only evaluated for SDR HD sequences
No picture reordering between input and output
• Encoding constraints
No pre-processing, post-processing only within the coding loop
Static quantizer setting with one-time change to meet target bitrate
Relevant optimization methods to be reported
Joint Call for Proposals (CfP) on Video Compression with Capability beyond HEVC
UHD = Ultra High Definition, HD = High Definition, HLG = Hybrid Log Gamma, PQ = Perceptive Quantization (ITU-T BT2020), ERP = Equirectangular Projection
15. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
15
• Category-specific submissions (total 46):
SDR: 22 submissions (8 of which are registered only in this category)
HDR: 12 submissions
360°: 12 submissions (2 of which are registered only in this category)
For all categories: HEVC anchors (HM) and JEM anchors
• Proposals described in input documents JVET-J0011...JVET-J0033
Participation of 32 institutions
• Evaluation: Double stimulus test
Rate points: lowest rate was typically less than "fair" quality for HEVC, but still possible to code
Three ways of judging benefit:
Mean MOS over all test cases (28x4 test points: 23x4 C1, 5x4 C2 )
Count cases where a proposal was visually better/worse than JEM
Count cases where a proposal was visually better than HEVC (HEVC at higher rate point)
• Reports: Input subjective test [JVET-J0080], output CfP results [JVET-J1003]
VVC CfP Responses
16. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
16
• Objective performance: best performers report
>40% bit rate reduction compared to HEVC,
>10% compared to JEM (for SDR case)
2 proposals used some degree of subjective optimization
1 proposal used large-segment multipass encoding
Similar ranges for HDR and 360°
Obviously, proposals with more elements show better performance
Nevertheless, some proposals show similar performance as JEM with significant complexity/run time
reduction vs. JEM
• Subjective tests generally show similar (or even better) tendency
Benefit over HEVC very clear
Benefit over JEM visible at various points
Performance
17. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
17
• JVET-J1003:
Report of subjective
evaluation contains
28 plots as shown,
one per sequence
• Count significant
cases of positive/
negative benefit
with non-overlapping
confidence interval
against JEM
Performance
HM
JEM
Rate1...4
Proposals ranked by MOS (per rate point)
+1 credit
-1 credit
18. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
18
• "Mean" and "significance-count"
method suggested at least 7
proposals that were obviously
better than JEM
Performance SDR
Pxx 10
Pxx 8
Pxx 8
Pxx 6
Pxx 6
Pxx 6
Pxx 6
Pnn 3
Pnn 3
Pnn 2
Pnn 2
Pnn 1
Pnn 1
JEM 0
Pnn 0
Pnn -1
Pnn -1
Pnn -1
Pnn -2
Pnn -2
Pnn -2
Pnn -3
Pnn -4
HM -36
Pxx 6,53
Pxx 6,46
Pxx 6,41
Pxx 6,37
Pxx 6,33
Pxx 6,33
Pxx 6,26
Pnn 6,23
Pnn 6,17
Pnn 6,15
Pnn 6,13
Pnn 6,11
Pnn 6,04
Pnn 6,04
Pnn 6,03
Pnn 6,03
Pnn 6,01
JEM 6,01
Pnn 6,00
Pnn 5,96
Pnn 5,94
Pnn 5,88
Pnn 5,86
HM 4,57
Mean MOS Significance vs. JEM
60 ... +60
19. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
19
• Similar
tendency
in HDR
and 360°
categories
• Mostly same
coding tools
as in SDR
provide good
benefit
Performance HDR / 360°
Mean MOS Signif. vs. JEM
Pxx 6,04
Pxx 6,00
Pxx 5,94
Pxx 5,93
Pxx 5,86
Pnn 5,85
Pnn 5,80
Pnn 5,67
JEM 5,62
Pnn 5,60
Pnn 5,59
Pnn 5,45
Pnn 5,11
HM 4,14
Pxx 7
Pxx 3
Pxx 2
Pxx 2
Pxx 2
Pnn 1
Pnn 1
JEM 0
Pnn 0
Pnn 0
Pnn -1
Pnn -1
Pnn -6
HM -20
32 ... +32
Mean MOS Signif. vs. JEM
Pxx 6,20
Pxx 6,19
Pxx 6,06
Pxx 6,03
Pxx 5,99
Pxx 5,96
Pxx 5,86
Pnn 5,69
Pnn 5,67
Pnn 5,51
Pnn 5,45
JEM 5,11
HM 3,79
Pnn 3,45
Pxx 9
Pxx 9
Pxx 8
Pnn 7
Pxx 7
Pxx 6
Pxx 5
Pxx 4
Pnn 2
Pnn 1
Pnn 1
JEM 0
HM -9
Pnn -12
20 ... +20HDR 360°
20. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
20
• Comparison of proposals to HEVC at higher rate points
Subjective quality of best performing proposals always equal or even better (about 1/3 of cases) than HEVC
at next higher rate point, over all categories (with approx. 40% less rate)
Subjective quality of best performing proposals always equal or even better (about 1/5 of cases) than HEVC
at second next higher rate point, in SDR-UHD category (with approx. 65% less rate)
• Highest rate point HEVC may be close to transparent quality in many cases, difficult to become better
• Though not always the same proposal performing best at a given rate point, it can be anticipated that merits of
different proposals could be combined
50% (or more) bit rate reduction with same quality will probably be achievable
Performance compared to HEVC
21. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
21
CODING TOOLS
22. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
22
• In terms of large architecture: Most proposals similar, no deviation from
hybrid coding mainstream
• Most improvements from further refinements of well-known building
blocks of HEVC and JEM
Partititioning: Quad/binary, augmented by ternary tree and finer
Intra prediction using
directional modes, DC and planar
sample smoothing with various adaptation
inheritance of chroma modes and chroma sample prediction from luma
Inter prediction: advanced motion vector prediction, affine models, sub-
block partitioning, switchable primary transforms, mostly DCT/DST variants
Secondary transforms targeting specific cases of prediction residual
characteristics
Adaptive loop filter based on local classification, some new variants
Quantization / context-adaptive arithmetic coding
CfP analysis: What was proposed?
23. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
23
• New elements (some come with high complexity):
Decoder side estimation for mode/MV derivation and sample prediction both in intra and inter coding (JEM)
Finer partitioning: Asymmetric, geometric
Neural networks for prediction, loop filtering, upsampling, (encoder control)
Additional elements using template matching
Intra block copy / current picture referencing
Additional non-linear, de-noising and statistics-based loop filters
Additional linear and non-linear elements in prediction
• HDR specific:
New adaptive reshaping and quantization, also in-loop
HDR-specific modifications of existing tools, e.g. deblocking
• 360-video specific:
Variants of projection formats, geometry-corrected face boundary padding
Modification and disabling of existing tools at face boundaries
CfP analysis: What was proposed?
24. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
24
• VVC Working Draft 1 / Test Model 1 (VTM1): basic approach
• VTM Block structure
Unified tree (coding block unites prediction and transform)
CTU size 128x128, rectangular blocks (dyadic sizes),
smallest luma size 4x4
Maximum transform size 64x64
• VTM: Some removed elements of HEVC:
Mode dependent transform (DST-VII), mode dependent scan
Strong intra smoothing
Sign data hiding in transform coding
Unnecessary high-level syntax (e.g. VPS)
Tiles and wavefront
Quantization weighting
VVC Test Model and Benchmark Set
• Benchmark Set defined in addition to
VTM, including the following well-known
JEM tools:
• 65 intra prediction modes
• Coefficient coding
• AMT + 4x4 NSST
• Affine motion
• GALF
• Subblock merge candidate (ATMVP)
• Adaptive motion vector precision
• Decoder motion vector refinement
• LM Chroma mode
Purpose: testing benefit of technology
against better performing set
25. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
25
• Prediction block partitioning of a 2N×2N CB
• Transform block partitioning of a CB
Quadtree partitioning of CB → Residual Quad Tree (RQT)
Transform size 4×4 to 32×32
TB size 4×4 to 64×64
PB boundaries inside TBs allowed
HEVC: Prediction Blocks (PBs) and Transform Blocks (TBs)
26. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
27
• Simple ternary-tree split was used in several proposals, can be alternated
with binary split
• Further proposed variants of partitioning included
Asymmetric binary split modes
Diagonal and geometric (wedge-shaped) split modes
Block Partitioning: Quadtree – Ternary Tree – Binary Tree
Example:
(source: JVET-J1002)
27. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
28
TOOLS PROPOSED BY RWTH
28. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
29
• Motivation: Towards object-oriented coding
Follow object boundaries more closely
Less coding artifacts where it matters
• Prediction, transform and coding driven by actual object
shape under RD-constraint
Inter- and intra-predicted segments for handling of
disocclusions
Overlapped wedge based filtering at partition boundary
Shape-adaptive DCT for spatially localized transform
coding
RWTH Proposal: Geometric Partitioning (GEO)
Source: M. Bläser, J. Sauer, and M. Wien, “Description of SDR and 360o video coding technology proposal
by RWTH Aachen University,” Doc. JVET-J0023, Joint Video Experts Team of ITU-T VCEG and ISO/IEC MPEG, San Diego, USA, 10th meeting, Apr. 2018
29. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
30
• GEO available for all block sizes ≥ 8×8 luma samples
• Partitioning is represented by two coordinate points 𝑃0 and 𝑃1 on the block boundary
• Prediction of two coordinate points 𝑃0 and 𝑃1 from 16 pre-defined templates (scaled for non-square blocks)
Alternative: Spatial or temporal prediction
Refinement: block size dependent offset
• Integration with AMVP, MERGE, FRUC
(no AFFINE (yet))
GEO: Partitioning Coding and Prediction
30. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
31
• No transform-tree in JEM 7.0 localization of residual error for larger blocks required
• ΔSA-DCT adapted from MPEG 4 software for blocks up to 128×128
• Currently floating point implementation – integer transform targeted
• SA-DCT signaled as additional transform choice next to full block DCT ( 4 total GEO transform modes)
• Coding of transform coefficients (TSBs, significance flags) with regard to shape
GEO: Shape-Adaptive DCT for Geometric Partitions
Segment with high prediction error
Segment with low prediction error
Example of 64×32 residual block
31. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
32
Results for GEO
JEM 7.0 JEM 7.0 + GEO
• Visual improvements at object boundaries
Sharper contours
Less staircase-effect
More background details
• Objective gains (BD-rate savings)
Against HEVC: ~33% on C1, ~25% on C2
Against JEM: ~0.8% for both, C1 and C2
JEM 7.0
32. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
33
Results for GEO
JEM 7.0 JEM 7.0 + GEO
• Visual improvements at object boundaries
Sharper contours
Less staircase-effect
More background details
• Objective gains (BD-rate savings)
Against HEVC: ~33% on C1, ~25% on C2
Against JEM: ~0.8% for both, C1 and C2
JEM 7.0 + GEO
33. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
34
• Motivation: Special characteristics of 360 content
360° symmetry not exploited by current codecs
Motion across face boundaries possible
Geometric distortions
Motion compensation suboptimal
Not correctly treated by loop filters
Here shown for cube, but similar problems for
other coding formats
• Proposal: Doing things correctly that “broke” for
360°content
Face extension for motion estimation and
compensation
Loop filtering over continuous boundaries
according to 3D arrangement
RWTH Proposal: 360°Coding Tools
Source: M. Bläser, J. Sauer, and M. Wien, “Description of SDR and 360o video coding technology proposal by RWTH Aachen University,”
Doc. JVET-J0023, Joint Video Experts Team of ITU-T VCEG and ISO/IEC MPEG, San Diego, USA, 10th meeting, Apr. 2018
34. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
35
360°coding tools - Face extension for cube projection (EAC/CMP/ACP)
𝑯 𝐵2𝐴 =
0 0 𝑓2
0 𝑓 0
−1 0 0
𝑓 =
face width
2
• Approach can be transferred to other coding formats
35. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
36
• Reference samples of blocks at face
boundaries changed
Original: Samples from top or left block are
used
Modified: Samples are chosen according to
3D cube geometry
• Approach can be transferred to other coding
formats
360°coding tools - Corrected deblocking filter (DBF)
36. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
37
• Objective gains (BD-rate savings)
Against HEVC anchor: ~31%
E2E WS-PSNR
Agains JEM (same projection
format): ~1.6%
Gains higher for sequences with
high motion
Results for 360°coding tools
JEM deblocking Proposed deblocking
37. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
38
SUMMARY AND OUTLOOK
38. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
39
• Report of Results from the Call for Proposals on Video Compression with Capability beyond HEVC
[JVET-J1003]
Documentation of results per sequence, marking HM and JEM anchors, not identifying individual proponents
Assessment of qualitative (and as far as possible quantitative) benefit of submitted technology compared to
anchors
• Working Draft 1 of Versatile Video Coding [JVET-J1001]
"Reduced" HEVC plus quad/binary/ternary tree structure
• Test Model 1 of Versatile Video Coding (VTM 1) [JVET-J1002]
Corresponding encoder and algorithm description
Documents issued after CfP Results
39. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
40
• CE1: Partitioning
• CE2: In-loop filters
• CE3: Intra prediction and mode coding
• CE4: Inter prediction and MV coding
• CE5: Arithmetic coding engine
• CE6: Transforms and transform signalling
• CE7: Quantization and coefficient coding
• CE8: Current picture referencing
• CE9: Decoder side MV derivation
• CE10: Combined and multi-hypothesis prediction
• CE11: Composite reference pictures
• CE12: Mapping for HDR content
• CE13: Projection formats
Core Experiments defined by JVET
40. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
41
• Call for Proposals demonstrated availability of significant compression benefit
HEVC out-performed by virtually all proposals
Subjective results suggest initial rate savings of 40+% over HEVC at starting point
• Versatile Video Coding (VVC): First Working Draft and Test Model defined
Reduced initial tool set
Step-by-step integration of tools
Evaluation of concurring variants of tools
Consideration of algorithmic complexity
Further fast progress expected, goal: finalization 2020
Summary and Outlook
41. Versatile Video Coding – Video Compression beyond HEVC: Coding Tools for SDR and 360°Video |
Mathias Wien | RWTH Aachen University | Institut für Nachrichtentechnik | 14.05.2018
42
• Document archives (publicly accessible)
JVET / VVC:
http://phenix.it-sudparis.eu/jvet
http://ftp3.itu.ch/av-arch/jvet-site
JCT-VC / HEVC:
http://phenix.it-sudparis.eu/jct
http://ftp3.itu.ch/av-arch/jctvc-site
• Software for HEVC, JEM, and 360 Video (publicly accessible):
https://jvet.hhi.fraunhofer.de/svn/svn_VVCSoftware_VTM
https://jvet.hhi.fraunhofer.de/svn/svn_VVCSoftware_BMS
https://jvet.hhi.fraunhofer.de/svn/svn_HMJEMSoftware/
https://jvet.hhi.fraunhofer.de/svn/svn_360Lib/
https://hevc.hhi.fraunhofer.de/svn/svn_HEVCSoftware/
Further Information