This document summarizes a research paper that proposes a new method to improve the efficiency of the HEVC video coding standard. The proposed method uses particle swarm optimization (PSO) to optimize the coding unit partitioning patterns in HEVC in order to reduce computational complexity. Compared to existing algorithms, the proposed PSO-based method achieves a 12% increase in PSNR video quality while reducing computational complexity by 40% and overall encoding delay. The method integrates this optimized intra-frame prediction into the HEVC encoding and decoding process.
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors.
Optimal coding unit decision for early termination in high efficiency video c...IJECEIAES
Video compression is an emerging research topic in the field of block based video encoders. Due to the growth of video coding technologies, high efficiency video coding (HEVC) delivers superior coding performance. With the increased encoding complexity, the HEVC enhances the rate-distortion (RD) performance. In the video compression, the out-sized coding units (CUs) have higher encoding complexity. Therefore, the computational encoding cost and complexity remain vital concerns, which need to be considered as an optimization task. In this manuscript, an enhanced whale optimization algorithm (EWOA) is implemented to reduce the computational time and complexity of the HEVC. In the EWOA, a cosine function is incorporated with the controlling parameter A and two correlation factors are included in the WOA for controlling the position of whales and regulating the movement of search mechanism during the optimization and search processes. The bit streams in the Luma-coding tree block are selected using EWOA that defines the CU neighbors and is used in the HEVC. The results indicate that the EWOA achieves best bit rate (BR), time saving, and peak signal to noise ratio (PSNR). The EWOA showed 0.006-0.012 dB higher PSNR than the existing models in the real-time videos.
IEEE MMSP'21: INCEPT: Intra CU Depth Prediction for HEVCVignesh V Menon
Abstract: High Efficiency Video Coding (HEVC) improves the encoding efficiency by utilizing sophisticated tools such as flexible Coding Tree Units(CTUs) partitioning. TheCodingUnits(CUs) can be split recursively into four equally sized CUs ranging from 64×64 to 8×8 pixels. At each depth level (or CU size), intra prediction via exhaustive mode search was exploited in HEVC to improve the encoding efficiency and result in a very high encoding time complexity. This paper proposes an Intra CU Depth Prediction(INCEPT) algorithm, which limits Rate-Distortion Optimization(RDO) for each CTU in HEVC by utilizing the spatial correlation with the neighboring CTUs, which is computed using a DCT energy-based feature. Thus, INCEPT reduces the number of candidate CU sizes required to be considered for each CTU in HEVC intra coding. Experimental results show that the INCEPT algorithm achieves a better trade-off between the encoding efficiency and encoding time saving (i.e., BDR/∆T) than the benchmark algorithms. While BDR/∆T is12.35% and 9.03% for the benchmark algorithms, it is 5.49% for the proposed algorithm. As a result, INCEPT achieves a 23.34% reduction in encoding time on average while incurring only a 1.67% increase in bitrate than the original coding in the x265 HEVC open-source encoder.
High Efficiency Video Coding (HEVC) improves the encoding efficiency by utilizing sophisticated tools such as flexible Coding Tree Unit (CTU) partitioning. The Coding Unit (CU) can be split recursively into four equally sized CUs ranging from 64×64 to 8×8 pixels. At each depth level (or CU size), intra prediction via exhaustive mode search was exploited in HEVC to improve the encoding efficiency and result in a very high encoding time complexity. This paper proposes an Intra CU Depth Prediction (INCEPT) algorithm, which limits Rate-Distortion Optimization (RDO) for each CTU in HEVC by utilizing the spatial correlation with the neighboring CTUs, which is computed using a DCT energy-based feature. Thus, INCEPT reduces the number of candidate CU sizes required to be considered for each CTU in HEVC intra coding. Experimental results show that the INCEPT algorithm achieves a better trade-off between the encoding efficiency and encoding time saving (i.e., BDR/∆T) than the benchmark algorithms. While BDR/∆T is 12.35% and 9.03% for the benchmark algorithms, it is 5.49% for the proposed algorithm. As a result, INCEPT achieves a 23.34% reduction in encoding time on average while incurring only a 1.67% increase in bit rate than the original coding in the x265 HEVC open-source encoder.
Video streaming using light-weight transcoding and in-network intelligenceMinh Nguyen
In this paper, we introduce a novel approach, LwTE, which reduces streaming costs in HTTP Adaptive Streaming (HAS) by enabling light-weight transcoding at the edge. In LwTE, during encoding of a video segment in the origin server, a metadata is generated which stores the optimal encoding decisions. LwTE enables us to store only the highest bitrate plus corresponding metadata (of very small size) for unpopular video segments/bitrates. Since metadata is of very small size, replacing unpopular video segments/bitrates with their metadata results in considerable saving in the storage costs. The metadata is reused at the edge servers to reduce the required time and computational resources for on-the-fly transcoding.
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors.
Optimal coding unit decision for early termination in high efficiency video c...IJECEIAES
Video compression is an emerging research topic in the field of block based video encoders. Due to the growth of video coding technologies, high efficiency video coding (HEVC) delivers superior coding performance. With the increased encoding complexity, the HEVC enhances the rate-distortion (RD) performance. In the video compression, the out-sized coding units (CUs) have higher encoding complexity. Therefore, the computational encoding cost and complexity remain vital concerns, which need to be considered as an optimization task. In this manuscript, an enhanced whale optimization algorithm (EWOA) is implemented to reduce the computational time and complexity of the HEVC. In the EWOA, a cosine function is incorporated with the controlling parameter A and two correlation factors are included in the WOA for controlling the position of whales and regulating the movement of search mechanism during the optimization and search processes. The bit streams in the Luma-coding tree block are selected using EWOA that defines the CU neighbors and is used in the HEVC. The results indicate that the EWOA achieves best bit rate (BR), time saving, and peak signal to noise ratio (PSNR). The EWOA showed 0.006-0.012 dB higher PSNR than the existing models in the real-time videos.
IEEE MMSP'21: INCEPT: Intra CU Depth Prediction for HEVCVignesh V Menon
Abstract: High Efficiency Video Coding (HEVC) improves the encoding efficiency by utilizing sophisticated tools such as flexible Coding Tree Units(CTUs) partitioning. TheCodingUnits(CUs) can be split recursively into four equally sized CUs ranging from 64×64 to 8×8 pixels. At each depth level (or CU size), intra prediction via exhaustive mode search was exploited in HEVC to improve the encoding efficiency and result in a very high encoding time complexity. This paper proposes an Intra CU Depth Prediction(INCEPT) algorithm, which limits Rate-Distortion Optimization(RDO) for each CTU in HEVC by utilizing the spatial correlation with the neighboring CTUs, which is computed using a DCT energy-based feature. Thus, INCEPT reduces the number of candidate CU sizes required to be considered for each CTU in HEVC intra coding. Experimental results show that the INCEPT algorithm achieves a better trade-off between the encoding efficiency and encoding time saving (i.e., BDR/∆T) than the benchmark algorithms. While BDR/∆T is12.35% and 9.03% for the benchmark algorithms, it is 5.49% for the proposed algorithm. As a result, INCEPT achieves a 23.34% reduction in encoding time on average while incurring only a 1.67% increase in bitrate than the original coding in the x265 HEVC open-source encoder.
High Efficiency Video Coding (HEVC) improves the encoding efficiency by utilizing sophisticated tools such as flexible Coding Tree Unit (CTU) partitioning. The Coding Unit (CU) can be split recursively into four equally sized CUs ranging from 64×64 to 8×8 pixels. At each depth level (or CU size), intra prediction via exhaustive mode search was exploited in HEVC to improve the encoding efficiency and result in a very high encoding time complexity. This paper proposes an Intra CU Depth Prediction (INCEPT) algorithm, which limits Rate-Distortion Optimization (RDO) for each CTU in HEVC by utilizing the spatial correlation with the neighboring CTUs, which is computed using a DCT energy-based feature. Thus, INCEPT reduces the number of candidate CU sizes required to be considered for each CTU in HEVC intra coding. Experimental results show that the INCEPT algorithm achieves a better trade-off between the encoding efficiency and encoding time saving (i.e., BDR/∆T) than the benchmark algorithms. While BDR/∆T is 12.35% and 9.03% for the benchmark algorithms, it is 5.49% for the proposed algorithm. As a result, INCEPT achieves a 23.34% reduction in encoding time on average while incurring only a 1.67% increase in bit rate than the original coding in the x265 HEVC open-source encoder.
Video streaming using light-weight transcoding and in-network intelligenceMinh Nguyen
In this paper, we introduce a novel approach, LwTE, which reduces streaming costs in HTTP Adaptive Streaming (HAS) by enabling light-weight transcoding at the edge. In LwTE, during encoding of a video segment in the origin server, a metadata is generated which stores the optimal encoding decisions. LwTE enables us to store only the highest bitrate plus corresponding metadata (of very small size) for unpopular video segments/bitrates. Since metadata is of very small size, replacing unpopular video segments/bitrates with their metadata results in considerable saving in the storage costs. The metadata is reused at the edge servers to reduce the required time and computational resources for on-the-fly transcoding.
Distributed Video Coding (DVC) has become increasingly popular in recent times among the researchers in video coding due to its attractive and promising features. DVC primarily has a modified complexity balance between the encoder and decoder, in contrast to conventional video codecs. However, Most of the reported DVC schemes have a high time-delay in decoder which hinders its practical application in real-time systems. In this work, we focus on speed up the Side Information(SI) generation module in DVC, which is a major function in the DVC coding algorithm and one of the time-consuming factor at the decoder. By applied it through Compute Unified Device Architecture (CUDA) based on General-Purpose Graphics Processing Unit (GPGPU), the experimental results show that a considerable speedup can be obtained by using the proposed parallelized SI generation algorithm.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
.
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERcscpconf
This paper proposes about motion estimation in H.264/AVC encoder. Compared with standards
such as MPEG-2 and MPEG-4 Visual, H.264 can deliver better image quality at the same
compressed bit rate or at a lower bit rate. The increase in compression efficiency comes at the
expense of increase in complexity, which is a fact that must be overcome. An efficient Co-design
methodology is required, where the encoder software application is highly optimized and
structured in a very modular and efficient manner, so as to allow its most complex and time
consuming operations to be offloaded to dedicated hardware accelerators. The Motion
Estimation algorithm is the most computationally intensive part of the encoder which is simulated using MATLAB. The hardware/software co-simulation is done using system generator tool and implemented using Xilinx FPGA Spartan 3E for different scanning methods.
Machine learning-based energy consumption modeling and comparing of H.264 and...IJECEIAES
Advancement of the prediction models used in a variety of fields is a result of the contribution of machine learning approaches. Utilizing such modeling in feature engineering is exceptionally imperative and required. In this research, we show how to utilize machine learning to save time in research experiments, where we save more than five thousand hours of measuring the energy consumption of encoding recordings. Since measuring the energy consumption has got to be done by humans and since we require more than eleven thousand experiments to cover all the combinations of video sequences, video bit rate, and video encoding settings, we utilize machine learning to model the energy consumption utilizing linear regression. VP8 codec has been offered by Google as a free video encoder in an effort to replace the popular H.264 video encoder standard. This research model energy consumption and describes the major differences between H.264/AVC and VP8 encoders based on of energy consumption and performance through experiments that are machine learning-based modeling. Twentynine uncompressed video segments from a standard data-set are used, with several sizes, details, and dynamics, where the frame sizes ranging from QCIF(176x144) to 2160p(3840x2160). For fairness in comparison analysis, we use seven settings in VP8 encoder and fifteen types of tuning in H.264/AVC. The settings cover various video qualities. The performance metrics include video qualities, encoding time, and encoding energy consumption.
Video Compression Algorithm Based on Frame Difference Approaches ijsc
The huge usage of digital multimedia via communications, wireless communications, Internet, Intranet and cellular mobile leads to incurable growth of data flow through these Media. The researchers go deep in developing efficient techniques in these fields such as compression of data, image and video. Recently, video compression techniques and their applications in many areas (educational, agriculture, medical …) cause this field to be one of the most interested fields. Wavelet transform is an efficient method that can be used to perform an efficient compression technique. This work deals with the developing of an efficient video compression approach based on frames difference approaches that concentrated on the calculation of frame near distance (difference between frames). The
selection of the meaningful frame depends on many factors such as compression performance, frame details, frame size and near distance between frames. Three different approaches are applied for removing the lowest frame difference. In this paper, many videos are tested to insure the efficiency of this technique, in addition a good performance results has been obtained.
A Novel Approaches For Chromatic Squander Less Visceral Coding Techniques Usi...IJERA Editor
Recent advances in video capturing and display technologies, along with the exponentially increasing demand of
video services, challenge the video coding research community to design new algorithms able to significantly
improve the compression performance of the current H.264/AVC standard. This target is currently gaining
evidence with the standardization activities in the High Efficiency Video Coding (HEVC) project. The distortion
models used in HEVC are mean squared error (MSE) and sum of absolute difference (SAD). However, they are
widely criticized for not correlating well with perceptual image quality. The structural similarity (SSIM) index
has been found to be a good indicator of perceived image quality. Meanwhile, it is computationally simple
compared with other state-of-the-art perceptual quality measures and has a number of desirable mathematical
properties for optimization tasks. We propose a perceptual video coding method to improve upon the current
HEVC based on an SSIM-inspired divisive normalization scheme as an attempt to transform the DCT domain
frame prediction residuals to a perceptually uniform space before encoding.
Based on the residual divisive normalization process, we define a distortion model for mode selection and show
that such a divisive normalization strategy largely simplifies the subsequent perceptual rate-distortion
optimization procedure. We further adjust the divisive normalization factors based on local content of the video
frame. Experiments show that the scheme can achieve significant gain in terms of rate-SSIM performance and
better visual quality when compared with HEVC
A Novel Approaches For Chromatic Squander Less Visceral Coding Techniques Usi...IJERA Editor
Recent advances in video capturing and display technologies, along with the exponentially increasing demand of
video services, challenge the video coding research community to design new algorithms able to significantly
improve the compression performance of the current H.264/AVC standard. This target is currently gaining
evidence with the standardization activities in the High Efficiency Video Coding (HEVC) project. The distortion
models used in HEVC are mean squared error (MSE) and sum of absolute difference (SAD). However, they are
widely criticized for not correlating well with perceptual image quality. The structural similarity (SSIM) index
has been found to be a good indicator of perceived image quality. Meanwhile, it is computationally simple
compared with other state-of-the-art perceptual quality measures and has a number of desirable mathematical
properties for optimization tasks. We propose a perceptual video coding method to improve upon the current
HEVC based on an SSIM-inspired divisive normalization scheme as an attempt to transform the DCT domain
frame prediction residuals to a perceptually uniform space before encoding.
Based on the residual divisive normalization process, we define a distortion model for mode selection and show
that such a divisive normalization strategy largely simplifies the subsequent perceptual rate-distortion
optimization procedure. We further adjust the divisive normalization factors based on local content of the video
frame. Experiments show that the scheme can achieve significant gain in terms of rate-SSIM performance and
better visual quality when compared with HEVC
H2B2VS (HEVC hybrid broadcast broadband video services) – Building innovative...Raoul Monnier
Broadcast and broadband networks continue to be separate worlds in the video consumption business. Some initiatives such as HbbTV have built a bridge between both worlds, but its application is almost limited to providing links over the broadcast channel to content providers’ applications such as Catch-up TV services. When it comes to reality, the user is using either one network or the other.
H2B2VS is a Celtic-Plus project aiming at exploiting the potential of real hybrid networks by implementing efficient synchronization mechanisms and using new video coding standard such as High Efficiency Video Coding (HEVC). The goal is to develop successful hybrid network solutions that enable value added services with an optimum bandwidth usage in each network and with clear commercial applications. An example of the potential of this approach is the transmission of Ultra-HD TV by sending the main content over the broadcast channel and the required complementary information over the broadband network. This technology can also be used to improve the life of handicapped persons: Deaf people receive through the broadband network a sign language translation of a programme sent over the broadcast channel; the TV set then displays this translation in an inset window.
One of the most important contributions of the project is developing and testing synchronization methods between two different networks that offer unequal qualities of service with significant differences in delay and jitter.
In this paper, the main technological project contributions are described, including SHVC, the scalable extension of HEVC and a special focus on the synchronization solution adopted by MPEG and DVB. The paper also presents some of the implemented practical use cases, such as the sign language translation described above, and their performance results so as to evaluate the commercial application of this type of solution.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
Low complexity video coding for sensor networkeSAT Journals
Abstract Modern video codecs such as H.264/AVC give state-of-the-art compression performance. However, extensive use of optimization tools makes them highly complex and hence not suitable for wireless video sensor network. In this paper an efficient video codec with substantially reduced complexity is proposed. Simulation result shows that the proposed video codec gives comparable compression performance compared to H.264/AVC but at substantially reduced computational complexity. Keywords—Low complexity coding, Sensor network, Video coding, Wavelet transform.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Fpga implementation of (15,7) bch encoder and decoder for text messageeSAT Journals
Abstract In a communication channel, noise and interferences are the two main sources of errors occur during the transmission of the message. Thus, to get the error free communication error control codes are used. This paper discusses, FPGA implementation of (15, 7) BCH Encoder and Decoder for text message using Verilog Hardware Description Language. Initially each character in a text message is converted into binary data of 7 bits. These 7 bits are encoded into 15 bit codeword using (15, 7) BCH encoder. If any 2 bit error in any position of 15 bit codeword, is detected and corrected. This corrected data is converted back into an ASCII character. The decoder is implemented using the Peterson algorithm and Chine’s search algorithm. Simulation was carried out by using Xilinx 12.1 ISE simulator, and verified results for an arbitrarily chosen message data. Synthesis was successfully done by using the RTL compiler, power and area is estimated for 180nm Technology. Finally both encoder and decoder design is implemented on Spartan 3E FPGA. Index Terms: BCH Encoder, BCH Decoder, FPGA, Verilog, Cadence RTL compiler
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
A REAL-TIME H.264/AVC ENCODER&DECODER WITH VERTICAL MODE FOR INTRA FRAME AND ...csandit
The video coding standards are being developed to satisfy the requirements of applications for
various purposes, better picture quality, higher coding efficiency, and more error robustness.
The new international video coding standard H.264 /AVC aims at having significant
improvements in coding efficiency, and error robustness in comparison with the previous
standards such as MPEG-2, H261, H263,and H264. Video stream needs to be processed from
several steps in order to encode and decode the video such that it is compressed efficiently with
available limited resources of hardware and software. All advantages and disadvantages of
available algorithms should be known to implement a codec to accomplish final requirement.
The purpose of this project is to implement all basic building blocks of H.264 video encoder and
decoder. The significance of the project is the inclusion of all components required to encode
and decode a video in MatLab .
Distributed Video Coding (DVC) has become increasingly popular in recent times among the researchers in video coding due to its attractive and promising features. DVC primarily has a modified complexity balance between the encoder and decoder, in contrast to conventional video codecs. However, Most of the reported DVC schemes have a high time-delay in decoder which hinders its practical application in real-time systems. In this work, we focus on speed up the Side Information(SI) generation module in DVC, which is a major function in the DVC coding algorithm and one of the time-consuming factor at the decoder. By applied it through Compute Unified Device Architecture (CUDA) based on General-Purpose Graphics Processing Unit (GPGPU), the experimental results show that a considerable speedup can be obtained by using the proposed parallelized SI generation algorithm.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
.
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERcscpconf
This paper proposes about motion estimation in H.264/AVC encoder. Compared with standards
such as MPEG-2 and MPEG-4 Visual, H.264 can deliver better image quality at the same
compressed bit rate or at a lower bit rate. The increase in compression efficiency comes at the
expense of increase in complexity, which is a fact that must be overcome. An efficient Co-design
methodology is required, where the encoder software application is highly optimized and
structured in a very modular and efficient manner, so as to allow its most complex and time
consuming operations to be offloaded to dedicated hardware accelerators. The Motion
Estimation algorithm is the most computationally intensive part of the encoder which is simulated using MATLAB. The hardware/software co-simulation is done using system generator tool and implemented using Xilinx FPGA Spartan 3E for different scanning methods.
Machine learning-based energy consumption modeling and comparing of H.264 and...IJECEIAES
Advancement of the prediction models used in a variety of fields is a result of the contribution of machine learning approaches. Utilizing such modeling in feature engineering is exceptionally imperative and required. In this research, we show how to utilize machine learning to save time in research experiments, where we save more than five thousand hours of measuring the energy consumption of encoding recordings. Since measuring the energy consumption has got to be done by humans and since we require more than eleven thousand experiments to cover all the combinations of video sequences, video bit rate, and video encoding settings, we utilize machine learning to model the energy consumption utilizing linear regression. VP8 codec has been offered by Google as a free video encoder in an effort to replace the popular H.264 video encoder standard. This research model energy consumption and describes the major differences between H.264/AVC and VP8 encoders based on of energy consumption and performance through experiments that are machine learning-based modeling. Twentynine uncompressed video segments from a standard data-set are used, with several sizes, details, and dynamics, where the frame sizes ranging from QCIF(176x144) to 2160p(3840x2160). For fairness in comparison analysis, we use seven settings in VP8 encoder and fifteen types of tuning in H.264/AVC. The settings cover various video qualities. The performance metrics include video qualities, encoding time, and encoding energy consumption.
Video Compression Algorithm Based on Frame Difference Approaches ijsc
The huge usage of digital multimedia via communications, wireless communications, Internet, Intranet and cellular mobile leads to incurable growth of data flow through these Media. The researchers go deep in developing efficient techniques in these fields such as compression of data, image and video. Recently, video compression techniques and their applications in many areas (educational, agriculture, medical …) cause this field to be one of the most interested fields. Wavelet transform is an efficient method that can be used to perform an efficient compression technique. This work deals with the developing of an efficient video compression approach based on frames difference approaches that concentrated on the calculation of frame near distance (difference between frames). The
selection of the meaningful frame depends on many factors such as compression performance, frame details, frame size and near distance between frames. Three different approaches are applied for removing the lowest frame difference. In this paper, many videos are tested to insure the efficiency of this technique, in addition a good performance results has been obtained.
A Novel Approaches For Chromatic Squander Less Visceral Coding Techniques Usi...IJERA Editor
Recent advances in video capturing and display technologies, along with the exponentially increasing demand of
video services, challenge the video coding research community to design new algorithms able to significantly
improve the compression performance of the current H.264/AVC standard. This target is currently gaining
evidence with the standardization activities in the High Efficiency Video Coding (HEVC) project. The distortion
models used in HEVC are mean squared error (MSE) and sum of absolute difference (SAD). However, they are
widely criticized for not correlating well with perceptual image quality. The structural similarity (SSIM) index
has been found to be a good indicator of perceived image quality. Meanwhile, it is computationally simple
compared with other state-of-the-art perceptual quality measures and has a number of desirable mathematical
properties for optimization tasks. We propose a perceptual video coding method to improve upon the current
HEVC based on an SSIM-inspired divisive normalization scheme as an attempt to transform the DCT domain
frame prediction residuals to a perceptually uniform space before encoding.
Based on the residual divisive normalization process, we define a distortion model for mode selection and show
that such a divisive normalization strategy largely simplifies the subsequent perceptual rate-distortion
optimization procedure. We further adjust the divisive normalization factors based on local content of the video
frame. Experiments show that the scheme can achieve significant gain in terms of rate-SSIM performance and
better visual quality when compared with HEVC
A Novel Approaches For Chromatic Squander Less Visceral Coding Techniques Usi...IJERA Editor
Recent advances in video capturing and display technologies, along with the exponentially increasing demand of
video services, challenge the video coding research community to design new algorithms able to significantly
improve the compression performance of the current H.264/AVC standard. This target is currently gaining
evidence with the standardization activities in the High Efficiency Video Coding (HEVC) project. The distortion
models used in HEVC are mean squared error (MSE) and sum of absolute difference (SAD). However, they are
widely criticized for not correlating well with perceptual image quality. The structural similarity (SSIM) index
has been found to be a good indicator of perceived image quality. Meanwhile, it is computationally simple
compared with other state-of-the-art perceptual quality measures and has a number of desirable mathematical
properties for optimization tasks. We propose a perceptual video coding method to improve upon the current
HEVC based on an SSIM-inspired divisive normalization scheme as an attempt to transform the DCT domain
frame prediction residuals to a perceptually uniform space before encoding.
Based on the residual divisive normalization process, we define a distortion model for mode selection and show
that such a divisive normalization strategy largely simplifies the subsequent perceptual rate-distortion
optimization procedure. We further adjust the divisive normalization factors based on local content of the video
frame. Experiments show that the scheme can achieve significant gain in terms of rate-SSIM performance and
better visual quality when compared with HEVC
H2B2VS (HEVC hybrid broadcast broadband video services) – Building innovative...Raoul Monnier
Broadcast and broadband networks continue to be separate worlds in the video consumption business. Some initiatives such as HbbTV have built a bridge between both worlds, but its application is almost limited to providing links over the broadcast channel to content providers’ applications such as Catch-up TV services. When it comes to reality, the user is using either one network or the other.
H2B2VS is a Celtic-Plus project aiming at exploiting the potential of real hybrid networks by implementing efficient synchronization mechanisms and using new video coding standard such as High Efficiency Video Coding (HEVC). The goal is to develop successful hybrid network solutions that enable value added services with an optimum bandwidth usage in each network and with clear commercial applications. An example of the potential of this approach is the transmission of Ultra-HD TV by sending the main content over the broadcast channel and the required complementary information over the broadband network. This technology can also be used to improve the life of handicapped persons: Deaf people receive through the broadband network a sign language translation of a programme sent over the broadcast channel; the TV set then displays this translation in an inset window.
One of the most important contributions of the project is developing and testing synchronization methods between two different networks that offer unequal qualities of service with significant differences in delay and jitter.
In this paper, the main technological project contributions are described, including SHVC, the scalable extension of HEVC and a special focus on the synchronization solution adopted by MPEG and DVB. The paper also presents some of the implemented practical use cases, such as the sign language translation described above, and their performance results so as to evaluate the commercial application of this type of solution.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
Low complexity video coding for sensor networkeSAT Journals
Abstract Modern video codecs such as H.264/AVC give state-of-the-art compression performance. However, extensive use of optimization tools makes them highly complex and hence not suitable for wireless video sensor network. In this paper an efficient video codec with substantially reduced complexity is proposed. Simulation result shows that the proposed video codec gives comparable compression performance compared to H.264/AVC but at substantially reduced computational complexity. Keywords—Low complexity coding, Sensor network, Video coding, Wavelet transform.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Fpga implementation of (15,7) bch encoder and decoder for text messageeSAT Journals
Abstract In a communication channel, noise and interferences are the two main sources of errors occur during the transmission of the message. Thus, to get the error free communication error control codes are used. This paper discusses, FPGA implementation of (15, 7) BCH Encoder and Decoder for text message using Verilog Hardware Description Language. Initially each character in a text message is converted into binary data of 7 bits. These 7 bits are encoded into 15 bit codeword using (15, 7) BCH encoder. If any 2 bit error in any position of 15 bit codeword, is detected and corrected. This corrected data is converted back into an ASCII character. The decoder is implemented using the Peterson algorithm and Chine’s search algorithm. Simulation was carried out by using Xilinx 12.1 ISE simulator, and verified results for an arbitrarily chosen message data. Synthesis was successfully done by using the RTL compiler, power and area is estimated for 180nm Technology. Finally both encoder and decoder design is implemented on Spartan 3E FPGA. Index Terms: BCH Encoder, BCH Decoder, FPGA, Verilog, Cadence RTL compiler
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
A REAL-TIME H.264/AVC ENCODER&DECODER WITH VERTICAL MODE FOR INTRA FRAME AND ...csandit
The video coding standards are being developed to satisfy the requirements of applications for
various purposes, better picture quality, higher coding efficiency, and more error robustness.
The new international video coding standard H.264 /AVC aims at having significant
improvements in coding efficiency, and error robustness in comparison with the previous
standards such as MPEG-2, H261, H263,and H264. Video stream needs to be processed from
several steps in order to encode and decode the video such that it is compressed efficiently with
available limited resources of hardware and software. All advantages and disadvantages of
available algorithms should be known to implement a codec to accomplish final requirement.
The purpose of this project is to implement all basic building blocks of H.264 video encoder and
decoder. The significance of the project is the inclusion of all components required to encode
and decode a video in MatLab .
Similar to IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME PREDICTION (20)
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
A review on techniques and modelling methodologies used for checking electrom...nooriasukmaningtyas
The proper function of the integrated circuit (IC) in an inhibiting electromagnetic environment has always been a serious concern throughout the decades of revolution in the world of electronics, from disjunct devices to today’s integrated circuit technology, where billions of transistors are combined on a single chip. The automotive industry and smart vehicles in particular, are confronting design issues such as being prone to electromagnetic interference (EMI). Electronic control devices calculate incorrect outputs because of EMI and sensors give misleading values which can prove fatal in case of automotives. In this paper, the authors have non exhaustively tried to review research work concerned with the investigation of EMI in ICs and prediction of this EMI using various modelling methodologies and measurement setups.
Low power architecture of logic gates using adiabatic techniquesnooriasukmaningtyas
The growing significance of portable systems to limit power consumption in ultra-large-scale-integration chips of very high density, has recently led to rapid and inventive progresses in low-power design. The most effective technique is adiabatic logic circuit design in energy-efficient hardware. This paper presents two adiabatic approaches for the design of low power circuits, modified positive feedback adiabatic logic (modified PFAL) and the other is direct current diode based positive feedback adiabatic logic (DC-DB PFAL). Logic gates are the preliminary components in any digital circuit design. By improving the performance of basic gates, one can improvise the whole system performance. In this paper proposed circuit design of the low power architecture of OR/NOR, AND/NAND, and XOR/XNOR gates are presented using the said approaches and their results are analyzed for powerdissipation, delay, power-delay-product and rise time and compared with the other adiabatic techniques along with the conventional complementary metal oxide semiconductor (CMOS) designs reported in the literature. It has been found that the designs with DC-DB PFAL technique outperform with the percentage improvement of 65% for NOR gate and 7% for NAND gate and 34% for XNOR gate over the modified PFAL techniques at 10 MHz respectively.
Online aptitude test management system project report.pdfKamal Acharya
The purpose of on-line aptitude test system is to take online test in an efficient manner and no time wasting for checking the paper. The main objective of on-line aptitude test system is to efficiently evaluate the candidate thoroughly through a fully automated system that not only saves lot of time but also gives fast results. For students they give papers according to their convenience and time and there is no need of using extra thing like paper, pen etc. This can be used in educational institutions as well as in corporate world. Can be used anywhere any time as it is a web based application (user Location doesn’t matter). No restriction that examiner has to be present when the candidate takes the test.
Every time when lecturers/professors need to conduct examinations they have to sit down think about the questions and then create a whole new set of questions for each and every exam. In some cases the professor may want to give an open book online exam that is the student can take the exam any time anywhere, but the student might have to answer the questions in a limited time period. The professor may want to change the sequence of questions for every student. The problem that a student has is whenever a date for the exam is declared the student has to take it and there is no way he can take it at some other time. This project will create an interface for the examiner to create and store questions in a repository. It will also create an interface for the student to take examinations at his convenience and the questions and/or exams may be timed. Thereby creating an application which can be used by examiners and examinee’s simultaneously.
Examination System is very useful for Teachers/Professors. As in the teaching profession, you are responsible for writing question papers. In the conventional method, you write the question paper on paper, keep question papers separate from answers and all this information you have to keep in a locker to avoid unauthorized access. Using the Examination System you can create a question paper and everything will be written to a single exam file in encrypted format. You can set the General and Administrator password to avoid unauthorized access to your question paper. Every time you start the examination, the program shuffles all the questions and selects them randomly from the database, which reduces the chances of memorizing the questions.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressions
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME PREDICTION
1. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
DOI:10.5121/ijma.2020.12401 1
IMPROVING PSNR AND PROCESSING
SPEED FOR HEVC USING HYBRID
PSO FOR INTRA FRAME PREDICTION
Swati Vinod Sakhare and Upena D. Dalal
Department of Electronics Engineering, The Sardar Vallabhbhai
National Institute of Technology, Surat, Gujarat, India
ABSTRACT
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors..
KEYWORDS
Fast Encoding, PSO, High Efficiency Video Coding, Quadtree-Based Coding Unit Partitioning.
1. INTRODUCTION
These days, video dissemination for different objects is multiplying over the Internet with the
guides of helpful correspondence systems and brilliant cell phones. Furthermore, video buyers
progressively request top quality (HD) and ultra-top notch (UHD) recordings to encounter better
visual quality. Accordingly, the conveyance of HD/UHD video to the cell phone clients over the
Internet is turning into a well-known pattern. Be that as it may, the information amount for
HD/UHD video is tremendous because of the higher video goals and edge rate. The information
size of a 10-second video with 3840 × 2160 goals at an edge pace of 60casings for each
second arrives at almost 15 GB. Because of this, the conveyance of HD/UHD
video requests a bigger measure of system transmission capacity and information stockpiling
contrasted with the lower goals standard definition (SD) recordings.
With respect to saving money on organize assets and capacity prerequisite, a productive
pressure system is vitally significant. Joint Collaborative Team on video coding (JCT-
VC),the communitarian venture gathering of ITU-T Video Coding Experts Group (VCEG) and
the ISO/IEC Moving Picture Expert Group (MPEG), has executed a profoundly proficient
videocoding standard called High Efficiency Video Coding (HEVC)/H.265 [1] as an answer for
the issue of expanded video goals. ITU-T and ISO/IEC are the principle institutionalization
2. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
2
bodies which have institutionalized all HEVC’s predecessor guidelines in numerous years.
They have used a 16 × 16 macro block as an essential handling unit in HEVC’s precursor. Each
edge is part into macro blocks. Each macro block includes one 16 × 16 square of lumaparts to
speak to the splendour and two 8 × 8 squares of chroma segments to allude the shading in the
4:2:0 chrominance sub sampling groups. Along these lines, the macro block is the
biggest square size to demonstrate the anticipated data of intra-outline or between outline
expectation in past video coding norms. In any case, ordinary HD and UHD recordings have
numerous bigger edge areas than the macro block, and those districts can speak
to the equivalent moving data. On the off chance that the macroblock is utilized as an essential
preparing unit for regular HD and UHD recordings, a lot of bits are important to flag the
expectation data. Correspondingly, the change square size is bigger than the macro block size.
Consequently, HEVC bolsters a bigger square size as a fundamental handling unit called CTU
for intra-outline or between outline expectation and change coding. Albeit a huge square size is
sufficient for high goals video, it’s anything but a decent decision for low goals video. To be
good with both high-and low-goals recordings, HEVC can deftly parcel the video outline into a
few square CTUs of 2L×2 L tests, where L ∈ {4, 5, 6}. The encoder deftly picks are asonable
estimation of L for proposed application to have the best ex- change off between coding
execution and cost, for example, memory stockpiling, encoding time, and
postponement. Be that as it may, utilizing bigger square for choosing whether intra-mode or
between mode at the forecast stage can’t ensure to get a decent RD execution for expectation
organize. To accomplish better coding proficiency, HEVC presented another essential
handling unit, called CU and an adaptable quadtree apportioning from CTU to CU. Thusly,CU
size can be 64× 64, 32 × 32, 16 × 16, and 8 × 8 at profundity 0, profundity 1, profundity2, and
profundity 3,individually. To characterize CU size or profundity, HEVC begins a preliminary
encoding which incorporates two primary capacities called the RD cost estimation and
correlation in top-down and base up way, individually, as referenced in Section I. In the
top-down RD cost computation of a 64× 64 CTU, the RD costs for all conceivable 85 CUs
are determined in a preorder traversal of the quadtree, if the greatest CU profundity is 3. In
subtleties, there are 1, 4, 16, and 64 CUs at profundity 0, profundity 1,profundity 2, and
profundity 3, separately, and the complete number of CUs is P3 i=0 4 I = 85CUs. Subsequent
to computing the RD costs for four kids CUs of each parent CU, HEVCgoes to the RD cost
correlation with choose whether a parent CU is part or not by looking at the RD cost of parting
and non-parting states of parent CU. At that point, HEVC changes to the RD cost estimation or
performs correlation once more, contingent upon the situation of parent CU. In this way, there
are 85 computations and 21 examinations in the top-down RD cost estimation and base up RD
cost correlation of a 64 × 64 CTU, individually. After at long last looking at a root CU at
profundity 0 with its four kids CUs at profundity 1, the best CU quadtree structure of a CTU
with the least RD cost is picked among 83,522 potential quadtree structures. The CU dividing
examples of casing portrayal of picture request tally(POC) 40 of grouping ”Blowing
Bubbles” looked by a thorough RDO search of HM variant16.5 (HM16.5). The preliminary
encoding of HEVC finds the best CU segment structure of each CTU after a
comprehensive RDO search. In this manner, picking an ideal CU apportioning
structure can be demonstrated as an advancement issue and can be discovered an answer by a
softly appropriate enhancement apparatus to look through a space of conceivable CU segment
arrangements. For little space advancement process, customary thorough procedures
are suitable to discover the arrangement [20]. In any case, the procedures dependent
on man-made consciousness (AI) are productive for a tremendous inquiry space and PSO is one
of AI systems to look through a decent arrangement proficiently.
The major objectives of this text are,
3. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
3
To design a reduced complexity intra-frame predictor using soft computing
To design a reduced complexity intra-frame predictor using soft computing
To optimize the time complexity of this predictor using cellular automata rules
To integrate the reduced time complexity and reduced computational
complexity predictor into HEVC encoding and decoding process.
The next section describes various techniques for improving HEVC performance followed by
the proposed predictors. This text concludes with some interesting observations about our
results and some recommendations that can be researched to further optimize the prediction
performance.
1.1 DUAL TREE COMPLEX WAVELET TRANSFORM
DT-DWT is the advanced design of DWT. Unlike DWT, DTDWT can obtain better shift
invariance and directional selectivity [21, 22]. DTDWT is also known as complex transform
since it includes the real and imaginary part of six oriented wavelet coefficients. Figure 1 shows
the filter band tree structure of the DTDWT. As shown in the figure, the top tree R generates the
real parts of the DTDWT coefficient and bottom tree I generates the imaginary parts of the DT-
DWT coefficient. * denotes a convolution operation, ↓2 means a down sampling by 2. lfr and hfr
are low-pass filter and high-pass filter, which form a Hilbert transformed pair to insure the
perfect reconstruction of the discrete wavelet transform. With the filter band tree structure, the
following wavelet sub-bands which oriented at ±75°, ±150
and ±450 are produced.
Figure1. Filter band Tree Structure of the DT-DWT
y
x
y
x
,
1 (LH Wavelet)
y
x
y
x
,
2
(HL Wavelet)
y
x
y
x
,
3
(HH Wavelet)
From the above equations,
x
and
y
represent the low pass filter along with first and second
dimension. Similarly,
x
and
y
represent the high pass filter along with first and second
dimension. Also, LH and HL sub-bands are oriented at vertical and horizontal directions
respectively. The HH sub-band is simultaneously oriented along the +45° and−45° diagonal
directions. It is also denoted as HHp (positive oriented direction) and HHn (negative oriented
4. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
4
direction). Similarly, two LH sub-bands and two HL sub-bands are denoted as LHp, LHn, HLp and
HLn. With these six oriented sub-bands, best angular mode is estimated among the 35 angular
modes for intra prediction process.
2. RELATED WORK
HEVC accomplishes the bit rate sparing of almost half under the equivalent visual quality
contrasted with the H.264/AVC. Thus, HEVC turns into a famous video codec. Variants of
HEVC like H.264, H.265, and others are iterations over the base HEVC codec that improve its
efficiency by adding computationally optimum algorithms during the inter-frame and
intra-frame prediction conditions. Researchers from different fields including video
processing specialists, Mathematicians, signal processing experts, to name a few have forayed
into this field in order to further optimize the efficiency of HEVC processing. For instance the
work in [2] adds watermarking capabilities to HEVC by matrix encoding in the DCT (discrete
cosine transform) block of HEVC. This paper achieves data hiding with minimum distortions in
the output video. This indicates that HEVC has some inherent redundancies which can be
reduced in order to further optimize the video encoding/decoding performance.
These redundancies are in terms of inter and intra-frame prediction co-efficient values. This basic
property paves as the motivation for the work in this paper. A motion density-based scheme with
unequal error protection (UEP) is proposed in [3], wherein it is seen that motion density schemes
outperform the existing inter-frame & intra-frame prediction schemes of HEVC.This
performance is evaluated in terms of the capability of the algorithm to find out important frames
from the input video. A higher value is an indicative of better performance for the system. The
proposed approach in [3] outperforms other control unit (CU) based strategies by more than10%.
This approach can be used to evaluate the best quality frames or key-frames.These key-frames
serve as the base-line for inter and intra-prediction in HEVC. Another approach similar to [2], but
directed towards H.265 codec is proposed in [4]. Wherein, the synchronization error is
reduced after two-stage re-compression in H.265 codec. This approach uses spatial
texture analysis for finding out the most suitable embedding blocks. These blocks are then used
in representation mode in order to find out the best pixels for watermarking. The identified
pixels contain some level of redundancy, and thus can be reduced (compressed) without
any significant loss in video quality. This can improve the frame rate and the efficiency of the
HEVC system when operating in the H.265 mode. It is seen that the proposed algorithm performs
well in the presence of any kind of noise, and there by can be used for further redundancy
reduction of HEVC.
The work in [5] is inspired by these approaches in [2,4], and uses a concept of just notable
distortion (JND). A good quality video can be encoded and decoded using the JND concept.
These identified JND points in the video frames can reduce the size of HEVC data by more than
13% on average, and up-to 39% for certain video sequences. The mean opinion score(MOS) was
evaluated for different videos, and it is observed that the approach in [5] has similar performance
to original HEVC algorithm in terms of visual quality, but it has a reduced compressed video
size. A similar work is proposed in [6], wherein the concept of classical secretary problem (CSP)
is used in the rough-mode-decision module of HEVC.Moreover, the CSP is modified using a
dynamic stopping criterion that further enhances the performance by reducing the encoding
delay and marginally increasing the bit-rate performance. It uses the concept of mode
reduction with the help of redundancy evaluation. A similar concept is proposed in [2], [4] & [5],
and is also the base for this research.
HEVC can be extended to 3D videos. The concept of fast-depth map for intra-mode selection in
3D videos is given in [7], wherein the depth is analysed from the different dimensions of the 3D
5. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
5
video. This depth map is used for prediction of intra-mode redundancies, and finally a
compressed video is obtained. Various depth modelling models are proposed in [7], some of them
also use tensor features for homogeneity detection. Due to the use of depth maps, there is a large
reduction in encoding delay, which further improves the encoding and decoding performance for
3D videos. The approach in [7], can further utilize deep learning methods like deep neural
networks as proposed in [8] to optimize its performance. It is observed from[8], that deep net
models like convolution neural networks (CNN) can be trained with different videos to
identify the redundancies in them. This trained model can then be applied to new videos to
optimize their redundancies with minimal computational complexity and improved bandwidth.
They further observe that specialized models like IPCNN can be trained to specifically reduce the
intra-frame redundancies in order to optimize the quality of service(QoS) for HEVC. An
approach that can be facilitated by CNNs is proposed in [9], where in metrics like rate distortion
are evaluated to reduce the complexity of encoding and decoding process. They have used
texture homogeneity between inter-frames and spatio-temporalcorrelation between intra-
frames in order to reduce the encoding time by more than 70% than normal HEVC. Though the
results seem promising, it is advised that researchers perform adue diligence before using this
research in their applications. The work proposes development of fast coding unit and fast
prediction unit in order to improve the efficiency of the HEVC system. While most of these
research models are based on lossy HEVC performance improvement, the work in [10]
uses lossless HEVC using context-based angular & planarintra predictions. It also uses
redundancy reduction in HEVC videos by identifying redundant edges, textures, colours, and
other parameters between neighbouring pixels. They use pixel-level processing for edge and
texture redundancy optimization without increasing any computational complexity. Due to
removal of redundant edges and textures, the resulting video is completely lossless. It can achieve
performance improvement of up to 10% when compared with other standard HEVC models.
Another 3D video optimization algorithm is mentioned in [11], that uses
dynamically configurable depth maps similar to [7]. In [10] the depth maps are not generated
using tensors or hyper-planes, but they use the concept of Rough Mode Decision (RMD). It is
known that RMD is inherited from the texture maps, rather than the depth maps. This RMD
affects the block distortion and the rate distortion of HEVC, and thus can be used for better
HEVC performance. The proposed work achieves 0.1% improvement in Bjontegaard Delta-rate
(BDRate), which indicates that the compression performance is high when compared to normal
HEVC encoding. A similar method like [11] is given in [12], wherein methods like bipartition
modes, intra-picture skip, and DC-only are used to optimise depth map processing. Their work
indicates that depth map processing to identify redundancies using these approaches can reduce
the encoding delay by more than 20%. They also propose that reduction in texture and depth can
be combined to further improve the HEVC performance.
A fast and adaptive mode decision HEVC algorithm can be seen in [13], which uses coding unit
partition for early termination of intra-prediction process. The proposed work in [13]forms the
base our work, wherein this paper also utilizes mode reduction technique similar to[13] for a
better HEVC performance. They have reduced the number of modes from 35 to 11,which
improve the Bjontegaard delta rate by 1.7%, but reduce the average delay by more than50%.
Thus, giving a big bump in terms of final video performance. The work in [13] also utilizes CU
partitioning based on number of coding bits, which further helps in improving the system
performance. This work can further be improved by addition of RD cost as a measure for mode
reduction as proposed in [14]. RD cost can be an early prediction metric for reducing the
number of intra modes from HEVC. Due to inclusion of RD cost in evaluation of mode
reduction, a performance improvement of more than 25% can be expected when
compared to usual HEVC system, which can further be improved by adding machine learning
mechanisms like the one proposed in [15] for adaptive CU size decisions. The work
6. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
6
i[15]proposes the use of complexity classification for training the machine learning model. This
complexity classification method uses parameters like CU size, CU partitions and
rate distortion to train a support vector machine (SVM) algorithm. This SVM algorithm solves
a2-class classification problem, and classifies each intra-frame into required and non-required.
All the non-required frames are dropped, and finally we get the compressed video with minimal
complexity. The proposed ML algorithm reduces the complexity by more than 60%,and thereby
speeding up the entire process of HEVC compression and decompression. The approach in [15]
can further be modified using the techniques mentioned in [16]. From there view done in [16], we
can observe that dynamic support vector machines (DSVMs), which can be destroyed and re-
created for every inter-frame and intra-frame model prediction are the best option for HEVC
encoding. These models must be integrated with existing HEVC approaches to further improve
their efficiency. The SVM models can also be used as the final flat layer for CNN models
described in [17]. This replacement can enhance the performance of the existing CNN models by
more than 20%, and also reduce the complexity of processing the HEVC videos. Moreover, the
CNNs can be replaced by deep CNNs, as proposed in [18]to further reduce the intra-mode
redundancies. These redundancies are easily analysed by deep CNNs, and thereby can be further
reduced with the help of models like GoogLe Net or VG GNet. A combination of layers
like convolutional, ReLU, convolutional, ReLU,convolutional, Max Pooling, fully
connected, ReLU and finally fully connected can be used for a better prediction performance.
Another experimental work is described in [19], where in a SAD unit is proposed to compress
ultra HD 8K videos. This can be used as a future work for deep CNN models. The next section
describes our proposed PSO-based approach for HEVC processing
3. PROPOSED PSO COMBINED WITH CELLULAR AUTOMATA
MODEL FOR INTRA FRAME PREDICTION
PSO is applied inside the intra-frame prediction process in order to optimize the PSNR at the
decoding side. This model requires a certain amount of delay for the first one or two searches, but
it is compensated as the number of frames are processed. Due to this self-learning nature of the
algorithm, it can be integrated inside the intra-frame prediction block of HEVC. The CA
technique further optimizes the performance of the existing PSO. It does so by reducing the
randomized search space of the PSO via rules of CA. The application of CA to PSO is done with
the help of the following rules in CA.
Figure 2. CA rules
7. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
7
From the figure 2, we can observe that based on the input patterns the CA selects one output.
This output in our case is the limited range of global and local best particles. Using CA the range
of the local and global best particles is limited. The following algorithm describes the working of
PSO and CA with HEVC.
Initialize the PSO and CA parameters, namely,
o Number of particles = P
o Number of iterations = I
o PSO constants C1 and C2
o CA_LIMIT_PBEST – Is the limited number of HEVC modes to be used for
learning from PBEST (should be less than half modes)
o CA_LIMIT_GBEST – Is the limited number of HEVC modes to be used for
learning from GBEST (can be more than half modes)
Let the structure for the 64x64 CTU block be defined as follows,
Figure 3. Block division process
Here ‘a’ is the main block, ‘b0 … b3’ are the divided blocks, and ‘c0 … c15’ are the sub-
divided blocks and finally ‘d0 … d63’ are the 64 CTU blocks
Let’s call this combination a particle, and in each solution generate random particles for
operation
Considering the 3-level depth as shown in the figure, the particle will have 21 bits, as
follows,
P = a0,b0,b1,b2,b3,c0,c1,c2…c15
where,
a0 = 0, when CU is not split, else a0=1
bi= 0, when a0=0 or CU is not split, else bi=1
ci=0, when a0=0, corresponding b=0, or CU is not split, else ci=1
Here a, b, and c represent the splitting decisions for depth 0, depth 1, and depth 2,
respectively
The possible values for a is 0 (non-splitting) and 1 (splitting). The possible values for b
are null if a is 0, 0 (non-splitting), and 1 (splitting). The possible values for c are null if
its corresponding parent b is 0, 0 (non-splitting), and 1 (splitting). It should be noted that
the proposed data structure is composed of a group of dependent genes. Therefore, the
total number of possible CU partitioning patterns P is calculated as,
where d ∈ {1, 2, 3} is the maximum CU depth and the mod is the modulo operation for
finding the remainder. If the maximum CU depth is 2 and 3, the total number of possible
partitioning patterns is only 17 as shown in the following figure, and 83,522 even there
are five genes and 21 genes to represent the CU partitioning pattern of a 64 × 64 CTU,
respectively.
8. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
8
Figure 4. Frame division process
Initialize the particles using randomized values of a, b and c (from the figure 4 shown in
the previous step)
For each iteration perform the following,
o For each particle, find the fitness of the particle using the following equation,
where F is the RD cost-oriented fitness function, a, bi, and cj are the values of one gene,
four genes, and sixteen genes of each chromosome to represent the splitting decision at
depth 0, 1, 2, and 3, respectively. RDCost(a), RDCost(bi), RDCost(cj), and RDCost(dk)
are the RD cost values of one CU, 4 CUs, 16 CUs, and 64 CUs at depth 0, 1, 2, and 3,
respectively.
o If the fitness value is better than the best fitness value (pBest) in history, then set current
value as the new pBest
o Choose the particle with the best fitness value of all the particles as the gBest
o For each particle follow the given steps,
Evaluate the velocity of the particle using the following equation,
v = v + C1 * random (CA_LIMIT_PBEST) * (pBest - currentFitness) + C2 * random
(CA_LIMIT_GBEST) * (gBest - currentFitness)
Update the position of the particle using the following equation,
presentParticle = presentParticle + v
9. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
9
At the end of the last iteration, use the particle with gBest fitness value as intra-frame
prediction particle.
The best particle is replaced in the output stream of HEVC as the encoded block. Once the
particle selection is done then a dual-tree complex wavelet transform block is used in order to
reduce the modes of the system from 35 modes to 8 modes. Initially, Dual Tree Discrete Wavelet
Transform (DT-DWT) [25] is applied to the optimum selected block by PSO algorithm. By
applying this transform, six oriented wavelet sub-bands are generated. Among the sub-bands, two
LH and two HL sub-bands are used find the direction or angle of the texture in a block. Polarity
of the texture angle is estimated with two HH Sub-bands. With the direction and the angle of the
texture, a mode is determined that is closer to the actual best mode. For best mode selection, the
four modes around this determined mode (Modedeter) and also DC and planar modes are
considered as a final candidate list. This candidate list is forwarded to the process of Rate
Distortion Optimization (RDO). With the RDO, the mode with minimum rate distortion cost is
selected as the best mode (Modebest). Based on this selected Modebest , the encoder encodes the
video frames. Decoder decodes it and output of the decoder is analyzed for performance.
The flow diagram for the proposed system can be observed in the figure 5. From the figure we
can observe that the output of PSO and CA system is given to the DT-CWT based HEVC
encoder thereby hybridizing the system with the already existing high efficiency encoder.
Moreover, the output of the decoder is used for performing result evaluation of PSNR and delay.
These outputs and their analysis are showcased in the next section.
Due to the combination of the proposed intra-frame prediction model with the dual tree complex
wavelet transform (DTCWT), the overall effectiveness of the system is improved. The combined
model is able to reduce the search space, and also reduce the number of modes needed for
encoding. Thereby, giving a dual level advantage to the system under test.
Figure5. Overall flow of the system
Usually rate distortion optimization (RDO) is done in DTCWT. In this work, the RDO is not
done, but the RD cost is evaluated. We would request readers not to get confused between the
two processes. As PSO is only using RD cost in order to select the best blocks for
encoding/decoding, while DTCWT uses RD cost for optimization with the help of mode
reduction.
10. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
10
4. RESULT EVALUATION
In this paper, HEVC intra prediction algorithm using PSO as well as PSO with CA is proposed.
The results are compared with [25] which use dual tree complex wavelet transform for intra
prediction. Simulation is carried out in JAVA Net Beans Software. The results were compared
for delay and PSNR values for different videos. These results were compared for HEVC DT-
CWT [25] , HEVC with PSO and HEVC with PSO+CA as shown in Table1 which showcases
the delay results obtained for the videos on the given algorithms. Author in [25] proposes a novel
approach to reduce the modes from 35 to 8 and then selecting a optimum mode. We have
proposed PSO with CA to reduce time consumption in decision making process along with the
further application of dual tree complex wavelet transform to reduce the
computational complexity. Similarly, a comparison of PSNR for these algorithms was performed,
and the results are tabulated in Table 2
.
Table 1. Comparison of Time delays for the proposed algorithm
From the results we can observe that the proposed algorithm is able to reduce the delay and
improve the PSNR of the existing dual-tree based HEVC system. We also evaluated the average
values of PSNR and delay for both the algorithms, and observed the following results.
11. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
11
Table 2. Comparison of PSNR for algorithms
The average results of the algorithm are tabulated in table 3.
Table 3. Average results Time Delays and PSNR of the algorithm
Comparison of Computational complexity in terms of time delay as well as video quality in terms
of PSNR are tabulated in Table 4 for different testing video sequences. From the
comparison table, it is very clear that both the proposed algorithm that is HEVC processing using
PSO as well as using PSO and CA provides better video quality and reduction in time
complexity. The conclusion and some interesting observations from these results are mentioned
in the next section.
Table 4. Comparison among stat-of-the-art algorithms
12. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
12
Figure 6. Comparison of PSNR Results
As shown in Figure 6, it is clear that proposed PSO and PSO with CA provides almost similar
video quality which is comparatively better to the existing algorithm using dual tree complex
wavelet transform.
Figure 7. Comparison of Time Delay Results
13. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
13
As shown in the Figure 7, proposed PSO with CA algorithm provides better time complexity
reduction. The conclusion and some interesting observations from these results are mentioned in
the next section
5. CONCLUSION AND FUTURE WORK
The increased computational complexity in HEVC is a major problem especially for power
constrained devices or real-time applications especially for high-resolution videos. Therefore, it is
highly desirable to optimize the encoding process for computational complexity reduction while
maintaining the coding efficiency of HEVC. Fast intra prediction algorithm using PSO with CA
is proposed in this paper. The experimental results are conducted for various test video. The
results are evaluated based on encoding time and peak signal to noise ratio (PSNR). The
proposed PSO+CA based HEVC performs faster than the existing HEVC algorithm in terms of
overall delay by 40%. It also outperforms the existing method by 12 % in terms of PSNR. The
results of comparative experiments demonstrate that the proposed algorithm can no doubt
effectively reduce the computational complexity of HEVC Encoder while maintaining good
video quality.
All these advantages are evident due to the extensive intra-frame prediction phase, where in most
of the mapping process and calculations are pre-dominantly done. Another reason for such a huge
bump in performance is the presence of the light weighted execution phase.
There are many other ways to explore in the CU early termination, mode reduction and fast
intraprediction in the intra prediction area as suggested by literature. In future, many of these
methods can be combined, or if needed, one method may be replaced by a new method and
encoding time gains can be explored. Convolution Neural Network Model, SVM machine
learning approach can also be applied in order to reduce time complexity. Similar Intra
Prediction algorithms can be developed for fast inter-prediction resulting in lessen coding time
and reduced complexity. Future research can be conducted to reduce computational
complexity in Quad tree structure means dividing CTUs up to CU and PU both for the intra and
inter coding can be improved to obtain much higher reduction of encoding time, better bit rate
and PSNR. The aim should be to reduce the overall complexity of HEVC encoder suitable for
hand held devices as well as transmission with limited computing resources.
In future this work can be further improved by evaluating the performance for higher bit rate
videos. These videos are a bit complex to map, and thus might be a need of multiple pre –
execution steps before a required level of efficiency is achieved. Moreover, in order to really
optimize the performance further, researchers can use quantum computing for processing, and
develop quantum computational layers in order to evaluate its performance, and apply the
proposed machine.
REFERENCES
[1] G. Sullivan et al., “Overview of the high efficiency video coding (HEVC) standard”, IEEE
Transactions on Circuits and Systems for Video Technology, Vol. 22, No. 12, pp 1649-1668,
December 2012.
[2] T. Wiegand, G. J. Sullivan, G. Bjontegaard, and A. Luthra, ‘‘Overview of the H.264/AVC video
coding standard,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, no. 7,
pp. 560–576, Jul. 2003.
[3] JCT-VC, HM Software. Accessed: Nov. 5, 2016. [Online]. Available:
https://hevc.hhi.fraunhofer.de/svn/svn_HEVCSoftware/tags/HM-16.5/
14. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
14
[4] S. Cho and M. Kim, ‘‘Fast CU splitting and pruning for suboptimal CU partitioning in HEVC intra
coding,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 23, no. 9, pp. 1555–
1564, Sep. 2013.
[5] B. Min and R. C. C. Cheung, ‘‘A fast CU size decision algorithm for the HEVC intra encoder,’’ IEEE
Transactions on Circuits and Systems for Video Technology, vol. 25, no. 5, pp. 892–896, May 2015.
[6] T. Li, M. Xu, and X. Deng, ‘‘A deep convolutional neural network approach for complexity reduction
on intra-mode HEVC,’’ in Proc. IEEE Int. Conf. Multimedia Expo (ICME), Jul. 2017, pp. 1255–
1260.
[7] T. Zhang, M.-T. Sun, D. Zhao, and W. Gao, ‘‘Fast intra-mode and CU size decision for HEVC,’’
IEEE Transactions on Circuits and Systems for Video Technology, vol. 27, no. 8, pp. 1714–1726,
Aug. 2017.
[8] L. Shen, Z. Liu, X. Zhang, W. Zhao, and Z. Zhang, ‘‘An effective CU size decision method for
HEVC encoders,’’ IEEE Trans. on Multimedia, vol. 15, no. 2, pp. 465–470, Feb. 2013.
[9] X. Hou and Y. Xue, ‘‘Fast coding unit partitioning algorithm for HEVC,’’ in Proc. IEEE Int. Conf.
Consum. Electron. (ICCE), Jan. 2014, pp. 7–10.
[10] I. Zupancic, S. G. Blasi, E. Peixoto, and E. Izquierdo, ‘‘Inter-prediction optimizations for video
coding using adaptive coding unit visiting order,’’ IEEE Trans.on Multimedia, vol. 18, no. 9, pp.
1677–1690, Sep. 2016.
[11] X. Shen, L. Yu, and J. Chen, ‘‘Fast coding unit size selection for HEVC based on Bayesian decision
rule,’’ in Proc. of Picture Coding Symp., May 2012, pp. 453–456.
[12] K. Duan, P. Liu, K. Jia, and Z. Feng, ‘‘An adaptive quad-tree depth range prediction mechanism for
HEVC,’’ IEEE Access, vol. 6, pp. 54195–54206, 2018.
[13] X. Shen and L. Yu, ‘‘CU splitting early termination based on weighted SVM,’’ EURASIP Journal
proceedings of Image and Video Processing, vol. 2013, no. 1, p. 4, Dec. 2013.
[14] G. Corrêa, P. A. Assuncao, L. V. Agostini, and L. A. da Silva Cruz, ‘‘Fast HEVC encoding decisions
using data mining,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 25, no.
4, pp. 660–673, Apr. 2015.
[15] Y. Zhang, S. Kwong, X. Wang, H. Yuan, Z. Pan, and L. Xu, ‘‘Machine learning-based coding unit
depth decisions for flexible complexity allocation in high efficiency video coding,’’ IEEE
Transaction on Image Processing, vol. 24, no. 7, pp. 2225–2238, Jul. 2015.
[16] A. Heindel, T. Haubner, and A. Kaup, ‘‘Fast CU split decisions for HEVC inter coding using support
vector machines,’’ in Proc. Picture Coding Symp. (PCS), Dec. 2016, pp. 1–5.
[17] L. Zhu, Y. Zhang, S. Kwong, X. Wang, and T. Zhao, ‘‘Fuzzy SVM-based coding unit decision in
HEVC,’’ IEEE Transaction on Broadcasting, vol. 64, no. 3, pp. 681–694, Sep. 2018.
[18] H.-S. Kim and R.-H. Park, ‘‘Fast CU partitioning algorithm for HEVC using an online-learning-
based Bayesian decision rule,’’ IEEE Transactions on Circuits and Systems for Video Technology,
vol. 26, no. 1, pp. 130–138, Jan. 2016.
[19] L. Zhu, Y. Zhang, Z. Pan, R. Wang, S. Kwong, and Z. Peng, ‘‘Binary and multi-class learning based
low complexity optimization for HEVC encoding,’’ IEEE Transaction on Broadcasting, vol. 63, no.
3, pp. 547–561, Sep. 2017.
[20] M. Xu, T. Li, Z. Wang, X. Deng, R. Yang, and Z. Guan, ‘‘Reducing complexity of HEVC: A deep
learning approach,’’ IEEE Transaction on Image Processing, vol. 27, no. 10, pp. 5044–5059, Oct.
2018.
[21] Z. Pan, S. Kwong, M.-T. Sun, and J. Lei, ‘‘Early MERGE mode decision based on motion estimation
and hierarchical depth correlation for HEVC,’’ IEEE Transaction on Broadcasting, vol. 60, no. 2, pp.
405–412, Jun. 2014.
[22] Z. Pan, J. Lei, Y. Zhang, X. Sun, and S. Kwong, ‘‘Fast motion estimation based on content property
for low-complexity H.265/HEVC encoder,’’ IEEE Transaction on Broadcasting, vol. 62, no. 3, pp.
675–684, Sep. 2016.
[23] M. Mitchell, An Introduction to Genetic Algorithms. Cambridge, MA, USA: MIT Press, 1996.
[24] Z. Michalewicz, Genetic Algorithms+Data Structures=Evolution Programs, 3rd ed. Berlin, Germany:
Springer-Verlag, 1996.
[25] Farhad Pakdaman, Mahmoud-Reza Hashemi, and Mohammad Ghanbari, “Fast and Efficient Intra
Mode Decision for HEVC, Based on Dual-Tree Complex Wavelet,” in Proc. Of Article in
Multimedia Tools and Applications , May 2016
[26] Romberg J-K, Hyeokho C, Baraniuk R-G (2001) Multiscale edge grammars for complex wavelet
transforms. Proc. International Conference on Image Processing (ICIP), p. 614–617
15. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
15
[27] Selesnick I-W, Baraniuk R-G, Kingsbury N-C (2005) The dual-tree complex wavelet transform.
IEEE Signal Process Mag 22(6):123–151
AUTHORS
Swati Vinod Sakhare
She is a PhD scholar at The Sardar Vallabhbhai Patel National Institute of
Technology (SVNIT), Surat. She received her graduate degree in Electronics and
Telecommunication from Government College of Engineering, Amravati and
postgraduate degree in Electronics from B. D. College of Engineering Sevagram,
Wardha. Her research interests include Video processing, Digital Image Processing,
Microprocessors and Optical Fiber Systems. She is working in G. H. Raisoni College
of Engineering and Management, Amravati since 2012. She has guided several PG
students for their project work and coordinated conference, STTPs, workshops and
several technical programs for students, faculty members and working professional.
Dr. Mrs. Upena D. Dalal
She is currently Professor at Electronics and Communication Engineering department,
at The Sardar Vallabhbhai Patel National Institute of Technology (SVNIT), Surat. She
received her doctoral degree in era of wireless communication and post graduation in
Electronics and Communication Systems in which she was gold medalist. She has
vast teaching experience of 24 years and her major subjects are Cellular Technology,
Wireless Communication and Fiber optic networks. She is author of well-known book
on “Wire- less Communication” by Oxford University press (2009) and “Selectively
Imposed Pilot based Channel Estimation” by VDM publications, Germany (2010).
She is also co-author of book on “WiMAX Developments” from Intec web, Viena,
Austria (2010). She has guided several PhD and MTech students and coordinated
conference, STTPs, workshops and numerous technical programs for betterment of
students, faculty and research fraternity