High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors.
This document summarizes a research paper that proposes using genetic algorithms and differential evolution to optimize the interleaver design for turbo codes. Turbo codes achieve performance close to the theoretical limit by using parallel concatenation of recursive systematic convolutional codes. The interleaver permutes the input bits, which affects turbo code performance. This research aims to use evolutionary algorithms to find higher performing turbo code interleavers compared to conventional designs. It compares the proposed approaches to the traditional genetic algorithm method and finds that differential evolution performs well for optimizing turbo code interleaver design.
1) The document describes a modification to the Huffman coding used in JPEG image compression. It proposes pairing each non-zero DCT coefficient with the run-length of subsequent (rather than preceding) zero coefficients.
2) This allows using separate optimized Huffman code tables for each DCT coefficient position, improving compression by 10-15% over standard JPEG coding.
3) The decoding procedure is not changed and no end-of-block marker is needed, providing advantages with no increase in complexity.
This document analyzes the RC4 encryption algorithm and examines how its performance is affected by changing parameters like encryption key length and file size. Experimental tests were conducted to measure encryption time for different key lengths and file types. The results show encryption time increases with longer keys and larger files, and are modeled mathematically. The document also provides background on encryption methods, how RC4 works, and compares stream and block ciphers.
Hardware implementation of (63, 51) bch encoder and decoder for wban using lf...ijitjournal
Error Correcting Codes are required to have a reliable communication through a medium that has an
unacceptable bit error rate and low signal to noise ratio. In IEEE 802.15.6 2.4GHz Wireless Body Area
Network (WBAN), data gets corrupted during the transmission and reception due to noises and
interferences. Ultra low power operation is crucial to prolong the life of implantable devices. Hence simple
block codes like BCH (63, 51, 2) can be employed in the transceiver design of 802.15.6 Narrowband PHY.
In this paper, implementation of BCH (63, 51, t = 2) Encoder and Decoder using VHDL is discussed. The
incoming 51 bits are encoded into 63 bit code word using (63, 51) BCH encoder. It can detect and correct
up to 2 random errors. The design of an encoder is implemented using Linear Feed Back Shift Register
(LFSR) for polynomial division and the decoder design is based on syndrome calculator, inversion-less
Berlekamp-Massey algorithm (BMA) and Chien search algorithm. Synthesis and simulation were carried
out using Xilinx ISE 14.2 and ModelSim 10.1c. The codes are implemented over Virtex 4 FPGA device and
tested on DN8000K10PCIE Logic Emulation Board. To the best of our knowledge, it is the first time an
implementation of (63, 51) BCH encoder and decoder carried out.
D I G I T A L C O M M U N I C A T I O N S J N T U M O D E L P A P E R{Wwwguest3f9c6b
This document contains questions from a digital communications exam for a B.Tech course. The questions cover topics like PCM systems, delta modulation, digital modulation techniques, error probability analysis, information theory concepts, channel capacity, block codes and conventional codes. There are 8 questions in total with sub-questions on analyzing and comparing communication systems and coding schemes.
Hybrid ldpc and stbc algorithms to improve ber reduction in ofdmIAEME Publication
This document discusses improving bit error rate (BER) performance in orthogonal frequency division multiplexing (OFDM) systems using hybrid low-density parity check (LDPC) codes and space-time block coding (STBC). It first describes MMSE and LS channel estimation techniques in OFDM and an adaptive modulation and coding WiMAX system. It then proposes using LDPC codes with STBC across multiple transmit antennas to achieve coding gain and diversity gain. Simulation results show the hybrid LDPC-STBC method provides better BER performance than STBC alone, especially at low signal-to-noise ratios in flat Rayleigh fading channels.
Memory Based Hardware Efficient Implementation of FIR FiltersDr.SHANTHI K.G
The document summarizes memory-based hardware efficient implementations of finite impulse response (FIR) filters. FIR filters are commonly used in digital signal processing systems. The paper explores memory-based realization of FIR filters using techniques like direct memory implementation and distributed arithmetic. Direct memory implementation replaces multiplications with filter coefficients with pre-computed values stored in a read-only memory (ROM) or lookup table (LUT). Distributed arithmetic implements MAC operations using LUT accesses and shift-accumulation, making it well-suited for field-programmable gate arrays. The paper compares different memory-based architectures for FIR filters in terms of ROM size, delay, and throughput to assist in selecting the best architecture for a given application.
This document summarizes the porting and optimization of the ITU-T G.729.1 speech coding algorithm to the SC3850 DSP core. The authors modified the reference G.729.1 code to fix-point C and optimized it for the SC3850 architecture through techniques like replacing functions with intrinsics, inlining small functions, and restructuring loops and memory accesses. They achieved over 90% improvement in cycles per second (MCPS) through these optimizations while maintaining bit-exact accuracy against the ITU test vectors. The G.729.1 codec is an 8-32 kbps scalable wideband speech codec standardized by ITU, useful for applications like VoIP, audio conferencing, and
This document summarizes a research paper that proposes using genetic algorithms and differential evolution to optimize the interleaver design for turbo codes. Turbo codes achieve performance close to the theoretical limit by using parallel concatenation of recursive systematic convolutional codes. The interleaver permutes the input bits, which affects turbo code performance. This research aims to use evolutionary algorithms to find higher performing turbo code interleavers compared to conventional designs. It compares the proposed approaches to the traditional genetic algorithm method and finds that differential evolution performs well for optimizing turbo code interleaver design.
1) The document describes a modification to the Huffman coding used in JPEG image compression. It proposes pairing each non-zero DCT coefficient with the run-length of subsequent (rather than preceding) zero coefficients.
2) This allows using separate optimized Huffman code tables for each DCT coefficient position, improving compression by 10-15% over standard JPEG coding.
3) The decoding procedure is not changed and no end-of-block marker is needed, providing advantages with no increase in complexity.
This document analyzes the RC4 encryption algorithm and examines how its performance is affected by changing parameters like encryption key length and file size. Experimental tests were conducted to measure encryption time for different key lengths and file types. The results show encryption time increases with longer keys and larger files, and are modeled mathematically. The document also provides background on encryption methods, how RC4 works, and compares stream and block ciphers.
Hardware implementation of (63, 51) bch encoder and decoder for wban using lf...ijitjournal
Error Correcting Codes are required to have a reliable communication through a medium that has an
unacceptable bit error rate and low signal to noise ratio. In IEEE 802.15.6 2.4GHz Wireless Body Area
Network (WBAN), data gets corrupted during the transmission and reception due to noises and
interferences. Ultra low power operation is crucial to prolong the life of implantable devices. Hence simple
block codes like BCH (63, 51, 2) can be employed in the transceiver design of 802.15.6 Narrowband PHY.
In this paper, implementation of BCH (63, 51, t = 2) Encoder and Decoder using VHDL is discussed. The
incoming 51 bits are encoded into 63 bit code word using (63, 51) BCH encoder. It can detect and correct
up to 2 random errors. The design of an encoder is implemented using Linear Feed Back Shift Register
(LFSR) for polynomial division and the decoder design is based on syndrome calculator, inversion-less
Berlekamp-Massey algorithm (BMA) and Chien search algorithm. Synthesis and simulation were carried
out using Xilinx ISE 14.2 and ModelSim 10.1c. The codes are implemented over Virtex 4 FPGA device and
tested on DN8000K10PCIE Logic Emulation Board. To the best of our knowledge, it is the first time an
implementation of (63, 51) BCH encoder and decoder carried out.
D I G I T A L C O M M U N I C A T I O N S J N T U M O D E L P A P E R{Wwwguest3f9c6b
This document contains questions from a digital communications exam for a B.Tech course. The questions cover topics like PCM systems, delta modulation, digital modulation techniques, error probability analysis, information theory concepts, channel capacity, block codes and conventional codes. There are 8 questions in total with sub-questions on analyzing and comparing communication systems and coding schemes.
Hybrid ldpc and stbc algorithms to improve ber reduction in ofdmIAEME Publication
This document discusses improving bit error rate (BER) performance in orthogonal frequency division multiplexing (OFDM) systems using hybrid low-density parity check (LDPC) codes and space-time block coding (STBC). It first describes MMSE and LS channel estimation techniques in OFDM and an adaptive modulation and coding WiMAX system. It then proposes using LDPC codes with STBC across multiple transmit antennas to achieve coding gain and diversity gain. Simulation results show the hybrid LDPC-STBC method provides better BER performance than STBC alone, especially at low signal-to-noise ratios in flat Rayleigh fading channels.
Memory Based Hardware Efficient Implementation of FIR FiltersDr.SHANTHI K.G
The document summarizes memory-based hardware efficient implementations of finite impulse response (FIR) filters. FIR filters are commonly used in digital signal processing systems. The paper explores memory-based realization of FIR filters using techniques like direct memory implementation and distributed arithmetic. Direct memory implementation replaces multiplications with filter coefficients with pre-computed values stored in a read-only memory (ROM) or lookup table (LUT). Distributed arithmetic implements MAC operations using LUT accesses and shift-accumulation, making it well-suited for field-programmable gate arrays. The paper compares different memory-based architectures for FIR filters in terms of ROM size, delay, and throughput to assist in selecting the best architecture for a given application.
This document summarizes the porting and optimization of the ITU-T G.729.1 speech coding algorithm to the SC3850 DSP core. The authors modified the reference G.729.1 code to fix-point C and optimized it for the SC3850 architecture through techniques like replacing functions with intrinsics, inlining small functions, and restructuring loops and memory accesses. They achieved over 90% improvement in cycles per second (MCPS) through these optimizations while maintaining bit-exact accuracy against the ITU test vectors. The G.729.1 codec is an 8-32 kbps scalable wideband speech codec standardized by ITU, useful for applications like VoIP, audio conferencing, and
On the Performance Analysis of Multi-antenna Relaying System over Rayleigh Fa...IDES Editor
In this work, the end-to-end performance of an
amplify-and-forward multi-antenna infrastructure-based relay
(fixed relay) system over flat Rayleigh fading channel is
investigated. New closed form expressions for the statistics of
the received signal-to-noise ratio (SNR) are presented and
applied for studying the outage probability and the average
bit error rate of the digital receivers. The results reveal that
the system performance improves significantly (roughly 3 dB)
for M=2 over that for M=1 in both low and high signal-tonoise
ratio. However, little additional performance
improvement can be achieved for M>2 relative to M=2 at high
SNR.
This document summarizes an implementation of a data error corrector using VLSI techniques. It describes a convolutional encoder and Viterbi decoder with a constraint length of 9 and code rate of 1/2, realized using Verilog HDL. Convolutional codes are used for forward error correction in digital data transmission by adding redundant bits. The Viterbi algorithm performs maximum likelihood decoding by using a trellis structure to calculate path metrics and select the most probable transmitted sequence. The implemented Viterbi decoder contains branch metric, path metric, and survivor memory units to decode a received bit stream encoded with a convolutional code. It was simulated and synthesized using Xilinx 13.1i tools.
International Journal of Engineering Research and Development is an international premier peer reviewed open access engineering and technology journal promoting the discovery, innovation, advancement and dissemination of basic and transitional knowledge in engineering, technology and related disciplines.
The document describes a video watermarking scheme based on discrete wavelet transform (DWT) and principal component analysis (PCA) for copyright protection. The scheme embeds a binary logo watermark into video frames by applying DWT to decompose frames into sub-bands, then applying block-based PCA on sub-blocks of low and high frequency sub-bands. The watermark is embedded into the principal components of the sub-blocks. Algorithms are provided for applying DWT, PCA transforms, and embedding and extracting the watermark. The scheme aims to provide imperceptibility, robustness against attacks, and ownership protection for digital video content.
1. The document presents an audio watermarking algorithm based on discrete wavelet transform and least significant bit insertion (DWT-LSB). The DWT decomposes the audio signal into sub-bands to identify locations for embedding watermark bits. The watermark bits are embedded in the high-resolution sub-band using LSB insertion.
2. Testing showed SNR values between 53.7-59.6 dB for different audio signals, indicating imperceptibility of the embedded watermark. The DWT-LSB method provides satisfactory robustness and imperceptibility for audio watermarking.
3. Digital watermarking techniques like the one proposed can help protect copyright of digital audio by embedding invisible signatures as
This document summarizes a research paper on interference mitigation using a CDMA RAKE receiver with Walsh-Hadamard sequences. It proposes designing and implementing a CDMA RAKE receiver to achieve better performance than a conventional CDMA receiver. The CDMA RAKE receiver uses multiple correlators to separate multipath signals. It then aligns and combines the despread signals before making a decision. Walsh-Hadamard sequences are used for spreading and achieve better orthogonality. Simulation results show the CDMA RAKE receiver achieves a lower bit error rate than the conventional receiver by mitigating multipath interference through maximum ratio combining.
Hybrid Approach for Robust Digital Video WatermarkingIJSRD
With the growing popularity of internet and digital media, digital watermarking techniques have been developed to protect the copyright of multimedia objects such as text, audio, video, etc. So, we have proposed a hybrid video watermarking technique which takes the advantages of different transforms like DWT, DCT, SVD and Arnold Transform, which enhances more security and provides robustness to the watermark. In this paper method, video is divided into several groups of frames, and one of the frames is selected where watermark will be embedded. Before embedding watermark in a selected frame it will be pre-processed with Arnold Transform which will provide security to it. The selected plane of video frame are decomposed using DWT and high frequency band HH, middle frequency bands LH, HL are transformed with DCT. The DCT coefficients are SVD transformed which are embedded with corresponding transformed coefficients of watermarks along with Arnold Transform. The embedded watermark is extracted with inverse process of embedding. The proposed algorithm is tested with various video sequences using MATLAB 2013a. The distortion quality of original image and watermark is controlled by the Peak Signal to Noise Ratio, Signal to Noise Ratio and Mean square error of the watermarked frame with original frame.
Survey on Error Control Coding TechniquesIJTET Journal
This document discusses various error control coding techniques used to ensure correct data transmission over noisy channels. It describes automatic repeat request and forward error correction as the two main approaches. Specific coding schemes covered include parity codes, Hamming codes, BCH codes, Reed-Solomon codes, LDPC codes, convolutional codes, and turbo codes. Reed-Solomon codes can correct multiple burst errors with high code rates. LDPC codes provide performance close to the Shannon limit with lower complexity than turbo codes. The document provides an overview of the coding techniques and their encoding and decoding processes.
Efficient FPGA implementation of high speed digital delay for wideband beamfor...journalBEEI
In this paper, the authors present an FPGA implementation of a digital delay for beamforming applications. The digital delay is based on a Parallel Farrow Filter. Such architecture allows to reach a very high processing rate with wideband signals and it is suitable to be used with Time-Interleaved Analog to Digital Converters (TI-ADC). The proposed delay has been simulated in MATLAB, implemented on FPGA and characterized in terms of amplitude and phase response, maximum clock frequency and area.
This document discusses reliable data communication over mobile ad hoc networks using the WEAC protocol with ARQ technique. It first describes the WEAC protocol and mobile ad hoc network architecture. It then discusses different ARQ techniques like Stop-and-Wait, Go-Back-N, and Selective Repeat. Go-Back-N protocol is analyzed in more detail. The document also discusses ARQ feedback types and presents simulation results on uplink throughput with and without errors and ARQ. It concludes that ARQ can improve network throughput over mobile ad hoc networks by reducing dropped packets at relay nodes.
Error resilient for multiview video transmissions with gop analysisijma
The work in this paper examines the effects of group of pictures on H.264 multiview video coding bitstream
over an erroneous network with different error rates. The study considers analyzing the bitrate
performance for different GOP and error rates to see the effects on the quality of the reconstructed
multiview video. However, by analyzing the multiview video content it is possible to identify an optimum
GOP size depending on the type of application used. In a comparison test, the H.264 data partitioning and
the multi-layer data partitioning technique with different error rates and GOP are evaluated in terms of
quality perception. The results of the simulation confirm that Multi-layer data partitioning technique shows
a better performance at higher error rates with different GOP. Further experiments in this work have
shown the effects of GOP in terms of visual quality and bitrate for different multiview video sequences.
For wireless communication, the demand for high speed, low power and low cost Viterbi decoding are always required. Convolutional coding with Viterbi decoding is a very powerful method for forward error correction and detection method. It has been popularly used in many wireless communication systems to improve the limited capacity of its communication channels. VLSI technology in advance is using low power, less area and high speed constraints is often used for encoding and decoding of data.
ERROR RESILIENT FOR MULTIVIEW VIDEO TRANSMISSIONS WITH GOP ANALYSIS ijma
The work in this paper examines the effects of group of pictures on H.264 multiview video coding bitstream
over an erroneous network with different error rates. The study considers analyzing the bitrate
performance for different GOP and error rates to see the effects on the quality of the reconstructed
multiview video. However, by analyzing the multiview video content it is possible to identify an optimum
GOP size depending on the type of application used. In a comparison test, the H.264 data partitioning and
the multi-layer data partitioning technique with different error rates and GOP are evaluated in terms of
quality perception. The results of the simulation confirm that Multi-layer data partitioning technique shows
a better performance at higher error rates with different GOP. Further experiments in this work have
shown the effects of GOP in terms of visual quality and bitrate for different multiview video sequences
DWT Based Audio Watermarking Schemes : A Comparative Study ijcisjournal
The main problem encountered during multimedia transmission is its protection against illegal distribution
and copying. One of the possible solutions for this is digital watermarking. Digital audio watermarking is
the technique of embedding watermark content to the audio signal to protect the owner copyrights. In this
paper, we used three wavelet transforms i.e. Discrete Wavelet Transform (DWT), Double Density DWT
(DDDWT) and Dual Tree DWT (DTDWT) for audio watermarking and the performance analysis of each
transform is presented. The key idea of the basic algorithm is to segment the audio signal into two parts,
one is for synchronization code insertion and other one is for watermark embedding. Initially, binary
watermark image is scrambled using chaotic technique to provide secrecy. By using QuantizationIndex
Modulation (QIM), this method works as a blind technique. The comparative analysis of the three methods
is made by conducting robustness and imperceptibility tests are conducted on five benchmark audio
signals.
This document proposes a bi-level/full-color video combination scheme to enable video communication across a wide range of bandwidths. Bi-level video uses 1 bit per pixel and works well below 56 Kbps, while full-color video has higher quality but requires over 33.6 Kbps. The scheme uses bandwidth estimation to switch between the two formats in the 33.6-56 Kbps range for smooth adaptation. It estimates available bandwidth from receiver feedback on packet loss and round-trip time, then adjusts the video format accordingly.
SECURED TEXT MESSAGE TRANSMISSION WITH IMPLEMENTATION OF CONCATENATED CFB CRY...cscpconf
In the present simulated system, text message transmission has been secured with concatenated
implementation of Cipher Feedback(CFB) cryptographic algorithm. It is anticipated from the numerical
results that the pre-ZF channel equalization based MIMO OFDM wireless communication
system outperforms in QAM digital modulation and BCH channel coding under AWGN and
Raleigh fading channels .In Pre-MMSE/pre-ZF channel equalization scheme, the system shows
comparatively worst performance in convolutional channel coding scheme with QAM/QPSK digital
modulation. It has been observed from the present study that the system performance deteriorates with
increase in noise power as compared to signal power. study of a secured MIMO Orthogonal
Frequency-Division Multiplexing wireless communication system with implementation of two pre
channel equalization techniques such as Pre-Minimum Mean Square Error (Pre-MMSE) and PreZero
Forcing(Pre-ZF) under QPSK and QAM digital modulations.
This document discusses the design of low-density parity-check (LDPC) codes that provide unequal error protection (UEP) when used with higher order constellations (HOCs) like 8-PSK and 64-QAM. The proposed method divides the variable node degree distribution into sub-distributions corresponding to protection classes from the source coding and modulation classes of the HOC. An iterative linear programming approach is used to optimize the sub-distributions to enhance the UEP capability of the code for different signal-to-noise ratios, while reducing the overall bit error rate by accounting for the different bit error probabilities of the HOC. Simulation results show significant bit error rate reductions compared to codes optimized for binary phase-shift keying
Data detection with a progressive parallel ici canceller in mimo ofdmeSAT Publishing House
The document describes a progressive parallel interference canceller (PPIC) for use in a MIMO-OFDM system to suppress inter-carrier interference (ICI). PPIC is compared to parallel interference canceller (PIC) and shows lower complexity and better performance. PPIC architecture is simpler than PIC and more suitable for implementation in wireless communication systems requiring high data rates and mobility. Simulation results show that PPIC combined with LDPC coding achieves lower bit error rates than PIC combined with LDPC coding.
A fast pu mode decision algorithm for h.264 avc to hevc transcodingcsandit
The document proposes a fast algorithm for transcoding video from H.264/AVC to HEVC. The algorithm reuses motion vectors, residuals, and modes from H.264/AVC blocks to predict coding units in HEVC. It includes a fast prediction unit decision and fast motion estimation. For prediction unit decision, it analyzes motion vector variance and residuals to skip modes. For depths 0-1, it handles larger HEVC blocks differently than depths 2-3 which match H.264/AVC block sizes. Fast motion estimation adapts the search range based on motion vector differences and predicts ranges using transition probabilities between standards. Simulation results show up to 53% encoding time savings while maintaining rate-dist
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
This document summarizes a research paper that proposes a new method to improve the efficiency of the HEVC video coding standard. The proposed method uses particle swarm optimization (PSO) to optimize the coding unit partitioning patterns in HEVC in order to reduce computational complexity. Compared to existing algorithms, the proposed PSO-based method achieves a 12% increase in PSNR video quality while reducing computational complexity by 40% and overall encoding delay. The method integrates this optimized intra-frame prediction into the HEVC encoding and decoding process.
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors..
International Journal of Engineering Research and DevelopmentIJERD Editor
The document summarizes an emerging VP8 video codec that is designed for mobile devices. It aims to significantly reduce computational complexity through several techniques while maintaining good video quality. The key techniques include a predictive algorithm for motion estimation that reduces computation by 18.5-20x compared to full search, using integer discrete cosine transform instead of floating point to achieve 2.6-3.5x speed improvement, and skipping DCT and quantization for some macroblocks to reduce computations. Experimental results on test sequences show negligible quality degradation of 0.2-0.5dB for integer DCT and 0.5dB on average for the full codec, while achieving real-time encoding rates on mobile devices. The proposed low-complexity
On the Performance Analysis of Multi-antenna Relaying System over Rayleigh Fa...IDES Editor
In this work, the end-to-end performance of an
amplify-and-forward multi-antenna infrastructure-based relay
(fixed relay) system over flat Rayleigh fading channel is
investigated. New closed form expressions for the statistics of
the received signal-to-noise ratio (SNR) are presented and
applied for studying the outage probability and the average
bit error rate of the digital receivers. The results reveal that
the system performance improves significantly (roughly 3 dB)
for M=2 over that for M=1 in both low and high signal-tonoise
ratio. However, little additional performance
improvement can be achieved for M>2 relative to M=2 at high
SNR.
This document summarizes an implementation of a data error corrector using VLSI techniques. It describes a convolutional encoder and Viterbi decoder with a constraint length of 9 and code rate of 1/2, realized using Verilog HDL. Convolutional codes are used for forward error correction in digital data transmission by adding redundant bits. The Viterbi algorithm performs maximum likelihood decoding by using a trellis structure to calculate path metrics and select the most probable transmitted sequence. The implemented Viterbi decoder contains branch metric, path metric, and survivor memory units to decode a received bit stream encoded with a convolutional code. It was simulated and synthesized using Xilinx 13.1i tools.
International Journal of Engineering Research and Development is an international premier peer reviewed open access engineering and technology journal promoting the discovery, innovation, advancement and dissemination of basic and transitional knowledge in engineering, technology and related disciplines.
The document describes a video watermarking scheme based on discrete wavelet transform (DWT) and principal component analysis (PCA) for copyright protection. The scheme embeds a binary logo watermark into video frames by applying DWT to decompose frames into sub-bands, then applying block-based PCA on sub-blocks of low and high frequency sub-bands. The watermark is embedded into the principal components of the sub-blocks. Algorithms are provided for applying DWT, PCA transforms, and embedding and extracting the watermark. The scheme aims to provide imperceptibility, robustness against attacks, and ownership protection for digital video content.
1. The document presents an audio watermarking algorithm based on discrete wavelet transform and least significant bit insertion (DWT-LSB). The DWT decomposes the audio signal into sub-bands to identify locations for embedding watermark bits. The watermark bits are embedded in the high-resolution sub-band using LSB insertion.
2. Testing showed SNR values between 53.7-59.6 dB for different audio signals, indicating imperceptibility of the embedded watermark. The DWT-LSB method provides satisfactory robustness and imperceptibility for audio watermarking.
3. Digital watermarking techniques like the one proposed can help protect copyright of digital audio by embedding invisible signatures as
This document summarizes a research paper on interference mitigation using a CDMA RAKE receiver with Walsh-Hadamard sequences. It proposes designing and implementing a CDMA RAKE receiver to achieve better performance than a conventional CDMA receiver. The CDMA RAKE receiver uses multiple correlators to separate multipath signals. It then aligns and combines the despread signals before making a decision. Walsh-Hadamard sequences are used for spreading and achieve better orthogonality. Simulation results show the CDMA RAKE receiver achieves a lower bit error rate than the conventional receiver by mitigating multipath interference through maximum ratio combining.
Hybrid Approach for Robust Digital Video WatermarkingIJSRD
With the growing popularity of internet and digital media, digital watermarking techniques have been developed to protect the copyright of multimedia objects such as text, audio, video, etc. So, we have proposed a hybrid video watermarking technique which takes the advantages of different transforms like DWT, DCT, SVD and Arnold Transform, which enhances more security and provides robustness to the watermark. In this paper method, video is divided into several groups of frames, and one of the frames is selected where watermark will be embedded. Before embedding watermark in a selected frame it will be pre-processed with Arnold Transform which will provide security to it. The selected plane of video frame are decomposed using DWT and high frequency band HH, middle frequency bands LH, HL are transformed with DCT. The DCT coefficients are SVD transformed which are embedded with corresponding transformed coefficients of watermarks along with Arnold Transform. The embedded watermark is extracted with inverse process of embedding. The proposed algorithm is tested with various video sequences using MATLAB 2013a. The distortion quality of original image and watermark is controlled by the Peak Signal to Noise Ratio, Signal to Noise Ratio and Mean square error of the watermarked frame with original frame.
Survey on Error Control Coding TechniquesIJTET Journal
This document discusses various error control coding techniques used to ensure correct data transmission over noisy channels. It describes automatic repeat request and forward error correction as the two main approaches. Specific coding schemes covered include parity codes, Hamming codes, BCH codes, Reed-Solomon codes, LDPC codes, convolutional codes, and turbo codes. Reed-Solomon codes can correct multiple burst errors with high code rates. LDPC codes provide performance close to the Shannon limit with lower complexity than turbo codes. The document provides an overview of the coding techniques and their encoding and decoding processes.
Efficient FPGA implementation of high speed digital delay for wideband beamfor...journalBEEI
In this paper, the authors present an FPGA implementation of a digital delay for beamforming applications. The digital delay is based on a Parallel Farrow Filter. Such architecture allows to reach a very high processing rate with wideband signals and it is suitable to be used with Time-Interleaved Analog to Digital Converters (TI-ADC). The proposed delay has been simulated in MATLAB, implemented on FPGA and characterized in terms of amplitude and phase response, maximum clock frequency and area.
This document discusses reliable data communication over mobile ad hoc networks using the WEAC protocol with ARQ technique. It first describes the WEAC protocol and mobile ad hoc network architecture. It then discusses different ARQ techniques like Stop-and-Wait, Go-Back-N, and Selective Repeat. Go-Back-N protocol is analyzed in more detail. The document also discusses ARQ feedback types and presents simulation results on uplink throughput with and without errors and ARQ. It concludes that ARQ can improve network throughput over mobile ad hoc networks by reducing dropped packets at relay nodes.
Error resilient for multiview video transmissions with gop analysisijma
The work in this paper examines the effects of group of pictures on H.264 multiview video coding bitstream
over an erroneous network with different error rates. The study considers analyzing the bitrate
performance for different GOP and error rates to see the effects on the quality of the reconstructed
multiview video. However, by analyzing the multiview video content it is possible to identify an optimum
GOP size depending on the type of application used. In a comparison test, the H.264 data partitioning and
the multi-layer data partitioning technique with different error rates and GOP are evaluated in terms of
quality perception. The results of the simulation confirm that Multi-layer data partitioning technique shows
a better performance at higher error rates with different GOP. Further experiments in this work have
shown the effects of GOP in terms of visual quality and bitrate for different multiview video sequences.
For wireless communication, the demand for high speed, low power and low cost Viterbi decoding are always required. Convolutional coding with Viterbi decoding is a very powerful method for forward error correction and detection method. It has been popularly used in many wireless communication systems to improve the limited capacity of its communication channels. VLSI technology in advance is using low power, less area and high speed constraints is often used for encoding and decoding of data.
ERROR RESILIENT FOR MULTIVIEW VIDEO TRANSMISSIONS WITH GOP ANALYSIS ijma
The work in this paper examines the effects of group of pictures on H.264 multiview video coding bitstream
over an erroneous network with different error rates. The study considers analyzing the bitrate
performance for different GOP and error rates to see the effects on the quality of the reconstructed
multiview video. However, by analyzing the multiview video content it is possible to identify an optimum
GOP size depending on the type of application used. In a comparison test, the H.264 data partitioning and
the multi-layer data partitioning technique with different error rates and GOP are evaluated in terms of
quality perception. The results of the simulation confirm that Multi-layer data partitioning technique shows
a better performance at higher error rates with different GOP. Further experiments in this work have
shown the effects of GOP in terms of visual quality and bitrate for different multiview video sequences
DWT Based Audio Watermarking Schemes : A Comparative Study ijcisjournal
The main problem encountered during multimedia transmission is its protection against illegal distribution
and copying. One of the possible solutions for this is digital watermarking. Digital audio watermarking is
the technique of embedding watermark content to the audio signal to protect the owner copyrights. In this
paper, we used three wavelet transforms i.e. Discrete Wavelet Transform (DWT), Double Density DWT
(DDDWT) and Dual Tree DWT (DTDWT) for audio watermarking and the performance analysis of each
transform is presented. The key idea of the basic algorithm is to segment the audio signal into two parts,
one is for synchronization code insertion and other one is for watermark embedding. Initially, binary
watermark image is scrambled using chaotic technique to provide secrecy. By using QuantizationIndex
Modulation (QIM), this method works as a blind technique. The comparative analysis of the three methods
is made by conducting robustness and imperceptibility tests are conducted on five benchmark audio
signals.
This document proposes a bi-level/full-color video combination scheme to enable video communication across a wide range of bandwidths. Bi-level video uses 1 bit per pixel and works well below 56 Kbps, while full-color video has higher quality but requires over 33.6 Kbps. The scheme uses bandwidth estimation to switch between the two formats in the 33.6-56 Kbps range for smooth adaptation. It estimates available bandwidth from receiver feedback on packet loss and round-trip time, then adjusts the video format accordingly.
SECURED TEXT MESSAGE TRANSMISSION WITH IMPLEMENTATION OF CONCATENATED CFB CRY...cscpconf
In the present simulated system, text message transmission has been secured with concatenated
implementation of Cipher Feedback(CFB) cryptographic algorithm. It is anticipated from the numerical
results that the pre-ZF channel equalization based MIMO OFDM wireless communication
system outperforms in QAM digital modulation and BCH channel coding under AWGN and
Raleigh fading channels .In Pre-MMSE/pre-ZF channel equalization scheme, the system shows
comparatively worst performance in convolutional channel coding scheme with QAM/QPSK digital
modulation. It has been observed from the present study that the system performance deteriorates with
increase in noise power as compared to signal power. study of a secured MIMO Orthogonal
Frequency-Division Multiplexing wireless communication system with implementation of two pre
channel equalization techniques such as Pre-Minimum Mean Square Error (Pre-MMSE) and PreZero
Forcing(Pre-ZF) under QPSK and QAM digital modulations.
This document discusses the design of low-density parity-check (LDPC) codes that provide unequal error protection (UEP) when used with higher order constellations (HOCs) like 8-PSK and 64-QAM. The proposed method divides the variable node degree distribution into sub-distributions corresponding to protection classes from the source coding and modulation classes of the HOC. An iterative linear programming approach is used to optimize the sub-distributions to enhance the UEP capability of the code for different signal-to-noise ratios, while reducing the overall bit error rate by accounting for the different bit error probabilities of the HOC. Simulation results show significant bit error rate reductions compared to codes optimized for binary phase-shift keying
Data detection with a progressive parallel ici canceller in mimo ofdmeSAT Publishing House
The document describes a progressive parallel interference canceller (PPIC) for use in a MIMO-OFDM system to suppress inter-carrier interference (ICI). PPIC is compared to parallel interference canceller (PIC) and shows lower complexity and better performance. PPIC architecture is simpler than PIC and more suitable for implementation in wireless communication systems requiring high data rates and mobility. Simulation results show that PPIC combined with LDPC coding achieves lower bit error rates than PIC combined with LDPC coding.
A fast pu mode decision algorithm for h.264 avc to hevc transcodingcsandit
The document proposes a fast algorithm for transcoding video from H.264/AVC to HEVC. The algorithm reuses motion vectors, residuals, and modes from H.264/AVC blocks to predict coding units in HEVC. It includes a fast prediction unit decision and fast motion estimation. For prediction unit decision, it analyzes motion vector variance and residuals to skip modes. For depths 0-1, it handles larger HEVC blocks differently than depths 2-3 which match H.264/AVC block sizes. Fast motion estimation adapts the search range based on motion vector differences and predicts ranges using transition probabilities between standards. Simulation results show up to 53% encoding time savings while maintaining rate-dist
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
This document summarizes a research paper that proposes a new method to improve the efficiency of the HEVC video coding standard. The proposed method uses particle swarm optimization (PSO) to optimize the coding unit partitioning patterns in HEVC in order to reduce computational complexity. Compared to existing algorithms, the proposed PSO-based method achieves a 12% increase in PSNR video quality while reducing computational complexity by 40% and overall encoding delay. The method integrates this optimized intra-frame prediction into the HEVC encoding and decoding process.
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME...ijma
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors..
International Journal of Engineering Research and DevelopmentIJERD Editor
The document summarizes an emerging VP8 video codec that is designed for mobile devices. It aims to significantly reduce computational complexity through several techniques while maintaining good video quality. The key techniques include a predictive algorithm for motion estimation that reduces computation by 18.5-20x compared to full search, using integer discrete cosine transform instead of floating point to achieve 2.6-3.5x speed improvement, and skipping DCT and quantization for some macroblocks to reduce computations. Experimental results on test sequences show negligible quality degradation of 0.2-0.5dB for integer DCT and 0.5dB on average for the full codec, while achieving real-time encoding rates on mobile devices. The proposed low-complexity
Optimal coding unit decision for early termination in high efficiency video c...IJECEIAES
Video compression is an emerging research topic in the field of block based video encoders. Due to the growth of video coding technologies, high efficiency video coding (HEVC) delivers superior coding performance. With the increased encoding complexity, the HEVC enhances the rate-distortion (RD) performance. In the video compression, the out-sized coding units (CUs) have higher encoding complexity. Therefore, the computational encoding cost and complexity remain vital concerns, which need to be considered as an optimization task. In this manuscript, an enhanced whale optimization algorithm (EWOA) is implemented to reduce the computational time and complexity of the HEVC. In the EWOA, a cosine function is incorporated with the controlling parameter A and two correlation factors are included in the WOA for controlling the position of whales and regulating the movement of search mechanism during the optimization and search processes. The bit streams in the Luma-coding tree block are selected using EWOA that defines the CU neighbors and is used in the HEVC. The results indicate that the EWOA achieves best bit rate (BR), time saving, and peak signal to noise ratio (PSNR). The EWOA showed 0.006-0.012 dB higher PSNR than the existing models in the real-time videos.
Video streaming using light-weight transcoding and in-network intelligenceMinh Nguyen
In this paper, we introduce a novel approach, LwTE, which reduces streaming costs in HTTP Adaptive Streaming (HAS) by enabling light-weight transcoding at the edge. In LwTE, during encoding of a video segment in the origin server, a metadata is generated which stores the optimal encoding decisions. LwTE enables us to store only the highest bitrate plus corresponding metadata (of very small size) for unpopular video segments/bitrates. Since metadata is of very small size, replacing unpopular video segments/bitrates with their metadata results in considerable saving in the storage costs. The metadata is reused at the edge servers to reduce the required time and computational resources for on-the-fly transcoding.
IEEE MMSP'21: INCEPT: Intra CU Depth Prediction for HEVCVignesh V Menon
The document presents the INCEPT algorithm for intra coding unit (CU) depth prediction in HEVC video encoding. INCEPT uses features like texture energy and neighboring CU statistics to predict the minimum and maximum depth for each CU. It achieved an average encoding time savings of 23.34% over benchmark algorithms with only a 1.28% increase in bitrate. The algorithm works well across different video types and quality levels. Future work could reduce PU mode decision time and extend the approach to the VVC standard.
High Efficiency Video Coding (HEVC) improves the encoding efficiency by utilizing sophisticated tools such as flexible Coding Tree Unit (CTU) partitioning. The Coding Unit (CU) can be split recursively into four equally sized CUs ranging from 64×64 to 8×8 pixels. At each depth level (or CU size), intra prediction via exhaustive mode search was exploited in HEVC to improve the encoding efficiency and result in a very high encoding time complexity. This paper proposes an Intra CU Depth Prediction (INCEPT) algorithm, which limits Rate-Distortion Optimization (RDO) for each CTU in HEVC by utilizing the spatial correlation with the neighboring CTUs, which is computed using a DCT energy-based feature. Thus, INCEPT reduces the number of candidate CU sizes required to be considered for each CTU in HEVC intra coding. Experimental results show that the INCEPT algorithm achieves a better trade-off between the encoding efficiency and encoding time saving (i.e., BDR/∆T) than the benchmark algorithms. While BDR/∆T is 12.35% and 9.03% for the benchmark algorithms, it is 5.49% for the proposed algorithm. As a result, INCEPT achieves a 23.34% reduction in encoding time on average while incurring only a 1.67% increase in bit rate than the original coding in the x265 HEVC open-source encoder.
This document discusses video quality analysis for H.264 based on the human visual system. It proposes an improved video quality assessment method that adds color comparison to structural similarity measurement. The method separates similarity measurement into four comparisons: luminance, contrast, structure, and color. Experimental results on video sets with two distortion types show the proposed method's quality scores are more consistent with visual quality than classical methods. It also discusses the H.264 video coding standard and provides examples of encoding and decoding experimental results.
Distributed Video Coding (DVC) has become increasingly popular in recent times among the researchers in video coding due to its attractive and promising features. DVC primarily has a modified complexity balance between the encoder and decoder, in contrast to conventional video codecs. However, Most of the reported DVC schemes have a high time-delay in decoder which hinders its practical application in real-time systems. In this work, we focus on speed up the Side Information(SI) generation module in DVC, which is a major function in the DVC coding algorithm and one of the time-consuming factor at the decoder. By applied it through Compute Unified Device Architecture (CUDA) based on General-Purpose Graphics Processing Unit (GPGPU), the experimental results show that a considerable speedup can be obtained by using the proposed parallelized SI generation algorithm.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
Requiring only half the bitrate of its predecessor, the new standard – HEVC or H.265 – will significantly reduce the need for bandwidth and expensive, limited spectrum. HEVC (H.265) will enable the launch of new video services and in particular ultra HD television (UHDTV).
State-of-the-art video compression techniques – HEVC/H.265 – can reduce the size of raw video by a factor of about 100 without any noticeable reduction in visual quality. With estimates indicating that compressed real-time video accounts for more than 50 percent of current network traffic, and this figure is set to rise to 90 percent within a few years, HEVC/H.265 will be a welcome relief for network operators.
New services, devices and changing viewing patterns are among the factors contributing to the growth in video traffic as people watch more and more traditional TV and video-streaming services on their mobile devices.
Ericsson has been heavily involved in the standardization of HEVC since it began in 2010, and this Ericsson Review article highlights some of the contributions that have led to the compression efficiency offered by HEVC.
.
Machine learning-based energy consumption modeling and comparing of H.264 and...IJECEIAES
Advancement of the prediction models used in a variety of fields is a result of the contribution of machine learning approaches. Utilizing such modeling in feature engineering is exceptionally imperative and required. In this research, we show how to utilize machine learning to save time in research experiments, where we save more than five thousand hours of measuring the energy consumption of encoding recordings. Since measuring the energy consumption has got to be done by humans and since we require more than eleven thousand experiments to cover all the combinations of video sequences, video bit rate, and video encoding settings, we utilize machine learning to model the energy consumption utilizing linear regression. VP8 codec has been offered by Google as a free video encoder in an effort to replace the popular H.264 video encoder standard. This research model energy consumption and describes the major differences between H.264/AVC and VP8 encoders based on of energy consumption and performance through experiments that are machine learning-based modeling. Twentynine uncompressed video segments from a standard data-set are used, with several sizes, details, and dynamics, where the frame sizes ranging from QCIF(176x144) to 2160p(3840x2160). For fairness in comparison analysis, we use seven settings in VP8 encoder and fifteen types of tuning in H.264/AVC. The settings cover various video qualities. The performance metrics include video qualities, encoding time, and encoding energy consumption.
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERcscpconf
This paper proposes about motion estimation in H.264/AVC encoder. Compared with standards
such as MPEG-2 and MPEG-4 Visual, H.264 can deliver better image quality at the same
compressed bit rate or at a lower bit rate. The increase in compression efficiency comes at the
expense of increase in complexity, which is a fact that must be overcome. An efficient Co-design
methodology is required, where the encoder software application is highly optimized and
structured in a very modular and efficient manner, so as to allow its most complex and time
consuming operations to be offloaded to dedicated hardware accelerators. The Motion
Estimation algorithm is the most computationally intensive part of the encoder which is simulated using MATLAB. The hardware/software co-simulation is done using system generator tool and implemented using Xilinx FPGA Spartan 3E for different scanning methods.
This document describes a project to design an H.264 video decoder using Verilog. It implements the key decoding blocks like Context-Based Adaptive Binary Arithmetic Coding (CABAC), inverse quantization, and inverse discrete cosine transform. CABAC is the entropy decoding method used in H.264 that is computationally intensive. The project develops hardware modules for these blocks to accelerate decoding and enable real-time performance. It presents the designs of the individual modules and simulation results showing their functionality. The goal is to improve on software implementations by using dedicated hardware for the critical decoding stages.
H.265ImprovedCE_over_H.264-HarmonicMay2014FinalDonald Pian
H.265/HEVC is a video compression standard that achieves around 50% higher compression efficiency than its predecessor H.264. It introduces new coding tools like larger coding units (64x64 vs 16x16 in H.264), additional filters, and more flexible block partitioning. Subjective comparisons of original and compressed video are important and can involve viewing them side-by-side, alternating between them, or viewing a difference image alongside the compressed video to detect artifacts. When developing technology for Hollywood, it is important to preserve the director's artistic intent, use proper color spaces, and avoid introducing artifacts without permission.
This document presents a real-time H.264 video compression algorithm for underwater channels. It proposes a hardware/software encoder design using a multi-core processor and graphics accelerator. The software uses uneven multi-hexagonal search for efficient motion estimation and Trellis 1 quantization. The encoder was tested on an underwater video achieving a 210:1 compression ratio and 41.37dB PSNR at 17.5 frames/second, satisfying real-time underwater video transmission requirements.
Video Compression Algorithm Based on Frame Difference Approaches ijsc
The huge usage of digital multimedia via communications, wireless communications, Internet, Intranet and cellular mobile leads to incurable growth of data flow through these Media. The researchers go deep in developing efficient techniques in these fields such as compression of data, image and video. Recently, video compression techniques and their applications in many areas (educational, agriculture, medical …) cause this field to be one of the most interested fields. Wavelet transform is an efficient method that can be used to perform an efficient compression technique. This work deals with the developing of an efficient video compression approach based on frames difference approaches that concentrated on the calculation of frame near distance (difference between frames). The
selection of the meaningful frame depends on many factors such as compression performance, frame details, frame size and near distance between frames. Three different approaches are applied for removing the lowest frame difference. In this paper, many videos are tested to insure the efficiency of this technique, in addition a good performance results has been obtained.
H2B2VS (HEVC hybrid broadcast broadband video services) – Building innovative...Raoul Monnier
Broadcast and broadband networks continue to be separate worlds in the video consumption business. Some initiatives such as HbbTV have built a bridge between both worlds, but its application is almost limited to providing links over the broadcast channel to content providers’ applications such as Catch-up TV services. When it comes to reality, the user is using either one network or the other.
H2B2VS is a Celtic-Plus project aiming at exploiting the potential of real hybrid networks by implementing efficient synchronization mechanisms and using new video coding standard such as High Efficiency Video Coding (HEVC). The goal is to develop successful hybrid network solutions that enable value added services with an optimum bandwidth usage in each network and with clear commercial applications. An example of the potential of this approach is the transmission of Ultra-HD TV by sending the main content over the broadcast channel and the required complementary information over the broadband network. This technology can also be used to improve the life of handicapped persons: Deaf people receive through the broadband network a sign language translation of a programme sent over the broadcast channel; the TV set then displays this translation in an inset window.
One of the most important contributions of the project is developing and testing synchronization methods between two different networks that offer unequal qualities of service with significant differences in delay and jitter.
In this paper, the main technological project contributions are described, including SHVC, the scalable extension of HEVC and a special focus on the synchronization solution adopted by MPEG and DVB. The paper also presents some of the implemented practical use cases, such as the sign language translation described above, and their performance results so as to evaluate the commercial application of this type of solution.
Similar to IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME PREDICTION (20)
Batteries -Introduction – Types of Batteries – discharging and charging of battery - characteristics of battery –battery rating- various tests on battery- – Primary battery: silver button cell- Secondary battery :Ni-Cd battery-modern battery: lithium ion battery-maintenance of batteries-choices of batteries for electric vehicle applications.
Fuel Cells: Introduction- importance and classification of fuel cells - description, principle, components, applications of fuel cells: H2-O2 fuel cell, alkaline fuel cell, molten carbonate fuel cell and direct methanol fuel cells.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
UNLOCKING HEALTHCARE 4.0: NAVIGATING CRITICAL SUCCESS FACTORS FOR EFFECTIVE I...amsjournal
The Fourth Industrial Revolution is transforming industries, including healthcare, by integrating digital,
physical, and biological technologies. This study examines the integration of 4.0 technologies into
healthcare, identifying success factors and challenges through interviews with 70 stakeholders from 33
countries. Healthcare is evolving significantly, with varied objectives across nations aiming to improve
population health. The study explores stakeholders' perceptions on critical success factors, identifying
challenges such as insufficiently trained personnel, organizational silos, and structural barriers to data
exchange. Facilitators for integration include cost reduction initiatives and interoperability policies.
Technologies like IoT, Big Data, AI, Machine Learning, and robotics enhance diagnostics, treatment
precision, and real-time monitoring, reducing errors and optimizing resource utilization. Automation
improves employee satisfaction and patient care, while Blockchain and telemedicine drive cost reductions.
Successful integration requires skilled professionals and supportive policies, promising efficient resource
use, lower error rates, and accelerated processes, leading to optimized global healthcare outcomes.
IMPROVING PSNR AND PROCESSING SPEED FOR HEVC USING HYBRID PSO FOR INTRA FRAME PREDICTION
1. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
DOI:10.5121/ijma.2020.12401 1
IMPROVING PSNR AND PROCESSING
SPEED FOR HEVC USING HYBRID
PSO FOR INTRA FRAME PREDICTION
Swati Vinod Sakhare and Upena D. Dalal
Department of Electronics Engineering, The Sardar Vallabhbhai
National Institute of Technology, Surat, Gujarat, India
ABSTRACT
High efficiency video coding (HEVC) is the newest video codec to increase significantly the coding
efficiency of its ancestor H.264/Advance Video Coding. However, the HEVC delivers a highly increased
computation complexity. In this paper, a coding unit partitioning pattern optimization method based on
particle swarm optimization (PSO) is proposed to reduce the computational complexity of hierarchical
quadtree-based coding unit partitioning. The required coding unit partitioning pattern for exhaustive
partitioning and the rate distortion cost are efficiently considered as the chromosome and the fitness
function of the PSO, respectively. To reduce the computational time, the cellular automata-based (CA)
rule based time limit is used in order to find out the best possible modes of operation. Compared to the
current state of the art algorithms, this scheme is computationally simple and achieves superior
reconstructed video quality (12% increase in PSNR compared to existing methods) at less computational
complexity (overall delay by 40%), Increasing the bandwidth and reducing the errors..
KEYWORDS
Fast Encoding, PSO, High Efficiency Video Coding, Quadtree-Based Coding Unit Partitioning.
1. INTRODUCTION
These days, video dissemination for different objects is multiplying over the Internet with the
guides of helpful correspondence systems and brilliant cell phones. Furthermore, video buyers
progressively request top quality (HD) and ultra-top notch (UHD) recordings to encounter better
visual quality. Accordingly, the conveyance of HD/UHD video to the cell phone clients over the
Internet is turning into a well-known pattern. Be that as it may, the information amount for
HD/UHD video is tremendous because of the higher video goals and edge rate. The information
size of a 10-second video with 3840 × 2160 goals at an edge pace of 60casings for each
second arrives at almost 15 GB. Because of this, the conveyance of HD/UHD
video requests a bigger measure of system transmission capacity and information stockpiling
contrasted with the lower goals standard definition (SD) recordings.
With respect to saving money on organize assets and capacity prerequisite, a productive
pressure system is vitally significant. Joint Collaborative Team on video coding (JCT-
VC),the communitarian venture gathering of ITU-T Video Coding Experts Group (VCEG) and
the ISO/IEC Moving Picture Expert Group (MPEG), has executed a profoundly proficient
videocoding standard called High Efficiency Video Coding (HEVC)/H.265 [1] as an answer for
the issue of expanded video goals. ITU-T and ISO/IEC are the principle institutionalization
2. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
2
bodies which have institutionalized all HEVC’s predecessor guidelines in numerous years.
They have used a 16 × 16 macro block as an essential handling unit in HEVC’s precursor. Each
edge is part into macro blocks. Each macro block includes one 16 × 16 square of lumaparts to
speak to the splendour and two 8 × 8 squares of chroma segments to allude the shading in the
4:2:0 chrominance sub sampling groups. Along these lines, the macro block is the
biggest square size to demonstrate the anticipated data of intra-outline or between outline
expectation in past video coding norms. In any case, ordinary HD and UHD recordings have
numerous bigger edge areas than the macro block, and those districts can speak
to the equivalent moving data. On the off chance that the macroblock is utilized as an essential
preparing unit for regular HD and UHD recordings, a lot of bits are important to flag the
expectation data. Correspondingly, the change square size is bigger than the macro block size.
Consequently, HEVC bolsters a bigger square size as a fundamental handling unit called CTU
for intra-outline or between outline expectation and change coding. Albeit a huge square size is
sufficient for high goals video, it’s anything but a decent decision for low goals video. To be
good with both high-and low-goals recordings, HEVC can deftly parcel the video outline into a
few square CTUs of 2L×2 L tests, where L ∈ {4, 5, 6}. The encoder deftly picks are asonable
estimation of L for proposed application to have the best ex- change off between coding
execution and cost, for example, memory stockpiling, encoding time, and
postponement. Be that as it may, utilizing bigger square for choosing whether intra-mode or
between mode at the forecast stage can’t ensure to get a decent RD execution for expectation
organize. To accomplish better coding proficiency, HEVC presented another essential
handling unit, called CU and an adaptable quadtree apportioning from CTU to CU. Thusly,CU
size can be 64× 64, 32 × 32, 16 × 16, and 8 × 8 at profundity 0, profundity 1, profundity2, and
profundity 3,individually. To characterize CU size or profundity, HEVC begins a preliminary
encoding which incorporates two primary capacities called the RD cost estimation and
correlation in top-down and base up way, individually, as referenced in Section I. In the
top-down RD cost computation of a 64× 64 CTU, the RD costs for all conceivable 85 CUs
are determined in a preorder traversal of the quadtree, if the greatest CU profundity is 3. In
subtleties, there are 1, 4, 16, and 64 CUs at profundity 0, profundity 1,profundity 2, and
profundity 3, separately, and the complete number of CUs is P3 i=0 4 I = 85CUs. Subsequent
to computing the RD costs for four kids CUs of each parent CU, HEVCgoes to the RD cost
correlation with choose whether a parent CU is part or not by looking at the RD cost of parting
and non-parting states of parent CU. At that point, HEVC changes to the RD cost estimation or
performs correlation once more, contingent upon the situation of parent CU. In this way, there
are 85 computations and 21 examinations in the top-down RD cost estimation and base up RD
cost correlation of a 64 × 64 CTU, individually. After at long last looking at a root CU at
profundity 0 with its four kids CUs at profundity 1, the best CU quadtree structure of a CTU
with the least RD cost is picked among 83,522 potential quadtree structures. The CU dividing
examples of casing portrayal of picture request tally(POC) 40 of grouping ”Blowing
Bubbles” looked by a thorough RDO search of HM variant16.5 (HM16.5). The preliminary
encoding of HEVC finds the best CU segment structure of each CTU after a
comprehensive RDO search. In this manner, picking an ideal CU apportioning
structure can be demonstrated as an advancement issue and can be discovered an answer by a
softly appropriate enhancement apparatus to look through a space of conceivable CU segment
arrangements. For little space advancement process, customary thorough procedures
are suitable to discover the arrangement [20]. In any case, the procedures dependent
on man-made consciousness (AI) are productive for a tremendous inquiry space and PSO is one
of AI systems to look through a decent arrangement proficiently.
The major objectives of this text are,
3. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
3
To design a reduced complexity intra-frame predictor using soft computing
To design a reduced complexity intra-frame predictor using soft computing
To optimize the time complexity of this predictor using cellular automata rules
To integrate the reduced time complexity and reduced computational
complexity predictor into HEVC encoding and decoding process.
The next section describes various techniques for improving HEVC performance followed by
the proposed predictors. This text concludes with some interesting observations about our
results and some recommendations that can be researched to further optimize the prediction
performance.
1.1 DUAL TREE COMPLEX WAVELET TRANSFORM
DT-DWT is the advanced design of DWT. Unlike DWT, DTDWT can obtain better shift
invariance and directional selectivity [21, 22]. DTDWT is also known as complex transform
since it includes the real and imaginary part of six oriented wavelet coefficients. Figure 1 shows
the filter band tree structure of the DTDWT. As shown in the figure, the top tree R generates the
real parts of the DTDWT coefficient and bottom tree I generates the imaginary parts of the DT-
DWT coefficient. * denotes a convolution operation, ↓2 means a down sampling by 2. lfr and hfr
are low-pass filter and high-pass filter, which form a Hilbert transformed pair to insure the
perfect reconstruction of the discrete wavelet transform. With the filter band tree structure, the
following wavelet sub-bands which oriented at ±75°, ±150
and ±450 are produced.
Figure1. Filter band Tree Structure of the DT-DWT
yxyx ,1 (LH Wavelet)
yxyx ,2
(HL Wavelet)
yxyx ,3
(HH Wavelet)
From the above equations, x and y represent the low pass filter along with first and second
dimension. Similarly, x and y represent the high pass filter along with first and second
dimension. Also, LH and HL sub-bands are oriented at vertical and horizontal directions
respectively. The HH sub-band is simultaneously oriented along the +45° and−45° diagonal
directions. It is also denoted as HHp (positive oriented direction) and HHn (negative oriented
4. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
4
direction). Similarly, two LH sub-bands and two HL sub-bands are denoted as LHp, LHn, HLp and
HLn. With these six oriented sub-bands, best angular mode is estimated among the 35 angular
modes for intra prediction process.
2. RELATED WORK
HEVC accomplishes the bit rate sparing of almost half under the equivalent visual quality
contrasted with the H.264/AVC. Thus, HEVC turns into a famous video codec. Variants of
HEVC like H.264, H.265, and others are iterations over the base HEVC codec that improve its
efficiency by adding computationally optimum algorithms during the inter-frame and
intra-frame prediction conditions. Researchers from different fields including video
processing specialists, Mathematicians, signal processing experts, to name a few have forayed
into this field in order to further optimize the efficiency of HEVC processing. For instance the
work in [2] adds watermarking capabilities to HEVC by matrix encoding in the DCT (discrete
cosine transform) block of HEVC. This paper achieves data hiding with minimum distortions in
the output video. This indicates that HEVC has some inherent redundancies which can be
reduced in order to further optimize the video encoding/decoding performance.
These redundancies are in terms of inter and intra-frame prediction co-efficient values. This basic
property paves as the motivation for the work in this paper. A motion density-based scheme with
unequal error protection (UEP) is proposed in [3], wherein it is seen that motion density schemes
outperform the existing inter-frame & intra-frame prediction schemes of HEVC.This
performance is evaluated in terms of the capability of the algorithm to find out important frames
from the input video. A higher value is an indicative of better performance for the system. The
proposed approach in [3] outperforms other control unit (CU) based strategies by more than10%.
This approach can be used to evaluate the best quality frames or key-frames.These key-frames
serve as the base-line for inter and intra-prediction in HEVC. Another approach similar to [2], but
directed towards H.265 codec is proposed in [4]. Wherein, the synchronization error is
reduced after two-stage re-compression in H.265 codec. This approach uses spatial
texture analysis for finding out the most suitable embedding blocks. These blocks are then used
in representation mode in order to find out the best pixels for watermarking. The identified
pixels contain some level of redundancy, and thus can be reduced (compressed) without
any significant loss in video quality. This can improve the frame rate and the efficiency of the
HEVC system when operating in the H.265 mode. It is seen that the proposed algorithm performs
well in the presence of any kind of noise, and there by can be used for further redundancy
reduction of HEVC.
The work in [5] is inspired by these approaches in [2,4], and uses a concept of just notable
distortion (JND). A good quality video can be encoded and decoded using the JND concept.
These identified JND points in the video frames can reduce the size of HEVC data by more than
13% on average, and up-to 39% for certain video sequences. The mean opinion score(MOS) was
evaluated for different videos, and it is observed that the approach in [5] has similar performance
to original HEVC algorithm in terms of visual quality, but it has a reduced compressed video
size. A similar work is proposed in [6], wherein the concept of classical secretary problem (CSP)
is used in the rough-mode-decision module of HEVC.Moreover, the CSP is modified using a
dynamic stopping criterion that further enhances the performance by reducing the encoding
delay and marginally increasing the bit-rate performance. It uses the concept of mode
reduction with the help of redundancy evaluation. A similar concept is proposed in [2], [4] & [5],
and is also the base for this research.
HEVC can be extended to 3D videos. The concept of fast-depth map for intra-mode selection in
3D videos is given in [7], wherein the depth is analysed from the different dimensions of the 3D
5. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
5
video. This depth map is used for prediction of intra-mode redundancies, and finally a
compressed video is obtained. Various depth modelling models are proposed in [7], some of them
also use tensor features for homogeneity detection. Due to the use of depth maps, there is a large
reduction in encoding delay, which further improves the encoding and decoding performance for
3D videos. The approach in [7], can further utilize deep learning methods like deep neural
networks as proposed in [8] to optimize its performance. It is observed from[8], that deep net
models like convolution neural networks (CNN) can be trained with different videos to
identify the redundancies in them. This trained model can then be applied to new videos to
optimize their redundancies with minimal computational complexity and improved bandwidth.
They further observe that specialized models like IPCNN can be trained to specifically reduce the
intra-frame redundancies in order to optimize the quality of service(QoS) for HEVC. An
approach that can be facilitated by CNNs is proposed in [9], where in metrics like rate distortion
are evaluated to reduce the complexity of encoding and decoding process. They have used
texture homogeneity between inter-frames and spatio-temporalcorrelation between intra-
frames in order to reduce the encoding time by more than 70% than normal HEVC. Though the
results seem promising, it is advised that researchers perform adue diligence before using this
research in their applications. The work proposes development of fast coding unit and fast
prediction unit in order to improve the efficiency of the HEVC system. While most of these
research models are based on lossy HEVC performance improvement, the work in [10]
uses lossless HEVC using context-based angular & planarintra predictions. It also uses
redundancy reduction in HEVC videos by identifying redundant edges, textures, colours, and
other parameters between neighbouring pixels. They use pixel-level processing for edge and
texture redundancy optimization without increasing any computational complexity. Due to
removal of redundant edges and textures, the resulting video is completely lossless. It can achieve
performance improvement of up to 10% when compared with other standard HEVC models.
Another 3D video optimization algorithm is mentioned in [11], that uses
dynamically configurable depth maps similar to [7]. In [10] the depth maps are not generated
using tensors or hyper-planes, but they use the concept of Rough Mode Decision (RMD). It is
known that RMD is inherited from the texture maps, rather than the depth maps. This RMD
affects the block distortion and the rate distortion of HEVC, and thus can be used for better
HEVC performance. The proposed work achieves 0.1% improvement in Bjontegaard Delta-rate
(BDRate), which indicates that the compression performance is high when compared to normal
HEVC encoding. A similar method like [11] is given in [12], wherein methods like bipartition
modes, intra-picture skip, and DC-only are used to optimise depth map processing. Their work
indicates that depth map processing to identify redundancies using these approaches can reduce
the encoding delay by more than 20%. They also propose that reduction in texture and depth can
be combined to further improve the HEVC performance.
A fast and adaptive mode decision HEVC algorithm can be seen in [13], which uses coding unit
partition for early termination of intra-prediction process. The proposed work in [13]forms the
base our work, wherein this paper also utilizes mode reduction technique similar to[13] for a
better HEVC performance. They have reduced the number of modes from 35 to 11,which
improve the Bjontegaard delta rate by 1.7%, but reduce the average delay by more than50%.
Thus, giving a big bump in terms of final video performance. The work in [13] also utilizes CU
partitioning based on number of coding bits, which further helps in improving the system
performance. This work can further be improved by addition of RD cost as a measure for mode
reduction as proposed in [14]. RD cost can be an early prediction metric for reducing the
number of intra modes from HEVC. Due to inclusion of RD cost in evaluation of mode
reduction, a performance improvement of more than 25% can be expected when
compared to usual HEVC system, which can further be improved by adding machine learning
mechanisms like the one proposed in [15] for adaptive CU size decisions. The work
6. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
6
i[15]proposes the use of complexity classification for training the machine learning model. This
complexity classification method uses parameters like CU size, CU partitions and
rate distortion to train a support vector machine (SVM) algorithm. This SVM algorithm solves
a2-class classification problem, and classifies each intra-frame into required and non-required.
All the non-required frames are dropped, and finally we get the compressed video with minimal
complexity. The proposed ML algorithm reduces the complexity by more than 60%,and thereby
speeding up the entire process of HEVC compression and decompression. The approach in [15]
can further be modified using the techniques mentioned in [16]. From there view done in [16], we
can observe that dynamic support vector machines (DSVMs), which can be destroyed and re-
created for every inter-frame and intra-frame model prediction are the best option for HEVC
encoding. These models must be integrated with existing HEVC approaches to further improve
their efficiency. The SVM models can also be used as the final flat layer for CNN models
described in [17]. This replacement can enhance the performance of the existing CNN models by
more than 20%, and also reduce the complexity of processing the HEVC videos. Moreover, the
CNNs can be replaced by deep CNNs, as proposed in [18]to further reduce the intra-mode
redundancies. These redundancies are easily analysed by deep CNNs, and thereby can be further
reduced with the help of models like GoogLe Net or VG GNet. A combination of layers
like convolutional, ReLU, convolutional, ReLU,convolutional, Max Pooling, fully
connected, ReLU and finally fully connected can be used for a better prediction performance.
Another experimental work is described in [19], where in a SAD unit is proposed to compress
ultra HD 8K videos. This can be used as a future work for deep CNN models. The next section
describes our proposed PSO-based approach for HEVC processing
3. PROPOSED PSO COMBINED WITH CELLULAR AUTOMATA
MODEL FOR INTRA FRAME PREDICTION
PSO is applied inside the intra-frame prediction process in order to optimize the PSNR at the
decoding side. This model requires a certain amount of delay for the first one or two searches, but
it is compensated as the number of frames are processed. Due to this self-learning nature of the
algorithm, it can be integrated inside the intra-frame prediction block of HEVC. The CA
technique further optimizes the performance of the existing PSO. It does so by reducing the
randomized search space of the PSO via rules of CA. The application of CA to PSO is done with
the help of the following rules in CA.
Figure 2. CA rules
7. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
7
From the figure 2, we can observe that based on the input patterns the CA selects one output.
This output in our case is the limited range of global and local best particles. Using CA the range
of the local and global best particles is limited. The following algorithm describes the working of
PSO and CA with HEVC.
Initialize the PSO and CA parameters, namely,
o Number of particles = P
o Number of iterations = I
o PSO constants C1 and C2
o CA_LIMIT_PBEST – Is the limited number of HEVC modes to be used for
learning from PBEST (should be less than half modes)
o CA_LIMIT_GBEST – Is the limited number of HEVC modes to be used for
learning from GBEST (can be more than half modes)
Let the structure for the 64x64 CTU block be defined as follows,
Figure 3. Block division process
Here ‘a’ is the main block, ‘b0 … b3’ are the divided blocks, and ‘c0 … c15’ are the sub-
divided blocks and finally ‘d0 … d63’ are the 64 CTU blocks
Let’s call this combination a particle, and in each solution generate random particles for
operation
Considering the 3-level depth as shown in the figure, the particle will have 21 bits, as
follows,
P = a0,b0,b1,b2,b3,c0,c1,c2…c15
where,
a0 = 0, when CU is not split, else a0=1
bi= 0, when a0=0 or CU is not split, else bi=1
ci=0, when a0=0, corresponding b=0, or CU is not split, else ci=1
Here a, b, and c represent the splitting decisions for depth 0, depth 1, and depth 2,
respectively
The possible values for a is 0 (non-splitting) and 1 (splitting). The possible values for b
are null if a is 0, 0 (non-splitting), and 1 (splitting). The possible values for c are null if
its corresponding parent b is 0, 0 (non-splitting), and 1 (splitting). It should be noted that
the proposed data structure is composed of a group of dependent genes. Therefore, the
total number of possible CU partitioning patterns P is calculated as,
where d ∈ {1, 2, 3} is the maximum CU depth and the mod is the modulo operation for
finding the remainder. If the maximum CU depth is 2 and 3, the total number of possible
partitioning patterns is only 17 as shown in the following figure, and 83,522 even there
are five genes and 21 genes to represent the CU partitioning pattern of a 64 × 64 CTU,
respectively.
8. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
8
Figure 4. Frame division process
Initialize the particles using randomized values of a, b and c (from the figure 4 shown in
the previous step)
For each iteration perform the following,
o For each particle, find the fitness of the particle using the following equation,
where F is the RD cost-oriented fitness function, a, bi, and cj are the values of one gene,
four genes, and sixteen genes of each chromosome to represent the splitting decision at
depth 0, 1, 2, and 3, respectively. RDCost(a), RDCost(bi), RDCost(cj), and RDCost(dk)
are the RD cost values of one CU, 4 CUs, 16 CUs, and 64 CUs at depth 0, 1, 2, and 3,
respectively.
o If the fitness value is better than the best fitness value (pBest) in history, then set current
value as the new pBest
o Choose the particle with the best fitness value of all the particles as the gBest
o For each particle follow the given steps,
Evaluate the velocity of the particle using the following equation,
v = v + C1 * random (CA_LIMIT_PBEST) * (pBest - currentFitness) + C2 * random
(CA_LIMIT_GBEST) * (gBest - currentFitness)
Update the position of the particle using the following equation,
presentParticle = presentParticle + v
9. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
9
At the end of the last iteration, use the particle with gBest fitness value as intra-frame
prediction particle.
The best particle is replaced in the output stream of HEVC as the encoded block. Once the
particle selection is done then a dual-tree complex wavelet transform block is used in order to
reduce the modes of the system from 35 modes to 8 modes. Initially, Dual Tree Discrete Wavelet
Transform (DT-DWT) [25] is applied to the optimum selected block by PSO algorithm. By
applying this transform, six oriented wavelet sub-bands are generated. Among the sub-bands, two
LH and two HL sub-bands are used find the direction or angle of the texture in a block. Polarity
of the texture angle is estimated with two HH Sub-bands. With the direction and the angle of the
texture, a mode is determined that is closer to the actual best mode. For best mode selection, the
four modes around this determined mode (Modedeter) and also DC and planar modes are
considered as a final candidate list. This candidate list is forwarded to the process of Rate
Distortion Optimization (RDO). With the RDO, the mode with minimum rate distortion cost is
selected as the best mode (Modebest). Based on this selected Modebest , the encoder encodes the
video frames. Decoder decodes it and output of the decoder is analyzed for performance.
The flow diagram for the proposed system can be observed in the figure 5. From the figure we
can observe that the output of PSO and CA system is given to the DT-CWT based HEVC
encoder thereby hybridizing the system with the already existing high efficiency encoder.
Moreover, the output of the decoder is used for performing result evaluation of PSNR and delay.
These outputs and their analysis are showcased in the next section.
Due to the combination of the proposed intra-frame prediction model with the dual tree complex
wavelet transform (DTCWT), the overall effectiveness of the system is improved. The combined
model is able to reduce the search space, and also reduce the number of modes needed for
encoding. Thereby, giving a dual level advantage to the system under test.
Figure5. Overall flow of the system
Usually rate distortion optimization (RDO) is done in DTCWT. In this work, the RDO is not
done, but the RD cost is evaluated. We would request readers not to get confused between the
two processes. As PSO is only using RD cost in order to select the best blocks for
encoding/decoding, while DTCWT uses RD cost for optimization with the help of mode
reduction.
10. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
10
4. RESULT EVALUATION
In this paper, HEVC intra prediction algorithm using PSO as well as PSO with CA is proposed.
The results are compared with [25] which use dual tree complex wavelet transform for intra
prediction. Simulation is carried out in JAVA Net Beans Software. The results were compared
for delay and PSNR values for different videos. These results were compared for HEVC DT-
CWT [25] , HEVC with PSO and HEVC with PSO+CA as shown in Table1 which showcases
the delay results obtained for the videos on the given algorithms. Author in [25] proposes a novel
approach to reduce the modes from 35 to 8 and then selecting a optimum mode. We have
proposed PSO with CA to reduce time consumption in decision making process along with the
further application of dual tree complex wavelet transform to reduce the
computational complexity. Similarly, a comparison of PSNR for these algorithms was performed,
and the results are tabulated in Table 2
.
Table 1. Comparison of Time delays for the proposed algorithm
From the results we can observe that the proposed algorithm is able to reduce the delay and
improve the PSNR of the existing dual-tree based HEVC system. We also evaluated the average
values of PSNR and delay for both the algorithms, and observed the following results.
11. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
11
Table 2. Comparison of PSNR for algorithms
The average results of the algorithm are tabulated in table 3.
Table 3. Average results Time Delays and PSNR of the algorithm
Comparison of Computational complexity in terms of time delay as well as video quality in terms
of PSNR are tabulated in Table 4 for different testing video sequences. From the
comparison table, it is very clear that both the proposed algorithm that is HEVC processing using
PSO as well as using PSO and CA provides better video quality and reduction in time
complexity. The conclusion and some interesting observations from these results are mentioned
in the next section.
Table 4. Comparison among stat-of-the-art algorithms
12. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
12
Figure 6. Comparison of PSNR Results
As shown in Figure 6, it is clear that proposed PSO and PSO with CA provides almost similar
video quality which is comparatively better to the existing algorithm using dual tree complex
wavelet transform.
Figure 7. Comparison of Time Delay Results
13. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
13
As shown in the Figure 7, proposed PSO with CA algorithm provides better time complexity
reduction. The conclusion and some interesting observations from these results are mentioned in
the next section
5. CONCLUSION AND FUTURE WORK
The increased computational complexity in HEVC is a major problem especially for power
constrained devices or real-time applications especially for high-resolution videos. Therefore, it is
highly desirable to optimize the encoding process for computational complexity reduction while
maintaining the coding efficiency of HEVC. Fast intra prediction algorithm using PSO with CA
is proposed in this paper. The experimental results are conducted for various test video. The
results are evaluated based on encoding time and peak signal to noise ratio (PSNR). The
proposed PSO+CA based HEVC performs faster than the existing HEVC algorithm in terms of
overall delay by 40%. It also outperforms the existing method by 12 % in terms of PSNR. The
results of comparative experiments demonstrate that the proposed algorithm can no doubt
effectively reduce the computational complexity of HEVC Encoder while maintaining good
video quality.
All these advantages are evident due to the extensive intra-frame prediction phase, where in most
of the mapping process and calculations are pre-dominantly done. Another reason for such a huge
bump in performance is the presence of the light weighted execution phase.
There are many other ways to explore in the CU early termination, mode reduction and fast
intraprediction in the intra prediction area as suggested by literature. In future, many of these
methods can be combined, or if needed, one method may be replaced by a new method and
encoding time gains can be explored. Convolution Neural Network Model, SVM machine
learning approach can also be applied in order to reduce time complexity. Similar Intra
Prediction algorithms can be developed for fast inter-prediction resulting in lessen coding time
and reduced complexity. Future research can be conducted to reduce computational
complexity in Quad tree structure means dividing CTUs up to CU and PU both for the intra and
inter coding can be improved to obtain much higher reduction of encoding time, better bit rate
and PSNR. The aim should be to reduce the overall complexity of HEVC encoder suitable for
hand held devices as well as transmission with limited computing resources.
In future this work can be further improved by evaluating the performance for higher bit rate
videos. These videos are a bit complex to map, and thus might be a need of multiple pre –
execution steps before a required level of efficiency is achieved. Moreover, in order to really
optimize the performance further, researchers can use quantum computing for processing, and
develop quantum computational layers in order to evaluate its performance, and apply the
proposed machine.
REFERENCES
[1] G. Sullivan et al., “Overview of the high efficiency video coding (HEVC) standard”, IEEE
Transactions on Circuits and Systems for Video Technology, Vol. 22, No. 12, pp 1649-1668,
December 2012.
[2] T. Wiegand, G. J. Sullivan, G. Bjontegaard, and A. Luthra, ‘‘Overview of the H.264/AVC video
coding standard,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, no. 7,
pp. 560–576, Jul. 2003.
[3] JCT-VC, HM Software. Accessed: Nov. 5, 2016. [Online]. Available:
https://hevc.hhi.fraunhofer.de/svn/svn_HEVCSoftware/tags/HM-16.5/
14. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
14
[4] S. Cho and M. Kim, ‘‘Fast CU splitting and pruning for suboptimal CU partitioning in HEVC intra
coding,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 23, no. 9, pp. 1555–
1564, Sep. 2013.
[5] B. Min and R. C. C. Cheung, ‘‘A fast CU size decision algorithm for the HEVC intra encoder,’’ IEEE
Transactions on Circuits and Systems for Video Technology, vol. 25, no. 5, pp. 892–896, May 2015.
[6] T. Li, M. Xu, and X. Deng, ‘‘A deep convolutional neural network approach for complexity reduction
on intra-mode HEVC,’’ in Proc. IEEE Int. Conf. Multimedia Expo (ICME), Jul. 2017, pp. 1255–
1260.
[7] T. Zhang, M.-T. Sun, D. Zhao, and W. Gao, ‘‘Fast intra-mode and CU size decision for HEVC,’’
IEEE Transactions on Circuits and Systems for Video Technology, vol. 27, no. 8, pp. 1714–1726,
Aug. 2017.
[8] L. Shen, Z. Liu, X. Zhang, W. Zhao, and Z. Zhang, ‘‘An effective CU size decision method for
HEVC encoders,’’ IEEE Trans. on Multimedia, vol. 15, no. 2, pp. 465–470, Feb. 2013.
[9] X. Hou and Y. Xue, ‘‘Fast coding unit partitioning algorithm for HEVC,’’ in Proc. IEEE Int. Conf.
Consum. Electron. (ICCE), Jan. 2014, pp. 7–10.
[10] I. Zupancic, S. G. Blasi, E. Peixoto, and E. Izquierdo, ‘‘Inter-prediction optimizations for video
coding using adaptive coding unit visiting order,’’ IEEE Trans.on Multimedia, vol. 18, no. 9, pp.
1677–1690, Sep. 2016.
[11] X. Shen, L. Yu, and J. Chen, ‘‘Fast coding unit size selection for HEVC based on Bayesian decision
rule,’’ in Proc. of Picture Coding Symp., May 2012, pp. 453–456.
[12] K. Duan, P. Liu, K. Jia, and Z. Feng, ‘‘An adaptive quad-tree depth range prediction mechanism for
HEVC,’’ IEEE Access, vol. 6, pp. 54195–54206, 2018.
[13] X. Shen and L. Yu, ‘‘CU splitting early termination based on weighted SVM,’’ EURASIP Journal
proceedings of Image and Video Processing, vol. 2013, no. 1, p. 4, Dec. 2013.
[14] G. Corrêa, P. A. Assuncao, L. V. Agostini, and L. A. da Silva Cruz, ‘‘Fast HEVC encoding decisions
using data mining,’’ IEEE Transactions on Circuits and Systems for Video Technology, vol. 25, no.
4, pp. 660–673, Apr. 2015.
[15] Y. Zhang, S. Kwong, X. Wang, H. Yuan, Z. Pan, and L. Xu, ‘‘Machine learning-based coding unit
depth decisions for flexible complexity allocation in high efficiency video coding,’’ IEEE
Transaction on Image Processing, vol. 24, no. 7, pp. 2225–2238, Jul. 2015.
[16] A. Heindel, T. Haubner, and A. Kaup, ‘‘Fast CU split decisions for HEVC inter coding using support
vector machines,’’ in Proc. Picture Coding Symp. (PCS), Dec. 2016, pp. 1–5.
[17] L. Zhu, Y. Zhang, S. Kwong, X. Wang, and T. Zhao, ‘‘Fuzzy SVM-based coding unit decision in
HEVC,’’ IEEE Transaction on Broadcasting, vol. 64, no. 3, pp. 681–694, Sep. 2018.
[18] H.-S. Kim and R.-H. Park, ‘‘Fast CU partitioning algorithm for HEVC using an online-learning-
based Bayesian decision rule,’’ IEEE Transactions on Circuits and Systems for Video Technology,
vol. 26, no. 1, pp. 130–138, Jan. 2016.
[19] L. Zhu, Y. Zhang, Z. Pan, R. Wang, S. Kwong, and Z. Peng, ‘‘Binary and multi-class learning based
low complexity optimization for HEVC encoding,’’ IEEE Transaction on Broadcasting, vol. 63, no.
3, pp. 547–561, Sep. 2017.
[20] M. Xu, T. Li, Z. Wang, X. Deng, R. Yang, and Z. Guan, ‘‘Reducing complexity of HEVC: A deep
learning approach,’’ IEEE Transaction on Image Processing, vol. 27, no. 10, pp. 5044–5059, Oct.
2018.
[21] Z. Pan, S. Kwong, M.-T. Sun, and J. Lei, ‘‘Early MERGE mode decision based on motion estimation
and hierarchical depth correlation for HEVC,’’ IEEE Transaction on Broadcasting, vol. 60, no. 2, pp.
405–412, Jun. 2014.
[22] Z. Pan, J. Lei, Y. Zhang, X. Sun, and S. Kwong, ‘‘Fast motion estimation based on content property
for low-complexity H.265/HEVC encoder,’’ IEEE Transaction on Broadcasting, vol. 62, no. 3, pp.
675–684, Sep. 2016.
[23] M. Mitchell, An Introduction to Genetic Algorithms. Cambridge, MA, USA: MIT Press, 1996.
[24] Z. Michalewicz, Genetic Algorithms+Data Structures=Evolution Programs, 3rd ed. Berlin, Germany:
Springer-Verlag, 1996.
[25] Farhad Pakdaman, Mahmoud-Reza Hashemi, and Mohammad Ghanbari, “Fast and Efficient Intra
Mode Decision for HEVC, Based on Dual-Tree Complex Wavelet,” in Proc. Of Article in
Multimedia Tools and Applications , May 2016
[26] Romberg J-K, Hyeokho C, Baraniuk R-G (2001) Multiscale edge grammars for complex wavelet
transforms. Proc. International Conference on Image Processing (ICIP), p. 614–617
15. The International Journal of Multimedia & Its Applications (IJMA) Vol.12, No. 1/2/3/4, August 2020
15
[27] Selesnick I-W, Baraniuk R-G, Kingsbury N-C (2005) The dual-tree complex wavelet transform.
IEEE Signal Process Mag 22(6):123–151
AUTHORS
Swati Vinod Sakhare
She is a PhD scholar at The Sardar Vallabhbhai Patel National Institute of
Technology (SVNIT), Surat. She received her graduate degree in Electronics and
Telecommunication from Government College of Engineering, Amravati and
postgraduate degree in Electronics from B. D. College of Engineering Sevagram,
Wardha. Her research interests include Video processing, Digital Image Processing,
Microprocessors and Optical Fiber Systems. She is working in G. H. Raisoni College
of Engineering and Management, Amravati since 2012. She has guided several PG
students for their project work and coordinated conference, STTPs, workshops and
several technical programs for students, faculty members and working professional.
Dr. Mrs. Upena D. Dalal
She is currently Professor at Electronics and Communication Engineering department,
at The Sardar Vallabhbhai Patel National Institute of Technology (SVNIT), Surat. She
received her doctoral degree in era of wireless communication and post graduation in
Electronics and Communication Systems in which she was gold medalist. She has
vast teaching experience of 24 years and her major subjects are Cellular Technology,
Wireless Communication and Fiber optic networks. She is author of well-known book
on “Wire- less Communication” by Oxford University press (2009) and “Selectively
Imposed Pilot based Channel Estimation” by VDM publications, Germany (2010).
She is also co-author of book on “WiMAX Developments” from Intec web, Viena,
Austria (2010). She has guided several PhD and MTech students and coordinated
conference, STTPs, workshops and numerous technical programs for betterment of
students, faculty and research fraternity