SlideShare a Scribd company logo
1 of 7
Download to read offline
IOSR Journal of Engineering (IOSRJEN) www.iosrjen.org 
ISSN (e): 2250-3021, ISSN (p): 2278-8719 
Vol. 04, Issue 08 (August. 2014), ||V4|| PP 01-07 
International organization of Scientific Research 1 | P a g e 
Video Quality Analysis for H.264 Based on Human Visual System Subrahmanyam.Ch1, Dr.D.Venkata Rao2, Dr.N.Usha Rani3 1(Research Scholar, Dept of ECE, Vignan University, Guntur, India) 2(Principal, Narasaraopeta Institute of Technology,Guntur, India) 3(Head of the Department, Dept of ECE, Vignan University, Guntur, India) Abstract: - The quality of video, such as the Mean Squared Error (MSE), Weighted Mean Squared Error (WMSE) or the Peak Signal-to-Noise Ratio (PSNR), not always correspond with visual observations. Structural Similarity based video quality assessment is proposed under the assumption that the Human Visual System (HVS) is highly adapted for extracting structural information from video. While the demand on high resolution quality increases in the media industry, Color loss will make the visual quality different. In this paper, we propose an improved quality assessment method by adding color comparison into structural similarity measurement system for evaluating video quality. Then we separate the task of similarity measurement into four comparisons: luminance, contrast, structure and color. Experimental results show that the predicted quality scores of the proposed method are more effective and consistent with visual quality than the classical methods using two different distortion types of video sets. Keywords: - MSE, WMSE, PSNR, HVS, luminance, contrast 
I. INTRODUCTION 
Visual information is very important for hearing-impaired people, because it allows them to communicate personally using the sign language. In our research, we focused on the fact that some parts of the person using the sign language are more important than others. We presented a visual attention model based on detection of low-level features such colour, intensity and texture and combination with the prior knowledge – in our case information about skin in image. Information about the visually relevant parts allows us to design an objective metric for this specific case. We presented an example of an objective metric based on human visual attention and detection of salient object in the observed scene. There is a huge use of semantic information in video. The ability to detect features is an interesting challenge by itself, but it takes on added importance to the extent it can serve as a reusable, extensible basis for query formation and search. Nowadays, the researchers focus mainly on solving the problems of finding the semantic information in video sequences. 
II. INTRODUCTION TO H.264 
H.264/AVC is the state-of-the-art video coding standard developed by Joint Video Team (which comprises of ITU-T VCEG and ISO/IEC MPEG). Primary aim of H.264/AVC project was to develop a standard which provides same video quality at about half the bit rates of previous standards like MPEG-2, H.263 etc. Additional goals were lowering the complexity and increasing the flexibility so that the standard can be applied to wide variety of applications over different networks and systems. Its first version was released in May 2003 under the name H.264/AVC by ITU-T. Since then many versions has been released which has improved the standard in a variety of ways. These new versions are known as Fidelity range extensions (FRExt). The draft for FRExt was completed in September 2004. Scalable video coding (SVC) is another aspect of H.264/AVC project which has gained a lot of popularity. The aim of SVC is to use a single bit-stream which could serve different networks and different devices. SVC extension was completed in November 2007. In sync with previous standards, H.264/AVC defines only the syntax of encoded bit-stream and the method of decoding it. But this limitation of standard provides maximum freedom to optimize implementations in a manner appropriate to specific applications. Fig 2.1 shows the scope of H.264/AVC standard.
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 2 | P a g e 
Fig 2.1: Scope of video coding standardization 
Since the first video coding standard (H.261), the functional blocks of the basic video coding algorithm 
remains the same. But the major difference lies in the details of each functional block. As a result each new 
video coding standard became more comprehensive and complex. Hence the video coding algorithm of 
H.264/AVC is also similar to previous standards but with an exception of de-blocking filter. Now let’s have a 
look at the codec (enCOder and DECoder pair) of H.264/AVC; starting with the encoder. 
2.1.1 Encoder: 
A video encoder is a device or software which performs compression over the bit-stream; as a result the 
size of the encoded bit stream greatly reduces in size. Before elaborating more about video encoder, some basic 
terms of video processing are introduced which will help a lot in explaining various functional blocks of the 
encoder afterwards. 
 Video Stream: A video stream consists of large number of still images, when these images are run at a rate 
of 25/30 frames per second a video is formed. These still images can be classified as frames or fields 
depending upon the procedure of scanning. 
 Macroblock (MB): Macroblock is special word in video coding used to describe a small region in a frame 
or field (usually the size of a MB is 16x16 pixels). Therefore, each frame/field is divided into large number 
of macroblocks. 
 Pixel (PICture ELement): is the smallest piece of information in a frame/ image. 
Fig 2.2 Basic encoding structure of H.264/AVC for a Macroblock
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 3 | P a g e 
Any video coding technique tries to reduce the size of the encoded bit stream by removing the redundancy present in the video sequence. Any video sequence has three types of redundancies temporal, spatial and statistical. To reduce temporal redundancy, similarities between the neighboring video frames are exploited by constructing a prediction of current frame. Normally a prediction is formed using one or two previous or future frames and is improved by compensating differences between different frames. The output of this whole process is a residual frame. These residues are processed by exploiting similarities between the neighboring pixels to remove spatial redundancy. Reduction is achieved by applying a transform (like DCT) and quantizing their results. The transform converts the sample into another domain where they are represented as transform coefficients; these coefficients are quantized to remove insignificant values, thus leaving a small number of coefficient values which provides more compact representation of residual frame. To remove statistical redundancy, an encoder uses various techniques (Huffman coding, Arithmetic coding etc) which assign short binary codes to commonly occurring vectors. Due to the reduction of redundancies the size of encoded bit- stream greatly reduces in size. Therefore, any video encoder consists of these three functional modules. Fig2.2 shows the H.264/AVC encoder, as is evident from the figure that H.264/AVC Encoder has a decoder loop incorporated in it. The main reason for including the decoder is to provide reference pictures which are very necessary for the prediction process. Therefore the encoder has two data flow paths, forward path and backward or reconstruction path. 
 Forward path: Here the normal encoding process is executed. An input frame is processed in terms of MB’s (16x16). Depending upon the type of MB, prediction technique is selected. There are two types of prediction techniques, intra prediction and inter prediction. In intra prediction, prediction is calculated from the samples of same frame which are previously encoded, decoded and reconstructed, whereas in inter prediction, prediction of current frame is calculated from the Motion compensation prediction of reference pictures which could be either previous or future frames. So when an ‘I’ MB is processed intra prediction is selected and when a ‘P/B’ MB is processed inter prediction is selected. Inter prediction techniques relies upon the motion estimation and motion compensation techniques for accurate prediction. These techniques are very instrumental in reducing the temporal redundancy between the frames. 
The goal of motion estimation technique is to find a best match for a block of current frame in reference frame. To find the best match, a search is performed in the reference frame. This search is carried out by comparing the block in current frame with the blocks in search area (usually centered on and around the current block position) of reference frame. The block or area which gives the lowest residual energy is considered as the best match region. Thus chosen best match region is called as ‘predictor’ for the block in current frame. In motion compensation, predictor is subtracted from the current block resulting in residues. So obtained residues are block transformed, scaled and quantized to obtain quantized coefficients. These quantized coefficients along with control data and motion vector information is entropy coded. Entropy encoded coefficients along with side information necessary for decoding each macro block at the decoder constitutes the encoded bit stream. 
 Reconstruction path: Here decoding and reconstruction of encoded frames is performed. To reconstruct the encoded frames, quantized coefficients are scaled (De-quantized) and inverse transformed to obtain residues which are added with the prediction to obtain unfiltered frame. This unfiltered frame contains blocking artifacts; blocking artifacts arise mainly due to the usage of block based integer DCT (Discrete Cosine Transform). To smooth out these blocking artifacts, de-blocking filter is applied over the frame. These filtered frames are stored as reference frames. 
2.1.2 Decoder: 
When the decoder receives the encoded bit stream, it entropy decodes them to produce a set of quantized coefficients which are reordered, scaled and inverse block transformed to obtain residues. Using the header information received in the encoded bit-stream, decoder generates the prediction. Residues are added to the prediction to obtain the unfiltered version of the original frame. Unfiltered frame contains blocking artifacts which deteriorates the video quality. De-blocking filter is applied over the unfiltered frame to remove the blocking artifacts and enhance the video quality of filtered frame. Filtered frame are stored as reference frames which are utilized in prediction process.
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 4 | P a g e 
Entropy 
decoding 
Scaling Reorder 
Inverse 
Transform + 
De-blocking 
filter 
Motion 
Compensation 
Intra 
pediction 
Reference 
frames 
Reconstructed 
frames + 
+ 
Intra 
Inter 
Prediction 
Residues 
NAL 
Unfiltered 
frame 
Fig 2.3: H.264 Decoder 
III. PROPOSED REFERNEC MODEL FOR VIDEO QUALITY MONITORING 
Three methodologies representing different measurement strategies for the assessment of the quality of video 
have been defined: 
– methodology using the complete video reference (double-ended); 
– methodology using reduced reference information (double-ended); 
– methodology using no reference signal (single-ended). 
The design and the development of a video quality monitor should consider a general structure of the 
measurement procedure for reduced reference and single-ended methodologies. The reference model is 
composed of the following four layers: 
– Measurement methodology defines the class or the strategy relative to the application requirement; 
– Measurement method is composed of a set of modules, algorithmic and associated ones, implemented to 
process inputs such as original signals or processed reference data, and provide output results such as 
processed reference data, level of impairment or final quality notation; 
– Algorithmic module(s) is the basic block of signal processing functions composing the method. It composes 
the core of the method from which the final objective qualification is delivered; 
– Associated module(s) is an additional function that aids the algorithmic module(s) in its operation by 
addressing such issues as dating, synchronization, presentation of data, etc. 
Each one of those methodologies offers a specific approach relative to its operational use. Each methodology is 
well adapted to a specific field of applications. This is due to its technical implementation constraints: 
– relevance of the measured data, 
– availability of the reference, 
– synchronization of original and impaired signals or data, 
– transmission channel for reference data, 
– real-time implementation, 
– in-service use, etc. 
Post-processing algorithms are the most popular methods for improving the quality of the image and 
video and eliminate the annoying blocking artifact. On the other hand, the post-processing algorithms can 
achieve deblocking without the original image and video, so the standard need not to be modified. 
Reduction of Blocking Artifacts in DCT Domain 
We introduce a post-processing filtering algorithm in DCT 
domain. We define the block , 
, ( , ) k l 
m n b u v and , 
, ( , ) k l 
m n B u v first. , ( , ) m n b u v is the (m,n)-th 8х8 block in the 
compressed image, and , ( , ) m n B u v is the DCT coefficients of , ( , ) m n b u v . , 
, ( , ) k l 
m n b u v is the shifted block with 
displacement k pixel in the x direction and displacement l pixels in the y direction with respective to block
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 5 | P a g e 
, ( , ) m n b u v , and , 
, ( , ) k l 
m n B u v is the DCT coefficients of the block , 
, ( , ) k l 
m n b u v . One example is shown in Fig. 
3.1. 
m 1,n b  
m,n b 
m 1,n 1 b   
m,n 1 b  
1, 1 
m,n b  
Fig. 3.1 Example of shifted block 
, 
, ( , ) k l 
m n b u v 
In the original image, the neighboring DCT coefficients at the same frequency are very similar and do 
not vary radically within a small range. Thus, we can apply low pass filter to the DCT coefficients at each 
frequency to filter the high frequency parts resulting from blocking effect. However, this method may blur the 
real edges in the original image, so we must have the mechanism to detect activity of the block and apply the 
filter with corresponding strength. 
DCT-domain filtering is applied to revise the block , ( , ) m n B u v to obtain the new DCT coefficients 
 
Bm,n (i, j) . 
 , 
, , , 
1 
( , ) ( , ) 
h h 
k l 
m n k l m n 
k h l h 
B i j w B u v 
W   
   
, 
h h 
k l 
k h l h 
W w 
  
   
The post-filtering works in different ways for the blocks with different activities. 
For blocks with low activity, the blocking artifact is more noticeable, so we apply strong filtering to smooth the 
high frequency components. The filter coefficients are defined. 
, 1, , 2,..., 2 k l w  k l   
For blocks with high activity, the blocking artifact is less noticeable, so we apply filtering with less strength 
to smooth blocking artifact and preserve the real edge. The filter coefficients are defined. 
, 
3, for ( , ) (0,0) 
1, k l 
k l 
w 
otherwise 
  
 
IV. EXPERIMENTAL RESULTS 
In this, we will show some experimental results. The video used in the experiment is 
“xylophone.mpg”, which consists of original high-resolution YUV format video files. It includes upto 10 
frames of re sized video. The details of the video frame values are shown. 
Step wise Process: 
Encoding I-Frame: 1 Encoding P Frame: 2 Encoding P Frame: 3 
Encoding P Frame: 4 Encoding P Frame: 5 Encoding P Frame: 6 
Encoding P Frame: 7 Encoding P Frame: 8 Encoding P Frame: 9 
Encoding P Frame: 10
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 6 | P a g e 
Fig.4.1. Encoding Video process of xylophone.mpg video (10 frames) Decoding I-Frame: 1 Decoding P Frame: 2 Decoding P Frame: 3 Decoding P Frame: 4 Decoding P Frame: 5 Decoding P Frame: 6 Decoding P Frame: 7 Decoding P Frame: 8 Decoding P Frame: 9 Decoding P Frame: 10 Fig.4.2. Decoding Video process of xylophone.mpg video (10 frames) Fig.4.3. PSNR = 18.8015 Fig.4.4. PSNR = 1.2104 
V. CONCLUSION 
Emerging new technologies demand tools for efficient indexing, browsing and retrieval of video data, which causes rapid expansion of areas of research where the semantic information is used. New methods that work with semantic information in image, video, and audio are developed frequently these days, which means that our list of methods is not final. Nevertheless, we picked the most used ones and tested them. We brought an overview of the methods for detecting and describing interesting points in video, modeling visual attention and saliency. We have tested the H.264 video formats. We have evaluated the use of visual saliency for compression of video sequences.
Video Quality Analysis for H.264 Based on Human Visual System 
International organization of Scientific Research 7 | P a g e 
REFERENCES 
[1] A. C Bovik. Introduction to Image and Video Quality Assessment[OL]. [2009-3-17]. http://live.ece. utexas.edu /research/quality/intro.htm. [2] C J van den Branden Lambrecht, Ed. Special Issue on Image and Video Quality Metrics [J]. Signal Process., 1998, 70(3): 155-294. [3] VQEG. Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment [OL]. (2000-3-15) [2009-3-17]. http://www.its.bldrdoc.gov/vqeg/projects/ frtv_phaseI/COM-80E_final_report.pdf. [4] T N Pappas and R J Safranek. Perceptual Criteria for Image Quality Evaluation. In: A. C Bovik, ed. Handbook of Image &Video Processing [M]. San Diego: Academic Press, 2000: 669-684. [5] Z Wang and A C Bovik. A Universal Image Quality Index [J]. IEEE Signal Processing Letters, 2002, 9(3): 81-84. [6] G H Chen, C L Yang, L M Po, et al. Edge-based Structural Similarity for Image Quality Assessment [C]. ICASSP, Toulouse, France, 2006: II-933-II936. [7] Commission Internationale d'Eclairage. CIE Publication No.15.2, Colorimetry [S]. 2nd Ed. Vienna: CIE Central Bureau, 1986. [8] Commission Internationale d'Eclairage. CIE Publication No. TC1. 47, Hue and Lightness Correction to Industrial Colour Difference Evaluation [S]. Vienna: CIE Central Bureau, 2000. 
[9] VQEG. Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment, Phase Ⅱ VQEG [OL]. (2003-8-25) [2009-3-17]. http://www.its.bldrdoc.gov/ /projects/frtv- _phaseII/downloads/VQEGII_Final_Report.pdf.

More Related Content

What's hot

IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...
IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...
IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...IRJET Journal
 
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...TELKOMNIKA JOURNAL
 
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODER
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERHARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODER
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERcscpconf
 
10.1.1.184.6612
10.1.1.184.661210.1.1.184.6612
10.1.1.184.6612NITC
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Editor IJARCET
 
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...CSCJournals
 
Review On Different Feature Extraction Algorithms
Review On Different Feature Extraction AlgorithmsReview On Different Feature Extraction Algorithms
Review On Different Feature Extraction AlgorithmsIRJET Journal
 
Compression: Video Compression (MPEG and others)
Compression: Video Compression (MPEG and others)Compression: Video Compression (MPEG and others)
Compression: Video Compression (MPEG and others)danishrafiq
 
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSION
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSIONSECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSION
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSIONsipij
 
ANALYSING JPEG CODING WITH MASKING
ANALYSING JPEG CODING WITH MASKINGANALYSING JPEG CODING WITH MASKING
ANALYSING JPEG CODING WITH MASKINGijma
 
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...INFOGAIN PUBLICATION
 
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...csandit
 
Efficient video compression using EZWT
Efficient video compression using EZWTEfficient video compression using EZWT
Efficient video compression using EZWTIJERA Editor
 

What's hot (16)

H43044145
H43044145H43044145
H43044145
 
IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...
IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...
IRJET- Segmenting and Classifying the Moving Object from HEVC Compressed Surv...
 
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...
 
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODER
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODERHARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODER
HARDWARE SOFTWARE CO-SIMULATION OF MOTION ESTIMATION IN H.264 ENCODER
 
10.1.1.184.6612
10.1.1.184.661210.1.1.184.6612
10.1.1.184.6612
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124
 
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...
Rate Distortion Performance for Joint Source Channel Coding of JPEG image Ove...
 
Review On Different Feature Extraction Algorithms
Review On Different Feature Extraction AlgorithmsReview On Different Feature Extraction Algorithms
Review On Different Feature Extraction Algorithms
 
Compression: Video Compression (MPEG and others)
Compression: Video Compression (MPEG and others)Compression: Video Compression (MPEG and others)
Compression: Video Compression (MPEG and others)
 
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSION
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSIONSECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSION
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSION
 
ANALYSING JPEG CODING WITH MASKING
ANALYSING JPEG CODING WITH MASKINGANALYSING JPEG CODING WITH MASKING
ANALYSING JPEG CODING WITH MASKING
 
A0540106
A0540106A0540106
A0540106
 
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...
 
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...
Corrosion Detection Using A.I : A Comparison of Standard Computer Vision Tech...
 
Efficient video compression using EZWT
Efficient video compression using EZWTEfficient video compression using EZWT
Efficient video compression using EZWT
 

Similar to A04840107

Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Ijripublishers Ijri
 
Paper id 2120148
Paper id 2120148Paper id 2120148
Paper id 2120148IJRAT
 
absorption, Cu2+ : glass, emission, excitation, XRD
absorption, Cu2+ : glass, emission, excitation, XRDabsorption, Cu2+ : glass, emission, excitation, XRD
absorption, Cu2+ : glass, emission, excitation, XRDIJERA Editor
 
Improved Error Detection and Data Recovery Architecture for Motion Estimation...
Improved Error Detection and Data Recovery Architecture for Motion Estimation...Improved Error Detection and Data Recovery Architecture for Motion Estimation...
Improved Error Detection and Data Recovery Architecture for Motion Estimation...IJERA Editor
 
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVC
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVCEfficient Architecture for Variable Block Size Motion Estimation in H.264/AVC
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVCIDES Editor
 
Design and Analysis of Quantization Based Low Bit Rate Encoding System
Design and Analysis of Quantization Based Low Bit Rate Encoding SystemDesign and Analysis of Quantization Based Low Bit Rate Encoding System
Design and Analysis of Quantization Based Low Bit Rate Encoding Systemijtsrd
 
Video Compression Using Block By Block Basis Salience Detection
Video Compression Using Block By Block Basis Salience DetectionVideo Compression Using Block By Block Basis Salience Detection
Video Compression Using Block By Block Basis Salience DetectionIRJET Journal
 
1 state of-the-art and trends in scalable video
1 state of-the-art and trends in scalable video1 state of-the-art and trends in scalable video
1 state of-the-art and trends in scalable videoYogananda Patnaik
 
Scanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecScanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecMuthu Samy
 
Scanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecScanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecMuthu Samy
 
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCSVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCIRJET Journal
 
Effective Compression of Digital Video
Effective Compression of Digital VideoEffective Compression of Digital Video
Effective Compression of Digital VideoIRJET Journal
 
Complexity Analysis in Scalable Video Coding
Complexity Analysis in Scalable Video CodingComplexity Analysis in Scalable Video Coding
Complexity Analysis in Scalable Video CodingWaqas Tariq
 
Low complexity video coding for sensor network
Low complexity video coding for sensor networkLow complexity video coding for sensor network
Low complexity video coding for sensor networkeSAT Publishing House
 
Low complexity video coding for sensor network
Low complexity video coding for sensor networkLow complexity video coding for sensor network
Low complexity video coding for sensor networkeSAT Journals
 
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...sipij
 

Similar to A04840107 (20)

Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...
 
C0161018
C0161018C0161018
C0161018
 
C0161018
C0161018C0161018
C0161018
 
Paper id 2120148
Paper id 2120148Paper id 2120148
Paper id 2120148
 
absorption, Cu2+ : glass, emission, excitation, XRD
absorption, Cu2+ : glass, emission, excitation, XRDabsorption, Cu2+ : glass, emission, excitation, XRD
absorption, Cu2+ : glass, emission, excitation, XRD
 
Improved Error Detection and Data Recovery Architecture for Motion Estimation...
Improved Error Detection and Data Recovery Architecture for Motion Estimation...Improved Error Detection and Data Recovery Architecture for Motion Estimation...
Improved Error Detection and Data Recovery Architecture for Motion Estimation...
 
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVC
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVCEfficient Architecture for Variable Block Size Motion Estimation in H.264/AVC
Efficient Architecture for Variable Block Size Motion Estimation in H.264/AVC
 
Design and Analysis of Quantization Based Low Bit Rate Encoding System
Design and Analysis of Quantization Based Low Bit Rate Encoding SystemDesign and Analysis of Quantization Based Low Bit Rate Encoding System
Design and Analysis of Quantization Based Low Bit Rate Encoding System
 
Video Compression Using Block By Block Basis Salience Detection
Video Compression Using Block By Block Basis Salience DetectionVideo Compression Using Block By Block Basis Salience Detection
Video Compression Using Block By Block Basis Salience Detection
 
1 state of-the-art and trends in scalable video
1 state of-the-art and trends in scalable video1 state of-the-art and trends in scalable video
1 state of-the-art and trends in scalable video
 
Scanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecScanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codec
 
Scanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codecScanned document compression using block based hybrid video codec
Scanned document compression using block based hybrid video codec
 
40120140504006
4012014050400640120140504006
40120140504006
 
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVCSVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
SVM Based Saliency Map Technique for Reducing Time Complexity in HEVC
 
Effective Compression of Digital Video
Effective Compression of Digital VideoEffective Compression of Digital Video
Effective Compression of Digital Video
 
C04841417
C04841417C04841417
C04841417
 
Complexity Analysis in Scalable Video Coding
Complexity Analysis in Scalable Video CodingComplexity Analysis in Scalable Video Coding
Complexity Analysis in Scalable Video Coding
 
Low complexity video coding for sensor network
Low complexity video coding for sensor networkLow complexity video coding for sensor network
Low complexity video coding for sensor network
 
Low complexity video coding for sensor network
Low complexity video coding for sensor networkLow complexity video coding for sensor network
Low complexity video coding for sensor network
 
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...
VIDEO QUALITY ASSESSMENT USING LAPLACIAN MODELING OF MOTION VECTOR DISTRIBUTI...
 

More from IOSR-JEN

More from IOSR-JEN (20)

C05921721
C05921721C05921721
C05921721
 
B05921016
B05921016B05921016
B05921016
 
A05920109
A05920109A05920109
A05920109
 
J05915457
J05915457J05915457
J05915457
 
I05914153
I05914153I05914153
I05914153
 
H05913540
H05913540H05913540
H05913540
 
G05913234
G05913234G05913234
G05913234
 
F05912731
F05912731F05912731
F05912731
 
E05912226
E05912226E05912226
E05912226
 
D05911621
D05911621D05911621
D05911621
 
C05911315
C05911315C05911315
C05911315
 
B05910712
B05910712B05910712
B05910712
 
A05910106
A05910106A05910106
A05910106
 
B05840510
B05840510B05840510
B05840510
 
I05844759
I05844759I05844759
I05844759
 
H05844346
H05844346H05844346
H05844346
 
G05843942
G05843942G05843942
G05843942
 
F05843238
F05843238F05843238
F05843238
 
E05842831
E05842831E05842831
E05842831
 
D05842227
D05842227D05842227
D05842227
 

Recently uploaded

Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 

Recently uploaded (20)

Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping Elbows
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 

A04840107

  • 1. IOSR Journal of Engineering (IOSRJEN) www.iosrjen.org ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 08 (August. 2014), ||V4|| PP 01-07 International organization of Scientific Research 1 | P a g e Video Quality Analysis for H.264 Based on Human Visual System Subrahmanyam.Ch1, Dr.D.Venkata Rao2, Dr.N.Usha Rani3 1(Research Scholar, Dept of ECE, Vignan University, Guntur, India) 2(Principal, Narasaraopeta Institute of Technology,Guntur, India) 3(Head of the Department, Dept of ECE, Vignan University, Guntur, India) Abstract: - The quality of video, such as the Mean Squared Error (MSE), Weighted Mean Squared Error (WMSE) or the Peak Signal-to-Noise Ratio (PSNR), not always correspond with visual observations. Structural Similarity based video quality assessment is proposed under the assumption that the Human Visual System (HVS) is highly adapted for extracting structural information from video. While the demand on high resolution quality increases in the media industry, Color loss will make the visual quality different. In this paper, we propose an improved quality assessment method by adding color comparison into structural similarity measurement system for evaluating video quality. Then we separate the task of similarity measurement into four comparisons: luminance, contrast, structure and color. Experimental results show that the predicted quality scores of the proposed method are more effective and consistent with visual quality than the classical methods using two different distortion types of video sets. Keywords: - MSE, WMSE, PSNR, HVS, luminance, contrast I. INTRODUCTION Visual information is very important for hearing-impaired people, because it allows them to communicate personally using the sign language. In our research, we focused on the fact that some parts of the person using the sign language are more important than others. We presented a visual attention model based on detection of low-level features such colour, intensity and texture and combination with the prior knowledge – in our case information about skin in image. Information about the visually relevant parts allows us to design an objective metric for this specific case. We presented an example of an objective metric based on human visual attention and detection of salient object in the observed scene. There is a huge use of semantic information in video. The ability to detect features is an interesting challenge by itself, but it takes on added importance to the extent it can serve as a reusable, extensible basis for query formation and search. Nowadays, the researchers focus mainly on solving the problems of finding the semantic information in video sequences. II. INTRODUCTION TO H.264 H.264/AVC is the state-of-the-art video coding standard developed by Joint Video Team (which comprises of ITU-T VCEG and ISO/IEC MPEG). Primary aim of H.264/AVC project was to develop a standard which provides same video quality at about half the bit rates of previous standards like MPEG-2, H.263 etc. Additional goals were lowering the complexity and increasing the flexibility so that the standard can be applied to wide variety of applications over different networks and systems. Its first version was released in May 2003 under the name H.264/AVC by ITU-T. Since then many versions has been released which has improved the standard in a variety of ways. These new versions are known as Fidelity range extensions (FRExt). The draft for FRExt was completed in September 2004. Scalable video coding (SVC) is another aspect of H.264/AVC project which has gained a lot of popularity. The aim of SVC is to use a single bit-stream which could serve different networks and different devices. SVC extension was completed in November 2007. In sync with previous standards, H.264/AVC defines only the syntax of encoded bit-stream and the method of decoding it. But this limitation of standard provides maximum freedom to optimize implementations in a manner appropriate to specific applications. Fig 2.1 shows the scope of H.264/AVC standard.
  • 2. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 2 | P a g e Fig 2.1: Scope of video coding standardization Since the first video coding standard (H.261), the functional blocks of the basic video coding algorithm remains the same. But the major difference lies in the details of each functional block. As a result each new video coding standard became more comprehensive and complex. Hence the video coding algorithm of H.264/AVC is also similar to previous standards but with an exception of de-blocking filter. Now let’s have a look at the codec (enCOder and DECoder pair) of H.264/AVC; starting with the encoder. 2.1.1 Encoder: A video encoder is a device or software which performs compression over the bit-stream; as a result the size of the encoded bit stream greatly reduces in size. Before elaborating more about video encoder, some basic terms of video processing are introduced which will help a lot in explaining various functional blocks of the encoder afterwards.  Video Stream: A video stream consists of large number of still images, when these images are run at a rate of 25/30 frames per second a video is formed. These still images can be classified as frames or fields depending upon the procedure of scanning.  Macroblock (MB): Macroblock is special word in video coding used to describe a small region in a frame or field (usually the size of a MB is 16x16 pixels). Therefore, each frame/field is divided into large number of macroblocks.  Pixel (PICture ELement): is the smallest piece of information in a frame/ image. Fig 2.2 Basic encoding structure of H.264/AVC for a Macroblock
  • 3. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 3 | P a g e Any video coding technique tries to reduce the size of the encoded bit stream by removing the redundancy present in the video sequence. Any video sequence has three types of redundancies temporal, spatial and statistical. To reduce temporal redundancy, similarities between the neighboring video frames are exploited by constructing a prediction of current frame. Normally a prediction is formed using one or two previous or future frames and is improved by compensating differences between different frames. The output of this whole process is a residual frame. These residues are processed by exploiting similarities between the neighboring pixels to remove spatial redundancy. Reduction is achieved by applying a transform (like DCT) and quantizing their results. The transform converts the sample into another domain where they are represented as transform coefficients; these coefficients are quantized to remove insignificant values, thus leaving a small number of coefficient values which provides more compact representation of residual frame. To remove statistical redundancy, an encoder uses various techniques (Huffman coding, Arithmetic coding etc) which assign short binary codes to commonly occurring vectors. Due to the reduction of redundancies the size of encoded bit- stream greatly reduces in size. Therefore, any video encoder consists of these three functional modules. Fig2.2 shows the H.264/AVC encoder, as is evident from the figure that H.264/AVC Encoder has a decoder loop incorporated in it. The main reason for including the decoder is to provide reference pictures which are very necessary for the prediction process. Therefore the encoder has two data flow paths, forward path and backward or reconstruction path.  Forward path: Here the normal encoding process is executed. An input frame is processed in terms of MB’s (16x16). Depending upon the type of MB, prediction technique is selected. There are two types of prediction techniques, intra prediction and inter prediction. In intra prediction, prediction is calculated from the samples of same frame which are previously encoded, decoded and reconstructed, whereas in inter prediction, prediction of current frame is calculated from the Motion compensation prediction of reference pictures which could be either previous or future frames. So when an ‘I’ MB is processed intra prediction is selected and when a ‘P/B’ MB is processed inter prediction is selected. Inter prediction techniques relies upon the motion estimation and motion compensation techniques for accurate prediction. These techniques are very instrumental in reducing the temporal redundancy between the frames. The goal of motion estimation technique is to find a best match for a block of current frame in reference frame. To find the best match, a search is performed in the reference frame. This search is carried out by comparing the block in current frame with the blocks in search area (usually centered on and around the current block position) of reference frame. The block or area which gives the lowest residual energy is considered as the best match region. Thus chosen best match region is called as ‘predictor’ for the block in current frame. In motion compensation, predictor is subtracted from the current block resulting in residues. So obtained residues are block transformed, scaled and quantized to obtain quantized coefficients. These quantized coefficients along with control data and motion vector information is entropy coded. Entropy encoded coefficients along with side information necessary for decoding each macro block at the decoder constitutes the encoded bit stream.  Reconstruction path: Here decoding and reconstruction of encoded frames is performed. To reconstruct the encoded frames, quantized coefficients are scaled (De-quantized) and inverse transformed to obtain residues which are added with the prediction to obtain unfiltered frame. This unfiltered frame contains blocking artifacts; blocking artifacts arise mainly due to the usage of block based integer DCT (Discrete Cosine Transform). To smooth out these blocking artifacts, de-blocking filter is applied over the frame. These filtered frames are stored as reference frames. 2.1.2 Decoder: When the decoder receives the encoded bit stream, it entropy decodes them to produce a set of quantized coefficients which are reordered, scaled and inverse block transformed to obtain residues. Using the header information received in the encoded bit-stream, decoder generates the prediction. Residues are added to the prediction to obtain the unfiltered version of the original frame. Unfiltered frame contains blocking artifacts which deteriorates the video quality. De-blocking filter is applied over the unfiltered frame to remove the blocking artifacts and enhance the video quality of filtered frame. Filtered frame are stored as reference frames which are utilized in prediction process.
  • 4. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 4 | P a g e Entropy decoding Scaling Reorder Inverse Transform + De-blocking filter Motion Compensation Intra pediction Reference frames Reconstructed frames + + Intra Inter Prediction Residues NAL Unfiltered frame Fig 2.3: H.264 Decoder III. PROPOSED REFERNEC MODEL FOR VIDEO QUALITY MONITORING Three methodologies representing different measurement strategies for the assessment of the quality of video have been defined: – methodology using the complete video reference (double-ended); – methodology using reduced reference information (double-ended); – methodology using no reference signal (single-ended). The design and the development of a video quality monitor should consider a general structure of the measurement procedure for reduced reference and single-ended methodologies. The reference model is composed of the following four layers: – Measurement methodology defines the class or the strategy relative to the application requirement; – Measurement method is composed of a set of modules, algorithmic and associated ones, implemented to process inputs such as original signals or processed reference data, and provide output results such as processed reference data, level of impairment or final quality notation; – Algorithmic module(s) is the basic block of signal processing functions composing the method. It composes the core of the method from which the final objective qualification is delivered; – Associated module(s) is an additional function that aids the algorithmic module(s) in its operation by addressing such issues as dating, synchronization, presentation of data, etc. Each one of those methodologies offers a specific approach relative to its operational use. Each methodology is well adapted to a specific field of applications. This is due to its technical implementation constraints: – relevance of the measured data, – availability of the reference, – synchronization of original and impaired signals or data, – transmission channel for reference data, – real-time implementation, – in-service use, etc. Post-processing algorithms are the most popular methods for improving the quality of the image and video and eliminate the annoying blocking artifact. On the other hand, the post-processing algorithms can achieve deblocking without the original image and video, so the standard need not to be modified. Reduction of Blocking Artifacts in DCT Domain We introduce a post-processing filtering algorithm in DCT domain. We define the block , , ( , ) k l m n b u v and , , ( , ) k l m n B u v first. , ( , ) m n b u v is the (m,n)-th 8х8 block in the compressed image, and , ( , ) m n B u v is the DCT coefficients of , ( , ) m n b u v . , , ( , ) k l m n b u v is the shifted block with displacement k pixel in the x direction and displacement l pixels in the y direction with respective to block
  • 5. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 5 | P a g e , ( , ) m n b u v , and , , ( , ) k l m n B u v is the DCT coefficients of the block , , ( , ) k l m n b u v . One example is shown in Fig. 3.1. m 1,n b  m,n b m 1,n 1 b   m,n 1 b  1, 1 m,n b  Fig. 3.1 Example of shifted block , , ( , ) k l m n b u v In the original image, the neighboring DCT coefficients at the same frequency are very similar and do not vary radically within a small range. Thus, we can apply low pass filter to the DCT coefficients at each frequency to filter the high frequency parts resulting from blocking effect. However, this method may blur the real edges in the original image, so we must have the mechanism to detect activity of the block and apply the filter with corresponding strength. DCT-domain filtering is applied to revise the block , ( , ) m n B u v to obtain the new DCT coefficients  Bm,n (i, j) .  , , , , 1 ( , ) ( , ) h h k l m n k l m n k h l h B i j w B u v W      , h h k l k h l h W w      The post-filtering works in different ways for the blocks with different activities. For blocks with low activity, the blocking artifact is more noticeable, so we apply strong filtering to smooth the high frequency components. The filter coefficients are defined. , 1, , 2,..., 2 k l w  k l   For blocks with high activity, the blocking artifact is less noticeable, so we apply filtering with less strength to smooth blocking artifact and preserve the real edge. The filter coefficients are defined. , 3, for ( , ) (0,0) 1, k l k l w otherwise    IV. EXPERIMENTAL RESULTS In this, we will show some experimental results. The video used in the experiment is “xylophone.mpg”, which consists of original high-resolution YUV format video files. It includes upto 10 frames of re sized video. The details of the video frame values are shown. Step wise Process: Encoding I-Frame: 1 Encoding P Frame: 2 Encoding P Frame: 3 Encoding P Frame: 4 Encoding P Frame: 5 Encoding P Frame: 6 Encoding P Frame: 7 Encoding P Frame: 8 Encoding P Frame: 9 Encoding P Frame: 10
  • 6. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 6 | P a g e Fig.4.1. Encoding Video process of xylophone.mpg video (10 frames) Decoding I-Frame: 1 Decoding P Frame: 2 Decoding P Frame: 3 Decoding P Frame: 4 Decoding P Frame: 5 Decoding P Frame: 6 Decoding P Frame: 7 Decoding P Frame: 8 Decoding P Frame: 9 Decoding P Frame: 10 Fig.4.2. Decoding Video process of xylophone.mpg video (10 frames) Fig.4.3. PSNR = 18.8015 Fig.4.4. PSNR = 1.2104 V. CONCLUSION Emerging new technologies demand tools for efficient indexing, browsing and retrieval of video data, which causes rapid expansion of areas of research where the semantic information is used. New methods that work with semantic information in image, video, and audio are developed frequently these days, which means that our list of methods is not final. Nevertheless, we picked the most used ones and tested them. We brought an overview of the methods for detecting and describing interesting points in video, modeling visual attention and saliency. We have tested the H.264 video formats. We have evaluated the use of visual saliency for compression of video sequences.
  • 7. Video Quality Analysis for H.264 Based on Human Visual System International organization of Scientific Research 7 | P a g e REFERENCES [1] A. C Bovik. Introduction to Image and Video Quality Assessment[OL]. [2009-3-17]. http://live.ece. utexas.edu /research/quality/intro.htm. [2] C J van den Branden Lambrecht, Ed. Special Issue on Image and Video Quality Metrics [J]. Signal Process., 1998, 70(3): 155-294. [3] VQEG. Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment [OL]. (2000-3-15) [2009-3-17]. http://www.its.bldrdoc.gov/vqeg/projects/ frtv_phaseI/COM-80E_final_report.pdf. [4] T N Pappas and R J Safranek. Perceptual Criteria for Image Quality Evaluation. In: A. C Bovik, ed. Handbook of Image &Video Processing [M]. San Diego: Academic Press, 2000: 669-684. [5] Z Wang and A C Bovik. A Universal Image Quality Index [J]. IEEE Signal Processing Letters, 2002, 9(3): 81-84. [6] G H Chen, C L Yang, L M Po, et al. Edge-based Structural Similarity for Image Quality Assessment [C]. ICASSP, Toulouse, France, 2006: II-933-II936. [7] Commission Internationale d'Eclairage. CIE Publication No.15.2, Colorimetry [S]. 2nd Ed. Vienna: CIE Central Bureau, 1986. [8] Commission Internationale d'Eclairage. CIE Publication No. TC1. 47, Hue and Lightness Correction to Industrial Colour Difference Evaluation [S]. Vienna: CIE Central Bureau, 2000. [9] VQEG. Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment, Phase Ⅱ VQEG [OL]. (2003-8-25) [2009-3-17]. http://www.its.bldrdoc.gov/ /projects/frtv- _phaseII/downloads/VQEGII_Final_Report.pdf.