This document discusses automatic view synthesis from stereoscopic 3D video through image domain warping. It begins with an introduction to stereoscopic 3D cinema and television, and the need for multi-view auto stereoscopic displays that allow glasses-free 3D viewing. It then describes image domain warping, which synthesizes new views from 2-view video using sparse disparity features and warping images to enforce the disparities, rather than using depth maps. The document outlines the image warping process and view synthesis algorithm, which extracts sparse disparity features, calculates warps to enforce the disparities for intermediate views, and warps the images to synthesize the output views.
This document describes a proposed system for implementing 3D video calling using holographic views. The system would use a prism placed on the receiving user's smartphone screen to generate a 3D holographic view of the calling user. It works by taking the 2D video frames from the call, processing them using color masking and perspective filtering techniques as part of a 2D to 3D conversion system, and then displaying modified frames on each side of the prism screen to create the illusion of a 3D hologram when the light rays refract through the prism. The system aims to provide a more realistic communication experience compared to traditional 2D video calls. Future work may include modifying the system to allow two-
This document discusses tele-immersion technology, which allows users in different locations to interact in a simulated holographic environment. It provides a history of the concept dating back to 1965, and describes how tele-immersion works using camera clusters to capture 3D environments, reconstruct 3D models, compress the data, transmit it over networks, and allow remote users to interact in the virtual space in real-time. Basic requirements like high-performance displays, computers, tracking sensors and networks are needed to support the technology. Potential applications include remote collaboration and future developments may enable touch interactions through haptic sensors.
IRJET- Photo Restoration using Multiresolution Texture Synthesis and Convolut...IRJET Journal
The document discusses techniques for removing haze and fog from images. It presents a technique called IDeRS that uses an iterative dehazing model to remove haze and fog from remote sensing images. The IDeRS technique estimates atmospheric light independently of haze-opaque regions using a haze-line prior method. It then computes a transmission map using the dark channel prior model to estimate a raw transmission map. The technique achieves high signal-to-noise ratios and improves on other methods that did not completely remove haze and suffered from artifacts.
Tele-immersion is a technology that allows users in different locations to collaborate in real-time as if they are in the same room. It uses cameras, displays, and tracking systems to reconstruct 3D scenes and environments. The idea was proposed in 1997 and researchers have since developed systems using multiple cameras, 3D modeling, and compression algorithms to transmit immersive virtual experiences over networks. However, challenges remain around bandwidth requirements, data transmission quality, and cost.
Tele immersion is a new arrising technology enables users to communicate with each other which are located remotely as they are together in same room or place...
This ppt includes all the basic aspect of tele immersion...
This document discusses image processing and provides examples of its applications. It covers the following key points:
1) Image processing involves analyzing and transforming images and can be used to extract information. The Mars Exploration Rover mission used image processing to compress and send images back to operators on Earth.
2) There are three main types of image processing: image-to-image, image-to-information, and information-to-image. Color spaces and compression techniques like Huffman coding are also discussed.
3) Huffman coding assigns variable length codes to characters based on their frequency, allowing for more common characters to be encoded with fewer bits and improving compression without loss of information. It has numerous applications including in
Hand gesture recognition using support vector machinetheijes
1) The document describes a system for hand gesture recognition using support vector machines. It uses Canny's edge detection algorithm and histogram of gradients (HOG) for feature extraction from input images of hand gestures.
2) The system is trained using a dataset of predefined hand gestures. During testing, it compares the features extracted from new input images to those in the training dataset and classifies the gesture using an SVM classifier.
3) Experimental results found the system could accurately recognize 20 different static hand gestures in complex backgrounds. However, the authors note that future work could focus on real-time gesture recognition and reducing complexity for faster processing.
This document discusses digital image processing and image compression. It covers 5 units: digital image fundamentals, image transforms, image enhancement, image filtering and restoration, and image compression. Image compression aims to reduce the size of image data and is important for applications like facsimile transmission and CD-ROM storage. There are two types of compression - lossless, where the original and reconstructed data are identical, and lossy, which allows some loss for higher compression ratios. Factors to consider for compression method selection include whether lossless or lossy is needed, coding efficiency, complexity tradeoffs, and the application.
This document describes a proposed system for implementing 3D video calling using holographic views. The system would use a prism placed on the receiving user's smartphone screen to generate a 3D holographic view of the calling user. It works by taking the 2D video frames from the call, processing them using color masking and perspective filtering techniques as part of a 2D to 3D conversion system, and then displaying modified frames on each side of the prism screen to create the illusion of a 3D hologram when the light rays refract through the prism. The system aims to provide a more realistic communication experience compared to traditional 2D video calls. Future work may include modifying the system to allow two-
This document discusses tele-immersion technology, which allows users in different locations to interact in a simulated holographic environment. It provides a history of the concept dating back to 1965, and describes how tele-immersion works using camera clusters to capture 3D environments, reconstruct 3D models, compress the data, transmit it over networks, and allow remote users to interact in the virtual space in real-time. Basic requirements like high-performance displays, computers, tracking sensors and networks are needed to support the technology. Potential applications include remote collaboration and future developments may enable touch interactions through haptic sensors.
IRJET- Photo Restoration using Multiresolution Texture Synthesis and Convolut...IRJET Journal
The document discusses techniques for removing haze and fog from images. It presents a technique called IDeRS that uses an iterative dehazing model to remove haze and fog from remote sensing images. The IDeRS technique estimates atmospheric light independently of haze-opaque regions using a haze-line prior method. It then computes a transmission map using the dark channel prior model to estimate a raw transmission map. The technique achieves high signal-to-noise ratios and improves on other methods that did not completely remove haze and suffered from artifacts.
Tele-immersion is a technology that allows users in different locations to collaborate in real-time as if they are in the same room. It uses cameras, displays, and tracking systems to reconstruct 3D scenes and environments. The idea was proposed in 1997 and researchers have since developed systems using multiple cameras, 3D modeling, and compression algorithms to transmit immersive virtual experiences over networks. However, challenges remain around bandwidth requirements, data transmission quality, and cost.
Tele immersion is a new arrising technology enables users to communicate with each other which are located remotely as they are together in same room or place...
This ppt includes all the basic aspect of tele immersion...
This document discusses image processing and provides examples of its applications. It covers the following key points:
1) Image processing involves analyzing and transforming images and can be used to extract information. The Mars Exploration Rover mission used image processing to compress and send images back to operators on Earth.
2) There are three main types of image processing: image-to-image, image-to-information, and information-to-image. Color spaces and compression techniques like Huffman coding are also discussed.
3) Huffman coding assigns variable length codes to characters based on their frequency, allowing for more common characters to be encoded with fewer bits and improving compression without loss of information. It has numerous applications including in
Hand gesture recognition using support vector machinetheijes
1) The document describes a system for hand gesture recognition using support vector machines. It uses Canny's edge detection algorithm and histogram of gradients (HOG) for feature extraction from input images of hand gestures.
2) The system is trained using a dataset of predefined hand gestures. During testing, it compares the features extracted from new input images to those in the training dataset and classifies the gesture using an SVM classifier.
3) Experimental results found the system could accurately recognize 20 different static hand gestures in complex backgrounds. However, the authors note that future work could focus on real-time gesture recognition and reducing complexity for faster processing.
This document discusses digital image processing and image compression. It covers 5 units: digital image fundamentals, image transforms, image enhancement, image filtering and restoration, and image compression. Image compression aims to reduce the size of image data and is important for applications like facsimile transmission and CD-ROM storage. There are two types of compression - lossless, where the original and reconstructed data are identical, and lossy, which allows some loss for higher compression ratios. Factors to consider for compression method selection include whether lossless or lossy is needed, coding efficiency, complexity tradeoffs, and the application.
The document describes a system for 3D modeling using hand gestures as input. It uses a vision-based tracking system to recognize hand gestures without any instruments attached to the hands. The system supports basic modeling tasks like selection, translation, rotation, and scaling of 3D objects using just five static hand gestures. Visual feedback is provided to help users perceive interactions. The goal is to provide an intuitive interface for 3D modeling that requires little or no training.
This document discusses digital image fundamentals including:
- The structure and function of the human eye and vision system.
- How images are represented digitally as matrices of pixel values.
- Factors that determine the resolution of a digital image such as sampling rate, quantization, and number of bits per pixel.
- Basic relationships between pixels such as connectivity and labeling of connected components.
AN EMERGING TREND OF FEATURE EXTRACTION METHOD IN VIDEO PROCESSINGcscpconf
Recently the progress in technology and flourishing applications open up new forecast and defy
for the image and video processing community. Compared to still images, video sequences
afford more information about how objects and scenarios change over time. Quality of video is
very significant before applying it to any kind of processing techniques. This paper deals with
two major problems in video processing they are noise reduction and object segmentation on
video frames. The segmentation of objects is performed using foreground segmentation based
and fuzzy c-means clustering segmentation is compared with the proposed method Improvised
fuzzy c – means segmentation based on color. This was applied in the video frame to segment
various objects in the current frame. The proposed technique is a powerful method for image
segmentation and it works for both single and multiple feature data with spatial information.
The experimental result was conducted using various noises and filtering methods to show which is best suited among others and the proposed segmentation approach generates good quality segmented frames.
The document discusses the key components of an image processing system, including image sensing, digitization, storage, and display. It covers common image sensing devices like cameras, scanners, and MRI systems. It also describes digitizers, different types of digital storage, and principal display devices. Finally, it discusses concepts like spatial and gray-level resolution, sampling and quantization, and interpolation methods used for zooming and shrinking digital images.
Image processing involves manipulating digital images through algorithms implemented on computers. A digital image is composed of picture elements called pixels arranged in a grid. Each pixel represents a color or intensity value. Common image processing tasks include computer vision, optical character recognition, medical imaging, and more. Key concepts in image processing include pixels, resolution, color depth, and filtering/manipulating pixel values.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
nternational Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Digital 3D imaging can be accelerated using advances in VLSI technology. High-resolution 3D images can be captured using laser-based vision systems, which produce 3D information insensitive to background illumination and surface texture. Complete images of featureless surfaces invisible to the human eye can be generated. Sensors for 3D digitization include position sensitive detectors and laser sensors. Continuous response position sensitive detectors provide precise centroid measurement while discrete response detectors are slower but more accurate. An integrated sensor architecture is proposed using a combination of these sensors to simultaneously measure color and 3D.
1) The document presents a real-time static hand gesture recognition system for the Devanagari number system using two feature extraction techniques: Discrete Cosine Transform (DCT) and Edge Oriented Histogram (EOH).
2) The system captures an image using a webcam, performs pre-processing, extracts the region of interest, then extracts features using DCT or EOH before matching against a training database to recognize the gesture.
3) An experiment tested 20 images and found DCT achieved a higher recognition accuracy of 18 gestures compared to 15 for EOH.
Sensors on 3 d digitization seminar reportVishnu Prasad
The document discusses sensors for 3D digitization. It describes two main strategies for 3D vision - passive vision which analyzes ambient light, and active vision which structures light using techniques like laser range cameras. It then discusses an auto-synchronized scanner that can provide registered 3D surface maps and color data by scanning a laser spot across a scene and detecting the reflected light with a linear sensor, producing registered images with spatial and color information.
1. Digital image processing focuses on improving images for human interpretation and machine perception. It involves digitizing an image using sensors and processors then displaying the digital image.
2. The key stages of digital image processing are enhancement, restoration, compression, and registration. Registration involves mapping image frames for tasks like object recognition.
3. Common processing techniques include contrast intensification to improve poor contrast, smoothing to reduce noise, and sharpening to enhance blurred details.
Image processing techniques can be used for face recognition applications. The process involves decomposing face images into subbands using discrete wavelet transform. The mid-frequency subband is selected and principal component analysis is applied to extract representational bases. These bases are stored for training images and used to translate probe images into representations which are classified to identify faces by matching with training representations. This approach segments discriminatory facial features to recognize identities despite variations in illumination, pose, expression and other factors.
The document summarizes digital image processing. It discusses the origins of digital images in the 1920s for transmitting newspaper photographs via transatlantic cable. The field grew with early computer processing of images from space missions in the 1960s. The document outlines fundamental steps in digital image processing, including image acquisition, enhancement, restoration, color processing, wavelets, compression, morphology, segmentation, and representation. It provides an abstract for a seminar report on digital image processing.
Device for text to speech production and to braille scriptIAEME Publication
The document describes a proposed system to convert text to both speech and Braille script for blind or deaf individuals. The system would take an image of text as input, perform image processing techniques like enhancement, filtering, and edge detection, then segment and recognize characters. The recognized text would be converted to speech output using text-to-speech synthesis or to Braille script by mapping characters to Braille codes and outputting to a tactile display. The goal is to make learning materials more accessible for blind or deaf individuals by converting textbook images to audio or Braille formats.
Retrieving Of Color Images Using SDS TechniqueEditor IJMTER
How data can be shared from one part of the world to the other in near real time came with
the arrival of internet. Along with this they have introduced new challenges like maintaining the
confidentiality of transmitting the data. This gave a boost to the research area related to cryptography.
Firstly, Encryption of images with the accepted encryption algorithms had significant downside as key
management was complicated and limited. Secondly, introduction to new area for encrypting images
was splitting the image at its pixel level in to multiple shares. But the major drawback of this approach
was that the recovered image had a poor quality. To overcome these mentioned drawbacks we have
proposed a new approach which does not attempt to use any type of keys for encryption.
teleimmersion is integration of haptics communication and image processing technique with virtual reality to introduce the sense of touch in virtual images .this is infact going to be the communication with exchange of emotion in real time
An Introduction to Image Processing and Artificial IntelligenceWasif Altaf
This document provides an introduction to image processing and artificial intelligence. It defines what an image is from different perspectives including in literature, general terms, and in computer science as an exact replica of a storage device. It describes image processing as analyzing and manipulating images with three main steps: importing an image, manipulating or analyzing it, and outputting the result. It also discusses what noise is in images, methods to remove noise, color enhancement techniques, sharpening images to increase contrast, and segmentation and edge detection.
This document summarizes a framework for precise object tracking under image deformation. It discusses various types of image deformation including noise, scaling and rotation, blurring, and illumination changes. It proposes techniques to address each deformation type, including denoising filters, linear and radon transformations for rotation, deblurring algorithms, and using color models like YCbCr that are robust to illumination changes. It also describes morphological operations, geometrical modeling for 3D pose estimation, and using an FIR model for motion prediction to track objects under deformation. The framework is aimed at applications like robot vision, surveillance, medical imaging and more.
This document discusses a structural similarity based approach for efficient multi-view video coding. It begins with an introduction to multi-view video coding and the structural similarity index metric. It then proposes using structural similarity to exploit structural information between different video views. The method uses structural similarity for rate distortion optimization in encoding. Experimental results show the left and right views of a video, their structural similarity image, the decoded 3D video, and the achieved minimum distortion level. The document aims to improve multi-view video quality by using structural similarity during the encoding process.
This document summarizes recent research on regenerated silk fibroin fibers produced through wet-spinning and electrospinning techniques. It discusses how silk fibroin is obtained from silkworm cocoons and its composition. The degumming and dissolution processes to prepare silk fibroin solutions for fiber spinning are described. Wet-spinning involves extruding a silk fibroin dope solution through a spinneret into a non-solvent bath, while electrospinning uses electric fields to spin nano-to-micrometer diameter fibers from silk fibroin solutions. The properties of regenerated silk fibers can be tailored for applications as tissue engineering scaffolds.
This document describes an audio steganography technique that aims to increase security by introducing randomness. It discusses how traditional least significant bit (LSB) modification is vulnerable to attacks. The proposed technique randomly selects both the bit position (1st, 2nd, or 3rd LSB) and audio sample for embedding secret message bits. This is intended to prevent attackers from detecting the embedding pattern. The technique uses character encoding like Huffman coding before message bits are hidden in an audio file using the modified LSB method. Experimental results showed the stego audio maintained quality while providing improved security over fixed LSB techniques.
The document describes a system for 3D modeling using hand gestures as input. It uses a vision-based tracking system to recognize hand gestures without any instruments attached to the hands. The system supports basic modeling tasks like selection, translation, rotation, and scaling of 3D objects using just five static hand gestures. Visual feedback is provided to help users perceive interactions. The goal is to provide an intuitive interface for 3D modeling that requires little or no training.
This document discusses digital image fundamentals including:
- The structure and function of the human eye and vision system.
- How images are represented digitally as matrices of pixel values.
- Factors that determine the resolution of a digital image such as sampling rate, quantization, and number of bits per pixel.
- Basic relationships between pixels such as connectivity and labeling of connected components.
AN EMERGING TREND OF FEATURE EXTRACTION METHOD IN VIDEO PROCESSINGcscpconf
Recently the progress in technology and flourishing applications open up new forecast and defy
for the image and video processing community. Compared to still images, video sequences
afford more information about how objects and scenarios change over time. Quality of video is
very significant before applying it to any kind of processing techniques. This paper deals with
two major problems in video processing they are noise reduction and object segmentation on
video frames. The segmentation of objects is performed using foreground segmentation based
and fuzzy c-means clustering segmentation is compared with the proposed method Improvised
fuzzy c – means segmentation based on color. This was applied in the video frame to segment
various objects in the current frame. The proposed technique is a powerful method for image
segmentation and it works for both single and multiple feature data with spatial information.
The experimental result was conducted using various noises and filtering methods to show which is best suited among others and the proposed segmentation approach generates good quality segmented frames.
The document discusses the key components of an image processing system, including image sensing, digitization, storage, and display. It covers common image sensing devices like cameras, scanners, and MRI systems. It also describes digitizers, different types of digital storage, and principal display devices. Finally, it discusses concepts like spatial and gray-level resolution, sampling and quantization, and interpolation methods used for zooming and shrinking digital images.
Image processing involves manipulating digital images through algorithms implemented on computers. A digital image is composed of picture elements called pixels arranged in a grid. Each pixel represents a color or intensity value. Common image processing tasks include computer vision, optical character recognition, medical imaging, and more. Key concepts in image processing include pixels, resolution, color depth, and filtering/manipulating pixel values.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
nternational Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Digital 3D imaging can be accelerated using advances in VLSI technology. High-resolution 3D images can be captured using laser-based vision systems, which produce 3D information insensitive to background illumination and surface texture. Complete images of featureless surfaces invisible to the human eye can be generated. Sensors for 3D digitization include position sensitive detectors and laser sensors. Continuous response position sensitive detectors provide precise centroid measurement while discrete response detectors are slower but more accurate. An integrated sensor architecture is proposed using a combination of these sensors to simultaneously measure color and 3D.
1) The document presents a real-time static hand gesture recognition system for the Devanagari number system using two feature extraction techniques: Discrete Cosine Transform (DCT) and Edge Oriented Histogram (EOH).
2) The system captures an image using a webcam, performs pre-processing, extracts the region of interest, then extracts features using DCT or EOH before matching against a training database to recognize the gesture.
3) An experiment tested 20 images and found DCT achieved a higher recognition accuracy of 18 gestures compared to 15 for EOH.
Sensors on 3 d digitization seminar reportVishnu Prasad
The document discusses sensors for 3D digitization. It describes two main strategies for 3D vision - passive vision which analyzes ambient light, and active vision which structures light using techniques like laser range cameras. It then discusses an auto-synchronized scanner that can provide registered 3D surface maps and color data by scanning a laser spot across a scene and detecting the reflected light with a linear sensor, producing registered images with spatial and color information.
1. Digital image processing focuses on improving images for human interpretation and machine perception. It involves digitizing an image using sensors and processors then displaying the digital image.
2. The key stages of digital image processing are enhancement, restoration, compression, and registration. Registration involves mapping image frames for tasks like object recognition.
3. Common processing techniques include contrast intensification to improve poor contrast, smoothing to reduce noise, and sharpening to enhance blurred details.
Image processing techniques can be used for face recognition applications. The process involves decomposing face images into subbands using discrete wavelet transform. The mid-frequency subband is selected and principal component analysis is applied to extract representational bases. These bases are stored for training images and used to translate probe images into representations which are classified to identify faces by matching with training representations. This approach segments discriminatory facial features to recognize identities despite variations in illumination, pose, expression and other factors.
The document summarizes digital image processing. It discusses the origins of digital images in the 1920s for transmitting newspaper photographs via transatlantic cable. The field grew with early computer processing of images from space missions in the 1960s. The document outlines fundamental steps in digital image processing, including image acquisition, enhancement, restoration, color processing, wavelets, compression, morphology, segmentation, and representation. It provides an abstract for a seminar report on digital image processing.
Device for text to speech production and to braille scriptIAEME Publication
The document describes a proposed system to convert text to both speech and Braille script for blind or deaf individuals. The system would take an image of text as input, perform image processing techniques like enhancement, filtering, and edge detection, then segment and recognize characters. The recognized text would be converted to speech output using text-to-speech synthesis or to Braille script by mapping characters to Braille codes and outputting to a tactile display. The goal is to make learning materials more accessible for blind or deaf individuals by converting textbook images to audio or Braille formats.
Retrieving Of Color Images Using SDS TechniqueEditor IJMTER
How data can be shared from one part of the world to the other in near real time came with
the arrival of internet. Along with this they have introduced new challenges like maintaining the
confidentiality of transmitting the data. This gave a boost to the research area related to cryptography.
Firstly, Encryption of images with the accepted encryption algorithms had significant downside as key
management was complicated and limited. Secondly, introduction to new area for encrypting images
was splitting the image at its pixel level in to multiple shares. But the major drawback of this approach
was that the recovered image had a poor quality. To overcome these mentioned drawbacks we have
proposed a new approach which does not attempt to use any type of keys for encryption.
teleimmersion is integration of haptics communication and image processing technique with virtual reality to introduce the sense of touch in virtual images .this is infact going to be the communication with exchange of emotion in real time
An Introduction to Image Processing and Artificial IntelligenceWasif Altaf
This document provides an introduction to image processing and artificial intelligence. It defines what an image is from different perspectives including in literature, general terms, and in computer science as an exact replica of a storage device. It describes image processing as analyzing and manipulating images with three main steps: importing an image, manipulating or analyzing it, and outputting the result. It also discusses what noise is in images, methods to remove noise, color enhancement techniques, sharpening images to increase contrast, and segmentation and edge detection.
This document summarizes a framework for precise object tracking under image deformation. It discusses various types of image deformation including noise, scaling and rotation, blurring, and illumination changes. It proposes techniques to address each deformation type, including denoising filters, linear and radon transformations for rotation, deblurring algorithms, and using color models like YCbCr that are robust to illumination changes. It also describes morphological operations, geometrical modeling for 3D pose estimation, and using an FIR model for motion prediction to track objects under deformation. The framework is aimed at applications like robot vision, surveillance, medical imaging and more.
This document discusses a structural similarity based approach for efficient multi-view video coding. It begins with an introduction to multi-view video coding and the structural similarity index metric. It then proposes using structural similarity to exploit structural information between different video views. The method uses structural similarity for rate distortion optimization in encoding. Experimental results show the left and right views of a video, their structural similarity image, the decoded 3D video, and the achieved minimum distortion level. The document aims to improve multi-view video quality by using structural similarity during the encoding process.
This document summarizes recent research on regenerated silk fibroin fibers produced through wet-spinning and electrospinning techniques. It discusses how silk fibroin is obtained from silkworm cocoons and its composition. The degumming and dissolution processes to prepare silk fibroin solutions for fiber spinning are described. Wet-spinning involves extruding a silk fibroin dope solution through a spinneret into a non-solvent bath, while electrospinning uses electric fields to spin nano-to-micrometer diameter fibers from silk fibroin solutions. The properties of regenerated silk fibers can be tailored for applications as tissue engineering scaffolds.
This document describes an audio steganography technique that aims to increase security by introducing randomness. It discusses how traditional least significant bit (LSB) modification is vulnerable to attacks. The proposed technique randomly selects both the bit position (1st, 2nd, or 3rd LSB) and audio sample for embedding secret message bits. This is intended to prevent attackers from detecting the embedding pattern. The technique uses character encoding like Huffman coding before message bits are hidden in an audio file using the modified LSB method. Experimental results showed the stego audio maintained quality while providing improved security over fixed LSB techniques.
This document discusses material selection for the structural design of a mini milling machine. It begins by analyzing the existing cast iron structure and then proposes a hybrid structure using both casting and fabricated parts. A methodology is presented involving material selection criteria, CAD modeling, and analysis. Various materials are considered for different components, including steel, cast iron, and polymer composites for the base. An analytic hierarchy process is used to rank materials based on properties like strength, damping, and cost. Steel, cast iron, and a polymer composite are compared for the base, with the composite ranking highest based on its properties and manufacturability.
This document summarizes statistical disclosure control techniques for protecting private data, specifically microaggregation. Microaggregation involves clustering individual records into small groups to anonymize the data before release. It aims to minimize information loss while preventing re-identification of individuals. The document discusses challenges with multivariate microaggregation and reviews different heuristic approaches. It also covers related topics like k-anonymity algorithms, various clustering techniques for microaggregation like k-means, and using genetic algorithms to handle large datasets.
The document discusses an algorithm called Adaptive Multichannel Component Analysis (AMMCA) for separating image sources from mixtures using adaptively learned dictionaries. It begins by reviewing image denoising using learned dictionaries, then extends this to image separation from single mixtures. The key contribution is applying this approach to separating sources from multichannel mixtures by learning local dictionaries for each source during the separation process. The algorithm is described and simulated results are shown separating two images from a noisy mixture using the learned dictionaries. In conclusion, AMMCA is able to separate sources without prior knowledge of their sparsity domains by fusing dictionary learning into the separation process.
This document describes a proposed advanced car automation and security system. The system would allow for multiple user profiles to be saved, with iris recognition used for authentication. When an authorized user's iris is recognized, the car settings would automatically adjust based on that user's saved profile. If an unauthorized person tries to access the car, a security message would be sent to the car owner. The system aims to provide convenience by automatically adjusting settings for different users, and enhance security by monitoring unauthorized access and notifying the owner. It would use an FPGA for automation control, MATLAB for authentication, and a GSM module for communication.
This document describes an algorithmic approach for detecting car accidents using smartphones. It proposes using sensors in smartphones like GPS, accelerometers, and microphones to detect accidents. The algorithm uses an 11-tuple model including factors like acceleration, sound, and speed to predict accidents. If acceleration and sound thresholds are exceeded while speed is above a minimum, or if movement is below a distance threshold after speed drops, an accident is detected. The algorithm aims to provide rapid emergency notification by detecting accidents and alerting emergency contacts.
This document summarizes the RedTacton technology, which enables data transfer between two devices through physical contact with the human body. RedTacton uses electric fields generated by the human body as a transmission medium. It allows for connectivity between various personal devices through natural physical interactions like handshakes. The technology works by placing sensors on the body that can detect minute electric fields used to transmit data in a point-to-point way. RedTacton provides a new way of connecting devices through a human-centered approach and establishes a type of network called a Human Area Network. Future applications could include uses in healthcare, security, and other areas where device-to-device communication through touch is useful.
The document analyzes the performance of M-ary modulations through the human body area channel. It simulates M-ary PAM and M-ary BOK modulation schemes at different carrier frequencies to obtain bit error rates. The simulations show that a carrier frequency of 2400MHz provides the best performance for both 16-PAM and 32-PAM modulation, with minimum bit error rates achieved using a selective rake receiver. Partial rake receivers performed poorer than selective rake receivers for all modulation schemes and frequencies tested.
This document describes a hybrid energy management system based on a fuzzy logic controller for power distribution. The system uses four power sources - wind power, photovoltaics, fuel cells, and electric power - connected to a common DC bus. An automatic energy management system provides load sharing between the power sources based on the load demand. Experimental results show that the system can successfully meet different load levels of 1000W, 2000W, and 3000W by distributing power from the sources according to the fuzzy logic controller and without wasted power.
This document summarizes research on defeating denial-of-service (DoS) attacks in wireless networks in the presence of jammers. It describes common types of jamming attacks like constant, deceptive, random, and reactive jammers. Detection techniques for jammers and methods to reduce the impact of DoS attacks are discussed. The objective is to detect jammers, lessen the effect of DoS attacks, and improve wireless communication security. Key jamming criteria like energy efficiency, detection probability, denial-of-service level, and strength against physical layer techniques are also outlined.
This document summarizes and evaluates scheduling algorithms for wireless IP networks that support multiclass traffic. It begins by describing the challenges of providing quality of service (QoS) over wireless networks due to time-varying transmission quality and location-dependent errors. It then reviews existing scheduling algorithms like weighted fair queuing (WFQ) and discusses their limitations in wireless environments. The document proposes a new scheduling mechanism that differentiates service between traffic classes and subclasses, allows compensation for non-real time traffic, and adjusts weights of real-time flows in error states to maintain throughput. Overall, the scheduling algorithm aims to provide QoS, fairness between flows, and flexibility to adapt to changing wireless channel conditions.
This document presents an Intelligent Vertical Handoff Algorithm (IVHA) that uses fuzzy logic to improve handoff decisions between heterogeneous wireless networks. The algorithm has two phases: 1) The handoff initialization phase uses fuzzy logic to adaptively set the handoff threshold based on RSSI, SINR, and data rate to trigger handoffs at the right time. 2) The handoff decision phase uses fuzzy logic to select the best network among available options based on bandwidth, network load, coverage, and user velocity. The algorithm aims to improve quality of service by reducing problems like ping-ponging during handoffs.
This document summarizes security issues related to mobile devices, networks, and communication. It discusses how mobile devices store sensitive data and access various networks, raising security concerns. Issues addressed include unauthorized access of data on lost or stolen devices, insecure communication channels, and vulnerabilities in mobile networks like cellular networks. The document also examines existing security measures and the need for improved solutions to address issues like authentication, encryption, and access control across mobile technologies.
This document reviews various video steganography methods that use neural networks. It discusses how neural networks can be used for steganalysis to detect hidden data in digital media. The document provides an overview of different neural network approaches that have been used for video watermarking and audio digital watermarking. These include using neural networks to preferentially allocate watermarks to motion coefficients in video and memorizing watermarks in the neurons of a counterpropagation neural network for audio. The conclusion states that neural network techniques can help improve the performance of various video steganography methods.
This document summarizes a study on landslides in the Western Ghat region of Maharashtra, India. It discusses the causes of landslides including heavy rainfall, erosion, and human activities like deforestation and construction. The major types of landslides addressed are creep, slump slides, debris avalanches, earth flows, and rock falls. The document then provides details on specific landslide events in the study area and suggests mitigation approaches like drainage control, stabilizing slope materials, planting vegetation, and setting structures back from slopes to minimize landslide risks and impacts.
This study investigated the use of low-cost agricultural waste materials as biosorbents for removing chromium (VI) from wastewater. Batch experiments were conducted using sweetlime fruit skin and bagasse to adsorb chromium (VI) at different concentrations, pH levels, and adsorbent amounts. The results showed that adsorption was most effective at lower chromium (VI) concentrations and acidic pH levels. Sweetlime fruit skin achieved 65% removal at 40 μg/L chromium (VI) and pH 2.5, while bagasse achieved 75% removal at the same concentration and pH 5. The study suggests that locally available agricultural wastes have potential as low-cost biosorbents for wastewater treatment.
This document summarizes a study on segmenting cysts in breast ultrasound images using texture features and an active contour method. The authors apply the Chan-Vese level set method to segment cyst regions based on texture features calculated from the images using different kernel sizes. Segmentation performance is evaluated using measures like area error rate, DICE coefficient, sensitivity and Hausdorff distance. The results show that mean texture features and preprocessing the images with Qui's mask produce more accurate segmentations with lower error rates compared to other texture features and kernels.
This document describes a proposed detection and warning system for railway tracks using wireless sensors. The system uses MEMS sensors, GPS, GSM, and ultrasonic sensors to monitor tracks and bridges for damage like cracks or structural issues. If a problem is detected, the system would immediately notify trains in the area through wireless communication. It discusses the technical components of the system in detail, including the microcontroller, sensors, GPS/GSM modules, and wireless transmission. The system aims to more quickly detect track issues and notify trains to prevent delays compared to existing systems. It provides block diagrams of the sensor network components and how they would function on the tracks and on trains.
This document summarizes a study on predicting creep life for gas turbine discs. Creep is an important design consideration for gas turbine components operating at high temperatures. The study models a typical gas turbine disc and analyzes temperature distribution, stresses, and creep strains using finite element analysis. An Inconel 718 disc is modeled and analyzed to determine creep life and emphasize creep considerations in gas turbine disc design. Thermal, stress, and creep analyses are performed to evaluate creep life and strains at different stress levels over time.
This document discusses digital stereoscopic imaging techniques. It presents a system for capturing multiple orientation stereo views of museum objects using a high resolution digital camera. Key aspects covered include the superior image quality of high-end digital cameras compared to film and lower-end digital cameras. It also examines various parameters that must be considered when capturing stereoscopic views, such as camera settings, image capture geometry, and background selection, in order to obtain high quality stereoscopic images.
The document discusses stereoscopic 3D production. It covers the differences between 3D and S3D, natural depth cues, depth perception, the business case for 3D, 3D storytelling techniques, stereoscopic technology formats, live 3D engineering challenges, Sky 3D broadcast models, S3D cinematography theory, S3D pre-production including depth budget and script, stereoscopic editing, and summarizes stereoscopic 3D tools.
Tele-immersion is a technology that uses 3D video conferencing to allow users in different locations to interact in a simulated holographic environment as if they are together in person. It was first proposed in 1997 and involves using arrays of cameras to capture 3D environments and people, transmitting the compressed 3D video streams over high-bandwidth networks, and displaying the streams so participants appear life-like and co-located. While challenging due to bandwidth and rendering needs, tele-immersion has applications for remote education, medical training, teleconferencing, and entertainment by reducing travel while enabling face-to-face style interactions from anywhere.
FPGA Based Pattern Generation and Synchonization for High Speed Structured Li...TELKOMNIKA JOURNAL
Recently, structured light 3D imaging devices have gained a keen attention due to their potential
applications to robotics, industrial manufacturing and medical imaging. Most of these applications require
high 3D precision yet high speed in image capturing for hard and/or soft real time environments. This
paper presents a method of high speed image capturing for structured light 3D imaging sensors with FPGA
based structured light pattern generation and projector-camera synchronization. Suggested setup reduces
the time for pattern projection and camera triggering to 16msec from 100msec that should be required by
conventional methods.
Automatic 2D to 3D Video Conversion For 3DTV'sRishikese MR
The seminar discuss about a little old technology still a main topic. Automatic 2D to 3D Video Conversion for 3DTV's. the slides have about 3Dtv, Need of 3Dtv, Various approaches to convert 2D to 3D, Extraction of scene depth information, Advantages & Disadvantages, Application of 3D TV, etc.
iMinds insights - 3D Visualization TechnologiesiMindsinsights
Transforming the way we deal with information - from consumption to interaction.
iMinds insights is a quarterly publication providing you with relevant tech updates based on interviews with academic and industry experts. iMinds is a digital research center and incubator based in Belgium.
The objective of a video communication system is to deliver the maximum of video data from the source to the destination through a communication channel using all of its available bandwidth. To achieve this objective, the source coding should compress the original video sequence as much as possible and the compressed video data should be robust and resilient to channel errors. However, while achieving a high coding efficiency, compression also makes the coded video bitstream vulnerable to transmission errors. Thus, the process of video data compression tends to work against the objectives of robustness and resilience to errors. Therefore, extra information that needs to be transmitted in 3-D video has brought new challenge and consumer applications will not gain more popularity unless the 3-D video coding problems are addressed.
The objective of a video communication system is to deliver the maximum of video data from the source to the destination through a communication channel using all of its available bandwidth. To achieve this objective, the source coding should compress the original video sequence as much as possible and the compressed video data should be robust and resilient to channel errors. However, while achieving a high coding efficiency, compression also makes the coded video bitstream vulnerable to transmission errors. Thus, the process of video data compression tends to work against the objectives of robustness and resilience to errors. Therefore, extra information that needs to be transmitted in 3-D video has brought new challenge and consumer applications will not gain more popularity unless the 3-D video coding problems are addressed.
PROPOSED SYSTEM FOR MID-AIR HOLOGRAPHY PROJECTION USING CONVERSION OF 2D TO 3...IAEME Publication
Holographic Projections is one of the world’s top trending technologies. As the technology has graphical interaction, it is quite intriguing. Due to the multifaceted nature, multinational corporations are using this technology to a wide extent for their various purposes. The primary advantage of the technology is that the computer vision is constantly evolving and it has made possible to view content in 3D. The technology has lot of scope to evolve such that it can be made feasible to the common people. The sole purpose of interaction with 3D holographic projection is to bring life to a virtual image in 3D world.
HA5 – COMPUTER ARTS BLOG ARTICLE – 3D: The Basicshamza_123456
This document discusses 3D modeling techniques for movies versus games. It explains that movie models can have millions of polygons while game models need to be more efficient to maintain performance. Game models often use techniques like normal mapping to add detail without increasing polygons. It also discusses differences in level of detail models and how not everything needs to be modeled in movies.
This document discusses 3D modeling techniques for movies versus games. It explains that movies can use higher polygon counts and various modeling techniques, while games need more efficient, lower polygon models to maintain performance. Techniques like normal mapping are used to add detail to game assets without increasing polygon counts. Level of detail (LOD) modeling is also discussed where lower resolution models are used at a distance. The document also covers differences in what needs to be modeled, such as only modeling visible parts for movies but full 360 degree models for games.
HA5 – COMPUTER ARTS BLOG ARTICLE – 3D: The Basicshamza_123456
This document provides information on 3D modeling techniques for movies versus games. It discusses how movie models can have millions of polygons while game models need to be more efficient, often using fewer than 10,000 polygons. Normal maps are described as a technique to add surface detail without adding polygons. Level of detail (LOD) modeling is discussed for both movies and games. Overall, the techniques differ due to movies having no interactivity or frame rate requirements, while games need efficient, real-time rendering.
This document presents a novel approach for jointly optimizing spatial prediction and transform coding in video compression. It aims to improve performance and reduce complexity compared to existing techniques. The proposed method uses singular value decomposition (SVD) to compress images. SVD decomposes an image matrix into three matrices, allowing the image to be approximated using only a few singular values. This achieves compression by removing redundant information. The document outlines the SVD approach for image compression and measures compression performance using compression ratio and mean squared error between the original and compressed images. It then discusses trends in image and video coding, including combining natural and synthetic content. Finally, it provides a block diagram of the proposed system and compares its compression performance to existing discrete cosine transform-
DISTRIBUTED SYSTEM FOR 3D REMOTE MONITORING USING KINECT DEPTH CAMERAScscpconf
This article describes the design and development ofa system for remote indoor 3D monitoring
using an undetermined number of Microsoft® Kinect sensors. In the proposed client-server
system, the Kinect cameras can be connected to different computers, addressing this way the
hardware limitation of one sensor per USB controller. The reason behind this limitation is the
high bandwidth needed by the sensor, which becomes also an issue for the distributed system
TCP/IP communications. Since traffic volume is too high, 3D data has to be compressed before
it can be sent over the network. The solution consists in self-coding the Kinect data into RGB
images and then using a standard multimedia codec to compress color maps. Information from
different sources is collected into a central client computer, where point clouds are transformed
to reconstruct the scene in 3D. An algorithm is proposed to conveniently merge the skeletons
detected locally by each Kinect, so that monitoring of people is robust to self and inter-user
occlusions. Final skeletons are labeled and trajectories of every joint can be saved for event
reconstruction or further analysis.
CT-SVD and Arnold Transform for Secure Color Image WatermarkingAM Publications,India
Watermarking is used for protecting copyright of digital images. In this paper, we propose a novel technique for watermarking using Contourlet Transform (CT) and Singular Value Decomposition (SVD). CT ensures imperceptibility of the watermark and SVD ensures its robustness against attacks. Arnold transform is used for scrambling watermark pixels to ensure watermark security. Watermark extraction is semi-blind, which avoids the need for original image for extraction. Both watermark and cover image are color images. Performance of the system is judged by using PSNR and Correlation Coefficient (CC) values. System shows good robustness against noise, JPEG compression, filtering and cropping
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
AN ADAPTIVE REMOTE DISPLAY FRAMEWORK TO IMPROVE POWER EFFICIENCY cscpconf
As computing performance and network technology have evolved, mobile device users can enjoy high quality multimedia more easily. Remote Display - the technology which mirrors the screen of one device to another device - allows handheld mobile devices to share their screen contents with larger-sized display devices such as TVs. However, there is general concern about high power consumption caused by complex computation for encoding and continuous data transmission in the mobile devices. In this paper, we present an adaptive remote display framework considering and utilizing the processing capability of display device. By supporting the Content Mirroring Mode, we can skip unnecessary steps and perform core activities to improve power efficiency and extend overall processing capability.
This document summarizes a research paper that examines pricing strategy in a two-stage supply chain consisting of a supplier and retailer. The supplier offers a credit period to the retailer, who then offers credit to customers. A mathematical model is formulated to maximize total profit for the integrated supply chain system. The model considers three cases based on the relative lengths of the credit periods offered at each stage. Equations are developed to represent the profit functions for the supplier, retailer and overall system in each case. The goal is to determine the optimal selling price that maximizes total integrated profit.
The document discusses melanoma skin cancer detection using a computer-aided diagnosis system based on dermoscopic images. It begins with an introduction to skin cancer and melanoma. It then reviews existing literature on automated melanoma detection systems that use techniques like image preprocessing, segmentation, feature extraction and classification. Features extracted in other studies include asymmetry, border irregularity, color, diameter and texture-based features. The proposed system collects dermoscopic images and performs preprocessing, segmentation, extracts 9 features based on the ABCD rule, and classifies images using a neural network classifier to detect melanoma. It aims to develop an automated diagnosis system to eliminate invasive biopsy procedures.
This document summarizes various techniques for image segmentation that have been studied and proposed in previous research. It discusses edge-based, threshold-based, region-based, clustering-based, and other common segmentation methods. It also reviews applications of segmentation in medical imaging, plant disease detection, and other fields. While no single technique can segment all images perfectly, hybrid and adaptive methods combining multiple approaches may provide better results. Overall, image segmentation remains an important but challenging task in digital image processing and computer vision.
This document presents a test for detecting a single upper outlier in a sample from a Johnson SB distribution when the parameters of the distribution are unknown. The test statistic proposed is based on maximum likelihood estimates of the four parameters (location, scale, and two shape) of the Johnson SB distribution. Critical values of the test statistic are obtained through simulation for different sample sizes. The performance of the test is investigated through simulation, showing it performs well at detecting outliers when the contaminant observation represents a large shift from the original distribution parameters. An example application to census data is also provided.
This document summarizes a research paper that proposes a portable device called the "Disha Device" to improve women's safety. The device has features like live location tracking, audio/video recording, automatic messaging to emergency contacts, a buzzer, flashlight, and pepper spray. It is designed using an Arduino microcontroller connected to GPS and GSM modules. When the button is pressed, it sends an alert message with the woman's location, sets off an alarm, activates the flashlight and pepper spray for self-defense. The goal is to provide women a compact, one-click safety system to help them escape dangerous situations or call for help with just a single press of a button.
- The document describes a study that constructed physical fitness norms for female students attending social welfare schools in Andhra Pradesh, India.
- Researchers tested 339 students in classes 6-10 on speed, strength, agility and flexibility tests. Tests included 50m run, bend and reach, medicine ball throw, broad jump, shuttle run, and vertical jump.
- The results showed that 9th class students had the best average time for the 50m run. 10th class students had the highest flexibility on average. Strength and performance generally improved with increased class level.
This document summarizes research on downdraft gasification of biomass. It discusses how downdraft gasifiers effectively convert solid biomass into a combustible producer gas. The gasification process involves pyrolysis and reactions between hot char and gases that produce CO, H2, and CH4. Downdraft gasifiers are well-suited for biomass gasification due to their simple design and ability to manage the gasification process with low tar production. The document also reviews previous studies on gasifier configuration upgrades and their impact on performance, and the principles of downdraft gasifier operation.
This document summarizes the design and manufacturing of a twin spindle drilling attachment. Key points:
- The attachment allows a drilling machine to simultaneously drill two holes in a single setting, improving productivity over a single spindle setup.
- It uses a sun and planet gear arrangement to transmit power from the main spindle to two drilling spindles.
- Components like gears, shafts, and housing were designed using Creo software and manufactured. Drill chucks, bearings, and bits were purchased.
- The attachment was assembled and installed on a vertical drilling machine. It is aimed at improving productivity in mass production applications by combining two drilling operations into one setup.
The document presents a comparative study of different gantry girder profiles for various crane capacities and gantry spans. Bending moments, shear forces, and section properties are calculated and tabulated for 'I'-section with top and bottom plates, symmetrical plate girder, 'I'-section with 'C'-section top flange, plate girder with rolled 'C'-section top flange, and unsymmetrical plate girder sections. Graphs of steel weight required per meter length are presented. The 'I'-section with 'C'-section top flange profile is found to be optimized for biaxial bending but rolled sections may not be available for all spans.
This document summarizes research on analyzing the first ply failure of laminated composite skew plates under concentrated load using finite element analysis. It first describes how a finite element model was developed using shell elements to analyze skew plates of varying skew angles, laminations, and boundary conditions. Three failure criteria (maximum stress, maximum strain, Tsai-Wu) were used to evaluate first ply failure loads. The minimum load from the criteria was taken as the governing failure load. The research aims to determine the effects of various parameters on first ply failure loads and validate the numerical approach through benchmark problems.
This document summarizes a study that investigated the larvicidal effects of Aegle marmelos (bael tree) leaf extracts on Aedes aegypti mosquitoes. Specifically, it assessed the efficacy of methanol extracts from A. marmelos leaves in killing A. aegypti larvae (at the third instar stage) and altering their midgut proteins. The study found that the leaf extract achieved 50% larval mortality (LC50) at a concentration of 49 ppm. Proteomic analysis of larval midguts revealed changes in protein expression levels after exposure to the extract, suggesting its bioactive compounds can disrupt the midgut. The aim is to identify specific inhibitor proteins in the midg
This document presents a system for classifying electrocardiogram (ECG) signals using a convolutional neural network (CNN). The system first preprocesses raw ECG data by removing noise and segmenting the signals. It then uses a CNN to extract features directly from the ECG data and classify arrhythmias without requiring complex feature engineering. The CNN architecture contains 11 convolutional layers and is optimized using techniques like batch normalization and dropout. The system was tested on ECG datasets and achieved classification accuracy of over 93%, demonstrating its effectiveness at automated ECG classification.
This document presents a new algorithm for extracting and summarizing news from online newspapers. The algorithm first extracts news related to the topic using keyword matching. It then distinguishes different types of news about the same topic. A term frequency-based summarization method is used to generate summaries. Sentences are scored based on term frequency and the highest scoring sentences are selected for the summary. The algorithm was evaluated on news datasets from various newspapers and showed good performance in intrinsic evaluation metrics like precision, recall and F-score. Thus, the proposed method can effectively extract and summarize online news for a given keyword or topic.
1. E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
i
AUTOMATIC VIEW SYNTHESIS FROM
STEREOSCOPIC 3D
BY IMAGE DOMAIN WARPING
Mrs. Mayuri T. Deshmukh Miss.Ashwini T. Sharnagat
Assistant Professor ME Ist
Digital Electronic
S.S.B.T’s C.O.E.T. Bambhori. S.S.B.T’s C.O.E.T. Bambhori.
Email id:mayuri.deshmukh08@gmail.com Email id:aashwini179@gmail.com
ABSTARCT- This paper focuses on the automatic view synthesis by image domain warping is presented
that automatically synthesizes new views directly from S3D video and functions completely. Nowadays,
stereoscopic 3D (S3D) cinema is already conventional and almost all new display devices for the home
support S3D content. S3D giving out communications to the home is already established partly in the
form of 3D Blu-ray discs, video on demand services, or television channels. The need to wear glasses is,
however, often considered as an obstacle, which hinders broader acceptance of this technology in the
home. Multiviewautosterescopic displays make possible a glasses free perception of S3D content for
several observers simultaneously and support head motion parallax in a limited range. To verify the
entire system the result is being observed on 3D television, 3D cinema’s for view the video allow a glasses
free perception.
Index Terms—Three dimensional TV, auto stereoscopic displays, 3D Blu-ray discs, head motion parallax
1. INTRODUCTION
STEREOSCOPIC 3D (S3D) cinema and television are in the process of changing the scenery of entertainment.
Primarily responsible for the change is the fact that technologies ranging from 3D content creation, to data
compression and transmission, to 3D display devices are progressively improving and adapted to enable a rich
and higher quality 3D experience. However, the necessity to wear glasses is often regarded as a main obstacle of
today’s conventional stereoscopic 3D display systems. Multi-view auto stereoscopic displays (MAD) overcome
this problem. They allow glasses free stereo viewing by emitting several images at the same time.
Stereoscopic 3D can expand users’ experiences beyond traditional2D-TV broadcasting by
contribution programs with depth idea of the observed scenes. IN fact, 3D has been successfully commercialized
as stereo movies, such as those by IMAX, for people to watch in the cinema, using special devices. Given that
the popularity of 3D programs has dramatically increased, 3D-TV has been known as a possible break through
for conventional TV technologies to satisfy the coming need for watching 3D programs at home. Typical MADs
which are on the market today require 8-views, 9-views or even 28-views as input. Because of the different
number of input views required by different MADs, no unique display format exists for such displays.
According to the formats involved in the distribution chain (Fig.1.), the transmission format has to enable such a
decoupling. Hence, a good transmission format has to fulfil the following requirements
• An automatic conversion from production to transmission format has to be possible. For live broadcast
applications also real-time conversion is required.
2. Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at:
• The transmission format has to allow an automatic and rea
format.
• The transmission format has to be well compressible to
and consumer 3D content production
stereoscopic displays. It is believed t
Stefanoski,2013]
Fig.1. High level view on the usual format conversions
Research communities and standardization bodies continue to
and enable efficient generation of novel views as required by MADs. Such formats can be divided into two
classes: video only formats like S3D and multiview video in general, and depth enhanced formats like si
view video plus depth and multiview video plus depth. The per
enhanced formats provides information on the depth structure of the 3D scene. Such depth data can be used for
view synthesis by depth-image bas
2004].
However, high quality view synthesis with DIBR requires high quality depth data. There exist
stereo algorithms which can automatically compute depth maps from stereo images, or depth sensors which can
capture depth usually of too little
content productions. Thus, today highly accurate depth maps are usually generated in a semiautomatic
where stereo algorithms or depth sensors are used to estimate or capture initial depth ma
improved in an interactive process. The most simple and cost efficient conversion from S3D As a
format to a transmission format consists of conducting only low level conversion steps (like colour space, bit
depth, frame-rate, or image resolution conversions). Such a transmission format can be
compatible to the existing S3D distribution and stereoscopic display infrastructure. Thus, using a transmission
format without supplementary depth data prevents an i
presented synthesis method can be used at the decoder side to synthesize new views directly from transmitted
S3D content. for real-time view synthesis is presented and analysed.
Fig.2 Stereo pinhole camera model illustrating the projection of a
It shows a bird’s-eye view of a pin
cameras are located at positions CL
E
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
The transmission format has to allow an automatic and real-time conversion into any particular
• The transmission format has to be well compressible to save transmission band width. nowadays
and consumer 3D content production is dominated by S3D content, i.e. 2-view content which is
stereoscopic displays. It is believed that S3D as aproduction format will dominate over years
1. High level view on the usual format conversions required from content Production to display
Research communities and standardization bodies continue to investigate formats which are well compressible
and enable efficient generation of novel views as required by MADs. Such formats can be divided into two
classes: video only formats like S3D and multiview video in general, and depth enhanced formats like si
view video plus depth and multiview video plus depth. The per-pixel depth information included in the depth
enhanced formats provides information on the depth structure of the 3D scene. Such depth data can be used for
e based rendering (DIBR) [Masayuki T, Zhao Y, and C. Zhu
However, high quality view synthesis with DIBR requires high quality depth data. There exist
stereo algorithms which can automatically compute depth maps from stereo images, or depth sensors which can
accuracy to allow a high quality synthesis as required e.g. in professional
, today highly accurate depth maps are usually generated in a semiautomatic
where stereo algorithms or depth sensors are used to estimate or capture initial depth ma
improved in an interactive process. The most simple and cost efficient conversion from S3D As a
format to a transmission format consists of conducting only low level conversion steps (like colour space, bit
or image resolution conversions). Such a transmission format can be well
compatible to the existing S3D distribution and stereoscopic display infrastructure. Thus, using a transmission
format without supplementary depth data prevents an increase of content production and distribution costs. The
presented synthesis method can be used at the decoder side to synthesize new views directly from transmitted
time view synthesis is presented and analysed.
Stereo pinhole camera model illustrating the projection of a
3D point pinto the image planes of both cameras.
eye view of a pin-hole model, parallel stereo camera setup. Projection centers of both
and CR at a baseline distance of b and both cameras have a focal length of
E-ISSN: 2321–9637
International Journal of Research in Advent Technology
79
particular N-view display
width. nowadays, professional
view content which is watchable on
production format will dominate over years[Nikolce and
Production to display
investigate formats which are well compressible
and enable efficient generation of novel views as required by MADs. Such formats can be divided into two
classes: video only formats like S3D and multiview video in general, and depth enhanced formats like single
pixel depth information included in the depth
enhanced formats provides information on the depth structure of the 3D scene. Such depth data can be used for
, and C. Zhu, 2012], [Fehn C,
However, high quality view synthesis with DIBR requires high quality depth data. There exist
stereo algorithms which can automatically compute depth maps from stereo images, or depth sensors which can
s required e.g. in professional
, today highly accurate depth maps are usually generated in a semiautomatic process,
where stereo algorithms or depth sensors are used to estimate or capture initial depth maps which are then
improved in an interactive process. The most simple and cost efficient conversion from S3D As a production
format to a transmission format consists of conducting only low level conversion steps (like colour space, bit
well compressed and is
compatible to the existing S3D distribution and stereoscopic display infrastructure. Thus, using a transmission
ncrease of content production and distribution costs. The
presented synthesis method can be used at the decoder side to synthesize new views directly from transmitted
hole model, parallel stereo camera setup. Projection centers of both
and both cameras have a focal length of f
3. E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
80
projecting a 3D point P, which is located at a distance Z from the projection centers, into the
respective.projection planes gives the projected points pL and pR. Hence, the projected points have an image
disparity of
d = (pL − cL) − (pR − cR) = (1)
Obviously, a synthesis of a new view at a position
new = (1 − ) + C (2)
Corresponds to having a baseline distance
= (3)
with respect to the left view. Hence, that would give disparities
= , (4)
i.e. a linear rescaling of all previous disparities.
2. IMAGE-DOMAIN-WARPING
Automatic view synthesis technology which synthesizes new views from 2-view video is highly popular. Such
synthesis technology would be compatible to any S3D distribution infrastructure to the home. Image-domain
Warping (IDW) is a view synthesis method which is able to automatically synthesize new views based on
stereoscopic video input. In contrast to synthesis methods based on DIBR, which relies on dense disparity or
depth maps, IDW employs only sparse disparities to synthesize a novel view. It use the facts that our human
visual system is not able to very exactly estimate absolute depth and that it is not responsive to image distortions
up to a certain level as long as images remain visually plausible, e.g. image distortions can be hidden in non-
salient regions. They are used to compute an image warp which enforces desired sparse disparities in the final
synthesized image while distortions are hidden in non-salient regions. To find out which disparities have to be
enforced in a synthesized image.[ Yin Zhao and Ce Zhu ,2011]
Goal: “warp” the pixels of the image so that they appear in the correct place for a new viewpoint. An
advantage of IDW is, there is no need a geometric model of the object/environment can be done in time
proportional to screen size and (Mostly) independent of object/environment complexity. Very less Disadvantage
are their require Limited resolution and Excessive warping reveals several visual artifacts a linear rescaling of
all previous disparities d is required to synthesize the new view.
In general, a stereo image pair doesn’t contain sufficient information to completely describe an image captured
from a slightly different camera n position. IDW uses image saliency information to deal with this problem.[
Aliaga ,2010]
I. Image Warps: We define a warp as a function that deforms the parameter domain of an image
W:[0, W] × [0, H]→R2 (5)
Where W and H are the width and height of the image, respectively. Image warps have a long history of use in
computer vision and graphics based problems goal of the IDW method is to compute a warping of Each of the
initial stereo images that can be used to produce an output image meeting predefined properties (e.g. scaled
image disparities). To do this, we formulate a quadratic energy functional E (w). A warp w is then warps defined
at regular grid positions.
W [p, q] := w( ∆x p, ∆y q). (6)
4. Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at:
Fig.3 E
II. View Synthesis: The IDW algorithm computes N
separated into four modules, In the Warp Calculator only that warps are computed which are necessary for the
Synthesis of A view that is located
minimizing their respective error functional. Calculated warps are then used
Interpolator/Extrapolator module to interpolate or extrapolate the warps that are necessary to synthesize the
output views, as required by a particular multi
warped module, images are warped to synthesize the output images.
Fig. 4 Block diagram of the view synthesizer which converts 2
I. Data Extraction: First, a sparse set of disparity features is extracted. These sparse disparities are estimated
an automatic, accurate and robust way.Disparities of vertical image edges are particularly important for the
stereo sis, i.e. the perceived depth. For this reason, additional features and corresponding disparities are detected
such that features lay uniformly distributed on nearly vertical image edges. Disparities of detected features are
estimated using the Lucas-Kanade method. The availability of such features is also important to prevent salient
synthesis errors with IDW like bending edges in the synt
Fig. 5 Disparities
II. Warp Calculation: Two warps
camera position located in the center between the two input cameras. For a given sparse set of disparity features
(xL, xR) ∈F, Disparities b
= !
"
E
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
3 Example of a warping function that deforms an input image.
The IDW algorithm computes N-view video from 2-view video where
Warp Calculator only that warps are computed which are necessary for the
of A view that is located in the middle between two input views. These warps are calculated by
minimizing their respective error functional. Calculated warps are then used
Interpolator/Extrapolator module to interpolate or extrapolate the warps that are necessary to synthesize the
output views, as required by a particular multi-view autosterescopic display. Finally, in the Image
warped to synthesize the output images.
diagram of the view synthesizer which converts 2-view videoto N-view video.
First, a sparse set of disparity features is extracted. These sparse disparities are estimated
an automatic, accurate and robust way.Disparities of vertical image edges are particularly important for the
, i.e. the perceived depth. For this reason, additional features and corresponding disparities are detected
formly distributed on nearly vertical image edges. Disparities of detected features are
Kanade method. The availability of such features is also important to prevent salient
synthesis errors with IDW like bending edges in the synthesized image
Fig. 5 Disparities estimated at sparse feature positions
Two warps wL and wR are computed which warp images IL and IR
camera position located in the center between the two input cameras. For a given sparse set of disparity features
(7)
E-ISSN: 2321–9637
International Journal of Research in Advent Technology
81
view video where N >2. It can be
Warp Calculator only that warps are computed which are necessary for the
views. These warps are calculated by
minimizing their respective error functional. Calculated warps are then used in the Warp
Interpolator/Extrapolator module to interpolate or extrapolate the warps that are necessary to synthesize the N
view autosterescopic display. Finally, in the Image-domain
view video.
First, a sparse set of disparity features is extracted. These sparse disparities are estimatedin
an automatic, accurate and robust way.Disparities of vertical image edges are particularly important for the
, i.e. the perceived depth. For this reason, additional features and corresponding disparities are detected
formly distributed on nearly vertical image edges. Disparities of detected features are
Kanade method. The availability of such features is also important to prevent salient
IR, respectively, to a
camera position located in the center between the two input cameras. For a given sparse set of disparity features
)
5. Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at:
have to be enforced. Each warp is computed as the result of an energy minimization problem. Anenergy
functional E is defined, and minimizing it yields a warp
view synthesis. The energy functional is de
terms which are related to a particular type of constraint as described below. Each term is
weighted with a parameter λ
#($) ≔ &#
In Warp calculation 4 constraints are used
Smoothness, Constraints and Energy
equation system. The number of spatial and temporal smoothness cons
of freedom of the warp Wto be solved, while the number of disparity constraints depends on the number of
detected features, which is content dependent.
III. Warp Interpolation/Extrapolation
camera positions as input. The main reason for this restriction is to reduce the overall computational complexity
of the warp calculation. Furthermore
depend on the number of output views required by a particular display system.
Fig.6 Cameras, camera positions, and associated warps
IV. Image-Domain Warping: An output image at a position
based on the input image which is closer to the desired output position. Because warps are continuous, no holes
can occur in the synthesized image. In particular, open regions are implicitly unpainted by stretching unsealing
texture from the neighborhood into the region
synthesis results in practice as long as views are synthesizedwhich are in the range
However, if only one image is used for the s
output image.
3. TRANSMISSION OF WARPS AS SUPPLEMENTARY DATA
To reduce the computational complexity at the receiver side, we modify the transmission system which was
proposed in fig.7 The modified system is shown in fig.7 Thus, it is proposed to shift the warp extraction and
warp calculation part to the sending side, and, in addition to the multiview data, to efficiently compress and
transmit the warp calculation result, i.e. a restricted s
Fig.7
E
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
have to be enforced. Each warp is computed as the result of an energy minimization problem. Anenergy
is defined, and minimizing it yields a warp w that creates the desired change of disparities after
view synthesis. The energy functional is defined with help of the extracted data and consists of three additive
terms which are related to a particular type of constraint as described below. Each term is
#&(') + (#((') + )#)(') (8)
4 constraints are used Disparity Constraints, Spatial Smoothness Constraints
Smoothness, Constraints and Energy Minimization.This functional always represents an over
equation system. The number of spatial and temporal smoothness constraints is dense in the number of degrees
of freedom of the warp Wto be solved, while the number of disparity constraints depends on the number of
detected features, which is content dependent.
Warp Interpolation/Extrapolation: Multiview auto stereoscopic displays require many views from different
camera positions as input. The main reason for this restriction is to reduce the overall computational complexity
calculation. Furthermore, using this approach, the complexity of the warp computation does not
depend on the number of output views required by a particular display system.
Fig.6 Cameras, camera positions, and associated warps
An output image at a position α is synthesized according to
based on the input image which is closer to the desired output position. Because warps are continuous, no holes
can occur in the synthesized image. In particular, open regions are implicitly unpainted by stretching unsealing
re from the neighborhood into the regionWe noticed that this kind of implicit unpointing provides good
synthesis results in practice as long as views are synthesizedwhich are in the range −
However, if only one image is used for the synthesis, empty regions can occur on the left or r
TRANSMISSION OF WARPS AS SUPPLEMENTARY DATA
To reduce the computational complexity at the receiver side, we modify the transmission system which was
modified system is shown in fig.7 Thus, it is proposed to shift the warp extraction and
warp calculation part to the sending side, and, in addition to the multiview data, to efficiently compress and
transmit the warp calculation result, i.e. a restricted set of warps
Fig.7 Modified transmission and view synthesis system
E-ISSN: 2321–9637
International Journal of Research in Advent Technology
82
have to be enforced. Each warp is computed as the result of an energy minimization problem. Anenergy
that creates the desired change of disparities after
fined with help of the extracted data and consists of three additive
)
Disparity Constraints, Spatial Smoothness Constraints,Temporal
This functional always represents an over-constrained
traints is dense in the number of degrees
of freedom of the warp Wto be solved, while the number of disparity constraints depends on the number of
Multiview auto stereoscopic displays require many views from different
camera positions as input. The main reason for this restriction is to reduce the overall computational complexity
ty of the warp computation does not
to i.e. Iα is synthesized
based on the input image which is closer to the desired output position. Because warps are continuous, no holes
can occur in the synthesized image. In particular, open regions are implicitly unpainted by stretching unsealing
We noticed that this kind of implicit unpointing provides good
−0.5 ≤ α ≤ 1.5 (Fig.6).
ynthesis, empty regions can occur on the left or right border of the
To reduce the computational complexity at the receiver side, we modify the transmission system which was
modified system is shown in fig.7 Thus, it is proposed to shift the warp extraction and
warp calculation part to the sending side, and, in addition to the multiview data, to efficiently compress and
6. Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at:
I. Warp Coding with a Dedicated Warp Coder
successively. For each view, they are encoded separately and multiplexed into a single bit stream. Without loss
of simplification, the coding of a warp sequence assigned to one view is described in the foll
the warp at time instant f as w f . Each warp
resolution where each node of the grid is indexed with integer coordinates
∈R2 assigned
I. Spatial Partitioning: A partition consists of a set of 2D locations which we call a group of locations (GOLs).
Each warp w f is partitioned into GOLs using a quincunx resolution pyramid,
II.Intra-Warp and Inter-Warp Prediction:
corresponding prediction modes are supported: INTRA, INTER_P and INTER_B. After prediction, residuals
w[i, j] − ẃf [i, j ] are computed, uniformly quantized, and entropy coded. Quantized residuals of each GOL are
entropy coded independently from other GOLs. In the INTRA prediction mode, all locations of
row-wise and predicted in a closed loop DPCM from previo
shown in Fig.9. Locations of all other GOLs
spatially neigh boring locations in *
II. Warp Coding With Help of a Video Coder
video coding technology, we propose the wa
warp coding method shown.
I. Coding System: Similar to the coding with the dedicated warp coder, warps are encoded separately for each
view and then multiplexed into a single bit stream. To encode a warp, first, a Warp Quantize is used to convert
each warp into an 8-bit grayscale image representation.
E
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
a Dedicated Warp Coder: Warps of all time instances and views are encoded
successively. For each view, they are encoded separately and multiplexed into a single bit stream. Without loss
of simplification, the coding of a warp sequence assigned to one view is described in the foll
. Each warp w f is represented as a regular quad grid, in above
resolution where each node of the grid is indexed with integer coordinates i, j and has a 2D location
Fig. 8 Block-diagram of warp coder.
A partition consists of a set of 2D locations which we call a group of locations (GOLs).
is partitioned into GOLs using a quincunx resolution pyramid,
Warp Prediction: Similar to video coding standards, three warp coding types and
corresponding prediction modes are supported: INTRA, INTER_P and INTER_B. After prediction, residuals
are computed, uniformly quantized, and entropy coded. Quantized residuals of each GOL are
entropy coded independently from other GOLs. In the INTRA prediction mode, all locations of
wise and predicted in a closed loop DPCM from previously scanned spatially neighbouring locations,as it is
shown in Fig.9. Locations of all other GOLs d fl are predicted by the respective centroids computed from
*+,-
. -
.
/
as indicated in Fig.9.
th Help of a Video Coder: To take advantage of already existing and highly sophisticated
video coding technology, we propose the warp coding system show in Fig.10 as an alternative to the dedicated
the coding with the dedicated warp coder, warps are encoded separately for each
view and then multiplexed into a single bit stream. To encode a warp, first, a Warp Quantize is used to convert
bit grayscale image representation.
Fig.9. Warp coding system using a video coder.
E-ISSN: 2321–9637
International Journal of Research in Advent Technology
83
Warps of all time instances and views are encoded
successively. For each view, they are encoded separately and multiplexed into a single bit stream. Without loss
of simplification, the coding of a warp sequence assigned to one view is described in the following. We denote
in above (Fig.3) of fixed
and has a 2D location w f [i, j]
A partition consists of a set of 2D locations which we call a group of locations (GOLs).
Similar to video coding standards, three warp coding types and
corresponding prediction modes are supported: INTRA, INTER_P and INTER_B. After prediction, residuals
are computed, uniformly quantized, and entropy coded. Quantized residuals of each GOL are
entropy coded independently from other GOLs. In the INTRA prediction mode, all locations of d f 1 are scanned
usly scanned spatially neighbouring locations,as it is
are predicted by the respective centroids computed from
To take advantage of already existing and highly sophisticated
as an alternative to the dedicated
the coding with the dedicated warp coder, warps are encoded separately for each
view and then multiplexed into a single bit stream. To encode a warp, first, a Warp Quantize is used to convert
7. Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at:
4. DISCUSSION
1) Recently, the Moving Pictures Experts Group (MPEG) issued a Call for Proposals (CfP) on 3D Video Coding
technology with the goal to identify
• A 3D video format,
• A corresponding efficient compression technology,
• A view synthesis technology which enables an efficient synthesis of new views based on the 3D video
format.
Fig..11. Assessed quality of the MPEG in the Multiview
This includes Data Extraction, Warp Calculation, and Warp Interpolation/Extrapolation.
3D format with 2 views is already supported by existing consumer and professional stereo cameras. With each
proposal to the CfP, compressed bit streams, a decoder, and view synthesis software had to be provided. Bit
streams had to be compressed at predefined target bit rates Proposals were evaluat
the synthesized views through formal subjective testing on both stereoscopic and multiview au
displays. Fig.11 shows the quality assessed on a
results were also assessed on a stereoscopic display; the corresponding stereo sequences can be found here for
download. [Sikora Thomas,1997]
5. CONCLUSION
A view synthesis method based on
from stereoscopic 3D video. It relies on an automatic estimation of sparse disparities and image saliency
information, and enforces target disparities in the synthesized images using an
Image-domain Warping leads to high quality synthesis results without requiring depth map estimation and
transmission. However a reuse of existing video coding technology has the advantage of reduced development
and production costs. For this reason, JCT
coding based on HEVC, which will allow receivers to perform a synthesis of new views based on Image
domain- Warping While a dedicated warp coder can have a stronger coding e
video coding technology for warp coding lies the reduced development and production costs, i.e.in the reuse of
available video coding chips. For this reason,
extend the upcoming 3D-HEVC standard by
E
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
1) Recently, the Moving Pictures Experts Group (MPEG) issued a Call for Proposals (CfP) on 3D Video Coding
ient compression technology,
view synthesis technology which enables an efficient synthesis of new views based on the 3D video
Fig. 10 Transmission and view synthesis system.
. Assessed quality of the MPEG in the Multiview autosterescopic display test scenario
This includes Data Extraction, Warp Calculation, and Warp Interpolation/Extrapolation.
3D format with 2 views is already supported by existing consumer and professional stereo cameras. With each
proposal to the CfP, compressed bit streams, a decoder, and view synthesis software had to be provided. Bit
be compressed at predefined target bit rates Proposals were evaluated by assessing the quality of
he synthesized views through formal subjective testing on both stereoscopic and multiview au
shows the quality assessed on a multiview auto stereoscopic display. Qualitatively similar
results were also assessed on a stereoscopic display; the corresponding stereo sequences can be found here for
A view synthesis method based on Image-domain-Warping. Its approach automatically synthesizes new views
from stereoscopic 3D video. It relies on an automatic estimation of sparse disparities and image saliency
information, and enforces target disparities in the synthesized images using an image warping framework.
domain Warping leads to high quality synthesis results without requiring depth map estimation and
transmission. However a reuse of existing video coding technology has the advantage of reduced development
ts. For this reason, JCT-3V planes to extend the upcoming 3D-HEVC standard by warp
coding based on HEVC, which will allow receivers to perform a synthesis of new views based on Image
While a dedicated warp coder can have a stronger coding efficiency, the
coding lies the reduced development and production costs, i.e.in the reuse of
available video coding chips. For this reason, based on the evaluation results presented in it,
HEVC standard by warp coding based on HEVC. This will enable the transmission
E-ISSN: 2321–9637
International Journal of Research in Advent Technology
84
1) Recently, the Moving Pictures Experts Group (MPEG) issued a Call for Proposals (CfP) on 3D Video Coding
view synthesis technology which enables an efficient synthesis of new views based on the 3D video
display test scenario.
Please note that such a
3D format with 2 views is already supported by existing consumer and professional stereo cameras. With each
proposal to the CfP, compressed bit streams, a decoder, and view synthesis software had to be provided. Bit
ed by assessing the quality of
he synthesized views through formal subjective testing on both stereoscopic and multiview auto stereoscopic
multiview auto stereoscopic display. Qualitatively similar
results were also assessed on a stereoscopic display; the corresponding stereo sequences can be found here for
approach automatically synthesizes new views
from stereoscopic 3D video. It relies on an automatic estimation of sparse disparities and image saliency
image warping framework.
domain Warping leads to high quality synthesis results without requiring depth map estimation and
transmission. However a reuse of existing video coding technology has the advantage of reduced development
HEVC standard by warp
coding based on HEVC, which will allow receivers to perform a synthesis of new views based on Image-
, the advantage of reusing
coding lies the reduced development and production costs, i.e.in the reuse of
based on the evaluation results presented in it, JCT-3V planes to
warp coding based on HEVC. This will enable the transmission of
8. E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
85
multi-view video plus warp data with an internationalstandard, which will allow the use of IDW for view
synthesisat the receiver side.
REFERENCES
[1] Aliaga G. Daniel image warping CS635 Spring 2010
[2] Fehn C, “Depth-image-based rendering (DIBR), compression, and Transmission for a new approach on
3D- TV,” Proc. SPIE, vol. 5291, Vol. 93–104, May 2004
[3] Le Gall D.J, “The MPEG video compression algorithm,” Signal Processing: Image Commune.1992,
vol. 4, no. 4, pp. 129–140
[4] Muller K, Merkle P, and T. Wiegand, “3-D video representation using depth maps,” Proc. IEEE, vol.
99, no. 4, pp. 643–656, Apr. 2011.
[5] Masayuki T, Zhao Y, and C. Zhu, 3D-TV System with Depth- Image-Based Rendering: Architecture,
Techniques and Challenges, 1st
sed. New York, NY, USA: Springer-Verlag, 2012.
[6] Nikolce Stefanoski, Oliver Wang, Manuel Lang, Pierre Greisen, Simon Heinzle, and Aljosa Smolic
IEEE Transactions On Image Processing, Vol. 22, No. 9, September 2013
[7] Smolic A,“3D video and free viewpoint video—From capture to display,” Pattern Recognit., vol. 44,
no. 9, pp. 1958–1968, Sep. 2011.
[8] Sikora Thomas, The MPEG-4 Video Standard Verification Model, Senior Member, IEEE Transactions
On Circuits And Systems For Video Technology, Vol. 7, No. 1, February 1997
[9] Yin Zhao, Ce Zhu,Depth No-Synthesis-Error Model for View Synthesis in 3-D Video Senior Member,
IEEE, Zhenzhong Chen, Member, IEEE, and Lu Yu, Member, IEEE IEEE TRANSACTIONS ON
IMAGE PROCESSING, VOL. 20, NO. 8, AUGUST 2011
[10] Zilly F, Riechert C, P. Eisert, and P. Kauff, “Semantic kernels binarized—A feature descriptor for fast
and robust matching,” in Nov. 2011, pp. 39–48.