At the end of this lecture, you should be able to;
describe sharpening through spatial filters.
Identify usage of derivatives in Image Processing.
discuss edge detection techniques.
compare 1st & 2nd order derivatives used for sharpening.
Apply sharpening techniques for problem solving.
This document discusses various techniques for image enhancement in the frequency domain. It describes three types of low-pass filters for smoothing images: ideal low-pass filters, Butterworth low-pass filters, and Gaussian low-pass filters. It also discusses three corresponding types of high-pass filters for sharpening images: ideal high-pass filters, Butterworth high-pass filters, and Gaussian high-pass filters. The key steps in frequency domain filtering are also summarized.
This document discusses various methods for estimating noise parameters and filtering noise from images. It begins by explaining how to estimate noise parameters such as mean and variance by analyzing sample images. It then covers periodic noise reduction using frequency domain filtering like notch filters. Other filtering methods discussed include direct inverse filtering, Wiener filtering, constrained least squares filtering, and iterative nonlinear restoration using the Lucy-Richardson algorithm. Examples are provided to illustrate Wiener filtering and constrained least squares filtering.
Subband coding decomposes a source signal into constituent frequency bands using digital filters like low-pass and high-pass filters. This separation into subbands allows each frequency component to be encoded and decoded separately, improving compression performance over techniques that treat the whole signal as one. The basic subband coding algorithm involves analysis using filtering and decimation to separate the signal, quantization and coding of the subband signals, and synthesis by decoding, upsampling and reconstruction filtering to reconstruct the original signal. Applications of subband coding include speech coding, audio coding and image compression, with MPEG audio standards using subband coding with 32 filters and bandwidths of f/64.
COM2304: Intensity Transformation and Spatial Filtering – I (Intensity Transf...Hemantha Kulathilake
At the end of this lesson, you should be able to;
describe spatial domain of the digital image.
recognize the image enhancement techniques.
describe and apply the concept of intensity transformation.
express histograms and histogram processing.
describe image noise.
characterize the types of Noise.
describe concept of image restoration.
COM2304: Intensity Transformation and Spatial Filtering – II Spatial Filterin...Hemantha Kulathilake
At the end of this lecture, you should be able to;
describe the fundamentals of spatial filtering.
generating spatial filter masks.
identify smoothing via linear filters and non linear filters.
apply smoothing techniques for problem solving.
The document discusses various types of filters that can be used to reduce noise in digital images, including mean filters, median filters, and order statistics filters. Mean filters include arithmetic, geometric, and harmonic filters, which reduce noise by calculating the mean pixel value within a neighborhood. Median filters select the median pixel value within a neighborhood to reduce salt and pepper noise while retaining edges. Adaptive filters modify their behavior based on statistical properties of local regions in order to better reduce noise without excessive blurring.
This document summarizes techniques for least mean square filtering and geometric transformations. It discusses minimum mean square error (Wiener) filtering, constrained least squares filtering, and geometric mean filtering for noise removal. It also covers spatial transformations, nearest neighbor gray level interpolation, and bilinear interpolation for geometric correction of distorted images. Examples are provided to demonstrate geometric distortion, nearest neighbor interpolation, and bilinear transformation.
This document discusses image restoration and reconstruction techniques for noise removal. It begins by defining image restoration as attempting to reverse degradation processes to restore degraded images. Various noise models are described, including Gaussian, Rayleigh, Erlang, exponential, uniform, and impulse noise. Spatial domain filtering techniques like mean, median, and order statistics filters are covered for noise removal. Frequency domain filtering using band reject filters is also discussed, as well as adaptive filtering techniques. Examples are provided to demonstrate noise removal.
This document discusses various techniques for image enhancement in the frequency domain. It describes three types of low-pass filters for smoothing images: ideal low-pass filters, Butterworth low-pass filters, and Gaussian low-pass filters. It also discusses three corresponding types of high-pass filters for sharpening images: ideal high-pass filters, Butterworth high-pass filters, and Gaussian high-pass filters. The key steps in frequency domain filtering are also summarized.
This document discusses various methods for estimating noise parameters and filtering noise from images. It begins by explaining how to estimate noise parameters such as mean and variance by analyzing sample images. It then covers periodic noise reduction using frequency domain filtering like notch filters. Other filtering methods discussed include direct inverse filtering, Wiener filtering, constrained least squares filtering, and iterative nonlinear restoration using the Lucy-Richardson algorithm. Examples are provided to illustrate Wiener filtering and constrained least squares filtering.
Subband coding decomposes a source signal into constituent frequency bands using digital filters like low-pass and high-pass filters. This separation into subbands allows each frequency component to be encoded and decoded separately, improving compression performance over techniques that treat the whole signal as one. The basic subband coding algorithm involves analysis using filtering and decimation to separate the signal, quantization and coding of the subband signals, and synthesis by decoding, upsampling and reconstruction filtering to reconstruct the original signal. Applications of subband coding include speech coding, audio coding and image compression, with MPEG audio standards using subband coding with 32 filters and bandwidths of f/64.
COM2304: Intensity Transformation and Spatial Filtering – I (Intensity Transf...Hemantha Kulathilake
At the end of this lesson, you should be able to;
describe spatial domain of the digital image.
recognize the image enhancement techniques.
describe and apply the concept of intensity transformation.
express histograms and histogram processing.
describe image noise.
characterize the types of Noise.
describe concept of image restoration.
COM2304: Intensity Transformation and Spatial Filtering – II Spatial Filterin...Hemantha Kulathilake
At the end of this lecture, you should be able to;
describe the fundamentals of spatial filtering.
generating spatial filter masks.
identify smoothing via linear filters and non linear filters.
apply smoothing techniques for problem solving.
The document discusses various types of filters that can be used to reduce noise in digital images, including mean filters, median filters, and order statistics filters. Mean filters include arithmetic, geometric, and harmonic filters, which reduce noise by calculating the mean pixel value within a neighborhood. Median filters select the median pixel value within a neighborhood to reduce salt and pepper noise while retaining edges. Adaptive filters modify their behavior based on statistical properties of local regions in order to better reduce noise without excessive blurring.
This document summarizes techniques for least mean square filtering and geometric transformations. It discusses minimum mean square error (Wiener) filtering, constrained least squares filtering, and geometric mean filtering for noise removal. It also covers spatial transformations, nearest neighbor gray level interpolation, and bilinear interpolation for geometric correction of distorted images. Examples are provided to demonstrate geometric distortion, nearest neighbor interpolation, and bilinear transformation.
This document discusses image restoration and reconstruction techniques for noise removal. It begins by defining image restoration as attempting to reverse degradation processes to restore degraded images. Various noise models are described, including Gaussian, Rayleigh, Erlang, exponential, uniform, and impulse noise. Spatial domain filtering techniques like mean, median, and order statistics filters are covered for noise removal. Frequency domain filtering using band reject filters is also discussed, as well as adaptive filtering techniques. Examples are provided to demonstrate noise removal.
Basic Introduction about Image Restoration (Order Statistics Filters)
Median Filter
Max and Min Filter
MidPoint Filter
Alpha-trimmed Mean filter.
and Brief Introduction to Periodic Noise
Any Question contact kalyan.acharjya@gmail.com
The document discusses image restoration techniques. It describes how images can become degraded through phenomena like motion, improper camera focusing, and noise. The goal of image restoration is to recover the original high quality image from its degraded version using knowledge about the degradation process and types of noise. Common noise models include Gaussian, Rayleigh, Erlang, exponential, and impulse noise. Filtering techniques like mean, order statistics, and adaptive filters can be used for restoration by smoothing the image while preserving edges. The adaptive filters change based on local image statistics to better reduce noise with less blurring than regular filters.
The document discusses different methods for representing segmented image regions, including:
1) Representing regions based on their external (boundary-based) characteristics or internal (pixel-based) characteristics.
2) Common boundary representation methods are boundary following algorithms, chain codes, and polygon approximation.
3) Chain codes represent boundaries as sequences of line segments coded by direction. Polygon approximation finds the minimum perimeter polygon to capture a boundary shape using the fewest line segments.
Image compression involves reducing the size of image files to reduce storage space and transmission time. There are three main types of redundancy in images: coding redundancy, spatial redundancy between neighboring pixels, and irrelevant information. Common compression methods remove these redundancies, such as Huffman coding, arithmetic coding, LZW coding, and run length coding. Popular image file formats include JPEG for photos, PNG for web images, and TIFF, GIF, and DICOM for other uses.
JPEG is a lossy compression method for color or grayscale images. It works best on continuous-tone images where adjacent pixels have similar colors. The JPEG standard defines several modes of operation and uses various techniques like color space transformation, discrete cosine transformation (DCT), quantization, differential pulse-code modulation, run length encoding, and Huffman coding to achieve high compression ratios while maintaining good image quality. Key aspects of the JPEG process include converting images to luminance and chrominance color space, applying DCT, quantizing coefficients, encoding DC values with DPCM, and entropy coding remaining coefficients.
Digital image processing involves techniques to restore degraded images. Image restoration aims to recover the original undistorted image from a degraded observation. The degradation is typically modeled as the original image being operated on by a degradation function and additive noise. Common restoration techniques include spatial domain filters like mean, median and order-statistic filters to remove noise, and frequency domain filtering to reduce periodic noise. The choice of restoration method depends on the type and characteristics of degradation in the image.
Run-length encoding is a data compression technique that works by eliminating redundant data. It identifies repeating characters or values and replaces them with a code consisting of the character and the number of repeats. This compressed encoded data is then transmitted. At the receiving end, the code is decoded to reconstruct the original data. It is useful for compressing any type of repeating data sequences and is commonly used in image compression by encoding runs of black or white pixels. The compression ratio achieved depends on the amount of repetition in the original uncompressed data.
Here in the ppt a detailed description of Image Enhancement Techniques is given which includes topics like Basic Gray level Transformations,Histogram Processing.
Enhancement using Arithmetic/Logic Operations.
image averaging and image averaging methods.
Piecewise-Linear Transformation Functions
Color fundamentals and color models - Digital Image ProcessingAmna
This presentation is based on Color fundamentals and Color models.
~ Introduction to Colors
~ Color in Image Processing
~ Color Fundamentals
~ Color Models
~ RGB Model
~ CMY Model
~ CMYK Model
~ HSI Model
~ HSI and RGB
~ RGB To HSI
~ HSI To RGB
This document discusses various topics related to computer graphics including anti-aliasing, area sampling, the Koch curve, and the C curve. Anti-aliasing is a technique used to reduce jagged edges by blending pixels. Area sampling is an anti-aliasing method that treats pixels as areas and calculates color based on object overlap. The Koch curve is a fractal curve generated by recursively altering line segments. The C curve replaces lines with two shorter lines at 90 degrees to form triangles at each iteration.
Digital Image Processing denotes the process of digital images with the use of digital computer. Digital images are contains various types of noises which are reduces the quality of images. Noises can be removed by various enhancement techniques. Image smoothing is a key technology of image enhancement, which can remove noise in images.
To highlight the contribution made to the total image appearance by specific bits.i.e. Assuming that each pixel is represented by 8 bits, the image is composed of 8 1-bit planes.Useful for analyzing the relative importance played by each bit of the image.
This slide gives you the basic understanding of digital image compression.
Please Note: This is a class teaching PPT, more and detail topics were covered in the classroom.
Digital Image Processing covers intensity transformations that can be performed on images. These include basic transformations like negatives, log transformations, and power-law transformations. It also discusses image histograms, which measure the frequency of each intensity level in an image. Histogram equalization aims to improve contrast by mapping intensities to produce a uniform histogram. It works by spreading out the most frequent intensity values.
This document provides an overview of a research project on image compression. It discusses image compression techniques including lossy and lossless compression. It describes using discrete wavelet transform, lifting wavelet transform, and stationary wavelet transform for image transformation. Experiments were conducted to compare the compression ratio and processing time of different combinations of wavelet transforms, vector quantization, and Huffman/Arithmetic coding. The results were analyzed to evaluate the compression performance and efficiency of the different methods.
This presentation describes briefly about the image enhancement in spatial domain, basic gray level transformation, histogram processing, enhancement using arithmetic/ logical operation, basics of spatial filtering and local enhancements.
This document discusses spatial filtering techniques in image processing. It begins by defining different types of filters based on the frequencies they preserve, such as low-pass, high-pass, band-pass and band-reject. It then explains that spatial filters require defining a neighborhood/mask and an operation. The document focuses on smoothing/low-pass filters which reduce noise and eliminate small details, and sharpening/high-pass filters which highlight fine details. Common smoothing filters discussed include averaging, Gaussian, and median filtering, while common sharpening filters include unsharp masking, high boost filtering, and filters based on image derivatives like gradient and Laplacian. Examples are provided to illustrate the effects of different filters.
This document summarizes digital image processing techniques including algebraic approaches to image restoration and inverse filtering. It discusses:
1) Unconstrained and constrained restoration, with unconstrained having no knowledge of noise and constrained using knowledge of noise.
2) Inverse filtering which is a direct method that minimizes error between degraded and original images using matrix operations, but can be unstable due to noise or near-zero filter values.
3) Pseudo-inverse filtering which adds a threshold to the inverse filter to avoid instability, working better for noisy images by not amplifying high frequency noise.
This document discusses noise in image processing and various methods for noise removal. It defines noise as unwanted signals that can corrupt an image's quality and originality. Common sources of noise include poor image sensors, lens defects, and low light levels. The document outlines different types of noises like Gaussian noise and impulse noise. It then describes various linear and non-linear filters that can be used for noise removal, such as averaging filters, Gaussian filters, median filters, and Wiener filters. The median filter is effective for salt and pepper noise while preserving edges. Adaptive filters can discriminate between corrupted and clean pixels for better noise removal.
The document discusses noise models and methods for removing additive noise from digital images. It describes several types of noise that can affect images, such as Gaussian, impulse, uniform, Rayleigh, gamma and exponential noise. It also presents various noise filters that can be used to remove noise, including mean filters like arithmetic, geometric and harmonic filters, and order statistics filters such as median, max, min and midpoint filters. The filters aim to reduce noise while retaining image detail as much as possible.
At the end of this lecture students should be able to;
Define the C standard functions for managing file input output.
Apply taught concepts for writing programs.
At the end of this lesson, you should be able to;
identify color formation and how color visualize.
describe primary and secondary colors.
describe display on CRT and LCD.
comprehend RGB, CMY, CMYK and HSI color models.
Basic Introduction about Image Restoration (Order Statistics Filters)
Median Filter
Max and Min Filter
MidPoint Filter
Alpha-trimmed Mean filter.
and Brief Introduction to Periodic Noise
Any Question contact kalyan.acharjya@gmail.com
The document discusses image restoration techniques. It describes how images can become degraded through phenomena like motion, improper camera focusing, and noise. The goal of image restoration is to recover the original high quality image from its degraded version using knowledge about the degradation process and types of noise. Common noise models include Gaussian, Rayleigh, Erlang, exponential, and impulse noise. Filtering techniques like mean, order statistics, and adaptive filters can be used for restoration by smoothing the image while preserving edges. The adaptive filters change based on local image statistics to better reduce noise with less blurring than regular filters.
The document discusses different methods for representing segmented image regions, including:
1) Representing regions based on their external (boundary-based) characteristics or internal (pixel-based) characteristics.
2) Common boundary representation methods are boundary following algorithms, chain codes, and polygon approximation.
3) Chain codes represent boundaries as sequences of line segments coded by direction. Polygon approximation finds the minimum perimeter polygon to capture a boundary shape using the fewest line segments.
Image compression involves reducing the size of image files to reduce storage space and transmission time. There are three main types of redundancy in images: coding redundancy, spatial redundancy between neighboring pixels, and irrelevant information. Common compression methods remove these redundancies, such as Huffman coding, arithmetic coding, LZW coding, and run length coding. Popular image file formats include JPEG for photos, PNG for web images, and TIFF, GIF, and DICOM for other uses.
JPEG is a lossy compression method for color or grayscale images. It works best on continuous-tone images where adjacent pixels have similar colors. The JPEG standard defines several modes of operation and uses various techniques like color space transformation, discrete cosine transformation (DCT), quantization, differential pulse-code modulation, run length encoding, and Huffman coding to achieve high compression ratios while maintaining good image quality. Key aspects of the JPEG process include converting images to luminance and chrominance color space, applying DCT, quantizing coefficients, encoding DC values with DPCM, and entropy coding remaining coefficients.
Digital image processing involves techniques to restore degraded images. Image restoration aims to recover the original undistorted image from a degraded observation. The degradation is typically modeled as the original image being operated on by a degradation function and additive noise. Common restoration techniques include spatial domain filters like mean, median and order-statistic filters to remove noise, and frequency domain filtering to reduce periodic noise. The choice of restoration method depends on the type and characteristics of degradation in the image.
Run-length encoding is a data compression technique that works by eliminating redundant data. It identifies repeating characters or values and replaces them with a code consisting of the character and the number of repeats. This compressed encoded data is then transmitted. At the receiving end, the code is decoded to reconstruct the original data. It is useful for compressing any type of repeating data sequences and is commonly used in image compression by encoding runs of black or white pixels. The compression ratio achieved depends on the amount of repetition in the original uncompressed data.
Here in the ppt a detailed description of Image Enhancement Techniques is given which includes topics like Basic Gray level Transformations,Histogram Processing.
Enhancement using Arithmetic/Logic Operations.
image averaging and image averaging methods.
Piecewise-Linear Transformation Functions
Color fundamentals and color models - Digital Image ProcessingAmna
This presentation is based on Color fundamentals and Color models.
~ Introduction to Colors
~ Color in Image Processing
~ Color Fundamentals
~ Color Models
~ RGB Model
~ CMY Model
~ CMYK Model
~ HSI Model
~ HSI and RGB
~ RGB To HSI
~ HSI To RGB
This document discusses various topics related to computer graphics including anti-aliasing, area sampling, the Koch curve, and the C curve. Anti-aliasing is a technique used to reduce jagged edges by blending pixels. Area sampling is an anti-aliasing method that treats pixels as areas and calculates color based on object overlap. The Koch curve is a fractal curve generated by recursively altering line segments. The C curve replaces lines with two shorter lines at 90 degrees to form triangles at each iteration.
Digital Image Processing denotes the process of digital images with the use of digital computer. Digital images are contains various types of noises which are reduces the quality of images. Noises can be removed by various enhancement techniques. Image smoothing is a key technology of image enhancement, which can remove noise in images.
To highlight the contribution made to the total image appearance by specific bits.i.e. Assuming that each pixel is represented by 8 bits, the image is composed of 8 1-bit planes.Useful for analyzing the relative importance played by each bit of the image.
This slide gives you the basic understanding of digital image compression.
Please Note: This is a class teaching PPT, more and detail topics were covered in the classroom.
Digital Image Processing covers intensity transformations that can be performed on images. These include basic transformations like negatives, log transformations, and power-law transformations. It also discusses image histograms, which measure the frequency of each intensity level in an image. Histogram equalization aims to improve contrast by mapping intensities to produce a uniform histogram. It works by spreading out the most frequent intensity values.
This document provides an overview of a research project on image compression. It discusses image compression techniques including lossy and lossless compression. It describes using discrete wavelet transform, lifting wavelet transform, and stationary wavelet transform for image transformation. Experiments were conducted to compare the compression ratio and processing time of different combinations of wavelet transforms, vector quantization, and Huffman/Arithmetic coding. The results were analyzed to evaluate the compression performance and efficiency of the different methods.
This presentation describes briefly about the image enhancement in spatial domain, basic gray level transformation, histogram processing, enhancement using arithmetic/ logical operation, basics of spatial filtering and local enhancements.
This document discusses spatial filtering techniques in image processing. It begins by defining different types of filters based on the frequencies they preserve, such as low-pass, high-pass, band-pass and band-reject. It then explains that spatial filters require defining a neighborhood/mask and an operation. The document focuses on smoothing/low-pass filters which reduce noise and eliminate small details, and sharpening/high-pass filters which highlight fine details. Common smoothing filters discussed include averaging, Gaussian, and median filtering, while common sharpening filters include unsharp masking, high boost filtering, and filters based on image derivatives like gradient and Laplacian. Examples are provided to illustrate the effects of different filters.
This document summarizes digital image processing techniques including algebraic approaches to image restoration and inverse filtering. It discusses:
1) Unconstrained and constrained restoration, with unconstrained having no knowledge of noise and constrained using knowledge of noise.
2) Inverse filtering which is a direct method that minimizes error between degraded and original images using matrix operations, but can be unstable due to noise or near-zero filter values.
3) Pseudo-inverse filtering which adds a threshold to the inverse filter to avoid instability, working better for noisy images by not amplifying high frequency noise.
This document discusses noise in image processing and various methods for noise removal. It defines noise as unwanted signals that can corrupt an image's quality and originality. Common sources of noise include poor image sensors, lens defects, and low light levels. The document outlines different types of noises like Gaussian noise and impulse noise. It then describes various linear and non-linear filters that can be used for noise removal, such as averaging filters, Gaussian filters, median filters, and Wiener filters. The median filter is effective for salt and pepper noise while preserving edges. Adaptive filters can discriminate between corrupted and clean pixels for better noise removal.
The document discusses noise models and methods for removing additive noise from digital images. It describes several types of noise that can affect images, such as Gaussian, impulse, uniform, Rayleigh, gamma and exponential noise. It also presents various noise filters that can be used to remove noise, including mean filters like arithmetic, geometric and harmonic filters, and order statistics filters such as median, max, min and midpoint filters. The filters aim to reduce noise while retaining image detail as much as possible.
At the end of this lecture students should be able to;
Define the C standard functions for managing file input output.
Apply taught concepts for writing programs.
At the end of this lesson, you should be able to;
identify color formation and how color visualize.
describe primary and secondary colors.
describe display on CRT and LCD.
comprehend RGB, CMY, CMYK and HSI color models.
This document discusses digital image processing and spatial filtering. It begins by explaining that spatial filtering operates on neighborhoods of pixels rather than individual pixels. It then provides examples of simple neighborhood operations like minimum, maximum, and median filters. It also shows how spatial filtering can be expressed as an equation. The document goes on to explain smoothing spatial filters, which average pixel values in a neighborhood. It provides an example of a 3x3 averaging filter and shows how it is applied to each pixel. Finally, it discusses weighted smoothing filters that give more importance to pixels closer to the center.
Spatial filtering using image processingAnuj Arora
(1) Spatial filtering is defined as operations performed on pixels within a neighborhood of an image using a mask or kernel. (2) Filters can be used to blur/smooth an image by reducing noise or sharpen an image by enhancing edges. (3) Common linear filtering methods include averaging, Gaussian, and derivative filters which are implemented using various mask patterns to modify pixels in the filtered image.
This document provides an overview of digital image processing techniques for image restoration. It defines image restoration as improving a degraded image using prior knowledge of the degradation process. The goal is to recover the original image by applying an inverse process to the degradation function. Common degradation sources are discussed, along with noise models like Gaussian, salt and pepper, and periodic noise. Spatial and frequency domain filtering techniques are presented for restoration, such as mean, median and inverse filters. The maximum mean square error or Wiener filter is also introduced as a way to minimize restoration error.
Spatial domain filtering involves modifying an image by applying a filter or kernel to pixels within a neighborhood region. There are two main types of spatial filters - smoothing/low-pass filters which blur an image, and sharpening/high-pass filters which enhance edges and details. Smoothing filters replace each pixel value with the average of neighboring pixels, reducing noise. Sharpening filters use derivatives of Gaussian kernels to highlight areas of rapid intensity change, increasing contrast along edges. The effects of filtering depend on the size and shape of the kernel, with larger kernels producing more blurring or sharpening.
This document provides an overview of image enhancement techniques. It discusses the objectives of image enhancement, which is to process an image to make it more suitable for a specific application or task. The document focuses on spatial domain techniques for image enhancement, specifically point processing methods and histogram processing. It categorizes image enhancement methods into two broad categories: spatial domain methods, which directly manipulate pixel values; and frequency domain methods, which first convert the image into the frequency domain before performing enhancements.
This document provides an overview of various image enhancement techniques. It begins with an introduction to image enhancement and its objectives. It then outlines and describes several categories of enhancement methods, including spatial-frequency domain methods, point operations, histogram operations, spatial operations, and transform operations. Specific techniques discussed in detail include contrast stretching, clipping, thresholding, median filtering, unsharp masking, and principal component analysis for multispectral images. The document also covers color image enhancement and techniques for pseudocoloring.
The document discusses various image enhancement techniques in digital image processing. It describes point operations like image negative, contrast stretching, thresholding, brightness enhancement, log transformation, and power law transformation. Contrast stretching expands the range of intensity levels and can be done by multiplying pixels with a constant, using a transfer function, or histogram equalization. Thresholding converts an image to binary by assigning pixel values above a threshold to one level and below to another. Log and power law transformations compress high intensity values and expand low values to enhance an image. Matlab code examples are provided for each technique.
Spatial filtering involves applying filters or kernels to images to enhance or modify pixel values based on neighboring pixel values. Linear spatial filtering involves taking a weighted sum of pixel values within the filter window. Common filters include averaging filters for noise reduction, median filters to reduce impulse noise while preserving edges, and sharpening filters like Laplacian filters and unsharp masking to enhance details.
Spatial domain image enhancement techniques operate directly on pixel values. Some common techniques include point processing using gray level transformations, mask processing using filters, and histogram processing. Histogram equalization aims to create a uniform distribution of pixel values by mapping the original histogram to a wider range. This improves contrast by distributing pixels more evenly across gray levels.
At the end of this lesson, you should be able to;
define segmentation.
Describe edge based in segmentation.
describe thresholding and its properties.
apply edge detection and thresholding as segmentation techniques.
At the end of this lecture, you should be able to;
describe the importance of morphological features in an image.
describe the operation of erosion, dilation, open and close operations.
identify the practical advantage of the morphological operations.
apply morphological operations for problem solving.
LAPLACE TRANSFORM SUITABILITY FOR IMAGE PROCESSINGPriyanka Rathore
Image processing techniques can involve converting images to digital form and applying transformations like the Laplace transform. The Laplace transform is useful for applications like image sharpening, edge detection, and blob detection. It involves calculating the second derivative of the image to help identify edges and other discontinuities. The zero crossings of the Laplace transform output are particularly useful for edge detection as they indicate where the slope of the image changes most rapidly. While the Laplace transform provides benefits like simpler implementation and reliable noise performance, it can also result in spaghetti-like edge effects with complex computations.
At the end of this lesson, you should be able to;
describe spatial resolution
describe intensity resolution
identify the effect of aliasing
describe image interpolation
describe relationships among the pixels
This document discusses various spatial filtering methods used in image processing. Spatial filters are defined by their neighborhood, which is usually a square window, and their operation, which processes pixels in the neighborhood. Linear filters include correlation and convolution, where the output is a linear combination of input pixels. Common filters are smoothing (low-pass) filters like averaging and Gaussian, which reduce noise and detail, and sharpening (high-pass) filters like unsharp masking and derivatives, which enhance details like edges. Derivatives like the gradient and Laplacian are used to detect edges.
At the end of this lesson, you should be able to;
describe Connected Components and Contours in image segmentation.
discuss region based segmentation method.
discuss Region Growing segmentation technique.
discuss Morphological Watersheds segmentation.
discuss Model Based Segmentation.
discuss Motion Segmentation.
implement connected components, flood fill, watershed, template matching and frame difference techniques.
formulate possible mechanisms to propose segmentation methods to solve problems.
This document provides an agenda and overview of topics related to intensity transformations and spatial filtering for image enhancement. It discusses piecewise-linear transformation functions including contrast stretching, intensity-level slicing, and bit-plane slicing. It also covers histogram processing techniques such as histogram equalization, histogram matching, and using histogram statistics. Finally, it outlines fundamentals of spatial filtering including the mechanics of spatial filtering, spatial correlation and convolution, and generating smoothing and sharpening spatial filters.
When Discrete Optimization Meets Multimedia Security (and Beyond)Shujun Li
This document summarizes research on recovering missing discrete cosine transform (DCT) coefficients in JPEG images. It begins by introducing the problem of missing DCT coefficients that can occur during selective encryption. It then describes early naive approaches and the USO method for recovering the DC coefficient. The document presents an improved method called FRM that minimizes underflow/overflow rates. It proposes a new global optimization model for any missing DCT coefficients and solves it using linear programming. Finally, it discusses limitations of the linear programming approach and pursuing faster algorithms.
This document discusses several topics in image enhancement and processing including:
1. Spatial filtering which involves applying a weighted mask over an image to replace pixel values.
2. Logarithmic transformation which expands darker pixel values more than brighter ones for enhancement.
3. Thresholding, logarithmic transformation, negative transformation, contrast stretching, and grey level slicing as common nonlinear image transformations.
4. Weighted average filtering which applies different coefficients to pixels to give more importance to some over others.
5. High boost filters and unsharp masking as types of high pass sharpening filters used to highlight fine image details.
SpatialEnhancement of course CE7491 of NTUlyumingzhi
This document summarizes key concepts in image intensity transformations and filtering. It discusses two classes of spatial domain processing: point processing and spatial filtering. Point processing involves transformations that modify pixel intensities without regard to neighboring pixels, such as contrast stretching and histogram equalization. Spatial filtering computes new pixel values based on neighboring pixels using techniques like smoothing and sharpening filters. Specific filters covered include averaging, Gaussian, Laplacian, and median filters.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
The automatic license plate recognition(alpr)eSAT Journals
Abstract Every country uses their own way of designing and allocating number plates to their country vehicles. This license number plate is then used by various government offices for their respective regular administrative task like- traffic police tracking the people who are violating the traffic rules, to identify the theft cars, in toll collection and parking allocation management etc. In India all motorized vehicle are assigned unique numbers. These numbers are assigned to the vehicles by district-level Regional Transport Office (RTO). In India the license plates must be kept in both front and back of the vehicle. These plates in general are easily readable by human due to their high level of intelligence on the contrary; it becomes an extremely difficult task for the computers to do the same. Many attributes like illumination, blur, background color, foreground color etc. will pose a problem. Index Terms: Automatic license plate recognition (ALPR) system, proposed methodology, reference
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...CSCJournals
Fractal image compressions of Color Standard Lena and Satellite imageries have been carried out for the variable size range block method. The image is partitioned by considering maximum and minimum size of the range block and transforming the RGB color image into YUV form. Affine transformation and entropy coding are applied to achieve fractal compression. The Matlab simulation has been carried out for three different cases of variable range block sizes. The image is reconstructed using iterative functions and inverse transforms. The results indicate that both color Lena and Satellite imageries with R max = 16 and R min = 8, shows higher Compression ratio (CR) and good Peak Signal to Noise Ratios (PSNR). For the color standard Lena image the achievable CR~13.9 and PSNR ~25.9 dB, for Satellite rural image of CR~ 16 and PSNR ~ 23 and satellite urban image CR~16.4 and PSNR~16.5. The results of the present analysis demonstrate that, for the fractal compression scheme with variable range method applied to both color and gray scale Lena and satellite imageries, show higher CR and PSNR values compared to fixed range block size of 4 and 4 iterations. The results are presented and discussed in the paper.
This document describes a project to design a ball catching manipulator using a 2 degree of freedom robotic arm and webcam vision. The project involves selecting a 2 DOF manipulator, developing the inverse kinematics, generating trajectories to catch the ball using projectile motion equations, detecting the ball using image processing in the webcam footage, and transferring coordinate frames. Future work includes implementing the simulation with a Dynamixel robotic arm and using a Kinect for 3D detection and depth sensing.
The document discusses image segmentation techniques including thresholding. Thresholding divides an image into foreground and background regions based on pixel intensity values. Global thresholding uses a single threshold value for the entire image, while adaptive or local thresholding uses variable thresholds that change across the image. Multilevel thresholding can extract objects within a specific intensity range using multiple threshold values. The Hough transform is also presented as a way to connect disjointed edge points and detect shapes like lines in an image.
This document discusses various spatial filtering and image enhancement techniques including intensity transformation, smoothing filters, sharpening filters, and combining multiple techniques. It covers linear and non-linear spatial filters, smoothing filters like averaging and median filters for noise reduction, and sharpening filters such as unsharp masking, high-boost filtering, and gradient-based methods. Examples are provided to demonstrate the use of various filters for tasks like noise removal, edge enhancement, and combining techniques for improved image quality.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Environment Detection and Path Planning Using the E-puck Robot IRJET Journal
This document describes a path planning project using an e-puck mobile robot. The key steps are:
1. The environment is detected through a camera mounted above the space. Image segmentation is used to identify boundaries, obstacles, and the robot's starting position.
2. Post-processing is done on the segmented image to smooth edges before detecting obstacle corners, which will serve as nodes in a roadmap.
3. A visibility graph roadmap representing all possible paths from start to goal is generated.
4. An A* search algorithm is implemented to find the optimal path through the roadmap, avoiding obstacles.
5. The robot is programmed to follow the optimal path from its starting position
The automatic license plate recognition(alpr)eSAT Journals
Abstract Every country uses their own way of designing and allocating number plates to their country vehicles. This license number plate is then used by various government offices for their respective regular administrative task like- traffic police tracking the people who are violating the traffic rules, to identify the theft cars, in toll collection and parking allocation management etc. In India all motorized vehicle are assigned unique numbers. These numbers are assigned to the vehicles by district-level Regional Transport Office (RTO). In India the license plates must be kept in both front and back of the vehicle. These plates in general are easily readable by human due to their high level of intelligence on the contrary; it becomes an extremely difficult task for the computers to do the same. Many attributes like illumination, blur, background color, foreground color etc. will pose a problem. Index Terms: Automatic license plate recognition (ALPR) system, proposed methodology, reference
Similar to COM2304: Intensity Transformation and Spatial Filtering – III Spatial Filters for Sharpening (20)
This document discusses parsing with context-free grammars. It begins by introducing context-free grammars and their use in parsing sentences. It then discusses parsing as a search problem, and presents top-down and bottom-up parsing algorithms. Top-down parsing builds trees from the root node down, while bottom-up parsing builds trees from the leaves up. Both approaches have advantages and disadvantages related to efficiency. The document also introduces probabilistic context-free grammars, which augment grammars with rule probabilities, and discusses how these can be used for disambiguation.
The document discusses context-free grammars for modeling English syntax. It introduces key concepts like constituency, grammatical relations, and subcategorization. Context-free grammars use rules and symbols to generate sentences. They consist of terminal symbols (words), non-terminal symbols (phrases), and rules to expand non-terminals. Context-free grammars can model syntactic knowledge and generate sentences in both a top-down and bottom-up manner through parsing.
Parts-of-speech can be divided into closed classes and open classes. Closed classes have a fixed set of members like prepositions, while open classes like nouns and verbs are continually changing with new words being created. Parts-of-speech tagging is the process of assigning a part-of-speech tag to each word using statistical models trained on tagged corpora. Hidden Markov Models are commonly used, where the goal is to find the most probable tag sequence given an input word sequence.
This document provides an overview of Markov models and hidden Markov models (HMMs). It begins by introducing Markov chains, which are probabilistic state machines where the probability of transitioning to the next state depends only on the current state. Hidden Markov models extend Markov chains by adding hidden states that are not directly observable. The key aspects of HMMs are defined, including the hidden states, observed outputs, transition probabilities, and output probabilities. The document then discusses how to compute the likelihood of an observed sequence given an HMM, including using the forward algorithm to efficiently sum over all possible state sequences. Overall, the document provides a conceptual introduction to Markov models and HMMs, focusing on their structure, assumptions, and the forward algorithm
The document discusses various techniques for smoothing N-gram language models, including Laplace smoothing, Good-Turing discounting, and backoff models. Laplace smoothing involves adding one to all counts to address unseen events. Good-Turing smoothing estimates probabilities for low counts based on the frequency of higher counts. Backoff models first use the highest available order N-gram model and only fallback to a lower order if the current count is zero.
This document discusses methods for evaluating language models, including intrinsic and extrinsic evaluation. Intrinsic evaluation involves measuring a model's performance on a test set using metrics like perplexity, which is based on how well the model predicts the test set. Extrinsic evaluation embeds the model in an application and measures the application's performance. The document also covers techniques for dealing with unknown words like replacing low-frequency words with <UNK> and estimating its probability from training data.
The document discusses N-gram language models, which assign probabilities to sequences of words. An N-gram is a sequence of N words, such as a bigram (two words) or trigram (three words). The N-gram model approximates the probability of a word given its history as the probability given the previous N-1 words. This is called the Markov assumption. Maximum likelihood estimation is used to estimate N-gram probabilities from word counts in a corpus.
The document discusses minimum edit distance and how it can be used to quantify the similarity between two strings. Minimum edit distance is defined as the minimum number of editing operations like insertion, deletion, substitution needed to transform one string into another. Levenshtein distance assigns a cost of 1 to each insertion, deletion, or substitution, and calculates the minimum edits between two strings using dynamic programming to build up solutions from sub-problems. The algorithm can also be modified to produce an alignment between the strings by storing back pointers and doing a backtrace.
Finite-state morphological parsing uses finite-state transducers to parse words into their morphological components like stems and affixes. It requires a lexicon of stems and affixes, morphotactic rules describing valid morpheme combinations, and orthographic rules for spelling changes. The parser is built as a cascade of finite-state automata representing the lexicon, morphotactics and spelling rules. It maps surface word forms onto their underlying lexical representations including stems and morphological features. This allows morphological analysis of both regular and irregular forms.
The document discusses English morphology, including singular and plural forms, morphological parsing and stemming, and the different types of morphology in English such as inflectional, derivational, compounding, and cliticization morphology. It provides examples to illustrate singular and plural forms, morphological rules for changing word endings, and how prefixes, suffixes, and other morphemes can be added to word stems to change word classes or derive new words.
The document discusses text normalization, which involves segmenting and standardizing text for natural language processing. It describes tokenizing text into words and sentences, lemmatizing words into their root forms, and standardizing formats. Tokenization involves separating punctuation, normalizing word formats, and segmenting sentences. Lemmatization determines that words have the same root despite surface differences. Sentence segmentation identifies sentence boundaries, which can be ambiguous without context. Overall, text normalization prepares raw text for further natural language analysis.
The document discusses finite state automata (FSA) and regular expressions. It provides examples of deterministic and non-deterministic FSA that recognize strings containing combinations of words. Non-deterministic FSA can have multiple possible state transitions for a given input, requiring strategies like backing up, look-ahead, or parallel processing to determine if a string is accepted. FSA and regular expressions can define regular languages by generating all matching strings.
This document discusses various techniques for text normalization and pattern matching using regular expressions. It covers tokenization, lemmatization, edit distance, and basic regular expression patterns including character classes, quantifiers, anchors, precedence, and disjunction. The key goals of text normalization are to standardize text into a convenient form for processing, while regular expressions provide a language for specifying text search patterns.
This document provides an introduction to natural language processing and the knowledge and techniques required. It discusses:
1) The goals of natural language processing including enabling human-machine communication, improving human-human communication, and processing text/speech.
2) The different types of linguistic knowledge needed such as phonetics, phonology, morphology, syntax, semantics, pragmatics, and discourse.
3) Common models used like state machines, rule systems, logic, probabilistic models, and vector spaces and related algorithms like search algorithms, machine learning algorithms, and dynamic programming.
At the end of this lecture students should be able to;
Define the declaration C strings.
Compare fixed length and variable length string.
Apply strings for functions.
Define string handling functions.
Apply taught concepts for writing programs.
At the end of this lecture students should be able to;
Define, initialize and access to the C stuctuers.
Develop programs using structures in arrays and functions.
Use structures within structures and structures as pointers.
Define, initialize and access to the C unions.
Compare and contrast C structures and unions.
Define memory allocation and de-allocation methods in C.
Develop programs using memory allocation functions.
Apply taught concepts for writing programs.
At the end of this lecture students should be able to;
Define the C standard functions for managing input output.
Apply taught concepts for writing programs.
Pointer variables allow programmers to indirectly access and manipulate the memory addresses where variables are stored. Pointers must be declared with a data type and initialized by assigning the address of an existing variable using the address-of operator (&). Pointer variables can then be used to read from and write to the memory location of the variable being pointed to using indirection (*). Pointers enable operations like traversing arrays, passing arguments by reference, and dynamically allocating memory. Key pointer concepts covered include declaration, initialization, dereferencing, arithmetic, comparisons, NULL pointers, and their usage with arrays.
At the end of this lecture students should be able to;
Describe the C arrays.
Practice the declaration, initialization and access linear arrays.
Practice the declaration, initialization and access two dimensional arrays.
Apply taught concepts for writing programs.
At the end of this lecture students should be able to;
Describe the looping structures in C programming language.
Practice the control flow of different looping structures in C programming language.
Practice the variants in control flow of different looping structures in C programming language.
Apply taught concepts for writing programs.
ESR spectroscopy in liquid food and beverages.pptxPRIYANKA PATEL
With increasing population, people need to rely on packaged food stuffs. Packaging of food materials requires the preservation of food. There are various methods for the treatment of food to preserve them and irradiation treatment of food is one of them. It is the most common and the most harmless method for the food preservation as it does not alter the necessary micronutrients of food materials. Although irradiated food doesn’t cause any harm to the human health but still the quality assessment of food is required to provide consumers with necessary information about the food. ESR spectroscopy is the most sophisticated way to investigate the quality of the food and the free radicals induced during the processing of the food. ESR spin trapping technique is useful for the detection of highly unstable radicals in the food. The antioxidant capability of liquid food and beverages in mainly performed by spin trapping technique.
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
ESPP presentation to EU Waste Water Network, 4th June 2024 “EU policies driving nutrient removal and recycling
and the revised UWWTD (Urban Waste Water Treatment Directive)”
Nucleophilic Addition of carbonyl compounds.pptxSSR02
Nucleophilic addition is the most important reaction of carbonyls. Not just aldehydes and ketones, but also carboxylic acid derivatives in general.
Carbonyls undergo addition reactions with a large range of nucleophiles.
Comparing the relative basicity of the nucleophile and the product is extremely helpful in determining how reversible the addition reaction is. Reactions with Grignards and hydrides are irreversible. Reactions with weak bases like halides and carboxylates generally don’t happen.
Electronic effects (inductive effects, electron donation) have a large impact on reactivity.
Large groups adjacent to the carbonyl will slow the rate of reaction.
Neutral nucleophiles can also add to carbonyls, although their additions are generally slower and more reversible. Acid catalysis is sometimes employed to increase the rate of addition.
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...AbdullaAlAsif1
The pygmy halfbeak Dermogenys colletei, is known for its viviparous nature, this presents an intriguing case of relatively low fecundity, raising questions about potential compensatory reproductive strategies employed by this species. Our study delves into the examination of fecundity and the Gonadosomatic Index (GSI) in the Pygmy Halfbeak, D. colletei (Meisner, 2001), an intriguing viviparous fish indigenous to Sarawak, Borneo. We hypothesize that the Pygmy halfbeak, D. colletei, may exhibit unique reproductive adaptations to offset its low fecundity, thus enhancing its survival and fitness. To address this, we conducted a comprehensive study utilizing 28 mature female specimens of D. colletei, carefully measuring fecundity and GSI to shed light on the reproductive adaptations of this species. Our findings reveal that D. colletei indeed exhibits low fecundity, with a mean of 16.76 ± 2.01, and a mean GSI of 12.83 ± 1.27, providing crucial insights into the reproductive mechanisms at play in this species. These results underscore the existence of unique reproductive strategies in D. colletei, enabling its adaptation and persistence in Borneo's diverse aquatic ecosystems, and call for further ecological research to elucidate these mechanisms. This study lends to a better understanding of viviparous fish in Borneo and contributes to the broader field of aquatic ecology, enhancing our knowledge of species adaptations to unique ecological challenges.
When I was asked to give a companion lecture in support of ‘The Philosophy of Science’ (https://shorturl.at/4pUXz) I decided not to walk through the detail of the many methodologies in order of use. Instead, I chose to employ a long standing, and ongoing, scientific development as an exemplar. And so, I chose the ever evolving story of Thermodynamics as a scientific investigation at its best.
Conducted over a period of >200 years, Thermodynamics R&D, and application, benefitted from the highest levels of professionalism, collaboration, and technical thoroughness. New layers of application, methodology, and practice were made possible by the progressive advance of technology. In turn, this has seen measurement and modelling accuracy continually improved at a micro and macro level.
Perhaps most importantly, Thermodynamics rapidly became a primary tool in the advance of applied science/engineering/technology, spanning micro-tech, to aerospace and cosmology. I can think of no better a story to illustrate the breadth of scientific methodologies and applications at their best.
BREEDING METHODS FOR DISEASE RESISTANCE.pptxRASHMI M G
Plant breeding for disease resistance is a strategy to reduce crop losses caused by disease. Plants have an innate immune system that allows them to recognize pathogens and provide resistance. However, breeding for long-lasting resistance often involves combining multiple resistance genes
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...University of Maribor
Slides from talk:
Aleš Zamuda: Remote Sensing and Computational, Evolutionary, Supercomputing, and Intelligent Systems.
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Inter-Society Networking Panel GRSS/MTT-S/CIS Panel Session: Promoting Connection and Cooperation
https://www.etran.rs/2024/en/home-english/
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
The binding of cosmological structures by massless topological defects
COM2304: Intensity Transformation and Spatial Filtering – III Spatial Filters for Sharpening
1. COMPUTER GRAPHICS & IMAGE PROCESSING
COM2403
Intensity Transformation and Spatial Filtering – III
Spatial Filters for Sharpening
K.A.S.H.Kulathilake
B.Sc. (Hons) IT (SLIIT), MCS (UCSC), M.Phil (UOM), SEDA(UK)
Rajarata University of Sri Lanka
Faculty of Applied Sciences
Department of Physical Sciences
2. Learning Outcomes
COM2304 - Computer Graphics & Image
Processing
• At the end of this lecture, you should be
able to;
– describe sharpening through spatial filters.
– Identify usage of derivatives in Image
Processing.
– discuss edge detection techniques.
– compare 1st & 2nd order derivatives used for
sharpening.
– Apply sharpening techniques for problem
solving.
2
3. Sharpening Spatial Filters
• The principle objective of sharpening is to highlight
transitions in intensity.
• In other terms, sharpening spatial filters seek to highlight
fine detail
– Remove blurring from images
– Highlight edges
• Sharpening filters are based on spatial differentiation.
• The strength of the response of a derivative operator is
proportional to the degree of intensity discontinuity of
the image at the point at which the operator is applied.
• Hence, image differentiation enhances edges and other
discontinuities ( such as noise) and deemphasizes areas
with slowly varying intensities.
COM2304 - Computer Graphics & Image
Processing
3
4. Sharpening Spatial Filters (Cont…)
• As the initial approach, it is better to get an idea about
certain properties of first order and second order
derivatives in the digital context.
• To simplify the explanation, it has been focused
attention on one dimensional derivatives.
• We are interested in the behavior of these derivatives
in areas of
– Constant intensity,
– At the onset and end of discontinuities (step and ramp
discontinuities), and
– Along the intensity ramps.
COM2304 - Computer Graphics & Image
Processing
4
5. Sharpening Spatial Filters (Cont…)
• Differentiation measures the rate of change of
a function.
• In digital context, values are finite and the
maximum possible intensity change is also
finite.
• The shortest distance over which that change
can occur is between the adjacent pixels.
• To understand this let’s discuss an example.
COM2304 - Computer Graphics & Image
Processing
5
6. Sharpening Spatial Filters (Cont…)
A B
Intensity transition
6
COM2304 - Computer Graphics & Image
Processing
7. Sharpening Spatial Filters (Cont…)
• The formula for the 1st derivative of a function is as
follows:
• It is just the difference between subsequent values
and measures the rate of change of the function.
• It should satisfy following conditions:
– Must be zero in the areas of constant intensity.
– Must be nonzero at the onset of an intensity step or
ramp.
– Must be nonzero along the ramp.
COM2304 - Computer Graphics & Image
Processing
7
)()1( xfxf
x
f
9. Sharpening Spatial Filters (Cont…)
• The formula for the 2nd derivative of a function is as follows:
• Simply takes into account the values both before and after the
current value.
• It should satisfy:
– Must be zero in the areas of constant intensity.
– Must be nonzero at the onset and end of an intensity
step or ramp.
– Must be zero along the ramp of constant slope.
COM2304 - Computer Graphics & Image
Processing
9
)(2)1()1(2
2
xfxfxf
x
f
11. Sharpening Spatial Filters (Cont…)
• The 2nd derivative is more useful for image
enhancement than the 1st derivative because;
– Stronger response to fine detail
– Simpler implementation
COM2304 - Computer Graphics & Image
Processing
11
12. The Laplacian Operation
• In this section, we discuss 2D, second order
derivatives used for image sharpening.
• It is an isotropic filter
– It means that the response is independent of the
direction of the discontinuities in the image to
which the filter is applied. / rotation invariant.
COM2304 - Computer Graphics & Image
Processing
12
13. The Laplacian Operation (Cont…)
The Laplacian is defined as follows:
where the partial 2nd order derivative in the x direction
is defined as follows:
and in the y direction as follows:
y
f
x
f
f 2
2
2
2
2
),(2),1(),1(2
2
yxfyxfyxf
x
f
),(2)1,()1,(2
2
yxfyxfyxf
y
f
13
COM2304 - Computer Graphics & Image
Processing
14. The Laplacian Operation (Cont…)
So, the Laplacian can be given as follows:
We can easily build a filter based on this
),1(),1([2
yxfyxff
)]1,()1,( yxfyxf
),(4 yxf
0 1 0
1 -4 1
0 1 0
This provides an isotropic
results for rotations in
increments of 90 degrees
0 -1 0
-1 4 -1
0 -1 0
14
COM2304 - Computer Graphics & Image
Processing
15. The Laplacian Operation (Cont…)
• We can incorporate diagonal directions to
form a laplacian filter;
COM2304 - Computer Graphics & Image
Processing
15
),1(),1([2
yxfyxff )]1,()1,( yxfyxf
),(8 yxf
)]1,1()1,1( yxfyxf)1,1()1,1([ yxfyxf
1 1 1
1 -8 1
1 1 1
-1 -1 -1
-1 8 -1
-1 -1 -1
This provides an isotropic
results for rotations in
increments of 45 degrees
16. The Laplacian Operation (Cont…)
Applying the Laplacian to an image we get a new image
that highlights edges and other discontinuities
Original
Image
Laplacian
Filtered Image
Laplacian
Filtered Image
Scaled for Display
16
COM2304 - Computer Graphics & Image
Processing
17. The Laplacian Operation (Cont…)
The result of a Laplacian filtering is
not an enhanced image
We have to do more work in order
to get our final image
Subtract the Laplacian result from
the original image to generate our
final sharpened enhanced image Laplacian
Filtered Image
Scaled for Display
fyxfyxg 2
),(),(
17
COM2304 - Computer Graphics & Image
Processing
18. The Laplacian Operation (Cont…)
Laplacian Image Enhancement: In the final
sharpened image edges and fine detail are much
more obvious
- =
Original
Image
Laplacian
Filtered Image
Sharpened
Image
18
COM2304 - Computer Graphics & Image
Processing
20. Image Gradient & Properties
• First derivatives in image processing are implemented using
the magnitude of the gradient.
• Gradient is a tool of choice for finding edge strength and
direction at location (x, y) of an image f and it is defined as a
vector.
• For a function f(x, y) the gradient of f at coordinates (x, y) is
given as the column vector:
COM2304 - Computer Graphics & Image
Processing
20
y
f
x
f
G
G
y
x
f
This vector has important
geometrical property that it
points in the direction of the
greatest rate of change of f at
location (x, y).
21. Image Gradient & Properties (Cont…)
• Gradient Operators:
– Obtaining the gradient of an image requires
computing the partial derivatives x and y at every
pixel location in the image.
– Therefore, digital approximation of the partial
derivatives over a neighborhood about a point is
required.
COM2304 - Computer Graphics & Image
Processing
21
),()1,(
),(
),(),1(
),(
yxfyxf
y
yxf
G
yxfyxf
x
yxf
G
y
x
22. Image Gradient & Properties (Cont…)
• There are various gradient operators available
namely; ID mask, Robert Cross, Prewitt, Sobel
• I-D Mask:
– Above equations can be implemented for all
pertinent values of x and y by filtering f(x,y) with
the 1-D masks.
COM2304 - Computer Graphics & Image
Processing
22
-1
1
-1 1
23. Image Gradient & Properties (Cont…)
• Roberts Cross-Gradient Operators:
• When diagonal edge direction is of interest, we need a 2
Dimensional mask.
• This operator cannot compute the edge direction
COM2304 - Computer Graphics & Image
Processing
23
)(
),(
)(
),(
68
59
zz
y
yxf
G
zz
x
yxf
G
y
x
z1 z2 z3
z4 z5 z6
z7 z8 z9
24. Image Gradient & Properties (Cont…)
• Prewitt Gradient Operators:
COM2304 - Computer Graphics & Image
Processing
24
)()(
),(
)()(
),(
741963
321987
zzzzzz
y
yxf
G
zzzzzz
x
yxf
G
y
x
z1 z2 z3
z4 z5 z6
z7 z8 z9
25. Image Gradient & Properties (Cont…)
• Sobel Gradient Operators:
– Using 2 in the center location provides image
smoothing; hence, it has better noise suppression.
COM2304 - Computer Graphics & Image
Processing
25
)2()2(
),(
)2()2(
),(
741963
321987
zzzzzz
y
yxf
G
zzzzzz
x
yxf
G
y
x
z1 z2 z3
z4 z5 z6
z7 z8 z9
27. Image Gradient & Properties (Cont…)
• Note that the coefficients of all the masks
sum to zero, thus giving a response of zero in
areas of constant intensity, as expected of a
derivative operator.
• To filter an image it is filtered using both
operators the results of which are added
together
• All the masks discussed above are used to
determine Gx, and Gy.
COM2304 - Computer Graphics & Image
Processing
27
28. Image Gradient & Properties (Cont…)
• Prewitt and Sobel masks give isotropic results
only for vertical and horizontal edges.
• It is possible to modify the 3*3 masks, so that
they have their strongest responses along the
diagonal directions as shown below;
COM2304 - Computer Graphics & Image
Processing
28
0 1 1
-1 0 1
-1 -1 0
-1 -1 0
-1 0 1
0 1 1
Modified Prewitt
masks
29. How Image Gradient Works?
• How Sobel works?
– The Sobel edge detector is a gradient based
method.
– It works with first order derivatives.
– It calculates the first derivatives of the image
separately for the X and Y axes.
– The derivatives are only approximations (because
the images are not continuous).
COM2304 - Computer Graphics & Image
Processing
29
30. How Image Gradient Works? (Cont…)
– To approximate them, the following kernels are used for
convolution:
– The kernel on the left approximates the derivative along the X
axis. The one on the right is for the Y axis.
– Using this information, we can calculate the following:
• Magnitude or "strength" of the edge:
• Approximate strength:
• The orientation of the edge:
COM2304 - Computer Graphics & Image
Processing
30
31. How Image Gradient Works? (Cont…)
The magnitude of this vector is given by:
For practical reasons this can be simplified as:
)f( magf
2
1
22
yx GG
2
1
22
y
f
x
f
yx GGf
31
COM2304 - Computer Graphics & Image
Processing
It is the value of the rate
of change in the
direction of the gradient
vector.
32. How Image Gradient Works? (Cont…)
• The direction of the gradient vector is given by
the angle:
• It is measured with respect to the x- axis.
• The direction of an edge at an arbitrary point
(x, y) is orthogonal to the direction, ᾳ(x, y) of
the gradient vector at the point.
COM2304 - Computer Graphics & Image
Processing
32
x
y
g
g
yx nta),(
33. How Image Gradient Works? (Cont…)
COM2304 - Computer Graphics & Image
Processing
33
Figure contains zoomed
straight edge segment.
Each square shown here
corresponds to a pixel.
We want to find strength
and direction of the edge
at the point highlighted
with a box?
Example
34. How Image Gradient Works? (Cont…)
• Method:
– We need to compute the derivatives of x and y
directions using a 3* 3 neighborhood.
– To get the partial derivatives in the x direction
subtract the pixels in the top row of the
neighborhood from the pixels in the bottom row.
– To get the partial derivatives in the y direction
subtract the pixels in the left column of the
neighborhood from the pixels in the right column.
– Suppose
COM2304 - Computer Graphics & Image
Processing
34
2/
2/
yf
xf
35. How Image Gradient Works? (Cont…)
• M(x,y) at that point
equals to 2+2
• Similarly, the direction
of gradient vector at
the same point equals
to:
• It is same as 1350
measured in positive
direction with respect
to the x-axis.
COM2304 - Computer Graphics & Image
Processing
35
0
45tan),(
x
y
G
G
yx
https://www.trigonometrytable.com/tan-inverse.php
36. How Image Gradient Works? (Cont…)
• Edge at a point is
orthogonal to the gradient
vector at that point.
• So the direction angle of
the edge in this example is
• All edge point in the figure
have the same gradient,
so that entire edge
segment is in the same
direction.
COM2304 - Computer Graphics & Image
Processing
36
00
4590
38. How Image Gradient Works? (Cont…)
Sobel filters are typically used for edge detection
An image of a contact
lens which is
enhanced in order to
make defects (at four
and five o’clock in the
image) more obvious
38
COM2304 - Computer Graphics & Image
Processing
39. Canny Edge Detection
• The Canny edge detector is a good approximation of the optimal
operator, i.e., the one that maximizes the product of signal-to-noise
ratio and localization.
• Basically, the Canny edge detector is the first derivative of a
Gaussian function.
• The algorithm runs in 5 separate steps:
– Smoothing: Blurring of the image to remove noise.
– Finding gradients: The edges should be marked where the gradients of
the image has large magnitudes.
– Non-maximum suppression: Only local maxima should be marked as
edges.
– Double thresholding: Potential edges are determined by thresholding.
– Edge tracking by hysteresis: Final edges are determined by suppressing
all edges that are not connected to a very certain (strong) edge.
COM2304 - Computer Graphics & Image
Processing
39
Refer Note
42. 1st & 2nd Derivatives
Comparing the 1st and 2nd derivatives we can
conclude the following:
– 1st order derivatives generally produce thicker
edges
– 2nd order derivatives have a stronger response to
fine detail e.g. thin lines
– 1st order derivatives have stronger response to
grey level step
– 2nd order derivatives produce a double response
at step changes in grey level
42
COM2304 - Computer Graphics & Image
Processing
43. Combining Spatial Enhancement
Methods
Successful image enhancement
is typically not achieved using a
single operation
Rather we combine a range of
techniques in order to achieve a
final result
This example will focus on
enhancing the bone scan to the
right
43
COM2304 - Computer Graphics & Image
Processing
44. Combining Spatial Enhancement
Methods (Cont…)
Laplacian filter of
bone scan (a)
Sharpened version of
bone scan achieved by
subtracting (a) and (b)
Sobel filter of bone
scan (a)
(a)
(b)
(c)
(d) 44
COM2304 - Computer Graphics & Image
Processing
45. Combining Spatial Enhancement
Methods (Cont…)
The product of (c) and
(e) which will be used
as a mask
Sharpened image
which is sum of (a)
and (f)
Result of applying a
power-law trans. to
(g)
(e)
(f)
(g)
(h)
Image (d) smoothed with a
5*5 averaging filter
45
COM2304 - Computer Graphics & Image
Processing
48. Learning Outcomes Revisit
• Now, you should be able to;
– describe sharpening through spatial filters.
– Identify usage of derivatives in Image
Processing.
– discuss edge detection techniques.
– compare 1st & 2nd order derivatives used for
sharpening.
– Apply sharpening techniques for problem
solving.
COM2304 - Computer Graphics & Image
Processing
48