Articulo
Journal of Computer Science & Technology; vol. 5, no. 3
The handwritten manuscript recognizing process belongs to a set of initiatives which lean to the preservation of cultural patrimony gathered in libraries and archives, where there exist a great wealth in documents and even handwritten cards that accompany incunabula books. This work is the starting point of a research and development project oriented to digitalization and recognition of manuscript materials. The paper presented here discuss different algorithms used in the first stage dedicated to "image noise-cleaning" in order to improve it before the character recognition process begins. In order to make the handwritten-text recognition and image digitalization process efficient, it must be preceded by a preprocessing stage of the image to be treated, which includes thresholding, noise cleaning, thinning, base-line alignment and image segmentation, among others. Each of these steps will allow us to reduce the injurious variability when recognizing manuscripts (noise, random gray levels, slanted characters, ink level in different zones), and so increasing the probability of obtaining a suitable text recognition. In this paper, two image thinning methods are considered, and implemented. Finally, an evaluation is carried out obtaining many conclusions related to efficiency, speed and requirements, as well as ideas for future implementations.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5521
Around the World In 80 Minutes Quiz Part II - 2013parthiben1
The document provides information about a quiz game that asks questions about locations around the world. It begins with rules for the game, stating there are 41 questions starting in India and traveling around the world. It notes that some questions involve identifying locations from photographs taken by an environmental artist. The questions then begin, asking the participant to identify places, countries, landmarks, and other details about locations globally.
The document discusses kindness to animals in Islam based on teachings from the Quran and stories of the Prophet Muhammad. It provides examples of how Islam advocates for treating all living creatures, no matter how small, with compassion and preventing cruelty. Specific teachings mentioned include not causing unnecessary pain, avoiding emotional and mental cruelty, providing food and water, not overburdening or imprisoning animals, and being gentle even in consumption of meat.
Around the World In 80 Minutes Quiz - 2013parthiben1
Auroville, near Pondicherry, is an "experimental" township in Viluppuram district, Tamil Nadu, created in 1968 based on the vision of Mirra Alfassa to create a place for all human beings of good will to live freely as citizens of the world. Barren Island, located in the Andaman Sea, is the only active volcano in South Asia and attracts tourists for this reason. Christmas Island was named after the day it was discovered and is home to the annual migration of millions of terrestrial red crabs.
Objeto de conferencia
III International Conference on New Horizons in Education (INTE) (Praga, República Checa)
This work presents an open source web environment to learn GPSS language in Modeling and Simulation courses. With this environment, students build their models by selecting entities and configuring them instead of programming GPSS codes from scratch. Teachers can also create models so that students can apply, analyze and interpret results. Thus, it includes a simulation engine that stores snapshots of models as they are executed, and allows students to navigate through these snapshots. The environment may be combined with existing learning management systems.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/25674
Around the World In 80 Minutes Quiz Part II - 2013parthiben1
The document provides information about a quiz game that asks questions about locations around the world. It begins with rules for the game, stating there are 41 questions starting in India and traveling around the world. It notes that some questions involve identifying locations from photographs taken by an environmental artist. The questions then begin, asking the participant to identify places, countries, landmarks, and other details about locations globally.
The document discusses kindness to animals in Islam based on teachings from the Quran and stories of the Prophet Muhammad. It provides examples of how Islam advocates for treating all living creatures, no matter how small, with compassion and preventing cruelty. Specific teachings mentioned include not causing unnecessary pain, avoiding emotional and mental cruelty, providing food and water, not overburdening or imprisoning animals, and being gentle even in consumption of meat.
Around the World In 80 Minutes Quiz - 2013parthiben1
Auroville, near Pondicherry, is an "experimental" township in Viluppuram district, Tamil Nadu, created in 1968 based on the vision of Mirra Alfassa to create a place for all human beings of good will to live freely as citizens of the world. Barren Island, located in the Andaman Sea, is the only active volcano in South Asia and attracts tourists for this reason. Christmas Island was named after the day it was discovered and is home to the annual migration of millions of terrestrial red crabs.
Objeto de conferencia
III International Conference on New Horizons in Education (INTE) (Praga, República Checa)
This work presents an open source web environment to learn GPSS language in Modeling and Simulation courses. With this environment, students build their models by selecting entities and configuring them instead of programming GPSS codes from scratch. Teachers can also create models so that students can apply, analyze and interpret results. Thus, it includes a simulation engine that stores snapshots of models as they are executed, and allows students to navigate through these snapshots. The environment may be combined with existing learning management systems.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/25674
The document provides guidance for students conducting research for a social sciences fair. It instructs students to select a topic and develop 3 research questions not in the form of "Do you like" or "Do you know". Students need at least one book and one internet source for their research. They are shown how to search the school catalog and databases like NetTrekker and Britannica Elementary to find relevant resources. The document emphasizes properly citing all sources used.
This document discusses a solar powered aircraft that recently completed its first cross-country flight across the US. It has the wingspan of a Boeing 747 but weighs as much as a Chevy Volt, and is powered by 10 horsepower electric motors and nearly 12,000 solar cells. The aircraft aims to promote renewable energy technologies and plans to fly around the world in 2015.
The document contains a Tamil language quote that translates to "Water today is not common to anyone in the world." It then states that approximately 1 in 8 people, or nearly 900 million people, lack access to clean water globally.
1) The document describes an eco quiz being held on June 5th, 2014 in Chennai, India. It provides the rules for the clockwise and anti-clockwise rounds, and lists 20 sample questions from the quiz.
2) The questions cover a range of topics including wildlife, the environment, and famous personalities known for conservation efforts.
3) Some of the questions identify famous tigers like Machli and provide details about their significance and conservation work.
Francis Ford Coppola was adapted from a novella and set during the Vietnam War.
The movies being referred to are:
1. Catch-22 directed by Mike Nichols
2. Apocalypse Now directed by Francis Ford Coppola
24)
This movie directed by David Fincher was based on a novel by
Patricia Highsmith. It had many twists and turns in its plotline
keeping the audience guessing throughout.
The movie had Oscar nominated performances by Gwyneth
Paltrow and Jeremy Northam in lead roles.
Identify the movie.
The Talented Mr. Ripley
25)
Around The World In 80 Minutes Quiz Part III - Islands Edition 2014parthiben1
- The document provides rules and instructions for a quiz game being played, including that there will be 24 questions in each clockwise and anti-clockwise round, questions are open to everyone, and the quiz master's decision is final.
- It explains the "pounce" and "bounce" rules where players can earn points for correct answers or lose points for incorrect answers during pounces, and get the next question if answering correctly during bounces.
- The clockwise round section indicates there will be 24 questions asked in that round.
This document outlines the rules and structure for a quiz competition. It will consist of two written rounds with 15 questions each, as well as clockwise and anti-clockwise verbal rounds also with 15 questions each. The quiz master's decision is final. Sample questions are provided about famous paintings, current events, geography, and other topics. Correct answers are also included.
The document discusses various details about Tamil cinema directors, actors, actresses and movies. It asks questions about the names associated with the details provided and asks the reader to identify them. Some key details discussed include the director who said he would have joined the LTTE if born in Tamil Eelam, the top heroine in the 80s who refused a role but now assists all of Maniratnam's movies, and the only actor to have his 100th film be a commercial hit.
1. The document discusses Vaali, a Tamil lyricist who sometimes acted in films. It notes that he often depicted critical moments realistically and shot most of his films within India rather than foreign locales. His female leads were typically naughty but good with hints of sophistication.
2. It describes a film from the British period involving railway promotion. Some Indians converted to Christianity to get promotions but reconverted after retirement. The English associated this group with deception for a long time.
3. It asks the reader to identify the name of a 1990s Tamil film with the protagonist's name prefixed by the word for this group, alluding to the film X-Y.
1) The document outlines the rules for an India Quiz, including details about different rounds, scoring, and topics that will be covered such as Indian cities, films, and current events.
2) It provides examples of questions that may be asked, such as identifying the inspiration for a song lyricist's work or the meaning behind a logo.
3) Samples of potential written questions are included, such as identifying monuments depicted in illustrations of Indian cities.
In imaging science, image processing is processing of images using mathematical operations by using any form of signal processing for which the input is an image, a series of images, or a video, such as a photograph or video frame; the output of image processing may be either an image or a set of characteristics or parameters related to the image. Most image-processing techniques involve treating the image as a two-dimensional signal and applying standard signal-processing techniques to it. Images are also processed as three-dimensional signals where the third-dimension being time or the z-axis.
Image processing usually refers to digital image processing, but optical and analog image processing also are possible. This article is about general techniques that apply to all of them. The acquisition of images (producing the input image in the first place) is referred to as imaging.
Closely related to image processing are computer graphics and computer vision. In computer graphics, images are manually made from physical models of objects, environments, and lighting, instead of being acquired (via imaging devices such as cameras) from natural scenes, as in most animated movies. Computer vision, on the other hand, is often considered high-level image processing out of which a machine/computer/software intends to decipher the physical contents of an image or a sequence of images (e.g., videos or 3D full-body magnetic resonance scans).
In modern sciences and technologies, images also gain much broader scopes due to the ever growing importance of scientific visualization (of often large-scale complex scientific/experimental data). Examples include microarray data in genetic research, or real-time multi-asset portfolio trading in finance.
This document discusses single object tracking and velocity determination. It begins with an introduction and objectives of the project which is to develop an algorithm for tracking a single object and determining its velocity in a sequence of video frames. It then provides details on preprocessing techniques like mean filtering, Gaussian smoothing and median filtering to reduce noise. It describes segmentation methods including histogram-based, single Gaussian background and frame difference approaches. Feature extraction methods like edges, bounding boxes and color are explained. Object detection using optical flow and block matching is covered. Finally, it discusses tracking and calculating velocity of the moving object. MATLAB is introduced as a technical computing language for solving these types of problems.
This document provides an overview of key concepts in image processing, including:
- Digital image representation as a 2D array of integer pixel values
- Image acquisition through illumination of a scene and absorption of reflected energy by image sensors
- Sampling and quantization to convert a continuous image into discrete digital values
- Spatial and intensity resolution which determine image quality
- Common image file formats and operations like filtering, arithmetic/logical operations, and basic geometric transformations including translation, rotation, scaling and shearing.
This document provides an overview of digital image fundamentals and operations. It defines what a digital image is, how it is represented as a matrix, and common image types like RGB, grayscale, and binary. Pixels, resolution, neighborhoods, and basic relationships between pixels are discussed. The document also covers different types of image operations including point, local, and global operations as well as examples like arithmetic, logical, and geometric transformations. Finally, it introduces concepts of linear and nonlinear operations and announces the topic of the next lecture on image enhancement in the spatial domain.
This document proposes using spectral clustering based on the normalized graph Laplacian spectrum to solve problems in community detection and handwritten digit recognition. It summarizes the key concepts in graph signal processing and introduces spectral clustering. The paper provides a mathematical proof that the signs of the second eigenvector components of the normalized graph Laplacian can accurately partition a graph into two communities. It then applies this spectral clustering method to community detection and digit recognition, comparing results to other popular algorithms to demonstrate the advantages of the spectral clustering approach.
Introducing the Concept of Back-Inking as an Efficient Model for Document Ret...IJITCA Journal
Today, many institutions and organizations are facing serious problem due to the tremendously increasing
size of documents, and this problem is further triggering the storage and retrieval problems due to the
continuously growing space and efficiency requirements. This problem is becoming more complex with
time and the increase in the size and number of documents in an organization. Therefore, there is a
growing demand to address this problem. This demand and challenge can be met by developing a
technique to enable specialized document imaging people to use when there is a need for storing
documents images. Thus, we need special and efficient storage techniques for this type of information
storage (IS) systems.
In this paper, we present an efficient storage technique for electronic documents. The proposed technique
uses the Information Pixels concept to make the technique more efficient for certain image formats. In
addition, we shall see how Storing Information Pixels Addresses ( SIPA ) method is an efficient method for
document storage and as a result makes the document image storage relatively efficient for most image
formats.
The document discusses a method for 3D object recognition from 2D images using centroidal representation. It involves several steps: filtering and binarizing the image, detecting edges, calculating the object center point, extracting features around the centroid, and creating mathematical models using wavelet transforms and autoregression. Centroidal samples represent distances from the center to the boundary every 45 degrees. Wavelet transforms and autoregression are used to create scale and position invariant representations of the object for recognition.
Sign Language Recognition Using Image Processing For Mute Peoplepaperpublications3
Abstract: Computer recognition of sign language is an important research problem for enabling communication with mute people. This project introduces an efficient and fast algorithm for identification of the number of fingers opened in a gesture representing an alphabet of the Binary Sign Language.
The system does not require the hand to be perfectly aligned to the camera. The project uses image processing system to identify, especially English alphabetic sign language used by the mute people to communicate. The basic objective of this project is to develop a computer based intelligent system that will enable mute people significantly to communicate with all other people using their natural hand gestures.
The idea consisted of designing and building up an intelligent system using image processing, machine learning and artificial intelligence concepts to take visual inputs of sign language’s hand gestures and generate easily recognizable form of outputs.
Hence the objective of this project is to develop an intelligent system which can act as a translator between the sign language and the spoken language dynamically and can make the communication between people with mute and normal people both effective and efficient. The system is we are implementing for Binary sign language but it can detect any sign language with prior image processing.
Removal of Fixed-valued Impulse Noise based on Probability of Existence of t...IJECEIAES
This paper proposes a new approach for restoring images distorted by fixedvalued impulse noise. The detection process is based on finding the probability of existence of the image pixel. Extensive investigations indicate that the probability of existence of a pixel in an original image is bounded and has a maximum limit. The tested pixel is judged as original if it has probability of existence less than the threshold boundary. In many tested images, the proposed method indicates that the noisy pixels are detected efficiently. Moreover, this method is very fast, easy to implement and has an outstanding performance when compared with other well-known methods. Therefore, if the proposed filter is added as a preliminary stage to many filters, the final results will be improved.
This document summarizes a research paper that proposes segmenting images into objects by representing the image as a graph and using the Content Map Equation (CME) algorithm. The CME treats each pixel as a node in the graph and partitions the nodes to segment the image. The algorithm is tested on simple black and white images with added noise. Results show the CME achieves higher accuracy than K-means clustering in segmenting objects, though it sometimes partitions single objects across multiple segments. Future work involves merging CME segments to improve object segmentation.
This document discusses digital image processing concepts including:
- Image acquisition and representation, including sampling and quantization of images. CCD arrays are commonly used in digital cameras to capture images as arrays of pixels.
- A simple image formation model where the intensity of a pixel is a function of illumination and reflectance at that point. Typical ranges of illumination and reflectance are provided.
- Image interpolation techniques like nearest neighbor, bilinear, and bicubic interpolation which are used to increase or decrease the number of pixels in a digital image. Examples of applying these techniques are shown.
- Basic relationships between pixels including adjacency, paths, regions, boundaries, and distance measures like Euclidean, city block, and
The document provides guidance for students conducting research for a social sciences fair. It instructs students to select a topic and develop 3 research questions not in the form of "Do you like" or "Do you know". Students need at least one book and one internet source for their research. They are shown how to search the school catalog and databases like NetTrekker and Britannica Elementary to find relevant resources. The document emphasizes properly citing all sources used.
This document discusses a solar powered aircraft that recently completed its first cross-country flight across the US. It has the wingspan of a Boeing 747 but weighs as much as a Chevy Volt, and is powered by 10 horsepower electric motors and nearly 12,000 solar cells. The aircraft aims to promote renewable energy technologies and plans to fly around the world in 2015.
The document contains a Tamil language quote that translates to "Water today is not common to anyone in the world." It then states that approximately 1 in 8 people, or nearly 900 million people, lack access to clean water globally.
1) The document describes an eco quiz being held on June 5th, 2014 in Chennai, India. It provides the rules for the clockwise and anti-clockwise rounds, and lists 20 sample questions from the quiz.
2) The questions cover a range of topics including wildlife, the environment, and famous personalities known for conservation efforts.
3) Some of the questions identify famous tigers like Machli and provide details about their significance and conservation work.
Francis Ford Coppola was adapted from a novella and set during the Vietnam War.
The movies being referred to are:
1. Catch-22 directed by Mike Nichols
2. Apocalypse Now directed by Francis Ford Coppola
24)
This movie directed by David Fincher was based on a novel by
Patricia Highsmith. It had many twists and turns in its plotline
keeping the audience guessing throughout.
The movie had Oscar nominated performances by Gwyneth
Paltrow and Jeremy Northam in lead roles.
Identify the movie.
The Talented Mr. Ripley
25)
Around The World In 80 Minutes Quiz Part III - Islands Edition 2014parthiben1
- The document provides rules and instructions for a quiz game being played, including that there will be 24 questions in each clockwise and anti-clockwise round, questions are open to everyone, and the quiz master's decision is final.
- It explains the "pounce" and "bounce" rules where players can earn points for correct answers or lose points for incorrect answers during pounces, and get the next question if answering correctly during bounces.
- The clockwise round section indicates there will be 24 questions asked in that round.
This document outlines the rules and structure for a quiz competition. It will consist of two written rounds with 15 questions each, as well as clockwise and anti-clockwise verbal rounds also with 15 questions each. The quiz master's decision is final. Sample questions are provided about famous paintings, current events, geography, and other topics. Correct answers are also included.
The document discusses various details about Tamil cinema directors, actors, actresses and movies. It asks questions about the names associated with the details provided and asks the reader to identify them. Some key details discussed include the director who said he would have joined the LTTE if born in Tamil Eelam, the top heroine in the 80s who refused a role but now assists all of Maniratnam's movies, and the only actor to have his 100th film be a commercial hit.
1. The document discusses Vaali, a Tamil lyricist who sometimes acted in films. It notes that he often depicted critical moments realistically and shot most of his films within India rather than foreign locales. His female leads were typically naughty but good with hints of sophistication.
2. It describes a film from the British period involving railway promotion. Some Indians converted to Christianity to get promotions but reconverted after retirement. The English associated this group with deception for a long time.
3. It asks the reader to identify the name of a 1990s Tamil film with the protagonist's name prefixed by the word for this group, alluding to the film X-Y.
1) The document outlines the rules for an India Quiz, including details about different rounds, scoring, and topics that will be covered such as Indian cities, films, and current events.
2) It provides examples of questions that may be asked, such as identifying the inspiration for a song lyricist's work or the meaning behind a logo.
3) Samples of potential written questions are included, such as identifying monuments depicted in illustrations of Indian cities.
In imaging science, image processing is processing of images using mathematical operations by using any form of signal processing for which the input is an image, a series of images, or a video, such as a photograph or video frame; the output of image processing may be either an image or a set of characteristics or parameters related to the image. Most image-processing techniques involve treating the image as a two-dimensional signal and applying standard signal-processing techniques to it. Images are also processed as three-dimensional signals where the third-dimension being time or the z-axis.
Image processing usually refers to digital image processing, but optical and analog image processing also are possible. This article is about general techniques that apply to all of them. The acquisition of images (producing the input image in the first place) is referred to as imaging.
Closely related to image processing are computer graphics and computer vision. In computer graphics, images are manually made from physical models of objects, environments, and lighting, instead of being acquired (via imaging devices such as cameras) from natural scenes, as in most animated movies. Computer vision, on the other hand, is often considered high-level image processing out of which a machine/computer/software intends to decipher the physical contents of an image or a sequence of images (e.g., videos or 3D full-body magnetic resonance scans).
In modern sciences and technologies, images also gain much broader scopes due to the ever growing importance of scientific visualization (of often large-scale complex scientific/experimental data). Examples include microarray data in genetic research, or real-time multi-asset portfolio trading in finance.
This document discusses single object tracking and velocity determination. It begins with an introduction and objectives of the project which is to develop an algorithm for tracking a single object and determining its velocity in a sequence of video frames. It then provides details on preprocessing techniques like mean filtering, Gaussian smoothing and median filtering to reduce noise. It describes segmentation methods including histogram-based, single Gaussian background and frame difference approaches. Feature extraction methods like edges, bounding boxes and color are explained. Object detection using optical flow and block matching is covered. Finally, it discusses tracking and calculating velocity of the moving object. MATLAB is introduced as a technical computing language for solving these types of problems.
This document provides an overview of key concepts in image processing, including:
- Digital image representation as a 2D array of integer pixel values
- Image acquisition through illumination of a scene and absorption of reflected energy by image sensors
- Sampling and quantization to convert a continuous image into discrete digital values
- Spatial and intensity resolution which determine image quality
- Common image file formats and operations like filtering, arithmetic/logical operations, and basic geometric transformations including translation, rotation, scaling and shearing.
This document provides an overview of digital image fundamentals and operations. It defines what a digital image is, how it is represented as a matrix, and common image types like RGB, grayscale, and binary. Pixels, resolution, neighborhoods, and basic relationships between pixels are discussed. The document also covers different types of image operations including point, local, and global operations as well as examples like arithmetic, logical, and geometric transformations. Finally, it introduces concepts of linear and nonlinear operations and announces the topic of the next lecture on image enhancement in the spatial domain.
This document proposes using spectral clustering based on the normalized graph Laplacian spectrum to solve problems in community detection and handwritten digit recognition. It summarizes the key concepts in graph signal processing and introduces spectral clustering. The paper provides a mathematical proof that the signs of the second eigenvector components of the normalized graph Laplacian can accurately partition a graph into two communities. It then applies this spectral clustering method to community detection and digit recognition, comparing results to other popular algorithms to demonstrate the advantages of the spectral clustering approach.
Introducing the Concept of Back-Inking as an Efficient Model for Document Ret...IJITCA Journal
Today, many institutions and organizations are facing serious problem due to the tremendously increasing
size of documents, and this problem is further triggering the storage and retrieval problems due to the
continuously growing space and efficiency requirements. This problem is becoming more complex with
time and the increase in the size and number of documents in an organization. Therefore, there is a
growing demand to address this problem. This demand and challenge can be met by developing a
technique to enable specialized document imaging people to use when there is a need for storing
documents images. Thus, we need special and efficient storage techniques for this type of information
storage (IS) systems.
In this paper, we present an efficient storage technique for electronic documents. The proposed technique
uses the Information Pixels concept to make the technique more efficient for certain image formats. In
addition, we shall see how Storing Information Pixels Addresses ( SIPA ) method is an efficient method for
document storage and as a result makes the document image storage relatively efficient for most image
formats.
The document discusses a method for 3D object recognition from 2D images using centroidal representation. It involves several steps: filtering and binarizing the image, detecting edges, calculating the object center point, extracting features around the centroid, and creating mathematical models using wavelet transforms and autoregression. Centroidal samples represent distances from the center to the boundary every 45 degrees. Wavelet transforms and autoregression are used to create scale and position invariant representations of the object for recognition.
Sign Language Recognition Using Image Processing For Mute Peoplepaperpublications3
Abstract: Computer recognition of sign language is an important research problem for enabling communication with mute people. This project introduces an efficient and fast algorithm for identification of the number of fingers opened in a gesture representing an alphabet of the Binary Sign Language.
The system does not require the hand to be perfectly aligned to the camera. The project uses image processing system to identify, especially English alphabetic sign language used by the mute people to communicate. The basic objective of this project is to develop a computer based intelligent system that will enable mute people significantly to communicate with all other people using their natural hand gestures.
The idea consisted of designing and building up an intelligent system using image processing, machine learning and artificial intelligence concepts to take visual inputs of sign language’s hand gestures and generate easily recognizable form of outputs.
Hence the objective of this project is to develop an intelligent system which can act as a translator between the sign language and the spoken language dynamically and can make the communication between people with mute and normal people both effective and efficient. The system is we are implementing for Binary sign language but it can detect any sign language with prior image processing.
Removal of Fixed-valued Impulse Noise based on Probability of Existence of t...IJECEIAES
This paper proposes a new approach for restoring images distorted by fixedvalued impulse noise. The detection process is based on finding the probability of existence of the image pixel. Extensive investigations indicate that the probability of existence of a pixel in an original image is bounded and has a maximum limit. The tested pixel is judged as original if it has probability of existence less than the threshold boundary. In many tested images, the proposed method indicates that the noisy pixels are detected efficiently. Moreover, this method is very fast, easy to implement and has an outstanding performance when compared with other well-known methods. Therefore, if the proposed filter is added as a preliminary stage to many filters, the final results will be improved.
This document summarizes a research paper that proposes segmenting images into objects by representing the image as a graph and using the Content Map Equation (CME) algorithm. The CME treats each pixel as a node in the graph and partitions the nodes to segment the image. The algorithm is tested on simple black and white images with added noise. Results show the CME achieves higher accuracy than K-means clustering in segmenting objects, though it sometimes partitions single objects across multiple segments. Future work involves merging CME segments to improve object segmentation.
This document discusses digital image processing concepts including:
- Image acquisition and representation, including sampling and quantization of images. CCD arrays are commonly used in digital cameras to capture images as arrays of pixels.
- A simple image formation model where the intensity of a pixel is a function of illumination and reflectance at that point. Typical ranges of illumination and reflectance are provided.
- Image interpolation techniques like nearest neighbor, bilinear, and bicubic interpolation which are used to increase or decrease the number of pixels in a digital image. Examples of applying these techniques are shown.
- Basic relationships between pixels including adjacency, paths, regions, boundaries, and distance measures like Euclidean, city block, and
The document is a lab manual for a course on Computer Graphics and Multimedia. It contains:
1. A table of contents listing various sections like the time table, university scheme, syllabus, list of books, and list of programs.
2. The time table, university scheme, and syllabus provide details about the course schedule, assessment scheme, and topics to be covered.
3. The list of books and list of programs provide resources for students to refer to for the course and experiments to be performed in the lab.
INTRODUCING THE CONCEPT OF BACKINKING AS AN EFFICIENT MODEL FOR DOCUMENT RETR...IJITCA Journal
Today, many institutions and organizations are facing serious problem due to the tremendously increasing size of documents, and this problem is further triggering the storage and retrieval problems due to the continuously growing space and efficiency requirements. This increase in the size and number of documents is becoming a complex problem in most offices. Therefore, there is a demand to address this challenging problem. This can be met by developing a technique to enable specialized document imaging people to use when there is a need for storing documents images. Thus, there is a need for an efficient retrieval technique for this type of information retrieval (IR) systems.
INTRODUCING THE CONCEPT OF INFORMATION PIXELS AND THE SIPA (STORING INFORMATI...IJITCA Journal
Today, many institutions and organizations are facing serious problem due to the tremendously increasing size of documents, and this problem is further triggering the storage and retrieval problems due to the continuously growing space and efficiency requirements. This problem is becoming more complex with time and the increase in the size and number of documents in an organization. Therefore, there is a growing demand to address this problem. This demand and challenge can be met by developing a technique to enable specialized document imaging people to use when there is a need for storing documents images. Thus, we need special and efficient storage techniques for this type of information storage (IS) systems. In this paper, we present an efficient storage technique for electronic documents. The proposed technique uses the Information Pixels concept to make the technique more efficient for certain image formats. In addition, we shall see how Storing Information Pixels Addresses ( SIPA ) method is an efficient method for document storage and as a result makes the document image storage relatively efficient for most image formats.
INTRODUCING THE CONCEPT OF BACKINKING AS AN EFFICIENT MODEL FOR DOCUMENT RET...IJITCA Journal
This document presents an efficient technique called Back-Inking for document retrieval and image reconstruction from stored document images. Back-Inking works by replacing white pixels in an initialized blank image with black pixels, based on the addresses of information pixels stored in an array. The technique is tested on documents in different formats (GIF, BMP, JPG, TIF) and resolutions, showing that higher resolution images take longer to reconstruct but have better quality, while smaller partial document images reconstruct faster than full pages. Back-Inking provides an efficient way to retrieve images from stored address data.
Estrazione automatica delle linee in un'immagine digitalefrancescapadoin
This document discusses techniques for automatically extracting lines from a digital image. It introduces two common edge detection operators - Sobel and Canny - to identify edges in an image that can then be input to the Hough transform for line detection. It also explores using different filtering masks prior to edge detection and adaptive thresholding to improve results. The identified lines from the Hough transform are then intersected and clustered using k-means to estimate vanishing points for perspective calculation. Experimental validation of the results is also discussed.
Robust Fuzzy Data Clustering In An Ordinal Scale Based On A Similarity MeasureIJRES Journal
This paper is devoted to processing data given in an ordinal scale. A new objective function of a
special type is introduced. A group of robust fuzzy clustering algorithms based on the similarity measure is
introduced.
Identification system of characters in vehicular platesIJRES Journal
In this paper, is proposed a system that identifies the location and dimensions of the characters
relative to the image of a vehicular plate, when the location of the plate has not been accurate. The system is
divided into 4 stages, each with a specific purpose. Which are: binarization by thresholding, morphological
filtering, identification of the largest area and segmentation similarity. The first stage is used to find at the extent
if possible, the region occupied by the plate relative to the rest of the image. Then, in the filtering step, it seeks
to eliminate as far as possible the noise that interferes with the identification of the characters. The third stage is
used to identify the region occupied by the plate. Finally, in the fourth and last stage the segmentation by
similarity is used to identify position and dimension of the characters in the image, in stage a Kohonen neural
network is used.
Articulo
Journal of Computer Science & Technology; vol. 6, no. 2
The image recognizing process requires the identification of every logical object that compose every image, which first implies to recognize it as an object (segmentation) and then identify which object is, or at least which is the most likely one from the universe of objects that can be recognized (recognition). During the segmentation process, the aim is to identify as many objects that compose the images as possible. This process must be adapted to the universe of all objects that are looked for, which can vary from printed or manuscript characters to fruits or animals, or even fingerprints. Once all objects have been obtained, the system must carry on to the next step, which is the identification of the objects based on the called universe. In other words, if the system is looking for fruits, it must identify univocally fruits from apples and oranges; if they are characters, it must identify the character "a" from the rest of the alphabet, and soon. In this document, the character recognition step has been studied. More specifically, which methods to obtain characteristics exist (advantages and disadvantages, implementations, costs). There is also an overview about the feature vector, in which all features are stored and analyzed in order to perform the character recognition itself.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5520
Multimodal Searching and Semantic Spaces: ...or how to find images of Dalmati...Jonathon Hare
Tutorial at the "Reality of the Semantic Gap in Image Retrieval" tutorial at the first international conference on Semantics And digital Media Technology (SAMT 2006). 6th December 2006.
Similar to Manuscript document digitalization and recognition: a first approach (20)
Objeto de conferencia
Ciclo de conferencias en la Universidad Nacional Experimental del Táchira (UNET) (Venezuela, 2012)
Objetivo:
• Compartir la experiencia del SEDICI en todas las áreas que hacen al quehacer del repositorio: edición, catalogación, comunicación y difusión, software de soporte e interoperabilidad, servicios asociados y cuestiones legales, entre otras.
• Crear conciencia sobre el acceso abierto en todas sus formas.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/37866
Tesis de doctorado
De Giusti, Marisa Raquel; Gordillo, Silvia; Castro, Silvia; Suppi, Remo; Baldasarri, Sandra
Doctor en Ciencias Informáticas; Facultad de Informática
La enseñanza en el área de simulación de eventos discretos requiere integrar una variedad de conceptos teóricos y ponerlos en práctica a través de la creación y ejecución de modelos abstractos de simulación, con el objetivo de recopilar información que pueda traspolarse hacia los sistemas reales. Para construir modelos, ejecutarlos y analizar los resultados de cada ejecución se utilizan herramientas de software cada vez más sofisticadas que permiten expresar los elementos de los modelos en términos de entidades abstractas y relaciones, y que recopilan gran cantidad de datos y estadísticas sobre cada una de estas entidades del modelo. GPSS es una de estas herramientas, y se compone de un lenguaje de programación por bloques y un motor de simulación que traduce estos bloques en distintas entidades del modelo. A pesar de que su primera versión data de 1961, GPSS es aún muy utilizado por profesionales y empresas, y es una de las herramientas más utilizadas para la enseñanza de simulación de eventos discretos por instituciones académicas de todo el mundo.
El avance de la capacidad de cómputo de las computadoras ha permitido incorporar una mayor cantidad de herramientas y funciones a las distintas implementaciones de GPSS. Mientras que esto representa una ventaja para sus usuarios, requiere también un cada vez mayor esfuerzo por parte de los docentes para enseñar a sus estudiantes a aprovechar todo su potencial. Muchos docentes e investigadores han buscado optimizar la enseñanza de simulación de eventos discretos desde múltiples ángulos: la organización del curso y la metodología de enseñanza, la creación de elementos de aprendizaje que ayuden a aplicar los distintos elementos teóricos, la generación de herramientas para construir modelos GPSS, y la construcción de herramientas para comprender el motor de simulación por dentro.
En esta tesis se introduce una herramienta de software que permite construir modelos GPSS de manera interactiva, cuyo diseño fue pensado para integrar los elementos teóricos del curso con los objetos y entidades de GPSS. Esta herramienta también permite ejecutar estos modelos y analizar con alto nivel de detalle su evolución a través del tiempo de simulación, lo que permite a los estudiantes comprender cómo funciona el motor de simulación y cómo interactúan las distintas entidades entre sí. Se incluye también una propuesta de enseñanza basada en una fuerte participación de los estudiantes, que, por medio de esta nueva herramienta, les permite incorporar los conceptos más fácilmente. Esta propuesta de enseñanza fue puesta a prueba con alumnos del área de sistemas, quienes tomaron un curso que contiene los m
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/29753
Objeto de conferencia
The Benefits of Model-Driven Development in Institutional Repositories
II Conferencia Internacional Acceso Abierto, Comunicación Científica y Preservación Digital (BIREDIAL) (Colombia, 2012)
Los Repositorios Institucionales (RI) se han consolidado en las instituciones en las áreas científicas y académicas, así lo demuestran los directorios de repositorios existentes de acceso abierto y en los depósitos diarios de artículos o documentos realizados por diferentes vías, tales como el autoarchivo por parte de los usuarios registrados y las catalogaciones por parte de los bibliotecarios. Los sistemas RI se basan en diversos modelos conceptuales, por lo que en este trabajo se realiza un relevamiento bibliográfico del Desarrollo de Software Dirigido por Modelos (MDD) en los sistemas y aplicaciones para los RI con el propósito de exponer los beneficios de la aplicación del MDD en los RI. El MDD es un paradigma de construcción de software que asigna a los modelos un rol central y activo bajo el cual se derivan modelos que van desde los más abstractos a los concretos, este proceso se realiza a través de transformaciones sucesivas. Este paradigma proporciona un marco de trabajo que permite a los interesados compartir sus puntos de vista y manipular directamente las representaciones de las entidades de este dominio. Por ello, se presentan los beneficios agrupados según los actores que están presentes, a saber, desarrolladores, dueños de negocio y expertos del dominio. En conclusión, estos beneficios ayudan a que todo el entorno del dominio de los RI se concentre en implementaciones de software más formales, generando una consolidación de tales sistemas, donde los principales beneficiarios serán los usuarios finales a través de los múltiples servicios que son y serán ofrecidos por estos sistemas.; The Institutional Repositories (IR) have been consolidated into the institutions in scientific and academic areas, as shown by the directories existing open access repositories and the deposits daily of articles made by different ways, such as by self-archiving of registered users and the cataloging by librarians. IR systems are based on various conceptual models, so in this paper a bibliographic survey Model-Driven Development (MDD) in systems and applications for RI in order to expose the benefits of applying MDD in IR. The MDD is a paradigm for building software that assigns a central role models and active under which derive models ranging from the most abstract to the concrete, this is done through successive transformations. This paradigm provides a framework that allows interested parties to share their views and directly manipulate representations of the entities of this domain. Therefore, the benefits are grouped by actors that are present, namely, developers, business owners and domain experts. In conclusion, t
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/26044
Objeto de conferencia
I Conferencia sobre Bibliotecas y Repositorios Digitales (BIREDIAL) (Colombia, 2011)
El Servicio de Difusión de la Creación Intelectual es un proyecto de Repositorio Digital Institucional creado dentro de la UNLP, orientado a funcionar como el punto central de difusión de toda la producción académica generada dentro de la institución. Dada la tendencia vista en las principales instituciones académicas del mundo de hacer pública su producción científica a través de repositorios digitales de acceso abierto, el SeDiCI pasa a ser una herramienta estratégica para la jerarquización de la institución.
Desde su creación en el año 2003, el SeDiCI ha afrontado diversas dificultades que han influído directa o indirectamente en su desarrollo y crecimiento, pero aún a pesar de estas problemáticas, actualmente el SeDiCI cuenta con una base de datos documental que supera los 14000 recursos académicos propios (de la UNLP) expuestos bajo las políticas del Acceso Abierto. Esto convierte al SeDiCI en uno de los principales exponentes en su tipo, tanto a nivel nacional como regional (América Latina). En este documento se presentan experiencias y desafíos que el SeDiCI ha enfrentado, describiendo en cada caso el problema, su contexto y las vías de acción tomadas para superarlo. Los principales tópicos son: necesidad de apoyo institucional, reglas y metodologías de catalogación, mejoras en los servicios provistos a los usuarios, importación de recursos, entre otros.
Adicionalmente se describen algunos de los desafíos actuales, diferentes líneas de investigación y desarrollo orientadas a resolver los retos y a expandir y mejorar los servicios proporcionados a la comunidad de usuarios. Entre estos se encuentran: mecanismos y herramientas de harvesting, gestión de grandes volúmenes de información, ontologías y repositorios semánticos, legislación relacionada al Acceso Abierto, Diseminación Selectiva de la Información, Autoarchivo, etc.
El principal objetivo de este documento es exponer la experiencia adquirida a partir de este proyecto, con la intención de que resulte de utilidad para aquellas instituciones que se encuentran en el proceso de creación de sus propios repositorios institucionales, o bien que se encuentren frente a problemáticas similares a las aquí expuestas.; The Intellectual Creation Dissemination Service is the Institutional Digital Repository of La Plata National University.
The project is intended to be the main distribution source of all the academic work produced inside UNLP. In view of main worldwide institutions's trend towards the publication of academic resources through open access digital repositories, SeDiCI has pointed to become a strategic tool to bring relevance to the University.
Since its creation in the year 2003, SeDiCI has faced up many challenges and difficulties. Th
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5528
Preprint
DYNA; Edition 184
Los Repositorios Institucionales (RI) se han consolidado en la academia, prueba de ello es el crecimiento en número de registros en los directorios existentes realizado por diferentes vías: autoarchivo por parte de autores, la incorporación de material a cargo de bibliotecarios, entre otras. En este trabajo se hace un relevamiento bibliográfico sobre el uso del enfoque de Desarrollo de Software Dirigido por Modelos (MDD) en los sistemas de RI con el propósito de establecer una relación entre ellos. El MDD es un paradigma de construcción de software que asigna a los modelos un rol central y se derivan modelos que van desde los más abstractos a los más concretos. Este paradigma, además, proporciona un marco de trabajo que permite a los interesados compartir sus puntos de vista y manipular las representaciones de las entidades del dominio. En conclusión, el seguimiento de las diferentes investigaciones relevadas y lo aquí expuesto permiten incentivar implementaciones de software para los RI.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/35601
Articulo
e-colabora; vol. 1, no. 2
Desde su creación en el año 2003, el Servicio de Difusión de la Creación Intelectual ha afrontado diversas dificultades que han influido directa o indirectamente en su desarrollo y crecimiento. En este documento se presentan algunas de estas experiencias, describiendo en cada caso el problema, su contexto y las vías de acción tomadas para superarlo. Adicionalmente se describen algunas de las diferentes líneas de investigación y desarrollo actuales, orientadas a expandir y mejorar los servicios proporcionados a la comunidad de usuarios. De ahí que el objetivo principal de este trabajo sea exponer la experiencia adquirida, con la intención de que resulte de utilidad para aquellas instituciones que se encuentren en el proceso de creación de sus propios repositorios.; Since its creation in the year 2003, the Intellectual Creation Dissemination Service has faced up many difficulties. Thus the initiative has been directly and indirectly affected during its development and growth. This work presents some of these experiencies, describing problems, context and solution approaches. This document also describes some of the new and most recent challenges and the current research and development trends, which are oriented to improve and extend the services provided by SeDiCI. The main purpose of this document is to share the lived experiences with this project, which may be useful to other institutions working on their own digital repositories.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5527
Informe tecnico
America Learning & Media; Edición 027
Celsius es el software utilizado por los miembros de ISTEC que participan de la iniciativa LibLink para gestionar los pedidos de material bibliográfico de sus usuarios, atender solicitudes de provisión desde otras instituciones participantes, facilitar el intercambio de los documentos y generar estadísticas que permiten transparentar el intercambio y evaluar la calidad del trabajo de los participantes. Este software es desarrollado por la UNLP, y su primera versión data del año 2001. Celsius es ofrecido a todas las instituciones participantes de manera gratuita, a quienes también se les brinda documentación actualizada y asistencia personalizada para realizar la instalación y mantenimiento, instalar actualizaciones y formar al equipo de personas que utilizarán esta herramienta en cada institución. El proyecto Celsius3 tiene como característica principal la gestión centralizada de todas las instancias de Celsius de los miembros de LibLink. Esto implica, por un lado, la creación de instancias a medidas que se incorporan nuevos miembros, y por el otro la centralización todas las instalaciones existentes de Celsius 1.x y 2.x, lo que implica a su vez las migraciones y normalizaciones de sus respectivas bases de datos. Cabe aclarar que, si bien se busca contar con una instalación única y centralizada de esta plataforma, las instituciones seguirán contando con su propia instancia de Celsius: usuarios, pedidos, administradores, comunicaciones y estadísticas de cada instancia serán datos que gestionará cada institución de manera independiente del resto. Además la plataforma está pensada para que puedan mantener sus dominios actuales de acceso a Celsius.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/34504
Objeto de conferencia
PKP International Scholarly Publishing Conferences (Mexico, 2013)
La Universidad Nacional de La Plata tiene como objetivo prioritario la difusión de todo el conocimiento generado en la institución, a fin de devolver a la sociedad parte del esfuerzo invertido en la Universidad pública. Para alcanzar este objetivo se han generado diversas iniciativas desde la gestión, la creación de nuevos servicios y la implementación de nuevas líneas de investigación y desarrollo para potenciar tales servicios. El Servicio de Difusión de la Creación Intelectual es el Repositorio Institucional central de la UNLP, y posee en la actualidad todo tipo de materiales científicos y académicos producidos desde la Universidad, incluyendo artículos de revistas científicas, publicaciones en congresos, tesis de posgrado, tesinas, normativas y ordenanzas, libros y libros electrónicos, documentos de audio, materiales educativos, y fotografías de piezas de museos, entre otros. Este repositorio institucional, el mayor de la Argentina y uno de los principales de América Latina, posee un rol central en la difusión de la producción intelectual de la UNLP y en la coordinación con otros servicios y desarrollos, entre los que se destacan el Portal de Revistas de la UNLP, soportado por el software Open Journal Systems, y el Portal de Congresos de la UNLP, soportado por el software Open Conference Systems. En este trabajo se detallan los distintos mecanismos que se han implementado desde la Universidad Nacional de La Plata para facilitar la interacción entre su repositorio institucional SEDICI -construido sobre el software Dspace- y estos portales, evitar la duplicación de esfuerzos y maximizar la difusión abierta del conocimiento. Dichos mecanismos incluyen desarrollos tecnológicos como la utilización de diversos protocolos (RSS/Atom, Sword, OAI-PMH) para establecer un camino unificado de comunicación entre los sistemas, o el desarrollo de plugins que permiten exportar la información desde las herramientas de gestión de producción académica, en un formato comprensible por otros sistemas. Todo esto, además, involucra el establecimiento de flujos de trabajo entre los equipos de personas que conforman los distintos servicios, con la finalidad de obtener un mecanismo unificado que contemple los métodos y las herramientas para exportar los datos e incorporarlos al repositorio. Finalmente, en este trabajo se incluyen otros esfuerzos generados tanto desde el SEDICI como desde la presidencia de la Universidad para asegurar la preservación de toda la producción intelectual y brindar nuevos mecanismos para maximizar la visibilidad de estos documentos científicos y académicos. Esto abarca a otros actores y direcciones de esta institución tales como la Editorial de la Universidad (EDULP), la Dirección de Educación a Distancia (EAD) y la Rad
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/27406
Objeto de conferencia
XX Asamblea General de ISTEC (Puebla, México, 2014)
Sumario de la presentación:
Parte 1 - Conceptos básicos. Repositorio, interoperabilidad, preservación, guías, proyectos
Parte 2 - Metadatos de preservación
Parte 3 - Directrices sobre preservación PREMIS, Modelo de datos PREMIS, METS Otros esquema de metadatos y más posibilidades en la preservación
Parte 4- OAIS
Parte 5- DSPACE Modelo de datos, OAIS en Dspace
Panel LibLink.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/34889
Objeto de conferencia
XX Asamblea General de ISTEC (Puebla, México, 2014)
Sumario:
- Innovación tecnológica
- Formación de RRHH
- Calidad de servicios
- Integración con otras iniciativas
- Administración y gestión
Sesión final LibLink.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/34941
Objeto de conferencia
III Conferencia Internacional de Biblioteca Digital y Educación a Distancia
Se presenta una plataforma de recolección destinada a relacionar y unificar información disponible en distintos lugares de la Web-que siguen diferentes convenciones-para crear un repositorio temático que puedan navegar los usuarios. La plataforma será usada en el Servicio de Difusión de la Creación Intelectual (SeDiCI) y utiliza de manera combinada ontologías y tesauros para brindar información mejor clasificada.
Actualmente, la información está diseminada en recursos de la Web y los motores de búsqueda tradicionales le devuelven al usuario listas rankeadas sin proveer ninguna relación semántica entre documentos. Los usuarios pasan gran cantidad de tiempo para vincular unos documentos con otros y saber cuáles atacan el dominio completo del problema; recién al localizar las semejanzas y las diferencias entre fragmentos de información éstas se trasladan a su trabajo y sirven para la creación de nuevo conocimiento.
La plataforma propuesta separa los módulos de funcionamiento de los diferentes dominios de interés (temas) para permitir su utilización en distintas áreas de conocimiento. El desarrollo incluye dos agentes que recorren las URLs almacenadas en una base de datos (uno responsable de poblar una ontología y otro de obtener URLs relacionadas), un módulo capaz de reconocer las páginas marcadas, interpretar las etiquetas y proveer las reglas para extraer la información y guardarla en un fichero RDF; tras esta etapa se aplica una homogeneización y la información así transformada se clasifica en función de una ontología de dominio.
La plataforma vuelve más eficientes los procesos de extracción automática y búsqueda de información en fuentes heterogéneas que representan los mismos conceptos siguiendo distintas convenciones.; Presentation of a web collection platform designed to relate and unify information available on different standard web sources with a view to creating a user-browseable thematic repository.
The platform will be used at the Servicio de Difusión de la Creación Intelectual (SeDiCI) [Intellectual Creation Diffusion Service] combined with ontologies and thesaurus to provide improved data sorting.
Data is currently spread on web resources and traditional search engines return ranked lists with no semantic relation among documents. Users have to spend a great deal of time relating documents and trying to figure out which ones fully address the issue domain. It is only after locating similarities and differences that information fragments are applied to the user's work, enabling knowledge creation.
The proposed platform sorts out the different theme domain functioning modules to allow their use in various knowledge areas. Development includes two agents that searches data base stored URLs, one is
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5555
Objeto de conferencia
III Simposio Internacional de Bibliotecas Digitales (San Pablo, Brasil)
El proceso de reconocimiento de la escritura manuscrita forma parte de las iniciativas que propenden a la preservación de patrimonio cultural resguardado en Bibliotecas y archivos donde existe una gran riqueza de documentos y hasta fichas manuscritas que acompañan libros incunables. Este trabajo es el punto de partida de un proyecto de investigación y desarrollo orientado a la digitalización y reconocimiento de material manuscrito y la ponencia que aquí se presenta discute diferentes algoritmos utilizados en una primera etapa dedicada a "limpiar" la imagen de ruido para mejorarla antes de comenzar el reconocimiento de caracteres. Dado que PrEBi-SeDiCI forman parte integrante de redes de bibliotecas que intercambian documentos digitalizados vía scanning, el presente desarrollo ha tenido una utilización adicional relacionada al mejoramiento de las imágenes de documentos de intercambio que presentaban problemas comunes en la digitalización: bordes, impurezas, descentrado, etc.., si bien no es esta la finalidad de esta investigación no por ello resulta una utilidad menor en el marco de intercambios de consorcios de bibliotecas. Para que el proceso de digitalización y reconocimiento de textos manuscritos sea eficiente debe estar precedido de una etapa de "preprocesamiento" de la imagen a tratar que incluye umbralización, limpieza de ruido, adelgazamiento, enderezamiento de la línea base y segmentación de la imagen entre otros. Cada uno de estos pasos permitirá reducir la variabilidad nociva al momento de reconocer los textos manuscritos (ruido, niveles aleatorios de grises, inclinación de caracteres, zonas con más y menos tinta), aumentando así la probabilidad de reconocer adecuadamente los textos. En este trabajo se consideran dos métodos de adelgazamiento de imágenes, se realiza la implementación y finalmente se lleva adelante una evaluación obteniendo conclusiones relativas a la eficiencia, velocidad y requerimientos, así como también ideas para futuras implementaciones. En la primera parte del documento, se presentan algunas definiciones relacionadas con los métodos utilizados, luego se muestran los resultados obtenidos sobre un mismo conjunto de imágenes aplicando las teorías propuestas y finalmente, se exponen algunas ideas para optimizar los algoritmos elegidos.; The handwritten manusctipt recognizing process belongs to the iniciatives which lean to cultural patrimony preservation shielded in Libraries and files where there exists a big wealth in documents and even handritten cards that accompany incunable books. This work is point to begin with a research and development proyect oriented to digitalization and recognition of manuscipt materials and the paper presented here discuss diferent algorithms used in the first stage ded
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5534
Este documento describe varias tareas de curación automatizadas para mejorar la calidad e integridad de los datos en repositorios digitales implementados con DSpace. Estas tareas incluyen verificar enlaces externos, conectar metadatos con vocabularios controlados, y generar metadatos de preservación a partir de los archivos. También propone nuevas estrategias para seleccionar recursos a procesar y ejecutar tareas de curación de manera más eficiente.
Objeto de conferencia
III Conferencia de Bibliotecas y Repositorios Digitales de América Latina (BIREDIAL) y VIII Simposio Internacional de Bibliotecas Digitales (SIBD) (Costa Rica, 2013)
Durante el año 2012, el equipo del Portal de Congresos de la UNLP desarrolló un plugin para el software OCS que permite automatizar la generación de actas de congresos, y que a la vez simplifica de manera considerable la integración de dichas actas con el SEDICI. Este plugin puede, en líneas generales, exportar información de un mismo congreso en diferentes formatos, según se requiera en cada caso. Se han implementado distintos formatos de exportación, como por ejemplo CSV (valores separados por coma), HTML e incluso documento de texto (Word).
Estas tres exportaciones toman los datos de todos los trabajos (título, autores, resumen, instituciones, etc.) y los combinan en un único archivo, que el gestor de cada congreso puede descargar. También se implementó la exportación de todos los artículos de cada congreso en un único archivo .zip, y se modificaron los nombres de los archivos incluidos en el archivo .zip para que adopten el título del trabajo al que corresponden. De esta forma, los gestores pueden generar los libros de resúmenes de manera muy simple, y el repositorio institucional puede cargar todos los trabajos de cada congreso de una forma mucho más rápida.
Si bien se han implementado hasta ahora algunos formatos de exportación, el plugin fue pensado desde el principio para ser extendido hacia nuevos formatos, mediante un diseño basado en objetos que implementa el patrón de diseño conocido tomo Template Method.
Esta herramienta fue puesta a disponibilidad para usuarios en varios congresos del Portal de Congresos de la UNLP a modo de pruebas y los resultados fueron satisfactorios. En la actualidad, la herramienta ya está siendo utilizada para incorporar las producciones de los congresos al repositorio institucional SEDICI.
Aún cuando este desarrollo permitió simplificar la generación de actas de resúmenes de congresos, puede optimizarse más la incorporación de los trabajos al repositorio institucional, a fin de evitar la carga manual de los mismos. Por este motivo, el equipo del Portal de Congresos se encuentra evaluando diversas tecnologías de interoperabilidad entre sistemas. Entre ellas se destaca el protocolo sword en su versión 2 como una opción muy prometedora y relativamente simple. Este protocolo permite a los gestores de los congresos realizar el depósito automático o semiautomático de los trabajos desde el Portal de Congresos en el repositorio institucional, soportado por DSpace. Dicho depósito puede realizarse sobre una colección privada, lo que permitirá que el equipo de SEDICI revise los trabajos antes de confirmar su incorporación definitiva al repositorio.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/30522
Objeto de conferencia
International Conference on Engineering Education ICEE-2011 (Irlanda)
The Ibero-American Science and Technology Education Consortium (ISTEC) is a non-profit organization comprised of educational, research, industrial, and multilateral organizations throughout the Americas and the Iberian Peninsula. The Consortium was established in 1990 to foster scientific, engineering, and technology education, joint international research and development efforts among its members, and to provide a cost-effective vehicle for the application and transfer of technology. After twenty years, ISTEC has established a presence in the region, but it also has experienced problems to interact with different cultures and interests. During 2010 it suffered important changes in its organization and big efforts were realized to accomplish new goals and to share worldwide expertise, to facilitate distributed problem solving, creating the local critical mass needed for the development of regional projects in areas such as: continuing education, libraries and repositories, globalization of the culture of quality and accreditation standards, R&D, intellectual property development, capital acquisition, and social responsibility, among others. ISTEC continues to be dedicated to the improvement of Science, Engineering, Technology, Math education, R&D, and Entrepreneurship. The Consortium will foster technology transfer and the development of social and business entrepreneurs through the implementation of a global network that pretends to reach other countries in the world creating clusters of businesses and institutions that share common interest, assisting in the establishment of strategic alliances/joint ventures, and the promotion of collaborative partnerships in general.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/27159
Objeto de conferencia
BIREDIAL - Conferencia Internacional Acceso Abierto, Comunicación Científica y Preservación Digital
La preservación digital se define como el conjunto de prácticas de naturaleza política, estratégica y acciones concretas, destinadas a asegurar el acceso a los objetos digitales a largo plazo. El desarrollo de los repositorios institucionales, el crecimiento de sus contenidos y el reconocimiento de que la actividad institucional se canaliza, cada vez más, en soporte digital, obliga a los repositorios a acompañar su desarrollo con actividades destinadas a la preservación. En este trabajo se presentan el estándar 14721 (OAIS), los metadatos PREMIS y las directrices para la preservación, en conjunto con el esquema METS, para finalmente, explorar los metadatos en esquemas muy utilizados en la tarea normal de un repositorio (MODS, DC) y señalar los que resultan útiles a los fines de la preservación, proponiendo su reutilización. Un segundo objetivo práctico es mostrar qué herramientas de preservación ofrece el desarrollo DSpace que sustenta al repositorio SeDiCI - UNLP.; Digital preservation is defined as a set of political and strategic practices and concrete actions deployed to ensure long term access to digital objects. The development of institutional repositories, growth of their content and awareness that institutional activity is increasingly channeled through digital media require the deployment of preservation activities to sustain the evolution of repositories. This work presents ISO standard 14721 (OAIS), PREMIS metadata and guidelines for preservation, along with the METS schema to finally explWore metadata in widely-used repository-related schema (MODS, DC) and point out those useful for preservation purposes, proposing their reuse. A second practical purpose is to show the preservation tools offered by the DSpace development which supports the SeDiCI - UNLP repository.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/26045
Objeto de conferencia
XX Asamblea General de ISTEC (México, 2014)
Ponencia presentada en la XX Asamblea General de ISTEC (Puebla, México), en la cual se describen los distintos mecanismos de interoperabilidad implementados entre el repositorio institucional (SEDICI) y distintos servicios en línea de la Universidad Nacional de La Plata.
Lugar: INAOE (Puebla, México).
Expositores a través de videoconferencia: Gonzalo Luján Villarreal y Franco Agustín Terruzzi.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/34200
Articulo
Jornada Virtual de Acceso Abierto
El Servicio de Difusión de la Creación Intelectual (SeDiCI) es el repositorio institucional de la Universidad Nacional de La Plata (UNLP), creado en el 2003 con el objetivo de dar visibilidad a la producción académica producida en esta casa de estudios considerando que el acceso libre posibilita un mayor número de citas y por tanto un mayor impacto, atendiendo al rol fundamental de una institución pública de socializar el conocimiento. Creado en el año 2003, actualmente SeDiCI se encuentra posicionado entre los primeros 10 principales repositorios digitales de América Latina según la Webometrics, y ocupa la primera posición en Argentina como repositorio institucional. En este trabajo se presentan algunas de las principales características y servicios ofrecidos por el portal, desde su fundación hasta la actualidad.
Publicado en CD-ROM en: <i>Jornada Virtual de Acceso Abierto (Argentina 2010)</i>.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/27160
Articulo
The Online Journal of New Horizons in Education; vol. 3, no. 1
This work presents an open source web environment to learn GPSS language in Modeling and Simulation courses. With this environment, students build their models by selecting entities and configuring them instead of programming GPSS codes from scratch. Teachers can also create models so that students can apply, analyze and interpret results. Thus, it includes a simulation engine that stores snapshots of models as they are executed, and allows students to navigate through these snapshots. The environment may be combined with existing learning management systems.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/25669
Objeto de conferencia
VI Simposio Internacional de Bibliotecas Digitales (Brasil)
Digital repositories acting as resource aggregators typically face different challenges, roughly classified in three main categories: extraction, improvement and storage. The first category comprises issues related to dealing with different resource collection protocols: OAI-PMH, web-crawling, webservices, etc and their representation: XML, HTML, database tuples, unstructured documents, etc. The second category comprises information improvements based on controlled vocabularies, specific date formats, correction of malformed data, etc. Finally, the third category deals with the destination of downloaded resources: unification into a common database, sorting by certain criteria, etc.
This paper proposes an ETL architecture for designing a software application that provides a comprehensive solution to challenges posed by a digital repository as resource aggregator.
Design and implementation aspects considered during the development of this tool are described, focusing especially on architecture highlights.
Ver registro completo en: http://sedici.unlp.edu.ar/handle/10915/5529
More from Servicio de Difusión de la Creación Intelectual (SEDICI) (20)
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
Manuscript document digitalization and recognition: a first approach
1. MANUSCRIPT DOCUMENT DIGITALIZATION AND RECOGNITION:
A FIRST APPROACH
Marisa R. De Giusti1
, Maria Marta Vila, Gonzalo Luján Villarreal
Proyecto de Enlace de Bibliotecas (PrEBi)
Servicio de Difusión de la Creación Intelectual (SeDiCI)
Universidad Nacional de La Plata
La Plata, Argentina.
marisadg@ing.unlp.edu.ar, {vilamm, gonetil}@sedici.unlp.edu.ar
Abstract. 1
The handwritten manuscript recognizing process belongs
to a set of initiatives which lean to the preservation of
cultural patrimony gathered in libraries and archives,
where there exist a great wealth in documents and even
handwritten cards that accompany incunabula books. This
work is the starting point of a research and development
project oriented to digitalization and recognition of
manuscript materials. The paper presented here discuss
different algorithms used in the first stage dedicated to
“image noise-cleaning” in order to improve it before the
character recognition process begins.
In order to make the handwritten-text recognition and
image digitalization process efficient, it must be preceded
by a preprocessing stage of the image to be treated, which
includes thresholding, noise cleaning, thinning, base-line
alignment and image segmentation, among others. Each of
these steps will allow us to reduce the injurious variability
when recognizing manuscripts (noise, random gray levels,
slanted characters, ink level in different zones), and so
increasing the probability of obtaining a suitable text
recognition. In this paper, two image thinning methods are
considered, and implemented. Finally, an evaluation is
carried out obtaining many conclusions related to
efficiency, speed and requirements, as well as ideas for
future implementations.
Keywords: patrimonial conservation, digitalization,
thinnig, connected components
1. INTRODUCTION.
During the last years, the digitalization of document pieces
watched over in libraries and archives of all over the world
has attained a great significance for its infinite
possibilities, both for specialists and a broader public. This
new social space, enabled by digital technologies, allows
us to discover a set of objects which constituted a highly
unknown cultural richness. However, the current
advantages of this technology have also represented a hard
learning, which has encompassed dramatic mistakes paid
with the destruction or considerable damage of original
documents. For this practical reason, the problem of
digitalization, and its own technological nature, has
opened an international reflection space which is mainly
characterized by the constant flow of specialized
information and by the introduction of a constant social
benefit discourse.
When digitized, documents are better preserved, though
the search and access to such information is rather slow,
sequential, and thus highly inefficient. Cataloguing such
documents may solve the problem, but requires a large
1
Investigador Comisión de Investigaciones Científicas de la
Provincia de Buenos Aires – CIC.
quantity of human and temporal resources to carry out this
process; the automatic indexation of digitized documents -
using both pattern recognition systems and database
management systems - allows accelerating the cataloguing
process as well as the searching and access processes,
turning it into a viable solution in terms of complexity,
times and costs.
In order to recognize manuscript handwriting in a digitized
document, a set of operations allowing the individual
characterization of each object, extracting the biggest
possible quantity of characteristics, should be carried out.
In order to make this process of feature extraction as
efficient as possible, a previous processing of the image
must be carried out, removing any ambiguous or
confusing component among similar objects. One of the
most important tasks is obtaining the skeleton of the
objects, which is achieved by thinning such objects,
keeping its size and shape.
2. PREVIOUS DEFINITIONS.
An image can be viewed as a set of points which have a
certain value indicating its color2
.2
In grayscale images,
those values range from 0 (black point – without
brightness) to 255 (white point).
The images we will work with have already been
binarized (thresholded), applying to them a previous
treatment in which, based on a threshold value, all the
pixels surpassing such value are assigned 0, and 1 is
assigned to those under such threshold. The new image
constituted by 1s and 0s allows us to work in handwriting
recognition eliminating the problem of colors or grayscale.
The selection of the threshold can be manual or automatic,
existing several methods for choosing the best threshold.
Even though a detailed description exceeds the purposes
of this paper, interested readers may consult [11].
The set of pixels (also called point p = (x,y) with x , y ∈
N) making up an image can be considered in two ways: 3
a) As a matrix, where I [ p ] = I [ (x,y) ] = k; with k
pixel value x,y (k ∈ [0,255] in gray images; k ∈ [0,1] in
binarized images)
b) As a function Pixel: (Image, point) -> value where
Pixel( I , p) = Pixel( I , (x,y) ) = k.
In order to thin an image keeping the topology (so as to
not alter the shapes), the points belonging to the border
should be removed, i.e. those points with the following
properties:
2
For colored images, each point has 3 values: red, green, and blue
value (RGB). Interested readers may consult [10][11].
3
The present work opts for the functional notation, using
mathematical expressions and, in some instances, including
matrices in the cases of concepts such as neighborhood and
adjacency.
JCS&T Vol. 5 No. 3 October 2005
158
2. 1) Black point (it is useless to remove white points)
2) Simple point
3) Final point exclusion
Where:
• A pixel P is considered as simple point if the set of
P’s black neighbors has a single connected component
adjacent to P.
• A pixel P is considered as final point if it has a single
black neighbor (end point of the image).
In order to eliminate the border points, several runs shall
be carried out (thus keeping the topology). In each run, all
the “removable” points should be marked, and then the
removal process should be fulfilled. Runs will be repeated
until the removal of at least one pixel is achieved.
Neighborhood Concept. Pixel’s neighbors are conditioned
to the type of meshing used for representing the image.
q-neighborhood is defined (also q-adjacency) as a set of
pixels neighbors of p. The value of q will depend on the
type of meshing. If a hexagonal mesh is used, each image
pixel (excluding the borders) has 6 neighbors, in which
case we are dealing with a 6-neighborhood. A
quadrangular mesh presents two possible options: a
4-neighbor (considering the pixels at the sides, up and
down of the pixel, as neighbors) or a 8-neighborhood
(idem 4-neighborhood plus 4 pixels of the diagonals).
(p,q) adjacency is defined as:
p-adjacency for black pixels
q-adjacency for white pixels44
Computation of connected components in an image. A
connected component (CC) of an image is a set of pixels
which fulfils the norm that, for each pair, there exist a
digital path linking them. A digital path of a pixel p to
another pixel p is a set of pixels Ppq = {pi i=0..n} so that:
1) ∀ pi ∈ Ppq, color(pi ) = color(p) = color(q)
2) p0 = p; pn = q
3) For all i=1,...,n-1, pi has exactly two neighbors in Ppq
which are pi-1 , pi+1
4) p0, pn have exactly one neighbor which are p1 and pn-1,
respectively.
Finally, a connected component is bounded if it does not
have any border pixel.
3. AN ALGORITHM FOR COMPUTING
CONNECTED COMPONENTS
(4-ADJACENCY IS USED).
Considerations: the image used has been binarized. Pixels
value 1 (black) or 0 (white). A matrix is used, of the same
size of the image which will store labels to characterize
the original image pixels according to its type of
equivalence.
An image is traced from left to right and upside down. For
each pixel p=(x,y), the neighbors p1=(x-1,y) and
p2=(x,y-1) are examined.
- If they both value 0 (are white), a new label is created
and is assigned to P in the label matrix.
- If only one is 0, the same label of the other (which is
not 0) is assigned to P.
- If none is 0, the label of any of them is assigned to P.
Two scenarios may be presented:
4
q-distance between two pixels is known as the length of the
shortest path linking them.
o The label s of p1 and p2 are equal, with which
the label of p will be the same in both.
o The labels of p1 and p2 are different, in this case p
receives the label of any of the two. Let’s suppose it is
assigned the label of p1. Then, it shall be recalled that,
even though p2 and p have different labels, they belong
to the same component.
If the labels are considered as types of equivalence to
which pixels belong, if p1 belongs to type 1 and p2 to type
2, and during the image processing p1 and p2 belong to
the same component, the type of equivalence of p1 is thus
the same as that of p2. In consequence, Type 1=Type 2. In
order to keep record of this, a vector in memory is used,
which has as many places as labels used. Labels are
integers, all of which allows us to use them to access
directly the vector as indexes.
Each position of the vector will represent the label, and its
content will represent the type of equivalence to which
such label belongs. For example, if we have the vector:
V = {1,2,3,2,3,1,4}
the first position of the vector is position 0, the last is the
n-1 (in our case, 6), and the semantics of vector V is:
“label ‘x’ belongs to the type of equivalence V[x]”, then:
• Labels 0 and 5 belong to the type of equivalence 1.
• Labels 1 and 3 belong to the type of equivalence 2.
• Labels 2 and 4 belong to the type of equivalence 3
• Label 6 belongs to the type of equivalence 4.
All of which means: let Eq :: Label -> Integer be a
function taken by a label and returns the type of
equivalence to which such label belongs;
Eq(x) = V[x];
With this procedure, we have obtained a matrix containing
the labels of all the pixels according to their types of
equivalence. Several labels belonging to the same type
make the matrix normalization mandatory, tracing it and
assigning its type of equivalence to each element.
This algorithm will also store all the elements belonging to
each type, reason why an option would be keeping this
matrix in memory. This would generate a great waste of
memory, since space for spaces in blank (those pixels with
no value and not belonging to any type of equivalence) is
also reserved. As counterpart, this solution allows us to
access directly the type of equivalence of a certain pixel.
Let p = (x,y) pixel ∈ I image.
Eq(p) = I(x,y);
(We shall recall that images can be viewed as functions
taking two whole values x and y, and return the value of
the pixel in position (x,y)).
Another solution consists in storing another vector in the
vector of the existing types of equivalence, where all the
points belonging to that type of equivalence are stored for
each position. The definition of type would be:
Type
Point = record
X,Y : integer // pixels
End;
VectorOfPoints = array of point;
Class = record
Points: PointVector
numberOfClass: integer
end;
JCS&T Vol. 5 No. 3 October 2005
159
3. Classes = array of Class
The main advantage of this implementation lies in the
memory saving, since only those points belonging to some
type of equivalence will be stored; the disadvantage is that
finding the class of equivalence of a certain point will
imply tracing the vector of types of equivalence, and for
each type, tracing the points vector belonging to it. In the
best of the cases, this will mean a single access (first class,
first point). In the worst: all the points of a given image I
belong to some type of equivalence and image I has x
pixels of height and y of width in a total of z = x*y. In the
worst of the cases, it shall be considered that the searched
point is found in the last position of the vector and in the
last type of equivalence, and the access to this point will
take z accesses to memory.
This worst case is not a hundred percent real, since it is
impossible that all the points of an image belong to some
type of equivalence, since this would mean that it has, for
instance, all of its pixels black. From the experience
acquired, it is possible to estimate (for the type of images
we have worked with) that less than the half of the pixels
are black and, in some cases, the number only reaches the
20%.
4. A ZHANG-SUEN’S THINNING ALGORITHM.
This method is fast and simple to implement; it counts
with two subiterations in which those pixels fulfilling the
rules defined for iteration are removed. Recalling that:
• Label 6 belongs to the type of equivalence 4.
• A pixel is a final point if it has a single black
neighbor, being the rest all white.
• A pixel’s connectivity is defined as the number of
objects it could connect in the original image, and is
computed turning round a pixel clockwise and counting
how many color changes are produced. The number of
changes will be the connectivity, i.e., the number of
regions it connects.
As a first step of the algorithm, an image smoothing is
applied, and all the pixels with two or more black
neighbors and with connectivity inferior to two5
are
deleted. Then, the two iterations are carried out. In order
to remove a pixel in the first iteration, it should have the
following properties:
• Count with connectivity 1.
• Have a quantity of black neighbors between 2 and 6
(included).
• Have at least one of the following pixels in white: [x-
1,y], [x,y+1], [x,y-1]
• Have at least one of the following pixels in white [x-
1,y], [x+1,y], [x,y-1]
In the second iteration, the pixels fulfilling the following
rules will be removed:
• Count with connectivity 1.
• Have a quantity of black neighbors between 2 and 6
(included).
• Have at least one of the following pixels in white:
[x-1,y], [x,y+1], [x+1,y]
• Have at least one of the following pixels in white:
[x,y+1], [x+1,y], [x,y-1]
As previously expressed, iterations will go on as points are
being eliminated. If the pre-processing of the image is not
carried out, the results are not so good; results are shown
5
Note that this requires an extra processing, reducing the
algorithm efficiency.
in Figure 1 and 2.
a) b) c)
Figure 1: a) Original Image b) Skeleton of the image
obtained with the algorithm based on Connected
Components c) Skeleton obtained with the Zhang-Suen
method.
a) b) c)
Figure 2: Similar to Figure 1) Here we can notice in detail
how the lines of each method are drawn; a greater thickness
is observed in the case of the Zhang-Suen method.
5. PROBLEM IN NOISY BORDERS.
According to the experiments carried out, the results have
not been so encouraging in some images, for instance, in
those cases in which the borders were dithering or have
random noise, case in which, the thinning process
produces short random lines, generating a distortion of the
image or a loss of its shape. The problems presented in
this type of images lead to drawbacks when characters are
to be recognized.
Those images presenting this characteristic have a “dirty”
border with points distributed at random and, when they
are thinned, the random points are not removed since they
fulfill the characteristics of the border point. Both
algorithms react in the same fashion before this
phenomenon, generating lines all along the border of the
image, similar to ramifications of its skeleton, in an
attempt to keep the border points together with such
skeleton. (See Figure 3).
Figure 3. Skeletons of the first image obtained with CC in the
first case and with Zhang-Suen in the second.
A solution with filters. This problem can be solved by
applying a mean filter to the image before it is thinned,
thus producing smoothest borders and without the
previous imperfections. In the results obtained (see Figure
4), we have attained clearly better image skeletons,
especially with the Zhang Suen algorithm, since with the
CC-based method we not always obtain the same
improvements. As disadvantage, this method adds an extra
processing cost.
JCS&T Vol. 5 No. 3 October 2005
160
4. Figure 4. Image to which the mean filter was applied;
Skeleton of the same image with CC; Skeleton of such image
with Zhang-Suen.
6. OPTIMIZATION WITH THE CONNECTED
COMPONENTS – BASED METHOD.
The problem of the thinning algorithm based on the
computation of connected components is its high
complexity and inefficiency. It can be upgraded by using
another method for the recognition of simple points (main
use of the algorithm) without the need of obtaining all the
connected components.
This technique has the same basis as the connected
components-based method: the removable points are those
which are simple and are not final point, though it requires
considerable lesser time to determine whether a point is
simple. The idea of this optimization consists in verifying
each point in a given side; this means that a point may be
viewed as a simple point in a direction if its neighbors
fulfill certain conditions in such direction.
The algorithm carries out several runs, verifying in each a
different size: north, south, east, and west sides. In each of
the runs, the image is traced, searching those black points
which are simple according to the sense of the tracing and
which are not a final point.
Let d be the direction in which point P is being analyzed.
Let Pd be the set of conditions indicated if point P is
simple in direction d. Of course, for each possible d
(north, south, east, and west), the set Pd will be different.
The elements of the set Pd are called ci functions
ci : (Image, Point) -> [F,T] (F = FALSE, T = TRUE)
The structure of the set Pd is similar for all of the cases:
there exists a set of conditions that should NOT be
fulfilled and others that should be necessarily fulfilled. A
common condition is that point P should have the black
value, which will be here ignored as member of the set
(since black points are always being analyzed).
Let I(x,y) : Image -> [0,1] be function returning the value
of point (x,y). (1= white, 0 = black)
Let Sn:: Point -> [F,T] Sn be the function indicating
whether a point if north simple. (F = FALSE, T = TRUE)
d = North (1)
• c1= I(x,y-1) =0 ^ I(x+1,y) =1 ^ I(x,y+1) =0
• c2= I(x,y-1) =1 ^ I(x-1,y-1) =0 ^ I(x-1,y) =1
• c3= I(x-1,y) =1 ^ I(x-1,y+1)=0 ^ I(x,y+1) =1
• c4= I(x,y+1)=1^ I(x+1,y+1)=0 ^ I(x+1,y) =1
• c5= I(x+1,y)=1^ I(x+1,y-1) =0 ^ I(x,y-1) =1
Figure 5 graphically shows the 5 conditions.
Black and white points are those points that should have
that value. Gray points are points whose value is of no
interest. The point marked with an (x) is the analyzed
point.
Figure 5: Cases allowing the identification of a north simple
point.
Similarly, Ss(p) (South Simple Point), Se(p)( EastSimple
Point) So(p) (West Simple Point) are defined as:
d = South (2)
• c1 = I(x-1,y)=1 ^ I(x,y+1)=0
• c2, c3, c4 and c5 are the same
d = Est (3)
• c1 = I(x-1,y-1)=1 ^ I(x-1,y)=0 ^ I(x+1,y)=0
• c2, c3, c4 and c5 are the same
d = West (4)
• c1 = I(x,y+1)=1 ^ I(x-1,y)=0 ^ I(x+1,y)=0
• c2, c3, c4 and c5 are the same
As observed, conditions c2,c3,c4 and c5 are always the
same and have been included in the set Pd for the sake of
the algorithm’s implementation, in which an or logic is
carried out among all the conditions of the set (then, such
result is rejected in order to represent the ¬∃ of each
function’s definition).
Once the points fulfilling the simple, and not the final,
point conditions (in the corresponding direction) are
found, they are removed changing their values into white.
This will be repeated, like in the previous algorithm, while
points fulfilling some conditions are being found.
7. DIFFERENT BORDERS.
Figure 6 shows the images, two of which have been
thinned with each of the described algorithms; the thinning
in both cases (b and c) is right, keeping the topology and
without unexpected effects, but if we observe in more
detail, we could see that the line generated by the CC
algorithm is thinner than the line generated by Zhang-
Suen.
Figure 6 shows three images. The first is a line thinned
using CC, and the third is the same line, though Zhang-
Suen was applied. The central image shows in gray the
new points emerging when using the latter algorithm.
T if I(x+1 , y)= 1 ^ ¬∃ ci∈ Pd
ci= T with i=1, 2,3,4,5
F otherwise
Ss(p)
T if I(x , y+1)= 1 ^ ¬∃ ci∈ Pd
ci= T with i=1, 2,3,4,5
F otherwise
Se(p)
T if I(x , y-1)= 1 ^ ¬∃ ci∈ Pd
ci= T with i=1, 2,3,4,5
F otherwise
So(p)
T if I(x-1 , y)= 1 ^ ¬∃ ci∈ Pd
ci= T with i=1, 2,3,4,5
F otherwise
Sn(p)
JCS&T Vol. 5 No. 3 October 2005
161
5. Figure 6: Lines 1 and 3 have been thinned ith CC and
his is due the fact that, by using 4-adjacency in CC, in
w
Zhang-Suen methods, respectively. The central line shows the
“difference” between both.
T
order to keep a line linked (within the same CC), only its
vertices are to be handled. While in this case the problem
does not get worse, in several instances it is desirable to
count with more defined lines. The solution to this
particular situation is quite simple: instead of taking 4-
adjacency, use 8-adjacency when finding the CC; i.e.,
instead of considering just two points, 4 points are to be
considered. Figure 7 shows in gray, in the first place, the
points considered with 4-adjancency, while next they are
presented with 8-adjacency.
Figure 7: Pixels to be analy we take 4-adjacency and
hen using 8-adjacency, we get similar results in both
8. CONCLUSIONS.
Even though both algorithms render proper results, the
zed if
8-adjacency.
W
algorithms.
Zhang-Suen’s algorithm proved to be simpler to
implement and, in terms of empirical efficiency, we
observed that this algorithm typically demands inferior
running times than the CC-based algorithm, even in its
improved version. Data can be seen in Figures 8 and 9,
where the times spent in the running of each algorithm in
its simple version, and after applying filters, are
graphically shown.
Figure 8 Analysis of the times for the different techniques
nother advantage of the Zhang-Suen’s algorithm is its
developed over the image of Figure 1.
A
treatment of borders with random noise. As previously
mentioned, when a mean filter is applied over the image,
we can obtain satisfactory results, while the CC-based
algorithm not always presents those results.
Figure 9: Time Scheme applied to Figure 3 and 4 with the
owever, this does not mean that it is always better to use
er hand, an image connected components are
REFERENCES.
[1] Elisabetta Bruzzone, Meri Cristina Coffetti; Elsag spa
Sargur N. Srihari; “Off-Line
n and W.
. Rath and R. Manmatha; Features for Word
different techniques
H
the former algorithm. One of the problems found with the
Zhang-Suen’s algorithm is that if there exist intersected
thick lines, when they are thinned, they generate an effect
in which such intersection produces a segment instead of a
simple point. This is known as “necking” effect; and in
order to solve it the image should be pre-processed,
focusing on its close angles. But this entails more
processing, which is translated in extra running time. The
“necking” effect, though it exists, is not so marked when
using CC.
On the oth
also used for other purposes; for instance, to know the
quantity of closed areas in an image (holes), reversing the
image (changing the whites to blacks and blacks to
whites). Then, all the points belonging to the same CC will
belong to the same closed area; and the total quantity of
closed areas will be the quantity of connected components
minus 1. The method is also used as basis for segmenting
the image in words, taking in principle elements of the
same connected component as simple elements.
If we take a global look to the system, using connected
components may benefit the efficiency, since they are
computed only once and they are then used as many times
as is required.
– R&P Departament. An algorithm for Extracting Cursive
Text Lines, Genova – Italy
[2] Radmilo M. Bozinovic,
Cursive Scrtipt Word Recognition”. IEEE Transactions on
Pattern Analysis and Machine Intelligence. Vol 11 No. 1
[3] K. Badie and M. Shimura. Machine recognition of
roman cursive script in Proc. 6th
. Int. Conf. Patternt
Recognition, Munich, West Germany, Oct 1982.
[4] R. Manmatha, Chengfeng Han, E. M. Risema
B. Croft; “Indexing Handwriting Using Word Matching”.
Center for Intelligent Information Retrieval, Computer
Science Department, Univeristy of Massachusetts,
Amherst.
[5] Toni M
JCS&T Vol. 5 No. 3 October 2005
162
6. Spotting in Historical Manuscipts. Multi-Media Indexing
and Retrieval Group. Center for Information Retrieval.
University of Massachusetts, Amherst.
[6] L. Fletcher and R. Kasturi. “A robust algorithm for text
asey. “Block segmentation
g and S. N. Srihari. “Classification of
string separation from mixed text/graphics images”. IEEE
Transactions on Pattern Analysis and Machine
Intelligence, 10:910-918. 1988
[7] F. Wahl, K. Wong and R. C
and text extraction in mixed text/image documents”.
Computer Vision Graphics and Image Processing, 20:375-
390,1982
[8] D. Wan
newspaper image blocks using texture analysis”.
Computer Vision Graphics and Image Processing, 47:329-
352
[9] Michel Weinfeld, “Reconnaissance d'ecriture
manuscrite: segmentation de mots”. Département
d'Enseignement et de Recherche en Informatique. École
Polytechnique, Paris, France
[10] William Prat, John Wiley & Sons, Digital Image
Processing, 1991, Second Edition
[11] Gonzalez Rafael, Woods, Addison-Wesley Digital
Image Processing, 1992, Second Edition
[12] T. M. Rath, R. Manmatha. “Word Spotting for
Handwritten Historical Document Retrieval”
JCS&T Vol. 5 No. 3 October 2005
163
Received: Jul. 2005. Accepted: Sep. 2005.