Abstract Recent advances in multimedia technologies allow the capture and storage of video data with relatively inexpensive computers. Furthermore, the new possibilities offered by the information highways have made a large amount of video data publicly available. However, without appropriate search techniques all these data are hardly usable. Users are not satisfied with the video retrieval systems that provide analogue VCR functionality. For example, a user analyses a soccer video will ask for specific events such as goals. Content-based search and retrieval of video data becomes a challenging and important problem. Therefore, the need for tools that can be manipulate the video content in the same way as traditional databases manage numeric and textual data is significant. Therefore, a more efficient method for video retrieval in WWW or within large lecture video archives is urgently needed. This project presents an approach for automated video indexing and video search in large lecture video archives. First of all, we apply automatic video segmentation and key-frame detection to offer a visual guideline for the video content navigation. Subsequently, we extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames and Automatic Speech Recognition on lecture audio tracks. Keywords—Feature extraction, video annotation, video browsing, video retrieval, video structure analysis
Unsupervised object-level video summarization with online motion auto-encoderNEERAJ BAGHEL
Unsupervised video summarization plays an important role on digesting, browsing, and searching the ever-growing videos every day.
Author investigate a pioneer research direction towards the unsupervised object-level video summarization.
It can be distinguished from existing pipelines in two aspects:
Extracting key motions of participated objects
Learning to summarize in an unsupervised and online manner.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Video content analysis and retrieval system using video storytelling and inde...IJECEIAES
Videos are used often for communicating ideas, concepts, experience, and situations, because of the significant advances made in video communication technology. The social media platforms enhanced the video usage expeditiously. At, present, recognition of a video is done, using the metadata like video title, video descriptions, and video thumbnails. There are situations like video searcher requires only a video clip on a specific topic from a long video. This paper proposes a novel methodology for the analysis of video content and using video storytelling and indexing techniques for the retrieval of the intended video clip from a long duration video. Video storytelling technique is used for video content analysis and to produce a description of the video. The video description thus created is used for preparation of an index using wormhole algorithm, guarantying the search of a keyword of definite length L, within the minimum worst-case time. This video index can be used by video searching algorithm to retrieve the relevant part of the video by virtue of the frequency of the word in the keyword search of the video index. Instead of downloading and transferring a whole video, the user can download or transfer the specifically necessary video clip. The network constraints associated with the transfer of videos are considerably addressed.
Unsupervised object-level video summarization with online motion auto-encoderNEERAJ BAGHEL
Unsupervised video summarization plays an important role on digesting, browsing, and searching the ever-growing videos every day.
Author investigate a pioneer research direction towards the unsupervised object-level video summarization.
It can be distinguished from existing pipelines in two aspects:
Extracting key motions of participated objects
Learning to summarize in an unsupervised and online manner.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Video content analysis and retrieval system using video storytelling and inde...IJECEIAES
Videos are used often for communicating ideas, concepts, experience, and situations, because of the significant advances made in video communication technology. The social media platforms enhanced the video usage expeditiously. At, present, recognition of a video is done, using the metadata like video title, video descriptions, and video thumbnails. There are situations like video searcher requires only a video clip on a specific topic from a long video. This paper proposes a novel methodology for the analysis of video content and using video storytelling and indexing techniques for the retrieval of the intended video clip from a long duration video. Video storytelling technique is used for video content analysis and to produce a description of the video. The video description thus created is used for preparation of an index using wormhole algorithm, guarantying the search of a keyword of definite length L, within the minimum worst-case time. This video index can be used by video searching algorithm to retrieve the relevant part of the video by virtue of the frequency of the word in the keyword search of the video index. Instead of downloading and transferring a whole video, the user can download or transfer the specifically necessary video clip. The network constraints associated with the transfer of videos are considerably addressed.
An Exploration based on Multifarious Video Copy Detection Strategiesidescitation
We co-exist in an era, where tonnes and tonnes of
videos are uploaded every day. Video copy detection has become
the need for the hour as most of them are user generated
Internet videos through popular sites such as YouTube. It acts
as a medium to restrain piracy and prove whether the contents
are legitimate. The usual procedure adopted in video copy
detection techniques include discovering whether a query
video is copied from a database of videos or not. This paper
acquaints different Video copy detection techniques that have
been adopted to ensure robust and secure videos along some
applications of video fingerprinting.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
consumers use a smartphone device to display the media contents for work and entertainment purposes, as well as watching online video. Online video streaming is the main cause that consume smartphone’s energy quickly. To overcome this problem, smartphone’s energy management is crucial. Thus, a hybrid energy-aware profiler is proposed. Basically, a profiler will monitor and manage the energy consumption in the smartphone devices. The hybrid energy-aware profiler will set up a protocol preference of both the user and the device. Then, it will estimates the energy consumption in smartphone. However, saving energy alone can contribute to the Quality of Experience (QoE) neglection, thus the proposed solution takes into account the client QoE. Even though there are several existing energy-aware profilers that have been developed to manage energy use in smartphones however, most energy-aware profilers does not consider QoE at the same time. The proposed solution consider both, the performance of the hybrid energy-aware profiler is compared with the baseline energy models against a variation of content adaptation according to the pre-defined variables. Three types of variables were determined; resolution, frame rate and energy consumption in smartphone devices. In this area, QoE subjective methods based on MOS (Mean Opinion Score) are the most commonly used approaches for defining and quantifying real video quality. Nevertheless, although these approaches have been established to consistently quantify users’ amounts of approval, they do not adequately realize which the criteria of video attribute that important are. In this paper, we conducted an experiment with a certain devices to measures user’s QoE and energy usage of video attribute in smartphone devices. Our results demonstrate that the list of possible solution is a relevant and useful video attribute that satify the users.
Abstract Recently, Video is becoming a catholic medium for e-learning. As per the popularity of online video information over the World Wide Web (WWW) is mostly dependent on user-assigned tags or specification, which is the system by which we can access such videos. However, this system have limitations for retrieval and frequently we want access to the content (pacify) of the video itself is directly matched against a user’s query except manually assigned tags or specifications. In e-lecturing videos it contains visual and aural medium: slides of presentation and speech. in this system, we are going to retrieve the text from the videos automatically. To abstract visible information, we apply video content analysis to detect slides and optical character recognition to obtain their text. We abstract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames and Automatic Speech Recognition (ASR) on lecture audio. The ASR and OCR translate and discover slide text line types are accept for keywords abstraction, in which video and fragment-level keywords are abstracted for video searching on the basis of contents. .Key Words: Video fragmentation, Frame Abstraction, video indexing, and etc
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Efficient and Robust Detection of Duplicate Videos in a Databaserahulmonikasharma
In this paper, the duplicate detection method is to retrieve the best matching model video for a given query video using fingerprint. We have used the Color Layout Descriptor method and Opponent Color Space to extract feature from frame and perform k-means based clustering to generate fingerprints which are further encoded by Vector Quantization. The model-to-query video distance is computed using a new distance measure to find the similarity. To perform efficient search coarse-to-fine matching scheme is used to retrieve best match. We perform experiments on query videos and real time video with an average duration of 60 sec; the duplicate video is detected with high similarity.
Efficient and Robust Detection of Duplicate Videos in a Databaserahulmonikasharma
In this paper, the duplicate detection method is to retrieve the best matching model video for a given query video using fingerprint. We have used the Color Layout Descriptor method and Opponent Color Space to extract feature from frame and perform k-means based clustering to generate fingerprints which are further encoded by Vector Quantization. The model-to-query video distance is computed using a new distance measure to find the similarity. To perform efficient search coarse-to-fine matching scheme is used to retrieve best match. We perform experiments on query videos and real time video with an average duration of 60 sec; the duplicate video is detected with high similarity.
Content Based Video Retrieval Using Integrated Feature Extraction and Persona...IJERD Editor
Traditional video retrieval methods fail to meet technical challenges due to large and rapid growth of
multimedia data, demanding effective retrieval systems. In the last decade Content Based Video Retrieval
(CBVR) has become more and more popular. The amount of lecture video data on the Worldwide Web (WWW)
is growing rapidly. Therefore, a more efficient method for video retrieval in WWW or within large lecture video
archives is urgently needed. This paper presents an implementation of automated video indexing and video
search in large videodatabase. First of all, we apply automatic video segmentation and key-frame detection to
extract the frames from video. At next, we extract textual keywords by applying on video i.e. Optical Character
Recognition (OCR) technology on key-frames and Automatic Speech Recognition (ASR) on audio tracks of that
video. At next, we also extractingcolour, texture and edge detector features from different method. At last, we
integrate all the keywords and features which has extracted from above techniques for searching
purpose.Finallysearch similarity measure is applied to retrieve the best matchingcorresponding videos are
presented as output from database. Additionally we are providing Re-ranking of results as per users interest in
original result.
An Exploration based on Multifarious Video Copy Detection Strategiesidescitation
We co-exist in an era, where tonnes and tonnes of
videos are uploaded every day. Video copy detection has become
the need for the hour as most of them are user generated
Internet videos through popular sites such as YouTube. It acts
as a medium to restrain piracy and prove whether the contents
are legitimate. The usual procedure adopted in video copy
detection techniques include discovering whether a query
video is copied from a database of videos or not. This paper
acquaints different Video copy detection techniques that have
been adopted to ensure robust and secure videos along some
applications of video fingerprinting.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
consumers use a smartphone device to display the media contents for work and entertainment purposes, as well as watching online video. Online video streaming is the main cause that consume smartphone’s energy quickly. To overcome this problem, smartphone’s energy management is crucial. Thus, a hybrid energy-aware profiler is proposed. Basically, a profiler will monitor and manage the energy consumption in the smartphone devices. The hybrid energy-aware profiler will set up a protocol preference of both the user and the device. Then, it will estimates the energy consumption in smartphone. However, saving energy alone can contribute to the Quality of Experience (QoE) neglection, thus the proposed solution takes into account the client QoE. Even though there are several existing energy-aware profilers that have been developed to manage energy use in smartphones however, most energy-aware profilers does not consider QoE at the same time. The proposed solution consider both, the performance of the hybrid energy-aware profiler is compared with the baseline energy models against a variation of content adaptation according to the pre-defined variables. Three types of variables were determined; resolution, frame rate and energy consumption in smartphone devices. In this area, QoE subjective methods based on MOS (Mean Opinion Score) are the most commonly used approaches for defining and quantifying real video quality. Nevertheless, although these approaches have been established to consistently quantify users’ amounts of approval, they do not adequately realize which the criteria of video attribute that important are. In this paper, we conducted an experiment with a certain devices to measures user’s QoE and energy usage of video attribute in smartphone devices. Our results demonstrate that the list of possible solution is a relevant and useful video attribute that satify the users.
Abstract Recently, Video is becoming a catholic medium for e-learning. As per the popularity of online video information over the World Wide Web (WWW) is mostly dependent on user-assigned tags or specification, which is the system by which we can access such videos. However, this system have limitations for retrieval and frequently we want access to the content (pacify) of the video itself is directly matched against a user’s query except manually assigned tags or specifications. In e-lecturing videos it contains visual and aural medium: slides of presentation and speech. in this system, we are going to retrieve the text from the videos automatically. To abstract visible information, we apply video content analysis to detect slides and optical character recognition to obtain their text. We abstract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames and Automatic Speech Recognition (ASR) on lecture audio. The ASR and OCR translate and discover slide text line types are accept for keywords abstraction, in which video and fragment-level keywords are abstracted for video searching on the basis of contents. .Key Words: Video fragmentation, Frame Abstraction, video indexing, and etc
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Efficient and Robust Detection of Duplicate Videos in a Databaserahulmonikasharma
In this paper, the duplicate detection method is to retrieve the best matching model video for a given query video using fingerprint. We have used the Color Layout Descriptor method and Opponent Color Space to extract feature from frame and perform k-means based clustering to generate fingerprints which are further encoded by Vector Quantization. The model-to-query video distance is computed using a new distance measure to find the similarity. To perform efficient search coarse-to-fine matching scheme is used to retrieve best match. We perform experiments on query videos and real time video with an average duration of 60 sec; the duplicate video is detected with high similarity.
Efficient and Robust Detection of Duplicate Videos in a Databaserahulmonikasharma
In this paper, the duplicate detection method is to retrieve the best matching model video for a given query video using fingerprint. We have used the Color Layout Descriptor method and Opponent Color Space to extract feature from frame and perform k-means based clustering to generate fingerprints which are further encoded by Vector Quantization. The model-to-query video distance is computed using a new distance measure to find the similarity. To perform efficient search coarse-to-fine matching scheme is used to retrieve best match. We perform experiments on query videos and real time video with an average duration of 60 sec; the duplicate video is detected with high similarity.
Content Based Video Retrieval Using Integrated Feature Extraction and Persona...IJERD Editor
Traditional video retrieval methods fail to meet technical challenges due to large and rapid growth of
multimedia data, demanding effective retrieval systems. In the last decade Content Based Video Retrieval
(CBVR) has become more and more popular. The amount of lecture video data on the Worldwide Web (WWW)
is growing rapidly. Therefore, a more efficient method for video retrieval in WWW or within large lecture video
archives is urgently needed. This paper presents an implementation of automated video indexing and video
search in large videodatabase. First of all, we apply automatic video segmentation and key-frame detection to
extract the frames from video. At next, we extract textual keywords by applying on video i.e. Optical Character
Recognition (OCR) technology on key-frames and Automatic Speech Recognition (ASR) on audio tracks of that
video. At next, we also extractingcolour, texture and edge detector features from different method. At last, we
integrate all the keywords and features which has extracted from above techniques for searching
purpose.Finallysearch similarity measure is applied to retrieve the best matchingcorresponding videos are
presented as output from database. Additionally we are providing Re-ranking of results as per users interest in
original result.
Content based video retrieval using discrete cosine transformnooriasukmaningtyas
A content based video retrieval (CBVR) framework is built in this paper.
One of the essential features of video retrieval process and CBVR is a color
value. The discrete cosine transform (DCT) is used to extract a query video
features to compare with the video features stored in our database. Average
result of 0.6475 was obtained by using the DCT after implementing it to the
database we created and collected, and on all categories. This technique was
applied on our database of video, check 100 database videos, 5 videos in
Keywords: each category.
Mechanical properties of hybrid fiber reinforced concrete for pavementseSAT Journals
Abstract
The effect of addition of mono fibers and hybrid fibers on the mechanical properties of concrete mixture is studied in the present
investigation. Steel fibers of 1% and polypropylene fibers 0.036% were added individually to the concrete mixture as mono fibers and
then they were added together to form a hybrid fiber reinforced concrete. Mechanical properties such as compressive, split tensile and
flexural strength were determined. The results show that hybrid fibers improve the compressive strength marginally as compared to
mono fibers. Whereas, hybridization improves split tensile strength and flexural strength noticeably.
Keywords:-Hybridization, mono fibers, steel fiber, polypropylene fiber, Improvement in mechanical properties.
Material management in construction – a case studyeSAT Journals
Abstract
The objective of the present study is to understand about all the problems occurring in the company because of improper application
of material management. In construction project operation, often there is a project cost variance in terms of the material, equipments,
manpower, subcontractor, overhead cost, and general condition. Material is the main component in construction projects. Therefore,
if the material management is not properly managed it will create a project cost variance. Project cost can be controlled by taking
corrective actions towards the cost variance. Therefore a methodology is used to diagnose and evaluate the procurement process
involved in material management and launch a continuous improvement was developed and applied. A thorough study was carried
out along with study of cases, surveys and interviews to professionals involved in this area. As a result, a methodology for diagnosis
and improvement was proposed and tested in selected projects. The results obtained show that the main problem of procurement is
related to schedule delays and lack of specified quality for the project. To prevent this situation it is often necessary to dedicate
important resources like money, personnel, time, etc. To monitor and control the process. A great potential for improvement was
detected if state of the art technologies such as, electronic mail, electronic data interchange (EDI), and analysis were applied to the
procurement process. These helped to eliminate the root causes for many types of problems that were detected.
Managing drought short term strategies in semi arid regions a case studyeSAT Journals
Abstract
Drought management needs multidisciplinary action. Interdisciplinary efforts among the experts in various fields of the droughts
prone areas are helpful to achieve tangible and permanent solution for this recurring problem. The Gulbarga district having the total
area around 16, 240 sq.km, and accounts 8.45 per cent of the Karnataka state area. The district has been situated with latitude 17º 19'
60" North and longitude of 76 º 49' 60" east. The district is situated entirely on the Deccan plateau positioned at a height of 300 to
750 m above MSL. Sub-tropical, semi-arid type is one among the drought prone districts of Karnataka State. The drought
management is very important for a district like Gulbarga. In this paper various short term strategies are discussed to mitigate the
drought condition in the district.
Keywords: Drought, South-West monsoon, Semi-Arid, Rainfall, Strategies etc.
Life cycle cost analysis of overlay for an urban road in bangaloreeSAT Journals
Abstract
Pavements are subjected to severe condition of stresses and weathering effects from the day they are constructed and opened to traffic
mainly due to its fatigue behavior and environmental effects. Therefore, pavement rehabilitation is one of the most important
components of entire road systems. This paper highlights the design of concrete pavement with added mono fibers like polypropylene,
steel and hybrid fibres for a widened portion of existing concrete pavement and various overlay alternatives for an existing
bituminous pavement in an urban road in Bangalore. Along with this, Life cycle cost analyses at these sections are done by Net
Present Value (NPV) method to identify the most feasible option. The results show that though the initial cost of construction of
concrete overlay is high, over a period of time it prove to be better than the bituminous overlay considering the whole life cycle cost.
The economic analysis also indicates that, out of the three fibre options, hybrid reinforced concrete would be economical without
compromising the performance of the pavement.
Keywords: - Fatigue, Life cycle cost analysis, Net Present Value method, Overlay, Rehabilitation
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materialseSAT Journals
Abstract
The issue of growing demand on our nation’s roadways over that past couple of decades, decreasing budgetary funds, and the need to
provide a safe, efficient, and cost effective roadway system has led to a dramatic increase in the need to rehabilitate our existing
pavements and the issue of building sustainable road infrastructure in India. With these emergency of the mentioned needs and this
are today’s burning issue and has become the purpose of the study.
In the present study, the samples of existing bituminous layer materials were collected from NH-48(Devahalli to Hassan) site.The
mixtures were designed by Marshall Method as per Asphalt institute (MS-II) at 20% and 30% Reclaimed Asphalt Pavement (RAP).
RAP material was blended with virgin aggregate such that all specimens tested for the, Dense Bituminous Macadam-II (DBM-II)
gradation as per Ministry of Roads, Transport, and Highways (MoRT&H) and cost analysis were carried out to know the economics.
Laboratory results and analysis showed the use of recycled materials showed significant variability in Marshall Stability, and the
variability increased with the increase in RAP content. The saving can be realized from utilization of recycled materials as per the
methodology, the reduction in the total cost is 19%, 30%, comparing with the virgin mixes.
Keywords: Reclaimed Asphalt Pavement, Marshall Stability, MS-II, Dense Bituminous Macadam-II
Laboratory investigation of expansive soil stabilized with natural inorganic ...eSAT Journals
Abstract
Soil stabilization has proven to be one of the oldest techniques to improve the soil properties. Literature review conducted revealed
that uses of natural inorganic stabilizers are found to be one of the best options for soil stabilization. In this regard an attempt has
been made to evaluate the influence of RBI-81 stabilizer on properties of black cotton soil through laboratory investigations. Black
cotton soil with varying percentages of RBI-81 viz., 0, 0.5, 1, 1.5, 2, and 2.5 percent were studied for moisture density relationships
and strength behaviour of soils. Also the effect of curing period was evaluated as literature review clearly emphasized the strength
gain of soils stabilized with RBI-81 over a period of time. The results obtained shows that the unconfined compressive strength of
specimens treated with RBI-81 increased approximately by 250% for a curing period of 28 days as compared to virgin soil. Further
the CBR value improved approximately by 400%. The studies indicated an increasing trend for soil strength behaviour with
increasing percentage of RBI-81 suggesting its potential applications in soil stabilization.
Influence of reinforcement on the behavior of hollow concrete block masonry p...eSAT Journals
Abstract
Reinforced masonry was developed to exploit the strength potential of masonry and to solve its lack of tensile strength. Experimental
and analytical studies have been carried out to investigate the effect of reinforcement on the behavior of hollow concrete block
masonry prisms under compression and to predict ultimate failure compressive strength. In the numerical program, three dimensional
non-linear finite elements (FE) model based on the micro-modeling approach is developed for both unreinforced and reinforced
masonry prisms using ANSYS (14.5). The proposed FE model uses multi-linear stress-strain relationships to model the non-linear
behavior of hollow concrete block, mortar, and grout. Willam-Warnke’s five parameter failure theory has been adopted to model the
failure of masonry materials. The comparison of the numerical and experimental results indicates that the FE models can successfully
capture the highly nonlinear behavior of the physical specimens and accurately predict their strength and failure mechanisms.
Keywords: Structural masonry, Hollow concrete block prism, grout, Compression failure, Finite element method,
Numerical modeling.
Influence of compaction energy on soil stabilized with chemical stabilizereSAT Journals
Abstract
Increase in traffic along with heavier magnitude of wheel loads cause rapid deterioration in pavements. There is a need to improve
density, strength of soil subgrade and other pavement layers. In this study an attempt is made to improve the properties of locally
available loamy soil using twin approaches viz., i) increasing the compaction of soil and ii) treating the soil with chemical stabilizer.
Laboratory studies are carried out on both untreated and treated soil samples compacted by different compaction efforts. Studies
show that increase in compaction effort results in increase in density of soil. However in soil treated with chemical stabilizer, rate of
increase in density is not significant. The soil treated with chemical stabilizer exhibits improvement in both strength and performance
properties.
Keywords: compaction, density, subgradestabilization, resilient modulus
Geographical information system (gis) for water resources managementeSAT Journals
Abstract
Water resources projects are inherited with overlapping and at times conflicting objectives. These projects are often of varied sizes
ranging from major projects with command areas of millions of hectares to very small projects implemented at the local level. Thus,
in all these projects there is seldom proper coordination which is essential for ensuring collective sustainability.
Integrated watershed development and management is the accepted answer but in turn requires a comprehensive framework that can
enable planning process involving all the stakeholders at different levels and scales is compulsory. Such a unified hydrological
framework is essential to evaluate the cause and effect of all the proposed actions within the drainage basins.
The present paper describes a hydrological framework developed in the form of a Hydrologic Information System (HIS) which is
intended to meet the specific information needs of the various line departments of a typical State connected with water related aspects.
The HIS consist of a hydrologic information database coupled with tools for collating primary and secondary data and tools for
analyzing and visualizing the data and information. The HIS also incorporates hydrological model base for indirect assessment of
various entities of water balance in space and time. The framework would be maintained and updated to reflect fully the most
accurate ground truth data and the infrastructure requirements for planning and management.
Keywords: Hydrological Information System (HIS); WebGIS; Data Model; Web Mapping Services
Forest type mapping of bidar forest division, karnataka using geoinformatics ...eSAT Journals
Abstract
The study demonstrate the potentiality of satellite remote sensing technique for the generation of baseline information on forest types
including tree plantation details in Bidar forest division, Karnataka covering an area of 5814.60Sq.Kms. The Total Area of Bidar
forest division is 5814Sq.Kms analysis of the satellite data in the study area reveals that about 84% of the total area is Covered by
crop land, 1.778% of the area is covered by dry deciduous forest, 1.38 % of mixed plantation, which is very threatening to the
environmental stability of the forest, future plantation site has been mapped. With the use of latest Geo-informatics technology proper
and exact condition of the trees can be observed and necessary precautions can be taken for future plantation works in an appropriate
manner
Keywords:-RS, GIS, GPS, Forest Type, Tree Plantation
Factors influencing compressive strength of geopolymer concreteeSAT Journals
Abstract
To study effects of several factors on the properties of fly ash based geopolymer concrete on the compressive strength and also the
cost comparison with the normal concrete. The test variables were molarities of sodium hydroxide(NaOH) 8M,14M and 16M, ratio of
NaOH to sodium silicate (Na2SiO3) 1, 1.5, 2 and 2.5, alkaline liquid to fly ash ratio 0.35 and 0.40 and replacement of water in
Na2SiO3 solution by 10%, 20% and 30% were used in the present study. The test results indicated that the highest compressive
strength 54 MPa was observed for 16M of NaOH, ratio of NaOH to Na2SiO3 2.5 and alkaline liquid to fly ash ratio of 0.35. Lowest
compressive strength of 27 MPa was observed for 8M of NaOH, ratio of NaOH to Na2SiO3 is 1 and alkaline liquid to fly ash ratio of
0.40. Alkaline liquid to fly ash ratio of 0.35, water replacement of 10% and 30% for 8 and 16 molarity of NaOH and has resulted in
compressive strength of 36 MPa and 20 MPa respectively. Superplasticiser dosage of 2 % by weight of fly ash has given higher
strength in all cases.
Keywords: compressive strength, alkaline liquid, fly ash
Experimental investigation on circular hollow steel columns in filled with li...eSAT Journals
Abstract
Composite Circular hollow Steel tubes with and without GFRP infill for three different grades of Light weight concrete are tested for
ultimate load capacity and axial shortening , under Cyclic loading. Steel tubes are compared for different lengths, cross sections and
thickness. Specimens were tested separately after adopting Taguchi’s L9 (Latin Squares) Orthogonal array in order to save the initial
experimental cost on number of specimens and experimental duration. Analysis was carried out using ANN (Artificial Neural
Network) technique with the assistance of Mini Tab- a statistical soft tool. Comparison for predicted, experimental & ANN output is
obtained from linear regression plots. From this research study, it can be concluded that *Cross sectional area of steel tube has most
significant effect on ultimate load carrying capacity, *as length of steel tube increased- load carrying capacity decreased & *ANN
modeling predicted acceptable results. Thus ANN tool can be utilized for predicting ultimate load carrying capacity for composite
columns.
Keywords: Light weight concrete, GFRP, Artificial Neural Network, Linear Regression, Back propagation, orthogonal
Array, Latin Squares
Experimental behavior of circular hsscfrc filled steel tubular columns under ...eSAT Journals
Abstract
This paper presents an outlook on experimental behavior and a comparison with predicted formula on the behaviour of circular
concentrically loaded self-consolidating fibre reinforced concrete filled steel tube columns (HSSCFRC). Forty-five specimens were
tested. The main parameters varied in the tests are: (1) percentage of fiber (2) tube diameter or width to wall thickness ratio (D/t
from 15 to 25) (3) L/d ratio from 2.97 to 7.04 the results from these predictions were compared with the experimental data. The
experimental results) were also validated in this study.
Keywords: Self-compacting concrete; Concrete-filled steel tube; axial load behavior; Ultimate capacity.
Evaluation of punching shear in flat slabseSAT Journals
Abstract
Flat-slab construction has been widely used in construction today because of many advantages that it offers. The basic philosophy in
the design of flat slab is to consider only gravity forces; this method ignores the effect of punching shear due to unbalanced moments
at the slab column junction which is critical. An attempt has been made to generate generalized design sheets which accounts both
punching shear due to gravity loads and unbalanced moments for cases (a) interior column; (b) edge column (bending perpendicular
to shorter edge); (c) edge column (bending parallel to shorter edge); (d) corner column. These design sheets are prepared as per
codal provisions of IS 456-2000. These design sheets will be helpful in calculating the shear reinforcement to be provided at the
critical section which is ignored in many design offices. Apart from its usefulness in evaluating punching shear and the necessary
shear reinforcement, the design sheets developed will enable the designer to fix the depth of flat slab during the initial phase of the
design.
Keywords: Flat slabs, punching shear, unbalanced moment.
Evaluation of performance of intake tower dam for recent earthquake in indiaeSAT Journals
Abstract
Intake towers are typically tall, hollow, reinforced concrete structures and form entrance to reservoir outlet works. A parametric
study on dynamic behavior of circular cylindrical towers can be carried out to study the effect of depth of submergence, wall thickness
and slenderness ratio, and also effect on tower considering dynamic analysis for time history function of different soil condition and
by Goyal and Chopra accounting interaction effects of added hydrodynamic mass of surrounding and inside water in intake tower of
dam
Key words: Hydrodynamic mass, Depth of submergence, Reservoir, Time history analysis,
Evaluation of operational efficiency of urban road network using travel time ...eSAT Journals
Abstract
Efficiency of the road network system is analyzed by travel time reliability measures. The study overlooks on an important measure of
travel time reliability and prioritizing Tiruchirappalli road network. Traffic volume and travel time were collected using license plate
matching method. Travel time measures were estimated from average travel time and 95th travel time. Effect of non-motorized vehicle
on efficiency of road system was evaluated. Relation between buffer time index and traffic volume was created. Travel time model has
been developed and travel time measure was validated. Then service quality of road sections in network were graded based on
travel time reliability measures.
Keywords: Buffer Time Index (BTI); Average Travel Time (ATT); Travel Time Reliability (TTR); Buffer Time (BT).
Estimation of surface runoff in nallur amanikere watershed using scs cn methodeSAT Journals
Abstract
The development of watershed aims at productive utilization of all the available natural resources in the entire area extending from
ridge line to stream outlet. The per capita availability of land for cultivation has been decreasing over the years. Therefore, water and
the related land resources must be developed, utilized and managed in an integrated and comprehensive manner. Remote sensing and
GIS techniques are being increasingly used for planning, management and development of natural resources. The study area, Nallur
Amanikere watershed geographically lies between 110 38’ and 110 52’ N latitude and 760 30’ and 760 50’ E longitude with an area of
415.68 Sq. km. The thematic layers such as land use/land cover and soil maps were derived from remotely sensed data and overlayed
through ArcGIS software to assign the curve number on polygon wise. The daily rainfall data of six rain gauge stations in and around
the watershed (2001-2011) was used to estimate the daily runoff from the watershed using Soil Conservation Service - Curve Number
(SCS-CN) method. The runoff estimated from the SCS-CN model was then used to know the variation of runoff potential with different
land use/land cover and with different soil conditions.
Keywords: Watershed, Nallur watershed, Surface runoff, Rainfall-Runoff, SCS-CN, Remote Sensing, GIS.
Estimation of morphometric parameters and runoff using rs & gis techniqueseSAT Journals
Abstract
Land and water are the two vital natural resources, the optimal management of these resources with minimum adverse environmental
impact are essential not only for sustainable development but also for human survival. Satellite remote sensing with geographic
information system has a pragmatic approach to map and generate spatial input layers of predicting response behavior and yield of
watershed. Hence, in the present study an attempt has been made to understand the hydrological process of the catchment at the
watershed level by drawing the inferences from moprhometric analysis and runoff. The study area chosen for the present study is
Yagachi catchment situated in Chickamaglur and Hassan district lies geographically at a longitude 75⁰52’08.77”E and
13⁰10’50.77”N latitude. It covers an area of 559.493 Sq.km. Morphometric analysis is carried out to estimate morphometric
parameters at Micro-watershed to understand the hydrological response of the catchment at the Micro-watershed level. Daily runoff
is estimated using USDA SCS curve number model for a period of 10 years from 2001 to 2010. The rainfall runoff relationship of the
study shows there is a positive correlation.
Keywords: morphometric analysis, runoff, remote sensing and GIS, SCS - method
-
Effect of variation of plastic hinge length on the results of non linear anal...eSAT Journals
Abstract The nonlinear Static procedure also well known as pushover analysis is method where in monotonically increasing loads are applied to the structure till the structure is unable to resist any further load. It is a popular tool for seismic performance evaluation of existing and new structures. In literature lot of research has been carried out on conventional pushover analysis and after knowing deficiency efforts have been made to improve it. But actual test results to verify the analytically obtained pushover results are rarely available. It has been found that some amount of variation is always expected to exist in seismic demand prediction of pushover analysis. Initial study is carried out by considering user defined hinge properties and default hinge length. Attempt is being made to assess the variation of pushover analysis results by considering user defined hinge properties and various hinge length formulations available in literature and results compared with experimentally obtained results based on test carried out on a G+2 storied RCC framed structure. For the present study two geometric models viz bare frame and rigid frame model is considered and it is found that the results of pushover analysis are very sensitive to geometric model and hinge length adopted. Keywords: Pushover analysis, Base shear, Displacement, hinge length, moment curvature analysis
Effect of use of recycled materials on indirect tensile strength of asphalt c...eSAT Journals
Abstract
Depletion of natural resources and aggregate quarries for the road construction is a serious problem to procure materials. Hence
recycling or reuse of material is beneficial. On emphasizing development in sustainable construction in the present era, recycling of
asphalt pavements is one of the effective and proven rehabilitation processes. For the laboratory investigations reclaimed asphalt
pavement (RAP) from NH-4 and crumb rubber modified binder (CRMB-55) was used. Foundry waste was used as a replacement to
conventional filler. Laboratory tests were conducted on asphalt concrete mixes with 30, 40, 50, and 60 percent replacement with RAP.
These test results were compared with conventional mixes and asphalt concrete mixes with complete binder extracted RAP
aggregates. Mix design was carried out by Marshall Method. The Marshall Tests indicated highest stability values for asphalt
concrete (AC) mixes with 60% RAP. The optimum binder content (OBC) decreased with increased in RAP in AC mixes. The Indirect
Tensile Strength (ITS) for AC mixes with RAP also was found to be higher when compared to conventional AC mixes at 300C.
Keywords: Reclaimed asphalt pavement, Foundry waste, Recycling, Marshall Stability, Indirect tensile strength.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Student information management system project report ii.pdfKamal Acharya
Our project explains about the student management. This project mainly explains the various actions related to student details. This project shows some ease in adding, editing and deleting the student details. It also provides a less time consuming process for viewing, adding, editing and deleting the marks of the students.
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
Review on content based video lecture retrieval
1. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 306
REVIEW ON CONTENT BASED VIDEO LECTURE RETRIEVAL
Deshmukh Bhagyashri D1
1
PVPIT, M.E. Comp. Engg., Savitribai Phule Pune University, India
Abstract
Recent advances in multimedia technologies allow the capture and storage of video data with relatively inexpensive computers.
Furthermore, the new possibilities offered by the information highways have made a large amount of video data publicly
available. However, without appropriate search techniques all these data are hardly usable. Users are not satisfied with the video
retrieval systems that provide analogue VCR functionality. For example, a user analyses a soccer video will ask for specific
events such as goals. Content-based search and retrieval of video data becomes a challenging and important problem. Therefore,
the need for tools that can be manipulate the video content in the same way as traditional databases manage numeric and textual
data is significant. Therefore, a more efficient method for video retrieval in WWW or within large lecture video archives is
urgently needed. This project presents an approach for automated video indexing and video search in large lecture video
archives. First of all, we apply automatic video segmentation and key-frame detection to offer a visual guideline for the video
content navigation. Subsequently, we extract textual metadata by applying video Optical Character Recognition (OCR)
technology on key-frames and Automatic Speech Recognition on lecture audio tracks.
Keywords—Feature extraction, video annotation, video browsing, video retrieval, video structure analysis.
-------------------------------------------------------------------***-------------------------------------------------------------------
1. INTRODUCTION
In the last decade e-lecturing has become more and more
popular. The more college students are trying to interact
with and trying to learn from this e-lecture. Digital videos
are becoming a popular storage and exchange medium due
to the rapid development in recording technology. A number
of universities and research institutions are taking the
opportunity to record their lectures and publish them online
for students to access independent of time and location. As a
result, there has been a huge amount of multimedia data on
the Web which is very difficult for user to judge whether a
video is useful by only glancing at the title and find desired
videos without a search function within video archive. The
user might thus want to find the piece of information he
requires without viewing the complete video. The problem
becomes how to retrieve the appropriate information in a
large lecture video archive more efficiently. The manually
provided metadata is typically brief, high level and
subjective. The next generation of video retrieval systems
apply automatically generated metadata by using video
analysis technologies.
Hypothesis: The relevant metadata can be automatically
gathered from lecture videos by using appropriate analysis
techniques. They can help a user to find and to understand
lecture contents more efficiently, and the learning
effectiveness can thus be improved.
Our research work mainly focus on those lecture videos
produced by using the screen grabbing method. Segmenting
two-scenes lecture videos can be achieved by only
processing slide video streams, which contain most of the
visual text metadata.
Key frame recognition by OCR technology:- The OCR
technique is nothing but the Optical Character Recognition
.The OCR technique is used for extracting textual metadata
.By using OCR we can convert different type of documents
such as paper document and image capture by digital camera
into editable and searchable data. With OCR the recognized
document looks like the original. The OCR software allows
you to save a lot of time and efforts when creating and
processing and repurchasing various document. The search
indices are created based on the global metadata obtained
from the video hosting website and texts extracted from
slide videos by using a standard OCR engine.
Speech recognition by AUTOMATIC SPEECH
RECOGNITION technique: The AUTOMATIC SPEECH
RECOGNITION technique is nothing but the Automatic
Speech Recognition .A can provide speech to text
information from lecture video .in computer science speech
recognization is the translation of spoken word into text .It is
also known as automatic speech recognization. Speech is
one of the most important carriers of information in video
lectures. The poor recognition results not only limit the
usability of speech transcript, but also affect the efficiency
of the further indexing process.
In our research, we intended to continuously improve the
AUTOMATIC SPEECH RECOGNITION result based on
the open-source AUTOMATIC SPEECH RECOGNITION
tool.
A large amount of textual metadata will be created by using
OCR and AUTOMATIC SPEECH RECOGNITION
method, which opens up the content of lecture videos. We
extract metadata from visual as well as audio resources of
lecture videos automatically by applying appropriate
analysis techniques. For evaluation purposes we developed
several automatic indexing functionalities in a large lecture
video portal, which can guide both visually- and text-
oriented users to navigate within lecture video.
2. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 307
We investigate the usability and the effectiveness of
proposed video indexing features. For visual analysis, we
propose a new method for slide video segmentation and
apply video OCR to gather text metadata. A more flexible
search function has been developed based on the structured
video text. In order to overcome the solidity and consistency
problems of a content-based video search system, we
propose a keyword ranking method for multimodal
information resources. In order to evaluate the usability, we
implemented this approach in a large lecture video portal. In
summary, the major contributions of this paper are the
following:
a) We extract metadata from visual as well as audio
resources of lecture videos automatically by applying
appropriate analysis techniques. For evaluation
purposes we developed several automatic indexing
functionalities in a large lecture video portal, which
can guide both visually- and text-oriented users to
navigate within lecture video. We conducted a user
study intended to verify the research hypothesis and to
investigate the usability and the effectiveness of
proposed video indexing features.
b) For visual analysis, we propose a new method for
slide video segmentation and apply video OCR to
gather text metadata. Furthermore, lecture outline is
extracted from OCR transcripts by using stroke
c) We propose a solution for automatic German
phonetic dictionary generation, which fills the gap in
open-source ASR domain. The dictionary software and
compiled speech corpus are provided for the further
research use.
d) In order to overcome the solidity and consistency
problems of a content-based video search system, we
propose a keyword ranking method for multimodal
information resources. In order to evaluate the
usability, we implemented this approach in a large
lecture video portal.
e) The developed video analysis methods have been
evaluated by using compiled test data sets as well as
opened benchmarks. All compiled test sets are publicly
available from our website for the further research
use.The rest of the paper is organized as follows:
Section 2 reviews related work in lecture video
retrieval and content-based video search
domain.Section3 describes our automatic video
indexing methods.
2. RELATED WORK
2.1 Lecture Video Retrieval
Wang et al. proposed an approach for lecture video indexing
based on automated video segmentation and OCR analysis
[9]. The proposed segmentation algorithm in their work is
based on the differential ratio of text and background
regions. Using thresholds they attempt to capture the slide
transition. The final segmentation results are determined by
synchronizing detected slide key-frames and related text
books, where the text similarity between them was
calculated as indicator. Grcar et al. introduced
videoLectures.net in [10] which is a digital archive for
multimedia presentations. Similar to [9], the authors also
apply a synchronization process between the recorded
lecture video and the slide file, which has to be provided by
presenters. Our system contrasts to these two approaches
since it directly analyzes the video, which is thus
independent of any hardware or presentation technology.
The constrained slide format and the synchronization with
an external document are not required. Furthermore, since
the animated content evolvement is often applied in the
slide, but has not been considered in [9] and [10], their
system might not work robustly when those effects occur in
the lecture video. In [9], the final segmentation result is
strongly dependent on the quality of the OCR result. It
might be less efficient and imply redundancies, when only
poor OCR result is obtained. Tuna et al. presented their
approach for lecture video indexing and search [11]. They
segment lecture videos into key frames by using global
frame differencing metrics. Then standard OCR software is
applied for gathering textual metadata from slide streams, in
which they utilize some image transformation techniques to
improve the OCR result. They developed a new video
player, in which the indexing, search and captioning
processes are integrated. Similar to [9], the used global
differencing metrics cannot give a sufficient segmentation
result when animations or content build-ups are used in the
slides. In that case, many redundant segments will be
created. Moreover, the used image transformations might be
still not efficient enough for recognizing frames with
complex content and background distributions. Making use
of text detection and segmentation procedures could achieve
much better results rather than applying image
transformations. Jeong et al. proposed a lecture video
segmentation method using Scale Invariant Feature
Transform (SIFT) feature and the adaptive threshold in [12].
In their work SIFT feature is applied to measure slides with
similar content. An adaptive threshold selection algorithm is
used to detect slide transitions. In their evaluation, this
approach achieved promising results for processing one-
scene lecture videos as illustrated in Fig. 1a.Recently,
collaborative tagging has become a popular functionality in
lecture video portals. Sack and Waitelon is[13] and Moritz
et al. [14] apply tagging data for lecture video retrieval and
video search. Beyond the keyword based tagging, Yu et al.
proposed an approach to annotate lecture video resources by
using Linked Data. Their framework enables users to
semantically annotate videos using vocabularies defined in
the Linked Data cloud. Then those semantically linked
educational resources are further adopted in the video
browsing and video recommendation procedures. However,
, the effort and cost needed by the user annotation-based
approach cannot satisfy the requirements for processing
large amounts of web video data with a rapid increasing
speed. Here, the automatic analysis is no doubt much more
suitable. Data to further automatically annotate the extracted
textual metadata opens a future research direction.ASR
provides speech-to-text information on spoken languages,
which is thus well suited for content-based lecture video
retrieval. The studies described in [5] and [15] are based on
out-of-the-box commercial speech recognition software.
3. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 308
Concerning such commercial software, to achieve satisfying
results for a special working domain an adaption process is
often required, but the custom extension is rarely possible.
The authors of [1] and [6] focus on English speech
recognition for Technology Entertainment and Design
(TED) lecture videos and webcasts. In their system, the
training dictionary is created manually, which is thus hard to
be extended or optimized periodically. Glass et al. proposed
a solution for improving ASR results of English lectures by
collecting new speech data from the rough lecture audio data
[3]. Inspired by their work, we developed an approach for
creating speech data from German lecture videos. Haubold
and Kender focus on multi-speaker presentation videos. In
their work speaker changes can be detected by applying a
speech analysis method. A topic phrases extraction
algorithm from highly imperfect ASR results (WER _ 75
percent) has been proposed by using lecture course-related
sources such as text books and slide files [4].Overall, most
of those lecture speech recognition systems have low
recognition rate, the WERs of audio lectures are
approximately 40-85 percent. The poor recognition results
limit the further indexing efficiency. Therefore, how to
continuously improve ASR accuracy for lecture videos is
still an unsolved problem. The speaker-gestures-based
information retrieval for lecture videos has been studied in
[16]. The author equipped the lecture speaker with special
gloves that enable the automatic detection and evaluation of
gestures. The experimental results show that 12 percent of
the lecture topic boundaries were correctly detected using
speaker-gestures. However, those gesture features are highly
dependent on the characteristics of speakers and topics. It
might have limited use in large lecture video archives with
massive amounts of speakers.
3. AUTOMATED LECTURE VIDEO INDEXING
In this chapter we will present four analysis processes for
retrieving relevant metadata from the two main parts of
lecture video, namely the visual screen and audio tracks.
From the visual screen we firstly detect the slide transitions
and extract each unique slide frame with its temporal scope
considered as the video segment. Then the video OCR
analysis is performed for retrieving textual metadata from
slide frames. Based on OCR results, we propose a novel
solution for lecture outline extraction by using stroke
width and geometric information of detected text lines.
In speech-to-text analysis we applied the open-source ASR
software CMU Sphinx.2 To build the acoustic and language
model, we collected speech training data from open-source
corpora and our lecture videos. As already mentioned, it
lacks method in open-source context for creating German
phonetic dictionary automatically. We thus developed a
solution to fill this gap and made it available for the further
research use.
3.1 Slide Video Segmentation
Video browsing can be achieved by segmenting video into
representative key frames. The selected key frames can pro-
vide a visual guideline for navigation in the lecture video
portal. Moreover, video segmentation and key-frame
selection is also often adopted as a pre-processing for other
analysis tasks such as video OCR, visual concept detection,
etc. often considered as a video segment. This can be
roughly determined by analyzing the temporal scope of
lecture. Many approaches (as, e.g., [17], [11]) make use of
global pixel-level-differencing metrics for capturing slide
transitions. A drawback of this kind of approach is that the
salt and pepper noise of video signal can affect the
segmentation accuracy. After observing the content of
lecture slides, we realize that the major content as, e.g., text
lines, figures, tables, etc., can be considered as Connected
Components (CCs). We therefore propose to use CC instead
of pixel as the basis element for the differencing analysis.
We call it component-level-differencing metric. This way
we are able to control the valid size of the CC, so that the
salt and pepper noise can be rejected from the differencing
process. For creating CCs from binary images, we apply the
algorithm according to which demonstrated an excellent
performance advantage. Another benefit of our segmentation
method is its robustness to animated content progressive
build-ups used within lecture slides. Only the most com-
plete unique slides are captured as video segments. Those
effects affect the most lecture video segmentation methods
mentioned in section2. Our segmentation method consists of
two steps (cf. Fig. 2):
In the first step, the entire slide video is analyzed. We try to
capture every knowledge change between adjacent frames,
for which we established an analysis interval of three
seconds by taking both accuracy and efficiency into account.
This means that segments with a duration smaller than three
seconds may be discarded in our system. Since there are
very few topic segments shorter than three seconds, this
setting is therefore not critical. Then we create canny edge
maps for adjacent frames and build the pixel differential
image from the edge maps. The CC analysis is subsequently
performed on this differential image and the number of CCs
is then used as a threshold for the segmentation. A new
segment is captured if the number exceeds Ts1. Here we
4. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 309
Fig.1: we detect the first and the last object line in Rc vertically and perform the CC differencing metric on those line
sequences from adjacent frames.
In frame (a) and (b), a same text line (top) can be found;
whereas in frame (b) and (c), the CC-based differences of
both text lines exceed the threshold Ts2. establish a
relatively small Ts1 to ensure that each newly emerging
knowledge point will be captured. Obviously, the first
segmentation step is sensitive to animations and progressive
build-ups. The result is thus too redundant for video
indexing. Hence, the process continues with the second
segmentation step based on the frames in the first step.
In the second segmentation step the real slide transitions will
be captured. The title and content region of a slide frame is
first defined. We established the con-tent distribution of
commonly used slide styles by analyzing a large amount of
lecture videos in our database. Let Rt and Rc denote the title
and content region which account for 23 and 70 percent of
the entire frame height respectively. In Rt we apply CC-
based differencing as described above with a small threshold
value of 1 for capturing the slide transi-tions. Here any small
changes within the title region may cause a slide transition.
For instance, two slides often differ from each other i n a
single chapter num-ber. If there is no difference found in Rt,
then we try to detect the first and the last bounding box
object in Rc vertically and perform the CC-based differenc-
ing within the object regions of two adjacent frames (cf. Fig.
2). In case that the difference value of both object regions
between adjacent frames exceed the threshold Ts2, a slide
transition is then captured. For detecting the content
progressive build-up horizon-tally, the process could be
repeated in a similar man-ner. In our experiment, Ts1 ¼ 20
and Ts2 ¼ 5 have proven to serve best for our training data.
Exact set-ting of these parameters is not critical.
Since the proposed method is designed for segment-ing slide
videos, it might be not suitable when videos with varying
genres have been embedded in the slides and are played
during the presentation. To solve this problem we have
extended the original algorithm by using a Support Vector
Machine (SVM) classier and image intensity histogram
features. We use the Radial Basis Function (RBF) as kernel.
In order to make a comparison, we also applied the
Histogram of Oriented Gradients (HOG) feature [20] in the
experiment. We adapted the HOG feature with eight
gradient directions, whereas the local region size was set to
64 _ 64. Moreover, in using this approach our video
segmentation method is also suitable for processing such
one-screen lecture videos with a frequent switch between
slide- and speaker-scene.
Fig. 2: Lecture video segmentation workflow. Step 1:
adjacent frames are compared with each other by
applying the CC analysis on their differential edge maps.
Step 2: slide transitions are captured by performing title
and content region analysis.
5. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 310
3.2 Video OCR for Lecture Videos
Texts in the lecture slides are closely related to the lecture
content, can thus provide important information for the
retrieval task. In our framework, we developed a novel
video OCR system for gathering video text.
For text detection, we developed a new localization-
verification scheme. In the detection stage, an edge-based
multi-scale text detector is used to quickly localize candidate
text regions with a low rejection rate. For the subsequent
text area verification, an image entropy-based adaptive
refinement algorithm not only serves to reject false positives
that expose low edge density, but also further splits the most
text- and non-text-regions into separate blocks. Then Stroke
Width Transform (SWT) [22]-based verification procedures
are applied to remove the non text blocks since the SWT
verifier is not able to correctly identify special non-text
patterns such as sphere, window-blocks, garden fence, we
adopted an additional SVM classifier to sort out these non-
text patterns in order to further improve the detection
accuracy. For text segmentation and recognition, we
developed a novel binarization approach, in which we utilize
image skeleton and edge maps to identify the text pixels.
The proposed method consists of three main steps: text
gradient direction analysis, seed pixel selection, and seed-
region growing. After the seed-region growing process, the
video text images are converted into a suitable format for
standard OCR engines. The subsequent spell-checking
process will further sort out incorrect words from the
recognition results. Our video OCR system has been
evaluated by applying several test data sets. Especially by
using the online evaluation of the opened benchmark
ICDAR 2011 competition test sets for born-digital images
[23], our text detection method achieved the second place,
and our text binarization and word recognition method
achieved the first place in the corresponding ranking list on
their website (last check 08/2013). An in-depth discussion of
the developed video OCR approach and the detailed
evaluation results can be found in [24]. By applying the
open-source print OCR engine tesseract-ocr,5 we achieved
recognition of 92 percent of all characters and 85 percent of
all words correctly for lecture video images. The compiled
test data set including 180 lecture video frames and respec-
tive manual annotations is available at [21].
Generally, in the lecture slide the content of title, subtitle
and key point have more significance than the normal slide
text, as they summarize each slide. Due to this fact, we
classify the type of text lines recognized from slide frames
by using geometrical information and stroke width feature.
The benefits of the structure analysis method can be sum-
marized as follows:
The lecture outline can be extracted using classified text
lines, it can provide a fast overview of a lecture video and
each outline item with the time stamp can in turn be adopted
for video browsing .
The structure of text lines can reflect their different
significance. This information is valuable for a
search/indexing engine. Similar to web search engines
which make use of the explicitly pre-defined HTML
structure (HTML-tags) for calculating the weight of texts in
web pages, our method further opens up the video content
and enables the search engine to give more accurate and
flexible search results based on the structured video text.
The process begins with a title line identification procedure.
A text line will be considered as a candidate title line when
it localizes in the upper third part of the frame, it has more
than three characters, it is one of three highest text lines and
has the uppermost vertical position. Then the corresponding
text line objects will be labeled as the title objects and we
repeat the process on the remaining text objects in the same
manner. The further detected title lines must have a similar
height (the tolerance is up to 10px) and stroke width value
(the tolerance is up to 5px) as the first one. For our purposes,
we allow up to three title lines to be detected for each slide
frame. All non-title text line objects are further classified
into three classes: content text, key-point and footline. The
classification is based on the height and the average stroke
width of the text line object, which is described as follows:
Keypoint if st > smean /ht >hmean
Footline if st < smean /ht <hmean /y =ymax
Content-text otherwise
where smean and hmean denote the average stroke width and
the average text line height of a slide frame, and ymax
denotes the maximum vertical position of a text line object
(starts from the top-left corner of the image).To further
extract the lecture outline, we firstly apply a spell checker to
sort out text line objects which do not satisfy the following
conditions:
_ a valid text line object must have more than three
characters,
_ a valid text line object must contain at least one
noun,
_ the textual character count of a valid text line object
must be more than 50 percent of the entire string
length.
The remaining objects will be labeled as the outline object.
Subsequently, we merge all title lines within the same slide
according to their position. Other text line objects from this
slide will be considered as the subitem of the title line. Then
we merge text line objects from adjacent frames with the
similar title content (with 90 percent content overlap).
The similarity is measured by calculating the amount of
same characters and same words. After a merging process
all duplicated text lines will be removed. Finally, the lecture
outline is created by assigning all valid outline objects into a
consecutive structure according to their occurrences.
In addition to video OCR, ASR can provide speech-to-text
information from lecture videos, which offers the chance to
improve the quantity of automatically generated metadata
6. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 311
dramatically. However, as mentioned, most lecture speech
recognition systems cannot achieve a sufficient recognition
rate. A model-adaption process is often required.
Furthermore, in the open-source context, it lacks method for
generating the German phonetic dictionary automatically.
Therefore, we developed a solution intended to fill this gap.
We decided to build acoustic models for our special use case
by applying the CMU Sphinx Toolkit6 and the German
Speech Corpus by Voxforge7 as a baseline. Unlike other
approaches that collect speech data by applying dictation in
a quiet environment, we gathered hours of speech data from
real lecture videos and created corresponding transcripts.
This way the real teaching environment can be involved in
the training process. For the language model training we
applied the collected text corpora from German daily news
corpus (radio programs, 1996-2000), Wortschatz- Leipzig8
and the audio transcripts of the collected speech corpora.
Fig. 5 depicts the workflow for creating the speech training
data. First of all, the recorded audio file is segmented into
smaller pieces and improper segments are sorted out.
For each remaining segment the spoken text is transcribed
manually, and added to the transcript file automatically. As
an intermediate step, a list of all used words in the transcript
file is created. In order to obtain the phonetic dictionary, the
pronunciation of each word has to be represented
phonetically.
A recorded lecture audio stream yields approximately 90
minutes of speech data, which is far too long to be processed
by the ASR trainer or the speech decoder at once. Shorter
speech segments are thus required. Manually collecting
appropriate speech segments and transcripts is rather time
consuming and costly. There are a number of steps to be
performed to acquire high quality input data for a ASR
trainer tool. Our current approach is to fully automate
segmentation (Fig. 6(2)) and partly automate selection (Fig.
6(3)) without suffering from quality drawbacks like dropped
word endings. The fundamental steps can be described as
follows: we first compute the absolute values of input
samples to get a loudness curve, which is then down
sampled (e.g., by factor 100). Then a blur filter (e.g.,radius
¼ 3) is applied to eliminate outliers and a loudness threshold
determines which areas are considered quiet.Non-quiet areas
with a specified maximum length (5 seconds has proven to
serve best for our training data) will serve as a potential
speech utterance. Finally, we retrieve speech utterances
from the original audio stream and save them into files.From
the experiments we have learned that all speech segments
used for the acoustic model training must meet certain
quality requirements. Experience has shown that
approximately 50 percent of the generated audio segments
have to be sorted out due to one of the following reasons:
_ the segment contains acoustical noise created by
objects and humans in the evironment around the
speaker, e.g., doors closing, chairs moving, students
talking,
_the lecturer mispronounces some words, so that they
are completely invalid from an objective point of view,
_ the speaker’s language is clear, but the segmentation
algorithm cuts off parts of a spoken word so that it
becomes invalid.
Therefore, the classification of audio segments in good or
bad quality is not yet solvable automatically, as the term
―good quality‖ is very subjective and strongly depends on
one’s personal perception. Nevertheless, the proposed
segmentation method can perform a preselection that speeds
up the manual transcription process significantly. The
experimental results show that the WER decreased by about
19 percent, when adding 7.2 hours of speech data from our
lecture videos to the training set.
3.3 Creation of the German Phonetic Dictionary
The phonetic dictionary is an essential part of every ASR
software. For each word that appears in the transcripts, it
defines one or more phonetic representations. As our speech
corpus is growing continuously, the extension and
maintenance of the dictionary becomes a common task. An
automatic generator is highly desired. Unfortunately it lacks
such a tool in the open-source context. Therefore, we have
built a phonetics generator by using a customized phonetic
alphabet, which contains 45 phonemes used in German
pronunciation.
We provide this tool and the compiled speech training data
for the further research use.
The generation algorithm operates on three layers: On word
level, we have defined a so-called exemption dictionary
which contains foreign words in particular and whose
pronunciation does not follow German rules, e.g., byte,
pheanomen.
First of all, a check is made as to whether the input word can
be found in the exemption dictionary. If this is the case, the
phonetic representation is completely read from this
dictionary and no further steps have to be performed.
Otherwise, we scale down to syllable level by applying an
extern hyphenation algorithm. On syllable level, we
examine the result of the hyphenation algorithm, as, e.g.,
computer for the input word computer. For many single
syllables (and also pairs of syllables),the syllable mapping
describes the corresponding phonetic representation, as, e.g.,
au f for the German prefix auf and n i: differ for the
disyllabic German prefix nieder.If there is such a
representation for our input syllable, then it is added to the
phonetic result immediately. Otherwise, we have to split the
syllable further into its characters and proceed on character
level.On character level, a set of German pronunciation rules
are applied to determine how the current single character is
pronounced, including character type (consonant or vowel),
neighbouring characters, relative position inside the
containing syllable, absolute position inside the whole word,
etc., First and foremost, a heuristic checks if the current and
the next 1–2 characters can be pronounced natively. If this is
not the case, or the word is only one character long, the
7. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 11 | Nov-2014, Available @ http://www.ijret.org 312
characters are pronounced as if they were spelled letter by
letter, as, e.g., the abbreviations abc (for alphabet) and ZDF
(a German TV channel). In the next step, we determine the
character type (consonant or vowel) in order to apply the
correct pronunciation rules. The conditions of each of these
rules are verified, until one is true or all conditions are
verified and proven to be false. If the latter is the case, the
standard pronunciation is applied, which assumes closed
vowels. An evaluation with 20,000 words from transcripts
shows that 98.1 percent of all input words were processed
correctly without any manual amendment. The 1.9 percent
incorrect words mostly have an English pronunciation. They
are corrected manually and added to the exemption
dictionary subsequently.
4. CONCLUSION
In this paper, we presented an approach for content-based
lecture video indexing and retrieval in large lecture video
archives using text and audio information . In order to verify
the research hypothesis we apply visual as well as audio
resource of lecture videos for extracting content-based
metadata automatically.
REFERENCES
[1]. E. Leeuwis, M. Federico, and M. Cettolo, ―Language
modelling and transcription of the ted corpus lectures,‖ in
Proc. IEEE Int. Conf. Acoust., Speech Signal Process.,
2003, pp. 232–235.
[2]. D. Lee and G. G. Lee, ―A korean spoken document
retrieval system for lecture search,‖ in Proc. ACM Special
Interest Group Inf. RetrievalSearching Spontaneous
Conversational Speech Workshop, 2008.
[3]. J. Glass, T. J. Hazen, L. Hetherington, and C. Wang,
―Analysis and processing of lecture audio data: Preliminary
investigations,‖ in Proc. HLT-NAACL Workshop
Interdisciplinary Approaches Speech Indexing Retrieval,
2004, pp. 9–12.
[4]. A. Haubold and J. R. Kender, ―Augmented
segmentation and visualization for presentation videos,‖ in
Proc. 13th Annu. ACM Int. Conf. Multimedia, 2005, pp. 51–
60.
[5]. W. H€urst, T. Kreuzer, and M. Wiesenh€utter, ―A
qualitative study towards using large vocabulary automatic
speech recognition toindex recorded presentations for search
and access over the web,‖ in Proc. IADIS Int. Conf.
WWW/Internet, 2002, pp. 135–143.
[6]. C. Munteanu, G. Penn, R. Baecker, and Y. C. Zhang,
―Automatic speech recognition for webcasts: How good is
good enough andwhat to do when it isn’t,‖ in Proc. 8th Int.
Conf. Multimodal Interfaces,2006.
[7]. G. Salton and C. Buckley, ―Term-weighting approaches
in automatic text retrieval,‖ Inf. Process. Manage., vol. 24,
no. 5, pp. 513–523, 1988.
[8]. G. Salton, A. Wong, and C. S. Yang. (Nov. 1975). A
vector space model for automatic indexing, Commun. ACM,
18(11),pp. 613–620, [Online]. Available:
http://doi.acm.org/10.1145/361219.361220
[9]. T.-C. Pong, F. Wang, and C.-W. Ngo, ―Structuring low-
quality videotaped lectures for cross-reference browsing by
video textanalysis,‖ J. Pattern Recog., vol. 41, no. 10, pp.
3257–3269, 2008.
[10]. M. Grcar, D. Mladenic, and P. Kese, ―Semi-automatic
categorization of videos on videolectures.net,‖ in Proc. Eur.
Conf. Mach. Learn. Knowl. Discovery Databases, 2009, pp.
730–733.
[11]. T. Tuna, J. Subhlok, L. Barker, V. Varghese, O.
Johnson, and S. Shah. (2012), ―Development and evaluation
of indexed captioned searchable videos for stem
coursework,‖ in Proc. 43rd ACM Tech. Symp. Comput. Sci.
Educ., pp. 129–134. [Online]. Available:
http://doi.acm.org/10.1145/2157136.2157177.
[12]. H. J. Jeong, T.-E. Kim, and M. H. Kim.(2012), ―An
accurate lecture video segmentation method by using sift
and adaptive threshold,‖in Proc. 10th Int. Conf. Advances
Mobile Comput., pp. 285–288.[Online]. Available:
http://doi.acm.org/10.1145/2428955.2429011.
[13]. H. Sack and J. Waitelonis, ―Integrating social tagging
and document annotation for content-based search in
multimedia data,‖ inProc. 1st Semantic Authoring
Annotation Workshop, 2006.
[14]. C. Meinel, F. Moritz, and M. Siebert, ―Community
tagging in tele-teaching environments,‖ in Proc. 2nd Int.
Conf. e-Educ., e-Bus.,e-Manage. and E-Learn., 2011.
[15]. S. Repp, A. Gross, and C. Meinel, ―Browsing within
lecture videos based on the chain index of speech
transcription,‖ IEEE Trans.Learn. Technol., vol. 1, no. 3, pp.
145–156, Jul. 2008.
[16]. J. Eisenstein, R. Barzilay, and R. Davis. (2007).
―Turning lectures into comic books using linguistically
salient gestures,‖ in Proc. 22nd Nat. Conf. Artif. Intell., 1,
pp. 877–882. [Online]. Available:
http://dl.acm.org/citation.cfm?id=1619645.1619786.
[17]. J. Adcock, M. Cooper, L. Denoue, and H. Pirsiavash,
―Talkminer: A lecture webcast search engine,‖ in Proc.
ACM Int. Conf. Multimedia, 2010, pp. 241–250.
[18]. J. Nandzik, B. Litz, N. Flores Herr, A. L€ohden, I.
Konya, D.Baum, A. Bergholz, D. Sch€onfuß, C. Fey, J.
Osterhoff, J.Waitelonis, H. Sack, R. K€ohler, and P. Ndjiki-
Nya. (2012) ― Contentus—technologies for next generation
multimedia libraries, Multimedia Tools Appl., pp. 1–43,
[Online]. Available:http://dx.doi.org/10.1007/s11042-011-
0971-2.
[19]. F. Chang, C.-J. Chen, and C.-J. Lu, ―A linear-time
component labeling algorithm using contour tracing
technique,‖ Comput. Vis. Image Understanding, vol. 93, no.
2, pp. 206–220, Jan. 2004.
[20]. B. T. N. Dala, ―Histograms of oriented gradients for
humandetection,‖ in Proc. IEEE Conf. Comput. Vis. Pattern
Recog., 2005,vol. 1, pp. 886–893.
[21]. Ground truth data. (2013). [Online]. Available:
http://www.yanghaojin.com/research/videoOCR.html.
[22]. B. Epshtein, E. Ofek, and Y. Wexler, ―Detecting text
in natural scenes with stroke width transform,‖ in Proc. Int.
Conf. Comput.Vis. Pattern Recog., 2010, pp. 2963–2970..
[23]. H. Yang, B. Quehl, and H. Sack. (2012), ―A
framework for improved video text detection and
recognition,‖ Multimedia Tools Appl., pp. 1–29, [Online].
Available: http://dx.doi.org/10.1007/s11042-012-1250-6.
[24]. K. Toutanova, D. Klein, C. Manning, and Y. Singer,
―Feature-rich part-of-speech tagging with a cyclic
dependency network,‖ in Proc. Conf. North Amer. Chapter
Assoc. Comput. Linguistics: Human Lang. Technol., 2003,
pp. 252–259.