The document discusses the Metadata Encoding and Transmission Standard (METS), which is an XML schema for encoding descriptive, administrative, and structural metadata regarding objects within a digital library. It describes the characteristics and sections of a METS file, including the header, descriptive and administrative metadata, file and structural map sections. Current users of METS are also listed, such as libraries and universities. The purpose of METS is to provide a flexible structure for linking metadata and content about digital objects.
Metrics envelop number of subject domains, e.g., general relativity under physics, networking, mathematics, software analysis, etc. --- STATISTICS
Enumerated in the slides are the different metric fields in information science.
This document discusses the MEDLARS system, which is an online medical literature analysis and retrieval system started by the National Library of Medicine. It summarizes that MEDLARS began in 1960 and became operational in 1964 with the first computer-produced issue of Index Medicus. The document also notes that a study was conducted to evaluate MEDLARS by submitting 302 search requests from users and assessing recall and precision. The average recall was found to be 57.7% and precision was 50.4% based on the user relevance assessments. Recommendations were made to improve MEDLARS performance based on the results.
The Dublin Core Metadata Initiative began in 1994 to develop a core set of metadata semantics for describing and searching resources on the web. It consists of 15 basic metadata elements such as title, creator, and description. Dublin Core records can be expressed using XML and the Resource Description Framework (RDF), representing relationships between resources with subject-predicate-object triples. The Dublin Core Metadata Initiative maintains guidelines and models to define implementation of Dublin Core metadata.
This document provides an overview of metadata standards, including their purpose and types. It describes the MARC 21 and Dublin Core metadata standards in detail. MARC 21 is the predominant bibliographic standard, with formats for bibliographic data, holdings, and authority data. It exists in both MARC 21 and MARCXML syntaxes. Dublin Core is a simpler standard for resource discovery with 15 basic elements. It includes both simple and qualified versions with controlled vocabularies. The document lists several metadata standards and development organizations.
construction of a call number by computer
artificial intelligence
able to identify the subject and sub-subjects of the document
doubt about the capability of computers for classification
similar automatic production of title indexes or keyword enhanced indexes
attempts to design a powerful automatic
The document discusses the International Standard Bibliographic Description (ISBD), which is a set of rules produced by IFLA to create standardized bibliographic descriptions. It provides a brief history of ISBD, noting it was developed in the 1960s-1970s in response to a need for standardized cataloging. The key elements and areas of description in ISBD for monographs and serials are outlined. Characteristics of ISBD include its comprehensiveness, fixed order of data elements, and use of punctuation to delimit elements. The document serves as an introduction to ISBD.
DESIDOC is the Defence Scientific Information and Documentation Centre established in 1970 in Delhi, India. It operates under the Defence Research and Development Organization (DRDO) to disseminate science and technology information on cutting edge defence technologies. DESIDOC's vision is to be a centre of excellence for knowledge management in DRDO. It provides library resources and access to databases for DRDO headquarters and laboratories. DESIDOC conducts training programs and publishes various periodicals related to defence research.
DESIDOC is the central scientific information and documentation center of DRDO. It was established in 1958 as the Scientific Information Bureau and collects, processes, and disseminates technical information to DRDO scientists. DESIDOC maintains the Defence Science Library, publishes various journals and newsletters, provides training programs, and carries out other scientific information services like database development and publishing DRDO works.
Metrics envelop number of subject domains, e.g., general relativity under physics, networking, mathematics, software analysis, etc. --- STATISTICS
Enumerated in the slides are the different metric fields in information science.
This document discusses the MEDLARS system, which is an online medical literature analysis and retrieval system started by the National Library of Medicine. It summarizes that MEDLARS began in 1960 and became operational in 1964 with the first computer-produced issue of Index Medicus. The document also notes that a study was conducted to evaluate MEDLARS by submitting 302 search requests from users and assessing recall and precision. The average recall was found to be 57.7% and precision was 50.4% based on the user relevance assessments. Recommendations were made to improve MEDLARS performance based on the results.
The Dublin Core Metadata Initiative began in 1994 to develop a core set of metadata semantics for describing and searching resources on the web. It consists of 15 basic metadata elements such as title, creator, and description. Dublin Core records can be expressed using XML and the Resource Description Framework (RDF), representing relationships between resources with subject-predicate-object triples. The Dublin Core Metadata Initiative maintains guidelines and models to define implementation of Dublin Core metadata.
This document provides an overview of metadata standards, including their purpose and types. It describes the MARC 21 and Dublin Core metadata standards in detail. MARC 21 is the predominant bibliographic standard, with formats for bibliographic data, holdings, and authority data. It exists in both MARC 21 and MARCXML syntaxes. Dublin Core is a simpler standard for resource discovery with 15 basic elements. It includes both simple and qualified versions with controlled vocabularies. The document lists several metadata standards and development organizations.
construction of a call number by computer
artificial intelligence
able to identify the subject and sub-subjects of the document
doubt about the capability of computers for classification
similar automatic production of title indexes or keyword enhanced indexes
attempts to design a powerful automatic
The document discusses the International Standard Bibliographic Description (ISBD), which is a set of rules produced by IFLA to create standardized bibliographic descriptions. It provides a brief history of ISBD, noting it was developed in the 1960s-1970s in response to a need for standardized cataloging. The key elements and areas of description in ISBD for monographs and serials are outlined. Characteristics of ISBD include its comprehensiveness, fixed order of data elements, and use of punctuation to delimit elements. The document serves as an introduction to ISBD.
DESIDOC is the Defence Scientific Information and Documentation Centre established in 1970 in Delhi, India. It operates under the Defence Research and Development Organization (DRDO) to disseminate science and technology information on cutting edge defence technologies. DESIDOC's vision is to be a centre of excellence for knowledge management in DRDO. It provides library resources and access to databases for DRDO headquarters and laboratories. DESIDOC conducts training programs and publishes various periodicals related to defence research.
DESIDOC is the central scientific information and documentation center of DRDO. It was established in 1958 as the Scientific Information Bureau and collects, processes, and disseminates technical information to DRDO scientists. DESIDOC maintains the Defence Science Library, publishes various journals and newsletters, provides training programs, and carries out other scientific information services like database development and publishing DRDO works.
INSPEC is a major indexing database of scientific and technical literature published by the Institution of Engineering and Technology (IET). It contains nearly 12 million records covering topics in electronics, computer science, physics, engineering and related interdisciplinary fields. INSPEC provides comprehensive indexing and abstracts of journal articles, conference papers, books and reports for easy searching and retrieval of relevant information.
The document discusses the Dewey Decimal Classification system created by Melvil Dewey in 1876. It provides an overview of Dewey and the development of the DDC over 23 editions. The core features of the DDC are explained, including its enumerative scheme, hierarchical notation, and regular updating process. The DDC is now the most widely used library classification system globally.
This document discusses different modes of subject formation in library science, as proposed by Ranganathan in the 1950s. It outlines seven modes of subject formation - loose assemblage, lamination, dissection, fusion, distillation, agglomeration, and cluster. These modes represent systematic relationships that guide how ideas are organized into subjects. Understanding these formation modes helps classify subjects in documents to efficiently retrieve and disseminate knowledge to library users.
Standards to facilitate information exchange has always been a subject of concern.
To provide a flexible exchange format that could be used for converting data from libraries and information services of all types, UNESCO developed the Common Communication Format (CCF). The main aim of this format was to produce a method of organising bibliographic descriptions which could be exchanged between institutions. This format was to act as a link between the databases produced in different internal formats of libraries.
The document discusses the Sears List of Subject Headings (SLSH), a controlled vocabulary used for subject cataloging in small to medium sized libraries. It provides an overview of the history and purpose of SLSH, describes some of its key features like new subject headings added in the 21st edition, and outlines its underlying principles of direct, specific, and consistent subject entries based on common usage. The structure of SLSH is also briefly explained as an alphabetical list of subject terms for the entire range of knowledge.
Post coordinate indexing .. Library and information scienceharshaec
This document discusses post-coordinate indexing and uniterm indexing. Post-coordinate indexing allows searchers to freely combine terms at search time, which can increase recall but decrease precision. Uniterm indexing involves analyzing documents into individual concepts represented as single terms. These terms are recorded on cards divided into numbered columns, and searching involves finding common numbers across cards to retrieve documents. While flexible, post-coordinate and uniterm indexing can be time-consuming and lack context.
National social science documentation centre (nassdoc )GordonAmidu
NASSDOC was established in 1969 as a branch of the Indian Council of Social Science Research to provide library and information support services to social science researchers. It assists libraries of ICSSR research institutes, provides study grants to PhD candidates, and offers financial aid for documentation projects. NASSDOC also provides consultation services, reference services, literature searches, and document delivery to researchers.
POPSI (Postulate based permuted subject indexing) is a pre-coordinate indexing system developed by G. Bhattacharyya that uses an analytic-synthetic method and permutation of terms to approach documents from different perspectives. It is based on Ranganathan's postulates and classification principles. POPSI helps formulate subject headings, derive index entries, determine subject queries, and formulate search strategies. The main POPSI table contains notation used in the indexing process. Key steps include analysis, formalization, modulation, standardization, and generating organized and associative classification entries and references.
The document discusses the United Nations International Scientific Information System (UNISIST). It provides a history of UNISIST, describing how it was established through cooperation between UNESCO and ICSU to study the feasibility of a world science information system. The key aims and objectives of UNISIST are to coordinate trends toward cooperation in scientific information, act as a catalyst for necessary development, and facilitate access to world information resources through the establishment of standards and an interconnected network. UNISIST seeks to improve tools for system intercommunication and strengthen components of the information transfer chain.
The document provides information about AGRIS (International System for Agricultural Science and Technology), which is a global public database maintained by FAO containing over 8 million bibliographic records related to agriculture. It describes AGRIS's history and development since 1975, its objectives to support decision making and knowledge sharing, the various data sources and literature it covers, and its services including online access and information provision. It also discusses AGRIS's role in India and future plans to expand its coverage and develop decision support systems.
This document discusses the importance of scientific documentation and information centers for economic growth. It outlines the history and evolution of documentation centers in India, from early concepts put forth by Paul Otlet and E. Hymansas to the establishment of specific centers like SENDOC and NI-MSME. SENDOC and NI-MSME provide information services, training programs, publications, and more to support micro, small and medium enterprises in India, which employ over 30 million people and contribute significantly to exports and industrial production. The document concludes by emphasizing the objectives and services of documentation centers in promoting and developing MSMEs in India.
RDA is a new cataloging standard designed to replace AACR2 and provide guidelines for describing digital resources. It is based on FRBR and FRAD which are models that organize information by user tasks and relationships between entities like works, expressions, manifestations and items. RDA aims to be more intuitive for users by providing more detailed descriptions of resources and is being tested by various libraries and organizations before its full implementation. However, some questions remain regarding its costs and benefits compared to AACR2.
This document discusses the basics of subject indexing in libraries. It defines subject indexing as providing subject access to microdocuments like journal articles and research reports by assigning appropriate subject terms. The key points covered are:
- Subject indexing allows users to identify documents on a given subject and find related documents.
- Indexes are helpful for retrieving information from both print and digital collections. They provide subject access through assigned terms.
- Effective subject indexing requires identifying the main concepts in a document and re-expressing them as index terms so the document and terms express the same concepts.
- Principles of indexing include using terminology familiar to users and bringing related documents together under consistent, unambiguous headings. Specificity and exhaustivity must
when new subject come into existence ,we have to give a place among already existing subject. this ppt will help to how can we assign a place to particular subject.it will helpful for all the students whom are pursuing their master in library science ans information management
Greenstone Digital Library Software is an open source software suite for building and distributing digital library collections. It was developed by the New Zealand Digital Library Project and is distributed in cooperation with UNESCO. Greenstone allows for the creation of collections from a variety of material, provides tools for searching and browsing collections, and supports publishing collections on the web or CD-ROM. It has been widely adopted internationally since its initial release in 2000.
The document discusses bibliometrics and bibliometric laws. It defines bibliometrics as the application of mathematical and statistical methods to bibliographic information, and discusses its origins and objectives. Three important bibliometric laws are covered: Lotka's law describes the frequency of publication by authors, Bradford's law explains the scattering of articles across journals, and Zipf's law concerns the frequency of word usage. In conclusion, bibliometrics is an important tool in library and information science for studying information processes and flows.
The Indian Association of Special Libraries and Information Centers (IASLIC) is a non-profit organization founded in 1955 that aims to promote and develop special librarianship in India. IASLIC organizes conferences and seminars, publishes various journals and directories, offers training programs, and works to address issues impacting library professionals. Membership is open to both individuals and institutions and helps support IASLIC's work in advancing libraries.
This document discusses normative principles of classification as proposed by Dr. Ranganathan. It outlines six purposes that normative principles can serve, such as providing a scientific basis for classification scheme design. Ranganathan adopted conventions for terms like "laws", "canons", and "principles" depending on the level of classification. He recognized basic laws of interpretation, impartiality, symmetry, parsimony, local variation, and osmosis. The conclusion is that Ranganathan's theory provides a scientific basis for subject determination and classification.
This document provides information about enumerative and faceted classification schemes. It discusses enumerative classification schemes, noting that they involve assigning unique notations or symbols to predefined categories to create a hierarchical structure for organizing resources. Examples provided include the Library of Congress Classification and Dewey Decimal Classification. The benefits of enumerative schemes are also summarized, including efficient retrieval, consistency, subject browsing, and facilitating information organization. Faceted classification is then introduced as a system that organizes information based on multiple independent attributes or facets, allowing for more precise categorization. The components and benefits of faceted classification are outlined.
Information consolidation is defined as the process of evaluating and compressing relevant documents to provide users with reliable and concise information. It involves defining responsibility for analyzing documents and packaging information appropriately for users' needs, levels, and time constraints. The benefits of information consolidation include increasing the effectiveness and use of information for various activities, as well as expanding the circle of potential users by providing evaluated and synthesized information. The basic processes involve studying user needs, selecting relevant sources, evaluating and analyzing information, restructuring it into a new whole, and packaging and disseminating it to encourage use.
Meeting today’s dissemination challenges – Implementing International Standar...Jonathan Challener
This document discusses the .Stat system, which serves as a central repository for validated statistics and metadata. .Stat connects data production, sharing, and dissemination processes. It provides three key functional areas: a data upload engine, a data delivery engine, and a data browser. .Stat can be mapped to stages in the Generic Statistical Business Process Model and incorporates standards like SDMX for dissemination, data exchange, and internal data sharing. The document outlines .Stat's current role and future plans to further support SDMX artifacts, ingest, registries, and semantic web opportunities.
WHAT IS A DBMS? EXPLAIN DIFFERENT MYSQL COMMANDS AND CONSTRAINTS OF THE SAME.`Shweta Bhavsar
This document discusses database management systems (DBMS) and MySQL commands and constraints. It begins by defining a DBMS and describing their components and characteristics, including data models, query languages, and advantages like data integrity and sharing. It then explains common MySQL commands to create and manage databases, tables, and insert values. Constraints are also discussed as ways to define data types and validate values in tables.
INSPEC is a major indexing database of scientific and technical literature published by the Institution of Engineering and Technology (IET). It contains nearly 12 million records covering topics in electronics, computer science, physics, engineering and related interdisciplinary fields. INSPEC provides comprehensive indexing and abstracts of journal articles, conference papers, books and reports for easy searching and retrieval of relevant information.
The document discusses the Dewey Decimal Classification system created by Melvil Dewey in 1876. It provides an overview of Dewey and the development of the DDC over 23 editions. The core features of the DDC are explained, including its enumerative scheme, hierarchical notation, and regular updating process. The DDC is now the most widely used library classification system globally.
This document discusses different modes of subject formation in library science, as proposed by Ranganathan in the 1950s. It outlines seven modes of subject formation - loose assemblage, lamination, dissection, fusion, distillation, agglomeration, and cluster. These modes represent systematic relationships that guide how ideas are organized into subjects. Understanding these formation modes helps classify subjects in documents to efficiently retrieve and disseminate knowledge to library users.
Standards to facilitate information exchange has always been a subject of concern.
To provide a flexible exchange format that could be used for converting data from libraries and information services of all types, UNESCO developed the Common Communication Format (CCF). The main aim of this format was to produce a method of organising bibliographic descriptions which could be exchanged between institutions. This format was to act as a link between the databases produced in different internal formats of libraries.
The document discusses the Sears List of Subject Headings (SLSH), a controlled vocabulary used for subject cataloging in small to medium sized libraries. It provides an overview of the history and purpose of SLSH, describes some of its key features like new subject headings added in the 21st edition, and outlines its underlying principles of direct, specific, and consistent subject entries based on common usage. The structure of SLSH is also briefly explained as an alphabetical list of subject terms for the entire range of knowledge.
Post coordinate indexing .. Library and information scienceharshaec
This document discusses post-coordinate indexing and uniterm indexing. Post-coordinate indexing allows searchers to freely combine terms at search time, which can increase recall but decrease precision. Uniterm indexing involves analyzing documents into individual concepts represented as single terms. These terms are recorded on cards divided into numbered columns, and searching involves finding common numbers across cards to retrieve documents. While flexible, post-coordinate and uniterm indexing can be time-consuming and lack context.
National social science documentation centre (nassdoc )GordonAmidu
NASSDOC was established in 1969 as a branch of the Indian Council of Social Science Research to provide library and information support services to social science researchers. It assists libraries of ICSSR research institutes, provides study grants to PhD candidates, and offers financial aid for documentation projects. NASSDOC also provides consultation services, reference services, literature searches, and document delivery to researchers.
POPSI (Postulate based permuted subject indexing) is a pre-coordinate indexing system developed by G. Bhattacharyya that uses an analytic-synthetic method and permutation of terms to approach documents from different perspectives. It is based on Ranganathan's postulates and classification principles. POPSI helps formulate subject headings, derive index entries, determine subject queries, and formulate search strategies. The main POPSI table contains notation used in the indexing process. Key steps include analysis, formalization, modulation, standardization, and generating organized and associative classification entries and references.
The document discusses the United Nations International Scientific Information System (UNISIST). It provides a history of UNISIST, describing how it was established through cooperation between UNESCO and ICSU to study the feasibility of a world science information system. The key aims and objectives of UNISIST are to coordinate trends toward cooperation in scientific information, act as a catalyst for necessary development, and facilitate access to world information resources through the establishment of standards and an interconnected network. UNISIST seeks to improve tools for system intercommunication and strengthen components of the information transfer chain.
The document provides information about AGRIS (International System for Agricultural Science and Technology), which is a global public database maintained by FAO containing over 8 million bibliographic records related to agriculture. It describes AGRIS's history and development since 1975, its objectives to support decision making and knowledge sharing, the various data sources and literature it covers, and its services including online access and information provision. It also discusses AGRIS's role in India and future plans to expand its coverage and develop decision support systems.
This document discusses the importance of scientific documentation and information centers for economic growth. It outlines the history and evolution of documentation centers in India, from early concepts put forth by Paul Otlet and E. Hymansas to the establishment of specific centers like SENDOC and NI-MSME. SENDOC and NI-MSME provide information services, training programs, publications, and more to support micro, small and medium enterprises in India, which employ over 30 million people and contribute significantly to exports and industrial production. The document concludes by emphasizing the objectives and services of documentation centers in promoting and developing MSMEs in India.
RDA is a new cataloging standard designed to replace AACR2 and provide guidelines for describing digital resources. It is based on FRBR and FRAD which are models that organize information by user tasks and relationships between entities like works, expressions, manifestations and items. RDA aims to be more intuitive for users by providing more detailed descriptions of resources and is being tested by various libraries and organizations before its full implementation. However, some questions remain regarding its costs and benefits compared to AACR2.
This document discusses the basics of subject indexing in libraries. It defines subject indexing as providing subject access to microdocuments like journal articles and research reports by assigning appropriate subject terms. The key points covered are:
- Subject indexing allows users to identify documents on a given subject and find related documents.
- Indexes are helpful for retrieving information from both print and digital collections. They provide subject access through assigned terms.
- Effective subject indexing requires identifying the main concepts in a document and re-expressing them as index terms so the document and terms express the same concepts.
- Principles of indexing include using terminology familiar to users and bringing related documents together under consistent, unambiguous headings. Specificity and exhaustivity must
when new subject come into existence ,we have to give a place among already existing subject. this ppt will help to how can we assign a place to particular subject.it will helpful for all the students whom are pursuing their master in library science ans information management
Greenstone Digital Library Software is an open source software suite for building and distributing digital library collections. It was developed by the New Zealand Digital Library Project and is distributed in cooperation with UNESCO. Greenstone allows for the creation of collections from a variety of material, provides tools for searching and browsing collections, and supports publishing collections on the web or CD-ROM. It has been widely adopted internationally since its initial release in 2000.
The document discusses bibliometrics and bibliometric laws. It defines bibliometrics as the application of mathematical and statistical methods to bibliographic information, and discusses its origins and objectives. Three important bibliometric laws are covered: Lotka's law describes the frequency of publication by authors, Bradford's law explains the scattering of articles across journals, and Zipf's law concerns the frequency of word usage. In conclusion, bibliometrics is an important tool in library and information science for studying information processes and flows.
The Indian Association of Special Libraries and Information Centers (IASLIC) is a non-profit organization founded in 1955 that aims to promote and develop special librarianship in India. IASLIC organizes conferences and seminars, publishes various journals and directories, offers training programs, and works to address issues impacting library professionals. Membership is open to both individuals and institutions and helps support IASLIC's work in advancing libraries.
This document discusses normative principles of classification as proposed by Dr. Ranganathan. It outlines six purposes that normative principles can serve, such as providing a scientific basis for classification scheme design. Ranganathan adopted conventions for terms like "laws", "canons", and "principles" depending on the level of classification. He recognized basic laws of interpretation, impartiality, symmetry, parsimony, local variation, and osmosis. The conclusion is that Ranganathan's theory provides a scientific basis for subject determination and classification.
This document provides information about enumerative and faceted classification schemes. It discusses enumerative classification schemes, noting that they involve assigning unique notations or symbols to predefined categories to create a hierarchical structure for organizing resources. Examples provided include the Library of Congress Classification and Dewey Decimal Classification. The benefits of enumerative schemes are also summarized, including efficient retrieval, consistency, subject browsing, and facilitating information organization. Faceted classification is then introduced as a system that organizes information based on multiple independent attributes or facets, allowing for more precise categorization. The components and benefits of faceted classification are outlined.
Information consolidation is defined as the process of evaluating and compressing relevant documents to provide users with reliable and concise information. It involves defining responsibility for analyzing documents and packaging information appropriately for users' needs, levels, and time constraints. The benefits of information consolidation include increasing the effectiveness and use of information for various activities, as well as expanding the circle of potential users by providing evaluated and synthesized information. The basic processes involve studying user needs, selecting relevant sources, evaluating and analyzing information, restructuring it into a new whole, and packaging and disseminating it to encourage use.
Meeting today’s dissemination challenges – Implementing International Standar...Jonathan Challener
This document discusses the .Stat system, which serves as a central repository for validated statistics and metadata. .Stat connects data production, sharing, and dissemination processes. It provides three key functional areas: a data upload engine, a data delivery engine, and a data browser. .Stat can be mapped to stages in the Generic Statistical Business Process Model and incorporates standards like SDMX for dissemination, data exchange, and internal data sharing. The document outlines .Stat's current role and future plans to further support SDMX artifacts, ingest, registries, and semantic web opportunities.
WHAT IS A DBMS? EXPLAIN DIFFERENT MYSQL COMMANDS AND CONSTRAINTS OF THE SAME.`Shweta Bhavsar
This document discusses database management systems (DBMS) and MySQL commands and constraints. It begins by defining a DBMS and describing their components and characteristics, including data models, query languages, and advantages like data integrity and sharing. It then explains common MySQL commands to create and manage databases, tables, and insert values. Constraints are also discussed as ways to define data types and validate values in tables.
Secrets of Enterprise Data Mining: SQL Saturday 328 Birmingham ALMark Tabladillo
This document discusses secrets of enterprise data mining. It begins by defining data mining as the automated or semi-automated process of discovering patterns in data. It then discusses how data mining can be applied in various industries like telecommunications, oil and gas, and Volkswagen Group. Finally, it discusses how Microsoft offers solutions for enterprise data mining through SQL Server Analysis Services and Microsoft Azure Machine Learning.
Big data processing using - Hadoop TechnologyShital Kat
This document summarizes a report on Hadoop technology as a solution to big data processing. It discusses the big data problem, including defining big data, its characteristics and challenges. It then introduces Hadoop as a solution, describing its components HDFS for storage and MapReduce for parallel processing. Examples of common friend lists and word counting are provided. Finally, it briefly mentions some Hadoop projects and companies that use Hadoop.
IRJET- Generate Distributed Metadata using Blockchain Technology within HDFS ...IRJET Journal
This document proposes a new HDFS architecture that eliminates the single point of failure of the NameNode by distributing metadata storage using blockchain technology. In the traditional HDFS, the NameNode stores all metadata, but in the new architecture this is replaced by blockchain miners that securely store encrypted metadata across data nodes. Blockchain links data blocks in a serial manner with cryptographic hashes to ensure integrity. The key components are HDFS clients, data nodes for storage, and specially designated miner nodes that help create and store metadata blocks in an encrypted and distributed fashion similar to how transactions are recorded in a blockchain. This architecture aims to provide reliable, secure and faster metadata access without a single point of failure.
NOSQL Database Engines for Big Data Managementijtsrd
The document discusses NoSQL database engines and their use for managing large amounts of data. It describes that NoSQL databases were developed to address the challenges of scaling faced by traditional relational databases. The three main types of NoSQL databases discussed are key-value stores, document databases, and extensible record stores. MongoDB is provided as an example of a popular open source document database that is designed to store, retrieve, and manage semi-structured data through flexible document schemas.
Secrets of Enterprise Data Mining: SQL Saturday Oregon 201411Mark Tabladillo
If you have a SQL Server license (Standard or higher) then you already have the ability to start data mining. In this new presentation, you will see how to scale up data mining from the free Excel 2013 add-in to production use. Aimed at beginning to intermediate data miners, this presentation will show how mining models move from development to production. We will use SQL Server 2014 tools including SSMS, SSIS, and SSDT.
Big Data Processing with Hadoop : A ReviewIRJET Journal
1. This document provides an overview of big data processing with Hadoop. It defines big data and describes the challenges of volume, velocity, variety and variability.
2. Traditional data processing approaches are inadequate for big data due to its scale. Hadoop provides a distributed file system called HDFS and a MapReduce framework to address this.
3. HDFS uses a master-slave architecture with a NameNode and DataNodes to store and retrieve file blocks. MapReduce allows distributed processing of large datasets across clusters through mapping and reducing functions.
Review on Automation Tool for ERD NormalizationIRJET Journal
The document describes a methodology for automatically extracting entity relationship diagrams (ERDs) from English language descriptions. It involves using heuristic rules to extract entities, relationships, and attributes from text and then generating an ERD. The proposed system aims to provide a quick and easy way for database designers to generate ERDs from requirements documents. Existing normalization tools only operate on relational schemas and do not support conceptual-level normalization of ERDs. The proposed system would extract ERDs and then apply normalization techniques like 1NF and 2NF to remove ambiguities and inconsistencies from the initial ERD.
An Overview of General Data Mining ToolsIRJET Journal
This document provides an overview of several popular general data mining tools: Weka, Rapid Miner, IBM SPSS, Tanagra, KNIME, Orange, and R. It describes the key characteristics of each tool, including their interfaces, available algorithms, licensing, community support, and abilities for workflows and big data processing. Overall, the document evaluates and compares the capabilities of these seven major data mining software platforms.
Quality of Groundwater in Lingala Mandal of YSR Kadapa District, Andhraprades...IRJET Journal
This document provides an overview of several popular general data mining tools: Weka, Rapid Miner, IBM SPSS, Tanagra, KNIME, Orange, and R. It describes the key characteristics of each tool, including their interfaces, available algorithms, licensing, community support, and abilities for workflows and big data processing. Overall, the document finds that while each tool has strengths in different areas, Weka, Rapid Miner, IBM SPSS, KNIME, and Orange provide robust functionality for a wide range of general data mining tasks through their graphical user interfaces and available algorithms.
Database administration refers to the whole set of activities performed by a database administrator to ensure that a database is always available as needed. Other closely related tasks and roles are database security, database monitoring and troubleshooting, and planning for future growth
IRJET- Data Retrieval using Master Resource Description FrameworkIRJET Journal
This document discusses using a Master Resource Description Framework (MRDF) to improve data retrieval efficiency from databases. The MRDF combines multiple RDF files into a single framework to reduce the time needed for search engines to query each individual RDF file. It also describes using a user profile to track user interests and tailor query results accordingly for a personalized search experience. The MRDF approach is presented as improving search efficiency while retrieving data from databases.
Comparing and analyzing various method of data integration in big dataIRJET Journal
This document discusses various methods for integrating data from multiple heterogeneous sources in big data systems. It begins by defining data integration and explaining the challenges posed by big data's volume, velocity, variety and veracity. The document then examines different techniques for big data integration including Extract-Transform-Load (ETL), Enterprise Information Integration (EII), data migration, data consolidation, data propagation, data federation and change data capture. It analyzes the advantages and limitations of each technique. In conclusion, the document states that data integration is a major challenge and there is no single best method, as the appropriate solution depends on the specific use case and data characteristics.
The document discusses big data analysis and provides an introduction to key concepts. It is divided into three parts: Part 1 introduces big data and Hadoop, the open-source software framework for storing and processing large datasets. Part 2 provides a very quick introduction to understanding data and analyzing data, intended for those new to the topic. Part 3 discusses concepts and references to use cases for big data analysis in the airline industry, intended for more advanced readers. The document aims to familiarize business and management users with big data analysis terms and thinking processes for formulating analytical questions to address business problems.
Intro to big data and applications - day 2Parviz Vakili
The document provides an introduction and references for a presentation on big data and applications. It includes sections on data architecture, data governance, data modeling and design, and reference architectures for big data analytics. The presentation template was created by Slidesgo and credits are provided.
AUTOMATIC TRANSFER OF DATA USING SERVICE-ORIENTED ARCHITECTURE TO NoSQL DATAB...IRJET Journal
This document summarizes an academic paper that proposes a model for automatically migrating data from relational databases to NoSQL databases using service-oriented architecture. The model encapsulates popular NoSQL databases like MongoDB, Cassandra, and Neo4j as web services. This allows data to be efficiently migrated from a relational database like Apache Derby to a NoSQL database with minimal knowledge of how each database works. The document provides details of the proposed migration model and discusses its implementation and testing migrating data from Derby to the NoSQL databases successfully.
IRJET- Deduplication Detection for Similarity in Document Analysis Via Vector...IRJET Journal
The document discusses techniques for detecting similarity and deduplication in document analysis using vector analysis. It proposes analyzing documents by extracting abstract content, separating words and combining them in a word cloud to determine frequency. This approach aims to identify whether documents are duplicates by analyzing word vectors at the word, sentence and paragraph level while also applying techniques like stemming, stopping words and semantic similarity.
LAND USE LAND COVER AND NDVI OF MIRZAPUR DISTRICT, UPRAHUL
This Dissertation explores the particular circumstances of Mirzapur, a region located in the
core of India. Mirzapur, with its varied terrains and abundant biodiversity, offers an optimal
environment for investigating the changes in vegetation cover dynamics. Our study utilizes
advanced technologies such as GIS (Geographic Information Systems) and Remote sensing to
analyze the transformations that have taken place over the course of a decade.
The complex relationship between human activities and the environment has been the focus
of extensive research and worry. As the global community grapples with swift urbanization,
population expansion, and economic progress, the effects on natural ecosystems are becoming
more evident. A crucial element of this impact is the alteration of vegetation cover, which plays a
significant role in maintaining the ecological equilibrium of our planet.Land serves as the foundation for all human activities and provides the necessary materials for
these activities. As the most crucial natural resource, its utilization by humans results in different
'Land uses,' which are determined by both human activities and the physical characteristics of the
land.
The utilization of land is impacted by human needs and environmental factors. In countries
like India, rapid population growth and the emphasis on extensive resource exploitation can lead
to significant land degradation, adversely affecting the region's land cover.
Therefore, human intervention has significantly influenced land use patterns over many
centuries, evolving its structure over time and space. In the present era, these changes have
accelerated due to factors such as agriculture and urbanization. Information regarding land use and
cover is essential for various planning and management tasks related to the Earth's surface,
providing crucial environmental data for scientific, resource management, policy purposes, and
diverse human activities.
Accurate understanding of land use and cover is imperative for the development planning
of any area. Consequently, a wide range of professionals, including earth system scientists, land
and water managers, and urban planners, are interested in obtaining data on land use and cover
changes, conversion trends, and other related patterns. The spatial dimensions of land use and
cover support policymakers and scientists in making well-informed decisions, as alterations in
these patterns indicate shifts in economic and social conditions. Monitoring such changes with the
help of Advanced technologies like Remote Sensing and Geographic Information Systems is
crucial for coordinated efforts across different administrative levels. Advanced technologies like
Remote Sensing and Geographic Information Systems
9
Changes in vegetation cover refer to variations in the distribution, composition, and overall
structure of plant communities across different temporal and spatial scales. These changes can
occur natural.
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxEduSkills OECD
Iván Bornacelly, Policy Analyst at the OECD Centre for Skills, OECD, presents at the webinar 'Tackling job market gaps with a skills-first approach' on 12 June 2024
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
1. Metadata Encoding and Transmission Standard
Sandip Das
MS-LIS 2013-2015
Documentation Research And Training Centre (DRTC)
Indian Statistical Institute
Bangalore Centre
4. Metadata ?
Data about data
Metadata is structured information
Metadata can be used to find, gather, and maintain resources
over long periods of time.
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
4
5. METS ?
Metadata Encoding and Transmission Standard
A metadata standard for encoding descriptive, administrative,
and structural metadata regarding objects within a digital
library, expressed using the XML language of the World Wide
Web Consortium.
Produced by LC and DLF
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
5
6. XML?
An ISO standard
Interchangeable with other
applications
Easy to integrate
cataloguing
information with text
transcription, images etc.
Handles structural metadata
easily
Source : www.allzonedigital.com
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
6
7. Why We Use ?
All metadata format (like DC, EAD, VRA) utilize for bibliographic
description.
METS act as a container object for the many piece of metadata
needed to describe a single digital object.
A number of pieces of metadata are attached to each digital
object.
METS provides a method for building the object (i.e. metadata,
attached items etc.) together so they can be transferred.
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
7
9. Characteristics
Generally one METS file corresponds to one digital object (which may
incorporate many files)
All metadata encoded in single document
Each type is held in a separate section, linked by identifiers
All metadata and external data (eg. images, text, video) is either
referenced from METS file or can be held internally
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
9
11. HOW TO BUILD A METS DOCUMENT
METS
Header
Struct. Map
Admin MD
File Sec.
Struct. Link
Descrip. MD
Behaviors
Image Source : http://www.slideshare.net/Annie05/mets-short-overview-context-presentation
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
11
12. METS Header
The METS Header contains metadata describing the METS
document itself, including such information as creator,
editor, etc.
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
12
14. Descriptive Metadata
Can record all of the units of descriptive metadata pertaining to the digital
entity represented by METS document
– Descriptive Metadata may be
– External to the METS document
– Internal to the METS document
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
14
15. External Descriptive Metadata (Contd.)
External Descriptive metadata element in a METS document
may simply identify the type of descriptive metadata it
represents (MARC, EAD, etc), and point to this metadata in
its external location via a URI Descriptive Metadata
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
15
16. External Descriptive Metadata (Contd...)
<dmdSec ID="dmd001"> <mdRef LOCTYPE="URN"
MIMETYPE="application/xml" MDTYPE="EAD" LABEL="Berol Collection
Finding Aid">urn:x-nyu:fales1735</mdRef> </dmdSec>
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
16
17. Internal Descriptive Metadata
Internal descriptive metadata may be recorded internally in a METS
document in one of two ways
– As XML data using vocabulary and syntax specified in external XML
standard. For example, Dublin Core, MARC, MODS.
– As binary data. For example, a standard MARC record could simply be
incorporated as binary data into METS document.
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
17
19. Administrative metadata
This elements contain the administrative metadata pertaining to the files comprising a
digital library object, as well as that pertaining to the original source material used to
create the object. There are four main forms of administrative metadata provided for
in a METS document:
1. Technical Metadata (information regarding files' creation, format, and use
characteristics),
2. IPR Metadata (copyright and license information)
3. Source Metadata (descriptive and administrative metadata regarding the analog
source from which a digital library object derives)
4. Digital Provenance Metadata (information regarding source/destination
relationships between files)
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
19
20. File Section
The file section lists all files containing content which comprise the electronic
versions of the digital object.
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
20
21. File Section(contd…)
File groups
Files are organized into File Groups
based on the grouping you would
like to do. One way is to group by
format (tiff, hi-res jpeg, med-res
jpeg, gif, etc)
Feb 19, 2014
Source:www.loc.gov/standards/mets/presentations/METS.ppt
Sandip Das, DRTC, Indian Statistical Institute
21
22. File Section(contd…)
File
<file ID="FILE003" MIMETYPE="image/jpeg">
<Flocat LOCTYPE="URL">
http://dlib.nyu.edu/tamwag/beame.jpg
</FLocat>
</file>
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
22
23. Structural Map Section
The structural map is the heart of a METS document. It outlines a hierarchical
structure for the digital library object, and links the elements of that structure to
content files and metadata that pertain to each element.
Feb 19, 2014
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Sandip Das, DRTC, Indian Statistical Institute
23
25. Structural Link Section
The Structural Links section of METS allows METS creators to record
the existence of hyperlinks between nodes in the hierarchy
outlined in the Structural Map. This is of particular value in using
METS to archive Websites.
Image Source : www.loc.gov/standards/mets/presentations/METS.ppt
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
25
26. Structural Link Section
<div ID="P1" TYPE="page" LABEL="Page 1">
<fptr FILEID="HTMLF1"/>
<div ID="IMG1" TYPE="image" LABEL="Image Hyperlink to Page 2">
<fptr FILEID="JPGF1"/>
</div>
<div ID="P2" TYPE="page" LABEL="Page 2">
<fptr FILEID="HTMLF2"/> </div>
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
26
27. Behavior Section
A late addition to the METS standard, the behavior section allows
information to be recorded on how components of the digital
object should be rendered for the user, this may include
information on specific software packages to be used, or on
particular parameters to be used when rendering a file.
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
27
28. Introduction
Characteristics of METS
Sections of a METS
Current Users
Conclusion
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
28
29. Current users
– UC Berkeley (archival collections)
– Harvard (scanned print publications, e-journals)
– Library of Congress (audio/visual collections)
– EU Meta E project (historic newspapers)
– Michigan State (oral history collections)
– Univ. of Virginia (FEDORA digital objects)
– MIT (DSpace)
– RLG (RLG Cultural Materials)
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
29
33. Conclusion
• METS provides varied and flexible mechanisms for
– expressing structure or structures of a digital entity
– linking structure with simple and complex content
– linking structure with descriptive metadata
– linking structure and content files with administrative metadata
– linking behaviors with structure and content
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
33
34. Reference
METS official site:
http://www.loc.gov/standards/mets
Wikipedia
http://en.wikipedia.org/wiki/METS
UC Berkeley Library
http://ucblibrary4.berkeley.edu:8088/xdlib//calcultures/uci/mets
/cui_1_6_00004505.xml
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
34
35. Image Source : www.cornwall.gov.uk
Feb 19, 2014
Sandip Das, DRTC, Indian Statistical Institute
35