This document discusses network theory and its application to understanding and regulating the internet. It introduces concepts from complexity science like nodes, hubs, scale-free networks and power laws. Real world systems from citation networks to online communities demonstrate scale-free and small world properties. Understanding the internet as a complex, distributed network can help policymakers address issues like copyright, resilience to attacks, cybercrime and privacy in a way that accounts for emergent behaviors.
This document discusses the progress and next steps for Pharo, an open-source Smalltalk environment. Key points include:
- Since 2010, over 500 updates and 1000 issues closed through 5 sprints and a school on the VM.
- Versions 1.3 and 1.4 released with automated VM generation and testing.
- Hiring of full-time engineers to support development.
- Success in industry and community projects involving graphics, code model, serialization, and more.
- Upcoming work on a new canvas, compiler, source code model, serialization, scripting, and tools.
This document discusses crosswalking between Dublin Core and MARC metadata standards. It outlines key principles for crosswalking like embracing standards and exposing as much information content as practical. It then provides an illustrative mapping between Dublin Core elements and MARC 21 fields, showing how elements like Contributor map to 700 and 710 fields. The document also introduces CORC, a system that incorporates Dublin Core and MARC standards to automatically generate resource records and authority records and provide metadata in multiple formats.
This power point presentation gives an overview idea of basic features , structure and use of ISO 2709.Each of it's structural elements have been described with diagrams and the ultimate form of a ISO 2709 Bibliographic record has been illustrated with a picture.
Beyond the Record : OCLC & the Future of MARCtfons
1) The document discusses OCLC's role in metadata standards beyond MARC, including their crosswalk web service that translates between formats like MARC, MODS, DC, and ONIX.
2) It describes OCLC's involvement with RDA development and testing.
3) It introduces WorldCat Identities, which provides FRBRized views of bibliographic data clustered by work.
The document discusses the benefits and challenges of transitioning library data to linked data standards to make the data more accessible and interoperable on the web. It outlines principles of linked data and how library data could be transformed by assigning URIs to concepts, linking data sources, and storing data as RDF triples. Barriers include outdated library processes and standards like MARC that inhibit innovation, but initiatives like RDA, OpenLibrary, and data projects from the German National Library are helping advance the linked library data vision.
Todd Carpenter discusses NISO's Bibliographic Roadmap Initiative to develop a new bibliographic information ecosystem. The initiative aims to identify gaps, analyze economic impacts, and engage stakeholders in an open process to ensure the new system saves resources, improves services, and gains adoption. NISO held an initial meeting with librarians, publishers, and vendors to discuss priorities and gaps without designing specifications or favoring certain solutions. Next steps include prioritizing ideas, hosting discussion sessions, and releasing a report in the fall.
This document discusses network theory and its application to understanding and regulating the internet. It introduces concepts from complexity science like nodes, hubs, scale-free networks and power laws. Real world systems from citation networks to online communities demonstrate scale-free and small world properties. Understanding the internet as a complex, distributed network can help policymakers address issues like copyright, resilience to attacks, cybercrime and privacy in a way that accounts for emergent behaviors.
This document discusses the progress and next steps for Pharo, an open-source Smalltalk environment. Key points include:
- Since 2010, over 500 updates and 1000 issues closed through 5 sprints and a school on the VM.
- Versions 1.3 and 1.4 released with automated VM generation and testing.
- Hiring of full-time engineers to support development.
- Success in industry and community projects involving graphics, code model, serialization, and more.
- Upcoming work on a new canvas, compiler, source code model, serialization, scripting, and tools.
This document discusses crosswalking between Dublin Core and MARC metadata standards. It outlines key principles for crosswalking like embracing standards and exposing as much information content as practical. It then provides an illustrative mapping between Dublin Core elements and MARC 21 fields, showing how elements like Contributor map to 700 and 710 fields. The document also introduces CORC, a system that incorporates Dublin Core and MARC standards to automatically generate resource records and authority records and provide metadata in multiple formats.
This power point presentation gives an overview idea of basic features , structure and use of ISO 2709.Each of it's structural elements have been described with diagrams and the ultimate form of a ISO 2709 Bibliographic record has been illustrated with a picture.
Beyond the Record : OCLC & the Future of MARCtfons
1) The document discusses OCLC's role in metadata standards beyond MARC, including their crosswalk web service that translates between formats like MARC, MODS, DC, and ONIX.
2) It describes OCLC's involvement with RDA development and testing.
3) It introduces WorldCat Identities, which provides FRBRized views of bibliographic data clustered by work.
The document discusses the benefits and challenges of transitioning library data to linked data standards to make the data more accessible and interoperable on the web. It outlines principles of linked data and how library data could be transformed by assigning URIs to concepts, linking data sources, and storing data as RDF triples. Barriers include outdated library processes and standards like MARC that inhibit innovation, but initiatives like RDA, OpenLibrary, and data projects from the German National Library are helping advance the linked library data vision.
Todd Carpenter discusses NISO's Bibliographic Roadmap Initiative to develop a new bibliographic information ecosystem. The initiative aims to identify gaps, analyze economic impacts, and engage stakeholders in an open process to ensure the new system saves resources, improves services, and gains adoption. NISO held an initial meeting with librarians, publishers, and vendors to discuss priorities and gaps without designing specifications or favoring certain solutions. Next steps include prioritizing ideas, hosting discussion sessions, and releasing a report in the fall.
This document summarizes a presentation given at the NISO/BISG 7th Annual Forum on the Changing Standards Landscape. The presentation discusses the current bibliographic data ecosystem based on MARC standards and the movement toward linked data. It introduces NISO's Bibliographic Roadmap Initiative, which aims to identify gaps and engage stakeholders in developing a roadmap for the next generation of bibliographic data exchange to ensure interoperability, cost effectiveness, and adoption.
The document discusses the rise of big data and data science. It notes that while companies were dealing with medium sized data in the past, data is growing exponentially due to the internet and technologies like sensors. This growth is outpacing disk I/O performance, necessitating new database approaches like NoSQL and MapReduce. The key skills of data scientists are described as statistics, visualization, and data plumbing like cleaning, transforming and structuring large datasets. More data is also said to beat smart algorithms in many cases.
This document summarizes Todd Carpenter's presentation on designing a roadmap for a new bibliographic information ecosystem. It discusses how MARC has been the lingua franca of bibliographic data for decades but was originally designed to be efficient due to limited and expensive computer storage. However, much computer technology now pre-dates MARC. There is also growing movement toward linked bibliographic data. The presentation notes challenges in moving away from MARC, including lack of demonstrable benefits from a new system. NISO's Bibliographic Roadmap Initiative aims to identify gaps, engage stakeholders, and provide an open process to help assure the right approaches are taken to improve services and facilitate adoption.
Data Storage Considerations for the Tactical Field Collection of Digital Imag...Robert Watson
This document discusses data storage considerations for tactical field collection of digital imagery. It examines various storage media options including magnetic tape, optical disks, magnetic hard disks, and solid state drives. Optical disks provide high storage capacity and longevity, making them well-suited for long-term storage. However, hard disks have faster access times and have become more affordable and higher capacity over time. The document also covers video compression standards and metadata for imagery as well as software and hardware recommendations for imagery storage and retrieval systems.
Todd Carpenter gave a presentation on content distribution standards at the Council of Science Editors Conference on May 20, 2012. He discussed how NISO develops standards to reduce chaos in content distribution by bringing together publishers, libraries, and automation vendors. He provided examples of NISO standards like ISSN, DOI, and KBART that help identify and describe content. Carpenter emphasized that identifiers and metadata are key to standards and explained concepts like functional granularity. Overall, he highlighted NISO's role in developing interoperability standards to help distribute scientific content.
BUILDING A SCALABLE MULTIMEDIA WEB OBSERVATORYJonathon Hare
Web and Internet Science research group seminar series. University of Southampton. 13th March 2013.
The web is inherently multimedia in nature, and contains data and information in many different audio, visual and textual forms. To fully understand the nature of the web and the information contained within it, it is necessary to harness all modalities of data. Within the EU funded ARCOMEM project, we are building a platform for crawling and analysing samples of web and social-web data at scale. Whilst the project is ostensibly about issues related to intelligent web-archiving, the ARCOMEM software has features that make it ideal for use as a platform for a scalable Multimedia Web Observatory.
This talk will describe the ARCOMEM approach from data harvesting through to detailed content analysis and demonstrate how this approach relates to a multimedia web observatory. In addition to describing the overall framework, I'll show some of the research aspects of the system related specifically to multimodal multimedia data in small (>100GB) to medium-scale (multi-terabyte) web archives, and demonstrate how these are targeted to our Parliamentarian and Journalist end-users.
The document discusses Frictionless Data, an initiative by the Open Knowledge Foundation to make research data easier to share, consume, and analyze. It aims to introduce standards and tools to "containerize" datasets using simple specifications like Tabular Data Package. This would make data easier to integrate into tools and platforms, find, maintain quality for, and analyze. It discusses problems like lack of standards, tools to validate datasets are presented. Examples of early implementations that integrate validation checks and continuous validation are also provided.
Presented in May 2010
This presentation goes through the Wireshark network analyzer. It presents an overview of the different features that I've found useful while doing network performance analysis for ICS network protocols.
Internet infrastructure and the history of the world wide web presentation, f...Graham Garner
The document discusses the history and development of the Internet and World Wide Web. It outlines several key milestones from the creation of ARPANET in 1969 to the growth of social media and mobile internet. Charts show global internet usage has grown from 361 million users in 2000 to over 2.2 billion users in 2011, with Asia having the most users. The document also describes the original domain name system of country-code top-level domains and seven generic top-level domains.
The document discusses performance tuning of an ERP system like Banner, which is known to be difficult due to systemic issues across platforms. It provides an overview of concepts to help IT teams identify and solve internal datacenter problems, such as monitoring database, network, servers, applications, and gathering metrics on hardware and software configurations. The presentation also covers specific factors that affect Banner/Oracle performance, including product mix, queries, hardware settings, network saturation, processes, and user desktop tools. Metrics and usage patterns are monitored to analyze bottlenecks and prioritize areas for tuning the application, database, and hardware.
This document provides an overview of research data and the data lifecycle. It discusses the creation, processing, analysis, preservation, and reuse of data. It also addresses metadata, data repositories, and challenges around long-tail and big data. The key points are: research data goes through stages from creation to reuse; metadata is critical for documenting and defining data; data repositories curate data and facilitate access and preservation; and there are differences between standardized big data and more bespoke long-tail data. Effectively managing both is important for reproducibility and trust in scientific results.
This document provides an overview of research data and the data lifecycle. It discusses the creation, processing, analysis, preservation, and reuse of data. It also addresses metadata, data repositories, and challenges around long-tail and big data. The key points are: research data goes through stages from creation to reuse; metadata is critical for documenting and defining data; data repositories curate data and facilitate access and preservation; and there are differences between standardized big data and more bespoke long-tail data. Reproducibility, transparency, and ensuring data is well cared for are important responsibilities for scientists.
Requirements for Managing Unstructured DataDATAVERSITY
The document discusses requirements for managing unstructured data. It identifies common frustrations users have finding information across multiple document systems. It then outlines techniques to analyze current systems, including user surveys, process mapping, and analyzing metadata. The document stresses the need for standard metadata, centralized storage, and improved search tools. Mapping user needs to system capabilities could improve efficiency and save over $1 million annually by reducing wasted search time.
1) NISO held a webinar to discuss the outcomes of their Bibliographic Roadmap symposium in April 2013 which aimed to identify and prioritize topics for further exploration of bibliographic data exchange, identify gaps, and organize next steps.
2) The webinar recapped the in-person meeting and presented the results of a post-meeting survey ranking discussion topics.
3) Next steps discussed included identifying 3-5 projects for deeper discussion, convening expert groups on each topic, arranging calls with experts, a roundtable at ALA Midwinter, and releasing a final report in March 2014.
All Things Open 2014 - Day 1
Wednesday, October 22nd, 2014
Dan Bedard
Market Development Manager for iRODS Consortium, RENCI at UNC Chapel Hill
Lunch Session
Building the iRODS Consortium
Tim Malthus_Towards standards for the exchange of field spectral datasetsTERN Australia
This document discusses the development of standards for the exchange of field spectral datasets. It notes the importance of metadata for determining the quality and representativeness of spectral data obtained in the field. A workshop was held in 2012 to discuss best practices for data collection and exchange and key conclusions included the need for standards to facilitate accurate comparison across studies and the role of thorough metadata. Work is ongoing to enhance the SPECCHIO system for hosting spectral libraries and metadata and establishing it as the international tool for storage and exchange of spectral datasets.
This document provides an extensive summary of Wayne Schroeder's professional experience and expertise. It outlines over 38 years of experience in software engineering, data management, and scientific computing. Some key points include:
- He is an expert in iRODS and designed/implemented major components of the iRODS data management system.
- He has over 10 years of experience as a team lead and manager on projects like iRODS.
- His experience includes roles at organizations like the San Diego Supercomputer Center, Lawrence Livermore National Laboratory, and Entropia.
Wayne Schroeder is an expert in distributed data management and iRODS with over 38 years of experience in software engineering, data management, and scientific computing. He designed and implemented major components of iRODS and provided support to the international iRODS user community for over 12 years. He has extensive experience managing projects and leading teams, and currently owns a consulting business providing data management consulting services and products.
This document provides an introduction and overview of the INF2190 - Data Analytics course. It discusses the instructor, Attila Barta, details on where and when the course will take place. It then provides definitions and history of data analytics, discusses how the field has evolved with big data, and references enterprise data analytics architectures. It contrasts traditional vs. big data era data analytics approaches and tools. The objective of the course is described as providing students with the foundation to become data scientists.
This presentation was provided by Racquel Jemison, Ph.D., Christina MacLaughlin, Ph.D., and Paulomi Majumder. Ph.D., all of the American Chemical Society, for the second session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session Two: 'Expanding Pathways to Publishing Careers,' was held June 13, 2024.
This presentation was provided by Rebecca Benner, Ph.D., of the American Society of Anesthesiologists, for the second session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session Two: 'Expanding Pathways to Publishing Careers,' was held June 13, 2024.
More Related Content
Similar to 13 0213 w3c - carpenter - Designing a Roadmap to a New Bibliographic Information Ecosystem
This document summarizes a presentation given at the NISO/BISG 7th Annual Forum on the Changing Standards Landscape. The presentation discusses the current bibliographic data ecosystem based on MARC standards and the movement toward linked data. It introduces NISO's Bibliographic Roadmap Initiative, which aims to identify gaps and engage stakeholders in developing a roadmap for the next generation of bibliographic data exchange to ensure interoperability, cost effectiveness, and adoption.
The document discusses the rise of big data and data science. It notes that while companies were dealing with medium sized data in the past, data is growing exponentially due to the internet and technologies like sensors. This growth is outpacing disk I/O performance, necessitating new database approaches like NoSQL and MapReduce. The key skills of data scientists are described as statistics, visualization, and data plumbing like cleaning, transforming and structuring large datasets. More data is also said to beat smart algorithms in many cases.
This document summarizes Todd Carpenter's presentation on designing a roadmap for a new bibliographic information ecosystem. It discusses how MARC has been the lingua franca of bibliographic data for decades but was originally designed to be efficient due to limited and expensive computer storage. However, much computer technology now pre-dates MARC. There is also growing movement toward linked bibliographic data. The presentation notes challenges in moving away from MARC, including lack of demonstrable benefits from a new system. NISO's Bibliographic Roadmap Initiative aims to identify gaps, engage stakeholders, and provide an open process to help assure the right approaches are taken to improve services and facilitate adoption.
Data Storage Considerations for the Tactical Field Collection of Digital Imag...Robert Watson
This document discusses data storage considerations for tactical field collection of digital imagery. It examines various storage media options including magnetic tape, optical disks, magnetic hard disks, and solid state drives. Optical disks provide high storage capacity and longevity, making them well-suited for long-term storage. However, hard disks have faster access times and have become more affordable and higher capacity over time. The document also covers video compression standards and metadata for imagery as well as software and hardware recommendations for imagery storage and retrieval systems.
Todd Carpenter gave a presentation on content distribution standards at the Council of Science Editors Conference on May 20, 2012. He discussed how NISO develops standards to reduce chaos in content distribution by bringing together publishers, libraries, and automation vendors. He provided examples of NISO standards like ISSN, DOI, and KBART that help identify and describe content. Carpenter emphasized that identifiers and metadata are key to standards and explained concepts like functional granularity. Overall, he highlighted NISO's role in developing interoperability standards to help distribute scientific content.
BUILDING A SCALABLE MULTIMEDIA WEB OBSERVATORYJonathon Hare
Web and Internet Science research group seminar series. University of Southampton. 13th March 2013.
The web is inherently multimedia in nature, and contains data and information in many different audio, visual and textual forms. To fully understand the nature of the web and the information contained within it, it is necessary to harness all modalities of data. Within the EU funded ARCOMEM project, we are building a platform for crawling and analysing samples of web and social-web data at scale. Whilst the project is ostensibly about issues related to intelligent web-archiving, the ARCOMEM software has features that make it ideal for use as a platform for a scalable Multimedia Web Observatory.
This talk will describe the ARCOMEM approach from data harvesting through to detailed content analysis and demonstrate how this approach relates to a multimedia web observatory. In addition to describing the overall framework, I'll show some of the research aspects of the system related specifically to multimodal multimedia data in small (>100GB) to medium-scale (multi-terabyte) web archives, and demonstrate how these are targeted to our Parliamentarian and Journalist end-users.
The document discusses Frictionless Data, an initiative by the Open Knowledge Foundation to make research data easier to share, consume, and analyze. It aims to introduce standards and tools to "containerize" datasets using simple specifications like Tabular Data Package. This would make data easier to integrate into tools and platforms, find, maintain quality for, and analyze. It discusses problems like lack of standards, tools to validate datasets are presented. Examples of early implementations that integrate validation checks and continuous validation are also provided.
Presented in May 2010
This presentation goes through the Wireshark network analyzer. It presents an overview of the different features that I've found useful while doing network performance analysis for ICS network protocols.
Internet infrastructure and the history of the world wide web presentation, f...Graham Garner
The document discusses the history and development of the Internet and World Wide Web. It outlines several key milestones from the creation of ARPANET in 1969 to the growth of social media and mobile internet. Charts show global internet usage has grown from 361 million users in 2000 to over 2.2 billion users in 2011, with Asia having the most users. The document also describes the original domain name system of country-code top-level domains and seven generic top-level domains.
The document discusses performance tuning of an ERP system like Banner, which is known to be difficult due to systemic issues across platforms. It provides an overview of concepts to help IT teams identify and solve internal datacenter problems, such as monitoring database, network, servers, applications, and gathering metrics on hardware and software configurations. The presentation also covers specific factors that affect Banner/Oracle performance, including product mix, queries, hardware settings, network saturation, processes, and user desktop tools. Metrics and usage patterns are monitored to analyze bottlenecks and prioritize areas for tuning the application, database, and hardware.
This document provides an overview of research data and the data lifecycle. It discusses the creation, processing, analysis, preservation, and reuse of data. It also addresses metadata, data repositories, and challenges around long-tail and big data. The key points are: research data goes through stages from creation to reuse; metadata is critical for documenting and defining data; data repositories curate data and facilitate access and preservation; and there are differences between standardized big data and more bespoke long-tail data. Effectively managing both is important for reproducibility and trust in scientific results.
This document provides an overview of research data and the data lifecycle. It discusses the creation, processing, analysis, preservation, and reuse of data. It also addresses metadata, data repositories, and challenges around long-tail and big data. The key points are: research data goes through stages from creation to reuse; metadata is critical for documenting and defining data; data repositories curate data and facilitate access and preservation; and there are differences between standardized big data and more bespoke long-tail data. Reproducibility, transparency, and ensuring data is well cared for are important responsibilities for scientists.
Requirements for Managing Unstructured DataDATAVERSITY
The document discusses requirements for managing unstructured data. It identifies common frustrations users have finding information across multiple document systems. It then outlines techniques to analyze current systems, including user surveys, process mapping, and analyzing metadata. The document stresses the need for standard metadata, centralized storage, and improved search tools. Mapping user needs to system capabilities could improve efficiency and save over $1 million annually by reducing wasted search time.
1) NISO held a webinar to discuss the outcomes of their Bibliographic Roadmap symposium in April 2013 which aimed to identify and prioritize topics for further exploration of bibliographic data exchange, identify gaps, and organize next steps.
2) The webinar recapped the in-person meeting and presented the results of a post-meeting survey ranking discussion topics.
3) Next steps discussed included identifying 3-5 projects for deeper discussion, convening expert groups on each topic, arranging calls with experts, a roundtable at ALA Midwinter, and releasing a final report in March 2014.
All Things Open 2014 - Day 1
Wednesday, October 22nd, 2014
Dan Bedard
Market Development Manager for iRODS Consortium, RENCI at UNC Chapel Hill
Lunch Session
Building the iRODS Consortium
Tim Malthus_Towards standards for the exchange of field spectral datasetsTERN Australia
This document discusses the development of standards for the exchange of field spectral datasets. It notes the importance of metadata for determining the quality and representativeness of spectral data obtained in the field. A workshop was held in 2012 to discuss best practices for data collection and exchange and key conclusions included the need for standards to facilitate accurate comparison across studies and the role of thorough metadata. Work is ongoing to enhance the SPECCHIO system for hosting spectral libraries and metadata and establishing it as the international tool for storage and exchange of spectral datasets.
This document provides an extensive summary of Wayne Schroeder's professional experience and expertise. It outlines over 38 years of experience in software engineering, data management, and scientific computing. Some key points include:
- He is an expert in iRODS and designed/implemented major components of the iRODS data management system.
- He has over 10 years of experience as a team lead and manager on projects like iRODS.
- His experience includes roles at organizations like the San Diego Supercomputer Center, Lawrence Livermore National Laboratory, and Entropia.
Wayne Schroeder is an expert in distributed data management and iRODS with over 38 years of experience in software engineering, data management, and scientific computing. He designed and implemented major components of iRODS and provided support to the international iRODS user community for over 12 years. He has extensive experience managing projects and leading teams, and currently owns a consulting business providing data management consulting services and products.
This document provides an introduction and overview of the INF2190 - Data Analytics course. It discusses the instructor, Attila Barta, details on where and when the course will take place. It then provides definitions and history of data analytics, discusses how the field has evolved with big data, and references enterprise data analytics architectures. It contrasts traditional vs. big data era data analytics approaches and tools. The objective of the course is described as providing students with the foundation to become data scientists.
Similar to 13 0213 w3c - carpenter - Designing a Roadmap to a New Bibliographic Information Ecosystem (20)
This presentation was provided by Racquel Jemison, Ph.D., Christina MacLaughlin, Ph.D., and Paulomi Majumder. Ph.D., all of the American Chemical Society, for the second session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session Two: 'Expanding Pathways to Publishing Careers,' was held June 13, 2024.
This presentation was provided by Rebecca Benner, Ph.D., of the American Society of Anesthesiologists, for the second session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session Two: 'Expanding Pathways to Publishing Careers,' was held June 13, 2024.
This presentation was provided by Steph Pollock of The American Psychological Association’s Journals Program, and Damita Snow, of The American Society of Civil Engineers (ASCE), for the initial session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session One: 'Setting Expectations: a DEIA Primer,' was held June 6, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the closing segment of the NISO training series "AI & Prompt Design." Session Eight: Limitations and Potential Solutions, was held on May 23, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the seventh segment of the NISO training series "AI & Prompt Design." Session 7: Open Source Language Models, was held on May 16, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the sixth segment of the NISO training series "AI & Prompt Design." Session Six: Text Classification with LLMs, was held on May 9, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the fifth segment of the NISO training series "AI & Prompt Design." Session Five: Named Entity Recognition with LLMs, was held on May 2, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the fourth segment of the NISO training series "AI & Prompt Design." Session Four: Structured Data and Assistants, was held on April 25, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the third segment of the NISO training series "AI & Prompt Design." Session Three: Beginning Conversations, was held on April 18, 2024.
This presentation was provided by Kaveh Bazargan of River Valley Technologies, during the NISO webinar "Sustainability in Publishing." The event was held April 17, 2024.
This presentation was provided by Dana Compton of the American Society of Civil Engineers (ASCE), during the NISO webinar "Sustainability in Publishing." The event was held April 17, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, during the second segment of the NISO training series "AI & Prompt Design." Session Two: Large Language Models, was held on April 11, 2024.
This presentation was provided by Teresa Hazen of the University of Arizona, Geoff Morse of Northwestern University. and Ken Varnum of the University of Michigan, during the Spring ODI Conformance Statement Workshop for Libraries. This event was held on April 9, 2024
This presentation was provided by William Mattingly of the Smithsonian Institution, during the opening segment of the NISO training series "AI & Prompt Design." Session One: Introduction to Machine Learning, was held on April 4, 2024.
This presentation was provided by William Mattingly of the Smithsonian Institution, for the eight and final session of NISO's 2023 Training Series on Text and Data Mining. Session eight, "Building Data Driven Applications" was held on Thursday, December 7, 2023.
This presentation was provided by William Mattingly of the Smithsonian Institution, for the seventh session of NISO's 2023 Training Series on Text and Data Mining. Session seven, "Vector Databases and Semantic Searching" was held on Thursday, November 30, 2023.
This presentation was provided by William Mattingly of the Smithsonian Institution, for the sixth session of NISO's 2023 Training Series on Text and Data Mining. Session six, "Text Mining Techniques" was held on Thursday, November 16, 2023.
This presentation was provided by William Mattingly of the Smithsonian Institution, for the fifth session of NISO's 2023 Training Series on Text and Data Mining. Session five, "Text Processing for Library Data" was held on Thursday, November 9, 2023.
This presentation was provided by Todd Carpenter, Executive Director, during the NISO webinar on "Strategic Planning." The event was held virtually on November 8, 2023.
More from National Information Standards Organization (NISO) (20)
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Project Management Semester Long Project - Acuityjpupo2018
Acuity is an innovative learning app designed to transform the way you engage with knowledge. Powered by AI technology, Acuity takes complex topics and distills them into concise, interactive summaries that are easy to read & understand. Whether you're exploring the depths of quantum mechanics or seeking insight into historical events, Acuity provides the key information you need without the burden of lengthy texts.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
13 0213 w3c - carpenter - Designing a Roadmap to a New Bibliographic Information Ecosystem
1. Designing a Roadmap to a
New Bibliographic Information
Ecosystem
Understanding a complex
environment
Todd A. Carpenter, NISO Executive Director
eBooks: Great Expectations for Web Standards
February 12, 2013 - New York, NY
Tuesday, February 12, 13
2. About
• Non-profit industry trade association accredited by ANSI
with 150+ members
• Mission of developing and maintaining technical standards
related to information, documentation, discovery and
distribution of published materials and media
• Volunteer driven organization: 400+ spread out across
the world
• Represent US interests to ISO TC 46 & also serve as
Secretariat for ISO TC46/SC9 - Identification &
Description
• Responsible for standards like ISSN, DOI, Dublin Core
metadata, DAISY digital talking books, OpenURL, MARC
records, and ISBN (indirectly)
Tuesday, February 12, 13
4. 32% Libraries/Library 35 % Publishers/Publishing
Organizations Organizations
36 LSA Members
72 LSA Members
(non-voting)
ANSI
ISO
Other SDOs
33% Library Systems Suppliers,
Publishing Vendors & Intermediaries
NISO’s Community
Tuesday, February 12, 13
5. NISO Internationally
Actively participate internationally with ISO, EDItEUR,
IFLA, ICSTI, International STM Association, CODATA,
UK Serials Group, LIBER, Standards Australia, IETF,
ISO Registration Authorities
Tuesday, February 12, 13
6. Technical Committee (TC) 46
Information & Documentation
Subcommittees (SC):
4 – Systems Interoperability
8 – Performance Measurement
9 – Identification & Description
11 – Records Management
NISO manages the Secretariat of ISO TC 46, SC 9
Tuesday, February 12, 13
7. If you’ve never met MARC before...
01386cam 2200301 a
45000010008000000050017000080080041000250350021000669060045000879550027001320100017001590200015
00176040001800191043001200209050002200221082002100243110005500264245030800319260006700627300002
50069444000540071950000290077365000600080265000580086265000730092071000430099399100480103638568
5319951219150001.4881118s1989 nju 000 0 eng 9(DLC) 88029610 a7bcbccorignewd1eocipf19gy-
gencatlg aCIP ver. pv04 12-06-95 a 88029610 a0887389538 aDLCcDLCdDLC an-us---00aZ674.8b.N44
198900a021.6/5/09732192 aNational Information Standards Organization (U.S.)10aInformation retrieval service and
protocol :bAmerican national standard for information retrieval service definition and protocol specification for
library applications /capproved January 15, 1988 by American National Standards Institute ; developed by the National
Information Standards Organization. aNew Brunswick, N.J., U.S.A. :bTransaction Publishers,cc1989. axii, 50 p. ;c26
cm. 0aNational information standards series,x1041-5653 a"ANSI/NISO Z39.50-1988." 0aLibrary information
networksxStandardszUnited States. 0aComputer network protocolsxStandardszUnited States. 0aInformation storage
and retrieval systemsxStandardszUnited States.2 aAmerican National Standards Institute. bc-GenCollhZ674.8i.N44
1989tCopy 1wBOOKS
Tuesday, February 12, 13
8. If you’ve never met MARC before...
(formatted for your viewing pleasure)
Tuesday, February 12, 13
9. MARC Components
Encoding Structure
Z39.2
ISO 2709:2008 -- Format for information exchange
Format structure
Anglo-American Cataloging Rules (2nd Edition) AACR2
Resource Description & Access
Exchange System
Z39.50
SRU/SRW
Tuesday, February 12, 13
17. If you were building a network today
would you string copper everywhere?
Tuesday, February 12, 13
18. If you building a metadata ecosystem,
would you start here?
01386cam 2200301 a
45000010008000000050017000080080041000250350021000669060045000879550027001320100017001590200015
00176040001800191043001200209050002200221082002100243110005500264245030800319260006700627300002
50069444000540071950000290077365000600080265000580086265000730092071000430099399100480103638568
5319951219150001.4881118s1989 nju 000 0 eng 9(DLC) 88029610 a7bcbccorignewd1eocipf19gy-
gencatlg aCIP ver. pv04 12-06-95 a 88029610 a0887389538 aDLCcDLCdDLC an-us---00aZ674.8b.N44
198900a021.6/5/09732192 aNational Information Standards Organization (U.S.)10aInformation retrieval service and
protocol :bAmerican national standard for information retrieval service definition and protocol specification for
library applications /capproved January 15, 1988 by American National Standards Institute ; developed by the National
Information Standards Organization. aNew Brunswick, N.J., U.S.A. :bTransaction Publishers,cc1989. axii, 50 p. ;c26
cm. 0aNational information standards series,x1041-5653 a"ANSI/NISO Z39.50-1988." 0aLibrary information
networksxStandardszUnited States. 0aComputer network protocolsxStandardszUnited States. 0aInformation storage
and retrieval systemsxStandardszUnited States.2 aAmerican National Standards Institute. bc-GenCollhZ674.8i.N44
1989tCopy 1wBOOKS
Tuesday, February 12, 13
22. Movement toward
linked data
datahub.io - 5107 data stores
id.loc.gov
VIAF
OCLC WorldCat Linked Data Store
W3C Library Linked Data Incubator Group
Tuesday, February 12, 13
23. Organizations will not move away from a legacy system
unless the new system:
a) Is demonstrably cheaper
b) Is demonstrably more effective in producing results
(discovery, use, etc.)
c) Will make the organization demonstrably more
efficient (staff, management, sales, etc.)
OR
d) The legacy system becomes entirely
non-interoperable with other, more important systems
OR
e) The legacy system breaks and cannot be repaired
Tuesday, February 12, 13
24. Can we say a new
metadata management system
will be/do one of those things?
Tuesday, February 12, 13
26. Initiative coordination
Gap identification
Economic analysis
Engage diverse players
Open process
Tuesday, February 12, 13
27. Defining what you are by
what you are not
Not desinging a spec
Not picking winners/losers
Tuesday, February 12, 13
28. Open Community
Meeting
April 15-16, 2013
Baltimore/Washington
area
Tuesday, February 12, 13
29. Meeting Goals:
Identify/Discuss active projects
Brainstorm existing gaps
Prioritize 3-6 subtopics
Plan deeper exploration of issues
Diversify players in discussion
Tuesday, February 12, 13
30. What we are trying to avoid
Tuesday, February 12, 13
31. The world
makes way for
the man who
knows where
he is going.
- Ralph Waldo Emerson
Tuesday, February 12, 13
32. “If you don't know where
you're going, you might not
get there.”
- Yogi Berra
Tuesday, February 12, 13
33. Thank you!
Todd Carpenter, Executive Director
tcarpenter@niso.org
National Information Standards Organization (NISO)
One North Charles Street, Suite 1905
Baltimore, MD 21201 USA
+1 (301) 654-2512
www.niso.org
Tuesday, February 12, 13