This document provides guidance on preparing CSV files for importing archival descriptions into AtoM. It discusses:
- CSV templates and examples available from AtoM
- Column headers and field mappings between different description standards
- Using legacyId, parentId, and qubitParentSlug fields to represent hierarchical relationships
- Formatting dates, creators, locations, and other fields
- Linking imported data to existing authorities, accessions, or other records in AtoM
- Considerations for character encoding, line endings, and using tools like LibreOffice Calc to prepare CSV files
Proper preparation of the CSV file is important to ensure a successful import into AtoM without errors.
These slides accompanied a presentation by Dan Gillean and Sara Allain of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In this session, Dan Gillean covered some of key considerations when preparing to migrate data from a legacy source system into AtoM. The slides cover the steps involved, paying specific attention to AtoM's data model and behaviors that should be considered when preparing your data for import into AtoM.
These slides were first publicly shared as part of the 2019 AtoM Camp, held at the University of Westminster in London, UK, July 15-17, 2019.
The slides include tips, tricks, workarounds, warnings, and gotchas about AtoM that can help orient new users - and hopefully even teach experienced users a couple new things!
For more information about Access to Memory (AtoM), see:
* https://www.accesstomemory.org
These slides accompanied a presentation by Dan Gillean and Sara Allain of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In this session, we took a quick tour of some examples of how AtoM is being implemented by our global community of users. We looked specifically at interesting themes, customizations, or the creative use of existing features such as static pages or repository theming as a way of exploring some of the different ways AtoM can be used. Participants were then invited to come up and show off their AtoM site to the attendees as well.
Slides accompanying a day-long AtoM and Archivematica workshop, delivered October 17, 2016 at the VII Congresso Nacional de Arquivologia in Fortaleza, Brazil.
These slides introduce AtoM (short for Access to Memory), a flexible, open-source application for standards-based description and access in a multilingual, multi-repository environment.
For more information, visit https://accesstomemory.org/
Access to Memory (AtoM) is an open source web application for standards-based archival description and access - learn more at:
https://www.accesstomemory.org
These slides will provide users with an overview of how search works in AtoM, along with a detailed walkthrough of using the Advanced search panel, and performing Expert searches in AtoM.
The slides were originally created by Dan Gillean, AtoM Program Manager, for use in a series of training workshops delivered July 9-13, 2018 at the University of the Witswatersrand in Johannesburg, South Africa. The slides are based on current functionality in AtoM release 2.4.
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides provide advanced users with an overview of AtoM's data model, and demonstrate how a graphical user interface application such as MySQL Workbench can be used to explore the AtoM MySQL database from the back-end. The slides include a number of example queries.
Slides accompanying a day-long introduction to AtoM and Archivematica, presented by Dan Gillean and Justin Simpson at the UK National Archives as part of an AIM25 and Higher Education Archive Programme Network Meeting, December 2, 2016.
AtoM and Vagrant: Installing and Configuring the AtoM Vagrant Box for Local T...Artefactual Systems - AtoM
Â
These slides introduce AtoM users to Vagrant, and walk users through the process of installing the the AtoM Vagrant box for local testing and development on a home computer or laptop, regardless of what operating system you use.
WARNINGS:
These slides were last updated in May 2017, using the AtoM 2.4 Vagrant box, which is installed using Ubuntu 16.04. and PHP 7.0. Future versions of AtoM may use a different version of Ubuntu and PHP, which might change some of the command-line tasks used to update the box in Part 2. Be sure to check the AtoM documentation for the most up-to-date information: https://www.accesstomemory.org/docs/latest/
The AtoM Vagrant box is designed for local testing and development - it is NOT PRODUCTION READY and should not be used for long-term data storage. Please see the AtoM documentation for instructions on how to install AtoM on a server for use in your institution.
These slides accompanied a presentation by Dan Gillean and Sara Allain of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In this session, Dan Gillean covered some of key considerations when preparing to migrate data from a legacy source system into AtoM. The slides cover the steps involved, paying specific attention to AtoM's data model and behaviors that should be considered when preparing your data for import into AtoM.
These slides were first publicly shared as part of the 2019 AtoM Camp, held at the University of Westminster in London, UK, July 15-17, 2019.
The slides include tips, tricks, workarounds, warnings, and gotchas about AtoM that can help orient new users - and hopefully even teach experienced users a couple new things!
For more information about Access to Memory (AtoM), see:
* https://www.accesstomemory.org
These slides accompanied a presentation by Dan Gillean and Sara Allain of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In this session, we took a quick tour of some examples of how AtoM is being implemented by our global community of users. We looked specifically at interesting themes, customizations, or the creative use of existing features such as static pages or repository theming as a way of exploring some of the different ways AtoM can be used. Participants were then invited to come up and show off their AtoM site to the attendees as well.
Slides accompanying a day-long AtoM and Archivematica workshop, delivered October 17, 2016 at the VII Congresso Nacional de Arquivologia in Fortaleza, Brazil.
These slides introduce AtoM (short for Access to Memory), a flexible, open-source application for standards-based description and access in a multilingual, multi-repository environment.
For more information, visit https://accesstomemory.org/
Access to Memory (AtoM) is an open source web application for standards-based archival description and access - learn more at:
https://www.accesstomemory.org
These slides will provide users with an overview of how search works in AtoM, along with a detailed walkthrough of using the Advanced search panel, and performing Expert searches in AtoM.
The slides were originally created by Dan Gillean, AtoM Program Manager, for use in a series of training workshops delivered July 9-13, 2018 at the University of the Witswatersrand in Johannesburg, South Africa. The slides are based on current functionality in AtoM release 2.4.
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides provide advanced users with an overview of AtoM's data model, and demonstrate how a graphical user interface application such as MySQL Workbench can be used to explore the AtoM MySQL database from the back-end. The slides include a number of example queries.
Slides accompanying a day-long introduction to AtoM and Archivematica, presented by Dan Gillean and Justin Simpson at the UK National Archives as part of an AIM25 and Higher Education Archive Programme Network Meeting, December 2, 2016.
AtoM and Vagrant: Installing and Configuring the AtoM Vagrant Box for Local T...Artefactual Systems - AtoM
Â
These slides introduce AtoM users to Vagrant, and walk users through the process of installing the the AtoM Vagrant box for local testing and development on a home computer or laptop, regardless of what operating system you use.
WARNINGS:
These slides were last updated in May 2017, using the AtoM 2.4 Vagrant box, which is installed using Ubuntu 16.04. and PHP 7.0. Future versions of AtoM may use a different version of Ubuntu and PHP, which might change some of the command-line tasks used to update the box in Part 2. Be sure to check the AtoM documentation for the most up-to-date information: https://www.accesstomemory.org/docs/latest/
The AtoM Vagrant box is designed for local testing and development - it is NOT PRODUCTION READY and should not be used for long-term data storage. Please see the AtoM documentation for instructions on how to install AtoM on a server for use in your institution.
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides are intended for developers who are interested in modifying the default look and feel of AtoM - known as the Dominion theme - and developing a custom theme plugin. They include some theme examples, how to register a plugin in Symfony, and some ideas of the elements you can modify via theming, with examples.
These slides were presented by Dan Gillean at the first ever AtoM Camp, held at the SFU Harbour Centre in Vancouver, BC, Canada, March 20-22, 2017. For more information on the Camp, see:
https://wiki.accesstomemory.org/Community/Camps/SFU2017
These slides introduce some of the primary Command Line tasks available in Access to Memory (AtoM), which can be used to support site maintenance and troubleshooting. For further information on AtoM CLI tasks, see:
* https://www.accesstomemory.org/docs/latest/admin-manual/maintenance/cli-tools/
* https://www.youtube.com/playlist?list=PLZiwlG5eSMeyeETe15EsEBSu5htPLK-wm
The CLI tasks introduced in these slides are based on the 2.4 AtoM release - some tasks or task options may not be available in earlier releases; tasks may be changed in future releases.
Building the Future Together: AtoM3, Governance, and the Sustainability of Op...Artefactual Systems - AtoM
Â
Slides accompanying a presentation given by Dan Gillean on June 7th, 2018 at Open Repositories 2018, held in Bozeman, MT.
Access to Memory is a web-based open source application for standards based description and access. AtoM was first released in 2008 and much of the codebase is now relying on deprecated frameworks and libraries – and at the same time, new standards and technologies are changing how our profession approaches description and access. Currently Artefactual Systems, a Canadian based company, uses a services model to support the project. Artefactual is looking ahead to AtoM3, and considering building a linked data driven platform for archival description and access. As we consider AtoM's next generation, we are also examining governance and maintenance models to sustain the project and better empower our user community as Artefactual wasn't originally intended to be AtoM's organizational home. This presentation will offer some thoughts on existing open source project governance models, challenges, and possibilities for the future. How do we ensure community engagement and project sustainability over time?
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see: https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides introduce users to the Access to Memory codebase. You can find AtoM's code at: https://github.com/artefactual/atom/
Practical information for Alfresco integration with AOS (Sharepoint Protocol), Google Drive, Microsoft 365, ONLYOFFICE and Collabora Online.
Additionally ADW support for ONLYOFFICE is provided by https://github.com/atolcd/adf-onlyoffice-extension#installation
Slides accompanying a presentation delivered at the VII Congresso Nacional de Arquivologia in Fortaleza, Brazil, on October 19th, 2016. The slides provide an overview of the AtoM project's history, its maintenance by Artefactual, and its development philosophy, before proceeding to examine the application as a component used in a digital preservation ecosystem. Aspects of ISO 16363:2012, the Audit and Certification of Trustworthy Digital Repositories standard, are used to evaluate how AtoM can support description, management, administration, and access functions when used to maintain a chain of custody in a trustworthy digital repository ecosystem.
Dynamic Components using Single-Page-Application Concepts in AEM/CQNetcetera
Â
Dynamic components display content dependable on context, hence they cannot be cached.
Out of the box, Adobe Experience Manager doesn't give us many options for granular caching on a component level.
When faced with this problem, we usually resort to developing components that are leveraging Server-Side Includes or AJAX to get the HTML with dynamic data.
As an alternative solution, we have also developed dynamic components that use Single Page Application concepts, by using templates and JSON-responses, to provide the same dynamic behavior.
In this presentation we will cover all of the dynamic components types, compare the benefits and drawbacks of each, and state the use-cases where each can be effectively applied.
We will take a deeper look at the dynamic components done with SPA concepts, as they are rarity in the AEM world, and also provide a walk-through of the technologies used, how some common problems were solved, as well as the benefits that have been gained by their usage.
These slides accompany a 1.5 hour webinar sponsored by the Western New York Library Resources Council, presented by Dan Gillean of Artefactual Systems on February 15th, 2017.
The session was intended to introduce participants to some of the key standards, services, and tools available to support digital preservation planning and activities. Part 1 focused on DP101, and how to begin tackling digital preservation in your institution. Part 2 introduced the Archivematica project's history, philosophy, and aims, while Part 3 was a live demonstration of Archivematica in action.
Thank you to WNYLRC for sponsoring this event!
by Dhanraj Pondicherry, Sr. Solutions Architecture Manager, AWS
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze big data for a fraction of the cost of traditional data warehouses. In this session, we take an in-depth look at data warehousing with Amazon Redshift for big data analytics. We cover best practices to take advantage of Amazon Redshift's columnar technology and parallel processing capabilities to deliver high throughput and query performance. We also discuss how to design optimal schemas, load data efficiently, and use work load management. Level: 300
Achieving Continuous Availability for Your Applications with Oracle MAAMarkus Michalewicz
Â
First presented during Oracle Cloud World 2022, this presentation discusses how to "Achieving Continuous Availability for Your Applications with Oracle MAA". You will learn how Application Continuity and related technologies keep your applications available. Get technical insights into how Oracle Database can help protect your application workflows from interruptions caused by planned maintenance or unplanned outages. Hear from customers about which applications benefit the most right away without code changes—and when customization may be required.
This is the CSS Tutorial for Beginners that teach the basics of CSS. This tutorial will show the basic structure of a CSS style and will show 3 different methods to apply styles.
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze big data for a fraction of the cost of traditional data warehouses. By following a few best practices, you can take advantage of Amazon Redshift’s columnar technology and parallel processing capabilities to minimize I/O and deliver high throughput and query performance. This webinar will cover techniques to load data efficiently, design optimal schemas, and use work load management.
Learning Objectives:
• Get an inside look at Amazon Redshift's columnar technology and parallel processing capabilities
• Learn how to migrate from existing data warehouses, optimize schemas, and load data efficiently
• Learn best practices for managing workload, tuning your queries, and using Amazon Redshift's interleaved sorting features
Who Should Attend:
• Data Warehouse Developers, Big Data Architects, BI Managers, and Data Engineers
Moving Beyond Lambda Architectures with Apache KuduCloudera, Inc.
Â
-Kudu is a new storage layer for the Hadoop ecosystem that enables fast analytics on fast data; it splits the difference between the fast read/write of HBase and the fast scans of HDFS...while compromising minimally on performance. It can pair with Spark, Impala, or MapReduce.
-In the past, a lambda architecture was needed to run analytics on real-time data – that is, a complex architecture that created separate a “speed layer” for rapid availability/query/updates, and a “batch layer” for running analytics scans. This was complicated and took lots of tuning.
-With Kudu, the Apache ecosystem now has a simplified storage solution for analytic scans on rapidly updating data, eliminating the need for the aforementioned hybrid lambda architectures.
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides are intended for developers who are interested in modifying the default look and feel of AtoM - known as the Dominion theme - and developing a custom theme plugin. They include some theme examples, how to register a plugin in Symfony, and some ideas of the elements you can modify via theming, with examples.
These slides were presented by Dan Gillean at the first ever AtoM Camp, held at the SFU Harbour Centre in Vancouver, BC, Canada, March 20-22, 2017. For more information on the Camp, see:
https://wiki.accesstomemory.org/Community/Camps/SFU2017
These slides introduce some of the primary Command Line tasks available in Access to Memory (AtoM), which can be used to support site maintenance and troubleshooting. For further information on AtoM CLI tasks, see:
* https://www.accesstomemory.org/docs/latest/admin-manual/maintenance/cli-tools/
* https://www.youtube.com/playlist?list=PLZiwlG5eSMeyeETe15EsEBSu5htPLK-wm
The CLI tasks introduced in these slides are based on the 2.4 AtoM release - some tasks or task options may not be available in earlier releases; tasks may be changed in future releases.
Building the Future Together: AtoM3, Governance, and the Sustainability of Op...Artefactual Systems - AtoM
Â
Slides accompanying a presentation given by Dan Gillean on June 7th, 2018 at Open Repositories 2018, held in Bozeman, MT.
Access to Memory is a web-based open source application for standards based description and access. AtoM was first released in 2008 and much of the codebase is now relying on deprecated frameworks and libraries – and at the same time, new standards and technologies are changing how our profession approaches description and access. Currently Artefactual Systems, a Canadian based company, uses a services model to support the project. Artefactual is looking ahead to AtoM3, and considering building a linked data driven platform for archival description and access. As we consider AtoM's next generation, we are also examining governance and maintenance models to sustain the project and better empower our user community as Artefactual wasn't originally intended to be AtoM's organizational home. This presentation will offer some thoughts on existing open source project governance models, challenges, and possibilities for the future. How do we ensure community engagement and project sustainability over time?
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see: https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides introduce users to the Access to Memory codebase. You can find AtoM's code at: https://github.com/artefactual/atom/
Practical information for Alfresco integration with AOS (Sharepoint Protocol), Google Drive, Microsoft 365, ONLYOFFICE and Collabora Online.
Additionally ADW support for ONLYOFFICE is provided by https://github.com/atolcd/adf-onlyoffice-extension#installation
Slides accompanying a presentation delivered at the VII Congresso Nacional de Arquivologia in Fortaleza, Brazil, on October 19th, 2016. The slides provide an overview of the AtoM project's history, its maintenance by Artefactual, and its development philosophy, before proceeding to examine the application as a component used in a digital preservation ecosystem. Aspects of ISO 16363:2012, the Audit and Certification of Trustworthy Digital Repositories standard, are used to evaluate how AtoM can support description, management, administration, and access functions when used to maintain a chain of custody in a trustworthy digital repository ecosystem.
Dynamic Components using Single-Page-Application Concepts in AEM/CQNetcetera
Â
Dynamic components display content dependable on context, hence they cannot be cached.
Out of the box, Adobe Experience Manager doesn't give us many options for granular caching on a component level.
When faced with this problem, we usually resort to developing components that are leveraging Server-Side Includes or AJAX to get the HTML with dynamic data.
As an alternative solution, we have also developed dynamic components that use Single Page Application concepts, by using templates and JSON-responses, to provide the same dynamic behavior.
In this presentation we will cover all of the dynamic components types, compare the benefits and drawbacks of each, and state the use-cases where each can be effectively applied.
We will take a deeper look at the dynamic components done with SPA concepts, as they are rarity in the AEM world, and also provide a walk-through of the technologies used, how some common problems were solved, as well as the benefits that have been gained by their usage.
These slides accompany a 1.5 hour webinar sponsored by the Western New York Library Resources Council, presented by Dan Gillean of Artefactual Systems on February 15th, 2017.
The session was intended to introduce participants to some of the key standards, services, and tools available to support digital preservation planning and activities. Part 1 focused on DP101, and how to begin tackling digital preservation in your institution. Part 2 introduced the Archivematica project's history, philosophy, and aims, while Part 3 was a live demonstration of Archivematica in action.
Thank you to WNYLRC for sponsoring this event!
by Dhanraj Pondicherry, Sr. Solutions Architecture Manager, AWS
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze big data for a fraction of the cost of traditional data warehouses. In this session, we take an in-depth look at data warehousing with Amazon Redshift for big data analytics. We cover best practices to take advantage of Amazon Redshift's columnar technology and parallel processing capabilities to deliver high throughput and query performance. We also discuss how to design optimal schemas, load data efficiently, and use work load management. Level: 300
Achieving Continuous Availability for Your Applications with Oracle MAAMarkus Michalewicz
Â
First presented during Oracle Cloud World 2022, this presentation discusses how to "Achieving Continuous Availability for Your Applications with Oracle MAA". You will learn how Application Continuity and related technologies keep your applications available. Get technical insights into how Oracle Database can help protect your application workflows from interruptions caused by planned maintenance or unplanned outages. Hear from customers about which applications benefit the most right away without code changes—and when customization may be required.
This is the CSS Tutorial for Beginners that teach the basics of CSS. This tutorial will show the basic structure of a CSS style and will show 3 different methods to apply styles.
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze big data for a fraction of the cost of traditional data warehouses. By following a few best practices, you can take advantage of Amazon Redshift’s columnar technology and parallel processing capabilities to minimize I/O and deliver high throughput and query performance. This webinar will cover techniques to load data efficiently, design optimal schemas, and use work load management.
Learning Objectives:
• Get an inside look at Amazon Redshift's columnar technology and parallel processing capabilities
• Learn how to migrate from existing data warehouses, optimize schemas, and load data efficiently
• Learn best practices for managing workload, tuning your queries, and using Amazon Redshift's interleaved sorting features
Who Should Attend:
• Data Warehouse Developers, Big Data Architects, BI Managers, and Data Engineers
Moving Beyond Lambda Architectures with Apache KuduCloudera, Inc.
Â
-Kudu is a new storage layer for the Hadoop ecosystem that enables fast analytics on fast data; it splits the difference between the fast read/write of HBase and the fast scans of HDFS...while compromising minimally on performance. It can pair with Spark, Impala, or MapReduce.
-In the past, a lambda architecture was needed to run analytics on real-time data – that is, a complex architecture that created separate a “speed layer” for rapid availability/query/updates, and a “batch layer” for running analytics scans. This was complicated and took lots of tuning.
-With Kudu, the Apache ecosystem now has a simplified storage solution for analytic scans on rapidly updating data, eliminating the need for the aforementioned hybrid lambda architectures.
Introduction to Azure Data Lake and U-SQL for SQL users (SQL Saturday 635)Michael Rys
Â
Data Lakes have become a new tool in building modern data warehouse architectures. In this presentation we will introduce Microsoft's Azure Data Lake offering and its new big data processing language called U-SQL that makes Big Data Processing easy by combining the declarativity of SQL with the extensibility of C#. We will give you an initial introduction to U-SQL by explaining why we introduced U-SQL and showing with an example of how to analyze some tweet data with U-SQL and its extensibility capabilities and take you on an introductory tour of U-SQL that is geared towards existing SQL users.
slides for SQL Saturday 635, Vancouver BC, Aug 2017
Best Practices for Migrating your Data Warehouse to Amazon RedshiftAmazon Web Services
Â
You can gain substantially more business insights and save costs by migrating your existing data warehouse to Amazon Redshift. This session will cover the key benefits of migrating to Amazon Redshift, migration strategies, and tools and resources that can help you in the process. We’ll learn about AWS Database Migration Service and AWS Schema Migration Tool, which were recently enhanced to import data from six common data warehouse platforms.
End-to-end Data Governance with Apache Avro and AtlasDataWorks Summit
Â
Aeolus is Comcast’s new internal Big Data system for providing access to an integrated view of a wide variety of high-quality, near-real-time and batch data. Such integration can enable data scientists to uncover otherwise hidden trends, anomalies, and powerful predictors of business successes and failures. But integrating data across silos in a large enterprise is fraught with peril. There typically are few standards on naming conventions and data representation, and spotty documentation at best. The old rule of thumb often applies: 70% of the analysts’ time goes into data wrangling, while only 30% goes toward the actual analyses and simulations. The goal of the Athene Data Governance Platform within Aeolus is to invert this ratio. This talk will explain how Comcast is using Apache Avro and Atlas for end-to-end data governance, the challenges faced, and methods used to address these challenges.
Avro provides a lingua franca for data representation, data integration, and schema evolution. All data published for community consumption must have an associated avro schema in Atlas. Every step in its journey through Aeolus, in flight or at rest, is captured in Atlas. Atlas’ extensibility has allowed us to add or update various entity types (e.g., avro schemas, kafka topics, object store pseudo-directories) and lineage types (e.g., storing streaming data in object storage; embellishing and re-publishing streaming data; performing aggregations and other transformations on data at rest; and evolution of schemas with compatibility flags). Transformation services notify Atlas of lineage links via custom asynchronous kafka messaging.
Atlas provides self-service data discovery and lineage browsing and querying, via full-text search, DSL query language, or gremlin graph query language. Example queries: “Where is data from kafka topic X stored?” “Display the journey of data currently stored in pseudo-directory X since it entered the Aeolus system”. “Show me all earlier versions of schema S, and whether they are forward/backward compatible with each other.”
AWS Public Data Sets: How to Stage Petabytes of Data for Analysis in AWS (WPS...Amazon Web Services
Â
AWS hosts a variety of public data sets that anyone can access for free. Previously, large data sets such as satellite imagery or genomic data have required hours or days to locate, download, customize, and analyze. When data is made publicly available on AWS, anyone can analyze any volume of data without downloading or storing it themselves. In this session, the AWS Open Data Team shares tips and tricks, patterns and anti-patterns, and tools to help you effectively stage your data for analysis in the cloud.
These slides were presented by Dan Gillean, AtoM Program Manager with Artefactual Systems, to provide an update on the AtoM 2.5 release and long-term AtoM3 news to the UK regional AtoM users group meeting held in Glasgow on May 17, 2019. For more information on this group see:
https://wiki.accesstomemory.org/Community/Groups/UK
Fore more information on the AtoM 2.5 release, please see:
https://wiki.accesstomemory.org/Releases/Release_announcements/Release_2.5
Access to Memory (AtoM) is an open source web application for standards-based archival description and access - learn more at:
https://www.accesstomemory.org
To provide users with an easy to install local environment for testing and development, Artefactual maintains a version of AtoM that can be installed on a laptop or home computer, regardless of operating system. We have slides that will explain what Vagrant is and how to install the AtoM Vagrant box here:
http://bit.ly/AtoM-Vagrant
These slides will help users create a re-usable set of data for use in a local AtoM Vagrant environment. Having a set of data that can easily be reloaded will make the AtoM Vagrant box more useful to local testers and developers.
The slides were originally created by Dan Gillean, AtoM Program Manager, for use in a series of training workshops delivered July 9-13, 2018 at the University of the Witswatersrand in Johannesburg, South Africa. The slides are based on current functionality in AtoM release 2.4 - they have been tested in the AtoM 2.4.0.2 and 2.5.0.0 Vagrant boxes.
These slides accompanied a presentation by Dan Gillean and Sara Allain of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In this session, we looked at the history of the AtoM project, the current governance and maintenance model for AtoM that Artefactual Systems maintains, as well as several other open source business models and stakeholders as we consider ways in which governance and maintenance might look different in the future. Finally, we looked at the state of the application at present with an eye to the future - exploring how changes in international standards and web technologies provide an opportunity to consider what AtoM 3 might look like. The session was followed by open discussion.
These slides accompanied a presentation by Dan Gillean of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
In 2013, with the introduction of AtoM 2.0, Artefactual began a fresh rewrite of all the Access to Memory documentation, using Sphinx Documentation as the primary platform. These slides provide some context and background for this change, introduce Sphinx and reStructuredText, and go on to outline step by step how AtoM users can contribute to the AtoM documentation.
This worksheet accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
The presentation focused on the installation of Access to Memory, and the slides from it can be seen here:
https://www.slideshare.net/accesstomemory/installing-and-upgrading-atom
This handout showed attendees how Ansible can be used to automate the deployment process of AtoM, using the publicly available Ansible playbooks. You can find Artefactual's Ansible playbooks at:
https://github.com/artefactual/deploy-pub
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides are intended for programmers interested in developing new features for Access to Memory. They first provide an overview of the feature development proccess, and then look at the different types of development typically pursued (plugins, CLI tasks, Background jobs, and core features), with tips and suggestions for each.
These slides accompanied a presentation by Steve Breker of Artefactual Systems, delivered as part of AtoM Camp Cambridge, a three-day boot camp held at St John's College, Cambridge University, May 9-11, 2017 For more information, see:
https://wiki.accesstomemory.org/Community/Camps/SJC2017
These slides provide developers and system administrators with some suggestions on how to install, maintain, and upgrade Access to Memory. They are intended to supplement the offficial installation documentation, found in the Atom Administrator's Manual:
https://www.accesstomemory.org/docs/latest/
These slides were presented by Dan Gillean at the first ever AtoM Camp, held at the SFU Harbour Centre in Vancouver, BC, Canada, March 20-22, 2017. For more information on the Camp, see:
https://wiki.accesstomemory.org/Community/Camps/SFU2017
These slides introduce participants to basic commands in the linux/unix command-line interface, for navigation and basic tasks such as reading, creating, deleting, and moving files. The slides were intended as an initial introduction for archivists unfamiliar with using the command-line, prior to teaching them about the command-line tasks available in Access to Memory (AtoM), which can be used for maintenance and troubleshooting.
Slides accompanying a day-long AtoM and Archivematica workshop, delivered October 17, 2016 at the VII Congresso Nacional de Arquivologia in Fortaleza, Brazil.
These slides introduce Open Source and Free Software definitions, and proceed to outline Artefactual's approach to open source development and maintenance of the AtoM and Archivematica projects.
Technologie Proche: Imagining the Archival Systems of Tomorrow With the Tools...Artefactual Systems - AtoM
Â
These slides accompanied a June 4th, 2016 presentation made by Dan Gillean of Artefactual Systems at the Association of Canadian Archivists' 2016 Conference in Montreal, QC, Canada.
This presentation aims to examine several existing or emerging computing paradigms, with specific examples, to imagine how they might inform next-generation archival systems to support digital preservation, description, and access. Topics covered include:
- Distributed Version Control and git
- P2P architectures and the BitTorrent protocol
- Linked Open Data and RDF
- Blockchain technology
The session is part of an attempt by the ACA to create interactive "working sessions" at its conferences. Accompanying notes can be found at: http://bit.ly/tech-Proche
Participants were also asked to use the Twitter hashtag of #techProche for online interaction during the session.
Presentation made as part of a community update during Code4Arc, a day-long workshop organized as a part of Code4Lib 2016 in Philadelphia, Monday, March 7 2016.
See:
* http://2016.code4lib.org/
* http://code4lib.org/
* https://www.accesstomemory.org
Project Documentation with Sphinx (or, How I Learned to Stop Worrying and Lov...Artefactual Systems - AtoM
Â
Slides accompanying a lightning talk delivered at BC Code4Lib 2015, held in Vancouver, BC, Canada on November 26-27, 2015.
This talk covered our move from using wikis for project documentation with the Access to Memory (AtoM) project, to using Sphinx documentation with the AtoM 2.0.0 release.
More on AtoM: https://www.accesstomemory.org
More on Sphinx Documentation: http://sphinx-doc.org/
Code4Lib BC: http://wiki.code4lib.org/BC
Introductory slides about Access to Memory (AtoM), accompanying a 4-hour workshop session delivered at the 2015 Digital Library Federation Forum in Vancouver, BC, Canada (see: http://www.diglib.org/forums/2015forum/) on October 28, 2015.
For more information on AtoM, see: https://www.accesstomemory.org
Introducing Binder: A Web-based, Open Source Digital Preservation Management ...Artefactual Systems - AtoM
Â
Binder is a new digital preservation management application developed by Artefactual Systems in conjunction with the Museum of Modern Art (MoMA). This new system aims to facilitate digital collections care, management, and preservation for time-based media and born-digital artworks and is built from integrating functionality of both Archivematica and AtoM.
Artefactual Systems, working with the Museum of Modern Art, has worked to build the first open source digital repository designed to meet the needs and complex digital preservation requirements of museum collections. This new system aims to facilitate digital collections care, management, and preservation for time-based media and born-digital artworks and is built from integrating functionality of both Archivematica and AtoM. This lightning session will introduce the DRMC, outline its design/deployment process, features, and its relevance to archives, libraries, museums, and other cultural heritage institutions undertaking digital preservation.
Welcome to the first live UiPath Community Day Dubai! Join us for this unique occasion to meet our local and global UiPath Community and leaders. You will get a full view of the MEA region's automation landscape and the AI Powered automation technology capabilities of UiPath. Also, hosted by our local partners Marc Ellis, you will enjoy a half-day packed with industry insights and automation peers networking.
đź“• Curious on our agenda? Wait no more!
10:00 Welcome note - UiPath Community in Dubai
Lovely Sinha, UiPath Community Chapter Leader, UiPath MVPx3, Hyper-automation Consultant, First Abu Dhabi Bank
10:20 A UiPath cross-region MEA overview
Ashraf El Zarka, VP and Managing Director MEA, UiPath
10:35: Customer Success Journey
Deepthi Deepak, Head of Intelligent Automation CoE, First Abu Dhabi Bank
11:15 The UiPath approach to GenAI with our three principles: improve accuracy, supercharge productivity, and automate more
Boris Krumrey, Global VP, Automation Innovation, UiPath
12:15 To discover how Marc Ellis leverages tech-driven solutions in recruitment and managed services.
Brendan Lingam, Director of Sales and Business Development, Marc Ellis
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
Â
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
DevOps and Testing slides at DASA ConnectKari Kakkonen
Â
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Enhancing Performance with Globus and the Science DMZGlobus
Â
ESnet has led the way in helping national facilities—and many other institutions in the research community—configure Science DMZs and troubleshoot network issues to maximize data transfer performance. In this talk we will present a summary of approaches and tips for getting the most out of your network infrastructure using Globus Connect Server.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
Â
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Â
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
Â
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
Â
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
Â
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Â
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Â
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Â
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
5. Archival descriptions: overview
• 2 main templates: ISAD(G) and
RAD
• Column headers are the same
in each – there are just more in
RAD.
• You could combine columns from the 2
CSVs and it would import successfully –
it just may not display all data on a
single template!
• Column header names are
generic, based on database
names
Header names are generally written in
camelCase:
• scopeAndContent
• Scope and content (ISAD 3.3.1)
• Scope and content (RAD 1.7D)
• physicalCharacteristics
• Physical characteristics and technical
requirements (ISAD 3.4.4)
• Physical condition note (RAD 1.8B9a)
6. Crosswalking with CSV files
https://wiki.accesstomemory.org/Resources/CSV_templates
In the example CSV files from v2.2 on, we have included the relevant
content standard name and number in the sample data field. This
means you can import the CSV template to produce a sort of
“crosswalk” or key, showing you how fields in AtoM map to the column
headers, and to other standards templates (such as DACS, MODS, DC).
8. You can preview data in another
standard without having to edit, by
adding a semi-colon and the
standard’s abbreviation to the end
of its URL:
Standard template Add to URL
ISAD(G) ;isad
RAD ;rad
DACS ;dacs
Dublin Core ;dc
MODS ;mods
Crosswalking with CSV files
9. Character encoding and line endings
AtoM expects CSV files to use:
• UTF-8 character encoding
• Unix/linux style end of line
characters
UTF-8 (short for Unicode Transformation Format – 8-bit) is a
standard for encoding characters (such as letters, numbers,
punctuation, and special characters) in digital formats. UTF-8 is
backwards compatible with ASCII, the standard that predated it, and
has become the de facto encoding standard used on the web and
in email.
“In computing, a newline, also known as a line ending, end of line
(EOL), or line break, is a special character or sequence of characters
signifying the end of a line of text. The actual codes representing a
newline vary across operating systems, which can be a problem
when exchanging text files between systems with different newline
representations” (Wikipedia). These character are not typically made
visible in the document to users.
Here are some of the differences:
• Unix / Linux / FreeBSD / OS X use LF (line feed, n, 0x0A)
• Macs prior to OS X use CR (carriage return, r, 0x0D)
• Windows / DOS use CR+LF (carriage return followed by line feed, rn,
0x0D0A)
Failure to use these can lead to
import errors, such as strange
characters, or thousands of
blank descriptions being
created!
10. Character encoding and line endings
We strongly recommend
against using MS Excel as
your local spreadsheet
application when preparing
AtoM CSV data!
Microsoft likes to use its
own custom character
encoding (Windows-1252),
and makes it very difficult
to change the default
encoding to UTF-8.
Similarly, Excel tends to use
different line endings, which
can cause import issues.
LibreOffice Calc is an open source spreadsheet application that makes a
good alternative for AtoM CSV preparation.
When first opening a CSV file, it will allow you
to set the character set, separators, text
delimiter, and see a preview of the fields with
your selection to ensure everything is
rendering properly.
11. Hierarchical data: legacyId and parentId
• Can be used to manage hierarchies within a file
• legacyId values can be arbitrary – alphanumeric values. Should be unique within the
CSV file
• parentId values should use the legacyId value of the target parent record. For
example:
• IF a collection has a legacyId of 001
• AND the next row is a series that should be a child of the collection
• THEN the parentId value of the series should be 001
• If a record is a top-level description (such as a fonds or collection, etc), then leave the
parentId column blank
12. Hierarchical data: legacyId and parentId
IMPORTANT: Child records MUST appear AFTER (i.e. below) their parent records in the CSV row order,
otherwise the import will fail! Child and sibling order doesn’t matter, so long as the parent row is ABOVE
13. Hierarchical data: qubitParentSlug
• Can be used to link imported descriptions to existing descriptions in AtoM as children
• qubitParentSlug values should be the slug (or permalink – the unique part of the
URL after the slash) of the target parent description in AtoM
• WARNING: do not use parentId and qubitParentSlug in the same row!
• If both are present, AtoM will default to using the qubitParentSlug
• Can still use a mix of parentId and qubitParentSlug throughout the CSV – just
not in the same row
• All descriptions should still have a legacyId value added
14. Linking to Accession records on import
If you want to link incoming descriptions to an existing accession record,
you can add a new column to the description CSV, with the header:
accessionNumber
NOTE: if the accession number does not exist, AtoM will create a blank
stub accession with that number! Make sure you add the correct
Accession number :)
15. Dates and creators (events and actors)
Description edit templates have 3 date fields. The Display date is what the
end user will see - it is free text. The start and end dates must follow ISO
8601 (YYYY-MM-DD, etc) formatting. These fields are used to support
AtoM’s date range search.
• Display date
• Start date
• End date
16. Dates and creators (events and actors)
During CSV import, Creators and Dates are paired
Use the | pipe character to add multiple creators/dates.
You can use a literal NULL value in your CSV file to keep the spacing correct for
dates without actors or vice versa:
eventTypes: Sets the
type of event. Different
options based on default
template used in AtoM:
• ISAD: Creation, Accumulation
• RAD: Creation, Accumulation, Contribution, Collection, Broadcasting,
Manufacturing, Custody, Publication, Reproduction, Distribution
• DC: Creation, Contribution, Publication
• DACS: Creation, Publication Broadcasting, Record-keeping activity
• MODS: all
17. Dates and creators (events and actors)
eventActors and eventActorHistories are related to Authority records.
AtoM will first try to find an existing matching authority record, and will link to it if
found. If not, AtoM will create a new stub authority record. Matching criteria:
# Import type
Match on
name
Match on
repository
Match on history Outcome
1 Any Yes Yes (or blank) Yes (or blank) Linked (no updates)
2A
New or
delete/replace
Yes Yes No New actor created
2B Update Yes Yes (or blank) No Linked AND existing history updated
3 Any Yes No Yes (or blank) Linked (no updates)
4 Any Yes No No New actor created
5 Any No Yes (or blank) Yes (or blank) New actor created
18. Dates and creators (events and actors)
Remember if you pipe values to include more than 1 creator per row, you still need
to pipe across all event fields! Use NULL as a placeholder where needed.
IMPORTANT: AtoM includes creator inheritance! Meaning if you have added a creator at the top level, you
do NOT need to add them in the CSV at lower levels again – in fact this can cause performance and display
issues later! Instead, simply add the dates and use the NULL placeholder for the eventActor and
eventActorHistories fields:
19. Dates and creators (events and actors)
eventPlaces and eventDescriptions:
These fields are visible only in the RAD and
MODS templates in AtoM.
You could add these columns to an ISAD CSV
template and it would import without errors,
but they will not show in the ISAD template
user interface!
These values also need to be piped to match
up with the correct event when importing
multiple events per row.
20. Digital object columns in the description CSV
You can import digital objects at the same time you create new
descriptions.
Note that AtoM has a 1:1 relationship between a digital object and an
information object (i.e. archival description) – you can only attach one
object per description.
For local objects on the same server – use the digitalObjectPath
column (you can add it to the CSV if it is not in the template)
21. Digital object columns in the description CSV
You can also import remote digital objects available on the web, using the
digitalObjectURI column. AtoM will create local derivatives, but will
only store the path to the master, rather than retaining a local copy.
Requirements for a remote digital object import to succeed:
• Must be publicly available on the web (no logins, firewalls, VPN requirements, etc.)
• Must be an HTTP or HTTPS link (FTP links will not work)
• URI must end in the file extension (e.g. .jpg etc. Can’t link to a landing page such
as a YouTube video this way – AtoM needs to be able to fetch the actual object!)
22. Physical storage columns
Similarly to creators, you can link to or create new physical storage containers and
locations in a description CSV import
• For the import to succeed, all 3 fields must be populated for each row:
physicalObjectName, physicalObjectLocation, and
physicalObjectType
• Can pipe values in a row to add multiple locations
• AtoM will first attempt to match based on all 3 rows – if no match is found, a new
container and location will be created on import
• physicalObjectType is linked to the Physical Object Type taxonomy in AtoM
23. Culture, language and script fields
• culture: determines the source language of the description in AtoM. We
recommend populating for all rows. Expects ISO 639-1 two-letter language
code values (e.g. en, fr, es, etc.) Must have the culture added to Admin
>Settings > i18n languages for proper display!
• See a list of AtoM supported langs and codes here: bit.ly/AtoM-langs
• language (of material) and languageOfDescription both also
expect ISO 639-1 two-letter language code values. Can accept piped values
for multiple languages, e.g. en|fr|es
• script (of material) and scriptOfDescription expect four-letter
ISO 15924 script code values – e.g. Latn, Cyrl, etc. Can also pipe to enter
multiple values per row. Note that these codes are capitalized while
language and culture ones are not!
24. Other controlled value fields (linked to taxonomies)
• levelOfDescription: can enter new terms and will be created on
import (e.g. sub-sub-subseries, etc.)
• subjectAccessPoints, placeAccessPoints,
genreAccessPoints: can enter new terms or link to existing ones.
Can pipe these fields for multiple terms
• descriptionStatus: expects Draft, Revised, or Final
• New terms *may* import correctly, but better to add them to the taxonomy first
• levelOfDetail: expects Full, Partial, or Minimal
• New terms *may* import correctly, but better to add them to the taxonomy first
25. Other fields of note
• The nameAccessPoints field links to authority records, much like the eventActors for
creators. First it will look for an exact match on name, and then it will create a new stub
authority record if none is found. Be sure to check for typos, and use consistent naming
conventions with your authorities!
• The publicationStatus column sets the public visibility of your descriptions on import.
Expects values Draft or Public. If left blank, the default publication status (in Admin > Settings
> Global) will be used. Note that you can’t have a published description as a child of a Draft
description – this can cause display issues in AtoM!
• Alternative identifiers and their display labels can be imported
using the alternativeIdentifiers and
alternativeIdentifierLabels columns. Use pipe
separators to add multiple values. There should be a 1:1
relationship between the number of identifier values in the
alternativeIdentifiers column and corresponding
labels in the alternativeIdentifierLabels column.