(Jan 2011) Digital Curation (Guest Lecture)Carolyn Hank
Event: Guest lecture on introduction to digital curation for Prof. Elaine Menard's GLIS 639: Introduction to Museology class, School of Information Studies, McGill University (January 28, 2011)
Gabriel de la Iglesia is a highly motivated college graduate with a background in biological and earth sciences. He graduated from FIU with honors and was a member of multiple honor societies. His experience includes volunteering at Zoo Miami, interning with FIU's Nature Preserve studying herpetofauna and butterflies, assisting with seagrass research in Biscayne Bay, and aiding various FIU labs studying topics such as urban ecology, plant ecophysiology, and bird collisions. He is proficient in GIS and other software and participated in university sustainability efforts.
Esther Adamu submitted a document on her seminar topic of digital preservation. Digital preservation aims to ensure continued access to digital materials by addressing issues like format obsolescence. Key strategies include migration, emulation, and checksums. Examples of digital preservation efforts are the Internet Archive, Digital Public Library of America, and NASA Planetary Data System. Digital preservation is important for cultural heritage, knowledge continuity, and research. Challenges include format obsolescence, costs, and legal/ethical concerns. The field has evolved with technological advances and increased collaboration. Further changes will be needed to address new content types and legal/ethical issues.
Data hosting infrastructure for primary biodiversity dataPhil Cryer
Today, an unprecedented volume of primary biodiversity data are being generated worldwide, yet significant amounts of these data have been and will continue to be lost after the conclusion of the projects tasked with collecting them. To get the most value out of these data it is imperative to seek a solution whereby these data are rescued, archived and made available to the biodiversity community. To this end, the biodiversity informatics community requires investment in processes and infrastructure to mitigate data loss and provide solutions for long-term hosting and sharing of biodiversity data.
We review the current state of biodiversity data hosting and investigate the technological and sociological barriers to proper data management. We further explore the rescuing and re-hosting of legacy data, the state of existing toolsets and propose a future direction for the development of new discovery tools. We also explore the role of data standards and licensing in the context of data hosting and preservation. We provide five recommendations for the biodiversity community that will foster better data preservation and access: (1) encourage the community’s use of data standards, (2) promote the public domain licensing of data, (3) establish a community of those involved in data hosting and archival, (4) establish hosting centers for biodiversity data, and (5) develop tools for data discovery.
The community’s adoption of standards and development of tools to enable data discovery is essential to sustainable data preservation. Furthermore, the increased adoption of open content licensing, the establishment of data hosting infrastructure and the creation of a data hosting and archiving community are all necessary steps towards the community ensuring that data archival policies become standardized.
BMC Bioinformatics 2011, 12(Suppl 15):S5 doi:10.1186/1471-2105-12-S15-S5
http://www.biomedcentral.com/1471-2105/12/S15/S5
This presentation was given at the Winter Doctoral Seminar of the Chair of Supply Chain Management and Information Systems. It presents the achievements from my one-year research stay at the European Research Center for Information Systems of the Münster University in Germany under supervision of Prof. Dr.-Ing. Bernd Hellingrath.
This document contains 10 references cited in another work. The references are numbered and include the author(s), title, publication venue, and year. Topics covered include HTML5 application privilege separation, orthogonal security techniques, proxy re-encryption, database security and privacy, securing frame communication in browsers, content sniffing techniques, social network data anonymization, adapting Kerberos for browsers, and news articles regarding privacy and surveillance.
Global biodiversity data is critical for conservation, policymaking, and scientific research. However, most data is held by small, isolated publishers and is difficult to access. The Global Biodiversity Information Facility (GBIF) aims to mobilize this "small data" by creating common data standards and tools to publish data through its Integrated Publishing Toolkit. This allows data to be discovered through GBIF's portal and used for applications like predicting climate change impacts and invasive species spread. GBIF calls on all data holders to publish their data openly through its framework to build a comprehensive global resource for biodiversity data.
(Jan 2011) Digital Curation (Guest Lecture)Carolyn Hank
Event: Guest lecture on introduction to digital curation for Prof. Elaine Menard's GLIS 639: Introduction to Museology class, School of Information Studies, McGill University (January 28, 2011)
Gabriel de la Iglesia is a highly motivated college graduate with a background in biological and earth sciences. He graduated from FIU with honors and was a member of multiple honor societies. His experience includes volunteering at Zoo Miami, interning with FIU's Nature Preserve studying herpetofauna and butterflies, assisting with seagrass research in Biscayne Bay, and aiding various FIU labs studying topics such as urban ecology, plant ecophysiology, and bird collisions. He is proficient in GIS and other software and participated in university sustainability efforts.
Esther Adamu submitted a document on her seminar topic of digital preservation. Digital preservation aims to ensure continued access to digital materials by addressing issues like format obsolescence. Key strategies include migration, emulation, and checksums. Examples of digital preservation efforts are the Internet Archive, Digital Public Library of America, and NASA Planetary Data System. Digital preservation is important for cultural heritage, knowledge continuity, and research. Challenges include format obsolescence, costs, and legal/ethical concerns. The field has evolved with technological advances and increased collaboration. Further changes will be needed to address new content types and legal/ethical issues.
Data hosting infrastructure for primary biodiversity dataPhil Cryer
Today, an unprecedented volume of primary biodiversity data are being generated worldwide, yet significant amounts of these data have been and will continue to be lost after the conclusion of the projects tasked with collecting them. To get the most value out of these data it is imperative to seek a solution whereby these data are rescued, archived and made available to the biodiversity community. To this end, the biodiversity informatics community requires investment in processes and infrastructure to mitigate data loss and provide solutions for long-term hosting and sharing of biodiversity data.
We review the current state of biodiversity data hosting and investigate the technological and sociological barriers to proper data management. We further explore the rescuing and re-hosting of legacy data, the state of existing toolsets and propose a future direction for the development of new discovery tools. We also explore the role of data standards and licensing in the context of data hosting and preservation. We provide five recommendations for the biodiversity community that will foster better data preservation and access: (1) encourage the community’s use of data standards, (2) promote the public domain licensing of data, (3) establish a community of those involved in data hosting and archival, (4) establish hosting centers for biodiversity data, and (5) develop tools for data discovery.
The community’s adoption of standards and development of tools to enable data discovery is essential to sustainable data preservation. Furthermore, the increased adoption of open content licensing, the establishment of data hosting infrastructure and the creation of a data hosting and archiving community are all necessary steps towards the community ensuring that data archival policies become standardized.
BMC Bioinformatics 2011, 12(Suppl 15):S5 doi:10.1186/1471-2105-12-S15-S5
http://www.biomedcentral.com/1471-2105/12/S15/S5
This presentation was given at the Winter Doctoral Seminar of the Chair of Supply Chain Management and Information Systems. It presents the achievements from my one-year research stay at the European Research Center for Information Systems of the Münster University in Germany under supervision of Prof. Dr.-Ing. Bernd Hellingrath.
This document contains 10 references cited in another work. The references are numbered and include the author(s), title, publication venue, and year. Topics covered include HTML5 application privilege separation, orthogonal security techniques, proxy re-encryption, database security and privacy, securing frame communication in browsers, content sniffing techniques, social network data anonymization, adapting Kerberos for browsers, and news articles regarding privacy and surveillance.
Global biodiversity data is critical for conservation, policymaking, and scientific research. However, most data is held by small, isolated publishers and is difficult to access. The Global Biodiversity Information Facility (GBIF) aims to mobilize this "small data" by creating common data standards and tools to publish data through its Integrated Publishing Toolkit. This allows data to be discovered through GBIF's portal and used for applications like predicting climate change impacts and invasive species spread. GBIF calls on all data holders to publish their data openly through its framework to build a comprehensive global resource for biodiversity data.
ONLINE EDUCATIONAL INCIDENT RESPONSE TRAINING: VIRTUAL WORLDS AS A COST EFFEC...Maurice Dawson
Critical incident leadership and planning is an important topic in the field of homeland security. Within this field are many challenges. For the purposes of this forum only two shall be discussed. The two challenges related to critical incident leadership and planning is the ability to learn in a simulated environment and funding. The first challenge surrounds learning as there is a significant cost associated in providing a real life scenarios where employees can learn. Employees need to understand how they should react in a simulated process to ensure that they can perform in the real event. This simulated scenario also allows an independent group to measure the effectiveness of the incident response manager. An inexpensive and effective method to do this would be through the
utilization of Second Life (SL). SL is one of the most popular virtual world provides a unique and secure teaching and learning environment for instructors and students. This learning environment could be designed to be private for a specific response group or incident where each avatar assigned to a unique name could include the name of group name that the individual is belong to. The second issue,
which surrounds funding, directly correlates with the capabilities and readiness to respond to a group. To properly correct this one would need to find a method to assign funding needs to risks or events that are most likely to occur.
Digital preservation aims to maintain access to digital materials over time despite changes in technology. It faces challenges such as the obsolescence of storage technologies, instability of storage media, and maintaining the integrity of digital materials. Recommendations include starting preservation close to creation to ensure future access and increasing awareness of preservation techniques. While skepticism remains, institutions like the Library and Archives of Canada are mandated to preserve digital heritage for present and future generations through continued migration to new technologies and formats.
A survey on privacy preserving data publishingijcisjournal
Data mining is a computational process of analysing and extracting the data from large useful datasets. In
recent years, exchanging and publishing data has been common for their wealth of opportunities. Security,
Privacy and data integrity are considered as challenging problems in data
mining.Privacy is necessary to protect people’s interest in competitive situations. Privacy is an abilityto
create and maintain different sort of social relationships with people. Privacy Preservation is one of the
most important factor for an individual since he should not embarrassed by an adversary. The Privacy
Preservation is an important aspect of data mining to ensure the privacy by various methods. Privacy
Preservation is necessary to protect sensitive information associated with individual. This paper provides a
survey of key to success and an approach where individual’s privacy would to be non-distracted.
This document discusses challenges and opportunities for discovering and documenting biodiversity in the current information age. It argues that current taxonomic processes are too slow and that new approaches are needed to integrate distributed data sources and leverage community contributions. Specifically, it proposes:
1) Publishing new biodiversity data prior to formal documentation to accelerate discovery.
2) Developing automated workflows and online workspaces to integrate phylogenetic, distribution, and trait data.
3) Enabling community participation through open data sharing and collaborative annotation platforms.
This document discusses challenges and opportunities for discovering and documenting biodiversity in the current information age. It argues that current taxonomic processes are too slow and that new approaches are needed to integrate distributed data sources and leverage community sourcing. Specifically, it advocates for:
1) Publishing new biodiversity data prior to formal documentation to accelerate discovery.
2) Developing automated workflows and online workspaces to integrate phylogenetic, distribution, and trait data.
3) Enabling community participation in annotating and improving global biodiversity models and maps.
4) Changing incentives to value data sharing over individual "kudos" and prioritize the collective good of the scientific community.
The document discusses the issue of uneven distribution of biodiversity data around the world, with much of the data held by small publishers and citizen scientists. It notes that these "small data publishers" face challenges in discovering, accessing, managing and publishing their data according to standards. The document calls for developing standards and tools that make it easier for small data publishers to capture, organize and share their biodiversity data in order to help mobilize this important but hard to access data.
Meeting the NSF DMP Requirement June 13, 2012IUPUI
The document provides guidance on developing a data management plan (DMP) to meet requirements for National Science Foundation grant proposals. It discusses the context and rationale for federal data policies, defines the key elements required for a DMP, and provides examples of DMPs for different types of research data. The main points are: understanding the NSF data policy aims to increase research impact and data sharing/reuse; a DMP must address the types of data generated, metadata standards, data access/sharing plans, long-term preservation, and associated costs; and good planning helps ensure data remains accessible, usable and preserved into the future. Resources and guidance are available to help researchers develop robust and fundable DMPs.
VOLUME-7 ISSUE-8, AUGUST 2019 , International Journal of Research in Advent Technology (IJRAT) , ISSN: 2321-9637 (Online) Published By: MG Aricent Pvt Ltd
This document proposes a refinement of the slicing anonymization technique for privacy-preserving data mining. Slicing anonymization has been shown to effectively preserve data quality while achieving high data privacy. The proposed refinement aims to achieve even higher data utility and more secure data publishing through probabilistic non-homogeneous suppression and consideration of attribute correlations. The results of applying the technique to election data are analyzed using standard classification metrics to validate that the refined approach maintains high data quality and strong privacy preservation.
Session 4 KE4CAP Pres1 bo dwyer-ke4_cap_canadaweADAPT
This document discusses user-oriented climate adaptation platforms and approaches to integrating user needs. It defines climate services as providing climate products or advice to assist decision-making. Traditionally, climate services took a top-down approach, but now focus on understanding user contexts and specifying requirements. Meeting user needs involves tailoring information, evolving user capacities over time, and adapting to evolving policy requirements. Key approaches discussed are web-based information provision, capacity building, and direct policy support.
Resilience is introduced as the new security goal supported with security/safety-related information by data-centric services for predictive risk management in real-time. Secondary use of personal information is of essential importance. The problem is that data-centric services threaten resilience. Although privacy as a state of equilibrium and its enforcement with usable security by identity management aims actually at decreasing users’ own risk, its use by data-centric services for unilateral information flow control threatens privacy and resilience. Users lose control on their identity while at the same time competitiveness of in particular small and medium service providers is endangered due to reliable statements on authentication of derived information. Self-protection, however, depends on opposite security interests. This talk claims that Multilateral Security improves privacy and resilience by a multilateral secondary use of personal security-related information for distributed usage control. This kind of privacy is understood as informational self-determination whereas the key concept is non-linkable delegation of rights on secondary use of personal information.
presented at the workshop "Usable Security and Privacy" an event of "Mittelstand-Digital" of the Federal Ministry for Economic Affairs and Energy (BMWi) and HCI conference "Mensch und Computer 2015" in Stuttgart, Germany http://www.mittelstand-digital.de/DE/Service/suche,did=717526.html
Ensuring that an organisation's digital assets are safe, secure and accessible for the long term should (in theory) be an interesting, responsible and useful role for anyone in an organisation to accept. The critical importance of digital assets, the ubiquity of digital methods and the need for people in all walks of life to have effective means to refer to persistent sources of data reinforce this notion. How is it then that long-term asset management, information lifecycle management, data curation, digital preservation (call it what you will) is often regarded as a peripheral specialist activity that it is diffcult to resource, complex to carry out, and delivers benefits that are, at best, simply an insurance policy rather than an activity that adds value to an organisation?
This presentation will examine the importance of defining clear roles for those involved with digital preservation and will consider the importance of associating this professional activity with strategic and tactical frameworks. It is likely that automated services will increasingly be required to deal with the collosal amount of digital information that will be produced and consumed over the next century and whilst the type and nature of these services are yet to be defined, we can be fairly certain of one endurng requirement, namely, that human judgement will always be needed to curate interesting and useful content for future generations.
Open Source Software for Digital Preservation Repositories : A SurveyIJCSES Journal
In the digital age, the amount of data produced is growing exponentially. Governments and institutions can no longer rely on old methods for storing data and passing on the knowledge to future generations. Digital data preservation is a mandatory issue that needs proper strategies and tools. With this awareness, efforts are being made to create and perfect software solutions capable of responding to the challenge of properly preserving digital information. This paper focuses on the state-of-the-art in open-source software solutions for the digital preservation and curation field used to assimilate and disseminate information to designated audiences. Eleven open source projects for digital preservation are surveyed in areas such as supported standards and protocols, strategies for preservation, methodologies for reporting, dynamic of development, targeted operating systems, multilingual support and open source license. Furthermore, five of these open
source projects, are further analysed, with focus on features deemed important for the area. Along open source solutions, the paper also briefly surveys the standards and protocols relevant for digital data preservation. The area of digital data preservation repositories has several open source solutions, which can form the base to overcome the challenges to reach mature and reliable digital data preservation.
Open Source Software for Digital Preservation Repositories : A SurveyIJCSES Journal
In the digital age, the amount of data produced is growing exponentially. Governments and institutions can no longer rely on old methods for storing data and passing on the knowledge to future generations. Digital data preservation is a mandatory issue that needs proper strategies and tools. With this awareness, efforts are being made to create and perfect software solutions capable of responding to the challenge of properly preserving digital information. This paper focuses on the state-of-the-art in open-source software solutions for the digital preservation and curation field used to assimilate and disseminate information to designated audiences. Eleven open source projects for digital preservation are surveyed in areas such as supported standards and protocols, strategies for preservation, methodologies for reporting, dynamic of development, targeted operating systems, multilingual support and open source license. Furthermore, five of these open
source projects, are further analysed, with focus on features deemed important for the area. Along open source solutions, the paper also briefly surveys the standards and protocols relevant for digital data preservation. The area of digital data preservation repositories has several open source solutions, which can form the base to overcome the challenges to reach mature and reliable digital data preservation.
This document summarizes a presentation by Professor Jacqueline McGlade from the European Environment Agency on drivers of air pollution and trends in emissions. It discusses many factors influencing air quality like population growth, energy use, and urbanization. While air pollution has decreased in some countries, emissions exceed limits in 8 EU states. New science shows Europe is still catching up on air quality. The document outlines tools and data used to monitor changes and manage natural resources. It promotes sharing data through open access platforms to better support policy and engage citizens in monitoring air quality.
EDAM (Eco-stations Data Access Monitor) is a project at BIDS (Berkeley Institute of Data Science) Collaborative, an initiative of the institute aims at solving real-world data challenges
This presentation gives an overview of the key things that we need to consider before publishing data from the repository. It briefly discusses research data management, research data lifecycle, FAIR principles of research data management and then move on to key elements that should be considered while preparing datasets for publishing through repository.
Introduction to research data management; Lecture 01 for GRAD521Amanda Whitmire
Lesson 1: Introduction to research data management. From a series of lectures from a 10-week, 2-credit graduate-level course in research data management (GRAD521, offered at Oregon State University).
The course description is: "Careful examination of all aspects of research data management best practices. Designed to prepare students to exceed funder mandates for performance in data planning, documentation, preservation and sharing in an increasingly complex digital research environment. Open to students of all disciplines."
Major course content includes: Overview of research data management, definitions and best practices; Types, formats and stages of research data; Metadata (data documentation); Data storage, backup and security; Legal and ethical considerations of research data; Data sharing and reuse; Archiving and preservation.
See also, "Whitmire, Amanda (2014): GRAD 521 Research Data Management Lectures. figshare. http://dx.doi.org/10.6084/m9.figshare.1003835. Retrieved 23:25, Jan 07, 2015 (GMT)"
Crowdsourcing Approaches to Big Data Curation for Earth SciencesEdward Curry
The document discusses crowdsourcing approaches to data curation for earth sciences. It covers several topics including motivation for data curation, data quality and curation processes, crowdsourcing, case studies on crowdsourced data curation, and setting up a crowdsourced data curation process. Specifically, it describes challenges with data quality, defines data curation and the role of data curators. It also outlines different types of data curation approaches based on who performs the curation (individual curators, departments, communities) and how it is done (manually, automated, crowdsourced).
Towards A Differential Privacy and Utility Preserving Machine Learning Classi...Kato Mivule
Kato Mivule, Claude Turner, Soo-Yeon Ji, "Towards A Differential Privacy and Utility Preserving Machine Learning Classifier", Procedia Computer Science (Complex Adaptive Systems), 2012, Pages 176-181, Washington DC, USA.
Caring for Digital Collections in the AnthropoceneTrevor Owens
The craft of digital preservation and digital collections care is anchored in the past. It builds off the records, files, and works of those who came before us and those who designed and set up the systems that enable the creation, transmission, and rendering of their work. At the same time, the craft of digital preservation is also the work of a futurist. We must look to the past trends in the ebb and flow of the development of digital media and hedge our bets on how digital technologies of the future will play out. This talk explores key issues for exploring and imagining that future. We start with consideration of some key emerging technologies relevant to digital collections and then zoom out to consider the future of digital collections in the context of technologies of surveillance, precarity of both cultural heritage institutions and cultural heritage workers in the context of neoliberalism, and then explore the broad set of challenges facing the future of collections stemming from the increasing effects of anthropogenic climate change. Drawing on frameworks for maintenance, care, and repair this talk concludes with an opportunity to reflect on and consider how memory and information workers should approach the digital present and future of our institutions and professions.
Theory and Craft of Digital Preservation Lightning TalkTrevor Owens
I’m thrilled and honored to be a finalist for the Dutch Digital Heritage Network Award for Teaching and Communications for my book, The Theory and Craft of Digital Preservation. This is particularly significant to me for two reasons. First, I started out on this book directly as a teaching and communications effort and second because the international digital preservation community that DPC supports and encourages has been so vital in helping me develop and refine the ideas in this book. For this talk, I’m going to give a little context of where the book came from, how it was developed, and the overwhelming response I’ve received for it all of which I think make it a good fit for this particular award.
More Related Content
Similar to Digital Preservation: Understanding the Risks
ONLINE EDUCATIONAL INCIDENT RESPONSE TRAINING: VIRTUAL WORLDS AS A COST EFFEC...Maurice Dawson
Critical incident leadership and planning is an important topic in the field of homeland security. Within this field are many challenges. For the purposes of this forum only two shall be discussed. The two challenges related to critical incident leadership and planning is the ability to learn in a simulated environment and funding. The first challenge surrounds learning as there is a significant cost associated in providing a real life scenarios where employees can learn. Employees need to understand how they should react in a simulated process to ensure that they can perform in the real event. This simulated scenario also allows an independent group to measure the effectiveness of the incident response manager. An inexpensive and effective method to do this would be through the
utilization of Second Life (SL). SL is one of the most popular virtual world provides a unique and secure teaching and learning environment for instructors and students. This learning environment could be designed to be private for a specific response group or incident where each avatar assigned to a unique name could include the name of group name that the individual is belong to. The second issue,
which surrounds funding, directly correlates with the capabilities and readiness to respond to a group. To properly correct this one would need to find a method to assign funding needs to risks or events that are most likely to occur.
Digital preservation aims to maintain access to digital materials over time despite changes in technology. It faces challenges such as the obsolescence of storage technologies, instability of storage media, and maintaining the integrity of digital materials. Recommendations include starting preservation close to creation to ensure future access and increasing awareness of preservation techniques. While skepticism remains, institutions like the Library and Archives of Canada are mandated to preserve digital heritage for present and future generations through continued migration to new technologies and formats.
A survey on privacy preserving data publishingijcisjournal
Data mining is a computational process of analysing and extracting the data from large useful datasets. In
recent years, exchanging and publishing data has been common for their wealth of opportunities. Security,
Privacy and data integrity are considered as challenging problems in data
mining.Privacy is necessary to protect people’s interest in competitive situations. Privacy is an abilityto
create and maintain different sort of social relationships with people. Privacy Preservation is one of the
most important factor for an individual since he should not embarrassed by an adversary. The Privacy
Preservation is an important aspect of data mining to ensure the privacy by various methods. Privacy
Preservation is necessary to protect sensitive information associated with individual. This paper provides a
survey of key to success and an approach where individual’s privacy would to be non-distracted.
This document discusses challenges and opportunities for discovering and documenting biodiversity in the current information age. It argues that current taxonomic processes are too slow and that new approaches are needed to integrate distributed data sources and leverage community contributions. Specifically, it proposes:
1) Publishing new biodiversity data prior to formal documentation to accelerate discovery.
2) Developing automated workflows and online workspaces to integrate phylogenetic, distribution, and trait data.
3) Enabling community participation through open data sharing and collaborative annotation platforms.
This document discusses challenges and opportunities for discovering and documenting biodiversity in the current information age. It argues that current taxonomic processes are too slow and that new approaches are needed to integrate distributed data sources and leverage community sourcing. Specifically, it advocates for:
1) Publishing new biodiversity data prior to formal documentation to accelerate discovery.
2) Developing automated workflows and online workspaces to integrate phylogenetic, distribution, and trait data.
3) Enabling community participation in annotating and improving global biodiversity models and maps.
4) Changing incentives to value data sharing over individual "kudos" and prioritize the collective good of the scientific community.
The document discusses the issue of uneven distribution of biodiversity data around the world, with much of the data held by small publishers and citizen scientists. It notes that these "small data publishers" face challenges in discovering, accessing, managing and publishing their data according to standards. The document calls for developing standards and tools that make it easier for small data publishers to capture, organize and share their biodiversity data in order to help mobilize this important but hard to access data.
Meeting the NSF DMP Requirement June 13, 2012IUPUI
The document provides guidance on developing a data management plan (DMP) to meet requirements for National Science Foundation grant proposals. It discusses the context and rationale for federal data policies, defines the key elements required for a DMP, and provides examples of DMPs for different types of research data. The main points are: understanding the NSF data policy aims to increase research impact and data sharing/reuse; a DMP must address the types of data generated, metadata standards, data access/sharing plans, long-term preservation, and associated costs; and good planning helps ensure data remains accessible, usable and preserved into the future. Resources and guidance are available to help researchers develop robust and fundable DMPs.
VOLUME-7 ISSUE-8, AUGUST 2019 , International Journal of Research in Advent Technology (IJRAT) , ISSN: 2321-9637 (Online) Published By: MG Aricent Pvt Ltd
This document proposes a refinement of the slicing anonymization technique for privacy-preserving data mining. Slicing anonymization has been shown to effectively preserve data quality while achieving high data privacy. The proposed refinement aims to achieve even higher data utility and more secure data publishing through probabilistic non-homogeneous suppression and consideration of attribute correlations. The results of applying the technique to election data are analyzed using standard classification metrics to validate that the refined approach maintains high data quality and strong privacy preservation.
Session 4 KE4CAP Pres1 bo dwyer-ke4_cap_canadaweADAPT
This document discusses user-oriented climate adaptation platforms and approaches to integrating user needs. It defines climate services as providing climate products or advice to assist decision-making. Traditionally, climate services took a top-down approach, but now focus on understanding user contexts and specifying requirements. Meeting user needs involves tailoring information, evolving user capacities over time, and adapting to evolving policy requirements. Key approaches discussed are web-based information provision, capacity building, and direct policy support.
Resilience is introduced as the new security goal supported with security/safety-related information by data-centric services for predictive risk management in real-time. Secondary use of personal information is of essential importance. The problem is that data-centric services threaten resilience. Although privacy as a state of equilibrium and its enforcement with usable security by identity management aims actually at decreasing users’ own risk, its use by data-centric services for unilateral information flow control threatens privacy and resilience. Users lose control on their identity while at the same time competitiveness of in particular small and medium service providers is endangered due to reliable statements on authentication of derived information. Self-protection, however, depends on opposite security interests. This talk claims that Multilateral Security improves privacy and resilience by a multilateral secondary use of personal security-related information for distributed usage control. This kind of privacy is understood as informational self-determination whereas the key concept is non-linkable delegation of rights on secondary use of personal information.
presented at the workshop "Usable Security and Privacy" an event of "Mittelstand-Digital" of the Federal Ministry for Economic Affairs and Energy (BMWi) and HCI conference "Mensch und Computer 2015" in Stuttgart, Germany http://www.mittelstand-digital.de/DE/Service/suche,did=717526.html
Ensuring that an organisation's digital assets are safe, secure and accessible for the long term should (in theory) be an interesting, responsible and useful role for anyone in an organisation to accept. The critical importance of digital assets, the ubiquity of digital methods and the need for people in all walks of life to have effective means to refer to persistent sources of data reinforce this notion. How is it then that long-term asset management, information lifecycle management, data curation, digital preservation (call it what you will) is often regarded as a peripheral specialist activity that it is diffcult to resource, complex to carry out, and delivers benefits that are, at best, simply an insurance policy rather than an activity that adds value to an organisation?
This presentation will examine the importance of defining clear roles for those involved with digital preservation and will consider the importance of associating this professional activity with strategic and tactical frameworks. It is likely that automated services will increasingly be required to deal with the collosal amount of digital information that will be produced and consumed over the next century and whilst the type and nature of these services are yet to be defined, we can be fairly certain of one endurng requirement, namely, that human judgement will always be needed to curate interesting and useful content for future generations.
Open Source Software for Digital Preservation Repositories : A SurveyIJCSES Journal
In the digital age, the amount of data produced is growing exponentially. Governments and institutions can no longer rely on old methods for storing data and passing on the knowledge to future generations. Digital data preservation is a mandatory issue that needs proper strategies and tools. With this awareness, efforts are being made to create and perfect software solutions capable of responding to the challenge of properly preserving digital information. This paper focuses on the state-of-the-art in open-source software solutions for the digital preservation and curation field used to assimilate and disseminate information to designated audiences. Eleven open source projects for digital preservation are surveyed in areas such as supported standards and protocols, strategies for preservation, methodologies for reporting, dynamic of development, targeted operating systems, multilingual support and open source license. Furthermore, five of these open
source projects, are further analysed, with focus on features deemed important for the area. Along open source solutions, the paper also briefly surveys the standards and protocols relevant for digital data preservation. The area of digital data preservation repositories has several open source solutions, which can form the base to overcome the challenges to reach mature and reliable digital data preservation.
Open Source Software for Digital Preservation Repositories : A SurveyIJCSES Journal
In the digital age, the amount of data produced is growing exponentially. Governments and institutions can no longer rely on old methods for storing data and passing on the knowledge to future generations. Digital data preservation is a mandatory issue that needs proper strategies and tools. With this awareness, efforts are being made to create and perfect software solutions capable of responding to the challenge of properly preserving digital information. This paper focuses on the state-of-the-art in open-source software solutions for the digital preservation and curation field used to assimilate and disseminate information to designated audiences. Eleven open source projects for digital preservation are surveyed in areas such as supported standards and protocols, strategies for preservation, methodologies for reporting, dynamic of development, targeted operating systems, multilingual support and open source license. Furthermore, five of these open
source projects, are further analysed, with focus on features deemed important for the area. Along open source solutions, the paper also briefly surveys the standards and protocols relevant for digital data preservation. The area of digital data preservation repositories has several open source solutions, which can form the base to overcome the challenges to reach mature and reliable digital data preservation.
This document summarizes a presentation by Professor Jacqueline McGlade from the European Environment Agency on drivers of air pollution and trends in emissions. It discusses many factors influencing air quality like population growth, energy use, and urbanization. While air pollution has decreased in some countries, emissions exceed limits in 8 EU states. New science shows Europe is still catching up on air quality. The document outlines tools and data used to monitor changes and manage natural resources. It promotes sharing data through open access platforms to better support policy and engage citizens in monitoring air quality.
EDAM (Eco-stations Data Access Monitor) is a project at BIDS (Berkeley Institute of Data Science) Collaborative, an initiative of the institute aims at solving real-world data challenges
This presentation gives an overview of the key things that we need to consider before publishing data from the repository. It briefly discusses research data management, research data lifecycle, FAIR principles of research data management and then move on to key elements that should be considered while preparing datasets for publishing through repository.
Introduction to research data management; Lecture 01 for GRAD521Amanda Whitmire
Lesson 1: Introduction to research data management. From a series of lectures from a 10-week, 2-credit graduate-level course in research data management (GRAD521, offered at Oregon State University).
The course description is: "Careful examination of all aspects of research data management best practices. Designed to prepare students to exceed funder mandates for performance in data planning, documentation, preservation and sharing in an increasingly complex digital research environment. Open to students of all disciplines."
Major course content includes: Overview of research data management, definitions and best practices; Types, formats and stages of research data; Metadata (data documentation); Data storage, backup and security; Legal and ethical considerations of research data; Data sharing and reuse; Archiving and preservation.
See also, "Whitmire, Amanda (2014): GRAD 521 Research Data Management Lectures. figshare. http://dx.doi.org/10.6084/m9.figshare.1003835. Retrieved 23:25, Jan 07, 2015 (GMT)"
Crowdsourcing Approaches to Big Data Curation for Earth SciencesEdward Curry
The document discusses crowdsourcing approaches to data curation for earth sciences. It covers several topics including motivation for data curation, data quality and curation processes, crowdsourcing, case studies on crowdsourced data curation, and setting up a crowdsourced data curation process. Specifically, it describes challenges with data quality, defines data curation and the role of data curators. It also outlines different types of data curation approaches based on who performs the curation (individual curators, departments, communities) and how it is done (manually, automated, crowdsourced).
Towards A Differential Privacy and Utility Preserving Machine Learning Classi...Kato Mivule
Kato Mivule, Claude Turner, Soo-Yeon Ji, "Towards A Differential Privacy and Utility Preserving Machine Learning Classifier", Procedia Computer Science (Complex Adaptive Systems), 2012, Pages 176-181, Washington DC, USA.
Similar to Digital Preservation: Understanding the Risks (20)
Caring for Digital Collections in the AnthropoceneTrevor Owens
The craft of digital preservation and digital collections care is anchored in the past. It builds off the records, files, and works of those who came before us and those who designed and set up the systems that enable the creation, transmission, and rendering of their work. At the same time, the craft of digital preservation is also the work of a futurist. We must look to the past trends in the ebb and flow of the development of digital media and hedge our bets on how digital technologies of the future will play out. This talk explores key issues for exploring and imagining that future. We start with consideration of some key emerging technologies relevant to digital collections and then zoom out to consider the future of digital collections in the context of technologies of surveillance, precarity of both cultural heritage institutions and cultural heritage workers in the context of neoliberalism, and then explore the broad set of challenges facing the future of collections stemming from the increasing effects of anthropogenic climate change. Drawing on frameworks for maintenance, care, and repair this talk concludes with an opportunity to reflect on and consider how memory and information workers should approach the digital present and future of our institutions and professions.
Theory and Craft of Digital Preservation Lightning TalkTrevor Owens
I’m thrilled and honored to be a finalist for the Dutch Digital Heritage Network Award for Teaching and Communications for my book, The Theory and Craft of Digital Preservation. This is particularly significant to me for two reasons. First, I started out on this book directly as a teaching and communications effort and second because the international digital preservation community that DPC supports and encourages has been so vital in helping me develop and refine the ideas in this book. For this talk, I’m going to give a little context of where the book came from, how it was developed, and the overwhelming response I’ve received for it all of which I think make it a good fit for this particular award.
Planning for Digital Preservation in OrganizationsTrevor Owens
This document outlines the key areas of digital preservation that were discussed at a 2019 workshop in Mexico City. The presenter used the Levels of Digital Preservation as an assessment tool to help understand the risks in each area of digital preservation, including storage, integrity, control, metadata, and content.
Make it Last: Principals for Digital Preservation and ConservationTrevor Owens
The document provides an overview of principles for digital preservation and conservation from a guest lecture given by Trevor Owens. It defines digital preservation as working to ensure enduring access to digital content. It then outlines 16 axioms of digital preservation, such as institutions being key enablers of long-term preservation and how digital preservation requires ongoing work and resources. It concludes with six practical steps anyone can take to start improving digital preservation efforts.
Testing Our Assumptions: The Centrality of Design Thinking and Scholarship fo...Trevor Owens
This document outlines an approach for the future of library practice that centers design thinking and scholarship. It argues that libraries are vital infrastructure for developing and disseminating knowledge and must grow and learn through dynamic applied research. The document recommends adopting user-centered design as a guiding framework, holistically mapping library ecosystems, developing a well-defined toolkit for projects, and using an iterative approach like Scrum. It also emphasizes considering values and impacts on the entire system level. Next steps include reframing work as places for experimentation, making design the focus of scholarship, including values in practice, and exploring Scrum for organizing teams.
We Have Interesting Problems: Some Applied Grand Challenges from Digital Libr...Trevor Owens
Libraries, Archives and Museums now have massive digital
holdings. There is tremendous potential for library and
information science, computer science and computer engineering
researchers to partner with cultural heritage institutions and
make our digital cultural record more useful and usable. In
particular, there is a significant need to bridge basic research in
areas such as computer vision, crowdsourcing, natural language
processing, multilingual OCR, and machine learning to make this
work directly usable in the practices of cultural heritage
institutions. In this talk, I discuss a series of exemplar projects,
largely funded through the Institute of Museum and Library
Services National Digital Platform initiative, that illustrate some
key principles for building applied research partnerships with
cultural heritage institutions. Building on Ben Schniderman’s
The New ABCs of Research: Achieving Breakthrough
Collaborations, I focus specifically on why the public purpose
and missions of cultural heritage institutions are particularly
valuable in establishing new kinds of collaborations that can
simultaneously advance basic research and the ability for people
of the world to engage with their cultural record.
Start Today: Digital Stewardship Communities & CollaborationsTrevor Owens
The increasingly digital records of our communities and our organizations require all of us to become digital stewardship and digital preservation practitioners. The challenge seems daunting but the good news is we don’t have to do it alone. A distributed network of practitioners and learners across the country are increasingly finding ways to learn together and share and pool their resources to tackle these challenges and provide enduring access to our digital heritage. Owens’ talk will provide examples of how archivists are rising to the challenge and practical guidance for both digital preservation beginners and experts.
Scientists’ Hard Drives, Databases, and Blogs: Preservation Intent and Source...Trevor Owens
Carl Sagan’s WordPerfect files, simulations emailed to Edward Lorenz, a database application from the National Library of Medicine, a collection of science blogs, a database of interstellar distances; each of these digital artifacts have been acquired by archives and special collections. Born digital primary sources are no longer a future concern for archivists, librarians, curators and historians. As historians of science turn their attention to the late 20th and early 21st century, they will need to work from these born-digital primary sources. We have already accumulated a significant born digital past and it’s time for work with born digital primary sources to become mainstream. This presentation will give a quick tour of individual born digital artifacts toward two goals. First, I argue for the need for archivists, curators and librarians to reflexively develop approaches to establishing preservation intent for digital content grounded in a dialog with the nature of a given set of digital objects and it’s future research use. Second, for historians, I suggest how trends in computational analysis of information in the digital humanities should be combined with approaches from digital forensics and new media studies to establish historiographic practices for born-digital source criticism. I conclude by suggesting the kinds of technical skills archivists, librarians, curators and historians working with these materials are going to need to develop. Just as historians working with premodern documents require language and paleography skills, historians working with digital artefacts will increasingly need to understand the inscription processes of hard drives, the provenance created by web crawlers, and how to read relational databases of varying vintages.
Platform Thinking: Frameworks for a National Digital Platform State of MindTrevor Owens
Talk presented as a closing keynote to the Biodiversity Heritage Library's National Digital Stewardship Residency program meeting at the National Museum of Natural History. This talk reviews the National Digital Platform framework developed by US IMLS in collaboration with various library, archives and museum stakeholders and presents a series of additional conceptual frameworks on the role of software in society and psychology.
Digital Infrastructures that Embody Library Principles: The IMLS national dig...Trevor Owens
Digital library infrastructures must not simply work. They must also manifest the core principles of libraries and archives. Since 2014, the Institute of Museum and Library Services (IMLS) has engaged with stakeholders from diverse library communities to consider collaborative approaches to building digital library tools and services. The “national digital platform” for libraries, archives, and museums is the framework that resulted from these dialogs. One key feature of the national digital platform (NDP) is the anchoring of core library principles within the development of digital tools and services. This essay explores how NDP-funded projects enact library principles as part of the national framework.
The IMLS National Digital Platform & Your Library: Tools You Can UseTrevor Owens
As libraries increasingly use digital infrastructure to provide access to content and resources, there are more and more opportunities for collaboration around the tools and services that they use to meet their users’ needs. To this end, the Institute of Museum and Library Services (IMLS) is making substantial investments in developing collaborative and sustainable technical and social digital infrastructure for libraries through the National Digital Platform initiative. In this talk, you will learn about a series of digital tools, services, training opportunities and resources IMLS is funding through the National Leadership Grants for Libraries Program and the Laura Bush 21st Century Librarian Program. The presentation will focus on ongoing projects and efforts that you and your library can get involved in and make direct use of. It will also provide insight into how you could develop competitive proposals for projects that could be funded through this national effort.
Next Steps for IMLS's National Digital PlatformTrevor Owens
This keynote, at the Upper Midwest Digital Collections Conference, provides and update on the National Digital Platform and 20 projects supported to enhance it. The national digital platform is a way of thinking about and approaching the digital capability and capacity of libraries across the US. In this sense, it is the combination of software applications, social and technical infrastructure, and staff expertise that provide library content and services to all users in the US. As libraries increasingly use digital infrastructure to provide access to digital content and resources, there are more and more opportunities for collaboration around the tools and services that they use to meet their users’ needs. It is possible for each library in the country to leverage and benefit from the work of other libraries in shared digital services, systems, and infrastructure.
We need to bridge gaps between disparate pieces of the existing digital infrastructure, for increased efficiencies, cost savings, access, and services. To this end, IMLS is focusing on the national digital platform as an area of priority in the National Leadership Grants to Libraries program and the Laura Bush 21st Century Librarian program. We are eager to explore how this way of thinking and approaching infrastructure development can help states make the best use of the funds they receive through the Grants to States program. We’re also eager to work with other foundations and funders to maximize the impact of our federal investment
Next Steps for IMLS's National Digital PlatformTrevor Owens
This document summarizes projects funded by the Institute of Museum and Library Services (IMLS) related to developing a National Digital Platform. It describes 7 projects improving open source digital library software tools and communities, 4 projects focused on scaling up shared services, 2 applied research projects related to collections at scale, and 3 projects aimed at improving access for all and inclusion. It provides brief descriptions and links to more information for each of the 20 projects. The overall goal is to expand the digital capability and capacity of libraries across the United States by prioritizing promising digital tools and services.
People, Communities and Platforms: Digital Cultural Heritage and the WebTrevor Owens
Libraries, archives and museums are sites of community memory. The first public computerized bulletin board system was called community memory. Trevor’s talk will explore the connections between the development of the web as a global knowledge base, the open source software movement, and digital strategy for libraries, archives and museums. This keynote talk will synthesize research on the history of online community software with practical experience working on open source digital library projects. This exploration underscores the essential role cultural heritage institutions need to play in this era of the web and some important distinctions between how the concept of community is deployed in discussions of the web.
Macroscopes and Distant Reading: Implications for Infrastructures to Support ...Trevor Owens
A talk exploring the implications for digital library infrastructures in the face of developments in how humanities scholars are engaging in computational research of library collections.
Digital Preservation's Role in the Future of the Digital HumanitiesTrevor Owens
Slides from an invited presentation I gave to the University of Pittsburgh's iSchool.
"Ensuring long term access to digital information sounds like a technical problem. It seems like digital preservation should be a computer science problem. Far from it. In this lecture Trevor Owens, a digital archivist at the Library of Congress argues that digital preservation is in fact a core problem and issue at the heart of the future of the digital humanities. Bringing together perspectives from the history of technology, new media studies, public history, and archival theory, he suggests the critical role that humanities scholars and practitioners should play in framing and shaping the collection, organization, description, and modes of access to the historically contingent digital material records of contemporary society."
Slides for an invited presentation I gave to the the National Archives and Records Administration’s Online Public Access (OPA) Integrated Product Team in College Park, MD in 2013.
This document discusses a student project to create a mobile game interpreting ruins and a glass house. It provides links to student blogs about the project. It includes screenshots from a prototype and discusses what elements of the project are significant - the student reports, storyboards, and reflections, rather than the app itself. It also discusses how meaning and significance can be found in artifacts and their documentation.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Building RAG with self-deployed Milvus vector database and Snowpark Container...Zilliz
This talk will give hands-on advice on building RAG applications with an open-source Milvus database deployed as a docker container. We will also introduce the integration of Milvus with Snowpark Container Services.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Digital Preservation: Understanding the Risks
1. Trevor Owens, July 11 2019, Workshop:
Forecasting Costs for Preserving, Archiving,
and Promoting Access to Biomedical Data.
National Academy of Sciences. Washington
D.C.
2. DIGITAL PRESERVATION
Ensuring long term access to
significant digital content. In
practice, fundamentally a process of
identifying and responding to
potential risks of loss of access and
usability.
3. NDSA LEVELS OF
DIGITAL PRESERVATION
Phillips, M., Bailey, J., Goethals, A., & Owens, T. (2013). The NDSA Levels of
Digital Preservation: An Explanation and Uses. In Proceedings of the 2013
IS&T Archiving Conference.
5. STORAGE AND GEOGRAPHY
Risk: Damage to storage media could
result in total loss of data.
Mitigation Action: Manage multiple
copies, in different geographic regions
that face different disaster threats.
6. FILE FIXITY AND DATA INTEGRITY
Risk: Loss of data through use,
transactions, or bit rot.
Mitigation Action: Generate, track, log,
and manage fixity information across
copies and use data to repair bad
copies.
7. INFORMATION SECURITY
Risk: Loss of data through unauthorized
user actions.
Mitigation Action: Manage access
restrictions, log actions on files, audit
logs.
8. METADATA
Risk: Loss of usability of data or ability
to authenticate data.
Mitigation Action: Produce and manage
administrative, technical, descriptive,
and preservation metadata and
maintain non-co-located copies of
metadata.
9. FILE FORMATS
Risk: Loss of usability/renderability of
data.
Mitigation Actions: Articulate
preservation intent, limit format support
re sustainability factors, inventory
formats, validate files, produce
derivatives, virtualization & emulation.
10. PLANNING APPROACH
The best way to mitigate these risks
involves 1) hiring permanent trained
staff and 2) resourcing them as a
central cost center and to 3) plan on
a supporting a continual refresh
cycle for software and hardware.
11. LOOK AT THE LEDGER
If you want to know if an organization is
serious about digital preservation ask
their accountants. What part of core
operations resources are invested in
staffing, contracts, and software and
hardware dedicated to digital
preservation?