The document describes the Humanities Networked Infrastructure (HuNI) project. HuNI aims to create a virtual laboratory that integrates 28 Australian cultural datasets and enables new forms of humanities research. It will harvest data from partner organizations, transform it into a searchable format and linked open data, and develop tools for researchers to discover, analyze, annotate, and share collections across the integrated datasets. The project is led by Deakin University with funding from NeCTAR and contributions from partner organizations.
The document discusses the Humanities Networked Infrastructure (HuNI) project. HuNI aims to (1) integrate cultural data from 28 sources at a national level, (2) make this aggregated data accessible through a new national data service, and (3) connect the data to the Linked Data Cloud to allow for complex queries. The HuNI lab app will allow users to discover, explore, connect, curate and share data as well as save and import their own data. HuNI intends to change the nature of humanities research by enabling work with larger datasets and breaking down disciplinary data boundaries to promote sharing and collaboration.
Research My World: Pilot Project EvaluationDeb Verhoeven
This document provides an evaluation of a pilot crowdfunding project called "Research My World" conducted by Deakin University in Australia. The key findings were:
- 6 out of 8 research projects were successfully funded, raising over $50,000 with additional funds raised after.
- The projects generated over 200 media stories reaching over 1.4 million people and over 3,600 tweets.
- The researchers improved their digital and social media skills, and saw increased profiles and networks.
- Factors like the reach of Twitter networks and driving traffic to project websites correlated with funding success.
Big CInema Data: Analysing global cinema showtimesDeb Verhoeven
Looking at cinema exhibition and distribution at an international scale requires data beyond broad aggregates, it requires data that is specific to individual films and cinema venues in order to appreciate the intricate temporal and geographic aspects of flow and patterns. The Kinomatics Project has tracked the global flow of individual film screenings (down to date and time) for over 54,000 films for 30,000 venues throughout 48 countries internationally.
This presentation will highlight the importance of global scale analysis and data through three case studies. The first will track the spatial and temporal relationships of The Hobbit: an unexpected journey, highlighting the complexities of international cinema enterprises and the subtleties of contemporary releasing strategies. The second explores the relationship between remittance flows and the movement of film around the globe with a focus on Bollywood films. The thrid test dyadic relationships between countries. This presentation will introduce some methods for analysing and visualising data used in the three case studies.
Songification: Enhancing live music dataDeb Verhoeven
This document summarizes a presentation about using music gig data to create musical compositions. It discusses analyzing the location, date, and frequency of gigs played by several Australian bands from the 1960s-1970s. This data is converted into musical notes based on the frequency and distance of gigs from a central point. The resulting notes are played together to create songified versions of the bands' gig histories and movements around venues in that era. Examples of songs created from the gig data of bands like Max Merritt and the Meteors, Billy Thorpe and the Aztecs, and Doug Parkinson are presented.
The document summarizes the history and dispersal of the Phillipps manuscript collection. It was assembled by Sir Thomas Phillipps in the 19th century and grew to over 60,000 volumes, making it one of the largest medieval manuscript collections ever amassed by a private collector. After Phillipps' death, the collection was reorganized and portions were sold off over the subsequent decades to libraries and at auction. The document outlines the project to digitally represent the provenance of manuscripts in the collection using linked data approaches and visualize their transmission over time and between collectors.
Presentation by Toby Burrows and Deb Verhoeven to the Fifth National Forum of AeRO (the Australian eResearch Organization), held in Perth on 26 July 2013. The presentation gives an overview of the HuNI Project as at July 2013. Topics covered include: data ingest and alignment from 28 Australian humanities datasets; building HuNI’s discovery functionality; and designing Virtual Laboratory tools for researchers.
Presentation at the Australasian Consortium of Humanities Research Centres (ACHRC), July 2013. Panel description:
The Digital Humanities offers not only new tools to support what we do in the Humanities, but also new ways of thinking about what it is that we do. This panel will build upon Alan Liu’s keynote discussion of ideas for digital tools for humanities advocacy and speak to the way non-digital centres can benefit from digital humanities initiatives.
The document discusses the Humanities Networked Infrastructure (HuNI) project. HuNI aims to (1) integrate cultural data from 28 sources at a national level, (2) make this aggregated data accessible through a new national data service, and (3) connect the data to the Linked Data Cloud to allow for complex queries. The HuNI lab app will allow users to discover, explore, connect, curate and share data as well as save and import their own data. HuNI intends to change the nature of humanities research by enabling work with larger datasets and breaking down disciplinary data boundaries to promote sharing and collaboration.
Research My World: Pilot Project EvaluationDeb Verhoeven
This document provides an evaluation of a pilot crowdfunding project called "Research My World" conducted by Deakin University in Australia. The key findings were:
- 6 out of 8 research projects were successfully funded, raising over $50,000 with additional funds raised after.
- The projects generated over 200 media stories reaching over 1.4 million people and over 3,600 tweets.
- The researchers improved their digital and social media skills, and saw increased profiles and networks.
- Factors like the reach of Twitter networks and driving traffic to project websites correlated with funding success.
Big CInema Data: Analysing global cinema showtimesDeb Verhoeven
Looking at cinema exhibition and distribution at an international scale requires data beyond broad aggregates, it requires data that is specific to individual films and cinema venues in order to appreciate the intricate temporal and geographic aspects of flow and patterns. The Kinomatics Project has tracked the global flow of individual film screenings (down to date and time) for over 54,000 films for 30,000 venues throughout 48 countries internationally.
This presentation will highlight the importance of global scale analysis and data through three case studies. The first will track the spatial and temporal relationships of The Hobbit: an unexpected journey, highlighting the complexities of international cinema enterprises and the subtleties of contemporary releasing strategies. The second explores the relationship between remittance flows and the movement of film around the globe with a focus on Bollywood films. The thrid test dyadic relationships between countries. This presentation will introduce some methods for analysing and visualising data used in the three case studies.
Songification: Enhancing live music dataDeb Verhoeven
This document summarizes a presentation about using music gig data to create musical compositions. It discusses analyzing the location, date, and frequency of gigs played by several Australian bands from the 1960s-1970s. This data is converted into musical notes based on the frequency and distance of gigs from a central point. The resulting notes are played together to create songified versions of the bands' gig histories and movements around venues in that era. Examples of songs created from the gig data of bands like Max Merritt and the Meteors, Billy Thorpe and the Aztecs, and Doug Parkinson are presented.
The document summarizes the history and dispersal of the Phillipps manuscript collection. It was assembled by Sir Thomas Phillipps in the 19th century and grew to over 60,000 volumes, making it one of the largest medieval manuscript collections ever amassed by a private collector. After Phillipps' death, the collection was reorganized and portions were sold off over the subsequent decades to libraries and at auction. The document outlines the project to digitally represent the provenance of manuscripts in the collection using linked data approaches and visualize their transmission over time and between collectors.
Presentation by Toby Burrows and Deb Verhoeven to the Fifth National Forum of AeRO (the Australian eResearch Organization), held in Perth on 26 July 2013. The presentation gives an overview of the HuNI Project as at July 2013. Topics covered include: data ingest and alignment from 28 Australian humanities datasets; building HuNI’s discovery functionality; and designing Virtual Laboratory tools for researchers.
Presentation at the Australasian Consortium of Humanities Research Centres (ACHRC), July 2013. Panel description:
The Digital Humanities offers not only new tools to support what we do in the Humanities, but also new ways of thinking about what it is that we do. This panel will build upon Alan Liu’s keynote discussion of ideas for digital tools for humanities advocacy and speak to the way non-digital centres can benefit from digital humanities initiatives.
CENDARI Summer School July 2015 BurrowsToby Burrows
This document discusses how medieval studies can benefit from applying digital humanities approaches and linked open data principles. It notes that while there are many digitized resources for medieval texts and manuscripts, they generally exist independently without connections between them. The document advocates for assigning unique identifiers to medieval people, places, manuscripts and their components to facilitate linking related data across different projects and datasets. This would allow for more integrated searching, browsing and analysis of medieval materials by establishing semantic relationships between entities. Achieving widespread interoperability through linked open data approaches could help address the current problems of information overload and duplication of effort in medieval digital scholarship.
This document discusses crowdfunding for university research projects. It provides information on different types of crowdfunding platforms, including domain-specific, university-specific, and education platforms. Data is presented on the success of the Pozible crowdfunding platform in launching and fully funding research projects. Charts show funding amounts over time for individual projects and overall. Factors that contribute to project success are discussed, including the Twitter network of the project principal and social media engagement with the project website.
An information system is designed to capture, store, process, and provide access to information to support organizational processes and decision making. The document discusses the design of a resource registry information system to support a hybrid cloud-based infrastructure. The resource registry collects and manages metadata about software systems, resources, and their status to enable service discovery, monitoring, and elastic resource allocation. It implements an open model to flexibly support evolving resource types and management needs over the long lifespan of the infrastructure.
Ontologies and the humanities: some issues affecting the design of digital in...Toby Burrows
This document discusses issues related to designing digital infrastructure for the humanities using ontologies. It notes that there are many ongoing efforts to develop ontologies for different domains in digital humanities. However, it also acknowledges linguistic, semantic, and conceptual difficulties in representing humanities knowledge through ontologies. As an alternative, it discusses strategies like topic modeling, linked data, and conceptual spaces that may better capture humanistic perspectives on relationships, cognition, and meaning. It argues that future humanities research should look beyond ontologies alone and examine computational modeling from cognitive science and philosophy.
1) The document discusses the opportunity for technology to improve organizational efficiency and transition economies into a "smart and clean world."
2) It argues that aggregate efficiency has stalled at around 22% for 30 years due to limitations of the Second Industrial Revolution, but that digitizing transport, energy, and communication through technologies like blockchain can help manage resources and increase efficiency.
3) Technologies like precision agriculture, cloud computing, robotics, and autonomous vehicles may allow for "dematerialization" and do more with fewer physical resources through effects like reduced waste and need for transportation/logistics infrastructure.
NeCTAR is an Australian government initiative to enhance research collaboration through the development of eResearch infrastructure. The University of Melbourne has been appointed lead agent. Objectives include developing a national cloud computing infrastructure called the NeCTAR Research Cloud using OpenStack. The cloud aims to provide researchers from any discipline with access to computing resources and collaboration tools from multiple sites around Australia.
an overview about EMBL-ABR, including bioinformatics infrastructure initiatives at national and pan-national level across the globe and activities EMBL-ABR is currently doing.
Open science refers to making scientific research and data accessible to all. It includes open access to publications, open data, open source software, open notebooks, and citizen science. The European Union supports open science to increase transparency, collaboration and innovation in research. A workshop was held in South Africa to help develop an open science policy, with feedback that the policy will be finalized in September 2018 after additional workshops with European Union involvement. Open science aims to make the entire research process publicly available and reusable to maximize scientific progress.
The African Open Science Platform (AOSP) aims to promote open data and open science across Africa. It is funded by the National Research Foundation of South Africa and managed by the Academy of Science of South Africa. AOSP focuses on developing policy frameworks, infrastructure, capacity building initiatives, and incentives to encourage data sharing across four key areas. It has already held several workshops and events in its first two years and outlines further actions and deliverables to advance open science in Africa.
RDAP 15 EarthCollab: Connecting Scientific Information Sources using the Sema...ASIS&T
Research Data Access and Preservation Summit, 2015
Minneapolis, MN
April 22-23, 2015
Erica M. Johns, Jon Corson-Rikert, Huda J. Khan, Dean B. Krafft and Matthew S. Mayernik
Presentation given to the first conference Pubmet, Zadar, Croatia, 2014
For the live presentation having all the rich media, please access: http://kosson.ro/webpedia/presentationsnicolaiec/Croatia2014/#/
The document summarizes the African Open Science Platform (AOSP). It discusses that AOSP aims to coordinate open science activities across Africa to increase collaboration, data sharing and reuse. It is focused on four key areas: establishing an open data forum, funding research infrastructure initiatives, funding transdisciplinary research projects, and developing open data policies. The ultimate goals are to accelerate discovery through open data, attract more funding, and contribute to global knowledge. AOSP is funded by the National Research Foundation of South Africa and managed by the Academy of Science of South Africa.
The document discusses the African Open Science Platform (AOSP) project. It notes that AOSP aims to support the development of open science in Africa by providing a coordinating platform. Specifically, it will work to establish an African open data platform, fund research data infrastructure initiatives, co-design open data policies, and develop incentives and training for research data science. The ultimate goal is to help African research institutions better manage, share and reuse research data according to FAIR open data principles.
This document provides an overview of the collaboration between the Australian National Data Service (ANDS) and the Terrestrial Ecosystem Research Network (TERN) to share Australia's nationally significant terrestrial ecosystem data. It describes ANDS and TERN, the establishment of national data collections, and a case study of the TERN national collection. The presentation demonstrates how the TERN collection is represented in Research Data Australia and outlines future work, such as bringing in more related ecosystem data assets and services.
The document discusses the African Open Science Platform (AOSP) project, which aims to support the development of open science in Africa. Key points:
- AOSP is a 3-year pilot project starting in 2016 that is funded by the South African Department of Science and Technology to establish an open data platform and coordinate open science initiatives across Africa.
- It is being implemented by the Academy of Science of South Africa (ASSAf) in partnership with organizations like the Association of African Universities (AAU) and UbuntuNet Alliance.
- The project involves several work packages, including establishing open data policies, research data infrastructure, training programs, and a roadmap for African research data.
- JK
Presentation of current challenges of upgrading the intrasturcture for access and preservation of social science research data and worklow in Slovene social science data archive
The document discusses funding opportunities from the National Natural Science Foundation of China (NSFC) for Chinese and European researchers. It provides an overview of NSFC, including its mission to support basic research, foster talented researchers, and strengthen international cooperation. It then describes various NSFC funding programs for individual researchers and international joint projects. Specific funding opportunities for Chinese-European collaboration are highlighted, such as joint research projects supported through agreements between NSFC and research organizations in Europe.
CENDARI Summer School July 2015 BurrowsToby Burrows
This document discusses how medieval studies can benefit from applying digital humanities approaches and linked open data principles. It notes that while there are many digitized resources for medieval texts and manuscripts, they generally exist independently without connections between them. The document advocates for assigning unique identifiers to medieval people, places, manuscripts and their components to facilitate linking related data across different projects and datasets. This would allow for more integrated searching, browsing and analysis of medieval materials by establishing semantic relationships between entities. Achieving widespread interoperability through linked open data approaches could help address the current problems of information overload and duplication of effort in medieval digital scholarship.
This document discusses crowdfunding for university research projects. It provides information on different types of crowdfunding platforms, including domain-specific, university-specific, and education platforms. Data is presented on the success of the Pozible crowdfunding platform in launching and fully funding research projects. Charts show funding amounts over time for individual projects and overall. Factors that contribute to project success are discussed, including the Twitter network of the project principal and social media engagement with the project website.
An information system is designed to capture, store, process, and provide access to information to support organizational processes and decision making. The document discusses the design of a resource registry information system to support a hybrid cloud-based infrastructure. The resource registry collects and manages metadata about software systems, resources, and their status to enable service discovery, monitoring, and elastic resource allocation. It implements an open model to flexibly support evolving resource types and management needs over the long lifespan of the infrastructure.
Ontologies and the humanities: some issues affecting the design of digital in...Toby Burrows
This document discusses issues related to designing digital infrastructure for the humanities using ontologies. It notes that there are many ongoing efforts to develop ontologies for different domains in digital humanities. However, it also acknowledges linguistic, semantic, and conceptual difficulties in representing humanities knowledge through ontologies. As an alternative, it discusses strategies like topic modeling, linked data, and conceptual spaces that may better capture humanistic perspectives on relationships, cognition, and meaning. It argues that future humanities research should look beyond ontologies alone and examine computational modeling from cognitive science and philosophy.
1) The document discusses the opportunity for technology to improve organizational efficiency and transition economies into a "smart and clean world."
2) It argues that aggregate efficiency has stalled at around 22% for 30 years due to limitations of the Second Industrial Revolution, but that digitizing transport, energy, and communication through technologies like blockchain can help manage resources and increase efficiency.
3) Technologies like precision agriculture, cloud computing, robotics, and autonomous vehicles may allow for "dematerialization" and do more with fewer physical resources through effects like reduced waste and need for transportation/logistics infrastructure.
NeCTAR is an Australian government initiative to enhance research collaboration through the development of eResearch infrastructure. The University of Melbourne has been appointed lead agent. Objectives include developing a national cloud computing infrastructure called the NeCTAR Research Cloud using OpenStack. The cloud aims to provide researchers from any discipline with access to computing resources and collaboration tools from multiple sites around Australia.
an overview about EMBL-ABR, including bioinformatics infrastructure initiatives at national and pan-national level across the globe and activities EMBL-ABR is currently doing.
Open science refers to making scientific research and data accessible to all. It includes open access to publications, open data, open source software, open notebooks, and citizen science. The European Union supports open science to increase transparency, collaboration and innovation in research. A workshop was held in South Africa to help develop an open science policy, with feedback that the policy will be finalized in September 2018 after additional workshops with European Union involvement. Open science aims to make the entire research process publicly available and reusable to maximize scientific progress.
The African Open Science Platform (AOSP) aims to promote open data and open science across Africa. It is funded by the National Research Foundation of South Africa and managed by the Academy of Science of South Africa. AOSP focuses on developing policy frameworks, infrastructure, capacity building initiatives, and incentives to encourage data sharing across four key areas. It has already held several workshops and events in its first two years and outlines further actions and deliverables to advance open science in Africa.
RDAP 15 EarthCollab: Connecting Scientific Information Sources using the Sema...ASIS&T
Research Data Access and Preservation Summit, 2015
Minneapolis, MN
April 22-23, 2015
Erica M. Johns, Jon Corson-Rikert, Huda J. Khan, Dean B. Krafft and Matthew S. Mayernik
Presentation given to the first conference Pubmet, Zadar, Croatia, 2014
For the live presentation having all the rich media, please access: http://kosson.ro/webpedia/presentationsnicolaiec/Croatia2014/#/
The document summarizes the African Open Science Platform (AOSP). It discusses that AOSP aims to coordinate open science activities across Africa to increase collaboration, data sharing and reuse. It is focused on four key areas: establishing an open data forum, funding research infrastructure initiatives, funding transdisciplinary research projects, and developing open data policies. The ultimate goals are to accelerate discovery through open data, attract more funding, and contribute to global knowledge. AOSP is funded by the National Research Foundation of South Africa and managed by the Academy of Science of South Africa.
The document discusses the African Open Science Platform (AOSP) project. It notes that AOSP aims to support the development of open science in Africa by providing a coordinating platform. Specifically, it will work to establish an African open data platform, fund research data infrastructure initiatives, co-design open data policies, and develop incentives and training for research data science. The ultimate goal is to help African research institutions better manage, share and reuse research data according to FAIR open data principles.
This document provides an overview of the collaboration between the Australian National Data Service (ANDS) and the Terrestrial Ecosystem Research Network (TERN) to share Australia's nationally significant terrestrial ecosystem data. It describes ANDS and TERN, the establishment of national data collections, and a case study of the TERN national collection. The presentation demonstrates how the TERN collection is represented in Research Data Australia and outlines future work, such as bringing in more related ecosystem data assets and services.
The document discusses the African Open Science Platform (AOSP) project, which aims to support the development of open science in Africa. Key points:
- AOSP is a 3-year pilot project starting in 2016 that is funded by the South African Department of Science and Technology to establish an open data platform and coordinate open science initiatives across Africa.
- It is being implemented by the Academy of Science of South Africa (ASSAf) in partnership with organizations like the Association of African Universities (AAU) and UbuntuNet Alliance.
- The project involves several work packages, including establishing open data policies, research data infrastructure, training programs, and a roadmap for African research data.
- JK
Presentation of current challenges of upgrading the intrasturcture for access and preservation of social science research data and worklow in Slovene social science data archive
The document discusses funding opportunities from the National Natural Science Foundation of China (NSFC) for Chinese and European researchers. It provides an overview of NSFC, including its mission to support basic research, foster talented researchers, and strengthen international cooperation. It then describes various NSFC funding programs for individual researchers and international joint projects. Specific funding opportunities for Chinese-European collaboration are highlighted, such as joint research projects supported through agreements between NSFC and research organizations in Europe.
This document discusses open science and research. It defines open science as making research transparent and accessible at all stages of the research process through open access, open data, open source code and open notebooks. It outlines the key elements of open science like open access publishing, open data repositories, open source software, citizen science and more. It also discusses open science initiatives in Europe, Africa and South Africa and the need for urgent policy actions to promote open science.
The document summarizes the African Open Science Platform (AOSP), which aims to coordinate open science activities across Africa. It discusses open science principles and the rationale for an African platform to increase collaboration, data reuse, and accelerate discovery. The AOSP is funded by the NRF of South Africa and managed by ASSAf. It focuses on establishing an open data forum, funding research infrastructure and projects, developing open data policies and training, and creating incentives for data sharing. Initial actions include workshops on policy, capacity building, and surveys to inform the platform's development.
The Humanities Networked Infrastructure (HuNI) combines data from 30 Australian cultural websites into the largest humanities and creative arts database in Australia. It covers all disciplines and brings together information on people, works, events, organizations and places that make up Australia's rich cultural landscape. Researchers can search HuNI, save search results as virtual collections, refine collections by adding links between records, and share collections with other researchers.
The document contains checklists for researchers and universities to prepare for crowdfunding campaigns. The institutional checklist covers having a point person, payment methods, receipting procedures, PR and marketing support, and being prepared to support projects beyond the campaign. The researcher checklist covers having an existing project, promotion and networking strategies, production of a promotional video, dedicating time to promotion, and support from their school and faculty. The overall document provides guidance to help ensure researchers and their universities are ready to successfully conduct a crowdfunding campaign.
Visualizing Cinema Data: Presentation at HOMER (Prague 2013)Deb Verhoeven
Cinema data is characteristically complex, heterogeneous and interlinked. Rather than relying on simple information retrieval techniques, researchers are increasingly turning to the creative exploration and reapplication of data in order to more fully explore the meaning of newly available and diverse data sets. In this context, the cinema historian becomes the creator of visual texts which can be assessed for both their interpretive insight and their aesthetic qualities. This paper presents four research projects that develop different spatio-temporal visualisation techniques to understand the industrial dynamics of post-war film exhibition and distribution in Australia. The research integrates groundbreaking work by a group of inter-disciplinary investigators into the effectiveness of techniques such as dendritic mapping, Circos, time-series graphs, animation, cartogram mapping, and multivariate visualisation for the study of cinema circuits and operations at a number of scales.
Kinomatics: Presentation at HOMER (Prague 2013) Deb Verhoeven
The document discusses the Kinomatics project which uses large datasets of cinema data to analyze trends in the global film industry. The project is led by researchers from Deakin and RMIT Universities and collects data on movie showtimes, venues and box office earnings from 48 countries. It aims to use this "big data" to better understand factors influencing the film industry and enable predictive analysis. The volume of data collected each week is demonstrated and challenges around data veracity are also discussed.
Mapping the Australian Screen Content ProducerDeb Verhoeven
The document discusses a survey of Australian screen content producers. It aimed to map the culture, motivations, and aspirations of over 4,000 producers in defined populations. The survey included open and closed questions about classifications, projects, education, employment, industry sentiment, attitudes, and perceptions.
The document also discusses the role and responsibilities of a producer. Producers are involved in all aspects of a production from development through post-production and marketing. They oversee the creative process and make important decisions. While no single producer handles every task, they must perform a majority of producing functions.
Producers tend to work in industries like health, education, media, and finance. They are motivated more by intrinsic rewards than
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Things to Consider When Choosing a Website Developer for your Website | FODUUFODUU
Choosing the right website developer is crucial for your business. This article covers essential factors to consider, including experience, portfolio, technical skills, communication, pricing, reputation & reviews, cost and budget considerations and post-launch support. Make an informed decision to ensure your website meets your business goals.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
AI-Powered Food Delivery Transforming App Development in Saudi Arabia.pdfTechgropse Pvt.Ltd.
In this blog post, we'll delve into the intersection of AI and app development in Saudi Arabia, focusing on the food delivery sector. We'll explore how AI is revolutionizing the way Saudi consumers order food, how restaurants manage their operations, and how delivery partners navigate the bustling streets of cities like Riyadh, Jeddah, and Dammam. Through real-world case studies, we'll showcase how leading Saudi food delivery apps are leveraging AI to redefine convenience, personalization, and efficiency.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
2. CRICOS Provider Code: 00113B
NATIONAL E-RESEARCH
COLLABORATION TOOLS AND
RESOURCES (NECTAR)
NeCTAR is a $47 million dollar, Australian Government
project, conducted as part of the Super Science
initiative and financed by the Education Investment
Fund. The University of Melbourne is the lead
agent, chosen by the Commonwealth Government.
4. • Ensure that Australian cultural datasets and the research
associated with them become part of the emerging
international Linked Open Data environment.
• Enable research enquiries to move easily from: what is?
to where is?
• Support the role of annotation and metadata in discovery
of new knowledge or the means to elucidate new
knowledge
• Position the idea of data as both a subject and object of
analysis in humanities
• Contribute to debates around standards for development
and implementation
HuNI BROAD BENEFITS
5. • Enable humanities researchers to work with cultural datasets
more efficiently and effectively, and on a larger scale;
• Encourage the systematic sharing of research data between
humanities researchers (including the cultural dataset
curators themselves), the community and cultural
institutions;
• Encourage a greater level of cross-disciplinary and
interdisciplinary research, both within the
humanities/creative arts and between the
humanities/creative arts and other disciplines, and the wider
public;
• Support innovative methodologies such as network
analysis, game theory and ‘virtual history’ that rely on large-
scale datasets
HUNI: SPECIFIC BENEFITS
6. 1. Organisational level: the goals and processes of the institutions
involved
2. The semantic level: meaning of the exchanged digital resources
3. Technical level: implementing data interoperability requires
both data integration and data exchange processes as well as
enabling effective use of the data that becomes available
Pasquale Pagano, ‘Data Interoperability’ (GRDI2020)
4. Project level: The advent of more complex ‘big humanities’
projects requires multiple and multi-disciplinary personnel
which in turn entails the organization of different workflows
and expectations: e.g. challenge of developing a
comprehensive or consortial approach, common definition of
project method etc.
INTEROPERABILITY
7. 1. A PARTNERSHIP
… a Deakin led consortium
• Cultural data providers (10) – project co-operators
• Humanities software developer (1) – project co-
developers
• eResearch organisations (2) – lead development
agencies
8. HUNI PARTNER DATASETS
AMHD
MAP
CAARP
Bonza
AFIRC
Circus Oz
AusStage
Media:
film, cinema, theatre, newspapers, magazines, advertis
ing, music, live performances
DAAO
AustLit
AWR
ADB
DoS
Biographical: artists, designers, writers, significant
people, scientists, Sydney demographics
EOAS
AUSTLANG
Mura
Indigenous languages
17. Welcome to the Cinema and Audiences Research Project (CAARP) database: An online encyclopaedia of
cinema-going in Australia.
Data
This site contains information on film screenings and venues in Australia.
430,137 screenings
10,256 films
1,978 cinemas
1,649 companies
From 1846 to now
18. • NeCTAR investment of $1.33M
• Partner contributions of $480,000
• Partner in-kind contributions amounting to >$1M
A FISCAL COLLABORATION
19. COMMUNITY BUILDING
• Collated user-stories (20)
• Online showcase events – next one is 4th September
2013
• Live link to the latest alpha prototype on huni.net.au;
feedback buttons
• Wider beta launch at eResearch Australasia in October
2013
• Stay up to date through our monthly Newsletter and
blog feed
• Follow us on twitter - @HuNIVL
20. Information design challenge to build an ontology and use
linked data and controlled vocabularies for data to be
aligned and related.
• Reading the data. Characteristics of the data determine
the ontological components selected and the major
“entities” (aka “access points”).
• Identified early as:
people, organisations, events, relationships, places, dates,
resources, and subjects.
• Components from ontologies already available and being
reused or kept in our sights: CIDOC-
CRM, FOAF, FRBR, FRBR-OO, BibFrame and PROV-O.
2. INTEGRATING MEANING
23. HUNI ONTOLOGY (all classes and
object properties)
cidoc:E41Appellation
cidoc:E49TimeAppellation
has subclass
cidoc:E44PlaceAppellation
has subclass
cidoc:E18PhysicalThing
cidoc:E24PhysicalManMadeThing
has subclass
cidoc:E19PhysicalObject
has subclass
frbr:F7Object
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)
frbr:F9Place
cidoc:P53hasCurrentOrFormerLocation (Domain>Range)
cidoc:P1isIdentifiedBy (Domain>Range)cidoc:E22Man-MadeObject
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)
cidoc:E52Time-Span
cidoc:E2TemporalEntity
has subclasscidoc:P4hasTimeSpan (Domain>Range)
cidoc:E4Period
has subclass
frbr:F22Self-Contained_Expression
frbr:F25Performance_Plan
has subclass
frbr:F26Recording
has subclass
frbr:F24Publication_Expression
has subclass
frbr:F15Complex_Work
frbr:F18Serial_Work
has subclass
cidoc:E21Person
frbr:F10Person
has subclass
cidoc:E67Birth
cidoc:P98iwasBorn (Domain>Range)
foaf:Person
has subclass
cidoc:E74Group
cidoc:P107iisCurrentOrFormerMemberOf (Domain>Range)
cidoc:E69Death
cidoc:P101idiedIn (Domain>Range)
cidoc:E7Activity
cidoc:P14iperformed (Domain>Range)
Thing
cidoc:E39Actor
has subclasscidoc:E15IdentifierAssignment
has subclass
huni:PrimaryTopic
has subclass
cidoc:E35Title
has subclass
cidoc:E71Man-MadeThing
has subclass
has subclass
cidoc:E53Place
has subclass
has subclass
huni:SKOS.Occupation
has subclass
has subclass
foaf:Group
has subclass
huni:SKOS.Role
has subclass
frbr:F6Concept
has subclass
frbr:F11Corporate_Body
has subclass
huni:SKOS.Collection
has subclass
cidoc:E42Identifier
has subclass
has subclass
frbr:F8Event
has subclass
huni:SKOS.Item
has subclass
has subclass
cidoc:E56Language
has subclass
has subclass
frbr:F13Identifier
has subclass
has subclass
cidoc:E55Type
has subclass
has subclassfrbr:F40Identifier_Assignment
has subclass
cidoc:P2hasType (Domain>Range)
cidoc:P11iparticipatedIn (Domain>Range)
has subclass
cidoc:P2HasType (Domain>Range)
has subclass has subclass
has subclass
has subclass
cidoc:E65Creation
has subclass
frbr:F31Performance
has subclasshas subclass
cidoc:E12Production
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)
cidoc:P1isIdentifiedBy (Domain>Range)
has subclass
huni:timeIsIdentifiedBy (Domain>Range)
cidoc:E5Event
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)cidoc:P1isIdentifiedBy (Domain>Range)
cidoc:P7tookPlaceAt (Domain>Range)cidoc:P1isIdentifiedBy (Domain>Range)
huni:hasOccupation (Domain>Range) huni:hasRole (Domain>Range)
cidoc:E48PlaceName
has subclass
frbr:F30Publication_Event
frbr:R24created (Domain>Range)frbr:F21Recording_Work
frbr:R23createdARealisationOf (Domain>Range)
frbr:F19Publication_Work
frbr:R24created (Domain>Range)
has subclass
cidoc:P1isIdentifiedBy (Domain>Range)cidoc:P1isIdentifiedBy (Domain>Range)
cidoc:P1isIdentifiedBy (Domain>Range)
huni:placeIsIdentifiedBy (Domain>Range)
frbr:F28Expression_Creation
has subclass cidoc:P108hasProduced (Domain>Range)
has subclassfrbr:F1Work
frbr:R19createdARealisationOf (Domain>Range)
frbr:F2Expression
frbr:R17created (Domain>Range)
frbr:F21Recording_Event
has subclass
cidoc:E28ConceptualObject
has subclass
has subclass
has subclass
cidoc:E89PropositionalObject
has subclass
frbr:F14Individual_Work
frbr:F17Aggregation_Work
has subclass
cidoc:P94hasCreated (Domain>Range)
frbr:f25Work_Conception
has subclass
cidoc:P102hasTitle (Domain>Range) huni:hasCollection (Domain>Range)
cidoc:P2hasType (Domain>Range)
has subclass
cidoc:P148hasComponent (Domain>Range)
cidoc:E73InformationObject
has subclass huni:hasItem (Domain>Range)
cidoc:P2HasType (Domain>Range) frbr:f16Container_Work
has subclass
has subclass
has subclass has subclass
frbr:F20Performance_Work
has subclasshas subclass has subclasscidoc:P72hasLanguage (Domain>Range) has subclass
cidoc:P2hasType (Domain>Range)cidoc:P2HasType (Domain>Range)
has subclass
frbr:R23createdARealisationOf (Domain>Range)
frbr:R24created (Domain>Range)
frbr:R24created (Domain>Range)
has subclass
cidoc:P102hasTitle (Domain>Range)
frbr:R12isRealisedIn (Domain>Range)
has subclass
has subclass
frbr:R16initiated (Domain>Range)
cidoc:P14iperformed (Domain>Range) has subclass
25. 3. HuNI DATA ARCHITECTURE
Data
integration
HuNI
side
Partner
side
Data harvest,
transform
and ingest
Solr Search Server
[HuNI Data]
RDF Triple Store
[HuNI Linked Data]
Data
analysis
and
mapping
HuNI Virtual Laboratory
Scholarly researcher workflow tasks Admin tasksPublic and citizen
researcher workflow tasks
Data
discovery
Data
analysis
Data
sharing
Analyse and annotate
collection
Export collection
Share collection and
analysis
Share search results
Corbicula
Registration and login
Profile management
History recording
Project management
Simple search
Advanced search
Save search results as
private collection
Refine / expand
collection
Simple search
Advanced search
Deep (SPARQL-based)
search
Data update
and
publish ADB DAAO CAARP AFIRC AusStage
26. A total of 28 Australian datasets are being harvested for integration into
HuNI
• Data gateway components, called HuNI Corbicula, deployed on the
NeCTAR Cloud to harvest the XML feed data and transforming it into
forms suitable for ingestion into two HuNI data aggregates: a Solr
search server [HuNI Data], and a Jena RDF Triple Store [HuNI Linked
Data]
DATA INTEGRATION
The harvesting process
requires:
• Live data feeds
deployed at the partner
sites to publish
updated partner data
as XML
Data
integration
HuNI
side
Partner
side
Data harvest,
transform
and ingest
Solr Search Server
[HuNI Data]
RDF Triple Store
[HuNI Linked Data]
Data
analysis
and
mapping
Corbicula
Data update
and
publish ADB DAAO CAARP AFIRC AusStage
28. TECHNOLOGY STACK
• front-end frameworks - AngularJS and Twitter
Bootstrap single page web app
• tools hosting framework - Open Social via Apache
Shindig
• back-end framework - SpringMVC via Roo.
• layer integration - RESTful web services
29. • Search the HuNI Data
• Save their search results as a
private collection
• Refine their collection through
additional searches
• Analyse and annotate their
collection with their own
assertions and commentary
• Export their collection for
further analysis
• Publish and share their
collection and research
RESEARCH ACTIVITIES
A researcher with a HuNI account will be able to:
HuNI Virtual Laboratory
Scholarly researcher workflow tasks Admin tasksPublic and citizen
researcher workflow tasks
Data
discovery
Data
analysis
Data
sharing
Analyse and annotate
collection
Export collection
Share collection and
analysis
Share search results
Registration and login
Profile management
History recording
Project management
Simple search
Advanced search
Save search results as
private collection
Refine / expand
collection
Simple search
Advanced search
Solr Search Server
[HuNI Data]
30. Scholarly researchers will also
be able to perform a “deep
search” of the graphs in RDF
Triple Store.
The large-scale aggregation of
Linked Data makes explicit the
relationships and connections
between related records across
all the partner
datasets, enabling the
researcher to construct more
complex semantic queries.
RESEARCH ACTIVITIES 2
HuNI Virtual Laboratory
Scholarly researcher workflow tasks Admin tasksPublic and citizen
researcher workflow tasks
Data
discovery
Data
analysis
Data
sharing
Registration and login
Profile management
History recording
Project management
Deep (SPARQL-based)
search
RDF Triple Store
[HuNI Linked Data]
39. 4. THE PROJECT
• project director/community liaison (20%)
• project manager (100%)
• technical coordinator (100%)
• information services coordinator (90%)
• community engagement (30%)
• communication coordinator (20%)
• administrative support (20%)
• software developer(s)
NeCTAR
Directorate
HuNI
Steering
Committee
Team HuNI
Technical
Working
Group
Expert
Advisory
Group
Expert Data
Group
42. HuNI: a virtual laboratory for the humanities
http://huni.net.au/@HuNIVL
Editor's Notes
Components of CIDOC-CRM, FOAF and FRBR-OO ontologies have been reused for the integration of the initial datasets. This is a means to encode people, their existence (birth and death events), their occupations and associations with organisations. More components have been added to record two further events, i.e. creation and production events, and to record works and expressions. Work is underway to plugin SKOS and structure vocabularies, using the data supplied (in EAC type schemas) to manage the range of terminology, e.g. recreational, vocational, professional and occupational. This draft is based on a portion of the data analysed and a "mud map" (based on an assessment of data available through web interfaces). See the draft as a line diagram. A view of the ontology generated in the tool Protege reveals FRBR-OO as an extension of CIDOC-CRM. Draft v0.3 using Initial DatasetsLimitations with using FOAF to handle personal names (culturally situated) have been found. The CIDOC component E41_Appellation and its subclasses will now be used, collections are being dealt with and further events are being added, e.g. E87_Curation_Activity to reflect actions of selection and collection development. Under discussion is: the inclusion of E90_Symbolic_Object to deal with citations (that are not feasible to strip apart and process but provide useful contextual information for an entity); the creation of "Floruit" as a time-related entity for E21_Person and E74_Group; categorising the datasets and collections as E89_Propositional_Object; and F3_Manifestation_Product_Type to deal with the disambiguation of portable and web formats of works.
This section of the HuNI ontology reveals the "joins" and class relationships, that reveal where the CIDOC-CRM and FRBR-OO ontologies align. The yellow-green bubbles record the CIDOC entities and the red bubbles record the FRBR entities. The bidirectional arrows indicate where there is a "sameAs" relationship, the unidirectional arrow indicate where there is a sub-class relationship.
The integration of partner data into HuNI requires two technical component:1. Live data feeds (at partner sites)Three technology options are available for the partners to publish their data as XML: jOAI, OAIcat and, for those who are not exposing their data via the OAI-PMH harvesting protocol, a custom-built solution that requires very little work to integrate at a provider’s site.We are not harvesting all the data – we are only harvesting the primary entity classes (and as much of the uniquely identifying information as possible for each class) that are common “touch points” across many of the partner data sites – people, places, events and objects. Therefore, the lowest common denominator for making the partner data harvestable is a flat XML file per class entity, together with the uniquely identifying information. For example, for the person class entity, uniquely identifying information will include first name, last name, date of birth/death, bio, occupation. 2. A data gateway component called CorbiculaTechnology is being deployed toharvest updated content from the partner XML data feeds and transform the data into forms suitable for ingestion into:A Solr search server: this aggregation of harvested XML records is referred to as ‘HuNI Data’ A Jena RDF Triple Store: this aggregation of stored RDF Graphs is referred to as ‘HuNI Linked Data’
Based on the data architecture as set out in the original RFP, there is a requirement to harvest, transform and ingest data each of the partner datasets into some sort of Linked Data store, and very early on in the technical decision making process, it was agreed that RDF (Resource Description Framework) – a metadata modeling specification - would be the lingua franca, and that all the technical components would be developed to work with this Linked Data specification.So we began by:Making some of the partner datasets harvestable to HuNI: by developing a harvest feed for those data providers who were technically able to publish their data in a standard export format/schema (EAC-CPF)Constructing the HuNI ontology and mapping partner data to this common data model. A number of standard cultural heritage ontologies were selected for examination because of their perceived close semantic fit to the nature and types of data in each of the 28 data sources. – CIDOC-CRM, FOAF, FRBR-OO, PROV-ODeploying a data gateway component – called Corbicula – on the NeCTAR Cloud, which is able to technically harvest and transforms the updated XML data from the partner feeds and ingest it into the RDF Triple Store. Once the mappings for a given data source are known, XSLT scripts are written to interpret the XML records and re-expresses (transforms) them as RDF graphs (essentially captures the relationship/link between records from all integrated data sets. But the integration into RDF has proven to be semantically complex and technically complex, because: The publishing format necessary to allow us to do the mappings is too high a technical barrier for most data custodians The data analysis and mapping to a common data model is proving time consuming and complexThe gateway component that harvests and transforms the data into RDF using XSLT has performance and memory issuesThe SPARQL-based search interface developments – where people can search and query the graphs – was proving too slowAs a result, after 10 months of development, only 6 partner data sources have completed their integration journey into the RDF Triple Store, and the search UI isn’t very performantSo back in May it was flagged that there is a as real project risk that we will not be able to fully transform all the partner data into Linked Data, and that only a small subset of partner datasets will be barely discoverable through the lab. This was a real problem, given that the main objective of HuNI is to provide a coSo the decision was made to exYour probably wondering – why have 2 data aggregates –why we mixed the data architectures – purely a project risk management decision – harvesting, mapping, transforming and ingesting into Linked Data is complex and time consuming, and there is a real danger that we won’t have a sufficient Linked Data layer in which to build the lab on – so in order to deliver some cross dataset search capability within the project timeframe, we introduced a new development strand which sees the accelerated harvesting and integration of data into the Solr aggregate So the decision has been made to continue populating the RDF store with partner data for the remainder of 2013, and work on UI in 2014To populate the Solr search server is easy, HuNI periodically harvests the updated XML records from the partner feeds, processes the XML content via a suitable transform, and submits the transformed XML data into the Solr search server. The transformation of partner XML records into HuNI Linked Data is complex and time consuming, and we’ve faced a number of technical issues, which isn’t surprising since we’re using a combination of largely unproven technologies, on the scale required for HuNI deployment First, the harvested data had to be cleaned and mapped to a core HuNI ontology. A range of cultural heritage ontologies were examined as the starting-point for building this core ontology framework. This has been an iterative process, determined by the nature of each data source and by the main types of data found in each source. The following standard ontologies are being aligned to create the HuNI Ontology:People and Organisations (using the CIDOC-CRM and FOAF ontologies) Items, Collections and Resources (using the PROV-O, CIDOC-CRM, FOAF and FRBR-OO ontologies) Events and Relations (using the PROV-O, CIDOC-CRM, FOAF and FRBR-OO ontologies) Place and Subject (using PROV-O, CIDOC-CRM, FOAF and FRBR-OO ontologies) Once the mappings to a common data model are known, the data needs to be technically transformed and ingested. This is made possible through the HuNI gateway component called Corbicula, which performs the following steps: Periodically harvests updated XML records from the source provider feedsUses XSLT to interpret the XML records and re-express (transform) them as RDF graphs.Stores the RDF graphs.The search feature needs to be based on the linked data, to take advantage of the semantic integration provided by the RDF aggregation
But of course this is a VL project and not a data integration project
Support the non-linear research methods practiced by humanities researchersHuNI is about inclusivity and not exclusivity – using 3rd party authentication for login - for the a community to form around HuNI, its user-base needs to extend beyond scholarly researchers. Also worth noting that any member of the general public interested in Australian culture can run a search across the related databases (the HuNI Data), and share their search results online – not just scholarly researchersThere are discovery limitations – whilst the context is given for each record found, what isn’t available are the known relationships between related records across the disparate data sources - so we’re currently working on a ‘Social Linked Data’ feature
Equipped with a full set of known facets and related data fields for each record type, researchers should be able to interact with, and construct complex queries of, the large-scale aggregation of Linked Data.
Link will be made available on huni.net.au soon
The lab is being designed to support the non-linear research methods practiced in the humanities and creative arts, and will support a workflow centred around discovery, analysis and sharing. As part of the discovery interface a researcher will be able to:Run a free text search across the aggregate and display their results Perform an advanced faceted browse of the aggregate by filtering their results by dataset and entity classes defined in the ontology: people, works, events, organisation, occupation/role, time, place, collections, language, objects. Narrow their search parameters at the start of their search by browsing for information within pre-defined access points. These are likely to be people, works and events since these entity classes are representative across all 28 data sources. Following the initial browse, the user can then filter their search results by dataset and the remaining entity classes.Run a SPARQL query to interrogate the underlying Linked Data The discovery interface is also going to enable serendipitous discovery (i.e. the ability to present information to users before they know what they want to search for):You might also be interested in... (based on the semantic relationships captured in the ontology)The notion of a generous interface is being included (based on some pre-defined daily query feeds), to give the researcher a sense of what is discoverable:On this day…Most popular searchesMost popular records The result sets will be displayed in a number of forms, with list being the default and map and timeline being optional. All search results will be displayed with hyperlinks that allow navigation to the source entity and will show the connections between records as per the ontology mappings
The LORE Tool (developed at UQ) will be made available in the lab where researchers will be able to:Display existing connections between relevant records held within their virtual collection, and Add further links between particular records, with commentary describing the relationship between them
Researchers will have the option to export their Virtual Collection as a .csv file so they can undertake further computational analysis outside of the HuNI lab and within their preferred tool environment.Whilst the lab will include a Tool Integration Framework specifying how third party tools can integrate within the lab and work with HuNI data, we recognize that tools come and go, and that researchers create their own relationship with their tools of choice. So offering an export function is crucial.
Researchers will have the option to share their virtual collection, and their analysis findings, via FB, twitter and email with other researchers
The development of HuNI is being managed as a projectHas a collaborative governance structure in place so that all key project decisions are only made as part of a consultative process Using Prince2 methodology in help manage the projectQuestion of consortial project management…Need to create best practice exemplars at the project management level…Staff in 4 states. Communication in skype or google hangout. Issues around discomfort with these communication technologies. Etc.