Presentation given by Rene Wiermer and Jeffrey van der Hoeven at ELAG2017 about automating rights decisions for digital content at the KB, National Library of the Netherlands
This document discusses several legal issues related to big data, including intellectual property rights, contract law and licensing, privacy and data protection, due process, liability, and jurisdictional problems. It notes that big data raises issues around rights clearing for copyrighted data, outdated exemptions, legal fragmentation across jurisdictions, enforceability of clickwrap licenses, definitions of personal data, ensuring transparency and auditing of decision-making logic, assigning liability in converged cloud services, and managing multiple applicable laws across jurisdictions. The document argues that harmonization of laws may help address some of the jurisdictional problems with applying legal frameworks to large-scale, multi-jurisdictional uses of big data.
A Network Centric Design For Relationship Based Rights Management (Tin180 Com)Tin180 VietNam
The document proposes a network-centric rights management framework called FIRM that uses "commpacts" (computational relationship objects) to manage digital relationships and enable rights management. It discusses how commpacts can bridge technical and social frameworks, provides examples of how relationships can be modeled and transactions authorized, and outlines goals of supporting extensible domains and usability through concepts like e-persons and forms. It also describes a RManage prototype that implements FIRM to experimentally manage digital contracts.
1) OpenSimulator is an open source platform for hosting virtual worlds that is compatible with Second Life but can also host alternative virtual worlds. It allows organizations to define their own terms of service.
2) Cloud computing provides scalable resources over the Internet like utilities. For this project, Amazon Web Services were used along with RightScale for infrastructure as a service.
3) OpenSimulator's cloud architecture uses multiple servers in a grid for larger worlds, dividing responsibilities among various servers like region, asset, and inventory servers.
KB Seminars: Working with Technology - Platforms; 10/13MDIF
This document provides an overview and agenda for a technology seminar discussing technology platforms and decision criteria. It will cover the purpose of platforms, the planning and decision making process, and do a comparison of major open source platforms. The document defines technology platforms and outlines various decision criteria to consider, including technical requirements, business factors like costs, and open source versus proprietary software pros and cons. Useful links are also provided.
Data Apps with the Lambda Architecture - with Real Work Examples on Merging B...Altan Khendup
The document discusses the Lambda architecture, which provides a common pattern for integrating real-time and batch processing systems. It describes the key components of Lambda - the batch layer, speed layer, and serving layer. The challenges of implementing Lambda are that it requires multiple systems and technologies to be coordinated. Real-world examples are needed to help practical application. The document also provides examples of medical and customer analytics use cases that could benefit from a Lambda approach.
The document describes the Dynamic On Demand Analysis Service (DODAS), which allows users to instantiate on-demand container-based clusters on any cloud provider with minimal effort. It provides batch systems, machine learning, and big data platforms as a service. The service architecture uses a PaaS orchestrator to deploy clusters on private and public clouds. Users can access deployed software applications through GUIs, REST APIs, or command line interfaces. The service offers configurable options for storage, compute, duration, and resources. Access is policy-based and funded through a market-driven payment model. Example use cases include opportunistic computing for CERN and a batch system for analyzing AMS satellite data. Documentation and tutorials are available online
A distributed system in its most simplest definition is a group of computers working together as to
appear as a single computer to the end-user. These machines have a shared state, operate
concurrently and can fail independently without affecting the whole system’s uptime.
This is in line with ever-growing technological expansion of the world, distributed systems are
becoming more and more widespread. Take a look at the increasing number of available
computer technologies/innovation around, this is sporadically increasing, and this result in
intense computational requirement.
Yeah, Moore’s law proposed more computing power by fitting more transistors (which
approximately doubles every two years) into a simple chip using cost-efficient approach - cool,
but over the past 5 years, there has been little deviation from this - ability to scale horizontally
and not just vertically alone.
This document discusses several legal issues related to big data, including intellectual property rights, contract law and licensing, privacy and data protection, due process, liability, and jurisdictional problems. It notes that big data raises issues around rights clearing for copyrighted data, outdated exemptions, legal fragmentation across jurisdictions, enforceability of clickwrap licenses, definitions of personal data, ensuring transparency and auditing of decision-making logic, assigning liability in converged cloud services, and managing multiple applicable laws across jurisdictions. The document argues that harmonization of laws may help address some of the jurisdictional problems with applying legal frameworks to large-scale, multi-jurisdictional uses of big data.
A Network Centric Design For Relationship Based Rights Management (Tin180 Com)Tin180 VietNam
The document proposes a network-centric rights management framework called FIRM that uses "commpacts" (computational relationship objects) to manage digital relationships and enable rights management. It discusses how commpacts can bridge technical and social frameworks, provides examples of how relationships can be modeled and transactions authorized, and outlines goals of supporting extensible domains and usability through concepts like e-persons and forms. It also describes a RManage prototype that implements FIRM to experimentally manage digital contracts.
1) OpenSimulator is an open source platform for hosting virtual worlds that is compatible with Second Life but can also host alternative virtual worlds. It allows organizations to define their own terms of service.
2) Cloud computing provides scalable resources over the Internet like utilities. For this project, Amazon Web Services were used along with RightScale for infrastructure as a service.
3) OpenSimulator's cloud architecture uses multiple servers in a grid for larger worlds, dividing responsibilities among various servers like region, asset, and inventory servers.
KB Seminars: Working with Technology - Platforms; 10/13MDIF
This document provides an overview and agenda for a technology seminar discussing technology platforms and decision criteria. It will cover the purpose of platforms, the planning and decision making process, and do a comparison of major open source platforms. The document defines technology platforms and outlines various decision criteria to consider, including technical requirements, business factors like costs, and open source versus proprietary software pros and cons. Useful links are also provided.
Data Apps with the Lambda Architecture - with Real Work Examples on Merging B...Altan Khendup
The document discusses the Lambda architecture, which provides a common pattern for integrating real-time and batch processing systems. It describes the key components of Lambda - the batch layer, speed layer, and serving layer. The challenges of implementing Lambda are that it requires multiple systems and technologies to be coordinated. Real-world examples are needed to help practical application. The document also provides examples of medical and customer analytics use cases that could benefit from a Lambda approach.
The document describes the Dynamic On Demand Analysis Service (DODAS), which allows users to instantiate on-demand container-based clusters on any cloud provider with minimal effort. It provides batch systems, machine learning, and big data platforms as a service. The service architecture uses a PaaS orchestrator to deploy clusters on private and public clouds. Users can access deployed software applications through GUIs, REST APIs, or command line interfaces. The service offers configurable options for storage, compute, duration, and resources. Access is policy-based and funded through a market-driven payment model. Example use cases include opportunistic computing for CERN and a batch system for analyzing AMS satellite data. Documentation and tutorials are available online
A distributed system in its most simplest definition is a group of computers working together as to
appear as a single computer to the end-user. These machines have a shared state, operate
concurrently and can fail independently without affecting the whole system’s uptime.
This is in line with ever-growing technological expansion of the world, distributed systems are
becoming more and more widespread. Take a look at the increasing number of available
computer technologies/innovation around, this is sporadically increasing, and this result in
intense computational requirement.
Yeah, Moore’s law proposed more computing power by fitting more transistors (which
approximately doubles every two years) into a simple chip using cost-efficient approach - cool,
but over the past 5 years, there has been little deviation from this - ability to scale horizontally
and not just vertically alone.
Presented Nov 18, 2015
Brief introduction to the types of cloud services, how libraries are using the cloud and the future of the cloud in libraries.
Includes presentation notes
The document summarizes the DALICC (Data Licenses Clearance Center) project. The project aims to develop a software framework that reduces the costs of clearing licenses for derivative works by providing tools to choose licenses, check compatibility, and resolve conflicts. It will represent licenses in RDF and use rules and semantics to reason about licenses and detect inconsistencies. The framework will include components for composing, annotating, and negotiating licenses through a license library and API. The goal is to increase productivity and reuse of data by easing license clearance.
Monitoring as an entry point for collaborationJulien Pivotto
This document summarizes a talk on using monitoring as an entry point for collaboration. It discusses using the Prometheus monitoring system to collect metrics and expose them using exporters. Grafana is then used to visualize the metrics and create dashboards focused on business metrics like requests, errors, and durations. These metrics provide observability across teams and enable alerting when business services are impacted.
The document discusses different content management systems (CMS), comparing features of open source and paid CMS like WordPress, Joomla, and Drupal. It outlines various factors to consider when choosing a CMS like installation, theming, plugins, ease of use, security, and scalability. The document also addresses stakeholder needs, project requirements, and questions to ask to determine the best CMS for a particular website.
Bitkom Cray presentation - on HPC affecting big data analytics in FSPhilip Filleul
High value analytics in FS are being enabled by Graph, machine learning and Spark technologies. To make these real at production scale HPC technologies are more appropriate than commodity clusters.
Distributed Systems: How to connect your real-time applicationsJaime Martin Losa
This document provides an overview of distributed systems and how to connect real-time applications using the Data Distribution Service (DDS) standard. It introduces DDS and its architecture, including topics, instances, keys, quality of service policies. It then demonstrates how to create a basic "hello world" publisher/subscriber example in both eProsima Fast RTPS and RTI Connext DDS middleware in 3 steps: defining the data type, generating code, and building/running the publisher and subscriber.
Cloud Busting: Understanding Cloud-based Digital ForensicsKerry Hazelton
Slide deck from the cloud forensics class taught by Tigran Terpandjian and me at BSides Charm 2018. To all who attended - thank you. To those who hung in there until the bitter end during the Cloud Forensic Challenge for those two RasPi 3B+ kits - you guys and gals really impressed us with your forensics skills! We hope to see all of you at the next event!
Bangalore Executive Seminar 2015: Case Study - Text Analysis on MongoDB for a...MongoDB
This document describes a case study of using text analytics on unstructured data from a multinational oil and gas company. The company needed a way to analyze communications data from email, chat and audio to help auditors identify deceitful transactions. Happiest Minds used MongoDB to store and query the semi-structured data. They developed a text analytics application using Java, Python and MongoDB that could consolidate the communications data, perform topic modeling using LDA, and generate reports to meet the company's needs. The solution helped auditors evaluate transactions faster and provided new insights into supplier relationships.
Apricot2017 Request tracing in distributed environmentHieu LE ☁
This document discusses logging and request tracing in distributed environments. It begins by introducing the context of distributed systems like cloud computing. It then reviews the current logging solution of ELK and Graylog and identifies pros and cons. Key requirements for tracing are outlined, including the need for end-to-end debugging. Approaches for workflow-centric tracing are surveyed, including explicit metadata propagation, schema-based, and black-box tracing. Examples of Magpie and Zipkin are provided. The presentation concludes with a demo of request tracing in OpenStack using OSProfiler.
This document provides an overview of big data analysis tools and methods presented by Ehsan Derakhshan of innfinision. It discusses what data and big data are, important questions about database selection, and several tools and solutions offered by innfinision including MongoDB, PyTables, Blosc, and Blaze. MongoDB is highlighted as a scalable and high performance document database. The advantages of these tools include optimized memory usage, rich queries, fast updates, and the ability to analyze and optimize queries.
The document discusses Software as a Service (SaaS) and the architectural considerations for developing SaaS applications. It covers the key aspects of SaaS including multi-tenancy, customization, scalability, and monetization options. The document also outlines some common architectural patterns for SaaS including using a shared database with customizable fields, meta-data services to manage customizations, and stateless and scalable application design.
Cloud computing allows users to access shared computing resources like servers, storage, databases, and applications over the internet. There are several types of cloud computing services including Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). Cloud computing provides benefits like reduced costs, increased storage, scalability, and mobility. However, concerns around security, data privacy, internet dependence, and availability levels remain limitations of cloud computing.
Big data is characterized by 3 V's - volume, velocity, and variety. It refers to large and complex datasets that are difficult to process using traditional database management tools. Key technologies to handle big data include distributed file systems, Apache Hadoop, data-intensive computing, and tools like MapReduce. Common tools used are infrastructure management tools like Chef and Puppet, monitoring tools like Nagios and Ganglia, and analytics platforms like Netezza and Greenplum.
Linking HPC to Data Management - EUDAT Summer School (Giuseppe Fiameni, CINECA)EUDAT
EUDAT and PRACE joined forces to help research communities gain access to high quality managed e-Infrastructures whose resources can be connected together to enable cross-utilization use cases and make them accessible without any technical barrier. The capability to couple data and compute resources together is considered one of the key factors to accelerate scientific innovation and advance research frontiers. The goal of this session was to present the EUDAT services, the results of the collaboration activity achieved so far and delivers a hands-on on how to write a Data Management Plan or DMP. The DMP is a useful instrument for researchers to reflect on and communicate about the way they will deal with their data. It prompts them to think about how they will generate, analyse and share data during their research project and afterwards.
Visit: https://www.eudat.eu/eudat-summer-school
The document discusses moving from a Data Ecosystem 1.0 to a Data Ecosystem 2.0. Data Ecosystem 1.0 had many compute frameworks and storage systems that were not co-located, leading to data silos and complexity. Data Ecosystem 2.0 proposes a virtual distributed file system (VDFS) that provides a unified namespace, API translation between systems, and intelligent caching to address these issues. Alluxio is presented as an example VDFS that provides benefits like global data access, accelerated analytics and machine learning, and reduced integration complexity across various industries. Real customer use cases demonstrate benefits of faster queries, improved productivity, and reduced costs.
Logging/Request Tracing in Distributed EnvironmentAPNIC
This document discusses logging and request tracing in distributed environments. It begins by introducing the context of distributed systems like cloud computing. It then reviews the current logging solution of ELK and Graylog and identifies pros and cons. Key requirements for tracing are outlined, including the need for end-to-end debugging. Approaches for workflow-centric tracing are surveyed, including explicit metadata propagation, schema-based, and black-box tracing. Examples of Magpie and Zipkin are provided. The presentation concludes with a demo of request tracing in OpenStack using OSProfiler.
Simple Workload and Application Portability (SWAP) for Cloud ComputingSam Johnston
The document proposes a new protocol called Simple Workload & Application Portability (SWAP) that aims to enable portability of workloads between cloud providers in a simple way, similar to how SMTP enabled email interoperability. SWAP would only address the minimum functionality required for workload portability without constraining providers' functionality or standardizing management interfaces. It leverages existing standards like HTTP and focuses only on allowing the programmatic transfer of workloads in open formats between servers, including public/private clouds and developer workstations.
Theo van Veen - Verrijkingen in Delpher: genereren, gebruiken en corrigerenKBNLResearch
Workshop given by Theo van Veen on enrichments in the platform Delpher (delpher.nl) during the launch event of the new website of the KB Lab on 11 April 2017.
Rene Voorburg - Using KB APIs to collect dataKBNLResearch
Part two of the workshop given by Rene Voorburg and Juliette Lonij at the launch event of the new website of the KB Lab on 11 April 2017. Part one of the workshop is available here; https://www.slideshare.net/KBNLResearch/rene-voorburg-and-juliette-lonij-search-download-and-analyse-kb-data-sets-a-practical-introduction
Presented Nov 18, 2015
Brief introduction to the types of cloud services, how libraries are using the cloud and the future of the cloud in libraries.
Includes presentation notes
The document summarizes the DALICC (Data Licenses Clearance Center) project. The project aims to develop a software framework that reduces the costs of clearing licenses for derivative works by providing tools to choose licenses, check compatibility, and resolve conflicts. It will represent licenses in RDF and use rules and semantics to reason about licenses and detect inconsistencies. The framework will include components for composing, annotating, and negotiating licenses through a license library and API. The goal is to increase productivity and reuse of data by easing license clearance.
Monitoring as an entry point for collaborationJulien Pivotto
This document summarizes a talk on using monitoring as an entry point for collaboration. It discusses using the Prometheus monitoring system to collect metrics and expose them using exporters. Grafana is then used to visualize the metrics and create dashboards focused on business metrics like requests, errors, and durations. These metrics provide observability across teams and enable alerting when business services are impacted.
The document discusses different content management systems (CMS), comparing features of open source and paid CMS like WordPress, Joomla, and Drupal. It outlines various factors to consider when choosing a CMS like installation, theming, plugins, ease of use, security, and scalability. The document also addresses stakeholder needs, project requirements, and questions to ask to determine the best CMS for a particular website.
Bitkom Cray presentation - on HPC affecting big data analytics in FSPhilip Filleul
High value analytics in FS are being enabled by Graph, machine learning and Spark technologies. To make these real at production scale HPC technologies are more appropriate than commodity clusters.
Distributed Systems: How to connect your real-time applicationsJaime Martin Losa
This document provides an overview of distributed systems and how to connect real-time applications using the Data Distribution Service (DDS) standard. It introduces DDS and its architecture, including topics, instances, keys, quality of service policies. It then demonstrates how to create a basic "hello world" publisher/subscriber example in both eProsima Fast RTPS and RTI Connext DDS middleware in 3 steps: defining the data type, generating code, and building/running the publisher and subscriber.
Cloud Busting: Understanding Cloud-based Digital ForensicsKerry Hazelton
Slide deck from the cloud forensics class taught by Tigran Terpandjian and me at BSides Charm 2018. To all who attended - thank you. To those who hung in there until the bitter end during the Cloud Forensic Challenge for those two RasPi 3B+ kits - you guys and gals really impressed us with your forensics skills! We hope to see all of you at the next event!
Bangalore Executive Seminar 2015: Case Study - Text Analysis on MongoDB for a...MongoDB
This document describes a case study of using text analytics on unstructured data from a multinational oil and gas company. The company needed a way to analyze communications data from email, chat and audio to help auditors identify deceitful transactions. Happiest Minds used MongoDB to store and query the semi-structured data. They developed a text analytics application using Java, Python and MongoDB that could consolidate the communications data, perform topic modeling using LDA, and generate reports to meet the company's needs. The solution helped auditors evaluate transactions faster and provided new insights into supplier relationships.
Apricot2017 Request tracing in distributed environmentHieu LE ☁
This document discusses logging and request tracing in distributed environments. It begins by introducing the context of distributed systems like cloud computing. It then reviews the current logging solution of ELK and Graylog and identifies pros and cons. Key requirements for tracing are outlined, including the need for end-to-end debugging. Approaches for workflow-centric tracing are surveyed, including explicit metadata propagation, schema-based, and black-box tracing. Examples of Magpie and Zipkin are provided. The presentation concludes with a demo of request tracing in OpenStack using OSProfiler.
This document provides an overview of big data analysis tools and methods presented by Ehsan Derakhshan of innfinision. It discusses what data and big data are, important questions about database selection, and several tools and solutions offered by innfinision including MongoDB, PyTables, Blosc, and Blaze. MongoDB is highlighted as a scalable and high performance document database. The advantages of these tools include optimized memory usage, rich queries, fast updates, and the ability to analyze and optimize queries.
The document discusses Software as a Service (SaaS) and the architectural considerations for developing SaaS applications. It covers the key aspects of SaaS including multi-tenancy, customization, scalability, and monetization options. The document also outlines some common architectural patterns for SaaS including using a shared database with customizable fields, meta-data services to manage customizations, and stateless and scalable application design.
Cloud computing allows users to access shared computing resources like servers, storage, databases, and applications over the internet. There are several types of cloud computing services including Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). Cloud computing provides benefits like reduced costs, increased storage, scalability, and mobility. However, concerns around security, data privacy, internet dependence, and availability levels remain limitations of cloud computing.
Big data is characterized by 3 V's - volume, velocity, and variety. It refers to large and complex datasets that are difficult to process using traditional database management tools. Key technologies to handle big data include distributed file systems, Apache Hadoop, data-intensive computing, and tools like MapReduce. Common tools used are infrastructure management tools like Chef and Puppet, monitoring tools like Nagios and Ganglia, and analytics platforms like Netezza and Greenplum.
Linking HPC to Data Management - EUDAT Summer School (Giuseppe Fiameni, CINECA)EUDAT
EUDAT and PRACE joined forces to help research communities gain access to high quality managed e-Infrastructures whose resources can be connected together to enable cross-utilization use cases and make them accessible without any technical barrier. The capability to couple data and compute resources together is considered one of the key factors to accelerate scientific innovation and advance research frontiers. The goal of this session was to present the EUDAT services, the results of the collaboration activity achieved so far and delivers a hands-on on how to write a Data Management Plan or DMP. The DMP is a useful instrument for researchers to reflect on and communicate about the way they will deal with their data. It prompts them to think about how they will generate, analyse and share data during their research project and afterwards.
Visit: https://www.eudat.eu/eudat-summer-school
The document discusses moving from a Data Ecosystem 1.0 to a Data Ecosystem 2.0. Data Ecosystem 1.0 had many compute frameworks and storage systems that were not co-located, leading to data silos and complexity. Data Ecosystem 2.0 proposes a virtual distributed file system (VDFS) that provides a unified namespace, API translation between systems, and intelligent caching to address these issues. Alluxio is presented as an example VDFS that provides benefits like global data access, accelerated analytics and machine learning, and reduced integration complexity across various industries. Real customer use cases demonstrate benefits of faster queries, improved productivity, and reduced costs.
Logging/Request Tracing in Distributed EnvironmentAPNIC
This document discusses logging and request tracing in distributed environments. It begins by introducing the context of distributed systems like cloud computing. It then reviews the current logging solution of ELK and Graylog and identifies pros and cons. Key requirements for tracing are outlined, including the need for end-to-end debugging. Approaches for workflow-centric tracing are surveyed, including explicit metadata propagation, schema-based, and black-box tracing. Examples of Magpie and Zipkin are provided. The presentation concludes with a demo of request tracing in OpenStack using OSProfiler.
Simple Workload and Application Portability (SWAP) for Cloud ComputingSam Johnston
The document proposes a new protocol called Simple Workload & Application Portability (SWAP) that aims to enable portability of workloads between cloud providers in a simple way, similar to how SMTP enabled email interoperability. SWAP would only address the minimum functionality required for workload portability without constraining providers' functionality or standardizing management interfaces. It leverages existing standards like HTTP and focuses only on allowing the programmatic transfer of workloads in open formats between servers, including public/private clouds and developer workstations.
Similar to Automating rights decision elag 2017 (20)
Theo van Veen - Verrijkingen in Delpher: genereren, gebruiken en corrigerenKBNLResearch
Workshop given by Theo van Veen on enrichments in the platform Delpher (delpher.nl) during the launch event of the new website of the KB Lab on 11 April 2017.
Rene Voorburg - Using KB APIs to collect dataKBNLResearch
Part two of the workshop given by Rene Voorburg and Juliette Lonij at the launch event of the new website of the KB Lab on 11 April 2017. Part one of the workshop is available here; https://www.slideshare.net/KBNLResearch/rene-voorburg-and-juliette-lonij-search-download-and-analyse-kb-data-sets-a-practical-introduction
Lightning talk of Lotte Wilms at the launch of the new website of the KB Lab on 11 April 2017 on the plans of the KB Lab, which is available at lab.kb.nl.
Lightning talk of dr. Martijn Kleppe at the launch of the new website of the KB Lab on 11 April 2017 on the KBK-1M dataset, which is available at http://lab.kb.nl/dataset/kbk-1m
Lightning talk by dr. Pim Huijnen at the launch of the new website of the KB Lab. Tools are available at http://lab.kb.nl/tool/keyword-generator and http://lab.kb.nl/tool/dictionary-viewer
Frank Harbers - Automatic genre classification of historical newspaper articles KBNLResearch
Lightning talk presentation of dr. Frank Harbers at the launch of the new website of the KB Lab on 11 April 2017 on the Genre classifier tool, presented by Juliette Lonij and Martijn Kleppe. The tool is available at http://lab.kb.nl/tool/genre-classifier
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
ESA/ACT Science Coffee: Diego Blas - Gravitational wave detection with orbita...Advanced-Concepts-Team
Presentation in the Science Coffee of the Advanced Concepts Team of the European Space Agency on the 07.06.2024.
Speaker: Diego Blas (IFAE/ICREA)
Title: Gravitational wave detection with orbital motion of Moon and artificial
Abstract:
In this talk I will describe some recent ideas to find gravitational waves from supermassive black holes or of primordial origin by studying their secular effect on the orbital motion of the Moon or satellites that are laser ranged.
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...Sérgio Sacani
Context. With a mass exceeding several 104 M⊙ and a rich and dense population of massive stars, supermassive young star clusters
represent the most massive star-forming environment that is dominated by the feedback from massive stars and gravitational interactions
among stars.
Aims. In this paper we present the Extended Westerlund 1 and 2 Open Clusters Survey (EWOCS) project, which aims to investigate
the influence of the starburst environment on the formation of stars and planets, and on the evolution of both low and high mass stars.
The primary targets of this project are Westerlund 1 and 2, the closest supermassive star clusters to the Sun.
Methods. The project is based primarily on recent observations conducted with the Chandra and JWST observatories. Specifically,
the Chandra survey of Westerlund 1 consists of 36 new ACIS-I observations, nearly co-pointed, for a total exposure time of 1 Msec.
Additionally, we included 8 archival Chandra/ACIS-S observations. This paper presents the resulting catalog of X-ray sources within
and around Westerlund 1. Sources were detected by combining various existing methods, and photon extraction and source validation
were carried out using the ACIS-Extract software.
Results. The EWOCS X-ray catalog comprises 5963 validated sources out of the 9420 initially provided to ACIS-Extract, reaching a
photon flux threshold of approximately 2 × 10−8 photons cm−2
s
−1
. The X-ray sources exhibit a highly concentrated spatial distribution,
with 1075 sources located within the central 1 arcmin. We have successfully detected X-ray emissions from 126 out of the 166 known
massive stars of the cluster, and we have collected over 71 000 photons from the magnetar CXO J164710.20-455217.
(June 12, 2024) Webinar: Development of PET theranostics targeting the molecu...Scintica Instrumentation
Targeting Hsp90 and its pathogen Orthologs with Tethered Inhibitors as a Diagnostic and Therapeutic Strategy for cancer and infectious diseases with Dr. Timothy Haystead.
Current Ms word generated power point presentation covers major details about the micronuclei test. It's significance and assays to conduct it. It is used to detect the micronuclei formation inside the cells of nearly every multicellular organism. It's formation takes place during chromosomal sepration at metaphase.
When I was asked to give a companion lecture in support of ‘The Philosophy of Science’ (https://shorturl.at/4pUXz) I decided not to walk through the detail of the many methodologies in order of use. Instead, I chose to employ a long standing, and ongoing, scientific development as an exemplar. And so, I chose the ever evolving story of Thermodynamics as a scientific investigation at its best.
Conducted over a period of >200 years, Thermodynamics R&D, and application, benefitted from the highest levels of professionalism, collaboration, and technical thoroughness. New layers of application, methodology, and practice were made possible by the progressive advance of technology. In turn, this has seen measurement and modelling accuracy continually improved at a micro and macro level.
Perhaps most importantly, Thermodynamics rapidly became a primary tool in the advance of applied science/engineering/technology, spanning micro-tech, to aerospace and cosmology. I can think of no better a story to illustrate the breadth of scientific methodologies and applications at their best.
Mending Clothing to Support Sustainable Fashion_CIMaR 2024.pdfSelcen Ozturkcan
Ozturkcan, S., Berndt, A., & Angelakis, A. (2024). Mending clothing to support sustainable fashion. Presented at the 31st Annual Conference by the Consortium for International Marketing Research (CIMaR), 10-13 Jun 2024, University of Gävle, Sweden.
The cost of acquiring information by natural selectionCarl Bergstrom
This is a short talk that I gave at the Banff International Research Station workshop on Modeling and Theory in Population Biology. The idea is to try to understand how the burden of natural selection relates to the amount of information that selection puts into the genome.
It's based on the first part of this research paper:
The cost of information acquisition by natural selection
Ryan Seamus McGee, Olivia Kosterlitz, Artem Kaznatcheev, Benjamin Kerr, Carl T. Bergstrom
bioRxiv 2022.07.02.498577; doi: https://doi.org/10.1101/2022.07.02.498577
2. The dream: In reality:
Open access to
everything
for
everybody!
Limited access
due to
copyright
&
contracts
3. Examples of restrictions (1)
1600 1930 1945 1980 2017
open closed
1400 1900 2017
open restricted
1995
Time ->
digitized
newspapers
digitized
books
no
download
4. Examples of restrictions (2)
Publisher AReading room only
Journal titels ->
open API key account
datasets
Scientific articles
Publisher B
Publisher Z
12. Needs 1: more information to the end user
- How do I get access ?
- What can I do with it ?
Improve UX with standardization of rights decisions
13. Needs 2: One system for multiple applications
- Several websites: Delpher, Geheugen van Nederland, Staten
Generaal Digitaal
- Several API’s: URN-Resolver, OAI-PMH, Search services …
Centralize access decisions for better compliance, management
and reporting
One change = immediately visible in each application
14. Needs 3: reducing our digitization backlog
- We have a lot of digital content that requires certain restrictions
- How can we make this accessible to anybody who is allowed to
see it ?
- We had an “on/off” infrastructure for most of our content
- Either accessible for everybody or not at all
- Not flexible enough, blocked workflows
Automation of rights decisions based on
- Metadata (Publication date, authors, publisher, type of
material..)
- Location (e.g. reading room)
- Type of user (e.g. researcher)
15.
16. Simple approach: extra metadata field ?
- For example
- <rights> FREE|RESTRICTED|CLOSED|... </rights>
- <license> CC0|CustomContract|... </license>
- Make decision based on the value of that field
- Works probably fine in a lot scenarios
- But:
- Does not scale with variation depending on context
- “Free for users of type researcher and visitors to the reading room, but not outside
of it”
- Needs maintenance over time
-Missing: why was this decision made ?
17. Instead: policies as code
- Policy: formalized set of rules regarding a collection of objects
- Decided at runtime -> decisions can change over time
- Follows general lines of thought of the organization: legal
obligations, contracts with publishers, management decisions
19. Still simple policy
Role-based access (from API-key, username/password auth…)
if (context.roles.contains("DS_METADATA_DTS"))
return Decision.permit();
Access based on publication date
static GregorianCalendar metadataFreeDate=new GregorianCalendar(1940,Calendar.JANUARY,1);
if (attributes.getMetadata().getPublicationDate()?.before(metadataFreeDate.getTime())) {
return Decision.permit();
}
Fallback
return Decision.denied();
20. Example: Books
Check for location
if (context.location.equals("READING_ROOM")) {
...
}
Demand measures to prevent downloads from frontend
if (attributes.listContainsValue("boeken-leeszaal-kopieerbeveiliging", "ppn",
attributes.getMetadata().getPpn()) ) {
return Decision.permit(new Obligation("DoNotDownload"),usageRights);
}
Check for death dates of all contributors
if (DateChecks.allAuthorsDeadLongerThan(attributes.getMetadata(),authorDeathDateLimit)) {
return Decision.permit(usageRights);
}
21. Decisions
Input: Identifier, Metadata, Location, Authorization
End result of a policy decisions:
- PERMIT
- DENIED
- NOT APPLICABLE
additional attributes:
- obligations: things the endpoint has to enforce
- advices: things the endpoint might need to improve UX
Ex: PERMIT (obligation:”DoNotDownload”, advice:”OnlyInReadingRoom”)
22. Diagram by David Brossard under a CC-BY 3.0 license
Enforce
Decide
Administer Metadata
Context
24. Architecture: XACML (sort of)
- Attribute Based Access Control (ABAC)
- Follows XACML reference architecture
- … but not the language (cumbersome, slow and restricted)
25. Technology
- Write the policies in an embedded scripting language (Groovy)
- Fast (in comparison to XACML language implementations)
- Able to be adopted/managed outside of core development team
- still: reuse of existing development toolchain
- Automated testing !
- Deployed as central REST service
- Serves multiple applications
28. Limitations
- Search filtering on access: combination with dynamic decisions
- Which objects am I allowed to use ?
- Export of access information to other systems (e.g. WorldCat)
Possible mitigations
- Compromises on dynamic decisions (short term)
- Move from slow ETL to event-based architectures (longer term)
29. Current status & results
- Stepwise in production since Mid 2016
- New objects are becoming available
- Copyright claims are easier to handle
- Clearer insight into current status of collection
- Better insight into needs for partnership contracts
- Impulses for better metadata storage/access infrastructure
175M requests per month
+/- 6 million a day
60+ million pages
under control by
access management
32. About
- Managing digital collections with multiple licenses and access
policies
- Technical choices that fit our organisational needs
Not about
- DRM and copy protection
- Usage of closed proprietary systems
33. Motivation
- As a public service organisation we want: access as far as
possible
- Limit of possibilities
- Licenses
- Contractual obligations
- Governmental and organisational policies
- Copyright status
- A simple yes or no is not always enough; we need
- a clear guideline for the user: what can I do with it and how do I get
access ?
- automation of management: we want to be able to scale and still be
compliant
34. Crossing the domains: communication
- Define your terms: Collection, policy, decision … make sure to
communicate them clearly
- Make sure contracts and managerial decisions can be translated to
the technical reality.
- Offer protection and guarantee options for future contracts
- Make compliance easier through monitoring + reporting
- Use of examples + flow diagrams
36. Our problems
- Multiple applications give access to collections
- ideally centralised decision making and reporting
- Decisions depend on context: user, location, time
- Flexible to allow for individual interventions
- Clearer insight necessary why things are hidden away