The document summarizes the role of Digital.CSIC in building a Current Research Information System (CRIS) at the Spanish National Research Council (CSIC). Digital.CSIC serves to monitor CSIC research outputs and support institutional research evaluation. It contains nearly 40,000 items deposited primarily through manual input. Efforts are underway to integrate Digital.CSIC with the ConCiencia system and automate deposits through APIs and batch imports from CSIC websites and databases. However, challenges remain around metadata schemas, author identities, and synchronizing updates between the two systems.
This document contains notes from a meeting between real estate agent Mike Rosenhahn and clients Steven and Lynn Denny. The goals of the meeting were to get to know the clients better, answer their questions, ask important questions, and determine if Mike is the right agent for them. Mike asked questions about the clients' current home and neighborhood as well as their reasons for moving and timelines. He provided information about his experience, approach, and marketing strategy. Mike emphasized service, knowledge, and integrity in working with clients.
The document discusses the rise of big data and NoSQL databases. It notes that organizations are drowning in large amounts of data from various sources like user-generated content. However, traditional relational databases struggle to handle this type and volume of semi-structured data in a distributed, scalable manner. This has led to the emergence of NoSQL databases that are more flexible and better suited for the distributed, large-scale requirements of big data.
Albert Simard - Mobilizing Knowledge: Acquisition, Analysis, and Action
Presentation at the Canadian Knowledge Mobilization Forum 2012, Ottawa, Ontario, http://www.kmbforum2012.org/
High speed PCB design faces several challenges: ensuring system timing, maintaining waveform integrity, avoiding crosstalk, and controlling power/ground stability and EMI. Cadence addresses these with a PCB design flow including functional verification, exploration, floorplanning, high-speed layout, and simulation using tools like SPECCTRAQuest and Allegro. The full flow supports both digital and analog design from IC to board.
1) QualysGuard Suite 7.0 includes updates and new features to QualysGuard's enterprise, SMB, and freemium security and compliance applications.
2) The update includes a new virtual scanner that can run on virtualized platforms, improved private cloud offerings including a virtual private SOC, and new asset management capabilities.
3) Qualys is also releasing new free services including external vulnerability scanning for websites, a browser security audit tool, and an SSL validation service to help companies assess their security posture without purchase.
Hadoop World 2011: Lily: Smart Data at Scale, Made EasyCloudera, Inc.
Lily is a repository made for the age of Data, and combines CDH, HBase and Solr in a powerful, high-level, developer-friendly backing store for content-centric application with ambition to scale. In this session, we highlight why we choose HBase as the foundation for Lily, and how Lily will allow users to not only store, index and search vast quantities of data, but also to track audience behaviour and generate recommendations, all in real-time.
The document summarizes an open source monitoring conference presented by the Icinga team. It provides an agenda for the conference that includes introductions to Icinga tools and platforms, comparisons to Nagios, architecture details, new features in Icinga's core, classic UI, documentation and web modules, and a live demo. It also discusses Icinga's team members, statistics on downloads and members, and a roadmap for future development.
This document contains notes from a meeting between real estate agent Mike Rosenhahn and clients Steven and Lynn Denny. The goals of the meeting were to get to know the clients better, answer their questions, ask important questions, and determine if Mike is the right agent for them. Mike asked questions about the clients' current home and neighborhood as well as their reasons for moving and timelines. He provided information about his experience, approach, and marketing strategy. Mike emphasized service, knowledge, and integrity in working with clients.
The document discusses the rise of big data and NoSQL databases. It notes that organizations are drowning in large amounts of data from various sources like user-generated content. However, traditional relational databases struggle to handle this type and volume of semi-structured data in a distributed, scalable manner. This has led to the emergence of NoSQL databases that are more flexible and better suited for the distributed, large-scale requirements of big data.
Albert Simard - Mobilizing Knowledge: Acquisition, Analysis, and Action
Presentation at the Canadian Knowledge Mobilization Forum 2012, Ottawa, Ontario, http://www.kmbforum2012.org/
High speed PCB design faces several challenges: ensuring system timing, maintaining waveform integrity, avoiding crosstalk, and controlling power/ground stability and EMI. Cadence addresses these with a PCB design flow including functional verification, exploration, floorplanning, high-speed layout, and simulation using tools like SPECCTRAQuest and Allegro. The full flow supports both digital and analog design from IC to board.
1) QualysGuard Suite 7.0 includes updates and new features to QualysGuard's enterprise, SMB, and freemium security and compliance applications.
2) The update includes a new virtual scanner that can run on virtualized platforms, improved private cloud offerings including a virtual private SOC, and new asset management capabilities.
3) Qualys is also releasing new free services including external vulnerability scanning for websites, a browser security audit tool, and an SSL validation service to help companies assess their security posture without purchase.
Hadoop World 2011: Lily: Smart Data at Scale, Made EasyCloudera, Inc.
Lily is a repository made for the age of Data, and combines CDH, HBase and Solr in a powerful, high-level, developer-friendly backing store for content-centric application with ambition to scale. In this session, we highlight why we choose HBase as the foundation for Lily, and how Lily will allow users to not only store, index and search vast quantities of data, but also to track audience behaviour and generate recommendations, all in real-time.
The document summarizes an open source monitoring conference presented by the Icinga team. It provides an agenda for the conference that includes introductions to Icinga tools and platforms, comparisons to Nagios, architecture details, new features in Icinga's core, classic UI, documentation and web modules, and a live demo. It also discusses Icinga's team members, statistics on downloads and members, and a roadmap for future development.
This document discusses the rapid growth of digital data and the challenges of analyzing large, unstructured datasets. It notes that in just one week in 2000, the Sloan Digital Sky Survey collected more data than had been collected in all of astronomy previously. Today, the Large Hadron Collider generates 40 terabytes per second and Twitter generates over 1 terabyte of tweets daily. By 2013, annual internet traffic was predicted to reach 667 exabytes. Hadoop provides a framework to analyze these vast and diverse datasets by distributing processing across commodity clusters close to where the data is stored.
The document discusses predictions for the future of web content management (WCM). It predicts that the future will be open, with open standards like JCR and CMIS driving ubiquity. It will be cloud-based, with hardware resources available on demand. WCM systems will be hybrid, using on-premises and cloud resources. The future will also be business-oriented, aligning WCM with business goals and stakeholders. Context and personalization will be important, driving user experiences. Systems will need to be agile to adapt to changing markets and data. Finally, the future will be mobile-centric, with built-in support for multiple channels and device detection.
Cms forum, future of Web Content Managementguest88136a
This document discusses predictions for the future of web content management (WCM). It makes 6 predictions: 1) The future is open with open standards like JCR and CMIS and open source software. 2) The future is cloudy with more content hosted in the cloud to reduce the need for on-premise hardware. 3) The future is hybrid with content hosted both on-premise and in the cloud. 4) The future is agile to adapt to changing markets and content that is likely to change. 5) The future is context-centric to better understand users and tailor experiences. 6) The future is mobile as more content is consumed on mobile devices.
This document discusses the new features and capabilities of QlikView 11, a business intelligence and data visualization platform. It highlights five main themes: 1) Social Business Discovery which enables collaboration and sharing of insights, 2) Mobile Business Discovery with optimized experiences for tablets and smartphones, 3) Comparative Analysis for comparing different data groupings, 4) Rapid Analytic App Platform with improved development and extensibility, and 5) Enterprise Platform with performance, administration, and security improvements. The document provides examples of how organizations are using QlikView to empower business users to access, analyze, and make decisions from data on their own.
This document provides an overview of NoSQL and Hadoop technologies. It discusses the trends driving these technologies like increasing data size, connectivity of data, semi-structured data, and decoupled service architectures. It introduces concepts from academic research like Amazon Dynamo, Google BigTable, and Brewer's CAP theorem. Specific technologies are explained like Hadoop for processing large datasets using MapReduce on the Hadoop Distributed File System.
The document describes a quality attribute workshop to help define and elaborate on non-functional requirements. The workshop involves building a quality attribute tree to characterize requirements, generating specific quality attribute scenarios, and mapping scenarios to architecture tactics. Quality attributes are refined from high-level goals into detailed scenarios. Scenarios are prioritized based on importance and difficulty. Example generic and specific scenarios are provided for attributes like performance and modifiability. Architecture tactics that could satisfy scenarios are also summarized. The goal is to systematically define and select the most important quality requirements to inform architectural decisions.
Cloud Computing through FCAPS Managed Services in a Virtualized Data Centervsarathy
This document discusses driving cloud computing through managed services in a virtualized data center. It introduces cloud computing and defines it as more than just an on-demand XaaS stack. The cloud must address issues like massive scalability, reliability, availability, performance optimization, and security. It also discusses how the cloud can help reduce rising data center complexity and costs which are getting out of control.
IC-SDV 2018: Harald Jenny (CENTREDOC) When Artificial Intelligence Joins Inte...Dr. Haxel Consult
Three years ago, we thought this would be impossible to accomplish. Today, it is a reality : Artificial Intelligence gains more and more importance in the value chain of Intellectual Property. The « red button » allowing to obtain, with very limited interaction from human intelligence, a full technology landscape while starting from a simple technical fact sheet, is at hand. Assessment of the progress made so far.
Sirris innovate2011 - Lily, Smart Data at scale made easy, Steven Noels, Oute...Sirris
Data growth is rapidly surpassing Moore's Law, as data sets are growing increasingly large, hence deriving insights from these large data sets is becoming more and more complex. Lily, a software product made by Outerthought, allows you to store, index and search vast quantities of data. In the next few years, successful business models will be based on monetization of data. Steven Noels will highlight the raison d'être of Lily, discussing challenges that every data-intensive organisation encounters.
Part 2 OCLC Strategic Presentation Bruce Crocco ACURIL 2011Antonio Alba
OCLC is building web-scale capabilities with libraries by focusing on data, community, and infrastructure. It is developing an open and extensible platform built on WorldCat that will provide responsive services at massive scale. This platform will allow libraries to innovate, expose and share applications, and benefit from others' innovations. Currently, OCLC provides global exposure through WorldCat.org, discovery through WorldCat Local, and circulation and acquisitions capabilities. It is also developing knowledge base management, license management, and an open platform for third-party development. OCLC aims to collaboratively build web-scale services with libraries to reduce costs and workflows through its cooperative approach.
Ordex Presentation at Nationaal Congres Open Data Eindhoven 20 april 2012Tom Zeppenfeldt IEC MSc
Presentation by Tom Zeppenfeldt from Ophileon on the Open Reporting Data Exchange, a platform to share open data which is integrated with a monitoring and reporting tool.
Ordex serves as a source for open data for visualization
In this age of Big Data, data volumes grow exceedingly larger while the technical problems and business scenarios become more complex. Compounding these complexities, data consumers are demanding faster analysis to common business questions asked of their Big Data. This session provides concrete examples of how to address this challenge. We will highlight the use of Big Data technologies—including Hadoop and Hive —with classic BI systems such as SQL Server Analysis Services.
Session takeaways:
• Understand the architectural components surrounding Hadoop, Hive, Classic BI, and the Tier-1 BI ecosystem
• Get strategies for addressing the technical issues when working with extremely large cubes
• See how to address the technical issues when working with Big Data systems from the DBA perspective
Enterprise GIS Implementation for Public Infrastructure and Integration with ...Michael Baker Jr., Inc.
The document describes how Prince William County Service Authority implemented an enterprise GIS system to integrate its water and wastewater utility data and operations. Key aspects included:
1) Migrating from an Autodesk MapGuide system to an ArcGIS Server-based system with Oracle Spatial database to serve authoritative utility data to web applications.
2) Designing a GIS web portal for searching and viewing utility asset data, work orders, customer information, and consumption data.
3) Integrating the GIS with work order management, customer information, and SCADA systems to improve operations and customer service functions like service requests and outage notifications.
4) Using the integrated system for asset management, condition
Migrating Netflix from Datacenter Oracle to Global CassandraAdrian Cockcroft
Netflix is migrating its datacenter infrastructure from Oracle databases to a globally distributed Apache Cassandra database on AWS. This will allow Netflix to scale more easily and deploy new features faster without being limited by the capacity of its own datacenters. The migration involves transitionally replicating data between Oracle and AWS services like SimpleDB while new services are deployed directly on Cassandra. This will cut Netflix's dependence on its existing datacenters and allow it to fully leverage the elasticity of the public cloud.
Douglas E. Wendelboe seeks a position as a Firmware Designer or Project Lead for biomedical electronic applications. He has over 20 years of experience designing software, firmware, and leading projects. He is experienced in developing biomedical devices according to FDA and ISO regulations. His skills include digital and analog hardware design, embedded systems, and programming languages like C, C++, and assembly.
Four Problems You Run into When DIY-ing a “Big Data” Analytics SystemTreasure Data, Inc.
The document discusses four common problems encountered when building a DIY big data analytics system: 1) how to collect and store data, 2) how to query data, 3) how different users access query results, and 4) how to scale the system. It introduces Treasure Data as a solution that handles all these problems, allowing users to collect, store, query, access, and scale their data easily without having to manage infrastructure. Treasure Data provides analytics as a service using Hadoop and has tools that support data collection, querying, sharing results between different roles, and automatic scaling as more data and queries are added.
Supporting Valorization of Cultural Heritage Documentation: TIVal ApproachGiuseppe Vizzari
Presentation of the paper "Supporting Valorization of Cultural Heritage Documentation:The TIVal Portal Approach" held in the context of the VSMM2012 Conference in Milano, Sept. 5, 2012
This document discusses key aspects of data warehousing design and implementation including:
- Storing historical data in a separate environment from operational data with a different structure optimized for analysis rather than transactions.
- Requirements gathering, data modeling, ETL, OLAP cube design, and report development to architect the analytical environment.
- Ongoing maintenance of the data warehouse including performance tuning, query optimization, quality assurance, and incremental enhancements.
This document discusses key aspects of data warehousing design and implementation including:
- Storing historical data in a separate environment from operational data with a different structure optimized for analysis.
- Dimensional data modeling using a star or snowflake schema to partition data into facts and dimensions.
- Extracting, transforming and loading data from source systems into the data warehouse through an ETL process.
- Designing OLAP cubes to allow for multi-dimensional reporting and analysis of the data.
- Developing reports and front-ends to allow users to access and analyze the data in the data warehouse.
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...EduSkills OECD
Andreas Schleicher, Director of Education and Skills at the OECD presents at the launch of PISA 2022 Volume III - Creative Minds, Creative Schools on 18 June 2024.
This document discusses the rapid growth of digital data and the challenges of analyzing large, unstructured datasets. It notes that in just one week in 2000, the Sloan Digital Sky Survey collected more data than had been collected in all of astronomy previously. Today, the Large Hadron Collider generates 40 terabytes per second and Twitter generates over 1 terabyte of tweets daily. By 2013, annual internet traffic was predicted to reach 667 exabytes. Hadoop provides a framework to analyze these vast and diverse datasets by distributing processing across commodity clusters close to where the data is stored.
The document discusses predictions for the future of web content management (WCM). It predicts that the future will be open, with open standards like JCR and CMIS driving ubiquity. It will be cloud-based, with hardware resources available on demand. WCM systems will be hybrid, using on-premises and cloud resources. The future will also be business-oriented, aligning WCM with business goals and stakeholders. Context and personalization will be important, driving user experiences. Systems will need to be agile to adapt to changing markets and data. Finally, the future will be mobile-centric, with built-in support for multiple channels and device detection.
Cms forum, future of Web Content Managementguest88136a
This document discusses predictions for the future of web content management (WCM). It makes 6 predictions: 1) The future is open with open standards like JCR and CMIS and open source software. 2) The future is cloudy with more content hosted in the cloud to reduce the need for on-premise hardware. 3) The future is hybrid with content hosted both on-premise and in the cloud. 4) The future is agile to adapt to changing markets and content that is likely to change. 5) The future is context-centric to better understand users and tailor experiences. 6) The future is mobile as more content is consumed on mobile devices.
This document discusses the new features and capabilities of QlikView 11, a business intelligence and data visualization platform. It highlights five main themes: 1) Social Business Discovery which enables collaboration and sharing of insights, 2) Mobile Business Discovery with optimized experiences for tablets and smartphones, 3) Comparative Analysis for comparing different data groupings, 4) Rapid Analytic App Platform with improved development and extensibility, and 5) Enterprise Platform with performance, administration, and security improvements. The document provides examples of how organizations are using QlikView to empower business users to access, analyze, and make decisions from data on their own.
This document provides an overview of NoSQL and Hadoop technologies. It discusses the trends driving these technologies like increasing data size, connectivity of data, semi-structured data, and decoupled service architectures. It introduces concepts from academic research like Amazon Dynamo, Google BigTable, and Brewer's CAP theorem. Specific technologies are explained like Hadoop for processing large datasets using MapReduce on the Hadoop Distributed File System.
The document describes a quality attribute workshop to help define and elaborate on non-functional requirements. The workshop involves building a quality attribute tree to characterize requirements, generating specific quality attribute scenarios, and mapping scenarios to architecture tactics. Quality attributes are refined from high-level goals into detailed scenarios. Scenarios are prioritized based on importance and difficulty. Example generic and specific scenarios are provided for attributes like performance and modifiability. Architecture tactics that could satisfy scenarios are also summarized. The goal is to systematically define and select the most important quality requirements to inform architectural decisions.
Cloud Computing through FCAPS Managed Services in a Virtualized Data Centervsarathy
This document discusses driving cloud computing through managed services in a virtualized data center. It introduces cloud computing and defines it as more than just an on-demand XaaS stack. The cloud must address issues like massive scalability, reliability, availability, performance optimization, and security. It also discusses how the cloud can help reduce rising data center complexity and costs which are getting out of control.
IC-SDV 2018: Harald Jenny (CENTREDOC) When Artificial Intelligence Joins Inte...Dr. Haxel Consult
Three years ago, we thought this would be impossible to accomplish. Today, it is a reality : Artificial Intelligence gains more and more importance in the value chain of Intellectual Property. The « red button » allowing to obtain, with very limited interaction from human intelligence, a full technology landscape while starting from a simple technical fact sheet, is at hand. Assessment of the progress made so far.
Sirris innovate2011 - Lily, Smart Data at scale made easy, Steven Noels, Oute...Sirris
Data growth is rapidly surpassing Moore's Law, as data sets are growing increasingly large, hence deriving insights from these large data sets is becoming more and more complex. Lily, a software product made by Outerthought, allows you to store, index and search vast quantities of data. In the next few years, successful business models will be based on monetization of data. Steven Noels will highlight the raison d'être of Lily, discussing challenges that every data-intensive organisation encounters.
Part 2 OCLC Strategic Presentation Bruce Crocco ACURIL 2011Antonio Alba
OCLC is building web-scale capabilities with libraries by focusing on data, community, and infrastructure. It is developing an open and extensible platform built on WorldCat that will provide responsive services at massive scale. This platform will allow libraries to innovate, expose and share applications, and benefit from others' innovations. Currently, OCLC provides global exposure through WorldCat.org, discovery through WorldCat Local, and circulation and acquisitions capabilities. It is also developing knowledge base management, license management, and an open platform for third-party development. OCLC aims to collaboratively build web-scale services with libraries to reduce costs and workflows through its cooperative approach.
Ordex Presentation at Nationaal Congres Open Data Eindhoven 20 april 2012Tom Zeppenfeldt IEC MSc
Presentation by Tom Zeppenfeldt from Ophileon on the Open Reporting Data Exchange, a platform to share open data which is integrated with a monitoring and reporting tool.
Ordex serves as a source for open data for visualization
In this age of Big Data, data volumes grow exceedingly larger while the technical problems and business scenarios become more complex. Compounding these complexities, data consumers are demanding faster analysis to common business questions asked of their Big Data. This session provides concrete examples of how to address this challenge. We will highlight the use of Big Data technologies—including Hadoop and Hive —with classic BI systems such as SQL Server Analysis Services.
Session takeaways:
• Understand the architectural components surrounding Hadoop, Hive, Classic BI, and the Tier-1 BI ecosystem
• Get strategies for addressing the technical issues when working with extremely large cubes
• See how to address the technical issues when working with Big Data systems from the DBA perspective
Enterprise GIS Implementation for Public Infrastructure and Integration with ...Michael Baker Jr., Inc.
The document describes how Prince William County Service Authority implemented an enterprise GIS system to integrate its water and wastewater utility data and operations. Key aspects included:
1) Migrating from an Autodesk MapGuide system to an ArcGIS Server-based system with Oracle Spatial database to serve authoritative utility data to web applications.
2) Designing a GIS web portal for searching and viewing utility asset data, work orders, customer information, and consumption data.
3) Integrating the GIS with work order management, customer information, and SCADA systems to improve operations and customer service functions like service requests and outage notifications.
4) Using the integrated system for asset management, condition
Migrating Netflix from Datacenter Oracle to Global CassandraAdrian Cockcroft
Netflix is migrating its datacenter infrastructure from Oracle databases to a globally distributed Apache Cassandra database on AWS. This will allow Netflix to scale more easily and deploy new features faster without being limited by the capacity of its own datacenters. The migration involves transitionally replicating data between Oracle and AWS services like SimpleDB while new services are deployed directly on Cassandra. This will cut Netflix's dependence on its existing datacenters and allow it to fully leverage the elasticity of the public cloud.
Douglas E. Wendelboe seeks a position as a Firmware Designer or Project Lead for biomedical electronic applications. He has over 20 years of experience designing software, firmware, and leading projects. He is experienced in developing biomedical devices according to FDA and ISO regulations. His skills include digital and analog hardware design, embedded systems, and programming languages like C, C++, and assembly.
Four Problems You Run into When DIY-ing a “Big Data” Analytics SystemTreasure Data, Inc.
The document discusses four common problems encountered when building a DIY big data analytics system: 1) how to collect and store data, 2) how to query data, 3) how different users access query results, and 4) how to scale the system. It introduces Treasure Data as a solution that handles all these problems, allowing users to collect, store, query, access, and scale their data easily without having to manage infrastructure. Treasure Data provides analytics as a service using Hadoop and has tools that support data collection, querying, sharing results between different roles, and automatic scaling as more data and queries are added.
Supporting Valorization of Cultural Heritage Documentation: TIVal ApproachGiuseppe Vizzari
Presentation of the paper "Supporting Valorization of Cultural Heritage Documentation:The TIVal Portal Approach" held in the context of the VSMM2012 Conference in Milano, Sept. 5, 2012
This document discusses key aspects of data warehousing design and implementation including:
- Storing historical data in a separate environment from operational data with a different structure optimized for analysis rather than transactions.
- Requirements gathering, data modeling, ETL, OLAP cube design, and report development to architect the analytical environment.
- Ongoing maintenance of the data warehouse including performance tuning, query optimization, quality assurance, and incremental enhancements.
This document discusses key aspects of data warehousing design and implementation including:
- Storing historical data in a separate environment from operational data with a different structure optimized for analysis.
- Dimensional data modeling using a star or snowflake schema to partition data into facts and dimensions.
- Extracting, transforming and loading data from source systems into the data warehouse through an ETL process.
- Designing OLAP cubes to allow for multi-dimensional reporting and analysis of the data.
- Developing reports and front-ends to allow users to access and analyze the data in the data warehouse.
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...EduSkills OECD
Andreas Schleicher, Director of Education and Skills at the OECD presents at the launch of PISA 2022 Volume III - Creative Minds, Creative Schools on 18 June 2024.
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...TechSoup
Whether you're new to SEO or looking to refine your existing strategies, this webinar will provide you with actionable insights and practical tips to elevate your nonprofit's online presence.
How to Download & Install Module From the Odoo App Store in Odoo 17Celine George
Custom modules offer the flexibility to extend Odoo's capabilities, address unique requirements, and optimize workflows to align seamlessly with your organization's processes. By leveraging custom modules, businesses can unlock greater efficiency, productivity, and innovation, empowering them to stay competitive in today's dynamic market landscape. In this tutorial, we'll guide you step by step on how to easily download and install modules from the Odoo App Store.
How to Manage Reception Report in Odoo 17Celine George
A business may deal with both sales and purchases occasionally. They buy things from vendors and then sell them to their customers. Such dealings can be confusing at times. Because multiple clients may inquire about the same product at the same time, after purchasing those products, customers must be assigned to them. Odoo has a tool called Reception Report that can be used to complete this assignment. By enabling this, a reception report comes automatically after confirming a receipt, from which we can assign products to orders.
1. “The role of Digital.CSIC in the building
of a CRIS at CSIC”
First International Workshop for Transfer of Information
for Innovation
Valencia, November 3, 2011
Isabel Bernal
Unit of Scientific Resources for Research (URICI)
Digital.CSIC Technical Office
Spanish National Research Council (CSIC) 1
2. Institutional data management and
dissemination: a complex landscape
Programación,
VICEPRESIDENCIA DE Seguimiento VICEPRESI
ORGANIZACIÓN Y y Documentación PCO DENCIA DE
RELACIONES Científica
ÍNVESTIGA
INSTITUCIONALES CIÓN
CIENTÍFICA
Y TÉCNICA
SGAI 2
3. • Digital.CSIC within “Frames” Action Line, Vice-
presidency of Organization and Institutional
Relations
• A part of CSIC System of Scientific Information
(SIC)
• A tool to monitor CSIC research outputs
• A role in the institutional research evaluation
exercise
• 75% of CSIC outputs accessible through the
repository by 2013
3
4. Digital.CSIC: Content growth
through manual input
• Nearly 40,000 items
• 85% deposits through
Mediated Archiving
Service
• Fully dedicated
Technical Office
• One repository
librarian in each of 77
CSIC libraries
• 900 item deposit
monthly on average 4
5. Automated deposits: first
experiences and challenges
• Batch imports from CSIC • Not that automated..:
centers web sites: Metadata crosswalks
Robotics and Industrial Syntax differences
Informatics Institute (IRII),
Institute for Cross- One time imports
Disciplinary Physics and No synchronization in
Complex Systems content population
(IFISC) Missing fields: manual input
• SWORD: BioMed • 1400+ collections in
Central, arXiv Digital.CSIC
5
6. CSIC scientific information
management system: mergers,
sinergies, automation
OUTPUTS
External sources
INSTITUTIONAL (Scopus, WoK, others)
INPUT SOURCES
Reports,
analysis,
exports etc
BDC
Scientific
Activities
CSIC
NBDC
Annual Report
Entities
and staff
ConCiencia
CVN Enriched CVN
CSIC
Action Plan Productivity
by Objectives
Accomplishment
Productivity
by Objectives Digital.CSIC
Accomplishment
6
7. Challenges on the way: CSIC
affiliations cleaning and Scopus APIs
functioning
7
9. ConCiencia, a new deposit process
for Digital.CSIC?
• Metadata of 130,000+ research outputs
• Monthly content updates through APIs and
batch imports
• 4,500+ full texts uploaded in 3 months
• A substitute for DSpace submission
process? 9
10. Making ConCiencia and
Digital.CSIC interact
ISSUES:
Autenticación
CAS
Correcta
Digital.CSIC
SO L
R
BUSCADOR (1)
SOLR
Conciencia
• Items duplicates
RESULTADOS
-Exportar excell
(2) • Synchronization of
DUPLICADOS data upload in both
SI
VALIDACIÓN
DUPLICADOS
(3)
platforms
NO
¿Fichero? DC (SI)
Registro
Validado
- Log actividad
• Metadata validation
- Grabar handle CC
DC (NO), CC (SI)
EDICIÓN (4)
• Author identities
PENDIENTE GRABACIÓN
check-up
SWORD ??
- Log actividad
SWORD
- Log actividad
• Copyright verification
- Marcado pendiente - Grabar handle CC
10
12. Where are we heading to for enhanced
access to knowledge?
• As a preliminary step for enhanced transfer of
knowledge, long standing issues for efficient internal
workflow remain:
-What metadata schema and granularity for CSIC research
outputs
-A unique persistent author identifier
-Discipline in CSIC affiliations names
-Inner dinamism and own initiatives within CSIC centers
-The challenge of CSIC mixed centers and institutes
-Linked data
12
13. Thanks for listening!
Isabel.bernal@bib.csic.es
Unit of Scientific Resources for Research (URICI)
Digital.CSIC Technical Office
Spanish National Research Council (CSIC)
13