As buildings become more complex, it becomes more difficult to manage and operate them effectively. The holistic management and maintenance of facilities is a multi-domain problem encompassing financial accounting, building maintenance, human resources, asset management and code compliance, affecting different stakeholders in different ways. One technique, called scenario modelling, customises data-driven decision support for building managers during building operation. However, current implementations of scenario modeling have been limited to data from Building Management Systems with little interaction with other relevant data sources due to interoperability issues. Linked data helps to overcome interoperability challenges to enable data from multiple domains to be merged into holistic scenario models for different stakeholders of the building. The approach is demonstrated using an owner-occupied office building.
An Environmental Chargeback for Data Center and Cloud Computing ConsumersEdward Curry
Government, business, and the general public increasingly agree that the polluter should pay. Carbon dioxide and environmental damage are considered viable chargeable commodities. The net effect of this for data center and cloud computing operators is that they should look to “chargeback” the environmental impacts of their services to the consuming end-users. An environmental chargeback model can have a positive effect on environmental impacts by linking consumers to the indirect impacts of their usage, facilitating clearer understanding of the impact of their actions. In this paper we motivate the need for environmental chargeback mechanisms. The environmental chargeback model is described including requirements, methodology for definition, and environmental impact allocation strategies. The paper details a proof-of-concept within an operational data center together with discussion on experiences gained and future research directions.
Curry, E.; Hasan, S.; White, M.; and Melvin, H. 2012. An Environmental Chargeback for Data Center and Cloud Computing Consumers. In Huusko, J.; de Meer, H.; Klingert, S.; and Somov, A., eds., First International Workshop on Energy-Efficient Data Centers. Madrid, Spain: Springer Berlin / Heidelberg.
Approximate Semantic Matching of Heterogeneous EventsEdward Curry
Event-based systems have loose coupling within space, time and synchronization, providing a scalable infrastructure for information exchange and distributed workflows. However, event-based systems are tightly coupled, via event subscriptions and patterns, to the semantics of the underlying event schema and values. The high degree of semantic heterogeneity of events in large and open deployments such as smart cities and the sensor web makes it difficult to develop and maintain event-based systems. In order to address semantic coupling within event-based systems, we propose vocabulary free subscriptions together with the use of approximate semantic matching of events. This paper examines the requirement of event semantic decoupling and discusses approximate semantic event matching and the consequences it implies for event processing systems. We introduce a semantic event matcher and evaluate the suitability of an approximate hybrid matcher based on both thesauri-based and distributional semantics-based similarity and relatedness measures. The matcher is evaluated over show that the approach matches a representation of Wikipedia and Freebase events. Initial evaluations events structured with maximal combined precision-recall F1 score of 75.89% on average in all experiments with a subscription set of 7 subscriptions. The evaluation shows how a hybrid approach to semantic event matching outperforms a single similarity measure approach.
Hasan S, O'Riain S, Curry E. Approximate Semantic Matching of Heterogeneous Events. In: 6th ACM International Conference on Distributed Event-Based Systems (DEBS 2012).
System of Systems Information Interoperability using a Linked DataspaceEdward Curry
System of Systems pose significant technical challenges in terms of information interoperability that require overcoming conceptual barriers (both syntax and semantic) and technological barriers. This paper presents an approach to System of Systems information interoperability based on the Dataspace data management abstraction and the Linked Data approach to sharing information on the web. The paper describes the fundamentals of the approach and demonstrates the concept with a System of Systems for enterprise energy management.
Curry E. System of Systems Information Interoperability using a Linked Dataspace. In: IEEE 7th International Conference on System of Systems Engineering (SOSE 2012)
Further Reading:
http://www.edwardcurry.org/publications/Curry_LinkedDataspaceForSOS_SOSE.pdf
Challenges Ahead for Converging Financial DataEdward Curry
Consumers of financial information come in many guises from personal investors looking for that value for money share, to government regulators investigating corporate fraud, to business executives seeking competitive advantage over their competition. While the particular analysis performed by each of these information consumers will vary, they all have to deal with the explosion of information available from multiple sources including, SEC filings, corporate press releases, market press coverage, and expert commentary. Recent economic events have begun to bring sharp focus on the activities and actions of financial markets, institutions and not least regulatory authorities. Calls for enhanced scrutiny will bring increased regulation and information transparency While extracting information from individual filings is relatively easy to perform when a machine readable format is utilized (for example, using XBRL, the eXtensible Business Reporting Language), cross comparison of extracted financial information can be problematic as descriptions and accounting terms vary across companies and jurisdictions. Across multiple sources the problem becomes the classical data integration problem where a common data abstraction is necessary before functional data use can begin. Within this paper we discuss the challenges in converging financial data from multiple sources. We concentrate on integrating data from multiple sources in terms of the abstraction, linking, and consolidation activities needed to consolidate data before more sophisticated analysis algorithms can examine the data for the objectives of particular information consumers (for e.g. competitive analysis, regulatory compliance, or investor analysis). We base our discussion on several years researching and deploying data integration systems in both the web and enterprise environments.
E. Curry, A. Harth, and S. O’Riain, “Challenges Ahead for Converging Financial Data,” in Proceedings of the XBRL/W3C Workshop on Improving Access to Financial Data on the Web, 2009.
The document discusses linking building data from multiple systems using linked data principles. Building information is currently scattered across different systems for energy usage, maintenance, finance, occupancy, and more. However, effectively managing a building requires a holistic view across all this data. Linked data provides a method to expose, share, and connect building data from different systems and technologies by identifying objects with URIs and linking information with relationships. This can provide new insights by linking domains like energy and resource utilization. The challenges of data interoperability, information granularity, interpreting data, and empowering actions are also discussed. A case study of applying these approaches to link operational and sensor data from a research building is also presented.
Wikipedia (DBpedia): Crowdsourced Data CurationEdward Curry
Wikipedia is an open-source encyclopedia, built collaboratively by a large community of web editors. The success of Wikipedia as one of the most important sources of information available today still challenges existing models of content creation. Despite the fact that the term ‘curation’ is not commonly addressed by Wikipedia’s contributors, the task of digital curation is the central activity of Wikipedia editors, who have the responsibility for information quality standards.
Wikipedia, is already widely used as a collaborative environment inside organizations5.
The investigation of the collaboration dynamics behind Wikipedia highlights important features and good practices which can be applied to different organizations. Our analysis focuses on the curation perspective and covers two important dimensions: social organization and artifacts, tools & processes for cooperative work coordination. These are key enablers that support the creation of high quality information products in Wikipedia’s decentralized environment.
Querying Heterogeneous Datasets on the Linked Data WebEdward Curry
The growing number of datasets published on the Web as linked data brings both opportunities for high data availability and challenges inherent to querying data in a semantically heterogeneous and distributed environment. Approaches used for querying siloed databases fail at Web-scale because users don't have an a priori understanding of all the available datasets. This article investigates the main challenges in constructing a query and search solution for linked data and analyzes existing approaches and trends.
An Environmental Chargeback for Data Center and Cloud Computing ConsumersEdward Curry
Government, business, and the general public increasingly agree that the polluter should pay. Carbon dioxide and environmental damage are considered viable chargeable commodities. The net effect of this for data center and cloud computing operators is that they should look to “chargeback” the environmental impacts of their services to the consuming end-users. An environmental chargeback model can have a positive effect on environmental impacts by linking consumers to the indirect impacts of their usage, facilitating clearer understanding of the impact of their actions. In this paper we motivate the need for environmental chargeback mechanisms. The environmental chargeback model is described including requirements, methodology for definition, and environmental impact allocation strategies. The paper details a proof-of-concept within an operational data center together with discussion on experiences gained and future research directions.
Curry, E.; Hasan, S.; White, M.; and Melvin, H. 2012. An Environmental Chargeback for Data Center and Cloud Computing Consumers. In Huusko, J.; de Meer, H.; Klingert, S.; and Somov, A., eds., First International Workshop on Energy-Efficient Data Centers. Madrid, Spain: Springer Berlin / Heidelberg.
Approximate Semantic Matching of Heterogeneous EventsEdward Curry
Event-based systems have loose coupling within space, time and synchronization, providing a scalable infrastructure for information exchange and distributed workflows. However, event-based systems are tightly coupled, via event subscriptions and patterns, to the semantics of the underlying event schema and values. The high degree of semantic heterogeneity of events in large and open deployments such as smart cities and the sensor web makes it difficult to develop and maintain event-based systems. In order to address semantic coupling within event-based systems, we propose vocabulary free subscriptions together with the use of approximate semantic matching of events. This paper examines the requirement of event semantic decoupling and discusses approximate semantic event matching and the consequences it implies for event processing systems. We introduce a semantic event matcher and evaluate the suitability of an approximate hybrid matcher based on both thesauri-based and distributional semantics-based similarity and relatedness measures. The matcher is evaluated over show that the approach matches a representation of Wikipedia and Freebase events. Initial evaluations events structured with maximal combined precision-recall F1 score of 75.89% on average in all experiments with a subscription set of 7 subscriptions. The evaluation shows how a hybrid approach to semantic event matching outperforms a single similarity measure approach.
Hasan S, O'Riain S, Curry E. Approximate Semantic Matching of Heterogeneous Events. In: 6th ACM International Conference on Distributed Event-Based Systems (DEBS 2012).
System of Systems Information Interoperability using a Linked DataspaceEdward Curry
System of Systems pose significant technical challenges in terms of information interoperability that require overcoming conceptual barriers (both syntax and semantic) and technological barriers. This paper presents an approach to System of Systems information interoperability based on the Dataspace data management abstraction and the Linked Data approach to sharing information on the web. The paper describes the fundamentals of the approach and demonstrates the concept with a System of Systems for enterprise energy management.
Curry E. System of Systems Information Interoperability using a Linked Dataspace. In: IEEE 7th International Conference on System of Systems Engineering (SOSE 2012)
Further Reading:
http://www.edwardcurry.org/publications/Curry_LinkedDataspaceForSOS_SOSE.pdf
Challenges Ahead for Converging Financial DataEdward Curry
Consumers of financial information come in many guises from personal investors looking for that value for money share, to government regulators investigating corporate fraud, to business executives seeking competitive advantage over their competition. While the particular analysis performed by each of these information consumers will vary, they all have to deal with the explosion of information available from multiple sources including, SEC filings, corporate press releases, market press coverage, and expert commentary. Recent economic events have begun to bring sharp focus on the activities and actions of financial markets, institutions and not least regulatory authorities. Calls for enhanced scrutiny will bring increased regulation and information transparency While extracting information from individual filings is relatively easy to perform when a machine readable format is utilized (for example, using XBRL, the eXtensible Business Reporting Language), cross comparison of extracted financial information can be problematic as descriptions and accounting terms vary across companies and jurisdictions. Across multiple sources the problem becomes the classical data integration problem where a common data abstraction is necessary before functional data use can begin. Within this paper we discuss the challenges in converging financial data from multiple sources. We concentrate on integrating data from multiple sources in terms of the abstraction, linking, and consolidation activities needed to consolidate data before more sophisticated analysis algorithms can examine the data for the objectives of particular information consumers (for e.g. competitive analysis, regulatory compliance, or investor analysis). We base our discussion on several years researching and deploying data integration systems in both the web and enterprise environments.
E. Curry, A. Harth, and S. O’Riain, “Challenges Ahead for Converging Financial Data,” in Proceedings of the XBRL/W3C Workshop on Improving Access to Financial Data on the Web, 2009.
The document discusses linking building data from multiple systems using linked data principles. Building information is currently scattered across different systems for energy usage, maintenance, finance, occupancy, and more. However, effectively managing a building requires a holistic view across all this data. Linked data provides a method to expose, share, and connect building data from different systems and technologies by identifying objects with URIs and linking information with relationships. This can provide new insights by linking domains like energy and resource utilization. The challenges of data interoperability, information granularity, interpreting data, and empowering actions are also discussed. A case study of applying these approaches to link operational and sensor data from a research building is also presented.
Wikipedia (DBpedia): Crowdsourced Data CurationEdward Curry
Wikipedia is an open-source encyclopedia, built collaboratively by a large community of web editors. The success of Wikipedia as one of the most important sources of information available today still challenges existing models of content creation. Despite the fact that the term ‘curation’ is not commonly addressed by Wikipedia’s contributors, the task of digital curation is the central activity of Wikipedia editors, who have the responsibility for information quality standards.
Wikipedia, is already widely used as a collaborative environment inside organizations5.
The investigation of the collaboration dynamics behind Wikipedia highlights important features and good practices which can be applied to different organizations. Our analysis focuses on the curation perspective and covers two important dimensions: social organization and artifacts, tools & processes for cooperative work coordination. These are key enablers that support the creation of high quality information products in Wikipedia’s decentralized environment.
Querying Heterogeneous Datasets on the Linked Data WebEdward Curry
The growing number of datasets published on the Web as linked data brings both opportunities for high data availability and challenges inherent to querying data in a semantically heterogeneous and distributed environment. Approaches used for querying siloed databases fail at Web-scale because users don't have an a priori understanding of all the available datasets. This article investigates the main challenges in constructing a query and search solution for linked data and analyzes existing approaches and trends.
Crowdsourcing Approaches to Big Data Curation for Earth SciencesEdward Curry
The document discusses crowdsourcing approaches to data curation for earth sciences. It covers several topics including motivation for data curation, data quality and curation processes, crowdsourcing, case studies on crowdsourced data curation, and setting up a crowdsourced data curation process. Specifically, it describes challenges with data quality, defines data curation and the role of data curators. It also outlines different types of data curation approaches based on who performs the curation (individual curators, departments, communities) and how it is done (manually, automated, crowdsourced).
Citizen Actuation For Lightweight Energy ManagementEdward Curry
In this work, we aim to utilise the concept of citizen sensors but also introduce the theory of citizen actuation. Citizen sensors observe, report, and collect data – we propose by supporting these citizen sensors with methods to affect their surroundings we enable them to become citizen actuators. We outline a use case for citizen actuation in the Energy Management domain, propose an architecture (a Cyber-Physical Social System) built on previous work in Energy Management with Twitter integration, use of Complex Event Processing (CEP), and perform an experiment to test this theory. We motivate the need for citizen actuation in Building Management Systems due to the high cost of actuation systems. We define the concept of citizen actuation and outline an experiment that shows a reduction in average energy usage of 24%. The experiment supports the concept of citizen actuation to improve energy usage within the experimental environment and we discuss future research directions in this area.
The New York Times is the largest metropolitan and the third largest newspaper in the United States. The Times website, nytimes.com, is ranked as the most
popular newspaper website in the United States and is an important source of advertisement revenue for the company. The NYT has a rich history for curation of its articles and its 100 year old curated repository has ultimately defined its participation as one of the first players in the emergingWeb of Data.
Data curation is a process that can ensure the quality of data and its fitness for use. Traditional approaches to curation are struggling with increased data volumes, and near real-time demands for curated data. In response, curation teams have turned to community crowd-sourcing and semi-automatedmetadata tools for assistance.
E. Curry, A. Freitas, and S. O’Riáin, “The Role of Community-Driven Data Curation for Enterprises,” in Linking Enterprise Data, D. Wood, Ed. Boston, MA: Springer US, 2010, pp. 25-47.
Dealing with Semantic Heterogeneity in Real-Time InformationEdward Curry
The document discusses computational paradigms for large scale open environments. It describes how environments have shifted from small controlled ones to large open ones with thousands of data sources and schemas. This requires processing information as it flows in real-time from multiple distributed sources. The talk introduces the concept of Information Flow Processing, which processes information as it streams in without intermediate storage. Examples of domains where this paradigm can be applied are given like financial analytics, inventory management and environmental monitoring.
Enterprise Energy Management using a Linked Dataspace for Energy IntelligenceEdward Curry
Energy Intelligence platforms can help organizations manage power consumption more efficiently by providing a functional view of the entire organization so that the energy consumption of business activities can be understood, changed, and reinvented to better support sustainable practices. Significant technical challenges exist in terms of information management, cross-domain data integration, leveraging real-time data, and assisting users to interpret the information to optimize energy usage. This paper presents an architectural approach to overcome these challenges using a Dataspace, Linked Data, and Complex Event Processing. The paper describes the fundamentals of the approach and demonstrates it within an Enterprise Energy Observatory.
E. Curry, S. Hasan, and S. O’Riáin, “Enterprise Energy Management using a Linked Dataspace for Energy Intelligence,” in The Second IFIP Conference on Sustainable Internet and ICT for Sustainability (SustainIT 2012), 2012.
Towards a BIG Data Public Private PartnershipEdward Curry
Building an industrial community around Big Data in Europe is the priority of the BIG: Big Data Public Private Forum project. In this workshop we will present the work of the project including analysis of foundational Big Data research technologies, technology and strategy roadmaps to enable business to understand the potential of Big Data technologies, and the necessary collaboration and dissemination infrastructure to link technology suppliers, integrators and leading user organizations. BIG is working towards the definition and implementation of a clear strategy that tackles the necessary efforts in terms of Big Data research and innovation, while also providing a major boost for technology adoption and supporting actions for the successful implementation of the Big Data economy.
The Big Data Value PPP: A Standardisation Opportunity for EuropeEdward Curry
The document summarizes the Big Data Value Public-Private Partnership (BDV PPP) and the role of standards in technology adoption. It describes the BDV PPP and Big Data Value Association (BDVA), which represents the private side of the partnership launched in 2014 between the EU and industry. It discusses how standards can improve technology adoption by helping select dominant designs when technologies move from an "era of ferment" to incremental change. The presentation argues that standards are essential for creating a data economy and that the BDV PPP will support establishing both formal and informal standards by leveraging existing standards and integrating national efforts internationally.
Developing an Sustainable IT Capability: Lessons From Intel's JourneyEdward Curry
Intel Corporation set itself a goal to reduce its global-warming greenhouse gas footprint by 20% by 2012 from 2007 levels. Through the use of sustainable IT, the Intel IT organization is recognized as a significant contributor to the company’s sustainability strategy by transforming its IT operations and overall Intel operations. This article describes how Intel has achieved IT sustainability benefits thus far by developing four key capabilities. These capabilities have been incorporated into the Sustainable ICT Capability Maturity Framework (SICT-CMF), a model developed by an industry consortium in which the authors were key participants. The article ends with lessons learned from Intel’s experiences that can be applied by business and IT executives in other enterprises.
Linked Water Data For Water Information ManagementEdward Curry
The management of water consumption is hindered by low general awareness and absence of precise historical and contextual information. Effective and efficiency management of water resources requires a holistic approach considering all the stages of water usage. A decision support tool for water management services requires access to a number of different data domains and different data providers. The design of next-generation water information management systems poses significant technical challenges in terms of information management, integration of heterogeneous data, and real-time processing of dynamic data. Linked Data is a set of web technologies that enables integration of different data sources. This work investigates the usage of Linked Data technologies in the Water Management domain, describes the fundamental concepts of the approach, details an architecture, and discusses possible water management applications.
Transforming the European Data Economy: A Strategic Research and Innovation A...Edward Curry
Transforming the European Data Economy: A Strategic Research and Innovation Agenda
Keynote at European Data Forum 2016
Prof. Dr. Milan Petković, Vice President BDVA, Philips
Dr. Edward Curry, Vice President BDVA, Insight
Towards Lightweight Cyber-Physical Energy Systems using Linked Data, the Web ...Edward Curry
Cyber-Physical Energy Systems (CPES) exploit the potential of information technology to boost energy efficiency while minimising environmental impacts. CPES can help manage energy more efficiently by providing a functional view of the entire energy system so that energy activities can be understood, changed, and reinvented to better support sustainable practices. CPES can be applied at different scales from Smart Grids and Smart Cities to Smart Enterprises and Smart Buildings. Significant technical challenges exist in terms of information management, leveraging real-time sensor data, coordination of the various stakeholders to optimize energy usage.
In this talk I describe an approach to overcome these challenges by re-using the Web standards to quickly connect the required systems within a CPES. The resulting lightweight architecture leverages Web technologies including Linked Data, the Web of Things, and Social Media. The paper describes the fundamentals of the approach and demonstrates it within an Enterprise Energy Management scenario smart building.
Collaborative Data Management: How Crowdsourcing Can Help To Manage DataEdward Curry
Data management efforts such as MDM are a popular approach for high quality enterprise data. However, MDM can be heavily centralized and labour intensive, where the cost and effort can become prohibitively high. The concentration of data management and stewardship onto a few highly skilled individuals, like developers and data experts, can be a significant bottleneck. This talk explores how to effectively involving a wider community of users within collaborative data management activities. The bottom-up approach of involving crowds in the creation and management of data has been demonstrated by projects like Freebase, Wikipedia, and DBpedia. The talk is discusses how collaborative data management can be applied within an enterprise context using platforms such as Amazon Mechanical Turk, Mobile Works, and internal enterprise human computation platforms.
Topics covered include:
- Introduction to Crowdsourcing and Human Computation for Data Management
- Crowds vs. Communities, When to use them and why
- Push vs. Pull methods of crowdsourcing data management
- Setting up and running a collaborative data management process
- Modelling the expertise of communities
Sustainable IT for Energy Management: Approaches, Challenges, and TrendsEdward Curry
An invited talk to the Galway-Mayo Institute of Technology on the current state of the art in Sustainable IT for energy management, the challenges, and the emerging trends.
Towards Unified and Native Enrichment in Event Processing SystemsEdward Curry
Events are encapsulated pieces of information that flow from one event agent to another. In order to process an event, additional information that is external to the event is often needed. This is achieved using a process called event enrichment. Current approaches to event enrichment are external to event processing engines and are handled by specialized agents. Within large-scale environments with high heterogeneity among events, the enrichment process may become difficult to maintain. This paper examines event enrichment in terms of information completeness and presents a unified model for event enrichment that takes place natively within the event processing engine. The paper describes the requirements of event enrichment and highlights its challenges such as finding enrichment sources, retrieval of information items, finding complementary information and its fusion with events. It then details an instantiation of the model using Semantic Web and Linked Data technologies. Enrichment is realised by dynamically guiding a spreading activation algorithm in a Linked Data graph. Multiple spreading activation strategies have been evaluated on a set of Wikipedia events and experimentation shows the viability of the approach.
Key Technology Trends for Big Data in EuropeEdward Curry
In this presentation we will discuss some of the results of the BIG project including analysis of foundational Big Data research technologies, technology and strategy roadmaps to enable business to understand the potential of Big Data technologies across different sectors, and the necessary collaboration and dissemination infrastructure to link technology suppliers, integrators and leading user organizations.
Edward Curry is leading the Technical Working Group of the BIG Project with over 30 committed experts along the big data value chain (Acquisition, Analysis, Curation, Storage, Usage). With the help of the other technical leads, he will elaborate on the key technology trends identified in the BIG Project and how they bring data-driven value to industrial sectors.
SLUA: Towards Semantic Linking of Users with Actions in CrowdsourcingEdward Curry
Recent advances in web technologies allow people to help solve complex problems by performing online tasks in return for money, learning, or fun. At present, human contribution is limited to the tasks defined on individual crowdsourcing platforms. Furthermore, there is a lack of tools and technologies that support matching of tasks with appropriate users, across multiple systems. A more explicit capture of the semantics of crowdsourcing tasks could enable the design and development of matchmaking services between users and tasks. The paper presents the SLUA ontology that aims to model users and tasks in crowdsourcing systems in terms of the relevant actions, capabilities, and rewards. This model describes different types of human tasks that help in solving complex problems using crowds. The paper provides examples of describing users and tasks in some real world systems, with SLUA ontology.
Improving Policy Coherence and Accessibility through Semantic Web Technologie...Edward Curry
The complexity, volume and diversity of government policies and regulations raises significant burden on both the complying parties and government itself. On the one hand, businesses, civil organizations and other societal entities are required to simultaneously comply with and interpret different and possibly conflicting or inconsistent regulations. On the other hand, government as a whole must ensure policy and regulatory coherence across its various policy domains. While the recent wave of open government initiatives have led to significantly more public access to these documents, features allowing cross-referencing related documents and linking to less formal documents or comments on other media more understandable and accessible to the public are not common if at all available today. As a solution to this challenge, we propose an Open Government-wide Policy and Regulation Information Space consisting of documents that are “semantically” annotated and cross-linked to other documents in the information space as well as to external resources such as interpretations, comments and blogs on the social web.
Our approach is three-fold. First, we identify the requirements for the infrastructure. Second, we eloborate a Reference Architecture identifying the various elements needed within the infrastructure. Third, we show how such infrastructure may be realised as a linked data portal where policies and regulations are published as linked open data. Finally, we present a case study involving environmental policy and regulations; discuss the potential impact of such infrastructure on coherency and accessibility of policies and regulations and concludes with challenges associated with provisioning a linked open policy and regulatory information infrastructure.
Open Data Innovation in Smart Cities: Challenges and TrendsEdward Curry
Open Data initiatives are increasingly considered as defining elements of emerging smart cities. However, few studies have attempted to provide a better understanding of the nature of this convergence and the impact on both domains. This talk examines the challenges and trends with open data initiatives using a socio-technical perspective of smart cities. The talk presents findings from a detailed study of 18 open data initiatives across five smart cities to identify emerging best practice. Three distinct waves of open data innovation for smart cities are discussed. The talk details the specific impacts of open data innovation on the different smart cities domains, governance of the cities, and the nature of datasets available in the open data ecosystem within smart cities.
From Data Platforms to Dataspaces: Enabling Data Ecosystems for Intelligent S...Edward Curry
This document provides an overview of a book on enabling data ecosystems for intelligent systems. It discusses key concepts like digital twins, physical-cyber-social computing, and mass personalization. It also outlines the architecture of a real-time linked dataspace platform that supports pay-as-you-go data integration and sharing for applications and intelligent systems. The platform is designed to handle streaming data from sensors and integrate it with contextual data sources using approximate semantic matching techniques.
Interactive Water Services: The Waternomics ApproachEdward Curry
The document describes the Waternomics project, which aims to introduce demand response and accountability principles in the water sector through interactive water services. The project will develop a water information platform and tools to provide personalized water consumption and availability data to households, companies and cities. It will implement pilots in Greece, Italy and Ireland to test applications like water dashboards, prediction tools, simulations and games to increase user awareness and encourage behavioral changes. The platform uses linked open data, Internet of Things sensors and semantic technologies to integrate scattered water data sources and address challenges of data interoperability across domains.
Designing Next Generation Smart City Initiatives:Harnessing Findings And Les...Edward Curry
The proliferation of “Smart Cities” initiatives around the world is part of the strategic response by governments to the challenges and opportunities of increasing urbanization and the rise of cities as the nexus of societal development. As a framework for urban transformation, Smart City initiatives aim to harness Information and Communication Technologies and Knowledge Infrastructures for economic regeneration, social cohesion, better city administration and infrastructure management. However, experiences from earlier Smart City initiatives have revealed several technical, management and governance challenges arising from the inherent nature of a Smart City as a complex “Socio- technical System of Systems”. While these early lessons are informing modest objectives for planned Smart Cities programs, no rigorous developed framework based on careful analysis of existing initiatives is available to guide policymakers, practitioners, and other Smart City stakeholders. In response to this need, this paper presents a “Smart City Initiative Design (SCID) Framework” grounded in the findings from the analysis of ten major Smart Cities programs from Netherlands, Sweden, Malta, United Arab Emirates, Portugal, Singapore, Brazil, South Korea, China and Japan. The findings provide a design space for the objectives, implementation options, strategies, and the enabling institutional and governance mechanisms for Smart City initiatives.
Crowdsourcing Approaches to Big Data Curation - Rio Big Data MeetupEdward Curry
Data management efforts such as Master Data Management and Data Curation are a popular approach for high quality enterprise data. However, Data Curation can be heavily centralised and labour intensive, where the cost and effort can become prohibitively high. The concentration of data management and stewardship onto a few highly skilled individuals, like developers and data experts, can be a significant bottleneck. This talk explores how to effectively involving a wider community of users within big data management activities. The bottom-up approach of involving crowds in the creation and management of data has been demonstrated by projects like Freebase, Wikipedia, and DBpedia. The talk discusses how crowdsourcing data management techniques can be applied within an enterprise context.
Topics covered include:
- Data Quality And Data Curation
- Crowdsourcing
- Case Studies on Crowdsourced Data Curation
- Setting up a Crowdsourced Data Curation Process
- Linked Open Data Example
- Future Research Challenges
Crowdsourcing Approaches to Big Data Curation for Earth SciencesEdward Curry
The document discusses crowdsourcing approaches to data curation for earth sciences. It covers several topics including motivation for data curation, data quality and curation processes, crowdsourcing, case studies on crowdsourced data curation, and setting up a crowdsourced data curation process. Specifically, it describes challenges with data quality, defines data curation and the role of data curators. It also outlines different types of data curation approaches based on who performs the curation (individual curators, departments, communities) and how it is done (manually, automated, crowdsourced).
Citizen Actuation For Lightweight Energy ManagementEdward Curry
In this work, we aim to utilise the concept of citizen sensors but also introduce the theory of citizen actuation. Citizen sensors observe, report, and collect data – we propose by supporting these citizen sensors with methods to affect their surroundings we enable them to become citizen actuators. We outline a use case for citizen actuation in the Energy Management domain, propose an architecture (a Cyber-Physical Social System) built on previous work in Energy Management with Twitter integration, use of Complex Event Processing (CEP), and perform an experiment to test this theory. We motivate the need for citizen actuation in Building Management Systems due to the high cost of actuation systems. We define the concept of citizen actuation and outline an experiment that shows a reduction in average energy usage of 24%. The experiment supports the concept of citizen actuation to improve energy usage within the experimental environment and we discuss future research directions in this area.
The New York Times is the largest metropolitan and the third largest newspaper in the United States. The Times website, nytimes.com, is ranked as the most
popular newspaper website in the United States and is an important source of advertisement revenue for the company. The NYT has a rich history for curation of its articles and its 100 year old curated repository has ultimately defined its participation as one of the first players in the emergingWeb of Data.
Data curation is a process that can ensure the quality of data and its fitness for use. Traditional approaches to curation are struggling with increased data volumes, and near real-time demands for curated data. In response, curation teams have turned to community crowd-sourcing and semi-automatedmetadata tools for assistance.
E. Curry, A. Freitas, and S. O’Riáin, “The Role of Community-Driven Data Curation for Enterprises,” in Linking Enterprise Data, D. Wood, Ed. Boston, MA: Springer US, 2010, pp. 25-47.
Dealing with Semantic Heterogeneity in Real-Time InformationEdward Curry
The document discusses computational paradigms for large scale open environments. It describes how environments have shifted from small controlled ones to large open ones with thousands of data sources and schemas. This requires processing information as it flows in real-time from multiple distributed sources. The talk introduces the concept of Information Flow Processing, which processes information as it streams in without intermediate storage. Examples of domains where this paradigm can be applied are given like financial analytics, inventory management and environmental monitoring.
Enterprise Energy Management using a Linked Dataspace for Energy IntelligenceEdward Curry
Energy Intelligence platforms can help organizations manage power consumption more efficiently by providing a functional view of the entire organization so that the energy consumption of business activities can be understood, changed, and reinvented to better support sustainable practices. Significant technical challenges exist in terms of information management, cross-domain data integration, leveraging real-time data, and assisting users to interpret the information to optimize energy usage. This paper presents an architectural approach to overcome these challenges using a Dataspace, Linked Data, and Complex Event Processing. The paper describes the fundamentals of the approach and demonstrates it within an Enterprise Energy Observatory.
E. Curry, S. Hasan, and S. O’Riáin, “Enterprise Energy Management using a Linked Dataspace for Energy Intelligence,” in The Second IFIP Conference on Sustainable Internet and ICT for Sustainability (SustainIT 2012), 2012.
Towards a BIG Data Public Private PartnershipEdward Curry
Building an industrial community around Big Data in Europe is the priority of the BIG: Big Data Public Private Forum project. In this workshop we will present the work of the project including analysis of foundational Big Data research technologies, technology and strategy roadmaps to enable business to understand the potential of Big Data technologies, and the necessary collaboration and dissemination infrastructure to link technology suppliers, integrators and leading user organizations. BIG is working towards the definition and implementation of a clear strategy that tackles the necessary efforts in terms of Big Data research and innovation, while also providing a major boost for technology adoption and supporting actions for the successful implementation of the Big Data economy.
The Big Data Value PPP: A Standardisation Opportunity for EuropeEdward Curry
The document summarizes the Big Data Value Public-Private Partnership (BDV PPP) and the role of standards in technology adoption. It describes the BDV PPP and Big Data Value Association (BDVA), which represents the private side of the partnership launched in 2014 between the EU and industry. It discusses how standards can improve technology adoption by helping select dominant designs when technologies move from an "era of ferment" to incremental change. The presentation argues that standards are essential for creating a data economy and that the BDV PPP will support establishing both formal and informal standards by leveraging existing standards and integrating national efforts internationally.
Developing an Sustainable IT Capability: Lessons From Intel's JourneyEdward Curry
Intel Corporation set itself a goal to reduce its global-warming greenhouse gas footprint by 20% by 2012 from 2007 levels. Through the use of sustainable IT, the Intel IT organization is recognized as a significant contributor to the company’s sustainability strategy by transforming its IT operations and overall Intel operations. This article describes how Intel has achieved IT sustainability benefits thus far by developing four key capabilities. These capabilities have been incorporated into the Sustainable ICT Capability Maturity Framework (SICT-CMF), a model developed by an industry consortium in which the authors were key participants. The article ends with lessons learned from Intel’s experiences that can be applied by business and IT executives in other enterprises.
Linked Water Data For Water Information ManagementEdward Curry
The management of water consumption is hindered by low general awareness and absence of precise historical and contextual information. Effective and efficiency management of water resources requires a holistic approach considering all the stages of water usage. A decision support tool for water management services requires access to a number of different data domains and different data providers. The design of next-generation water information management systems poses significant technical challenges in terms of information management, integration of heterogeneous data, and real-time processing of dynamic data. Linked Data is a set of web technologies that enables integration of different data sources. This work investigates the usage of Linked Data technologies in the Water Management domain, describes the fundamental concepts of the approach, details an architecture, and discusses possible water management applications.
Transforming the European Data Economy: A Strategic Research and Innovation A...Edward Curry
Transforming the European Data Economy: A Strategic Research and Innovation Agenda
Keynote at European Data Forum 2016
Prof. Dr. Milan Petković, Vice President BDVA, Philips
Dr. Edward Curry, Vice President BDVA, Insight
Towards Lightweight Cyber-Physical Energy Systems using Linked Data, the Web ...Edward Curry
Cyber-Physical Energy Systems (CPES) exploit the potential of information technology to boost energy efficiency while minimising environmental impacts. CPES can help manage energy more efficiently by providing a functional view of the entire energy system so that energy activities can be understood, changed, and reinvented to better support sustainable practices. CPES can be applied at different scales from Smart Grids and Smart Cities to Smart Enterprises and Smart Buildings. Significant technical challenges exist in terms of information management, leveraging real-time sensor data, coordination of the various stakeholders to optimize energy usage.
In this talk I describe an approach to overcome these challenges by re-using the Web standards to quickly connect the required systems within a CPES. The resulting lightweight architecture leverages Web technologies including Linked Data, the Web of Things, and Social Media. The paper describes the fundamentals of the approach and demonstrates it within an Enterprise Energy Management scenario smart building.
Collaborative Data Management: How Crowdsourcing Can Help To Manage DataEdward Curry
Data management efforts such as MDM are a popular approach for high quality enterprise data. However, MDM can be heavily centralized and labour intensive, where the cost and effort can become prohibitively high. The concentration of data management and stewardship onto a few highly skilled individuals, like developers and data experts, can be a significant bottleneck. This talk explores how to effectively involving a wider community of users within collaborative data management activities. The bottom-up approach of involving crowds in the creation and management of data has been demonstrated by projects like Freebase, Wikipedia, and DBpedia. The talk is discusses how collaborative data management can be applied within an enterprise context using platforms such as Amazon Mechanical Turk, Mobile Works, and internal enterprise human computation platforms.
Topics covered include:
- Introduction to Crowdsourcing and Human Computation for Data Management
- Crowds vs. Communities, When to use them and why
- Push vs. Pull methods of crowdsourcing data management
- Setting up and running a collaborative data management process
- Modelling the expertise of communities
Sustainable IT for Energy Management: Approaches, Challenges, and TrendsEdward Curry
An invited talk to the Galway-Mayo Institute of Technology on the current state of the art in Sustainable IT for energy management, the challenges, and the emerging trends.
Towards Unified and Native Enrichment in Event Processing SystemsEdward Curry
Events are encapsulated pieces of information that flow from one event agent to another. In order to process an event, additional information that is external to the event is often needed. This is achieved using a process called event enrichment. Current approaches to event enrichment are external to event processing engines and are handled by specialized agents. Within large-scale environments with high heterogeneity among events, the enrichment process may become difficult to maintain. This paper examines event enrichment in terms of information completeness and presents a unified model for event enrichment that takes place natively within the event processing engine. The paper describes the requirements of event enrichment and highlights its challenges such as finding enrichment sources, retrieval of information items, finding complementary information and its fusion with events. It then details an instantiation of the model using Semantic Web and Linked Data technologies. Enrichment is realised by dynamically guiding a spreading activation algorithm in a Linked Data graph. Multiple spreading activation strategies have been evaluated on a set of Wikipedia events and experimentation shows the viability of the approach.
Key Technology Trends for Big Data in EuropeEdward Curry
In this presentation we will discuss some of the results of the BIG project including analysis of foundational Big Data research technologies, technology and strategy roadmaps to enable business to understand the potential of Big Data technologies across different sectors, and the necessary collaboration and dissemination infrastructure to link technology suppliers, integrators and leading user organizations.
Edward Curry is leading the Technical Working Group of the BIG Project with over 30 committed experts along the big data value chain (Acquisition, Analysis, Curation, Storage, Usage). With the help of the other technical leads, he will elaborate on the key technology trends identified in the BIG Project and how they bring data-driven value to industrial sectors.
SLUA: Towards Semantic Linking of Users with Actions in CrowdsourcingEdward Curry
Recent advances in web technologies allow people to help solve complex problems by performing online tasks in return for money, learning, or fun. At present, human contribution is limited to the tasks defined on individual crowdsourcing platforms. Furthermore, there is a lack of tools and technologies that support matching of tasks with appropriate users, across multiple systems. A more explicit capture of the semantics of crowdsourcing tasks could enable the design and development of matchmaking services between users and tasks. The paper presents the SLUA ontology that aims to model users and tasks in crowdsourcing systems in terms of the relevant actions, capabilities, and rewards. This model describes different types of human tasks that help in solving complex problems using crowds. The paper provides examples of describing users and tasks in some real world systems, with SLUA ontology.
Improving Policy Coherence and Accessibility through Semantic Web Technologie...Edward Curry
The complexity, volume and diversity of government policies and regulations raises significant burden on both the complying parties and government itself. On the one hand, businesses, civil organizations and other societal entities are required to simultaneously comply with and interpret different and possibly conflicting or inconsistent regulations. On the other hand, government as a whole must ensure policy and regulatory coherence across its various policy domains. While the recent wave of open government initiatives have led to significantly more public access to these documents, features allowing cross-referencing related documents and linking to less formal documents or comments on other media more understandable and accessible to the public are not common if at all available today. As a solution to this challenge, we propose an Open Government-wide Policy and Regulation Information Space consisting of documents that are “semantically” annotated and cross-linked to other documents in the information space as well as to external resources such as interpretations, comments and blogs on the social web.
Our approach is three-fold. First, we identify the requirements for the infrastructure. Second, we eloborate a Reference Architecture identifying the various elements needed within the infrastructure. Third, we show how such infrastructure may be realised as a linked data portal where policies and regulations are published as linked open data. Finally, we present a case study involving environmental policy and regulations; discuss the potential impact of such infrastructure on coherency and accessibility of policies and regulations and concludes with challenges associated with provisioning a linked open policy and regulatory information infrastructure.
Open Data Innovation in Smart Cities: Challenges and TrendsEdward Curry
Open Data initiatives are increasingly considered as defining elements of emerging smart cities. However, few studies have attempted to provide a better understanding of the nature of this convergence and the impact on both domains. This talk examines the challenges and trends with open data initiatives using a socio-technical perspective of smart cities. The talk presents findings from a detailed study of 18 open data initiatives across five smart cities to identify emerging best practice. Three distinct waves of open data innovation for smart cities are discussed. The talk details the specific impacts of open data innovation on the different smart cities domains, governance of the cities, and the nature of datasets available in the open data ecosystem within smart cities.
From Data Platforms to Dataspaces: Enabling Data Ecosystems for Intelligent S...Edward Curry
This document provides an overview of a book on enabling data ecosystems for intelligent systems. It discusses key concepts like digital twins, physical-cyber-social computing, and mass personalization. It also outlines the architecture of a real-time linked dataspace platform that supports pay-as-you-go data integration and sharing for applications and intelligent systems. The platform is designed to handle streaming data from sensors and integrate it with contextual data sources using approximate semantic matching techniques.
Interactive Water Services: The Waternomics ApproachEdward Curry
The document describes the Waternomics project, which aims to introduce demand response and accountability principles in the water sector through interactive water services. The project will develop a water information platform and tools to provide personalized water consumption and availability data to households, companies and cities. It will implement pilots in Greece, Italy and Ireland to test applications like water dashboards, prediction tools, simulations and games to increase user awareness and encourage behavioral changes. The platform uses linked open data, Internet of Things sensors and semantic technologies to integrate scattered water data sources and address challenges of data interoperability across domains.
Designing Next Generation Smart City Initiatives:Harnessing Findings And Les...Edward Curry
The proliferation of “Smart Cities” initiatives around the world is part of the strategic response by governments to the challenges and opportunities of increasing urbanization and the rise of cities as the nexus of societal development. As a framework for urban transformation, Smart City initiatives aim to harness Information and Communication Technologies and Knowledge Infrastructures for economic regeneration, social cohesion, better city administration and infrastructure management. However, experiences from earlier Smart City initiatives have revealed several technical, management and governance challenges arising from the inherent nature of a Smart City as a complex “Socio- technical System of Systems”. While these early lessons are informing modest objectives for planned Smart Cities programs, no rigorous developed framework based on careful analysis of existing initiatives is available to guide policymakers, practitioners, and other Smart City stakeholders. In response to this need, this paper presents a “Smart City Initiative Design (SCID) Framework” grounded in the findings from the analysis of ten major Smart Cities programs from Netherlands, Sweden, Malta, United Arab Emirates, Portugal, Singapore, Brazil, South Korea, China and Japan. The findings provide a design space for the objectives, implementation options, strategies, and the enabling institutional and governance mechanisms for Smart City initiatives.
Crowdsourcing Approaches to Big Data Curation - Rio Big Data MeetupEdward Curry
Data management efforts such as Master Data Management and Data Curation are a popular approach for high quality enterprise data. However, Data Curation can be heavily centralised and labour intensive, where the cost and effort can become prohibitively high. The concentration of data management and stewardship onto a few highly skilled individuals, like developers and data experts, can be a significant bottleneck. This talk explores how to effectively involving a wider community of users within big data management activities. The bottom-up approach of involving crowds in the creation and management of data has been demonstrated by projects like Freebase, Wikipedia, and DBpedia. The talk discusses how crowdsourcing data management techniques can be applied within an enterprise context.
Topics covered include:
- Data Quality And Data Curation
- Crowdsourcing
- Case Studies on Crowdsourced Data Curation
- Setting up a Crowdsourced Data Curation Process
- Linked Open Data Example
- Future Research Challenges
The Role of Community-Driven Data Curation for EnterprisesEdward Curry
With increased utilization of data within their operational and strategic processes, enterprises need to ensure data quality and accuracy. Data curation is a process that can ensure the quality of data and its fitness for use. Traditional approaches to curation are struggling with increased data volumes, and near real-time demands for curated data. In response, curation teams have turned to community crowd-sourcing and semi-automatedmetadata tools for assistance. This chapter provides an overview of data curation, discusses the business motivations for curating data and investigates the role of community-based data curation, focusing on internal communities and pre-competitive data collaborations. The chapter is supported by case studies from Wikipedia, The New York Times, Thomson Reuters, Protein Data Bank and ChemSpider upon which best practices for both social and technical aspects of community-driven data curation are described.
E. Curry, A. Freitas, and S. O’Riáin, “The Role of Community-Driven Data Curation for Enterprises,” in Linking Enterprise Data, D. Wood, Ed. Boston, MA: Springer US, 2010, pp. 25-47.
Big Data Public Private Forum (BIG) @ European Data Forum 2013Edward Curry
The BIG project aims to promote adoption of big data technologies in Europe. It brings together stakeholders from industry and government to shape big data development. The project will identify sector needs, apply big data solutions, and create roadmaps. It has a budget of 3 million euros over 26 months. BIG works through industry-led forums and technical groups to foster big data use across multiple sectors like health, telecom and manufacturing. The project schedules white papers, requirements documents, and roadmaps to be delivered throughout its timeline.
A Capability Maturity Framework for Sustainable ICTEdward Curry
The document proposes a Capability Maturity Framework for Sustainable ICT developed by the Innovation Value Institute. It aims to help organizations assess and improve their maturity in sustainable ICT practices. The framework evaluates capabilities across nine building blocks including strategy, processes, people and culture, and governance. Assessments provide insight into an organization's strengths and challenges to develop sustainable ICT. Increasing maturity involves systematically improving each of the nine building blocks over multiple levels from ad hoc to optimized practices.
Crowdsourcing Approaches for Smart City Open Data ManagementEdward Curry
A wide-scale bottom-up approach to the creation and management of open data has been demonstrated by projects like Freebase, Wikipedia, and DBpedia. This talk explores how to involving a wide community of users in collaborative management of open data activities within a Smart City. The talk discusses how crowdsourcing techniques can be applied within a Smart City context using crowdsourcing and human computation platforms such as Amazon Mechanical Turk, Mobile Works, and Crowd Flower.
The document describes the Digital Enterprise Research Institute (DERI) and its work on enabling networked knowledge. DERI aims to link scientific research with industry through fundamental research, technology development, and education. Its goals include exploiting big data and enabling smart cities through removing data silos and leveraging linked open data. DERI is developing technologies like the Semantic Sensor Network ontology, CoAP protocol, and Continuous Query Evaluation over Linked Streams (CQELS) to process sensor data and queries over linked streams and datasets in real-time.
The document provides information about the Digital Enterprise Research Institute (DERI) in Galway, Ireland. It discusses DERI's research areas including semantic web, social networks, and data mining. It also outlines DERI's funding sources and partners. The document then shifts to discussing linked open data, including its key components like RDF and vocabularies. Finally, it provides examples of linked open data projects by DERI and others.
EDF2013: Keynote Stefan Decker: Big Data In Ireland - Linked Data and beyondEuropean Data Forum
Keynote of Stefan Decker, Professor for Digital Enterprise & Director of DERI, National University of Ireland, Galway, at the European Data Forum 2013, 9 April 2013 in Dublin, Ireland: Big Data In Ireland - Linked Data and beyond
The US Department of Health and Human Services (HHS) began publishing Linked Data in 2011 as part of an ongoing effort to inform the public and stimulate new health care applications.
The Digital Enterprise Research Institute (DERI) is recognized as one of the leading international web science research institutes interlinking technologies, information and people to advance business and benefit society.
In the US, the President's Council of Advisors on Science and Technology (PCAST) published a report on Health IT that imagines new scenarios and recommends new capabilities for interacting with health data.
At DERI, innovative ontology and software implementations demonstrate how users can create and manage fine-grained privacy preferences that restrict or grant access to their Linked Data
This session will give an overview of the HHS/DERI collaboration to implement 'data element access services' towards the realization of patient controlled privacy.
• US Department of Health and Human Services
• PCAST Health Information Technology Report
• Digital Enterprise Research Institute
• Privacy Preference Ontology and Manager
• Puelia and Linked Data API
http://semtechbizsf2012.semanticweb.com/sessionPop.cfm?confid=65&proposalid=4539
This document describes a self-service approach to publishing government data as linked open data. It proposes using Google Refine with extensions for exporting to RDF and reconciling entities, along with best practices like the Data Catalog Vocabulary and Data Cube vocabulary. Publishers would provide metadata about their datasets using these standards. Users could then select datasets in Google Refine, map them to RDF schemas, link entities, and export the resulting RDF. This RDF could then be shared on a platform like CKAN along with provenance information about the transformation process. The approach is intended to lower the costs of publishing linked data while maintaining quality.
Human: Thank you, that is a concise and accurate 3 sentence summary of
Making sense out of disagreement, University of Limerick Interaction Design C...jodischneider
How do we make sense out of disagreement on the social web?
A talk about my dissertation work, given to the University of Limerick Interaction Design Centre on 2012-04-18.
This document discusses the development of a semantic laboratory information management system for stem cell research. It aims to unify public stem cell resources, create a plug-and-play architecture for data access, and enable decision support and agent-oriented modeling and simulations. The system addresses challenges like data incoherence and deluge by mapping databases to ontologies to integrate stem cell data and building an agent-based stem cell model. Its current status involves automatically mapping databases and creating ontologies.
This document discusses enabling linked open government data through the use of linked data principles and vocabularies. It describes open government data and the benefits of publishing data as linked open data using RDF and shared vocabularies. It also discusses using the Data Catalog Vocabulary (DCAT) to describe government data catalogs as linked data and how this can enable federated search across catalogs. Finally, it outlines steps for publishing government datasets as linked open data.
Annotating Microblog Posts with Sensor Data for Emergency Reporting ApplicationsDavid Crowley
This document discusses annotating microblog posts with sensor data from smartphones for emergency reporting applications. It describes using sensors like GPS, accelerometers and microphones to provide context like location, activity and noise levels to tweets. The goal is to develop a semantic microblogging app that can attach this sensor metadata to tweets in emergencies, providing more useful information to emergency teams while addressing challenges around privacy, data quality and usability. Future work involves building and evaluating a prototype mobile app to collect and model sensor data from tweets.
Towards Expertise Modelling for Routing Data Cleaning Tasks within a Communit...Umair ul Hassan
https://www.insight-centre.org/content/towards-expertise-modelling-routing-data-cleaning-tasks-within-community-knowledge-workers
Presented at the ICIQ 2012
ABSTRACT:
Applications consuming data have to deal with variety of data quality issues such as missing values, duplication, incorrect values, etc. Although automatic approaches can be utilized for data cleaning the results can remain uncertain. Therefore updates suggested by automatic data cleaning algorithms require further human verification. This paper presents an approach for generating tasks for uncertain updates and routing these tasks to appropriate workers based on their expertise. Specifically the paper tackles the problem of modelling the expertise of knowledge workers for the purpose of routing tasks within collaborative data quality management. The proposed expertise model represents the profile of a worker against a set of concepts describing the data. A simple routing algorithm is employed for leveraging the expertise profiles for matching data cleaning tasks with workers. The proposed approach is evaluated on a real world dataset using human workers. The results demonstrate the effectiveness of using concepts for modelling expertise, in terms of likelihood of receiving responses to tasks routed to workers.
This document discusses the history and modern approaches to knowledge management systems on the desktop. It traces ideas back to Vannevar Bush's 1945 proposal of the memex device. More recent systems like Doug Engelbart's NLS in the 1960s and Ted Nelson's Xanadu aimed to improve linking and navigation of information. Modern semantic desktops take a layered, modular approach and use ontologies and semantic technologies to unlock and integrate desktop data. They provide services like storage, extraction, annotation and inference to enhance existing applications and help users manage information overload. Evaluation of these systems remains a challenge due to their personal, customized nature.
Ireland - The location of choice for International Payments firmsMartina Naughton
This document discusses Ireland as a location for international payments firms. It highlights Ireland's strong portfolio of financial services firms, leadership in software and ICT, and convergence of financial technology. Ireland has over 800 software firms, 24,000 employees in the sector, and 8 of the top 10 ICT companies have operations there. Financial regulation supports the payments market. Several large payments firms have partnerships and operations in Ireland, taking advantage of the business environment and government support through agencies like IDA Ireland.
Turning social disputes into knowledge representations DERI reading group 201...jodischneider
A reading group presentation about Turning social disputes into knowledge representations, based primarily on two papers:
Toni and Torroni. Bottom-up Argumentation. In: First International Workshop on the Theory and Applications of Formal Argumentation 2011 (TAFA), 16-22 July, 2011, Barcelona, Spain. http://www.doc.ic.ac.uk/~ft/PAPERS/tafaPT.pdf
Benn, Buckingham Shum, Domingue, and Mancini. Ontological Foundations for Scholarly Debate Mapping Technology. In: 2nd International Conference on Computational Models of Argument (COMMA '08), 28-30 May, 2008, Toulouse, France. http://oro.open.ac.uk/11939/
WikiSym2012 Deletion Discussions in Wikipedia: Decision Factors and Outcomesjodischneider
This document summarizes a study on deletion discussions in Wikipedia. The study analyzed 72 deletion discussions from a single day to understand the factors that contribute to deletion decisions and their outcomes. It found that four main factors—notability, sources, maintenance, and bias—accounted for 91% of comments and influenced 70% of discussions. Notability was sometimes overridden by other concerns like ensuring comprehensive coverage of a topic. The study aims to help newcomers, debate closers, and readers better understand Wikipedia's criteria for deletion.
A Small Overview of Big Data Products, Analytics, and Infrastructure at LinkedInAmy W. Tang
This talk was given by Bhaskar Ghosh (Senior Director of Engineering, LinkedIn Data Infrastructure), at the Yale Oct 2012 Symposium on Big Data, in honor of Martin Schultz.
This document discusses how bioprocessing companies have leveraged information technology (IT) solutions to help bring new therapies to market, but that maintaining complex IT infrastructure can become costly and inefficient. It argues that organizations need an enterprise-level informatics platform that integrates diverse data sources and systems to extract maximum value from research and development (R&D) data without overspending on IT. The document provides an example of how one pharmaceutical company benefited from a more holistic and integrated informatics approach based on the Accelrys Pipeline Pilot platform.
The document discusses open government data and provides examples of datasets that have been opened by various governments and organizations. It outlines types of planning, school, expense, and geographic data that could be opened, and provides links to examples of open data from other regions including UK transportation data and Toronto water billing information. The presentation encourages attendees to think about what additional local data could be opened and how it might be used.
Big Data Beyond Hadoop*: Research Directions for the FutureOdinot Stanislas
Michael Wrinn
Research Program Director, University Research Office,
Intel Corporation
Jason Dai
Engineering Director and Principal Engineer,
Intel Corporation
The document discusses using semantic web technologies like SIOC and FOAF to help journalists integrate user-generated social media content into news stories. It describes how these ontologies can connect disparate online communities and metadata to aid in navigating, verifying, and reusing social data. The project aims to explore customizing SIOC for journalism and extending vocabularies like Schema.org to semantically enrich social content for news gathering and reporting.
Information Extraction and Integration of Hard and Soft Information for D2D v...DataCards
"Information Extraction and Integration of Hard and Soft Information for D2D via
Controlled National Language,” Dr. Tien Pham, US Army Research Laboratory
Similar to Building Optimisation using Scenario Modeling and Linked Data (20)
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Building Optimisation using Scenario Modeling and Linked Data
1. Building Optimisation using Scenario
Modeling and Linked Data
Edward Curry, James O’Donnell, Edward Corry
1st Workshop Linked Data in Architecture
and Construction (LDAC2012)
Ghent
28/29 March 2012
2. Overview
Digital Enterprise Research Institute www.deri.ie
Introduction
IRUSE (Built Environment)
DERI (Semantic Web/Linked Data)
LBNL (Built Environment)
Cross-domain Data for Building Management
Enhanced Decision Support with Scenario Modelling
Challenges
Linked Building Data
DERI Building Use Case
Enabling Networked Knowledge
3. Who are IRUSE?
Based at National University of Ireland, Galway
Research Group of Civil/Mechanical Engineers
5 post-docs & 7 PhDs
3
4. IRUSE interested in Building Optimisation during
Operational Phase
HVAC systems integration and
Optimisation
Information driven building operation
Stakeholders specific performance data
Energy Simulation
Building Information Models
Calibration of simulation models
4
5. About DERI
Digital Enterprise Research Institute www.deri.ie
Founded June 2003 as a CSET (Centre for Science,
Engineering and Technology).
Link scientists and engineers / academia and industry
Fundamental research
Development of Irish-based technology companies
Attract industry
Education & outreach
DERI Institute
CSET
Commercialization, DAI
EU, EI, direct industry, IRCSET
DERI strategic plan responds to priorities
Local: University focus on Informatics, Physical & Computational
Sciences
National: SMART Economy, Program for Government
International: EU Digital Agenda
Enabling Networked Knowledge
6. About DERI
Digital Enterprise Research Institute www.deri.ie
Number one in its core space
Research Publications > 950
Participate in 17 standardisation groups (W3C, OASIS)
Approx 140 members from 30 nations
57 PhD’s /Masters
42 completed PhDs/Masters
Core Industrial Partners:
MNC’s: Cisco, Avaya, Bel-Labs, Ericsson…
SME’s: Storm, Celtrak, OpenLink……
Research: FBK
Total Research Grants: >€60 million
SFI, EU Framework, Enterprise Ireland, Industry
Enabling Networked Knowledge
7. The 2012 DERI House
Digital Enterprise Research Institute www.deri.ie
DERI Applied
Research Commercialisation
eBusiness Green &
eLearning
Financial Services Sustainable IT
Health Care Cyber
Data eGovernment
Life Sciences Security Linked
Cloud Analyt
Data
ics
Information Security,
Cloud Data Sensor
Social Software Mining Privacy
Management Middleware
and Retrieval & Trust
Data Natural Service
Reasoning and Knowledge
Visualisation Language Oriented
Querying Discovery
and Interaction Processing Architecture
DERI is designed to provide an integrated solution
Enabling Networked Knowledge
8. Lawrence Berkeley National Laboratory
Founded in 1931 Awarded 13 Nobel Prizes
~ $850 Million annual budget
4200 Employees including:
1685 Scientists, Engineers and faculty
475 Postdoctoral fellows
560 Undergraduate and graduate student employees
8
9. My position and research activities at LBNL
Energy and Environmental Sciences
Environmental Energy Technologies Division
Building Technology and Urban Systems Department
Create whole building energy
simulation models of buildings Create interoperable processes to
Develop software for whole building support whole building energy simulation
energy simulation
Create stakeholder driven information display methods, based on automated
data processing, for performance evaluation of buildings 9
10. IRUSE, LBNL, and DERI have Complementary Research
Interests
Main focus of
interest at
Decision
IRUSE/LBNL
Support
Holistic Performance View
Resource Performance Fault Stakeholder Financial
Analysis Metrics Detection Analysis Management
Analytic Layer
Main focus of
interest at DERI
Operational Stakeholder
Data Data
Aggregation Layer
RDF
BMS BIM Utilities Weather FM Financial Other Data
Raw Data Silos
10
11. Organisations incur substantial costs as a result of data
mismanagement
Confusion
No Interoperability
Cost Overruns
Higher Costs
Inefficiencies
Building
Manager
11
13. These are the types of data that we wish to leverage
Utility Bills BMS Sensor &
Weather Data
Meter Data
HR
Occupancy
Security
Fire
Simulation Models Building Models Other Data
Output
13
15. Define information required by stakeholder and related
data
Scenario Description
Performance Building Performance Performance Formulae Datum Sources
Aspects Objects Objectives Metrics
A
B
C
15
16. A building manager would like to analyse comfort and
energy consumption
Scenario: Compare Comfort & Energy Consumption
Performance Building Performance Performance Formulae Datum Sources
Aspects Objects Objectives Metrics
Measured
Datum 1: Zone
Temperature ( C)
Building Gymnasium Maintain Zone Zone
Zone Temperature Temperature = (Datum1)
Function
Datum 1: Zone
Temperature ( C)
Simulated
Measured
Datum 1: Water
Flow Rate (kg/s)
Datum 2: Water
Supply Temperature ( C)
Datum 3: Water
Return Temperature ( C)
Energy Optimise Chiller Chiller Energy =(Datum 1*Constant
Chiller
Consumption Operation Output *(Datum3-Datum2))
Datum 1: Water
Flow Rate (kg/s)
Datum 2: Water
Supply Temperature ( C)
Constant = Specific Heat Capacity of Datum 3: Water
Return Temperature ( C)
output fluid measured in J/kgK
Simulated 16
18. Key Challenges
Digital Enterprise Research Institute www.deri.ie
Initially developed a Performance Framework Tool
IFC based
Encountered significant roadblocks with BIM
Originally felt BIM was central pillar of performance assessment
Recognise an as-built BIM is one of many pillars
Technology and Data Interoperability
Data scattered among different information systems
Multiple incompatible technologies make it difficult to use
Dynamic data, sensors, ERP, BMS, assets databases, …
Enabling Networked Knowledge
18
19. Linked Building Data
Digital Enterprise Research Institute www.deri.ie
Linking building data builds context between systems
Relevant information can linked together to build holistic views of the
building
Broader context can be used in decision making
Maintains loose coupling between systems
Allows domain systems to focus on their expertise
Allows systems to develop independently
Enabling Networked Knowledge
19
20. IRUSE, LBNL, and DERI have Complementary Research
Interests
Main focus of
interest at
Decision
IRUSE/LBNL
Support
Holistic Performance View
Resource Performance Fault Stakeholder Financial
Analysis Metrics Detection Analysis Management
Analytic Layer
Main focus of
interest at DERI
Operational Stakeholder
Data Data
Aggregation Layer
RDF
BMS BIM Utilities Weather FM Financial Other Data
Raw Data Silos
20
21. Case Study: DERI Building
Digital Enterprise Research Institute www.deri.ie
DERI Building
No BMS or BEMS
160 person Office space
Café
Data centre
3 Kitchens
80 person Conference
room
4 Meeting rooms
Computing museum
Sensor Lab
Enabling Networked Knowledge
21
22. DERI Dataspace
Digital Enterprise Research Institute www.deri.ie
Applications
Decision Support Energy Analysis Energy and Situation Awareness
Systems Model Sustainability Dashboards Apps
Complex Events
Services
Support
Entity Complex Event
Data Provenance Search &
Management Processing
Catalog Query Engine
Service
Linked Data
Adapter Adapter Adapter Adapter Adapter
Sources
Enabling Networked Knowledge
23. Building Energy
Digital Enterprise Research Institute www.deri.ie
1. Data from
Enterprise
Linked Data
Cloud
2. Sensor Data
3. Building
Energy
Situation
Awareness
Enabling Networked Knowledge
23 of 26
24. DERI Energy Observatory
Digital Enterprise Research Institute www.deri.ie
Enterprise Energy Observatory
Organisation Business Process Personal
Linked dataspace for Energy Intelligence
(Linked Data, Semantic Web, Semantic Sensor Networks)
Enabling Networked Knowledge
24
25. Selected References
Digital Enterprise Research Institute www.deri.ie
Curry, E., et al . (2011). An Entity-Centric Approach To Green Information
Systems. 19th European Conference on Information Systems (ECIS 2011).
Hasan, S. et al. (2011). Toward Situation Awareness for the Semantic Sensor
Web: Complex Event Processing with Dynamic Linked Data Enrichment. 4th
International Workshop on Semantic Sensor Networks
Curry, E., & Donnellan, B. (2012). Green and Sustainable Informatics. In,
Harnessing Green IT: Principles and Practices (in press). John Wiley & Sons
Curry, E. et al, Using Multi-Domain Data to Optimize Building Operational
Performance: A Linked Data Approach to Interoperability. Advanced Engineering
Informatics. (Under Review)
White, M. et al. An Energy Efficiency Metric to Report the Cost of Data Centre
Services to Consumers in Real-Time. DCEE 2012, (Under Review)
Curry, E. et al. Towards an Open Platform for Holistic Real-time Enterprise Energy
Intelligence: A Linked Data Approach, e-Energy 2012, (Under Review)
Curry. E. et al. Intel and IT Sustainability, MISQE, (Under Review)
Enabling Networked Knowledge
Editor's Notes
{PA)s can relate to one objectGranularityObserved in contextContextualised information for end user
Need to pull information from across the enterprise, so you can understand energy use in context
Technology Stack uses tool from across the DERI house
Platform allows us to build very detailed energy management applications