Learn about how the Transforming Transport Lighthouse Project is helping to transform the Transport and Logistics domains using big data technologies. Lessons learned, pitfalls, innovation potential and business insights.
THIRUVANANTHAPURAM, JULY 19:
Marlabs, a Bangalore-based provider of IT services, is sponsoring a ‘Business Intelligence Technology’ conference at the Thiruvananthapuram Technopark on Friday.
The event will focus on emerging trends in Business Intelligence (BI) Technology, a Marlabs spokesman said.
It will feature eminent speakers from leading information technology companies including Marlabs, Infosys, UST Global, NeST and Kreara.
The conference will discuss latest developments in emerging BI areas such as predictive analytics, Big Data, mobile BI, social BI and advanced visualisations. It will also highlight the growing job opportunities for newly graduated software professionals in the Tier II and Tier III cities.
HITS: A History-Based Intelligent Transportation System IJDKP
Transportation is the driving force of any country. Today we are facing an explosion in the number of motor vehicles that affects our daily routines. Intelligent transportation systems (ITS) aim to provide efficient tools that solve traffic problems. Predicting route congestions during different day periods can help drivers choose better routes for their trips. In this paper we propose “HITS” a traffic control system that integrates moving object database techniques [30, 28] along with data warehousing techniques [15].
Our system uses historical traffic information to answer queries about moving objects on road network, and to analyze historical traffic conditions to enhance future traffic related decisions.
THIRUVANANTHAPURAM, JULY 19:
Marlabs, a Bangalore-based provider of IT services, is sponsoring a ‘Business Intelligence Technology’ conference at the Thiruvananthapuram Technopark on Friday.
The event will focus on emerging trends in Business Intelligence (BI) Technology, a Marlabs spokesman said.
It will feature eminent speakers from leading information technology companies including Marlabs, Infosys, UST Global, NeST and Kreara.
The conference will discuss latest developments in emerging BI areas such as predictive analytics, Big Data, mobile BI, social BI and advanced visualisations. It will also highlight the growing job opportunities for newly graduated software professionals in the Tier II and Tier III cities.
HITS: A History-Based Intelligent Transportation System IJDKP
Transportation is the driving force of any country. Today we are facing an explosion in the number of motor vehicles that affects our daily routines. Intelligent transportation systems (ITS) aim to provide efficient tools that solve traffic problems. Predicting route congestions during different day periods can help drivers choose better routes for their trips. In this paper we propose “HITS” a traffic control system that integrates moving object database techniques [30, 28] along with data warehousing techniques [15].
Our system uses historical traffic information to answer queries about moving objects on road network, and to analyze historical traffic conditions to enhance future traffic related decisions.
Drsp dimension reduction for similarity matching and pruning of time series ...IJDKP
Similarity matching and join of time series data streams has gained a lot of relevance in today’s world that
has large streaming data. This process finds wide scale application in the areas of location tracking,
sensor networks, object positioning and monitoring to name a few. However, as the size of the data stream
increases, the cost involved to retain all the data in order to aid the process of similarity matching also
increases. We develop a novel framework to addresses the following objectives. Firstly, Dimension
reduction is performed in the preprocessing stage, where large stream data is segmented and reduced into
a compact representation such that it retains all the crucial information by a technique called Multi-level
Segment Means (MSM). This reduces the space complexity associated with the storage of large time-series
data streams. Secondly, it incorporates effective Similarity Matching technique to analyze if the new data
objects are symmetric to the existing data stream. And finally, the Pruning Technique that filters out the
pseudo data object pairs and join only the relevant pairs. The computational cost for MSM is O(l*ni) and
the cost for pruning is O(DRF*wsize*d), where DRF is the Dimension Reduction Factor. We have
performed exhaustive experimental trials to show that the proposed framework is both efficient and
competent in comparison with earlier works.
Low-Cost Approximate and Adaptive Monitoring Techniques for the Internet of T...Demetris Trihinas
An overview of monitoring techniques used on the edge to lower big data and energy efficiency barriers for IoT. To achieve this we introduce the AdaM and ADMin frameworks. This presentation is from a talk given at the University of Cyprus (March 2017). If used, please cite one of the following:
- "Adam: An adaptive monitoring framework for sampling and filtering on IoT devices", D. Trihinas et al., IEEE BigData 2015, 10.1109/BigData.2015.7363816
- "ADMin: Adaptive Monitoring Dissemination for the Internet of Things", D. Trihinas et al., IEEE INFOCOM 2017, to appear
Enhancing Traffic Routing Inside a Network through IoT Technology & Network C...IJCNCJournal
IoT networking uses real items as stationary or mobile nodes. Mobile nodes complicate networking. Internet of Things (IoT) networks have a lot of control overhead messages because devices are mobile. These signals are generated by the constant flow of control data as such device identity, geographical positioning, node mobility, device configuration, and others. Network clustering is a popular overhead communication management method. Many cluster-based routing methods have been developed to address system restrictions. Node clustering based on the Internet of Things (IoT) protocol, may be used to cluster all network nodes according to predefined criteria. Each cluster will have a Smart Designated Node. SDN cluster management is efficient. Many intelligent nodes remain in the network. The network design spreads these signals. This paper presents an intelligent and responsive routing approach for clustered nodes in IoT networks. An existing method builds a new sub-area clustered topology. The Nodes Clustering Based on the Internet of Things (NCIoT) method improves message transmission between any two nodes. This will facilitate the secure and reliable interchange of healthcare data between professionals and patients. NCIoT is a system that organizes nodes in the Internet of Things (IoT) by grouping them together based on their proximity. It also picks SDN routes for these nodes. This approach involves selecting one option from a range of choices and preparing for likely outcomes problem addressing limitations on activities is a primary focus during the review process. Predictive inquiry employs the process of analyzing data to forecast and anticipate future events. This document provides an explanation of compact units. The Predictive Inquiry Small Packets (PISP) improved its backup system and partnered with SDN to establish a routing information table for each intelligent node, resulting in higher routing performance. Both principal and secondary roads are available for use. The simulation findings indicate that NCIoT algorithms outperform CBR protocols. Enhancements lead to a substantial 78% boost in network performance. In addition, the end-to-end latency dropped by 12.5%. The PISP methodology produces 5.9% more inquiry packets compared to alternative approaches. The algorithms are constructed and evaluated against academic ones.
Enhancing Traffic Routing Inside a Network through IoT Technology & Network C...IJCNCJournal
IoT networking uses real items as stationary or mobile nodes. Mobile nodes complicate networking. Internet of Things (IoT) networks have a lot of control overhead messages because devices are mobile. These signals are generated by the constant flow of control data as such device identity, geographical positioning, node mobility, device configuration, and others. Network clustering is a popular overhead communication management method. Many cluster-based routing methods have been developed to address system restrictions. Node clustering based on the Internet of Things (IoT) protocol, may be used to cluster all network nodes according to predefined criteria. Each cluster will have a Smart Designated Node. SDN cluster management is efficient. Many intelligent nodes remain in the network. The network design spreads these signals. This paper presents an intelligent and responsive routing approach for clustered nodes in IoT networks. An existing method builds a new sub-area clustered topology. The Nodes Clustering Based on the Internet of Things (NCIoT) method improves message transmission between any two nodes. This will facilitate the secure and reliable interchange of healthcare data between professionals and patients. NCIoT is a system that organizes nodes in the Internet of Things (IoT) by grouping them together based on their proximity. It also picks SDN routes for these nodes. This approach involves selecting one option from a range of choices and preparing for likely outcomes problem addressing limitations on activities is a primary focus during the review process. Predictive inquiry employs the process of analyzing data to forecast and anticipate future events. This document provides an explanation of compact units. The Predictive Inquiry Small Packets (PISP) improved its backup system and partnered with SDN to establish a routing information table for each intelligent node, resulting in higher routing performance. Both principal and secondary roads are available for use. The simulation findings indicate that NCIoT algorithms outperform CBR protocols. Enhancements lead to a substantial 78% boost in network performance. In addition, the end-to-end latency dropped by 12.5%. The PISP methodology produces 5.9% more inquiry packets compared to alternative approaches. The algorithms are constructed and evaluated against academic ones.
Dynamic Semantics for the Internet of Things PayamBarnaghi
Ontology Summit 2015 : Track A Session - Ontology Integration in the Internet of Things - Thu 2015-02-05,
http://ontolog-02.cim3.net/wiki/ConferenceCall_2015_02_05
SEAMLESS AUTOMATION AND INTEGRATION OF MACHINE LEARNING CAPABILITIES FOR BIG ...ijdpsjournal
The paper aims at proposing a solution for designing and developing a seamless automation and
integration of machine learning capabilities for Big Data with the following requirements: 1) the ability to
seamlessly handle and scale very large amount of unstructured and structured data from diversified and
heterogeneous sources; 2) the ability to systematically determine the steps and procedures needed for
analyzing Big Data datasets based on data characteristics, domain expert inputs, and data pre-processing
component; 3) the ability to automatically select the most appropriate libraries and tools to compute and
accelerate the machine learning computations; and 4) the ability to perform Big Data analytics with high
learning performance, but with minimal human intervention and supervision. The whole focus is to provide
a seamless automated and integrated solution which can be effectively used to analyze Big Data with highfrequency
and high-dimensional features from different types of data characteristics and different
application problem domains, with high accuracy, robustness, and scalability. This paper highlights the
research methodologies and research activities that we propose to be conducted by the Big Data
researchers and practitioners in order to develop and support seamless automation and integration of
machine learning capabilities for Big Data analytics.
SEAMLESS AUTOMATION AND INTEGRATION OF MACHINE LEARNING CAPABILITIES FOR BIG ...ijdpsjournal
The paper aims at proposing a solution for designing and developing a seamless automation and integration of machine learning capabilities for Big Data with the following requirements: 1) the ability to seamlessly handle and scale very large amount of unstructured and structured data from diversified and heterogeneous sources; 2) the ability to systematically determine the steps and procedures needed for
analyzing Big Data datasets based on data characteristics, domain expert inputs, and data pre-processing component; 3) the ability to automatically select the most appropriate libraries and tools to compute and accelerate the machine learning computations; and 4) the ability to perform Big Data analytics with high learning performance, but with minimal human intervention and supervision. The whole focus is to provide
a seamless automated and integrated solution which can be effectively used to analyze Big Data with highfrequency
and high-dimensional features from different types of data characteristics and different application problem domains, with high accuracy, robustness, and scalability. This paper highlights the research methodologies and research activities that we propose to be conducted by the Big Data researchers and practitioners in order to develop and support seamless automation and integration of machine learning capabilities for Big Data analytics.
A STUDY OF TRADITIONAL DATA ANALYSIS AND SENSOR DATA ANALYTICSijistjournal
The growth of smart and intelligent devices known as sensors generate large amount of data. These generated data over a time span takes such a large volume which is designated as big data. The data structure of repository holds unstructured data. The traditional data analytics methods well developed and used widely to analyze structured data and to limit extend the semi-structured data which involves additional processing over heads. The similar methods used to analyze unstructured data are different because of distributed computing approach where as there is a possibility of centralized processing in case of structured and semi-structured data. The under taken work is confined to analysis of both verities of methods. The result of this study is targeted to introduce methods available to analyze big data.
Asset information and data management smart railJames Nesbitt
The convergence of technology and infrastructure has the ability to transform our communities and economy, reduce emissions as well provide an opportunity for business leaders to optimise asset performance and reduce cost.
Asset information and data management will allow more precise decisions to be made to balance cost, risk and performance, supporting operational effectiveness and efficiency.
We will be addressing how the European rail sector are developing and implementing asset information strategies, managing data across multiple disparate systems and leveraging new technologies to succeed.
Predictive geospatial analytics using principal component regression IJECEIAES
Nowadays, exponential growth in geospatial or spatial data all over the globe, geospatial data analytics is absolutely deserved to pay attention in manipulating voluminous amount of geodata in various forms increasing with high velocity. In addition, dimensionality reduction has been playing a key role in high-dimensional big data sets including spatial data sets which are continuously growing not only in observations but also in features or dimensions. In this paper, predictive analytics on geospatial big data using Principal Component Regression (PCR), traditional Multiple Linear Regression (MLR) model improved with Principal Component Analysis (PCA), is implemented on distributed, parallel big data processing platform. The main objective of the system is to improve the predictive power of MLR model combined with PCA which reduces insignificant and irrelevant variables or dimensions of that model. Moreover, it is contributed to present how data mining and machine learning approaches can be efficiently utilized in predictive geospatial data analytics. For experimentation, OpenStreetMap (OSM) data is applied to develop a one-way road prediction for city Yangon, Myanmar. Experimental results show that hybrid approach of PCA and MLR can be efficiently utilized not only in road prediction using OSM data but also in improvement of traditional MLR model.
Big Data lay at the core of the strong data economy that is emerging in Europe. Although both large enterprises and SMEs acknowledge the potential of Big Data in disrupting the market and business models, this is not reflected in the growth of the data economy. The lack of trusted, secure, ethical-driven personal data platforms and privacy-aware analytics, hinders the growth of the data economy and creates concerns. The main considerations are related to the secure sharing of personal and proprietary/industrial data, and the definition of a fair remuneration mechanism that will be able to capture, produce, release and cash out the value of data, always for the benefit of all the involved stakeholders.
This webinar will focus on how such concerns that pertain to privacy, ethics and intellectual property rights can be tackled, by allowing individuals to take ownership and control of their data and share them at will, through flexible data sharing and fair compensation schemes with other entities (companies or not), as researched by the DataVaults project.
Big Data lay at the core of the strong data economy that is emerging in Europe. Although both large enterprises and SMEs acknowledge the potential of Big Data in disrupting the market and business models, this is not reflected in the growth of the data economy. The lack of trusted, secure, ethical-driven personal data platforms and privacy-aware analytics, hinders the growth of the data economy and creates concerns. The main considerations are related to the secure sharing of personal and proprietary/industrial data, and the definition of a fair remuneration mechanism that will be able to capture, produce, release and cash out the value of data, always for the benefit of all the involved stakeholders.
This webinar will focus on how such concerns that pertain to privacy, ethics and intellectual property rights can be tackled, by allowing individuals to take ownership and control of their data and share them at will, through flexible data sharing and fair compensation schemes with other entities (companies or not), as researched by the DataVaults project.
More Related Content
Similar to BDVe Webinar Series - TransformingTransport – Big Data in the Transport Domain
Drsp dimension reduction for similarity matching and pruning of time series ...IJDKP
Similarity matching and join of time series data streams has gained a lot of relevance in today’s world that
has large streaming data. This process finds wide scale application in the areas of location tracking,
sensor networks, object positioning and monitoring to name a few. However, as the size of the data stream
increases, the cost involved to retain all the data in order to aid the process of similarity matching also
increases. We develop a novel framework to addresses the following objectives. Firstly, Dimension
reduction is performed in the preprocessing stage, where large stream data is segmented and reduced into
a compact representation such that it retains all the crucial information by a technique called Multi-level
Segment Means (MSM). This reduces the space complexity associated with the storage of large time-series
data streams. Secondly, it incorporates effective Similarity Matching technique to analyze if the new data
objects are symmetric to the existing data stream. And finally, the Pruning Technique that filters out the
pseudo data object pairs and join only the relevant pairs. The computational cost for MSM is O(l*ni) and
the cost for pruning is O(DRF*wsize*d), where DRF is the Dimension Reduction Factor. We have
performed exhaustive experimental trials to show that the proposed framework is both efficient and
competent in comparison with earlier works.
Low-Cost Approximate and Adaptive Monitoring Techniques for the Internet of T...Demetris Trihinas
An overview of monitoring techniques used on the edge to lower big data and energy efficiency barriers for IoT. To achieve this we introduce the AdaM and ADMin frameworks. This presentation is from a talk given at the University of Cyprus (March 2017). If used, please cite one of the following:
- "Adam: An adaptive monitoring framework for sampling and filtering on IoT devices", D. Trihinas et al., IEEE BigData 2015, 10.1109/BigData.2015.7363816
- "ADMin: Adaptive Monitoring Dissemination for the Internet of Things", D. Trihinas et al., IEEE INFOCOM 2017, to appear
Enhancing Traffic Routing Inside a Network through IoT Technology & Network C...IJCNCJournal
IoT networking uses real items as stationary or mobile nodes. Mobile nodes complicate networking. Internet of Things (IoT) networks have a lot of control overhead messages because devices are mobile. These signals are generated by the constant flow of control data as such device identity, geographical positioning, node mobility, device configuration, and others. Network clustering is a popular overhead communication management method. Many cluster-based routing methods have been developed to address system restrictions. Node clustering based on the Internet of Things (IoT) protocol, may be used to cluster all network nodes according to predefined criteria. Each cluster will have a Smart Designated Node. SDN cluster management is efficient. Many intelligent nodes remain in the network. The network design spreads these signals. This paper presents an intelligent and responsive routing approach for clustered nodes in IoT networks. An existing method builds a new sub-area clustered topology. The Nodes Clustering Based on the Internet of Things (NCIoT) method improves message transmission between any two nodes. This will facilitate the secure and reliable interchange of healthcare data between professionals and patients. NCIoT is a system that organizes nodes in the Internet of Things (IoT) by grouping them together based on their proximity. It also picks SDN routes for these nodes. This approach involves selecting one option from a range of choices and preparing for likely outcomes problem addressing limitations on activities is a primary focus during the review process. Predictive inquiry employs the process of analyzing data to forecast and anticipate future events. This document provides an explanation of compact units. The Predictive Inquiry Small Packets (PISP) improved its backup system and partnered with SDN to establish a routing information table for each intelligent node, resulting in higher routing performance. Both principal and secondary roads are available for use. The simulation findings indicate that NCIoT algorithms outperform CBR protocols. Enhancements lead to a substantial 78% boost in network performance. In addition, the end-to-end latency dropped by 12.5%. The PISP methodology produces 5.9% more inquiry packets compared to alternative approaches. The algorithms are constructed and evaluated against academic ones.
Enhancing Traffic Routing Inside a Network through IoT Technology & Network C...IJCNCJournal
IoT networking uses real items as stationary or mobile nodes. Mobile nodes complicate networking. Internet of Things (IoT) networks have a lot of control overhead messages because devices are mobile. These signals are generated by the constant flow of control data as such device identity, geographical positioning, node mobility, device configuration, and others. Network clustering is a popular overhead communication management method. Many cluster-based routing methods have been developed to address system restrictions. Node clustering based on the Internet of Things (IoT) protocol, may be used to cluster all network nodes according to predefined criteria. Each cluster will have a Smart Designated Node. SDN cluster management is efficient. Many intelligent nodes remain in the network. The network design spreads these signals. This paper presents an intelligent and responsive routing approach for clustered nodes in IoT networks. An existing method builds a new sub-area clustered topology. The Nodes Clustering Based on the Internet of Things (NCIoT) method improves message transmission between any two nodes. This will facilitate the secure and reliable interchange of healthcare data between professionals and patients. NCIoT is a system that organizes nodes in the Internet of Things (IoT) by grouping them together based on their proximity. It also picks SDN routes for these nodes. This approach involves selecting one option from a range of choices and preparing for likely outcomes problem addressing limitations on activities is a primary focus during the review process. Predictive inquiry employs the process of analyzing data to forecast and anticipate future events. This document provides an explanation of compact units. The Predictive Inquiry Small Packets (PISP) improved its backup system and partnered with SDN to establish a routing information table for each intelligent node, resulting in higher routing performance. Both principal and secondary roads are available for use. The simulation findings indicate that NCIoT algorithms outperform CBR protocols. Enhancements lead to a substantial 78% boost in network performance. In addition, the end-to-end latency dropped by 12.5%. The PISP methodology produces 5.9% more inquiry packets compared to alternative approaches. The algorithms are constructed and evaluated against academic ones.
Dynamic Semantics for the Internet of Things PayamBarnaghi
Ontology Summit 2015 : Track A Session - Ontology Integration in the Internet of Things - Thu 2015-02-05,
http://ontolog-02.cim3.net/wiki/ConferenceCall_2015_02_05
SEAMLESS AUTOMATION AND INTEGRATION OF MACHINE LEARNING CAPABILITIES FOR BIG ...ijdpsjournal
The paper aims at proposing a solution for designing and developing a seamless automation and
integration of machine learning capabilities for Big Data with the following requirements: 1) the ability to
seamlessly handle and scale very large amount of unstructured and structured data from diversified and
heterogeneous sources; 2) the ability to systematically determine the steps and procedures needed for
analyzing Big Data datasets based on data characteristics, domain expert inputs, and data pre-processing
component; 3) the ability to automatically select the most appropriate libraries and tools to compute and
accelerate the machine learning computations; and 4) the ability to perform Big Data analytics with high
learning performance, but with minimal human intervention and supervision. The whole focus is to provide
a seamless automated and integrated solution which can be effectively used to analyze Big Data with highfrequency
and high-dimensional features from different types of data characteristics and different
application problem domains, with high accuracy, robustness, and scalability. This paper highlights the
research methodologies and research activities that we propose to be conducted by the Big Data
researchers and practitioners in order to develop and support seamless automation and integration of
machine learning capabilities for Big Data analytics.
SEAMLESS AUTOMATION AND INTEGRATION OF MACHINE LEARNING CAPABILITIES FOR BIG ...ijdpsjournal
The paper aims at proposing a solution for designing and developing a seamless automation and integration of machine learning capabilities for Big Data with the following requirements: 1) the ability to seamlessly handle and scale very large amount of unstructured and structured data from diversified and heterogeneous sources; 2) the ability to systematically determine the steps and procedures needed for
analyzing Big Data datasets based on data characteristics, domain expert inputs, and data pre-processing component; 3) the ability to automatically select the most appropriate libraries and tools to compute and accelerate the machine learning computations; and 4) the ability to perform Big Data analytics with high learning performance, but with minimal human intervention and supervision. The whole focus is to provide
a seamless automated and integrated solution which can be effectively used to analyze Big Data with highfrequency
and high-dimensional features from different types of data characteristics and different application problem domains, with high accuracy, robustness, and scalability. This paper highlights the research methodologies and research activities that we propose to be conducted by the Big Data researchers and practitioners in order to develop and support seamless automation and integration of machine learning capabilities for Big Data analytics.
A STUDY OF TRADITIONAL DATA ANALYSIS AND SENSOR DATA ANALYTICSijistjournal
The growth of smart and intelligent devices known as sensors generate large amount of data. These generated data over a time span takes such a large volume which is designated as big data. The data structure of repository holds unstructured data. The traditional data analytics methods well developed and used widely to analyze structured data and to limit extend the semi-structured data which involves additional processing over heads. The similar methods used to analyze unstructured data are different because of distributed computing approach where as there is a possibility of centralized processing in case of structured and semi-structured data. The under taken work is confined to analysis of both verities of methods. The result of this study is targeted to introduce methods available to analyze big data.
Asset information and data management smart railJames Nesbitt
The convergence of technology and infrastructure has the ability to transform our communities and economy, reduce emissions as well provide an opportunity for business leaders to optimise asset performance and reduce cost.
Asset information and data management will allow more precise decisions to be made to balance cost, risk and performance, supporting operational effectiveness and efficiency.
We will be addressing how the European rail sector are developing and implementing asset information strategies, managing data across multiple disparate systems and leveraging new technologies to succeed.
Predictive geospatial analytics using principal component regression IJECEIAES
Nowadays, exponential growth in geospatial or spatial data all over the globe, geospatial data analytics is absolutely deserved to pay attention in manipulating voluminous amount of geodata in various forms increasing with high velocity. In addition, dimensionality reduction has been playing a key role in high-dimensional big data sets including spatial data sets which are continuously growing not only in observations but also in features or dimensions. In this paper, predictive analytics on geospatial big data using Principal Component Regression (PCR), traditional Multiple Linear Regression (MLR) model improved with Principal Component Analysis (PCA), is implemented on distributed, parallel big data processing platform. The main objective of the system is to improve the predictive power of MLR model combined with PCA which reduces insignificant and irrelevant variables or dimensions of that model. Moreover, it is contributed to present how data mining and machine learning approaches can be efficiently utilized in predictive geospatial data analytics. For experimentation, OpenStreetMap (OSM) data is applied to develop a one-way road prediction for city Yangon, Myanmar. Experimental results show that hybrid approach of PCA and MLR can be efficiently utilized not only in road prediction using OSM data but also in improvement of traditional MLR model.
Big Data lay at the core of the strong data economy that is emerging in Europe. Although both large enterprises and SMEs acknowledge the potential of Big Data in disrupting the market and business models, this is not reflected in the growth of the data economy. The lack of trusted, secure, ethical-driven personal data platforms and privacy-aware analytics, hinders the growth of the data economy and creates concerns. The main considerations are related to the secure sharing of personal and proprietary/industrial data, and the definition of a fair remuneration mechanism that will be able to capture, produce, release and cash out the value of data, always for the benefit of all the involved stakeholders.
This webinar will focus on how such concerns that pertain to privacy, ethics and intellectual property rights can be tackled, by allowing individuals to take ownership and control of their data and share them at will, through flexible data sharing and fair compensation schemes with other entities (companies or not), as researched by the DataVaults project.
Big Data lay at the core of the strong data economy that is emerging in Europe. Although both large enterprises and SMEs acknowledge the potential of Big Data in disrupting the market and business models, this is not reflected in the growth of the data economy. The lack of trusted, secure, ethical-driven personal data platforms and privacy-aware analytics, hinders the growth of the data economy and creates concerns. The main considerations are related to the secure sharing of personal and proprietary/industrial data, and the definition of a fair remuneration mechanism that will be able to capture, produce, release and cash out the value of data, always for the benefit of all the involved stakeholders.
This webinar will focus on how such concerns that pertain to privacy, ethics and intellectual property rights can be tackled, by allowing individuals to take ownership and control of their data and share them at will, through flexible data sharing and fair compensation schemes with other entities (companies or not), as researched by the DataVaults project.
Big Data lay at the core of the strong data economy that is emerging in Europe. Although both large enterprises and SMEs acknowledge the potential of Big Data in disrupting the market and business models, this is not reflected in the growth of the data economy. The lack of trusted, secure, ethical-driven personal data platforms and privacy-aware analytics, hinders the growth of the data economy and creates concerns. The main considerations are related to the secure sharing of personal and proprietary/industrial data, and the definition of a fair remuneration mechanism that will be able to capture, produce, release and cash out the value of data, always for the benefit of all the involved stakeholders.
This webinar will focus on how such concerns that pertain to privacy, ethics and intellectual property rights can be tackled, by allowing individuals to take ownership and control of their data and share them at will, through flexible data sharing and fair compensation schemes with other entities (companies or not), as researched by the DataVaults project.
Intro - Three pillars for building a Smart Data Ecosystem: Trust, Security an...Big Data Value Association
Today’s data marketplaces are large, closed ecosystems that are in the hands of few established players or a consortium that decide on the rules, policies, etc.
Yet, the main barrier of the European data economy is the fact that current data spaces and marketplaces are “siloes”, without support for data exchange across their boundaries.
This webinar reveals how these boundaries can be overcome through the i3-MARKET “backplane”, which is an infrastructure able to connect all the stakeholders providing the suitable level of trust (consensus-based self-governing, auditability, reliability, verifiable credentials), security (P2P encryption, cryptographic proofs) and privacy (self-sovereign identity, zero-knowledge proof, explicit user consent).
Three pillars for building a Smart Data Ecosystem: Trust, Security and PrivacyBig Data Value Association
Today’s data marketplaces are large, closed ecosystems that are in the hands of few established players or a consortium that decide on the rules, policies, etc.
Yet, the main barrier of the European data economy is the fact that current data spaces and marketplaces are “siloes”, without support for data exchange across their boundaries.
This webinar reveals how these boundaries can be overcome through the i3-MARKET “backplane”, which is an infrastructure able to connect all the stakeholders providing the suitable level of trust (consensus-based self-governing, auditability, reliability, verifiable credentials), security (P2P encryption, cryptographic proofs) and privacy (self-sovereign identity, zero-knowledge proof, explicit user consent).
Market into context - Three pillars for building a Smart Data Ecosystem: Trus...Big Data Value Association
Today’s data marketplaces are large, closed ecosystems that are in the hands of few established players or a consortium that decide on the rules, policies, etc.
Yet, the main barrier of the European data economy is the fact that current data spaces and marketplaces are “siloes”, without support for data exchange across their boundaries.
This webinar reveals how these boundaries can be overcome through the i3-MARKET “backplane”, which is an infrastructure able to connect all the stakeholders providing the suitable level of trust (consensus-based self-governing, auditability, reliability, verifiable credentials), security (P2P encryption, cryptographic proofs) and privacy (self-sovereign identity, zero-knowledge proof, explicit user consent).
BDV Skills Accreditation - Future of digital skills in Europe reskilling and ...Big Data Value Association
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
BDV Skills Accreditation - Recognizing Data Science Skills with BDV Data Scie...Big Data Value Association
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
EIT label intro by Rroberto Prieto
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
Muluneh Oli (EIT Digital)
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
BDV Skills Accreditation - Definition and ensuring of digital roles and compe...Big Data Value Association
The objective of the workshop is to highlight the need for a pan European level skill recognition for Big Data that stimulates mobility and fulfils the definition of overarching Learning Objectives & Overarching Learning Impacts. It is also meant to get feedback on the formats that are being prepared namely, usage of Badges, Label and EIT Label for professionals.
BigDataPilotDemoDays - I BiDaaS Application to the Manufacturing Sector WebinarBig Data Value Association
The new data-driven industrial revolution highlights the need for big data technologies to unlock the potential in various application domains. To this end, BDV PPP projects I-BiDaaS, BigDataStack, Track & Know and Policy Cloud deliver innovative technologies to address the emerging needs of data operations and applications. To fully exploit the sustainability and take full advantage of the developed technologies, the projects onboarded pilots that exhibit their applicability in a wide variety of sectors. In the Big Data Pilot Demo Days, the projects will showcase the developed and implemented technologies to interested end-users from the industry as well as technology providers, for further adoption.
One of the main goals of the I-BiDaaS project is to provide a Big Data as a self-service solution that will empower the actual employees of European companies in targeted sectors (banking, manufacturing, telecom), i.e., the true decision-makers, with the insights and tools they need in order to make the right decisions in an agile way. In this big data pilot webinar, we will demonstrate in a step by step fashion the I-BiDaaS self-service solution and its application to the banking sector. In more detail, we will present an overview of the I-BiDaaS project focusing on the requirements of the CaixaBank pilot study, the I-BiDaaS architecture with its core technologies, and a step by step demo of the I-BiDaaS solution. Last but not least, we will show through CaixaBank's success story how I-BiDaaS can resolve data availability, data sharing, and breaking silos challenges in the banking domain.
At the heart of this DataBench webinar is the goal to share a benchmarking process helping European organisations developing Big Data Technologies to reach for excellence and constantly improve their performance, by measuring their technology development activity against parameters of high business relevance.
The webinar aims to provide the audience with a framework and tools to assess the performance and impact of Big Data and AI technologies, by providing real insights coming from DataBench. In addition, representatives from other projects part of the BDV PPP such as DeepHealth and They-Buy-for-You will participate to share the challenges and opportunities they have identified on the use of Big Data, Analytics, AI. The perspective of other projects that also have looked into benchmarking, such as Track&Now and I-BiDaaS will be introduced.
At the heart of this DataBench webinar is the goal to share a benchmarking process helping European organisations developing Big Data Technologies to reach for excellence and constantly improve their performance, by measuring their technology development activity against parameters of high business relevance.
The webinar aims to provide the audience with a framework and tools to assess the performance and impact of Big Data and AI technologies, by providing real insights coming from DataBench. In addition, representatives from other projects part of the BDV PPP such as DeepHealth and They-Buy-for-You will participate to share the challenges and opportunities they have identified on the use of Big Data, Analytics, AI. The perspective of other projects that also have looked into benchmarking, such as Track&Now and I-BiDaaS will be introduced.
Virtual BenchLearning - I-BiDaaS - Industrial-Driven Big Data as a Self-Servi...Big Data Value Association
At the heart of this DataBench webinar is the goal to share a benchmarking process helping European organisations developing Big Data Technologies to reach for excellence and constantly improve their performance, by measuring their technology development activity against parameters of high business relevance.
The webinar aims to provide the audience with a framework and tools to assess the performance and impact of Big Data and AI technologies, by providing real insights coming from DataBench. In addition, representatives from other projects part of the BDV PPP such as DeepHealth and They-Buy-for-You will participate to share the challenges and opportunities they have identified on the use of Big Data, Analytics, AI. The perspective of other projects that also have looked into benchmarking, such as Track&Now and I-BiDaaS will be introduced.
The problem of radicalisation is very high on the European agenda as increasing numbers of young European radicals return from Syria and use the internet to disseminate propaganda. To enable policy makers to design policies to address radicalisation effectively, Policy Cloud consortium will collect data from social media and other sources including the open-source Global Terrorism Database (GTD), the Onion City search engine which accesses data over the TOR dark web sites, and Twitter ( through Firehose). The data will be analysed using sentiment analysis and opinion mining software.
Policy Cloud Data Driven Policies against Radicalisation - Participatory poli...Big Data Value Association
The problem of radicalisation is very high on the European agenda as increasing numbers of young European radicals return from Syria and use the internet to disseminate propaganda. To enable policy makers to design policies to address radicalisation effectively, Policy Cloud consortium will collect data from social media and other sources including the open-source Global Terrorism Database (GTD), the Onion City search engine which accesses data over the TOR dark web sites, and Twitter ( through Firehose). The data will be analysed using sentiment analysis and opinion mining software.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
4. About TT
EU Horizon 2020 Big Data Value PPP Large Scale Pilot Action
• Demonstrates transformations big data has on mobility and logistics
• Part of
• 48 members - 18.7 MEUR budget - 30 months duration
4
6. TT Methodology
Rationale
• Each data set, domain, use case is different
• Diversity of data sources and infrastructures
• “No free lunch”
Each pilot
• Analytics solutions best suited for requirements and data
• Infrastructure best linked to data sources
• Big data pipelines and tools fit for purpose
Cross-cutting sharing of
• best practices, architecture patterns, KPIs, lessons learned, …
6
7. TT Methodology
3-Stage validation
and scale-up
Stage Embedding Scale of Data
Technology
Validation
Problem understanding and
validation of key solution ideas
(Historic) data pinpointing
problems and opportunities
Large-scale
Experiments
Controlled environment (not
productive environment)
Large historic and real-time data,
possibly anonymized / simulated
In-situ (on site)
trials
Trials in the field, involving actual
end-users
Real-time, live production data
complementing historic data
7
8. Transport Innovation via
Big Data
8
(IconSource:DHL/DETECON)
Efficiency
Customer
Experience
Business
Models
Smart Highways ++ ++ o
Sustainable Connected Vehicles ++ ++ o
Proactive Rail Infrastructures ++ + o
Ports as Intelligent Logistics Hubs ++ + o
Smart Airport Turnaround ++ + +
Integrated Urban Mobility ++ ++ o
Dynamic Supply Networks + + +
New
Business
Models
Improved
Operational
Efficiency
Better
Customer
Experience
Transport Domains
9. Transport Innovation via
Big Data
9
Run-time
visualization of
operations to
increase terminal
productivity
Deep Learning for
proactive transport
management
Enhanced decision
support for terminal
operators (risk and
reliability of
warnings)
Predictive analytics for proactive terminal process
management
@ duisport inland port terminal
10. Transport Innovation via
Big Data
10
Deep learning for proactive terminal management
Integrated data of container moves
(10,000 moves / month)
Data Integration
and Aggregation
(GPS / XYZ mapping;
from states to moves)
Data streams from terminal equipment
(1.3 mio states / month)
11. Transport Innovation via
Big Data
11
Deep learning for proactive terminal management
0
0,1
0,2
0,3
0,4
0,5
0,6
0,7
0 10 20 30 40 50 60 70 80 90
Diagrammtitel
nums2enoplannednopath mlp
Checkpoint [sequence prefix]
Accuracy[MCC]
RNN
MLP
Predicting Delays in Container Transport
(Recurrent Neural Networks)
+42%
Prediction Reliablity for Decision Support
(Ensemble Neural Networks / Bagging)
Cost Savings
Frequency
Cost savings of
14% on average
[Metzger & Föcker, “Predictive business process monitoring
considering reliability estimates”, CAiSE 2017]
[Metzger & Neubauer, “Considering non-sequential control flows for
process prediction with recurrent neural networks”, SEAA 2018]
12. Transport Innovation via
Big Data
12
Advanced analytics
solutions (Indra
HORUS) for improved
traffic distribution
along road corridor
Better information
and decision tools for
road users
Real-time incident
warnings based on
novel sensor
technology
Improved driving and travel experience
@ CINTRA/Ferrovial-managed highways
13. Transport Innovation via
Big Data
13
Real-time road incident warnings using novel sensor technology
Optical fiber-based sensor
(0.88 GB/sec)
Time
Distance
Filtered data
(1-5 GB/day)
Isolating Signals from Noise
(classification, adaptive
thresholds, clustering etc.)
= 3,500 virtual sensors
14. Transport Innovation via
Big Data
14
Real-time road incident warnings using novel sensor technology
Individual Mobility Pattern Detection
(trucks)
Aggregate Mobility Pattern Detection
(traffic jams)
15. Transport Innovation via
Big Data
Data-driven decision making in retailing
@ Athens International Airport
15
Advanced big data
analytics solutions
(Indra INPLAN) to
anticipate
passenger flow and
preferences
Adapt marketing to
expected passenger
typology per time
slot
Use data insights to
exploit market
niches
16. Conclusions
Opportunities
Deep learning
e.g., RNNs
Cross-sector data sharing
e.g., TT Data Portal
Challenges
Data protection
e.g., GDPR vs. IPR
Lack of skills
e.g., lack of up ½ million data
professionals in 2020 [IDC]
16
0
0,1
0,2
0,3
0,4
0,5
0,6
0,7
0 10 20 30 40 50 60 70 80 90
Diagrammtitel
nums2enoplannednopath mlp
Checkpoint
Accuracy
„deep“
„classical“
Commercial data: 68%
Personal data: 1%
17. Thank You!
17
This project received funding from the European Union’s Horizon 2020
research and innovation programme under grant agreement no. 731932
Contact:
Andreas Metzger
paluno
andreas.metzger@paluno.uni-due.de
Skype: ammetzger
http://www.transformingtransport.eu