Data Science as a Service: Intersection of Cloud Computing and Data SciencePouria Amirian
Dr. Pouria Amirian explains data science, steps in a data science workflow and show some experiments in AzureML. He also mentions about big data issues in a data science project and solutions to them.
Data Science as a Service: Intersection of Cloud Computing and Data SciencePouria Amirian
Dr. Pouria Amirian from the University of Oxford explains Data Science and its relationship with Big Data and Cloud Computing. Then he illustrates using AzureML to perform a simple data science analytics.
The widespread adoption of Information Technology systems and their
capability to trace data about process executions has made available Information
Technology data for the analysis of process executions. Meanwhile, at business
level, static and procedural knowledge, which can be exploited to analyze and rea-
son on data, is often available. In this paper we aim at providing an approach that,
combining static and procedural aspects, business and data levels and exploiting
semantic-based techniques allows business analysts to infer knowledge and use it
to analyze system executions. The proposed solution has been implemented using
current scalable Semantic Web technologies, that offer the possibility to keep the
advantages of semantic-based reasoning with non-trivial quantities of data.
Many HPC applications are massively parallel and can benefit from the spatial parallelism offered by reconfigurable logic. While modern memory technologies can offer high bandwidth, designers must craft advanced communication and memory architectures for efficient data movement and on-chip storage. Addressing these challenges requires to combine compiler optimizations, high-level synthesis, and hardware design.
In this talk, I will present challenges, solutions, and trends for generating massively parallel accelerators on FPGA for high-performance computing. These architectures can provide performance comparable to software implementations on high-end processors, and much higher energy efficiency thanks to logic customization.
Data Science as a Service: Intersection of Cloud Computing and Data SciencePouria Amirian
Dr. Pouria Amirian explains data science, steps in a data science workflow and show some experiments in AzureML. He also mentions about big data issues in a data science project and solutions to them.
Data Science as a Service: Intersection of Cloud Computing and Data SciencePouria Amirian
Dr. Pouria Amirian from the University of Oxford explains Data Science and its relationship with Big Data and Cloud Computing. Then he illustrates using AzureML to perform a simple data science analytics.
The widespread adoption of Information Technology systems and their
capability to trace data about process executions has made available Information
Technology data for the analysis of process executions. Meanwhile, at business
level, static and procedural knowledge, which can be exploited to analyze and rea-
son on data, is often available. In this paper we aim at providing an approach that,
combining static and procedural aspects, business and data levels and exploiting
semantic-based techniques allows business analysts to infer knowledge and use it
to analyze system executions. The proposed solution has been implemented using
current scalable Semantic Web technologies, that offer the possibility to keep the
advantages of semantic-based reasoning with non-trivial quantities of data.
Many HPC applications are massively parallel and can benefit from the spatial parallelism offered by reconfigurable logic. While modern memory technologies can offer high bandwidth, designers must craft advanced communication and memory architectures for efficient data movement and on-chip storage. Addressing these challenges requires to combine compiler optimizations, high-level synthesis, and hardware design.
In this talk, I will present challenges, solutions, and trends for generating massively parallel accelerators on FPGA for high-performance computing. These architectures can provide performance comparable to software implementations on high-end processors, and much higher energy efficiency thanks to logic customization.
Modeling Capabilities of Digital Twin Platforms: Old Wine in New Bottles?Daniel Lehner
Major companies such as Microsoft, Amazon, and Eclipse started to offer tooling support for Digital Twins a few years ago. Here, we investigate the modeling capabilities of these offered DT Platforms, and compare them with existing modeling approaches.
Presented at ECMFA 2022 by Jerome Pfeiffer.
Platform for Big Data Analytics and Visual Analytics: CSIRO use cases. Februa...Tomasz Bednarz
Presented at the ACEMS workshop at QUT in February 2015.
Credits: whole project team (names listed in the first slide).
Approved by CSIRO to be shared externally.
Accelerating open science and AI with automated, portable, customizable and r...Grigori Fursin
Validating experimental results from articles has finally become a norm at many systems and ML conferences. Nowadays, more than half of accepted papers pass artifact evaluation and share related code and data. Unfortunately, lack of a common experimental framework, common research methodology and common formats places an increasing burden on evaluators to validate a growing number of ad-hoc artifacts. Furthermore, having too many ad-hoc artifacts and Docker snapshots is almost as bad as not having any (!), since they cannot be easily reused, customized and built upon.
While overviewing more than 100 papers during artifact evaluation at PPoPP, CGO, PACT, Supercomputing and other conferences, we noticed that many of them use similar experimental setups, benchmarks, models, data sets, environments and platforms. This motivated us to develop Collective Knowledge (CK), an open workflow framework with a unified Python API to automate common researchers’ tasks such as detecting software and hardware dependencies, installing missing packages, downloading data sets and models, compiling and running programs, performing autotuning and co-design, crowdsourcing time-consuming experiments across computing resources provided by volunteers similar to SETI@home, applying statistical analysis and machine learning, validating results and plotting them on a common scoreboard for open and fair comparison, automatically generating interactive articles, and so on: http://cKnowledge.org.
In this presentation we will introduce CK concepts and present several real world use cases from General Motors and Arm
on collaborative benchmarking, autotuning and co-design of efficient software/hardware stacks for deep learning. We also present results and reusable CK components from the 1st ACM ReQuEST optimization tournament: http://cKnowledge.org/request. Finally, we introduce our latest initiative to create
an open repository of reusable research components and workflows to reboot and accelerate open science, quantum computing and AI!
Visual programming for hybrid user interfacesnisha thapa
This presentation is based on paper Visual Programming for Hybrid User Interfaces by Christian Pirchheim∗ Dieter Schmalstieg† Alexander Bornik‡
Graz University of Technology
Simplified Machine Learning Architecture with an Event Streaming Platform (Ap...Kai Wähner
Machine Learning is separated into model training and model inference. ML frameworks typically load historical data from a data store like HDFS or S3 to train models. This talk shows how you can completely avoid such a data store by ingesting streaming data directly via Apache Kafka from any source system into TensorFlow for model training and model inference using the capabilities of “TensorFlow I/O” add-on.
The talk compares this modern streaming architecture to traditional batch and big data alternatives and explains benefits like the simplified architecture, the ability of reprocessing events in the same order for training different models, and the possibility to build a scalable, mission-critical, real time ML architecture with muss less headaches and problems.
Key takeaways for the audience
• Scalable open source Machine Learning infrastructure
• Streaming ingestion into TensorFlow without the need for another data store like HDFS or S3 (leveraging TensorFlow I/O and its Kafka plugin)
• Stream Processing using analytic models in mission-critical deployments to act in Real Time
• Learn how Apache Kafka open source ecosystem including Kafka Connect, Kafka Streams and KSQL help to build, deploy, score and monitor analytic models
• Comparison and trade-offs between this modern streaming approach and traditional batch model training infrastructures
Case Study about BIM on GIS platform development project with the standard modelTae wook kang
To realize BIM on GIS technology for productivity of AEC industry, we should have some questions like these.
Questions
• What is the benefit from the fusion
between BIM and GIS as the viewpoint of
the public sector
• What do we should do first?
• What is the barrier to realize it?
• How to develop it?
Research and try to
• find the useful use-cases
• define the technology and the organization
including people etc
• survey the issues and define the considerations
• collaborate and research it with the institutes and
the industries
- 2013.5
Building a healthy data ecosystem around Kafka and Hadoop: Lessons learned at...Yael Garten
2017 StrataHadoop SJC conference talk. https://conferences.oreilly.com/strata/strata-ca/public/schedule/detail/56047
Description:
So, you finally have a data ecosystem with Kafka and Hadoop both deployed and operating correctly at scale. Congratulations. Are you done? Far from it.
As the birthplace of Kafka and an early adopter of Hadoop, LinkedIn has 13 years of combined experience using Kafka and Hadoop at scale to run a data-driven company. Both Kafka and Hadoop are flexible, scalable infrastructure pieces, but using these technologies without a clear idea of what the higher-level data ecosystem should be is perilous. Shirshanka Das and Yael Garten share best practices around data models and formats, choosing the right level of granularity of Kafka topics and Hadoop tables, and moving data efficiently and correctly between Kafka and Hadoop and explore a data abstraction layer, Dali, that can help you to process data seamlessly across Kafka and Hadoop.
Beyond pure technology, Shirshanka and Yael outline the three components of a great data culture and ecosystem and explain how to create maintainable data contracts between data producers and data consumers (like data scientists and data analysts) and how to standardize data effectively in a growing organization to enable (and not slow down) innovation and agility. They then look to the future, envisioning a world where you can successfully deploy a data abstraction of views on Hadoop data, like a data API as a protective and enabling shield. Along the way, Shirshanka and Yael discuss observations on how to enable teams to be good data citizens in producing, consuming, and owning datasets and offer an overview of LinkedIn’s governance model: the tools, process and teams that ensure that its data ecosystem can handle change and sustain #DataScienceHappiness.
Modeling Capabilities of Digital Twin Platforms: Old Wine in New Bottles?Daniel Lehner
Major companies such as Microsoft, Amazon, and Eclipse started to offer tooling support for Digital Twins a few years ago. Here, we investigate the modeling capabilities of these offered DT Platforms, and compare them with existing modeling approaches.
Presented at ECMFA 2022 by Jerome Pfeiffer.
Platform for Big Data Analytics and Visual Analytics: CSIRO use cases. Februa...Tomasz Bednarz
Presented at the ACEMS workshop at QUT in February 2015.
Credits: whole project team (names listed in the first slide).
Approved by CSIRO to be shared externally.
Accelerating open science and AI with automated, portable, customizable and r...Grigori Fursin
Validating experimental results from articles has finally become a norm at many systems and ML conferences. Nowadays, more than half of accepted papers pass artifact evaluation and share related code and data. Unfortunately, lack of a common experimental framework, common research methodology and common formats places an increasing burden on evaluators to validate a growing number of ad-hoc artifacts. Furthermore, having too many ad-hoc artifacts and Docker snapshots is almost as bad as not having any (!), since they cannot be easily reused, customized and built upon.
While overviewing more than 100 papers during artifact evaluation at PPoPP, CGO, PACT, Supercomputing and other conferences, we noticed that many of them use similar experimental setups, benchmarks, models, data sets, environments and platforms. This motivated us to develop Collective Knowledge (CK), an open workflow framework with a unified Python API to automate common researchers’ tasks such as detecting software and hardware dependencies, installing missing packages, downloading data sets and models, compiling and running programs, performing autotuning and co-design, crowdsourcing time-consuming experiments across computing resources provided by volunteers similar to SETI@home, applying statistical analysis and machine learning, validating results and plotting them on a common scoreboard for open and fair comparison, automatically generating interactive articles, and so on: http://cKnowledge.org.
In this presentation we will introduce CK concepts and present several real world use cases from General Motors and Arm
on collaborative benchmarking, autotuning and co-design of efficient software/hardware stacks for deep learning. We also present results and reusable CK components from the 1st ACM ReQuEST optimization tournament: http://cKnowledge.org/request. Finally, we introduce our latest initiative to create
an open repository of reusable research components and workflows to reboot and accelerate open science, quantum computing and AI!
Visual programming for hybrid user interfacesnisha thapa
This presentation is based on paper Visual Programming for Hybrid User Interfaces by Christian Pirchheim∗ Dieter Schmalstieg† Alexander Bornik‡
Graz University of Technology
Simplified Machine Learning Architecture with an Event Streaming Platform (Ap...Kai Wähner
Machine Learning is separated into model training and model inference. ML frameworks typically load historical data from a data store like HDFS or S3 to train models. This talk shows how you can completely avoid such a data store by ingesting streaming data directly via Apache Kafka from any source system into TensorFlow for model training and model inference using the capabilities of “TensorFlow I/O” add-on.
The talk compares this modern streaming architecture to traditional batch and big data alternatives and explains benefits like the simplified architecture, the ability of reprocessing events in the same order for training different models, and the possibility to build a scalable, mission-critical, real time ML architecture with muss less headaches and problems.
Key takeaways for the audience
• Scalable open source Machine Learning infrastructure
• Streaming ingestion into TensorFlow without the need for another data store like HDFS or S3 (leveraging TensorFlow I/O and its Kafka plugin)
• Stream Processing using analytic models in mission-critical deployments to act in Real Time
• Learn how Apache Kafka open source ecosystem including Kafka Connect, Kafka Streams and KSQL help to build, deploy, score and monitor analytic models
• Comparison and trade-offs between this modern streaming approach and traditional batch model training infrastructures
Case Study about BIM on GIS platform development project with the standard modelTae wook kang
To realize BIM on GIS technology for productivity of AEC industry, we should have some questions like these.
Questions
• What is the benefit from the fusion
between BIM and GIS as the viewpoint of
the public sector
• What do we should do first?
• What is the barrier to realize it?
• How to develop it?
Research and try to
• find the useful use-cases
• define the technology and the organization
including people etc
• survey the issues and define the considerations
• collaborate and research it with the institutes and
the industries
- 2013.5
Building a healthy data ecosystem around Kafka and Hadoop: Lessons learned at...Yael Garten
2017 StrataHadoop SJC conference talk. https://conferences.oreilly.com/strata/strata-ca/public/schedule/detail/56047
Description:
So, you finally have a data ecosystem with Kafka and Hadoop both deployed and operating correctly at scale. Congratulations. Are you done? Far from it.
As the birthplace of Kafka and an early adopter of Hadoop, LinkedIn has 13 years of combined experience using Kafka and Hadoop at scale to run a data-driven company. Both Kafka and Hadoop are flexible, scalable infrastructure pieces, but using these technologies without a clear idea of what the higher-level data ecosystem should be is perilous. Shirshanka Das and Yael Garten share best practices around data models and formats, choosing the right level of granularity of Kafka topics and Hadoop tables, and moving data efficiently and correctly between Kafka and Hadoop and explore a data abstraction layer, Dali, that can help you to process data seamlessly across Kafka and Hadoop.
Beyond pure technology, Shirshanka and Yael outline the three components of a great data culture and ecosystem and explain how to create maintainable data contracts between data producers and data consumers (like data scientists and data analysts) and how to standardize data effectively in a growing organization to enable (and not slow down) innovation and agility. They then look to the future, envisioning a world where you can successfully deploy a data abstraction of views on Hadoop data, like a data API as a protective and enabling shield. Along the way, Shirshanka and Yael discuss observations on how to enable teams to be good data citizens in producing, consuming, and owning datasets and offer an overview of LinkedIn’s governance model: the tools, process and teams that ensure that its data ecosystem can handle change and sustain #DataScienceHappiness.
Similar to A Model-Driven Platform for Engineering Holistic Digital Twins (20)
What's a Digital Twin - and why you need a better one?Daniel Lehner
The aim of this talk is to provide an understanding of (i) the relevance of Digital Twins, (ii) the reasons why we should buid more holistic Digital Twins, and (iii) which problems practitioners currently face when developing holistic Digital Twins.
I also present current research results on how to overcome these challenges.
Presentation given at the Lunch Time Seminar at JKU Linz on November 09, 2023.
Git-basiertes Qualitätsmonitoring von Systems Engineering ModellenDaniel Lehner
Lässt sich die Qualität von Systems Engineering messbar machen? Wenn ja, wie kann man diese Qualität visualisieren?
In diesem Vortrag geben wir Antworten aus diesen Fragen aus einem aktuellen Forschungsprojekt, und zeigen, wie durch eine Integration mit GitFlow hier außerdem Automatisierung möglich gemacht wird.
Präsentiert gemeinsam mit Konrad Wieland (LieberLieber Software GmbH) am Tag des Systems Engineering (TdSE) 2022
Towards Reactive Planning With Digital Twins and Model-Driven OptimizationDaniel Lehner
Presenting our work on enabling different Replanning strategies by combining Model-Driven Planning/Optimization with Digital Twins.
Talk given at ISOLA 2022.
Our experiences on Engineering Digital Twins, and and investigation of the tool support already offered by major companies.
Presented at the Live from Behind the Walls Event in 2022.
Model-based Detection of Runtime InconsistenciesDaniel Lehner
The final presentation (defense) of my master thesis.
The full thesis can also be found on researchgate and via the library of TU Vienna.
The implementation is also published on Github (links in presentation)
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Communications Mining Series - Zero to Hero - Session 1
A Model-Driven Platform for Engineering Holistic Digital Twins
1. Christian Doppler Laboratory for Model-Integrated Smart Production
Institute of Business Informatics – Software Engineering
Johannes Kepler University Linz
Altenberger Straße 69, Science Park 3
4040 Linz
Christian Doppler Laboratory for Model-Integrated Smart Production
CDL-MINT
A Model-Driven Platform for Engineering Holistic Digital Twins
Daniel Lehner
2. From Basic to Holistic Digital Twins
2
Digital
Twin (DT)
3. From Basic to Holistic Digital Twins
• Basic DT definition: Kritzinger [1]
• Existing DT platforms (Azure, AWS, …) support development of basic DTs
• How can we provide more functionality?
• Holistic DT: Basic DT + augment PT with additional functionality
3
[1] Kritzinger et al. "Digital Twin in manufacturing: A categorical literature review and classification." IFAC-PapersOnline 51.11 (2018): 1016-1022.
Simulation Platform
DT Platform
Automation Platform
communicates
communicates
communicates
Services
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
Simulation
Physical
Twin (PT)
Digital
Twin (DT)
communicates
4. Methodological Approach: Design Science [1]
4
[1] Wieringa, R. J. (2014). Design science methodology for information systems and software engineering. Springer.
(1) Problem investigation:
What are problems when engineering holistic DTs?
(2) Research Method
Use Cases
(4) Available Treatments
What are capabilities of current DT platforms?
What are existing MDE approaches for Digital Twins?
(5) Digital Twin Platforms
Method: Structured Review
(5) MDE4DT Approaches
Method: Systematic Mapping Study
(7) Requirements satifsfaction
Missing capabilities
• Reuse design time information
• Simulation support
• DT architecture integration support
(10) Validation
Comparing requirements satisfaction + required effort of
- New treatment
- Available treatments
(9) Prototype design and implementation
Contribution 1: Papers at ETFA 2021 and ECMFA 2022
Contribution 2: Paper at CASE and planned for T-ASE
Contribution 3: Paper planned for ECMFA 2024
(11) Research Method:
• Case Study
• Comparison Study
(8) New Treatment
DT++ Platform
5. Methodological Approach: Design Science [1]
5
[1] Wieringa, R. J. (2014). Design science methodology for information systems and software engineering. Springer.
(1) Problem investigation:
What are problems when engineering holistic DTs?
(2) Research Method
Use Cases
(4) Available Treatments
What are capabilities of current DT platforms?
What are existing MDE approaches for Digital Twins?
(5) Digital Twin Platforms
Method: Structured Review
(5) MDE4DT Approaches
Method: Systematic Mapping Study
(7) Requirements satifsfaction
Missing capabilities
• Reuse design time information
• Simulation support
• DT architecture integration support
(10) Validation
Comparing requirements satisfaction + required effort of
- New treatment
- Available treatments
(9) Prototype design and implementation
Contribution 1: Papers at ETFA 2021 and ECMFA 2022
Contribution 2: Paper at CASE and planned for T-ASE
Contribution 3: Paper planned for ECMFA 2024
(11) Research Method:
• Case Study
• Comparison Study
(8) New Treatment
DT++ Platform
6. Methodological Approach: Design Science [1]
6
[1] Wieringa, R. J. (2014). Design science methodology for information systems and software engineering. Springer.
(1) Problem investigation:
What are problems when engineering holistic DTs?
(2) Research Method
Use Cases
(4) Available Treatments
What are capabilities of current DT platforms?
What are existing MDE approaches for Digital Twins?
(5) Digital Twin Platforms
Method: Structured Review
(5) MDE4DT Approaches
Method: Systematic Mapping Study
(7) Requirements satifsfaction
Missing capabilities
• Reuse design time information
• Simulation support
• DT architecture integration support
(10) Validation
Comparing requirements satisfaction + required effort of
- New treatment
- Available treatments
(9) Prototype design and implementation
Contribution 1: Papers at ETFA 2021 and ECMFA 2022
Contribution 2: Paper at CASE and planned for T-ASE
Contribution 3: Paper planned for ECMFA 2024
(11) Research Method:
• Case Study
• Comparison Study
(8) New Treatment
DT++ Platform
7. Simulation Platform
DT Platform
Automation Platform
represents
represents
communicates
Problems in Engineering Holistic DTs
communicates communicates
7
Problem 1:
Redundant
specification
Problem 2: Manual
effort for connecting
services + simulations
Problem 3: Manual
effort for creating
architectures
Services
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
Simulation
Physical
Twin
Digital
Twin (DT)
DT Model
{twinType robot{
Attr posX, …
method gripItem(…)
…
Simulation Model
Engineering Model
Robot
posX: int
…
gripItem(…)
represents
communicates
Idle Moving
…
…
9. DT++ Platform
Automation
Platform
reuses
represents
represents
communicates
DT Workflow Model
uses
Architecture of the DT++ Platform
communicates
Engineering
Model
Robot
posX: int
…
gripItem(…)
communicates
9
Contribution 1:
Reuse design
time information
Contribution 2:
Automate integration
PT + Simulation => DT
Contribution 3:
Automate integration
DT + services => DT
architectures
DT Mega-Model
DT Model Monitor Config Model
Simulation Endpoint Config
PT Endpoint Config
Services
Monitoring Template Planning Template
…
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
Simulation
Physical
Twin
Twin
Manager
10. Contribution 1: Reuse by Transformation
10
Engineering Model Digital Twin Model Digital Twin
transformation
UML
AutomationML
Azure DTDL
Eclipse Vortolang
AWS TwinMaker-MM
Azure DT Service
Eclipse Ditto
AWS TwinMaker
Pfeiffer, J., Lehner, D., Wortmann, A., & Wimmer, M. (2022). Modeling capabilities of digital twin platforms-old wine in new
bottles?. J. Object Technol, 21(3), 3.
{
„twinTypes“: [{
„name“: „Robot“
„attributes“: [ … ],
„methods“: [ … ],
„relations“: [ … ]
],
twins: [
{„name“: „robot1“, „type“: „Robot“},
{„name“: „robot2“, „type“: „Robot“},
{„name“: „robot3“, „type“: „Robot“},
]
}
Robot
posX: int
…
gripItem(…)
robot3
robot2
robot1
configuration
robot1: Robot
robot2: Robot robot3: Robot
conformsTo
conformsTo
conformsTo
11. Contribution 1: Solution Approach
11
Pfeiffer, J., Lehner, D., Wortmann, A., & Wimmer, M. (2022). Modeling capabilities of digital twin platforms-old wine in new
bottles?. J. Object Technol, 21(3), 3.
Profile DTUML
«Stereotype»
DTProperty
Boolean isTimeSeries
String unit
String type
«Stereotype»
AWS_Property
Boolean isExternalId
Boolean isStoredLocally
List<String> allowedValues
String quantity
«Stereotype»
VortoLang_Property
Boolean isFault
String constraintRule
«Stereotype»
DTDL_Property
«Metaclass»
Property
Class type
Int lower
Int upper
Boolean readOnly
Property redefinedProperty
Value defaultValue
rule class2interface {
from
cl: UML!Class
to
interf: DTDL!Interface ()
do{
interf.displayName <- cl.name;
if(cl.ownedComment.notEmpty()){
interf.comment <- cl.ownedComment.first().body;
}
interf.contents <- Set{};
-- Create Commands
for(op in cl.ownedOperation) {
interf.contents <- interf.contents->including(thisModule.newCommand(op));
}
-- Create Components
-- …
-- Create Properties, Telemetries, Relationships and Components
-- …
}
}
}
UML profile for missing concepts ATL transformation for reuse
12. DT++ Platform
Automation
Platform
reuses
represents
represents
communicates
DT Workflow Model
uses
Architecture of the DT++ Platform
communicates
Engineering
Model
Robot
posX: int
…
gripItem(…)
communicates
12
Contribution 1:
Reuse design
time information
Contribution 2:
Automate integration
PT + Simulation => DT
Contribution 3:
Automate integration
DT + services => DT
architectures
DT Mega-Model
DT Model Monitor Config Model
Simulation Endpoint Config
PT Endpoint Config
Services
Monitoring Template Planning Template
…
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
Simulation
Physical
Twin
Twin
Manager
13. Contribution 2.1: Endpoints
13
Lehner, D., Gil, S., Mikkelsen, P., Larsen, P. & Wimmer, M. (2023). An Architectural Extension for Digital Twin Platforms to Leverage Behavioral Models, Proc.
of CASE.
<<abstract>>
Endpoint
Physical
Twin
Simulation
getAttributeValue(...): Object
executeOperation(...): void
14. DT Mega-Model
Contribution 2.2: TwinManager
14
Lehner, D., Gil, S., Mikkelsen, P., Larsen, P. & Wimmer, M. (2023). An Architectural Extension for Digital Twin Platforms to Leverage Behavioral Models, Proc.
of CASE.
Services Monitoring Template Planning Template
…
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
TwinManager
create/read/updateEndpoint(…)
executeOperation(List<Endpoint>, Time, …)
…
<<abstract>>
Endpoint
Physical
Twin
getAttributeValue(...): Object
executeOperation(...): void
communicates
communicates
Simulation
uses
15. Endpoint Config Model
WeBotEndpoint wbEndpoint implementing Robot at 137.654.1
gripAt(x, y, z) at /grip
data: current_x:curX, current_y:curY, current_z:curZ
Endpoint Config Model
ROSEndpoint rosEndpoint implementing Robot at 137.654.2
gripAt(x, y, z) at /grip
data: x:curX, y:curY, z:curZ
Service Config Model
Service Monitor at 127.0.0.1
in: monitorForDeviation(actual, simulation, curX), out: Deviation: bool
implements
implements
DT Mega-Model
twin niryoRobot: Robot
twins: actual at rosEndpoint, simulation at wbEndpoint
services: Monitor checkPosX
DT Model
Type Robot
int curX, curY, curZ
int targetX, targetY, targetZ
int torque
moveGripperTo(x, y, z)
gripAt(x, y, z)
placeAt(x, y, z)
uses
Contribution 2.3: DT Mega-Model
15
16. DT++ Platform
Automation
Platform
reuses
represents
represents
communicates
DT Workflow Model
uses
Architecture of the DT++ Platform
communicates
Engineering
Model
Robot
posX: int
…
gripItem(…)
communicates
16
Contribution 1:
Reuse design
time information
Contribution 2:
Automate integration
PT + Simulation => DT
Contribution 3:
Automate integration
DT + services => DT
architectures
DT Mega-Model
DT Model Monitor Config Model
Simulation Endpoint Config
PT Endpoint Config
Services
Monitoring Template Planning Template
…
SC-based
Planner
AI-based
Planner
Time-based
Monitor
Event-based
Monitor
Simulation
Physical
Twin
Twin
Manager
17. Contribution 3.1: DT Module Templates
17
Monitoring Template
Monitor
Monitor Config
Language
Event-based Monitor Module
Event-based
Monitor
Event
Monitoring
Language
LC
SC
LC
SC
Idea 1: Wrap software and language components into common DT modules
• Use MDE techniques to automate integration of different components
Idea 2: Provide templates of DT modules for reusability
• Integration is lifted on the reference architecture level
18. Contribution 3.2: DT Workflow Language
• Extends structural information of DT Mega-Model
• Describes interactions between serivces and DTs
18
DT Workflow Language
monitorForDeviation (actual, simulation, param):
every 10 seconds:
def = actual.getAttributeValue(param) –
simulation.getAttributeValue(param)
response def <= 1
Endpoint Model
Endpoint Model
Service Config Model
DT Mega-Model
DT Model
uses
20. Evaluation Plan: Case Study [1] (1/2)
Research Questions
▪ RQ1: Reuse of engineering models in DT++ Platform? (Contribution 1)
▪ RQ2: Effort of service connection in DT++ Platform? (Contribution 2)
▪ RQ3: Effort of architecture integration using DT++ Platform? (Contribution 3)
Considered Cases
20
Runeson, P., & Höst, M. (2009). Guidelines for conducting and reporting case study research in software engineering. Empirical software engineering, 14,
131-164.
Air Quality
Measurement
Stack Balancing Incubator Temperature
Managmeent
Smart Room
1
7
3
Stack
1
Stack
2
Stack
3
21. Initial Evaluation Results
21
RQ1 (Reuse potential) [1]
▪ Some adaptations required to existing languages
▪ Transformation reduces effort for setup + evolution by 50 %
RQ2 (Effort of service connection) [2]
▪ Effort for connecting services to different endpoints reduced by > 50 %
▪ Effort for switching the underlying DT of a service reduced by ~40 %
RQ3 (Effort of architecture integration)
▪ Increased scalability for integrating similar
services into a high number of different architectures
▪ Higher initial effort than baseline approaches
[1] Lehner, D., Sint, S., Vierhauser, M., Narzt, W., & Wimmer, M. (2021, September). AML4DT: a model-driven framework for developing and maintaining digital twins with
AutomationML. In 2021 26th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA) (pp. 1-8). IEEE.
[2] Pfeiffer, J., Lehner, D., Wortmann, A., & Wimmer, M. (2022). Modeling capabilities of digital twin platforms-old wine in new bottles?. J. Object Technol, 21(3), 3.
Effort 50 %
Effort 40-50 %
Initial Effort
Scalability
22. Current State + Ongoing Work
Problem Investigation + Existing Solutions
▪ DT Airquality Exemplar (MODDIT‘21)
▪ Reactive Planning using DTs (ISOLA‘22)
▪ Investigation of existing platforms (IEEE Software‘22)
▪ Systematic Mapping Study submitted to SOSYM
Solution + Validation
▪ Design of reuse mechanism (ETFA’21 and ECMFA‘22)
▪ Design of TwinManager (CASE‘23)
Ongoing Work
▪ Publication of Integration Method + DT Workflow Language (RQ3)
▪ Extension of existing evaluations to further use cases for generalizability
22
23. Conclusion + Potential Societal Impact
23
[1] Heithoff, M., Hellwig, A., Michael, J., & Rumpe, B. (2023). Digital twins for sustainable software systems. In Proc. of GREENS
Problems in Engineering holistic DTs
1. Redundancy
2. High effort for integrating components
Contributions of the DT++ Platform
1. Reuse
2. Automation
Potential impact on society
• Increased DT adoption enables innvoation
• Sustainable product development [1]
• Optimized use of human workforce
25. Evaluation Plan: Case Study [1] (2/2)
Metrics
25
Runeson, P., & Höst, M. (2009). Guidelines for conducting and reporting case study research in software engineering. Empirical software engineering, 14,
131-164.
Endpoint Model
WeBotEndpoint wbEndpoint implementing Robot at 137.654.1
gripAt(x, y, z) at /grip
data: current_x:curX, current_y:curY, current_z:curZ
1
2
3
Lines of Code
Twin Manager
Monitoring Template
# Operation Calls getAttributeValue(…)
TOD
O
metri
k für
jede
RQ
26. Conclusion + Potential Societal Impact
▪ Engineering holistic DTs is cumbersome and error-prone
▪ DT++ platform helps reduce this effort by
▪ Reusing existing design time information
▪ Automating the integration of PT and simulations into a holistic DT
▪ Automating the integration of DTs and services into DT architectures
26
TODO: Impact on Society
- Hoher Aufwand = weniger
Innovation möglich
- Weniger DTs = Potenziale bzgl.
Digitalisierung und CO2-
Messung weniger nutzbar
TODO: Too much text:
vllt Gegenüberstellung
Current State + DT++ mit
Problems + Contrubionts
darunter?
TODO: hier Referenz von
Keynote Judith reingeben
27. Contribution 3.1: A template-based method for DT architecture integration
27
Digital Twin
Planner DTPlatform
DT Product Line
Planner
DT Platform
Azure
Timed
Statecharts
Basic
Statecharts
AWS IoT
A
DTPlatform: Azure
Planner: Basic Statecharts
Azure
Basic
Statecharts
DTDL
Model
uses
SC
Models
uses
E
Phase 3: Architecture/Product Line Configuration
DT module
DTPlatform.data ->
Planner.data
…
Azure DT
SC Model
Execution Engine
DT service
data cmd
data
Timed
Statecharts
B
Basic
Statecharts
plan
Phase 2: Reference Architecture Definition
Planner
DT Platform
DTPlatform: …
Azure:
Azure.state -> DTPlatform.data
…
Phase 4: DT Architecture Generation
uses uses
generate
state
data
Azure
AWS IoT
D
uses
uses
DT module
wrapper
Connects DT templates
DT module and
template definition
model
Product Line
configuration
Reference architecture
definition model
DTPlatform Planner
DT template
bridge
C
F
G E
A
B B B
Phase 1: Component Definition
DT template
G
E
E
current
28. Holistic Digital Twins (DTs)
Holistic DT: provide functionality based on interactions with PT + simulation
- common functionality: anomaly detection, virtual experimentation, planing, …
Thesis goal
• reduce the effort for engineering holistic DTs
• by employing MDE techniques
• in order to make it easier for academics to adopt DTs
28
Holistic DT
- interactions with PT + simulation
- provide functionality
- anomaly detection
- virtual experimentation
- planning
Physical
Twin
Digital Twin
communicates
29. Evaluation Plan: Case Study [1]
Research Questions
▪ RQ1: Reuse potential (Contribution 1)
▪ RQ2: Effort of service connection (Contribution 2)
▪ RQ3: Effort of architecture integration (Contribution 3)
Cases
▪ Smart Room, incl. Air Quality Measurement
▪ Reactive Planning for Stack Balancing
▪ Incubator DT
Metrics
Count number of…
▪ … High-level operations
▪ … Change operations
▪ … Lines of Code
required to perform predefined scenarios
29
Runeson, P., & Höst, M. (2009). Guidelines for conducting and reporting case study research in software engineering. Empirical software engineering, 14,
131-164.
30. Automation
Platform
DT++ Platform
reuses
DT Mega-Model
represents describes
Monitoring Service
represents
DT Model Monitor Config Model
Simulation Endpoint
Configuration
communi
cates
PT Endpoint
Configuration
uses
Twin Manager
Contribution 2.1: Architectural Extensions of existing DT Platforms
Lehner, D., Gil, S., Mikkelsen, P., Larsen, P. & Wimmer, M. (2023). An Architectural Extension for Digital Twin Platforms to Leverage Behavioral Models, Proc.
of CASE.
Digital Twin
Physical
Twin
Simulation
communicates
Robot
posX: int
…
gripItem(…)
Engineering Model
communi
cates
Contribution 2.1: Endpoints
Abstraction of PT + Simulations
Contribution 2.2: TwinManager
Interface to PT + Simulations
Contribution 2.3: DT Mega-Model
Single source of configuration
30
31. Evaluation Plan: Case Study [1]
Research Questions
▪ RQ1: Reuse potential of DT++ Platform compared to existing platforms? (Contribution 1)
▪ RQ2: Effort of service connection in DT++ Platform, compared to existing platforms? (Contribution 2)
▪ RQ3: Effort of architecture integration using DT++ Platform, compared to existing platforms? (Contribution 3)
Cases
▪ Smart Room
▪ Air Quality Measurement
▪ Reactive Planning for Stack Balancing
▪ Heatbed Incubator
Metrics
Count number of…
▪ … High-level operations
▪ … Change operations
▪ … Lines of Code
required to perform predefined scenarios
31
Runeson, P., & Höst, M. (2009). Guidelines for conducting and reporting case study research in software engineering. Empirical software engineering, 14,
131-164.