In this presentation you will learn about our proprietary methodology to migrate CA 2E to PLEX. The process has three primary phases sandwiched between a time of preparation and the final project completion. Find out why Fortune 500 companies used CM First to spear head the project.
Cm First also compares the CA 2E model functionality with Plex model functionality. Understanding the several areas of similarity will help with the modernization process.
This document introduces BPMN 2.0 and how jBPM5 brings its power to business processes. BPMN 2.0 is a standard notation for modeling business processes that is understandable to various roles. It extends BPMN 1.2 with an XML-based notation, execution semantics, extensibility, choreography support, and conformance classes. jBPM5 supports the key BPMN 2.0 elements like activities, gateways, events, and their types, enabling the modeling of complex business processes.
This document discusses jBPM5, an open-source business process management project. It provides an overview of key BPM concepts like workflows and processes. jBPM5 aims to make processes ready for the future through support for BPMN 2.0, flexible processes, integration of rules and events. The roadmap includes adding features like business activity monitoring and service orchestration to the 5.x releases.
KrisVerlaenen gave a presentation on using business process management (BPM) to develop process-driven applications. BPM provides visibility, monitoring, and flexibility to help applications evolve over time. jBPM supports developing applications from embedded processes to remote services and enables customizing processes for different domains. The presentation demonstrated how to author and monitor processes using jBPM.
EPAM DevOps community meetup: Building CI/CD for microservice architectureKatherine Golovinova
Building small microservice based architecture is easy. When architecture consists of large set of microservices, you will face a number of challenges.
We will share our approach how to build CI/CD for a complex build, test and deploy procedures for microservice based architecures.
Building responsive and flexible applications with BPMKris Verlaenen
The presentation discusses building flexible applications using business process management (BPM) and case management. It provides an overview of case management, how it differs from structured processes by allowing ad-hoc choices and dynamic modifications. It also outlines how BPM can provide out-of-the-box building blocks to help develop applications that offer choice, flexibility and the ability to evolve over time through customization. Finally, it briefly discusses horizontal scaling of databases for BPM applications.
1. The document summarizes a migration of Partners Enterprise's Caché application from a Windows environment to HP-UX servers, including three phases: migrating the database tier, application tier, and disaster recovery configuration.
2. Key drivers for the migration were improving availability, scalability, and performance. The Windows platform had higher downtime and limitations on scalability compared to HP-UX.
3. The migration involved moving over 3.5 terabytes of data, upgrading software, addressing challenges around file access and printing between Windows and HP-UX, and implementing high availability clustering and disaster recovery.
Streamline your business processes and enhance productivity by using jBPMKris Verlaenen
The document is a presentation about jBPM 5 from Red Hat. It introduces jBPM 5 as a lightweight, embeddable business process management engine based on BPMN 2.0. It discusses jBPM 5's core engine, Eclipse plugin, Guvnor designer console and installer. The presentation demonstrates these components and outlines the roadmap for upcoming jBPM 5 releases. It also compares jBPM 5 to prior versions and provides resources for learning more about jBPM 5.
This document provides an overview and roadmap for case management, form modeling, and process designer improvements in the Red Hat JBoss BPM Suite product. Key points include:
1) Separating the business-central UI from the runtime kie-server and adding case management capabilities to the core engine.
2) Developing features for authoring, managing, and viewing cases at runtime including a case modeler, custom applications, and improved designer.
3) Addressing customer requests for advanced task routing, process/task administration APIs, and more fine-grained security.
This document introduces BPMN 2.0 and how jBPM5 brings its power to business processes. BPMN 2.0 is a standard notation for modeling business processes that is understandable to various roles. It extends BPMN 1.2 with an XML-based notation, execution semantics, extensibility, choreography support, and conformance classes. jBPM5 supports the key BPMN 2.0 elements like activities, gateways, events, and their types, enabling the modeling of complex business processes.
This document discusses jBPM5, an open-source business process management project. It provides an overview of key BPM concepts like workflows and processes. jBPM5 aims to make processes ready for the future through support for BPMN 2.0, flexible processes, integration of rules and events. The roadmap includes adding features like business activity monitoring and service orchestration to the 5.x releases.
KrisVerlaenen gave a presentation on using business process management (BPM) to develop process-driven applications. BPM provides visibility, monitoring, and flexibility to help applications evolve over time. jBPM supports developing applications from embedded processes to remote services and enables customizing processes for different domains. The presentation demonstrated how to author and monitor processes using jBPM.
EPAM DevOps community meetup: Building CI/CD for microservice architectureKatherine Golovinova
Building small microservice based architecture is easy. When architecture consists of large set of microservices, you will face a number of challenges.
We will share our approach how to build CI/CD for a complex build, test and deploy procedures for microservice based architecures.
Building responsive and flexible applications with BPMKris Verlaenen
The presentation discusses building flexible applications using business process management (BPM) and case management. It provides an overview of case management, how it differs from structured processes by allowing ad-hoc choices and dynamic modifications. It also outlines how BPM can provide out-of-the-box building blocks to help develop applications that offer choice, flexibility and the ability to evolve over time through customization. Finally, it briefly discusses horizontal scaling of databases for BPM applications.
1. The document summarizes a migration of Partners Enterprise's Caché application from a Windows environment to HP-UX servers, including three phases: migrating the database tier, application tier, and disaster recovery configuration.
2. Key drivers for the migration were improving availability, scalability, and performance. The Windows platform had higher downtime and limitations on scalability compared to HP-UX.
3. The migration involved moving over 3.5 terabytes of data, upgrading software, addressing challenges around file access and printing between Windows and HP-UX, and implementing high availability clustering and disaster recovery.
Streamline your business processes and enhance productivity by using jBPMKris Verlaenen
The document is a presentation about jBPM 5 from Red Hat. It introduces jBPM 5 as a lightweight, embeddable business process management engine based on BPMN 2.0. It discusses jBPM 5's core engine, Eclipse plugin, Guvnor designer console and installer. The presentation demonstrates these components and outlines the roadmap for upcoming jBPM 5 releases. It also compares jBPM 5 to prior versions and provides resources for learning more about jBPM 5.
This document provides an overview and roadmap for case management, form modeling, and process designer improvements in the Red Hat JBoss BPM Suite product. Key points include:
1) Separating the business-central UI from the runtime kie-server and adding case management capabilities to the core engine.
2) Developing features for authoring, managing, and viewing cases at runtime including a case modeler, custom applications, and improved designer.
3) Addressing customer requests for advanced task routing, process/task administration APIs, and more fine-grained security.
JBossOneDayTalk 2011: Using jBPM to bring more power to your business processesKris Verlaenen
1) jBPM5 is an open-source business process management project that offers a generic process engine supporting native BPMN 2.0 execution for developers and business users.
2) The core engine is a lightweight, embeddable, and extensible workflow engine in Java that supports BPMN 2.0 processes, rules integration, and various extra features like social, cloud, and mobile capabilities.
3) The roadmap includes enhancements to simulation/testing, no-code tooling, domain-specific capabilities, and improved form building and BPMN editing.
This document provides an overview of the architecture and functionality of Control-M, a mainframe job scheduling software. It describes the key components of Control-M including the Control-M Agent, Server, and Enterprise Management console. It also summarizes how Control-M is used to define, schedule, execute, monitor and manage jobs across platforms based on calendars, conditions, dependencies, resources and results.
Building advanced case-driven applicationsKris Verlaenen
The document discusses building case-driven applications using the jBPM framework. It describes how jBPM allows for both structured processes and adaptive cases that can be unpredictable. It highlights key foundations of jBPM including flexible BPMN 2.0 modeling, meta-driven extensions, and integrated processes, rules, and events. The document demonstrates how jBPM can be used to rapidly develop customized case-driven applications.
Srivalli Aparna - The Blueprints to SuccessNeotys_Partner
Since its beginning, the Performance Advisory Council aims to promote engagement between various experts from around the world, to create relevant, value-added content sharing between members. For Neotys, to strengthen our position as a thought leader in load & performance testing. During this event, 12 participants convened in Chamonix (France) exploring several topics on the minds of today’s performance tester such as DevOps, Shift Left/Right, Test Automation, Blockchain and Artificial Intelligence.
Empowering Business Users with Process Management ToolsKris Verlaenen
Demo of the various tools that are part of the upcoming jBPM6 release (and BPMS6 platform) that can be used by business users (business analysts, end users, etc.) to model, manage and analyze their business processes.
This document provides an overview and roadmap for case management capabilities in the jBPM Project, including:
1. Separating the business-central UI from the kie-server runtime and enabling custom applications.
2. Upcoming improvements to case management in the 7.x releases, including enhancements to the core engine, authoring tools, and runtime views.
3. Customer requests for additional process/task administration APIs and more advanced task routing strategies.
Enterprise Orchestration at UT (via Stonebranch) FINAL VERSION v3Jennifer Hubley
This document proposes a B2B Gateway managed file transfer service to replace UT's existing unmanaged file transfer service. The gateway would support both internal and external file transfers in a fault-tolerant manner using Stonebranch software. It describes problems with the current approach, assumptions that did not match reality, consumer needs and personas. Key features are managed internal file transfers with familiar protocols and a phased approach is outlined, starting with internal transfers and later adding external integrations and additional protocols. Success is defined in terms of new consumers, task reliability and usage metrics.
Automate workflows with leading open-source BPMKris Verlaenen
This document provides an overview of the JBoss BPM Suite 6.0 for automating workflows using open source business process management. It describes the core components and features of BPMS 6, including the execution server, development repository, knowledge management, integration with Fuse Service Works, and the business resource planner. The document also outlines the BPM lifecycle involving modeling, deploying, monitoring, and analyzing processes and tasks. It provides examples of using the runtime manager, REST API, and remote client to execute processes and tasks on the clustered execution server.
The document discusses jBPM5, an open source business process management platform that uses the BPMN 2.0 standard to model and execute business processes. It has a lightweight engine, full lifecycle support, and integrates with rules and events. The document outlines jBPM5's key features, architecture, roadmap for adaptive processes, and links for more information.
In these Masterclass slides, dedicated to IRIS Service Director customers, we take a detailed look at three areas, designed to save you time and money every day:
- Crystal Director module - used to automate email, document and report production.
- Scheduler integration with Outlook calendars.
- Web module for sub-contractors and for customers.
Accelerating time to delivery: Modernizing Application DevelopmentMicro Focus
The document discusses tools for accelerating application development including project templates, code snippets, cross-platform development, and application workflow modeling. It also covers Micro Focus's Enterprise Test Server for scalable enterprise testing without consuming mainframe resources.
From Domain-Specific Process Design to Execution and BackAdrian Mos
EclipseCON Europe 2014 talk during BPM Day.
"From Domain-Specific Process Design to Execution and Back" shows how to leverage monitoring data to enhance the understanding of the utilization of business concepts used in the definition of business processes.
Camunda for Modern Web Applications by Corinna Cohn and Sowmya Raghunathancamunda services GmbH
An overview of the architecture for an insurance claims loss notice application at Indiana Farm Bureau, using headless Camunda BPM to maintain application state and to drive a multi-screen and completely stateless Angular application. Workflow modeling can take users through multiple paths, and the sequence can be changed without altering the Angular application. This architecture decouples UI concerns from the BPM.
This document discusses event-driven business process management using jBPM and Drools. It provides an example of how software development tools and infrastructure can be combined with jBPM 5 and Drools to gain real-time understanding of processes and increase agility. Traditional BPM systems struggle with change, complexity, and flexibility, but event-driven BPM addresses these issues. The document also outlines a logistics company use case and how a rules-based solution could dynamically manage shipments and routing.
Steven King has over 20 years of experience in data processing and batch scheduling. He is currently a Production Control Analyst at Princess Cruises, where he is responsible for day-to-day batch workloads and file transfers between the company's fleet of ships, departments, sister companies, and vendors. Previously he worked as a Batch Scheduler and Computer Operator, gaining experience with scheduling packages like CA7, CA11, Zeke, Zebb, PeopleSoft, and Control-M.
The document describes two job openings for a 5-month project in Arnhem, Netherlands. The project involves adapting existing software systems to implement new functionality for transportation payment devices. The first position is for a software architect responsible for design and oversight. The second is for a C++ software engineer to review designs and adapt existing code. Both roles require experience with AIX, C++, DB2, and transaction processing systems.
Piyush Kumar Gupta is seeking career opportunities as a Linux System Administrator or Support Engineer in Delhi/NCR. He has over 7 years of experience in Linux administration and support roles. He is RHCE certified and has experience working with Red Hat Enterprise Linux 5/6/7 and CentOS 6. He is currently employed as a Linux Specialist at HCL Technologies and has previously worked at Dell International Services, Tetra Information Services, and INS E Solution.
This document discusses managing projects in new technologies and the impact on management and planning. It provides a brief history of technology from 1960 to the present day. It then covers various topics related to managing technologies including roles and responsibilities, project lifecycles, planning constraints, solution modeling, risk profiles, dependencies, and architecture types. The document discusses concepts like cloud computing, NoSQL databases, analytics, mobile applications, and emerging trends like the internet of things, big data, gamification, and social media.
Expert guidance on migrating from magento 1 to magento 2James Cowie
Migrating a Magento site is not just about code and data. Commerce platforms evolve over time and your Magento 1 solution is likely different today compared to the day you launched. Planning a successful migration means understanding what you have and where you are going before you can begin. In this session, architects from the Magento Expert Consulting Group will lay out best practices for defining your migration strategy, and share tips and techniques for code and data migration.
Did you know the average age of the top four banks is 175 years old? And those four banks are the end result of merging or acquiring 35 separate companies over the last 40 years? This means lots of different cultures, tech stacks, apps, processes, and controls – and SCRIPTS!!
Join Chris Nowak, Principal at FlowStates, and Avantika Mathur, ElectricFlow Product Manager, as they share hard-won lessons and emerging patterns in DevOps and release orchestration that can help any regulated organizations climb out of scripting hell. You’ll hear:
1) Why it’s important to prioritize delivery improvement across all your teams
2) How the theory of constraints is limiting your success
3) And how simple things like version control for artifacts can help improve success rates!
JBossOneDayTalk 2011: Using jBPM to bring more power to your business processesKris Verlaenen
1) jBPM5 is an open-source business process management project that offers a generic process engine supporting native BPMN 2.0 execution for developers and business users.
2) The core engine is a lightweight, embeddable, and extensible workflow engine in Java that supports BPMN 2.0 processes, rules integration, and various extra features like social, cloud, and mobile capabilities.
3) The roadmap includes enhancements to simulation/testing, no-code tooling, domain-specific capabilities, and improved form building and BPMN editing.
This document provides an overview of the architecture and functionality of Control-M, a mainframe job scheduling software. It describes the key components of Control-M including the Control-M Agent, Server, and Enterprise Management console. It also summarizes how Control-M is used to define, schedule, execute, monitor and manage jobs across platforms based on calendars, conditions, dependencies, resources and results.
Building advanced case-driven applicationsKris Verlaenen
The document discusses building case-driven applications using the jBPM framework. It describes how jBPM allows for both structured processes and adaptive cases that can be unpredictable. It highlights key foundations of jBPM including flexible BPMN 2.0 modeling, meta-driven extensions, and integrated processes, rules, and events. The document demonstrates how jBPM can be used to rapidly develop customized case-driven applications.
Srivalli Aparna - The Blueprints to SuccessNeotys_Partner
Since its beginning, the Performance Advisory Council aims to promote engagement between various experts from around the world, to create relevant, value-added content sharing between members. For Neotys, to strengthen our position as a thought leader in load & performance testing. During this event, 12 participants convened in Chamonix (France) exploring several topics on the minds of today’s performance tester such as DevOps, Shift Left/Right, Test Automation, Blockchain and Artificial Intelligence.
Empowering Business Users with Process Management ToolsKris Verlaenen
Demo of the various tools that are part of the upcoming jBPM6 release (and BPMS6 platform) that can be used by business users (business analysts, end users, etc.) to model, manage and analyze their business processes.
This document provides an overview and roadmap for case management capabilities in the jBPM Project, including:
1. Separating the business-central UI from the kie-server runtime and enabling custom applications.
2. Upcoming improvements to case management in the 7.x releases, including enhancements to the core engine, authoring tools, and runtime views.
3. Customer requests for additional process/task administration APIs and more advanced task routing strategies.
Enterprise Orchestration at UT (via Stonebranch) FINAL VERSION v3Jennifer Hubley
This document proposes a B2B Gateway managed file transfer service to replace UT's existing unmanaged file transfer service. The gateway would support both internal and external file transfers in a fault-tolerant manner using Stonebranch software. It describes problems with the current approach, assumptions that did not match reality, consumer needs and personas. Key features are managed internal file transfers with familiar protocols and a phased approach is outlined, starting with internal transfers and later adding external integrations and additional protocols. Success is defined in terms of new consumers, task reliability and usage metrics.
Automate workflows with leading open-source BPMKris Verlaenen
This document provides an overview of the JBoss BPM Suite 6.0 for automating workflows using open source business process management. It describes the core components and features of BPMS 6, including the execution server, development repository, knowledge management, integration with Fuse Service Works, and the business resource planner. The document also outlines the BPM lifecycle involving modeling, deploying, monitoring, and analyzing processes and tasks. It provides examples of using the runtime manager, REST API, and remote client to execute processes and tasks on the clustered execution server.
The document discusses jBPM5, an open source business process management platform that uses the BPMN 2.0 standard to model and execute business processes. It has a lightweight engine, full lifecycle support, and integrates with rules and events. The document outlines jBPM5's key features, architecture, roadmap for adaptive processes, and links for more information.
In these Masterclass slides, dedicated to IRIS Service Director customers, we take a detailed look at three areas, designed to save you time and money every day:
- Crystal Director module - used to automate email, document and report production.
- Scheduler integration with Outlook calendars.
- Web module for sub-contractors and for customers.
Accelerating time to delivery: Modernizing Application DevelopmentMicro Focus
The document discusses tools for accelerating application development including project templates, code snippets, cross-platform development, and application workflow modeling. It also covers Micro Focus's Enterprise Test Server for scalable enterprise testing without consuming mainframe resources.
From Domain-Specific Process Design to Execution and BackAdrian Mos
EclipseCON Europe 2014 talk during BPM Day.
"From Domain-Specific Process Design to Execution and Back" shows how to leverage monitoring data to enhance the understanding of the utilization of business concepts used in the definition of business processes.
Camunda for Modern Web Applications by Corinna Cohn and Sowmya Raghunathancamunda services GmbH
An overview of the architecture for an insurance claims loss notice application at Indiana Farm Bureau, using headless Camunda BPM to maintain application state and to drive a multi-screen and completely stateless Angular application. Workflow modeling can take users through multiple paths, and the sequence can be changed without altering the Angular application. This architecture decouples UI concerns from the BPM.
This document discusses event-driven business process management using jBPM and Drools. It provides an example of how software development tools and infrastructure can be combined with jBPM 5 and Drools to gain real-time understanding of processes and increase agility. Traditional BPM systems struggle with change, complexity, and flexibility, but event-driven BPM addresses these issues. The document also outlines a logistics company use case and how a rules-based solution could dynamically manage shipments and routing.
Steven King has over 20 years of experience in data processing and batch scheduling. He is currently a Production Control Analyst at Princess Cruises, where he is responsible for day-to-day batch workloads and file transfers between the company's fleet of ships, departments, sister companies, and vendors. Previously he worked as a Batch Scheduler and Computer Operator, gaining experience with scheduling packages like CA7, CA11, Zeke, Zebb, PeopleSoft, and Control-M.
The document describes two job openings for a 5-month project in Arnhem, Netherlands. The project involves adapting existing software systems to implement new functionality for transportation payment devices. The first position is for a software architect responsible for design and oversight. The second is for a C++ software engineer to review designs and adapt existing code. Both roles require experience with AIX, C++, DB2, and transaction processing systems.
Piyush Kumar Gupta is seeking career opportunities as a Linux System Administrator or Support Engineer in Delhi/NCR. He has over 7 years of experience in Linux administration and support roles. He is RHCE certified and has experience working with Red Hat Enterprise Linux 5/6/7 and CentOS 6. He is currently employed as a Linux Specialist at HCL Technologies and has previously worked at Dell International Services, Tetra Information Services, and INS E Solution.
This document discusses managing projects in new technologies and the impact on management and planning. It provides a brief history of technology from 1960 to the present day. It then covers various topics related to managing technologies including roles and responsibilities, project lifecycles, planning constraints, solution modeling, risk profiles, dependencies, and architecture types. The document discusses concepts like cloud computing, NoSQL databases, analytics, mobile applications, and emerging trends like the internet of things, big data, gamification, and social media.
Expert guidance on migrating from magento 1 to magento 2James Cowie
Migrating a Magento site is not just about code and data. Commerce platforms evolve over time and your Magento 1 solution is likely different today compared to the day you launched. Planning a successful migration means understanding what you have and where you are going before you can begin. In this session, architects from the Magento Expert Consulting Group will lay out best practices for defining your migration strategy, and share tips and techniques for code and data migration.
Did you know the average age of the top four banks is 175 years old? And those four banks are the end result of merging or acquiring 35 separate companies over the last 40 years? This means lots of different cultures, tech stacks, apps, processes, and controls – and SCRIPTS!!
Join Chris Nowak, Principal at FlowStates, and Avantika Mathur, ElectricFlow Product Manager, as they share hard-won lessons and emerging patterns in DevOps and release orchestration that can help any regulated organizations climb out of scripting hell. You’ll hear:
1) Why it’s important to prioritize delivery improvement across all your teams
2) How the theory of constraints is limiting your success
3) And how simple things like version control for artifacts can help improve success rates!
How did we move the mountain? - Migrating 1 trillion+ messages per day across...HostedbyConfluent
Have you ever migrated Kafka clusters from one data center to another being completely transparent to client applications?
At PayPal, as part of a massive datacenter migration initiative, Kafka team successfully moved all PayPal Kafka traffic across data centers. This initiative involved migrating 20+ Kafka clusters (1000+ broker and zookeeper nodes), as well as 60+ mirrormaker groups which seamlessly handle Kafka traffic volumes as high as 1 trillion messages per day. Throughout the course of this migration, applications required no modification, encountered 0% service outage, 0% message loss and duplicated messages. The whole migration process was fully transparent to Kafka applications.
In this session, you will learn the strategies, techniques and tools the PayPal Kafka team has utilized for managing the migration process. You will also learn the lessons and pitfalls they experienced during this exercise, as well as the secret sauce of making the migration successful.
This document discusses ADC Austin's M3 Modernization tool and process for modernizing legacy CA 2E environments. It provides an overview of the M3 methodology, which uses model-based migration to automate the modernization of the entire 2E model. A case study is presented on a customer migration project. The presentation concludes with a discussion of next steps organizations can take to evaluate and implement the M3 Modernization process.
The document discusses Microsoft System Center 2012 R2 and its components for managing IT infrastructure and automating processes. It provides an overview of System Center capabilities for data center and client automation. Key components described include System Center Configuration Manager for device management, Operations Manager for monitoring, Virtual Machine Manager for hypervisor management, and Service Manager for IT service management. The document demonstrates System Center's unified management capabilities and how customers can get started or advance their use of System Center.
Mixing d ps building architecture on the cross cutting examplecorehard_by
В рамках доклада мы поговорим о важности архитектурных решений, в том числе, для обеспечения высокого качества ПО при минимальных трудозатратах. Сквозной пример из области резервного копирования данных позволит лучше понять техническую, QA и общепроцессную составляющие подхода. Прошло достаточно времени, чтобы раскрыть технические детали без нарушения NDA, предложенный вариант на базе метрик, которые мы обязательно упомянем, был признан лучшим архитектурным решением в рамках компании – одного из лидеров отрасли, получил награду Microsoft, был «размножен» на смежные области. Приступаем: Builder, Decorator, Composite, Iterator и Visitor - как эти паттерны помогли решить нетривиальную С++ задачу.
The Top 5 Practices of a Highly Successful ChangeMan ZMF AdministratorSerena Software
The summary provides an overview of the Serena Software mainframe virtual user group meeting in October 2015. It discusses product updates to ChangeMan ZMF, Comparex, and ChangeMan SSM. It also covers best practices for highly successful ChangeMan ZMF administrators, including saying no to processes outside of ChangeMan, treating ChangeMan like a production system, taking upgrades seriously as projects, communicating with development, and continuing education. The meeting concluded with an opportunity for questions.
Here are the key terms from the document:
- Application Lifecycle Management (ALM): The product lifecycle management (governance, development, and maintenance) of application software.
- Collaborative Lifecycle Management (CLM): The coordination of software development activities across business and system requirements, design, development, build, test, and delivery.
- Jazz Team Server: The central server that integrates Rational CLM tools like Requirements Management, Quality/Test Management, Planning, Version Control, Work Items, and Build.
- Requirements Management: Tools to define and organize requirements with automated traceability.
- Quality/Test Management: Tools for comprehensive test planning, simplified lab management, and
Patterns and Pains of Migrating Legacy Applications to KubernetesQAware GmbH
Open Source Summit 2018, Vancouver (Canada): Talk by Josef Adersberger (@adersberger, CTO at QAware), Michael Frank (Software Architect at QAware) and Robert Bichler (IT Project Manager at Allianz Germany)
Abstract:
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud-native apps. But what to do if you’ve no shiny new cloud-native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can!
We’re facing the challenge of migrating hundreds of JEE legacy applications of a German blue chip company onto a Kubernetes cluster within one year.
The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way.
Patterns and Pains of Migrating Legacy Applications to KubernetesJosef Adersberger
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs, and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud native apps. But what to do if you’ve no shiny new cloud native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can!
We’re facing the challenge of migrating hundreds of JEE legacy applications of a German blue chip company onto a Kubernetes cluster within one year.
The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way.
Mark Cooper is a senior DevOps cloud engineer with over 15 years of experience in cloud development, DevOps, and cloud infrastructure. He has extensive skills in Java, Python, cloud technologies like OpenStack, and DevOps tools like Chef and Ansible. His background includes roles providing automated provisioning for IBM's cloud services, developing microservices for BlueMix, and managing deployments and pipelines. He aims to deliver high-quality solutions through an agile approach and effective communication skills.
The document discusses several software development life cycle (SDLC) models including waterfall, V-shaped, prototyping, rapid application development (RAD), incremental, spiral, agile, extreme programming (XP), feature driven design (FDD), dynamic systems development method (DSDM), and adaptive SDLC models. It provides an overview of the key phases, strengths, weaknesses, and scenarios where each model is best applied.
The document provides an overview of various software development life cycle (SDLC) models including Waterfall, V-Shaped, Prototyping, Rapid Application Development (RAD), Incremental, Spiral, Agile approaches like Extreme Programming (XP) and Feature Driven Development (FDD). It describes the key phases, strengths, weaknesses and scenarios where each model is best suited. The SDLC models range from traditional plan-driven to more adaptive approaches and the choice of model depends on project factors like requirements, risks, schedules and team preferences.
The document provides an overview of various software development life cycle (SDLC) models including Waterfall, V-Shaped, Prototyping, Rapid Application Development (RAD), Incremental, Spiral, Agile approaches like Extreme Programming (XP) and Feature Driven Development (FDD). It describes the key phases, strengths, weaknesses and scenarios where each model is best suited. The SDLC models range from traditional plan-driven to more adaptive approaches and the choice of model depends on project factors like requirements, risks, schedules and team preferences.
This presentation covers data migration challenges, Kovair's migration architecture and components, example scenarios and benefits derived from Omnibus solutions. To learn more contact sales@kovair.com
SoftwareCircus 2020 "The Past, Present, and Future of Cloud Native API Gateways"Daniel Bryant
An API gateway is at the core of how APIs are managed, secured, and presented within any web-based system. Although the technology has been in use for many years, it has not always kept pace with recent developments within the cloud native space, and many engineers are confused about how a cloud native API gateway relates to Kubernetes Ingress or a Service load balancer.
Join this session to learn about:
– The evolution of API gateways over the past ten years, and how the original problems they were solving have shifted in relation to cloud native technologies and workflow
– Current challenges of using an API gateway within Kubernetes: scaling the developer workflow; and supporting multiple architecture styles and protocols
– Strategies for exposing Kubernetes services and APIs at the edge of your system
– A brief guide to the (potential) future of cloud native API gateways
CloudStack and GitOps at Enterprise Scale - Alex Dometrius, Rene Glover - AT&TShapeBlue
The AT&T team recently embarked on a journey with CloudStack and has since deployed a solution which encompasses multiple data-centers. This talk focuses on how they are using open source tools like CloudStack, FreeIPA, and Metal as a Service (MaaS) to support KVM-based VM provisioning at an enterprise scale within a GitOps model.
-----------------------------------------
The CloudStack Collaboration Conference 2023 took place on 23-24th November. The conference, arranged by a group of volunteers from the Apache CloudStack Community, took place in the voco hotel, in Porte de Clichy, Paris. It hosted over 350 attendees, with 47 speakers holding technical talks, user stories, new features and integrations presentations and more.
This document discusses challenges with online patching in Oracle E-Business Suite release 12.2.5. It begins with an overview of the 12.2 architecture and how it enables features like file system editioning and database edition-based redefinition to allow patching while the application is online. It then covers the online patching cycle in detail and discusses options for developing custom code to be either fully or runtime compliant. The document concludes with lessons learned around areas like database object grants, the DB_Domain parameter, executing autoconfig, and administering application nodes. It also discusses some common challenges seen with online patching and useful utilities for monitoring and diagnosing issues.
DevOpsCon 2020: The Past, Present, and Future of Cloud Native API GatewaysDaniel Bryant
An API gateway is at the core of how APIs are managed, secured, and presented within any web-based system. Although the technology has been in use for many years, it has not always kept pace with recent developments within the cloud native space, and many engineers are confused about how a cloud native API gateway relates to Kubernetes Ingress or a Service load balancer.
Join this session to learn about:
– The evolution of API gateways over the past ten years, and how the original problems they were solving have shifted in relation to cloud native technologies and workflow
– Current challenges of using an API gateway within Kubernetes: scaling the developer workflow; and supporting multiple architecture styles and protocols
– Strategies for exposing Kubernetes services and APIs at the edge of your system
– A brief guide to the (potential) future of cloud native API gateways
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
CAKE: Sharing Slices of Confidential Data on BlockchainClaudio Di Ciccio
Presented at the CAiSE 2024 Forum, Intelligent Information Systems, June 6th, Limassol, Cyprus.
Synopsis: Cooperative information systems typically involve various entities in a collaborative process within a distributed environment. Blockchain technology offers a mechanism for automating such processes, even when only partial trust exists among participants. The data stored on the blockchain is replicated across all nodes in the network, ensuring accessibility to all participants. While this aspect facilitates traceability, integrity, and persistence, it poses challenges for adopting public blockchains in enterprise settings due to confidentiality issues. In this paper, we present a software tool named Control Access via Key Encryption (CAKE), designed to ensure data confidentiality in scenarios involving public blockchains. After outlining its core components and functionalities, we showcase the application of CAKE in the context of a real-world cyber-security project within the logistics domain.
Paper: https://doi.org/10.1007/978-3-031-61000-4_16
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
2. Who is CM First?
Enterprise Software Development Integrator
• Offices in Switzerland, USA, Italy, France, India
• Network of worldwide resale partners (Americas, EMEA, APJ, ANZ)
CA Primary Sales and Technology Partner
• CA 2E, CA Plex, CA Repository
• Footprint covers 20% of Installed Base
IBM Advanced Partner
• IBM Power Systems and zOS
App Dev Products
• CM Matchpoint ALM Suite
• CM WebClient
• CM Power (PHP) Suite
• CM M3 (CA Migration / Modernization)
• CM Meta Analytics
Semantic Design Partner
• DMS Code Analysis / Transformation
26. Next Steps
• Migration Questionnaire
• Start to have conversation on precise scope
• Sizing of environment, technical details
• Formal Analysis Proposal
• Analysis Project
• Migration Assessment Report
• Fixed Price Proposal
Editor's Notes
In this presentation, we will take a look at …
First, for those of you who are not familiar with CM First, here is an overview.
Who is CM First?
CM First is an Enterprise Software Development and Integrator with offices in Switzerland, US, Italy, France and India. We have a worldwide network of resale partners in the Americas, Europe, Middle East and Africa, Asia, Pacific and Japan and Australia and New Zeland.
Our primary business partnership is with CA. We primarily focus on CA 2E, CA Plex and CA Repository. We do a lot of work to enhance customer experience around these products.
CM First is also and IBM Advanced Partner on the IBM Power and zOS.
We also develop and maintain several of our own products. Matchpoint ALM Suite, WebClient, CM Power PHP Suite, M3 migration and Modernization and Meta Analytics.
CM First is also a partner with Symantic Design where we were together to develop DMS Code Analysis and Transformation
Select CM First / SD Modernization Customers
Here are some of our more recognized customers. We are pleased to be able to work with companies of many sized and assist companies in reaching specific goals.
6
Project Preparation
During the project preparation we are working together with the customer to get a successful foundation for the project. The cleaner the 2E model is the better the migration becomes. There is no need to send extra functions, fields and unwanted objects into Plex. This is the best time to make sure the unwanted objects are eliminated.
It is also important that the 2E application directly represents the 2E Model. To do this compile all the programs and fix any issues that come about. We find that over the years there are some objects that have been manipulated by hand or do not represent the current state of the 2E function definition. The best way to manage this is to deliver a freshly compiled application.
There will be a lot of testing going on at several stages of the process so it is important to get data that can be used to test all of the application functionality. The data will be used by several engineers to it is best to not take a copy of production and send that. CM First does enforce confidentiality but we want to focus on the migration and not on data security.
Once CM First gets a clean model we will install it on our IBMi for conversion purposes. If the model is 8.1 or earlier we will need to upgrade the model to a current version of 2E. It could take additional time if the model is from a version of 7 or earlier.
Before the migration begins it is important for CM First to receive a model that does not have a lot of invalid object definitions. The customer will find it beneficial as well because there is no need to migrate objects that are not actually part of the application anymore. Most models have collected field definitions, functions, files and messages that are no longer needed. These unused object definitions should be removed before sending the model to CM First.
The application should also be newly generated and compiled to make sure that the application directly reflects the model. This will eliminate the possibility of manually fixed programs, programs that have been deleted in 2E but not in the application and other potential issues.
CM First will need test data that can be used by several engineers to test the current application and the application once it has been migrated to Plex. This data should not be live confidential data.
Once CM First receives the save files of the model, application and data the model and application are installed on the CM First IBMi. If the model is prior to 8.5 it must be upgraded. The recommended 2E version to us is 8.6. If CM First has to convert the application from a 2E version prior to 8.1 there will be additional time and cost added to the project.
CM First does an extensive analysis of the 2E Model. We determine the number of objects in the model by type We review the user source by quantity and complexity. Execute user messages must be reviewed as they are much like user source.
There are options to look at when dealing with the user source.
CM First can be contracted to rewrite the source in Plex or java.
Customer can rewrite the source in Plex or Java.
Plex APIs can be used to call the native programs.
Each method has pros and cons. However, a decision must be made on how to handle the user source in order to complete the project successfully on budget and within an acceptable timeframe.
The full analysis that CM First completes helps to determine the expected length, cost and duration of the project. Results are used in creating an acceptable project timeline and assignment of appropriate resources to the tasks.
The migration process is run at least 2 times. First we do a preliminary model migration. During this migration we will identify any issues that came up. Issues will be looked at and a solution will be defined. In some cases it may mean the model needs to be adjusted while in other cases it could mean a modification of Plex patterns. By the time we are done with this phase we will expect a 99+ percent of the 2E model supported objects to be migrated into the Plex.
The final migration normally occurs with a freshly delivered model that is adjusted for the final migration. This model will provide the excellent results that we would expect.
Once the migration is completed we will validate that the objects expected to migrate were migrated as expected. Full testing of the Plex Model and generated application occur in Phase 3.
There are several key steps in the actual migration phase.
The 2E model is first converted to XML. We are not talking about just functions or fields but the entire 2E repository is converted to 5 different XML documents. These documents contain everything from field definitions to conditions to location of fields on the screens.
The XML files are then run through a transformation process that takes all the XML and converts it to object files that Plex will be able to import. This transformation is where the magic happens. Everthing needed to build the Plex Model from the 2E model is defined and refined so Plex can use the information.
Finally, a Plex API is used to import the transformed files into Plex. We have created some Plex patterns that help create a model that maintains some similarities to 2E in order to help 2E developers move into Plex development more quickly. At the end of this import you will have a fully functional Plex Model that represents your original 2E Model.
Once the migration is completed intensive testing must be done. The goal of this process is to make sure the Plex model and application have at least the same functionality as the 2E model and application. Testing is done at the model level to insure the model is migrated correctly to Plex. Testing is also done on the migrated application to demonstrate that the application has the same functionality that the original 2E application had.
This intense testing is accomplished with the Customer’s involvement. Once the Customer is satisfied with the migration it is considered complete.
Upon completion CM First delivers the customer a Plex Local Model. This model contains all the migrated information from 2E and has been thoroughly tested by both CM First and the customer. Our job is not done until the customer is happy.
We also provide training for developers to learn how to work with the new model in Plex.
CM First takes pride in providing outstanding support to all customers.
There is also a Guarantee that any supported feature will be guaranteed to work at no additional charge
16
Here is an example of the 2E database relations screen vs the Object browser in Plex.
You can see that in Plex all the Data files are placed under a Data Entity in Plex. This provides a similar organization in Plex as in 2E. Instead of doing a Z to get to the access paths the tree is simply expanded in Plex to easily find the file.
File definitions are also very similar between 2E and Plex. In 2E on the File Definition screen you can see a single file definition by entering an S on a file.
In Plex a similar display is seen by opening a file in the Model Editor. Here you see what are called Triples. These Triples have a similar look to a 2E file definition.
Note that Triples are a major tool used in Plex to define most objects.
Here you see the 2E screen editor and a Plex Panel editor. In the Panel editor you can move the buttons and fields around as needed to make the application look good. This panel can be generated in Java Swing directly from Plex or in an AJAX web screen by using Web Client.
Both 2E and Plex use action diagrams to add logic to a program. This example displays a common CASE statement in both action diagrams. Just like in 2E there are user points where logic can be added.
Here is a comparison of User points in 2E with some of the User points in Plex. We developed a pattern so that Action Diagrams User points were similar between Plex and 2E.
2E and Plex have similar methods of calling functions.
Here you see a 2E action diagram calling a function Saisie in file Article. When you do an FF on the function call the parameters are displayed.
Here you see a Call command in the Plex action diagram to the function Saisie in file Article. The parameters can also be seen in the function window.
These are just a few of the similarities between 2E and Plex. Plex does include a host of additional functionality that is not available to 2E. So once you get comfortable with the pieces of Plex that you need for your maintaining you 2E application you will be able to gain more efficiency and capabilities as you get into the world of Plex.
So where do we go from here?
Next Steps
We can provide you with a Migration Questionnaire to help in organizing thoughts on the scope of your migration. It is important to consider the size of your environment and to look into the details of your system.
The information from the questionnaire can be used to get you a formal analysis proposal.
Once this step is completed we can begin an analysis and create a custom Migration Assessment report. We feel confident that when we complete a full assessment we have sufficient information to make a realistic Fixed Price Proposal. This way you know exactly what to expect. We do not want you to have any unwanted surprises down the road.