This document provides an overview and summary of JD Edwards archiving and upgrades using ARCTOOLS. It summarizes a case study from DBG, a manufacturing company, that used ARCTOOLS to successfully archive data and complete an upgrade from JDE 8.10 to 9.1 on time over a weekend while virtualizing servers. The purge and archive using ARCTOOLS significantly reduced database sizes and improved system performance, backup times, and disaster recovery capabilities for DBG.
Migrating your traditional Data Warehouse to a Modern Data LakeAmazon Web Services
In this session, we discuss the latest features of Amazon Redshift and Redshift Spectrum, and take a deep dive into its architecture and inner workings. We share many of the recent availability, performance, and management enhancements and how they improve your end user experience. You also hear from 21st Century Fox, who presents a case study of their fast migration from an on-premises data warehouse to Amazon Redshift. Learn how they are expanding their data warehouse to a data lake that encompasses multiple data sources and data formats. This architecture helps them tie together siloed business units and get actionable 360-degree insights across their consumer base.
At the Open Repositories 2019 Terry Brady and Pascal Becker gave a workshop about DSpace and Docker. They gave an introduction into Docker, an overview over the DSpace specific Docker images, and exercises that help to learn how to use DSpace with Docker.
The Power of Two: Using IBM Standards Processing Engine for EDI Commerce or H...Brian Wilson
This is a presentation covering IBM's newest platform to handle standards processing, with current support for EDI X12 and EDIFACT, as well as HIPAA, within the IBM Integration Bus enterprise service bus offering.
Streaming Data Analytics with Amazon Redshift and Kinesis FirehoseAmazon Web Services
by Joyjeet Banerjee, Enterprise Solutions Architect, AWS
Evolving your analytics from batch processing to real-time processing can have a major business impact, but ingesting streaming data into your data warehouse requires building complex streaming data pipelines. Amazon Kinesis Firehose solves this problem by making it easy to transform and load streaming data into Amazon Redshift so that you can use existing analytics and business intelligence tools to extract information in near real-time and respond promptly. In this session, we will dive deep using Amazon Kinesis Firehose to load streaming data into Amazon Redshift reliably, scalably, and cost-effectively. Level: 200
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
Migrating your traditional Data Warehouse to a Modern Data LakeAmazon Web Services
In this session, we discuss the latest features of Amazon Redshift and Redshift Spectrum, and take a deep dive into its architecture and inner workings. We share many of the recent availability, performance, and management enhancements and how they improve your end user experience. You also hear from 21st Century Fox, who presents a case study of their fast migration from an on-premises data warehouse to Amazon Redshift. Learn how they are expanding their data warehouse to a data lake that encompasses multiple data sources and data formats. This architecture helps them tie together siloed business units and get actionable 360-degree insights across their consumer base.
At the Open Repositories 2019 Terry Brady and Pascal Becker gave a workshop about DSpace and Docker. They gave an introduction into Docker, an overview over the DSpace specific Docker images, and exercises that help to learn how to use DSpace with Docker.
The Power of Two: Using IBM Standards Processing Engine for EDI Commerce or H...Brian Wilson
This is a presentation covering IBM's newest platform to handle standards processing, with current support for EDI X12 and EDIFACT, as well as HIPAA, within the IBM Integration Bus enterprise service bus offering.
Streaming Data Analytics with Amazon Redshift and Kinesis FirehoseAmazon Web Services
by Joyjeet Banerjee, Enterprise Solutions Architect, AWS
Evolving your analytics from batch processing to real-time processing can have a major business impact, but ingesting streaming data into your data warehouse requires building complex streaming data pipelines. Amazon Kinesis Firehose solves this problem by making it easy to transform and load streaming data into Amazon Redshift so that you can use existing analytics and business intelligence tools to extract information in near real-time and respond promptly. In this session, we will dive deep using Amazon Kinesis Firehose to load streaming data into Amazon Redshift reliably, scalably, and cost-effectively. Level: 200
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
This is the presentation for the talk I gave at JavaDay Kiev 2015. This is about an evolution of data processing systems from simple ones with single DWH to the complex approaches like Data Lake, Lambda Architecture and Pipeline architecture
Cloudera - The Modern Platform for AnalyticsCloudera, Inc.
This presentation provides an overview of Cloudera and how a modern platform for Machine Learning and Analytics better enables a data-driven enterprise.
Polyglot Persistence vs Multi-Model DatabasesLuca Garulli
Many complex applications scale up by using several different databases, i.e. selecting the best DBMS for each use case. This tends to complicate modern architecture with many products by different vendors, no standards, and a lot of ETL which ultimately causes unpredictable results and a lot of headaches. Multi-Model DBMSs were created to make your life easier, giving you the option of using one NoSQL product with powerful multi-purpose engines capable of handling complex domains. Could one DBMS handle all your needs including speed and scalability in the times of Big Data? Luca will walk you through the benefits and trade-offs of multi-model DBMSs and will show you how easy it is to setup one open source database to handle many different use cases, saving you time and money.
Presented at Data Day Texas - Austin (TX) - USA
Want to see a high-level overview of the products in the Microsoft data platform portfolio in Azure? I’ll cover products in the categories of OLTP, OLAP, data warehouse, storage, data transport, data prep, data lake, IaaS, PaaS, SMP/MPP, NoSQL, Hadoop, open source, reporting, machine learning, and AI. It’s a lot to digest but I’ll categorize the products and discuss their use cases to help you narrow down the best products for the solution you want to build.
It is important to make architecture accessible in a good way so that it is accessible to all target groups as much as possible. At the same time, architectural languages such as ArchiMate are not fully focused on this. In addition, they are actually too abstract. You cannot express exactly enough what you are trying to model and the resulting model is open to several interpretations. Linked Data fits in well with these objectives and makes it easier to define and unlock more accessible and more targeted information.
In this presentation Danny Greefhorst will tell more about the motivation behind the idea, but also show a further elaboration. For example, he has made a mapping of ArchiMate to commonly used Linked Data vocabularies. He also made a demonstrator, in which you can see how you can define, enrich and publish ArchiMate models as Linked Data. He will also discuss how the European reference architecture EIRA is available as Linked Data.
Architect’s Open-Source Guide for a Data Mesh ArchitectureDatabricks
Data Mesh is an innovative concept addressing many data challenges from an architectural, cultural, and organizational perspective. But is the world ready to implement Data Mesh?
In this session, we will review the importance of core Data Mesh principles, what they can offer, and when it is a good idea to try a Data Mesh architecture. We will discuss common challenges with implementation of Data Mesh systems and focus on the role of open-source projects for it. Projects like Apache Spark can play a key part in standardized infrastructure platform implementation of Data Mesh. We will examine the landscape of useful data engineering open-source projects to utilize in several areas of a Data Mesh system in practice, along with an architectural example. We will touch on what work (culture, tools, mindset) needs to be done to ensure Data Mesh is more accessible for engineers in the industry.
The audience will leave with a good understanding of the benefits of Data Mesh architecture, common challenges, and the role of Apache Spark and other open-source projects for its implementation in real systems.
This session is targeted for architects, decision-makers, data-engineers, and system designers.
Processing the volume and variety of data that today’s organizations produce can be both challenging and costly – especially with a legacy data warehouse. Combining the scale and performance of the cloud with AWS and APN Partner solutions for migration, integration, analysis, and visualization can help overcome these obstacles. With a modern data warehouse architecture, organizations can store, process, and analyze massive volumes of data of virtually any type. Register for this upcoming webinar, where Pearson - an education and media conglomerate - will share in detail how they built a scalable and flexible business intelligence platform on the cloud, with Tableau and AWS.
Learn how you can seamlessly load and transform data in Amazon Redshift with Matillion ETL and analyze it with Tableau. Hear how 47Lining and NorthBay can provide insights to guide you through migration with ease. Tableau will discuss best practices to analyze your data on AWS and share new insights throughout your organization.
Google Cloud Storage is unified object storage for developers and enterprises, from live data serving to data analytics/ML to data archiving.
High performance, internet-scale, immutable BLOB (binary large object) storage
Simple Abstraction Storage buckets, immutable objects with mutable metadata and Globally unique URI identifiers for buckets, objects
Common storage for Google Cloud Platform services
The data lake has become extremely popular, but there is still confusion on how it should be used. In this presentation I will cover common big data architectures that use the data lake, the characteristics and benefits of a data lake, and how it works in conjunction with a relational data warehouse. Then I’ll go into details on using Azure Data Lake Store Gen2 as your data lake, and various typical use cases of the data lake. As a bonus I’ll talk about how to organize a data lake and discuss the various products that can be used in a modern data warehouse.
Data Democratization for Faster Decision-making and Business Agility (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3ogsO7F
Presented at 3rd Chief Digital Officer Asia Summit
The idea behind Data democratization is to enable every type of user in a company to have access to data and to ensure that there is no dependency on any single party that might create a bottleneck to data access. But this is easier said than done especially given the complex data management landscape that most organizations have today. Data virtualization is a modern data integration technique that not only delivers data in real time without replication but also simplifies data discovery, data exploration and navigating between related data sets.
In this on-demand session, you will understand how data virtualization enables enterprises to:
- Reduce up to 80% the time required to deliver data to the business adapted to the needs of each user
- Apply consistent security and governance policies across the self-service data delivery process
- Seamlessly implement the concept of 'Data Marketplace'
Definitive Guide to Select Right Data Warehouse (2020)Sprinkle Data Inc
Choosing the right data warehouse is a big challenge for organisations. In this doc, we have made an end to end comparison of leading data warehouses. Snowflake vs Redshift vs BigQuery vs Hive vs Athena
Sprinkledata.com
Solution architecture for big data projects
solution architecture,big data,hadoop,hive,hbase,impala,spark,apache,cassandra,SAP HANA,Cognos big insights
EDI Modernization for JD Edwards: Anatomy of an Integration ProjectCleo
While no business welcomes application replacement and upgrade pressures, the approaching end-of-life for older JD Edwards versions presents a rare opportunity to improve Electronic Data Interchange (EDI) productivity and automation levels at the same time.
In this presentation on EDI modernization for JD Edwards, we will focus on EDI implementation processes, methods, and best practices, including real-world JD Edwards project case studies.
See the recorded webinar at: http://www.extol.com/modernizejdeintegration
Inventive InfoTech INC is an Oracle Gold partner and align to the Oracle roadmap for its ERP products. We are proud members of Quest user group, which gives us a forum to understand our clients better and showcase our solutions worldwide. We provide JD Edwords Managed Services, CNC Managed Services and JD Edwords Enterpriseone services.
This is the presentation for the talk I gave at JavaDay Kiev 2015. This is about an evolution of data processing systems from simple ones with single DWH to the complex approaches like Data Lake, Lambda Architecture and Pipeline architecture
Cloudera - The Modern Platform for AnalyticsCloudera, Inc.
This presentation provides an overview of Cloudera and how a modern platform for Machine Learning and Analytics better enables a data-driven enterprise.
Polyglot Persistence vs Multi-Model DatabasesLuca Garulli
Many complex applications scale up by using several different databases, i.e. selecting the best DBMS for each use case. This tends to complicate modern architecture with many products by different vendors, no standards, and a lot of ETL which ultimately causes unpredictable results and a lot of headaches. Multi-Model DBMSs were created to make your life easier, giving you the option of using one NoSQL product with powerful multi-purpose engines capable of handling complex domains. Could one DBMS handle all your needs including speed and scalability in the times of Big Data? Luca will walk you through the benefits and trade-offs of multi-model DBMSs and will show you how easy it is to setup one open source database to handle many different use cases, saving you time and money.
Presented at Data Day Texas - Austin (TX) - USA
Want to see a high-level overview of the products in the Microsoft data platform portfolio in Azure? I’ll cover products in the categories of OLTP, OLAP, data warehouse, storage, data transport, data prep, data lake, IaaS, PaaS, SMP/MPP, NoSQL, Hadoop, open source, reporting, machine learning, and AI. It’s a lot to digest but I’ll categorize the products and discuss their use cases to help you narrow down the best products for the solution you want to build.
It is important to make architecture accessible in a good way so that it is accessible to all target groups as much as possible. At the same time, architectural languages such as ArchiMate are not fully focused on this. In addition, they are actually too abstract. You cannot express exactly enough what you are trying to model and the resulting model is open to several interpretations. Linked Data fits in well with these objectives and makes it easier to define and unlock more accessible and more targeted information.
In this presentation Danny Greefhorst will tell more about the motivation behind the idea, but also show a further elaboration. For example, he has made a mapping of ArchiMate to commonly used Linked Data vocabularies. He also made a demonstrator, in which you can see how you can define, enrich and publish ArchiMate models as Linked Data. He will also discuss how the European reference architecture EIRA is available as Linked Data.
Architect’s Open-Source Guide for a Data Mesh ArchitectureDatabricks
Data Mesh is an innovative concept addressing many data challenges from an architectural, cultural, and organizational perspective. But is the world ready to implement Data Mesh?
In this session, we will review the importance of core Data Mesh principles, what they can offer, and when it is a good idea to try a Data Mesh architecture. We will discuss common challenges with implementation of Data Mesh systems and focus on the role of open-source projects for it. Projects like Apache Spark can play a key part in standardized infrastructure platform implementation of Data Mesh. We will examine the landscape of useful data engineering open-source projects to utilize in several areas of a Data Mesh system in practice, along with an architectural example. We will touch on what work (culture, tools, mindset) needs to be done to ensure Data Mesh is more accessible for engineers in the industry.
The audience will leave with a good understanding of the benefits of Data Mesh architecture, common challenges, and the role of Apache Spark and other open-source projects for its implementation in real systems.
This session is targeted for architects, decision-makers, data-engineers, and system designers.
Processing the volume and variety of data that today’s organizations produce can be both challenging and costly – especially with a legacy data warehouse. Combining the scale and performance of the cloud with AWS and APN Partner solutions for migration, integration, analysis, and visualization can help overcome these obstacles. With a modern data warehouse architecture, organizations can store, process, and analyze massive volumes of data of virtually any type. Register for this upcoming webinar, where Pearson - an education and media conglomerate - will share in detail how they built a scalable and flexible business intelligence platform on the cloud, with Tableau and AWS.
Learn how you can seamlessly load and transform data in Amazon Redshift with Matillion ETL and analyze it with Tableau. Hear how 47Lining and NorthBay can provide insights to guide you through migration with ease. Tableau will discuss best practices to analyze your data on AWS and share new insights throughout your organization.
Google Cloud Storage is unified object storage for developers and enterprises, from live data serving to data analytics/ML to data archiving.
High performance, internet-scale, immutable BLOB (binary large object) storage
Simple Abstraction Storage buckets, immutable objects with mutable metadata and Globally unique URI identifiers for buckets, objects
Common storage for Google Cloud Platform services
The data lake has become extremely popular, but there is still confusion on how it should be used. In this presentation I will cover common big data architectures that use the data lake, the characteristics and benefits of a data lake, and how it works in conjunction with a relational data warehouse. Then I’ll go into details on using Azure Data Lake Store Gen2 as your data lake, and various typical use cases of the data lake. As a bonus I’ll talk about how to organize a data lake and discuss the various products that can be used in a modern data warehouse.
Data Democratization for Faster Decision-making and Business Agility (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3ogsO7F
Presented at 3rd Chief Digital Officer Asia Summit
The idea behind Data democratization is to enable every type of user in a company to have access to data and to ensure that there is no dependency on any single party that might create a bottleneck to data access. But this is easier said than done especially given the complex data management landscape that most organizations have today. Data virtualization is a modern data integration technique that not only delivers data in real time without replication but also simplifies data discovery, data exploration and navigating between related data sets.
In this on-demand session, you will understand how data virtualization enables enterprises to:
- Reduce up to 80% the time required to deliver data to the business adapted to the needs of each user
- Apply consistent security and governance policies across the self-service data delivery process
- Seamlessly implement the concept of 'Data Marketplace'
Definitive Guide to Select Right Data Warehouse (2020)Sprinkle Data Inc
Choosing the right data warehouse is a big challenge for organisations. In this doc, we have made an end to end comparison of leading data warehouses. Snowflake vs Redshift vs BigQuery vs Hive vs Athena
Sprinkledata.com
Solution architecture for big data projects
solution architecture,big data,hadoop,hive,hbase,impala,spark,apache,cassandra,SAP HANA,Cognos big insights
EDI Modernization for JD Edwards: Anatomy of an Integration ProjectCleo
While no business welcomes application replacement and upgrade pressures, the approaching end-of-life for older JD Edwards versions presents a rare opportunity to improve Electronic Data Interchange (EDI) productivity and automation levels at the same time.
In this presentation on EDI modernization for JD Edwards, we will focus on EDI implementation processes, methods, and best practices, including real-world JD Edwards project case studies.
See the recorded webinar at: http://www.extol.com/modernizejdeintegration
Inventive InfoTech INC is an Oracle Gold partner and align to the Oracle roadmap for its ERP products. We are proud members of Quest user group, which gives us a forum to understand our clients better and showcase our solutions worldwide. We provide JD Edwords Managed Services, CNC Managed Services and JD Edwords Enterpriseone services.
Practical Ontology For Enterprise Data ManagementRichard Green
As a methodologist and software engineer, I’ve designed CASE tools that collect and impose organization on meta-data at all levels of detail.
As a enterprise architect, I’ve worked in an environment that uses planning-level data. In a planning context, you need a method for quickly characterizing the problem space at the enterprise level. I’ve stumbled on to a way to do that.
What Customers Have to Say About Purge-it!TeamCain
The following is a compilation of customer quotes about the Oracle Validated Integration product that is a fully integrated purging and archiving solution for JD Edwards, Purge-it!
Collaborate 2016 - Modernizing the JD Edwards AP Process with Oracle Fusion M...AVIO Consulting
This presentation is from the AVIO and Land O'Lakes session at Collaborate 2016 (session 2007) called "Modernizing the JD Edwards AP Process".
This demo takes an Accounts Payable invoice through the complete lifecycle from scanning using Oracle Data Capture (ODC), extracting the metadata using Oracle Forms Recognition (OFR), storing the metadata an image using Oracle Imaging and Process Management (IPM), and Oracle Business Process Management (BPM).
Video of demo can be found at: https://www.youtube.com/watch?v=Lfk9fyMYiZI
OAUG Collaborate 2016 - Data Conversion & Maintenance made Easy!!Jamie Weingard
Do you want to see how the American Red Cross found an easy way to manage data conversions and maintenance? Join me at OAUG Collaborate 2016 at Mandalay Bay Resort & Casino, Las Vegas, NV - April 10 -14th. Session # 10299
For efficient and innovative use of big data, it is important to integrate multiple data bases across domains. For example, various public data bases are developed in life science, and how to find a novel scientific result using them is an essential technique. In social and business areas, open data strategies in many countries promote diversity of public data, how to combine big data and open data is a big challenge. That is, diversity of dataset is a problem to be solved for big data.
Ontology gives a systematized knowledge to integrate multiple datasets across domains with semantics of them. Linked Data also provides techniques to interlink datasets based on semantic web technologies. We consider that combinations of ontology and Linked Data based on ontological engineering can contribute to solution of diversity problem in big data.
In this talk, I discuss how ontological engineering could be applied to big data with some trial examples.
Here is a visual workflow of NetSuite’s ERP Order to Cash (O2C) and Purchase to Pay (P2P) processes, covering Order Management (Orders), Accounts Receivable (Receivables), Purchase Orders (Purchases) and Accounts Payable (Payables) plus an introduction to Customer Relationship Management (CRM).
A presentation on P2P (Procurement to Pay) process usually being followed by industry.
It's part of supply chain and PPT also shows -
- how the material inflow thru P2P is being balance by material outflow thru O2C
- how the cash outflow thru P2P is being balanced by cash inflow thru O2C
Intense competition and slow growth in mature markets have magnified uncertainty and put pressure on costs, just as regulators are escalating their demands. Research shows that CFOs and other senior finance executives believe that their function can play a key role but the ability to impact these challenges depends on levels of maturity and preparedness, which vary widely across companies and industries, as well by sub-functions. Here are the key findings from our research on how enterprises are driving transformation to achieve business impact.
Procure to Pay Transformation.
--
Presentation originally made at the SMART Supply Chain Conference, June 2013. This transformation won the Award for Procurement Excellence.
The presentation is Copyright protected and is for information purposes only and may not be reproduced in any form without written permission of authors.
New Approaches to Integrating Oracle and Postgres Database StrategiesEDB
Given today’s challenging economic environment, managing the costs of your Oracle databases is more important than ever. Open source alternatives like Postgres are a real option but the migration process is not always as simple as “rip and replace”. For many, it is a question of how to integrate the two databases.
This webinar will explore:
- The realities of Oracle and Postgres coexistence
- The migration continuum - identifying the best workloads for Postgres
- Strategies for integrating Oracle and Postgres databases
The move from Oracle database to Postgres is often not an "either/or" decision. Join us to understand how these two databases can peacefully coexist while reducing costs and managing risk.
A JDE Hat Trick – 3 Ways to Extend your JDE and Get Great EfficienciesTeamCain
Learn about a “hat trick” of solutions that can significantly extend the value of JDE for your organization – JDE purge and archive, automated and integrated Spreadsheet reporting, and integrated RF data collection. During the session, we will explain what each solution is, how it provides efficiencies for JDE customers, give at least two case studies for each product, and let you know the questions to ask yourself to see if the solution would help you out.
Originally presented at JDE INFOCUS 2013 (December 03, 2013)
Gary Carter, Director of Product Marketing presents the accelerating adoption of Postgres by large organizations interested in reducing their database costs.
He reviews the main drivers of Postgres adoption, as well as the profiles of several organizations currently reaping the benefits of Postgres, and explore a few strategies that your organization can leverage to make the move at your pace – quickly or slowly.
To see the Webcast recording – please visit EnterpriseDB > Resources > Webcasts. Questions? Contact sales@enterprisedb.com
NoSQL Now: Postgres - The NoSQL Cake You Can EatDATAVERSITY
The path to creating a single view of your customers or your transactional systems is overflowing with high costs and complexity. Major vendors have built massive, million-dollar systems that are too expensive and too complicated for most. NoSQL-only solutions seem to have promise, but simply do not necessarily have what you need. Learn what Postgres can do for you that NoSQL-only solutions can't.
Using a NoSQL-only solution and dumping gigabytes of data from multiple disparate systems into gigantic documents is complicated. And it forces tough choices—group all data by customer, by transaction, or by policy? You must choose, and this can be a hard process for some organizations. And almost always, organizations later learn they need relationships among the data, which NoSQL-only solutions cannot support.
Postgres eliminates the complexity and the pain of creating a single view of the customer. With recent advances, Postgres can support semi-structured, unstructured and structured data in the same environment, employing relational qualities and ACID compliance.
During this presentation, Marc Linster, SVP Products & Services, will review: ·
How to do more with Postgres
Open source alternative to RDBMS and more...
The NoSQL Conundrum
Why do developers like NoSQL Only solutions?
Problems and fallacies of NoSQL (only)
Data Standards
Data Islands
NoSQL Data Models include data access paths
Not Only SQL - Technology Innovation on a Robust Platform
Document Store
See JSON Examples
360 Degree view of the customer
Data Integration
Behavior-Driven Development (BDD) Testing with Apache Spark with Aaron Colcor...Databricks
This session presents a simple, human-based approach to create test suites targeting multiple points of contact in a data solution. Commonly, an enterprise will pick a data processing solution with heavy GUIs because it can make an easy to understand workflow around data. However, those solutions still are not able to verify the simplest use case, i.e. “If I put data into a solution to process data, then I should get a desired result.”
FIS will demonstrate and teach you how to build a unique testing solution on top of Apache Spark. Under its solution, FIS can actually prove to users in their organization that when they put data in, they get the correct result out. They can also enlist their entire team from product owner to developer to write complete unit tests. The type of flexibility Spark enables allows you to take unique paths in building robust, understandable data flows. The transformational element is the ability to do this in milliseconds, and not wait till the entire pipeline finishes.
How to Migrate from Oracle to EDB PostgresAshnikbiz
Are you planning to move from Oracle to a preferred enterprise open source database technology like Postgres? Watch this webinar to overcome the challenges you will face during the migration process.
How to migrate from Oracle to EDB PostgresAshnikbiz
Are you planning to move from Oracle to a preferred enterprise open source database technology like Postgres? Watch this webinar to overcome the challenges you will face during the migration process.
Watch full webinar here: https://www.ashnik.com/resources/webinars/migration-from-oracle-to-postgresql-need-vs-reality/
Postgres eliminates the complexity and the pain of creating a single view of the customer. With recent advances, Postgres can support semi-structured, unstructured and structured data in the same environment, employing relational qualities and ACID compliance.
This presentation reviews:
- How advances in Postgres enable it to match capabilities of NoSQL-only niche solutions·
- How the ETL process in Postgres is simple compared to undoing tables and schemas in order to transfer data to a NoSQL-only system
- How Foreign Data Wrappers – essentially pipelines between Postgres and other databases – work and how they help bridge the gap between disparate systems faster than an ETL process
Visit Enterprisedb.com and go to our Resources section, then Webcasts to listen to the presentation recording.
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...DATAVERSITY
Thirty years is a long time for a technology foundation to be as active as relational databases. Are their replacements here? In this webinar, we say no.
Databases have not sat around while Hadoop emerged. The Hadoop era generated a ton of interest and confusion, but is it still relevant as organizations are deploying cloud storage like a kid in a candy store? We’ll discuss what platforms to use for what data. This is a critical decision that can dictate two to five times additional work effort if it’s a bad fit.
Drop the herd mentality. In reality, there is no “one size fits all” right now. We need to make our platform decisions amidst this backdrop.
This webinar will distinguish these analytic deployment options and help you platform 2020 and beyond for success.
This is a high level presentation I delivered at BIWA Summit. It's just some high level thoughts related to today's NoSQL and Hadoop SQL engines (not deeply technical).
Jd edwards upgrade roundtable at innovate15 empire merchants case studyNERUG
An in-depth, interactive session addressing real world scenarios and the challenges that are often faced during an upgrade. Hear from customer and other "experts" who will share experiences, best practices and lessons learned and help eliminate "the gotchas!" If you’re planning or contemplating an upgrade, this session provides a great opportunity to understand what lies ahead -- and be much more prepared!
Sales Tax Compliance within Oracle E-Business Suite / JD Edwards / PeopleSoftNERUG
Sales and use tax compliance is difficult to manage, especially if your process is manual. Join Avalara, our partner in sales tax compliance as we walk you thru the best ways to quickly and reliably determine sales tax, filing returns and efficiently manage exempt sales within Oracle. Participants will learn:
• How to protect their business from audits
• How to save time and money on sales tax compliance
• Why zip codes mean zip when it comes to sales tax
• Nexus: Where do you currently have a physical presence
• Product and Service Taxability: How are your products and services taxed
The Business Case for Hosting JD Edwards in the CloudNERUG
This presentation will cover in detail the business case for hosting JD Edwards in the cloud. Hear from industry expert, John Bassett, CTO at GSI, Inc. During the presentation, John will address the following topics, comparing hosting to a more traditional in-house approach:
- Total cost of ownership (TCO)
- IT staffing and support costs
- Licensing costs
- Cost Predictability
- Security and compliance
- Scalability, performance and reliability
- Business continuity and redundancy
- System deployment
- System management, ongoing maintenance and upgrades
- Market adaptability, agility and innovation
PM 201: Emotional Intelligence for Project ManagersNERUG
One of our key learning objectives is to help project managers understand the importance and balance of SMART with HEART. Both are major components of Emotional Intelligence (EQ). We will discuss where Emotional Intelligence is required in order to manage change effectively, lead others to the vision, and produce the desired results. We will review the history of EQ theory, the evolution of EQ models, measurement techniques to assess your EQ, and how to form a cohesive team using the five basic behaviors. Recognizing and facing the different types/maturity levels in EQ can and will lead to a smoother project.
Take it to the Oracle ERP Cloud! Hitachi will profile a recent cloud ERP implementation. Hitachi will discuss why Oracle Cloud ERP was selected, the project approach, timeline and lessons learned. Hitachi will also discuss the multiple Oracle cloud solutions (Sales Cloud, Marketing Cloud, HCM Cloud, etc.) they have already deployed in-house to support their own business growth.
Implementation of eProfile and Benefits of using Talent Summary and Manager D...NERUG
This presentation focuses on eProfile as a collaborative application that enables employees to maintain their own profiles, thereby decreasing administrative time and costs and increasing data integrity. eProfile ensures that data changes comply with the organization's requirements
The session will focus on:
• Managing personal information
• Self-service functionality
• Manager Dashboard
• Talent Summary Dashboard
• Workflow user preferences
Using the PeopleSoft HCM 9.2 PUM (PeopleSoft Update Manager) for Upgrades and...NERUG
WorkStrategy will share key HR strategies for leveraging new PeopleSoft 9.2 functionality and delivery model. Come hear tips for installing the new PeopleSoft 9.2 Image on VirtualBox. Oracle offers pre-built PeopleSoft 9.2 VirtualBox virtual machines that enable you to have ready-made application environments for upgrade, maintenance and demonstrations in minutes rather than days.
This presentation will also review best practices for deploying the PeopleSoft Update Manager (PUM) patching process. PUM is a flexible, streamlined patching process that enables you to identify, understand, and selectively apply maintenance updates according to your requirements and your schedule.
Objectives:
1. Understand how to leverage new 9.2 functional enhancements
2. Hear about best practices for installing the PeopleSoft Update Manager (PUM)
3. Review tips for upgrades and applying maintenance updates
PeopleSoft 9.2 HCM Features and Functions Including Fluid MobileNERUG
WorkStrategy will highlight the benefits of upgrading to PeopleSoft HCM 9.2/8.54 including the new Fluid User Interface (mobile), Activity Guides (e.g. Benefits Life Events), Pivot Grids/Analytics, Paycheck Modeler, Talent Management Integrations, Dashboards, Workcenters, and other productivity tools. Presenters will also review steps for configuring new administrative tools including Pivot Grids/Analytics, Dashboard Quick Links, HR Notifications, and Secure Enterprise Search (SES).
The new PeopleSoft Fluid UI provides an intuitive and responsive user interface that provides native-application interactivity across a range of mobile operating systems and devices. PeopleSoft Fluid UI delivers two out-of-the box homepages (Employee Self Service and Manager Self Service) that provide employees, managers and executives with a familiar native user experience that helps maximize productivity and effectiveness.
Objectives:
· Understand how to leverage new 9.2 functional enhancements
· Review PeopleSoft Fluid User Interface enabled HCM transactions
· Learn configuration tips and tricks for new administrative tools
· Discuss upgrade planning and best practices
Cool Functionality and Workarounds to Make Your Life Easier - JD Edwards WorldNERUG
The JD Edwards software functions as one body but we tend to see the overall systems in parts and pieces because in IT and Support, like doctors, we specialize / troubleshoot in areas of expertise. Speaking software, you may only be licensed to only specific systems so perhaps when you go to the specialist you do not always learn what can make life easier in other systems. Can a general practitioner give you excellent advice? You bet!
Let us broaden your software knowledge for JD Edwards systems as a whole and bring together the new functionality and business savvy workarounds that make life in JD Edwards a healthy solution.
Upgrading PeopleSoft Financials from 8.9 to 9.2 – Advice You Need to Know fro...NERUG
Founded in 1893, the Visiting Nurse Service of New York (VNSNY) is the largest not-for-profit home and community-based health care organization in the United States, serving the five boroughs of New York City and Nassau, Suffolk, and Westchester Counties, as well as parts of upstate New York. As an early adopter, VNSNY will present a case study of how they were able to successfully upgrade from 8.9 to 9.2 in a limited time frame with limited resources and provide tips and advice on how others can achieve similar success. VNSNY was able to successfully upgrade GL, AM, AP, PO, and ePro from PeopleSoft 8.9/PeopleTools 8.49 to 9.2/PeopleTools 8.53 in just 7.5 months using Oracle’s newly released 8.9 to 9.2 path. VNSNY had only a very lean team of internal and external functional/technical resources. VNSNY was also able to successfully maximize the use of a non-full time external PeopleSoft Admin/Upgrader resource.
Hosted FAH Integrated with PeopleSoft for Deferred Revenue RecognitionNERUG
This presentation will cover the implementation at ETS of Financials Accounting Hub processing revenue recognition related detail journals for a web based order capture system integrated with a PeopleSoft General Ledger. The Financials Accounting Hub configuration will be discussed along with the integration points required to interface the resulting journal information into a PeopleSoft General Ledger.
Learn about the user experience enhancements in the next generations of PeopleTools and PeopleSoft. See how Oracle has delivered on streamlined navigation, efficiency, and easy of use. In PeopleSoft in particular, see the newly available PeopleSoft HCM self service capabilities, including direct reports org chart, manager dashboard, talent summary, and related actions. See a demonstration that showcases these new features.
What's New in the PeopleSoft 9.2 Accounts Payable Module?NERUG
Oracle's PeopleSoft Accounts Payable Release 9.2 can help organizations reduce costs, improve visibility into the invoicing and payment streams, and empower Accounts Payable to make decisions to drive bottom line savings. Learn how the new capabilities of PeopleSoft Accounts Payable Release 9.2 can help to streamline processes, reducing operations costs, and strengthen vendor relationships.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
3. 3
WELCOME!
Housekeeping:
- Visit us
- Pass the hat drawing – biz card or fill out a slip
- Show of hands:
- JDE? E-Biz? Demantra? DIY?
- AS400? Oracle? SQL?
4. Agenda
• Introductions to DBG and ARCTOOLS
• Why purge?
• Overview of Archiving in JDE
• Why is it soooooooo difficult?
• Case studies
• Upgrade Case Study - DBG Canada
• ARCTOOLS®
5. Definitions According to Webster’s
• Purge
• transitive verb
• “to make free of something unwanted”
• Archive
• noun
• “a place where records or historical documents are preserved”
• transitive verb
• “to file or collect in or as if in an archive”
According to ARCTOOLS:
Purging is removing records from the production environment.
Archiving is saving those records in an archive environment.
Purging and archiving together is removing records from production and adding them to the archive.
Destruction means permanently removing data from the archive (i.e. “purging from the archive”)
6. DBG
6
DBG is a privately owned supplier to the
auto and defense industries,
manufacturing metal components at 2
plants in Canada and 2 more in Mexico.
DBG has been running JDE since 1998.
7. DBG Canada
• Privately owned manufacturing company
• 35 years in business
• 2 facilities in Canada, 2 in Mexico, running 6x24
• Customers: truck, auto, military, consumer goods manufacturers
• Products:
• metal stampings
• stainless steel and carbon steel assemblies
• design and prototyping
8. ARCTOOLS Company History
• DCSoftware founded 1996, created ARCTOOLS/400
• David Shea, President
• Over 25 years of ERP experience
• Six years with a ‘Big Eight’ consultant
• Created JDE purge and archive methodology in collaboration with JDE and
several customers
• World – 2003. E1 on AS400 – 2004
• Created ARCTOOLS/EM (Oracle, SQL Server) 2005
• Barry Yelverton, Chief Product Architect, Oracle/SQL
• Locations: Boston, Denver, Paris, Limerick
• Over 400 JDE customers on AS400, Oracle, SQL Server
8
9. ARCTOOLS and JD Edwards
• Early 2001 – JDE customer makes introductions to Denver after a successful custom purge effort
• June 2001 – meetings with JDE at Quest Global / Focus
• July thru… developed (with JDE) ‘proof of concept’ purges
• June 2002 – SRO crowd at Quest Global presentation – over 300 customers
• Sept 2002 – development effort in Denver to work with JDE and several large customers
• Oct 2002 – beta site – Cargill
• Jan 2003 – first “JDE Module” released for World
• 2004 – JDE Module for XE on AS/400
• 2005 – Oracle / SQL
• OVI for World and EnterpriseOne
9
10. ARCTOOLS
• Mature, dependable product
• Over 400 Oracle JD Edwards customers world-wide
• A ‘true database solution’
• Not a UBE. Not ODBC. Processes run on the server, in the native
language of the box
• Multiple “accelerators” available for out-of-the-box archiving:
• JDE, Demantra, Oracle E-Business, “DIY”
• Easily tweaked for customized criteria, custom tables
• Use “DIY” capabilities for anything else on your server
10
11. Why the Purge & Archive Project?
• Swamped by data
• Ongoing savings and performance enhancement
• Upgrade on the horizon – good time to clean house
• Legal requirements
• Disaster Recovery
12. Some Benefits of Archiving
• DB Maintenance Savings
• Data Storage Savings
• Batch processing improvements
• Index build times
• Quicker Upgrades
• Faster backups
• DISASTER RECOVERY
BENEFITS MULTIPLIED THROUGHOUT OTHER
COPIES OF PRODUCTION
13. Archiving in JDE
• Pre-ARCTOOLS:
• JDE built-in purges are few and far between
• Cause significant issues – integrity reports, orphaned records
• Customer development efforts proved difficult, risky, painful
A fairly common issue with most packaged applications.
13
14. Why is Purging so Difficult? Data Integrity…
• Textbook:
• You would think it’s just a matter of headers and details, but you would be wrong…
• JDE specific issues:
• Cross functional area data relationships (PO, inventory, AP, GL…)
• “Batches” in AP, AR, GL
• Programs expecting data in other areas (ex: AP inquiry looking for F0911)
• JDE integrity reports (ex: AP by Offset)
• Complex criteria
• Complex scenarios
• “Unfortunate realities” in JDE database
15. Top-Down
Header
Child 1 Child 2
Child 2A
Child 3
Simple scenario – date, status, etc, criteria from the header.
This is what you would think it looks like…. But in reality it’s rarely this easy.
16. Bottom-Up
Header
Child 1 Child 2
Child 2A
Child 3
Other Table(s)
Complex scenario – multiple criteria across multiple files, cross references with other files… This is far more
common, and often far more complex than this.
17. Many to Many
Very complex scenario – multiple criteria across possibly all files, cross references with other files…
many to many relationships…
Voucher 1
Payment 1
Payment 2
Voucher 3
Voucher 4
Voucher 6
Voucher 5
Voucher 7
Payment 4
Payment 5
Payment 3
Voided after
horizon
Voucher 1
F0911
18. And Don’t Forget…
• All those ‘unfortunate realities’ in the JDE database:
• Bad data (converted… self-inflicted… JDE errors)
• Duplicate keys (World more so than E1)
• Orphan records
• Inconsistent field definitions – numeric versus alpha
• Inconsistent key sequences – kco/doc/dct?? doc/dct/kco???
• F00165 keys (pipes… truncated leading zeroes… inconsistent sequence, even within the same parent
document…..)
• Right justified text columns???
• Hooks into F0911…
• Doc type? What doc type? Why would you need a doc type?
• Moving targets (ex: doc type buried in a random text column… in different positions… for the same document type…)
• … etc… etc… etc…
19. Case Study – French Opticians Guild
• Reduced database maintenance time from 22 to 8 hours per weekend, leading to an average annual saving of > $100k / year
• Archived 430 GB of data from production alone
• Reduced the production database by 55%
• Savings multiplied across environments
ARCTOOLS Project Kick Off
Regular Scheduling of
Purge & Archive Jobs
20. Case Study – George P. Johnson
During the upgrade to 9.0 data conversion times improved over prior upgrade:
• 8.10 Data Conversion: 36 Hours
• Purge and archive using ARCTOOLS
• 9.0 Data Conversion: Less than 20 Hours
“The benefits can be seen not just during the upgrade process, but also on a daily basis by end users and the IT department –
with less data in the production environment, end-user queries run more efficiently and database maintenance is simplified.”
Brian Bartkowiak
Vice President, Information Technology
Worldwide at George P. Johnson
21. DBG Canada
• Privately owned manufacturing company
• 35 years in business
• 2 facilities in Canada, 2 in Mexico, running 6x24
• Customers: truck, auto, military, consumer goods manufacturers
• Products:
• metal stampings
• stainless steel and carbon steel assemblies
• design and prototyping
22. JDE at DBG
• On JDE since 1998
• Painful tables:
• F0911 - > 70 million
• F4111
• F3111
• F42199
• … others…
23. DBG – Long Running Processes
• MRP
• Canada: 4 hours
• Mexico: 8 hours
• Backups
• “daily” backup took more than 24 hours
• Concerns about restore time and disaster recovery
• Restore takes as long (or worse??)
• Any backup was at least more than a day old
• Long lead time from lightning strike to up and running
• What about data updates since backup?
24. DBG – 9.1 Upgrade
• 8.10 to 9.1
• Also implementing virtualization at the same time
• VMWare, server farm, etc.
• Virtualized other things, too, not just JDE (domain controllers, etc)
• Timeline:
• March – new hardware
• November – go live
• But in July / August, the realization…
25. DBG Mid-upgrade Reality Check
Data upgrade run time would
shut down plants!!!
(Four plants, 700 workers…)
26. DBG Canada – Upgrade and ARCTOOLS
• Without purging, DBG’s upgrade from 8.10 to 9.1 would have shut down plants
and idled hundreds of workers
• DBG looked at their options and decided to go with archiving. With 15 years of
data, they were ready…
27. DBG - ARCTOOLS
• Archiving decision made in AUGUST
• ARCTOOLS installed August 16th.
• “Dropped in”, vanilla, with only 10 weeks to go-live upgrade
• Decided to just “throw ARCTOOLS at it…”. “followed the script”
• Implemented in parallel with on-going upgrade project
• Accounts payable, accounts receivable, sales orders, work orders, purchase orders,
inventory, general ledger
• Significant data reductions
• Purge / archive a success
• Upgrade successful, on time
• Realized performance improvements, DR improvements
28. DBG – ARCTOOLS Project Team
• Senior analyst – “go figure it out”
• Light on the consulting required
• More on the project management side of parallel upgrade project
• Set a horizon date – ended up at ‘current plus three’
29. DBG – Stats…
• Driver run times – “minutes” (5…. 30…)
• Archive run times – “minutes” (15… 40…)
• Throughputs ranging from 2… 10… 40 million rows an hour
30. DBG – Benefits of the Purge/Archive
• Speed and Run Times – MRP runs each dropped to under an hour (from 4 and 8
hours each previously)
• Backup time dropped from more than a day to under two hours
• DR improvements
• Multiple copies of PD (PY, CRP, other sandboxes) all benefitted
• Increased enthusiasm for new version of JDE
31. DBG – Upgrade Weekend
• Four plants – 700 employees
• Shut down JDE on Friday 6pm
• Took a backup, restored to new system, ran table conversions
• Up and running Monday morning.
“We did it all in a weekend, which would have been impossible without
ARCTOOLS.”
- Julian Wainewright, DBG Canada
32. DBG – Going Forward
• Nothing formalized yet, but…
• Planning to purge every year or two
• Have some data issues that were made visible by the purge
33. Why Archive Before the Upgrade?
“When you’re doing an upgrade project and you unveil something new to people,
before they get to ‘hey this is new, this is cool’, they are going to try what they
know already. If it’s slower than what they had before, then in the first couple
minutes it will kill your project. They haven’t even gotten to the ‘whiz-bang’ stuff.
If they try the new system and it’s not faster it can really kill adoption.”
- Julian Wainewright, IT Manager, DBG Canada
34. Access to the Archive Data
• What you do with the archived data is up to you… you can
• Destroy it
• Back it off to tape
• Move it to another platform
• Leave it on the system but not configure JDE access to it
• Leave it on the system and configure JDE access to “archive only”
• Leave it on the system and configure JDE access to “live plus archive”
35. JDE Access to Live – E1
Access to Archived data
using the
Production ‘role’
Live
Data
PRODUCTION
36. JDE Access to Archive – E1
Access to Archived data
using the
Archive ‘role’
Live
Data
Archive
Data
ARCHIVE
37. JDE Access to Live + Archive – E1
Access to Archived data
using the
Union ‘role’
Live
Data
Archive
Data
41. A Wise Choice!!
• JDE Market leader - over ten years, more than 400 JDE customers on six continents
• Supports E1 (XE to current), World (a6.2 to current) and co-existence
• Runs on IBM i (a.k.a. AS/400), Oracle, SQL
• Quick implementation
• No expensive additional hardware required, no expensive middleware, no requirement to extract all
related master data
Advantages of ARCTOOLS®
42. Performance!!
• A true database solution
• NOT a UBE (UBEs could be suh-lowwwwwww….)
• Not ODBC
• Not a fat client
• No JDE services required
• Not a proprietary hardware / middleware / compression software solution
• ARCTOOLS code generator creates purge/archive programs at run-time in the native language of the
platform, and then executes the process on the server, close to the data
• Throughputs typically measured in millions of rows per hour
Advantages of ARCTOOLS®
43. Simplicity!!
• No JDE services required
• No ASU – we won’t add hundreds of objects to your E1 environment
• No ESUs – no hassles with every JDE upgrade
• No illogical archive table names – data is archived to original JDE table name – i.e. F4301 is archived to
F4301, not FYK57901
• Archive data is not co-mingled with production
Advantages of ARCTOOLS®
44. Flexibility!!!
• Not hard-coded. Completely flexible – add custom tables and custom criteria quickly and easily.
• Add F55 tables directly to the related purge process, not as a ‘custom’, ‘one-off’ or ‘orphan’ purge.
• Delivered pre-configured for JDE, but can purge any tables on your server inside or outside of JDE.
Advantages of ARCTOOLS®
45. Easy Access to Archived JDE Data!!!
• Data archived to proper table name (i.e. F4301 F4301), not some random custom table name (i.e.
F4301 FYK97xxx)
• Easy to set up with your existing JDE objects (screens and reports) – not mocked up, look-alike, cloned
UBEs (which would require even more mods, ESUs, etc…)
Advantages of ARCTOOLS®
46. Summarization!!!
• Can use JDE summarization for GL Detail F0911 (R09811 / P09811) for the archived data, or;
• ARCTOOLS provides custom summarization for F0911
• ARCTOOLS summarization routines are easily customized (versus E1 BSFN, UBE or RPG)
• JDE provides no summarization for Cardex F4111… but ARCTOOLS DOES
Advantages of ARCTOOLS®
47. • Gather control totals, run ‘before’ reports (aging, etc)
• Run driver process from JDE Module
• Run sanity checks on drivers
• Run purge/archive process from ARCTOOLS
• Gather control totals, run ‘after’ reports for comparison
How to Run a Purge/Archive
48. Not Just for JDE… Pre-configured Modules for:
• JDE World
• JDE E1
• Oracle E-Biz Financials
• BPCS
• JDA
• Demantra
• Agile
• Maximo
• Others in development…
ARCTOOLS®
50. For More Information
• www.arctools.com
• Case Studies
• White Papers
• Webinars, demo requests, speaking schedules, user groups
• www.jdealliance.com
• ARCTOOLS and other “best of breed” bolt-ons for JDE
• Joe DiPietro
• (+1) 508.435.8243 x201
• joe@arctools.com
50
51. Go to ARCTOOLS.com for more information
@jdearctools
facebook.com/ARCTOOLS
ARCTOOLS Channel
ARCTOOLS Group
Joe DiPietro
• 508.435.8243 x201
• joe@arctools.com
52. Please complete the session evaluation
We appreciate your feedback and insight
53. Join more than 500 other
JD Edwards users at
April 12-16, 2015
Mandalay Bay- Las Vegas, NV
• Keynote from Lyle Ekdahl,
Vice President of JD Edwards
• JD Edwards’ Roadmap and Strategy
(Including Big Announcements)
• Education around JD Edwards with Edge products, macro-level
content and high-level strategy
• More than 230 Solution Providers across several Oracle Products
register now!
QuestDirect.org/COLLABORATE