Containers are increasingly popular to package, ship and run applications or microservices with their completely configured runtime environment including platform components such as application server and data store.Continuous Delivery and automated DevOps hinge on containers. Docker Containers are widely used and Oracle has long been involved in the Docker community.This session introduces the Docker Container images published by Oracle for flagship products such as Database, WebLogic, Linux and Java and demonstrates how these can be used in environment provisioning, automated delivery pipelines and microservices architectures. The session shows how containers are built, shipped and run based on these images and shows the Oracle Container Cloud, as well as Wercker Cloud (for automated build and delivery pipelines) and Oracle Cloud Engine – the managed Kubernetes cloud service.
GCP - Continuous Integration and Delivery into Kubernetes with GitHub, Travis...Oleg Shalygin
Kubernetes provides an automated platform to deployment, scaling and operations of applications across a cluster of hosts. Complementing Kubernetes with a series of build scripts in conjunction with Travis-CI, GitHub, Artifactory, and Google Cloud Platform, we can take code from a merged pull request to a deployed environment with no manual intervention on a highly scaleable and robust infrastructure.
Jordi Mon Companys presents an overview of Weave GitOps Core for the Free GitOps Workshop on August 19, 2021.
Weave GitOps Core is a continuous delivery product to run apps in any Kubernetes. It is free and open source, and you can get started today!
https://www.weave.works/product/gitops-core/
Chat with us on our Slack channel! #weave-gitops http://bit.ly/WeaveGitOpsSlack
If you need to invite yourself to the Slack, visit https://slack.weave.works/
Francisco Javier Ramírez Urea - IT Architect, Hoplasoftware
Guillaume Morini - SE, Docker
The integration of Kubernetes orchestration into the Docker Enterprise Platform presents deployments with interesting new abstractions for application connectivity. Devs and Ops are often challenged with rationalizing how pod networking (with CNI plugins like Calico or Flannel), Services (via kube-proxy) and Ingress work in concert to enable application connectivity within and outside a cluster. Similarly, given the dynamic and transient nature of containerized microservice workloads, how to leverage scalable and declarative approaches like network policies to express segmentation and security primitives. This session provides an illustrative walkthrough of these core concepts by going through common deployment architectures providing design, operations, and scale considerations based on experience from numerous production deployments. We will discuss Kubernetes publishing methods and deep dive into Ingress Controllers. This session will also showcase how to complement application and operations workflows with policy-driven business, compliance and security controls typically required in enterprise production deployments including going further into limiting traffic to services, session persistence, rewriting, and activating container health checks.
Evénement Docker Paris: Anticipez les nouveaux business model et réduisez vos...Docker, Inc.
Au programme : la mise en place de plateformes agiles pour s’adapter aux nouveaux business models, l’optimisation des coûts IT dans le cadre de vos déploiements applicatifs, réussir la mise en oeuvre de Kubernetes, garantir la sécurité de vos applications tout au long de leur cycle de vie et bien plus encore.
GCP - Continuous Integration and Delivery into Kubernetes with GitHub, Travis...Oleg Shalygin
Kubernetes provides an automated platform to deployment, scaling and operations of applications across a cluster of hosts. Complementing Kubernetes with a series of build scripts in conjunction with Travis-CI, GitHub, Artifactory, and Google Cloud Platform, we can take code from a merged pull request to a deployed environment with no manual intervention on a highly scaleable and robust infrastructure.
Jordi Mon Companys presents an overview of Weave GitOps Core for the Free GitOps Workshop on August 19, 2021.
Weave GitOps Core is a continuous delivery product to run apps in any Kubernetes. It is free and open source, and you can get started today!
https://www.weave.works/product/gitops-core/
Chat with us on our Slack channel! #weave-gitops http://bit.ly/WeaveGitOpsSlack
If you need to invite yourself to the Slack, visit https://slack.weave.works/
Francisco Javier Ramírez Urea - IT Architect, Hoplasoftware
Guillaume Morini - SE, Docker
The integration of Kubernetes orchestration into the Docker Enterprise Platform presents deployments with interesting new abstractions for application connectivity. Devs and Ops are often challenged with rationalizing how pod networking (with CNI plugins like Calico or Flannel), Services (via kube-proxy) and Ingress work in concert to enable application connectivity within and outside a cluster. Similarly, given the dynamic and transient nature of containerized microservice workloads, how to leverage scalable and declarative approaches like network policies to express segmentation and security primitives. This session provides an illustrative walkthrough of these core concepts by going through common deployment architectures providing design, operations, and scale considerations based on experience from numerous production deployments. We will discuss Kubernetes publishing methods and deep dive into Ingress Controllers. This session will also showcase how to complement application and operations workflows with policy-driven business, compliance and security controls typically required in enterprise production deployments including going further into limiting traffic to services, session persistence, rewriting, and activating container health checks.
Evénement Docker Paris: Anticipez les nouveaux business model et réduisez vos...Docker, Inc.
Au programme : la mise en place de plateformes agiles pour s’adapter aux nouveaux business models, l’optimisation des coûts IT dans le cadre de vos déploiements applicatifs, réussir la mise en oeuvre de Kubernetes, garantir la sécurité de vos applications tout au long de leur cycle de vie et bien plus encore.
This presentation is to reflect on the amazing advancement of the open source community in the field of Cloud Computing and how does it now allow us to build reliable software components quickly within truly agile infrastructure.
We are on the cusp of a new era of application development software: instead of bolting on operations as an after-thought to the software development process, Kubernetes promises to bring development and operations together by design.
More and more businesses are requiring developers to own end to end delivery, including operational ownership. Weaveworks will share with you what GitOps means, and how easy it is to create cloud native applications, CICD pipelines, integrate operations and more, using GitOps.
Inherited from best practices going back 10-15 years, cloud native is making these practices more relevant today. At Weaveworks, they implement these principles in their product, Weave Cloud. This not only helps customers ship apps faster, it also helps them run their own cloud native stack. This presentation will show how Weaveworks does this, identify best practices and tools, and showcase some of Weaveworks’ use cases.
For the video of this presentation at Cloud Native London visit: https://skillsmatter.com/skillscasts/10506-keynote-by-alexis-richardson
To learn more about Weaveworks: www.weave.works
Fabio Ferrari | particles.io | PresentationFabio Ferrari
Results-driven, fully organized and qualified IT professional in cloud infrastructure design and automation for microservices architectures based on Docker containers and Kubernetes.
Google Certified Cloud Architect with a long experience in *nix systems management and administration, excellent skills on Google Cloud Platform and Google ecosystem integrations (Google SDK and Google API).
Detail-oriented DevOps Engineer accustomed to working as remote worker freelance in fast paced/multitasking
distributed environments.
How to build an event-driven, polyglot serverless microservices framework on ...Animesh Singh
Serverless cloud platforms are a major trend in 2016. Following on from Amazon’s Lambda service, released last year, this year has seen Google, IBM and Microsoft all launch their own solutions. Serverless microservices are executed on-demand, in milliseconds, rather than having to sit idle waiting. Users pay only for the raw computation time used.
In this talk detail how to build a distributed serverless, event-driven, microservices framework on OpenStack
The combination of StackPointCloud with NetApp creates NetApp Kubernetes Service, the industry’s first complete Kubernetes platform for multi-cloud deployments and a complete cloud-based stack for Azure, Google Cloud, AWS, and NetApp HCI. Further, Trident is a fully supported open source project maintained by NetApp, designed from the ground up to help meet the sophisticated persistence demands of containerized applications.
How to build your containerization strategyDocker, Inc.
The Docker Enterprise Edition platform helps customers deploy and manage applications faster and it secures the application pipeline at a lower cost than traditional application delivery models. But it takes more than just great technology to achieve the desired results. The organization and culture of your enterprise directly impacts what you transform, how it’s done, and who does it. Success requires a strategy for how you will govern the Docker EE container platform, how to assess your application estate, what your delivery pipeline will look like, and how to ensure developers, operators, security teams and others play nicely together.
In this talk I will cover topics such as different types of workloads (legacy, microservices, FaaS, big data, ...), how your org chart can influence whether you deploy a CaaS (Containers as a Service) vs CLaaS (Clusters as a Service), how "shifting left" can determine if you can outsource, centralized vs distributed CI/CD and how containers play a role, transforming your pets into cattle, how giant whale balloons are used for onboarding, and a prescriptive and comprehensive methodology for successfully deploying Docker in your enterprise.
DCEU 18: Desigual Transforms the In-Store Experience with Docker Enterprise C...Docker, Inc.
Mathias Kriegel - IT Operations, Desigual
Joan Anton Sances - Software Architect, Desigual
Desigual, a $1-billion-dollar fashion retailer headquartered in Barcelona, operates over 500 stores worldwide. The company is on a digital transformation journey touching every aspect of the customer experience. In this session, IT Operations and Software Architecture teams, will explain how Desigual built an in-store “assistant shopping” that transformed the customer experience adopting modern architecture models leveraging Docker Enterprise for containerization. In the session, you’ll learn: ● How Desigual is leveraging containers with Docker Enterprise, micro services, API´s, CI/CD and hybrid cloud to create an excellent customer experience. ● How to use a container platform to accelerate time-to-market for new applications. ● How Desigual changed its traditional IT operational model, focusing on bringing a PaaS like model for Developer teams, and what they learned along the way. ● How Dev and Ops teams aligned together in the process. ● How Developer productivity increased by adopting modern architecture models.
DCEU 18: Continuous Delivery with Docker Containers and Java: The Good, the B...Docker, Inc.
Daniel Bryant - Product Architect, Datawire
Implementing a continuous delivery (CD) pipeline for Java applications is not trivial, and the introduction of container technology to the development stack can introduce additional challenges and requirements. In this talk we will look at the high-level steps that are essential for creating an effective pipeline for creating and deploying Docker container-based Java applications. Key takeaways include: - The impact of using Docker containers on Java and Continuous Delivery - The benefits and challenges of packaging containerised Java applications - Options for adding metadata to container images - Validating nonfunctional/operational requirement changes imposed by executing Java applications within a container - Lessons learned the hard way (in production, at 3am, with lots of coffee)
DCEU 18: App-in-a-Box with Docker Application PackagesDocker, Inc.
Michael Irwin - Application Architect, Virginia Tech
Docker Application Packages is an experimental tool that makes it easy to share multi-service applications. Create a Compose file, package it in an image, and voilà! You now have an "app-in-a-box"! Not convinced yet? No worries! It took a while for me to be convinced too! In this session, we'll start off by diving into how Docker Application Packages actually works, which will help us understand the use cases. We'll see how dev environments can hook in to this app-in-a-box by replacing the service being worked on with a dev container. Then we'll move on to see how end-to-end functional tests are much easier to run. And, finally, we'll see how to maintain an "app-in-a-box" with the latest versions of each component in a CI/CD pipeline, allowing for a unique app-in-a-box for each feature branch under development. Lots of good material! And lots of live demos!
For years people have been using VM-based CI platforms where they are managing build nodes that run their CI workflows.
A few years ago, Codefresh revolutionized the CI/CD world and became the first container-native CI/CD platform.
**WATCH THE WEBINAR AT https://Codefresh.io/events **
In this webinar, we will look at the differences between VM-based CI pipelines and Docker-based CI pipelines, in terms of maintenance, upgrades, pipeline creation, caching, and speed!
Cloud native applications are popular these days. They promise superior reliability and almost arbitrary scalability. They follow three key principles: they are built and composed as microservices. They are packaged and distributed in containers. The containers are executed dynamically in the cloud. But which technology is best to build this kind of application? This talk will be your guidebook.
In this hands-on session, we will briefly introduce the core concepts and some key technologies of the cloud native stack and then show how to build, package, containerize, compose and orchestrate a cloud native showcase application on top of a cluster operating system such as Kubernetes or OpenShift. Throughout the session we will be using an off-the-shelf MIDI controller to visualize the concepts and to remote control the cluster.
Container Days 2017 conference. @ConDaysEU #CDS17 #qaware #CloudNativeNerd @LeanderReimer
Introduction into Docker Containers, the Oracle Platform and the Oracle (Nati...Lucas Jellema
Containers are increasingly popular to package, ship and run applications or microservices with their completely configured runtime environment including platform components such as application server and data store.Continuous Delivery and automated DevOps hinge on containers. Docker Containers are widely used and Oracle has long been involved in the Docker community.This session introduces the Docker Container images published by Oracle for flagship products such as Database, WebLogic, Linux and Java and demonstrates how these can be used in environment provisioning, automated delivery pipelines and microservices architectures. The session shows how containers are built, shipped and run based on these images and shows the Oracle Container Cloud, as well as Wercker Cloud (for automated build and delivery pipelines) and Oracle Cloud Engine - the managed Kubernetes cloud service..
Docker Concepts for Oracle/MySQL DBAs and DevOpsZohar Elkayam
Oracle Week 2017 Slides
Agenda:
Docker overview – why do we even need containers?
Installing Docker and getting started
Images and Containers
Docker Networks
Docker Storage and Volumes
Oracle and Docker
Docker tools, GUI and Swarm
This presentation is to reflect on the amazing advancement of the open source community in the field of Cloud Computing and how does it now allow us to build reliable software components quickly within truly agile infrastructure.
We are on the cusp of a new era of application development software: instead of bolting on operations as an after-thought to the software development process, Kubernetes promises to bring development and operations together by design.
More and more businesses are requiring developers to own end to end delivery, including operational ownership. Weaveworks will share with you what GitOps means, and how easy it is to create cloud native applications, CICD pipelines, integrate operations and more, using GitOps.
Inherited from best practices going back 10-15 years, cloud native is making these practices more relevant today. At Weaveworks, they implement these principles in their product, Weave Cloud. This not only helps customers ship apps faster, it also helps them run their own cloud native stack. This presentation will show how Weaveworks does this, identify best practices and tools, and showcase some of Weaveworks’ use cases.
For the video of this presentation at Cloud Native London visit: https://skillsmatter.com/skillscasts/10506-keynote-by-alexis-richardson
To learn more about Weaveworks: www.weave.works
Fabio Ferrari | particles.io | PresentationFabio Ferrari
Results-driven, fully organized and qualified IT professional in cloud infrastructure design and automation for microservices architectures based on Docker containers and Kubernetes.
Google Certified Cloud Architect with a long experience in *nix systems management and administration, excellent skills on Google Cloud Platform and Google ecosystem integrations (Google SDK and Google API).
Detail-oriented DevOps Engineer accustomed to working as remote worker freelance in fast paced/multitasking
distributed environments.
How to build an event-driven, polyglot serverless microservices framework on ...Animesh Singh
Serverless cloud platforms are a major trend in 2016. Following on from Amazon’s Lambda service, released last year, this year has seen Google, IBM and Microsoft all launch their own solutions. Serverless microservices are executed on-demand, in milliseconds, rather than having to sit idle waiting. Users pay only for the raw computation time used.
In this talk detail how to build a distributed serverless, event-driven, microservices framework on OpenStack
The combination of StackPointCloud with NetApp creates NetApp Kubernetes Service, the industry’s first complete Kubernetes platform for multi-cloud deployments and a complete cloud-based stack for Azure, Google Cloud, AWS, and NetApp HCI. Further, Trident is a fully supported open source project maintained by NetApp, designed from the ground up to help meet the sophisticated persistence demands of containerized applications.
How to build your containerization strategyDocker, Inc.
The Docker Enterprise Edition platform helps customers deploy and manage applications faster and it secures the application pipeline at a lower cost than traditional application delivery models. But it takes more than just great technology to achieve the desired results. The organization and culture of your enterprise directly impacts what you transform, how it’s done, and who does it. Success requires a strategy for how you will govern the Docker EE container platform, how to assess your application estate, what your delivery pipeline will look like, and how to ensure developers, operators, security teams and others play nicely together.
In this talk I will cover topics such as different types of workloads (legacy, microservices, FaaS, big data, ...), how your org chart can influence whether you deploy a CaaS (Containers as a Service) vs CLaaS (Clusters as a Service), how "shifting left" can determine if you can outsource, centralized vs distributed CI/CD and how containers play a role, transforming your pets into cattle, how giant whale balloons are used for onboarding, and a prescriptive and comprehensive methodology for successfully deploying Docker in your enterprise.
DCEU 18: Desigual Transforms the In-Store Experience with Docker Enterprise C...Docker, Inc.
Mathias Kriegel - IT Operations, Desigual
Joan Anton Sances - Software Architect, Desigual
Desigual, a $1-billion-dollar fashion retailer headquartered in Barcelona, operates over 500 stores worldwide. The company is on a digital transformation journey touching every aspect of the customer experience. In this session, IT Operations and Software Architecture teams, will explain how Desigual built an in-store “assistant shopping” that transformed the customer experience adopting modern architecture models leveraging Docker Enterprise for containerization. In the session, you’ll learn: ● How Desigual is leveraging containers with Docker Enterprise, micro services, API´s, CI/CD and hybrid cloud to create an excellent customer experience. ● How to use a container platform to accelerate time-to-market for new applications. ● How Desigual changed its traditional IT operational model, focusing on bringing a PaaS like model for Developer teams, and what they learned along the way. ● How Dev and Ops teams aligned together in the process. ● How Developer productivity increased by adopting modern architecture models.
DCEU 18: Continuous Delivery with Docker Containers and Java: The Good, the B...Docker, Inc.
Daniel Bryant - Product Architect, Datawire
Implementing a continuous delivery (CD) pipeline for Java applications is not trivial, and the introduction of container technology to the development stack can introduce additional challenges and requirements. In this talk we will look at the high-level steps that are essential for creating an effective pipeline for creating and deploying Docker container-based Java applications. Key takeaways include: - The impact of using Docker containers on Java and Continuous Delivery - The benefits and challenges of packaging containerised Java applications - Options for adding metadata to container images - Validating nonfunctional/operational requirement changes imposed by executing Java applications within a container - Lessons learned the hard way (in production, at 3am, with lots of coffee)
DCEU 18: App-in-a-Box with Docker Application PackagesDocker, Inc.
Michael Irwin - Application Architect, Virginia Tech
Docker Application Packages is an experimental tool that makes it easy to share multi-service applications. Create a Compose file, package it in an image, and voilà! You now have an "app-in-a-box"! Not convinced yet? No worries! It took a while for me to be convinced too! In this session, we'll start off by diving into how Docker Application Packages actually works, which will help us understand the use cases. We'll see how dev environments can hook in to this app-in-a-box by replacing the service being worked on with a dev container. Then we'll move on to see how end-to-end functional tests are much easier to run. And, finally, we'll see how to maintain an "app-in-a-box" with the latest versions of each component in a CI/CD pipeline, allowing for a unique app-in-a-box for each feature branch under development. Lots of good material! And lots of live demos!
For years people have been using VM-based CI platforms where they are managing build nodes that run their CI workflows.
A few years ago, Codefresh revolutionized the CI/CD world and became the first container-native CI/CD platform.
**WATCH THE WEBINAR AT https://Codefresh.io/events **
In this webinar, we will look at the differences between VM-based CI pipelines and Docker-based CI pipelines, in terms of maintenance, upgrades, pipeline creation, caching, and speed!
Cloud native applications are popular these days. They promise superior reliability and almost arbitrary scalability. They follow three key principles: they are built and composed as microservices. They are packaged and distributed in containers. The containers are executed dynamically in the cloud. But which technology is best to build this kind of application? This talk will be your guidebook.
In this hands-on session, we will briefly introduce the core concepts and some key technologies of the cloud native stack and then show how to build, package, containerize, compose and orchestrate a cloud native showcase application on top of a cluster operating system such as Kubernetes or OpenShift. Throughout the session we will be using an off-the-shelf MIDI controller to visualize the concepts and to remote control the cluster.
Container Days 2017 conference. @ConDaysEU #CDS17 #qaware #CloudNativeNerd @LeanderReimer
Introduction into Docker Containers, the Oracle Platform and the Oracle (Nati...Lucas Jellema
Containers are increasingly popular to package, ship and run applications or microservices with their completely configured runtime environment including platform components such as application server and data store.Continuous Delivery and automated DevOps hinge on containers. Docker Containers are widely used and Oracle has long been involved in the Docker community.This session introduces the Docker Container images published by Oracle for flagship products such as Database, WebLogic, Linux and Java and demonstrates how these can be used in environment provisioning, automated delivery pipelines and microservices architectures. The session shows how containers are built, shipped and run based on these images and shows the Oracle Container Cloud, as well as Wercker Cloud (for automated build and delivery pipelines) and Oracle Cloud Engine - the managed Kubernetes cloud service..
Docker Concepts for Oracle/MySQL DBAs and DevOpsZohar Elkayam
Oracle Week 2017 Slides
Agenda:
Docker overview – why do we even need containers?
Installing Docker and getting started
Images and Containers
Docker Networks
Docker Storage and Volumes
Oracle and Docker
Docker tools, GUI and Swarm
Intro to Docker at the 2016 Evans Developer relations conferenceMano Marks
Building large scale apps traditionally has traditionally meant building large monolithic apps to handle everything. In the new age of the cloud and on premise data centers, increasingly the world is looking to containers and microservices. This allows flexibility and agility. Individual teams can choose the tools they need and be assured they'll work in the environment they want. And it also has implications for how we do developer relations, making it easier to deploy samples without worrying about environment. This session will look at microservices and how they are changing both the enterprise, and our work in developer relations.
Automated Testing with Docker on Steroids - nlOUG TechExperience 2018 (Amersf...Lucas Jellema
Automated testing is important. We all know that we should do it. We also know that this can be painful, for many reasons. One of the most agonizing aspects of automated testing is the handling of the data. In order to run even the simplest of tests against the user interface, a service or API or even a PL/SQL unit typically requires that a proper starting point needs to be established in the database with respect to the data. Complex set up steps need to prepare various records to ensure the test can even start and afterwards in similarly complex tear down scripts we have to clean up after the test.
This session demonstrates how this hardship can be a thing of the past. Using snapshots of a test database in a Docker container with a managed test data set that supports all tests, we can create automated tests without any set up or tear down effort. These tests can run very fast, concurrently, and whenever and wherever you like them to run. This way of working opens up much higher test coverage and much increased productivity for developers and testers.
Docker Birthday #3 - Intro to Docker SlidesDocker, Inc.
High level overview of Docker + Birthday #3 overview (app and challenge portion)!
Learn more about Docker Birthday #3 celebrations here: https://www.docker.com/community/docker-birthday-3
Adopting Docker for production applications and services used to be hard. You had to hand-roll a lot of the underlying infrastructure and write lots of custom code for service discovery, load balancing, orchestration, desired state, etc. Today, with the rise of open source container orchestration platforms and cloud-native offerings, it's a lot easier to get up and running.
Github repo for demo: https://github.com/elabor8/dockertalk
Dot net platform and dotnet core fundamentalsLalit Kale
This is the presentation deck, I did for LimerickDotNet-Azure User group.
Event Url: https://www.meetup.com/Limerick-DotNet/events/240897689/
Session Details:
This session represented .NET journey of almost 17 years. Through this slid-deck, I narrated .NET platform progression till .NET Standards 2.0.
This session was accompanied by a small demo of running small dotnet program on alpine linux with docker container.
There's a new Docker release, and with it lots of changes. In this video Mano Marks, Docker Developer Relations Director, highlights some of the biggest new features.
Docker 1.9 Release Blog Post:
http://blog.docker.com/2015/11/docker-1-9-production-ready-swarm-multi-host-networking
Docker 1.9 Release Notes:
https://github.com/docker/docker/blob/master/CHANGELOG.md
Docker Swarm 1.0 Blog Post:
http://blog.docker.com/2015/11/swarm-1-0
Docker Multi Host Networking Documentation:
http://docs.docker.com/engine/userguide/networking/
Docker Swarm Documentation:
https://docs.docker.com/swarm
Docker Compose Documentation:
https://docs.docker.com/compose
Online Meetup on Multi Host Networking:
http://www.meetup.com/Docker-Online-Meetup/events/226522306/
Online Meetup on Swarm:
http://www.meetup.com/Docker-Online-Meetup/events/226520109/
Docker is an open platform for developers and system administrators to build, ship and run distributed applications. With Docker, IT organizations shrink application delivery from months to minutes, frictionlessly move workloads between data centers and the cloud and can achieve up to 20X greater efficiency in their use of computing resources. Inspired by an active community and by transparent, open source innovation, Docker containers have been downloaded more than 700 million times and Docker is used by millions of developers across thousands of the world’s most innovative organizations, including eBay, Baidu, the BBC, Goldman Sachs, Groupon, ING, Yelp, and Spotify. Docker’s rapid adoption has catalyzed an active ecosystem, resulting in more than 180,000 “Dockerized” applications, over 40 Docker-related startups and integration partnerships with AWS, Cloud Foundry, Google, IBM, Microsoft, OpenStack, Rackspace, Red Hat and VMware.
Docker is a tool designed to make it easier to create, deploy, and run applications
by using containers. Containers allow a developer to package up
an application with all of the parts it needs, such as libraries and other dependencies,
and ship it all out as one package. By doing so, thanks to the
container, the developer can rest assured that the application will run on
any other Linux machine regardless of any customized settings that machine
might have that could differ from the machine used for writing and testing
the code.
In a way, Docker is a bit like a virtual machine. But unlike a virtual
machine, rather than creating a whole virtual operating system, Docker allows
applications to use the same Linux kernel as the system that they’re
running on and only requires applications be shipped with things not already
running on the host computer. This gives a significant performance boost
and reduces the size of the application.
Introduction to web application development with Vue (for absolute beginners)...Lucas Jellema
In this slide deck I show you how you can easily and quickly create quite rich web applications with Vue 3 – without having to study complex concepts or understand many technical details. I have only recently learned how to work with Vue 3 myself and now is the best time for me to share my learning experience (and my enthusiasm) with you. I know what I found essential to understand and what most got me excited in these early steps (what was a little bit hard to grasp). I believe that I can present my steps and guide you to experience the same fun and have a similarly gratifying experience. I am not an expert in this subject – I have barely learned how to walk and that is why I can help you with these first steps with Vue.
In this deck, I do not explain how Vue works. I do not really know that. I will show you how to work with it and how to create web applications that are functional, appealing, fast and responsive.
The approach I am taking is straightforward:
• I will tell you a little bit about web development, browsers and reactive frameworks
• I will show the hello world of Vue applications
• I will explain about components and nesting, events, data binding and reactive behavior and demonstrate these concepts
• I will introduce Vue UI Component libraries – and with no effort at all we will launch our application to the next level – with rich components to explore, manipulate, visualize data collections
• We will publish the web application from our development environment to where the whole world could see it – using GitHub Pages
• As bonus topic – we discuss state management
At the end of this session you will be able to quickly create a simple yet rich web application with Vue 3. You have a starting point to further evolve your skills with the many online resources I am convinced that you will enjoy your newfound powers and the simplicity and power of Vue 3.
Note: a tutorial accompanies this slide deck - see https://github.com/lucasjellema/code-face-vue3-intro-reactiive-webapps-aug2023/blob/main/README.md
Making the Shift Left - Bringing Ops to Dev before bringing applications to p...Lucas Jellema
Designing, agreeing on, implementing and testing the application is our first challenge. But it does not end there. Applications require tender love and care when they are live. Application Operations needs to be in place along with the functionality of the application. AppOps is the process of making sure that the applications are executed as required and that any problems are detected, reported and dealt with. Some mechanisms used in AppOps: transaction tracing, log analysis, post-data-exchange-checks, health checking of all systems involved, in-production-testing of end-to-end data flows. Additionally, AppOps takes care of configuration management, scaling, cost management, technical life cycle management on solution components. In this session, we will take a closer look at what is required to keep those applications going and how we do ops by design from early on in the agile process.
Lightweight coding in powerful Cloud Development Environments (DigitalXchange...Lucas Jellema
Cloud Based Development environments allow software engineers to work in a new and refreshing way. The development environment runs in the cloud, based on a coded environment definition and with the sources from a specific branch in a Git repository. The environment can be quite powerful in memory, CPU and storage. Development can be done from a lightweight device such as a Chromebook or even a tablet. Switching between different environments becomes a breeze, collaborating in an environment is easily done. Using network tunneling, the IDE could run locally against the remote workspace and remote ports can be accessed on localhost. This session demonstrates both Gitpod and Github Codespaces - similar SaaS offerings with generous free tiers. They are great for quick investigation into new technologies, for working through tutorials and for contributing to open source projects. You will smile at the ease and elegance of engineering your software in this way.
Apache Superset - open source data exploration and visualization (Conclusion ...Lucas Jellema
Introducing Apache Superset - an open source platform for data exploration, visualization and analysis - co-starring Trino and Steampipe for providing SQL access to many non-SQL data sources.
CONNECTING THE REAL WORLD TO ENTERPRISE IT – HOW IoT DRIVES OUR ENERGY TRANSI...Lucas Jellema
Enterprise IT systems are deaf, blind and highly insensitive. They do not know what is going on in the outside world. Through Internet of Things technology, we provide eyes, ears and hands that allow enterprises to learn about and react in real time to events in the physical world. The energy transition at a major Dutch energy company (Eneco) is powered by IoT technology – to steer and sometimes curtail windmills and solar farms and to coordinate local energy production and trade. This session shows you how the physical world was connected to the customer portals and apps, asset management systems and Kafka platform through the Azure cloud based IoT Hub en Edge, digital twin, serverless functions, timeseries datastores and streaming data analysis. It is a story about technological innovation on top of existing foundations and of a vision for business and our society at large.
Help me move away from Oracle - or not?! (Oracle Community Tour EMEA - LVOUG...Lucas Jellema
I hear this aspiration from a growing number of organizations. Sometimes as a quite literal question. This however is merely half of a wish. Apparently, organizations want to quit with one thing — but have not yet stipulated what they desire instead. What is the objective that is pursued here? Only to get rid of Oracle? It will become clear why you should give a considerable thought about dropping Oracle, or any other vendors’ technology, when you’re not pleased with your current IT situation. You need to focus on the actual problems and objectives and define the suitable roadmap to fit your real needs. It turns out that the quest is usually for modernization and flexibility - and Oracle can very well be a part of that future.
Organizations with decades of investment in Oracle technology sometimes (and increasingly) express a wish to move away from Oracle. In this session, we will first explore where the desire to move away from Oracle might come from. Then we describe what the term Oracle represents — more than 2.000 products on all layers in the technology stack and in different business areas. Finally, we map out what the ‘moving away from’ consists of: defining where you ‘move to’ and subsequently actually going there.
It will become clear why you should give considerable thought about dropping Oracle, or any other vendors’ technology, when you’re not pleased with your current IT situation. You need to focus on the actual problems and objectives and define the suitable roadmap to fit your real needs. It turns out that the quest is usually for modernization and flexibility - and Oracle can very well be a part of that future.
Original storyline in this Medium Article: https://medium.com/real-vox/what-if-companies-say-help-me-move-away-from-oracle-ffbbc95afc4f
IoT - from prototype to enterprise platform (DigitalXchange 2022)Lucas Jellema
In 2019 the company started a small scale IoT project: smart meters in consumer homes, a cloud based IoT platform for device management, metrics collecting, monitoring and real time data processing. From the initial 12 devices and this single use case, the initiative has rapidly scaled, to tens of thousands devices - including entire wind parks and solar farms - and seven substantial business cases, not just for harvesting data but increasingly for real time actuation. The IoT Platform is feeding the brain at the heart of the enterprise - through an event streaming platform and an API platform. It supports complex operations with anomaly detection on metrics streams and device and communication monitoring. This session tells about the eye catching business cases - what are business objectives and results - and explains the journey since the start. It continues the story presented at DigitalXchange 2020 - discussing technical challenges and solutions as well as organizational aspects. Areas of particular interest: edge processing, data analytics and machine learning.
Who Wants to Become an IT Architect-A Look at the Bigger Picture - DigitalXch...Lucas Jellema
Pitch: The movie The Matrix made it clear: The Architect is powerful. How to be(come) and IT architect? What do you do, what do you need to know, is it fun and why? Using real world examples, core principles and useful tools, this session introduces the subtle art of designing and realizing flexible IT architectures. </p><p>Taking a step back to get and create an overview, frequently asking why to get to the real intention, bringing aspects such as cost, scale, time and change and business strategy into the design and bridging the gap between business owners, process managers and technical specialists. One way to define the responsibility of an IT architect. In this session, we will discuss what is expected of the architect and what you need to do for that and what you could use to get it done. How do you get started as an architect, how to grow in that role? We discuss a number of real life architectural challenges and solution design. And discuss a number of architecture principles, patterns, and powers to apply. Never stop programming - but perhaps rise to the architecture challenge too.
Notes: Many IT professionals aspire to become architects. Many architects wonder what it is they have to do. After 27 years in IT I find I have slowly and steadily moved into a role that I can probably use the label architect for, although still with some reluctance. What exactly does that mean - IT architect? While I may not have all answers and the ultimate truth and wisdom, I do have many architectural challenges to discuss and some core principles to share and a number of tips, tricks and tools to recommend that will help anyone get started or grow in a role as architect for software and IT systems. Elements that make an appearance include cloud, agile, DevOps, microservices, persistence, business, powers of persuasion, diagramming, cost, security, software engineering, data.
Outline: - two real world examples (one new business initiative, one running and struggling project) and how to approach them with an architect's mind - core principles to apply , patterns to us, what to unearth (the power question of WHY) - architecture products: what do you deliver as an architect; how do you ensure agility? - how to be effective? bringing your design to life - communication with stakeholders/powers of persuasion, monitoring adherence, being pragmatic but not lose grip; - anecdotal evidence from several small and large product teams - the good and also the ugly (architectural oversights and the consequences)
some specific answers to address - how much technical knowledge and programming skills does an architect require? What other knowledge is required and how to stay on top of your game? how to get going: first steps towards be(com)ing and architect?
Steampipe - use SQL to retrieve data from cloud, platforms and files (Code Ca...Lucas Jellema
Introduction to Steampipe - a tool for retrieving data and metadata about cloud resources, platform resources and file content - all through SQL. Data from clouds, files and platforms can be joined, filtered, sorted, aggregated using regular SQL. Steampipe offers a very convenient way to get hold of data that describes the environment in detail.
Automation of Software Engineering with OCI DevOps Build and Deployment Pipel...Lucas Jellema
Automation of software delivery has several advantages. Prevention of human error is certainly one. Consistent and complete execution of tried and tested build and deployment tasks as the only way to apply changes in the live environment. Once the pipelines have been set up, the engineers can focus on the software and applying the required changes to it. To bring that software all the way to production is a breeze. Oracle Cloud Infrastructure offers the DevOps service, introduced in the Summer of 2021. This service comes with git style code repositories, build servers and build pipelines, artifact repositories as well as deployment pipelines. This session introduces OCI DevOps and demonstrates how software can be built and deployed on OKE Kubernetes, Compute Instance VMs and Oracle Functions. From simple source code an application is put in production without manual intervention in the build and deployment process.
Introducing Dapr.io - the open source personal assistant to microservices and...Lucas Jellema
Dapr.io is an open source product, originated from Microsoft and embraced by a broad coalition of cloud suppliers (part of CNFC) and open source projects. Dapr is a runtime framework that can support any application and that especially shines with distributed applications - for example microservices - that run in containers, spread over clouds and / or edge devices.
With Dapr you give an application a "sidecar" - a kind of personal assistant that takes care of all kinds of common responsibilities. Capturing and retrieving state, publishing and consuming messages or events. Reading secrets and configuration data. Shielding and load balancing over service endpoints. Calling and subscribing to all kinds of SaaS and PaaS facilities. Logging traces across all kinds of application components and logically routing calls between microservices and other application components. Dapr provides generic APIs to the application (HTTP and gRPC) for calling all these generic services – and provides implementations of these APIs for all public clouds and dozens of technology components. This means that your application can easily make use of a wide range of relevant features - with a strict separation between the language the application uses for this (generic, simple) and the configuration of the specific technology (e.g. Redis, MySQL, CosmosDB, Cassandra, PostgreSQL, Oracle Database, MongoDB, Azure SQL etc) that the Dapr sidecar uses. Changing technology does not affect the application, but affects the configuration of the Sidecar. Dapr can be used from applications in any technology - from Java and C#/.NET to Go, Python, Node, Rust and PHP. Or whatever can talk HTTP (or gRPC).
In this Code Café I will introduce you to Dapr.io. I will show you what Dapr can do for you (application) and how you can Dapr-izen an application. I'll show you how an asynchronously collaborative system of microservices - implemented in different technologies - can be easily connected to Dapr, first to Redis as a Pub/Sub mechanism and then also to Apache Kafka without modifications. Then we do - with the interested parties - also a hands-on in which you will apply Dapr yourself . In a short time you get a good feel for how you can use Dapr for different aspects of your applications. And if nothing else, Dapr is a very easy way to get your code with Kafka, S3, Redis, Azure EventGrid, HashiCorp Consul, Twillio, Pulsar, RabbitMQ, HashiCorp Vault, AWS Secret Manager, Azure KeyVault, Cron, SMTP, Twitter, AWS SQS & SNS, GCP Pub/Sub and dozens of other technology components talk.
How and Why you can and should Participate in Open Source Projects (AMIS, Sof...Lucas Jellema
For a long time I have been reluctant to actively contribute to an open source project. I thought it would be rather complicated and demanding – and that I didn't have the knowledge or skills for it or at the very least that they (the project team) weren't waiting for me.
In December 2021, I decided to have a serious input into the Dapr.io project – and now finally to determine how it works and whether it is really that complicated. In this session I want to tell you about my experiences. How Fork, Clone, Branch, Push (and PR) is the rhythm of contributing to an open source project and how you do that (these are all Git actions against GitHub repositories). How to learn how such a project functions and how to connect to it; which tools are needed, which communication channels are used. I tell how the standards of the project – largely automatically enforced – help me to become a better software engineer, with an eye for readability and testability of the code.
How the review process is quite exciting once you have offered your contribution. And how the final "merge to master" of my contribution and then the actual release (Dapr 1.6 contains my first contribution) are nice milestones.
I hope to motivate participants in this session to also take the step yourself and contribute to an open source project in the form of issues or samples, documentation or code. It's valuable to the community and the specific project and I think it's definitely a valuable experience for the "contributer". I looked up to it and now that I've done it gives me confidence – and it tastes like more (I could still use some help with the work on Dapr.io, by the way).
Microservices, Apache Kafka, Node, Dapr and more - Part Two (Fontys Hogeschoo...Lucas Jellema
Apache Kafka is one of the best known enterprise grade message brokers – created at LinkedIn, donated to the Apache software foundation and used in an ever growing number of organizations to provide a backbone for asynchronous communication. This session introduces Apache Kafka – history, concepts, community and tooling. In a hands on lab, participants will create topics, publish and consume messages and get a general feel for Kafka. Simple microservices are developed in NodeJS – publishing to and consuming from Apache Kafka.
Dapr.io has support for Apache Kafka. Using Kafka through Dapr is very straightforward as is explained and demonstrated and applied in a second handson lab – with applications in various programming languages. Participants will even be able to exchange events across their laptops – through a cloud based Kafka broker.
Use of Apache Kafka in several architecture patterns is discussed – such as data integration, microservices, CQRS, Event Sourcing – along with a number of real world use cases from several well known organizations. The Kafka Connector framework is introduced – a set of adapters that allow us to easily connect Kafka to sources and sinks – where respectively change events are captured from and messages are published to.
Bonus Lab: Apache Kafka is ran on Kubernetes as is Dapr.io. Multiple mutually interacting microservices are deployed on the same local Kubernetes cluster.
Microservices, Node, Dapr and more - Part One (Fontys Hogeschool, Spring 2022)Lucas Jellema
This session does a quick recap of microservices: why do we want them, what problems do they solve and what are the principles around designing and implementing them? The Dapr.io runtime framework for distributed applications is introduced. Dapr provides a sidecar (almost like a personal assistant to a manager) to an application or microservice, a companion process that handles common tasks such as storing and retrieving state, consuming and publishing messages and events, invoking external services and other microservices as well as handling incoming requests. Participants will do a handson lab with Dapr.io and learn how to quickly implement interactions with various technologies, including Redis and MySQL.
Node(JS) is introduced – a server side JavaScript-based programming language that can be used well for implementing microservices. Some of the main characteristics of NodeJS are discussed (functional programming, asynchronous flows, NPM package manager) as well as common use cases (handle incoming HTTP requests, invoke REST APIs). In the second lab, Node and Dapr are used together to implement microservices that interact with databases and message brokers and each other – in a decoupled fashion.
6Reinventing Oracle Systems in a Cloudy World (RMOUG Trainingdays, February 2...Lucas Jellema
The cloud is changing many things. Even the decision to not (yet) adopt cloud is one to make explicitly. Now is a time for any organization to reconsider the IT landscape. For each system we should make a conscious ruling on its roadmap. The 6R model suggests six ways to move a system forward.
This session uses the 6R model and applies it specifically to Oracle technology based systems: what are the options and considerations for Oracle Database, Oracle Fusion Middleware, custom applications, and other red components? What future should we consider and how do we choose? The paths chosen by several Oracle-heavy users is presented to illustrate these options and the decision making process. Oracle Cloud Infrastructure and Autonomous Database play a role, as do Azure IaaS and Azure Managed Database as well as on premises systems. Latency, recovery, scalability, licenses, automation, lock-in, skills, and resources all make their appearance.
Help me move away from Oracle! (RMOUG Training Days 2022, February 2022)Lucas Jellema
Organizations with decades of investment in Oracle technology sometimes (and increasingly) express a wish to move away from Oracle. In this session, we will first explore where the desire to move away from Oracle might come from. Then we describe what the term Oracle represents -- more than 2.000 products on all layers in the technology stack and in different business areas. Finally, we map out what the 'moving away from' consists of: defining where you 'move to' and subsequently actually going there.
It will become clear why you should give considerable thought about dropping Oracle, or any other vendors' technology, when you're not pleased with your current IT situation. You need to focus on the actual problems and objectives and define the suitable roadmap to fit your real needs. It turns out that the quest is usually for modernization and flexibility - and Oracle can very well be a part of that future.
DevOps is a term used in many places and unfortunately also to mean many different things. This presentation (largely in Dutch) paints the DevOps picture. While it may not give a clear cut definition (there does not seem to be one) it certainly makes clear what DevOps is about, what objectives and origins are and which factors enable and drive DevOps.
Conclusion Code Cafe - Microcks for Mocking and Testing Async APIs (January 2...Lucas Jellema
Microcks is a tool for API Mocking and Testing. In this presentation an overview of the support in Microcks for asynchronous APIs - the event publishing and consuming behavior of services and applications
Cloud native applications offer scalability, flexibility, and optimal use of compute resources. Serverless functions interacting through events, leveraging cloud capabilities for persistent storage and automated operations take organization to the next level in IT. This session demonstrates polyglot Functions interacting with native cloud services for events and persistence (Object Storage and NoSQL Database) and leveraging the Key and Secrets Vault, Monitoring and Notifications services for operational control. A lightweight API Gateway is used to expose APIs to external consumers. Infrastructure as Code is the guiding principle in deploying both cloud resources and application components, through OCI CLI and Terraform. This session leverages many cloud native (enabling) services in Oracle Cloud Infrastructure. The session will introduce concepts, then spend most of the time on live demonstrations. All sources are shared with the audience, to allow participants to create the same application in their own cloud tenancy. What is so great about Cloud Native Applications? How do you create one? I will explain the first and demonstrate the second. On Oracle Cloud Infrastructure, using services that anyone can use for free, I will live create a cloud native application that streams, persists, notifies, scales, monitors Benefits: - get to know many different OCI services - understand the meaning, purpose and benefits of cloud native development - learn how to take your own first steps in OCI - for free!
Check out the webinar slides to learn more about how XfilesPro transforms Salesforce document management by leveraging its world-class applications. For more details, please connect with sales@xfilespro.com
If you want to watch the on-demand webinar, please click here: https://www.xfilespro.com/webinars/salesforce-document-management-2-0-smarter-faster-better/
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Enhancing Research Orchestration Capabilities at ORNL.pdfGlobus
Cross-facility research orchestration comes with ever-changing constraints regarding the availability and suitability of various compute and data resources. In short, a flexible data and processing fabric is needed to enable the dynamic redirection of data and compute tasks throughout the lifecycle of an experiment. In this talk, we illustrate how we easily leveraged Globus services to instrument the ACE research testbed at the Oak Ridge Leadership Computing Facility with flexible data and task orchestration capabilities.
Gamify Your Mind; The Secret Sauce to Delivering Success, Continuously Improv...Shahin Sheidaei
Games are powerful teaching tools, fostering hands-on engagement and fun. But they require careful consideration to succeed. Join me to explore factors in running and selecting games, ensuring they serve as effective teaching tools. Learn to maintain focus on learning objectives while playing, and how to measure the ROI of gaming in education. Discover strategies for pitching gaming to leadership. This session offers insights, tips, and examples for coaches, team leads, and enterprise leaders seeking to teach from simple to complex concepts.
Quarkus Hidden and Forbidden ExtensionsMax Andersen
Quarkus has a vast extension ecosystem and is known for its subsonic and subatomic feature set. Some of these features are not as well known, and some extensions are less talked about, but that does not make them less interesting - quite the opposite.
Come join this talk to see some tips and tricks for using Quarkus and some of the lesser known features, extensions and development techniques.
Launch Your Streaming Platforms in MinutesRoshan Dwivedi
The claim of launching a streaming platform in minutes might be a bit of an exaggeration, but there are services that can significantly streamline the process. Here's a breakdown:
Pros of Speedy Streaming Platform Launch Services:
No coding required: These services often use drag-and-drop interfaces or pre-built templates, eliminating the need for programming knowledge.
Faster setup: Compared to building from scratch, these platforms can get you up and running much quicker.
All-in-one solutions: Many services offer features like content management systems (CMS), video players, and monetization tools, reducing the need for multiple integrations.
Things to Consider:
Limited customization: These platforms may offer less flexibility in design and functionality compared to custom-built solutions.
Scalability: As your audience grows, you might need to upgrade to a more robust platform or encounter limitations with the "quick launch" option.
Features: Carefully evaluate which features are included and if they meet your specific needs (e.g., live streaming, subscription options).
Examples of Services for Launching Streaming Platforms:
Muvi [muvi com]
Uscreen [usencreen tv]
Alternatives to Consider:
Existing Streaming platforms: Platforms like YouTube or Twitch might be suitable for basic streaming needs, though monetization options might be limited.
Custom Development: While more time-consuming, custom development offers the most control and flexibility for your platform.
Overall, launching a streaming platform in minutes might not be entirely realistic, but these services can significantly speed up the process compared to building from scratch. Carefully consider your needs and budget when choosing the best option for you.
Understanding Globus Data Transfers with NetSageGlobus
NetSage is an open privacy-aware network measurement, analysis, and visualization service designed to help end-users visualize and reason about large data transfers. NetSage traditionally has used a combination of passive measurements, including SNMP and flow data, as well as active measurements, mainly perfSONAR, to provide longitudinal network performance data visualization. It has been deployed by dozens of networks world wide, and is supported domestically by the Engagement and Performance Operations Center (EPOC), NSF #2328479. We have recently expanded the NetSage data sources to include logs for Globus data transfers, following the same privacy-preserving approach as for Flow data. Using the logs for the Texas Advanced Computing Center (TACC) as an example, this talk will walk through several different example use cases that NetSage can answer, including: Who is using Globus to share data with my institution, and what kind of performance are they able to achieve? How many transfers has Globus supported for us? Which sites are we sharing the most data with, and how is that changing over time? How is my site using Globus to move data internally, and what kind of performance do we see for those transfers? What percentage of data transfers at my institution used Globus, and how did the overall data transfer performance compare to the Globus users?
Unleash Unlimited Potential with One-Time Purchase
BoxLang is more than just a language; it's a community. By choosing a Visionary License, you're not just investing in your success, you're actively contributing to the ongoing development and support of BoxLang.
In 2015, I used to write extensions for Joomla, WordPress, phpBB3, etc and I ...Juraj Vysvader
In 2015, I used to write extensions for Joomla, WordPress, phpBB3, etc and I didn't get rich from it but it did have 63K downloads (powered possible tens of thousands of websites).
How Recreation Management Software Can Streamline Your Operations.pptxwottaspaceseo
Recreation management software streamlines operations by automating key tasks such as scheduling, registration, and payment processing, reducing manual workload and errors. It provides centralized management of facilities, classes, and events, ensuring efficient resource allocation and facility usage. The software offers user-friendly online portals for easy access to bookings and program information, enhancing customer experience. Real-time reporting and data analytics deliver insights into attendance and preferences, aiding in strategic decision-making. Additionally, effective communication tools keep participants and staff informed with timely updates. Overall, recreation management software enhances efficiency, improves service delivery, and boosts customer satisfaction.
Exploring Innovations in Data Repository Solutions - Insights from the U.S. G...Globus
The U.S. Geological Survey (USGS) has made substantial investments in meeting evolving scientific, technical, and policy driven demands on storing, managing, and delivering data. As these demands continue to grow in complexity and scale, the USGS must continue to explore innovative solutions to improve its management, curation, sharing, delivering, and preservation approaches for large-scale research data. Supporting these needs, the USGS has partnered with the University of Chicago-Globus to research and develop advanced repository components and workflows leveraging its current investment in Globus. The primary outcome of this partnership includes the development of a prototype enterprise repository, driven by USGS Data Release requirements, through exploration and implementation of the entire suite of the Globus platform offerings, including Globus Flow, Globus Auth, Globus Transfer, and Globus Search. This presentation will provide insights into this research partnership, introduce the unique requirements and challenges being addressed and provide relevant project progress.
Providing Globus Services to Users of JASMIN for Environmental Data AnalysisGlobus
JASMIN is the UK’s high-performance data analysis platform for environmental science, operated by STFC on behalf of the UK Natural Environment Research Council (NERC). In addition to its role in hosting the CEDA Archive (NERC’s long-term repository for climate, atmospheric science & Earth observation data in the UK), JASMIN provides a collaborative platform to a community of around 2,000 scientists in the UK and beyond, providing nearly 400 environmental science projects with working space, compute resources and tools to facilitate their work. High-performance data transfer into and out of JASMIN has always been a key feature, with many scientists bringing model outputs from supercomputers elsewhere in the UK, to analyse against observational or other model data in the CEDA Archive. A growing number of JASMIN users are now realising the benefits of using the Globus service to provide reliable and efficient data movement and other tasks in this and other contexts. Further use cases involve long-distance (intercontinental) transfers to and from JASMIN, and collecting results from a mobile atmospheric radar system, pushing data to JASMIN via a lightweight Globus deployment. We provide details of how Globus fits into our current infrastructure, our experience of the recent migration to GCSv5.4, and of our interest in developing use of the wider ecosystem of Globus services for the benefit of our user community.
Innovating Inference - Remote Triggering of Large Language Models on HPC Clus...Globus
Large Language Models (LLMs) are currently the center of attention in the tech world, particularly for their potential to advance research. In this presentation, we'll explore a straightforward and effective method for quickly initiating inference runs on supercomputers using the vLLM tool with Globus Compute, specifically on the Polaris system at ALCF. We'll begin by briefly discussing the popularity and applications of LLMs in various fields. Following this, we will introduce the vLLM tool, and explain how it integrates with Globus Compute to efficiently manage LLM operations on Polaris. Attendees will learn the practical aspects of setting up and remotely triggering LLMs from local machines, focusing on ease of use and efficiency. This talk is ideal for researchers and practitioners looking to leverage the power of LLMs in their work, offering a clear guide to harnessing supercomputing resources for quick and effective LLM inference.
Software Engineering, Software Consulting, Tech Lead, Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Transaction, Spring MVC, OpenShift Cloud Platform, Kafka, REST, SOAP, LLD & HLD.
Climate Science Flows: Enabling Petabyte-Scale Climate Analysis with the Eart...Globus
The Earth System Grid Federation (ESGF) is a global network of data servers that archives and distributes the planet’s largest collection of Earth system model output for thousands of climate and environmental scientists worldwide. Many of these petabyte-scale data archives are located in proximity to large high-performance computing (HPC) or cloud computing resources, but the primary workflow for data users consists of transferring data, and applying computations on a different system. As a part of the ESGF 2.0 US project (funded by the United States Department of Energy Office of Science), we developed pre-defined data workflows, which can be run on-demand, capable of applying many data reduction and data analysis to the large ESGF data archives, transferring only the resultant analysis (ex. visualizations, smaller data files). In this talk, we will showcase a few of these workflows, highlighting how Globus Flows can be used for petabyte-scale climate analysis.
Climate Science Flows: Enabling Petabyte-Scale Climate Analysis with the Eart...
Intro to Docker Containers and the Oracle Platform – Database, WebLogic &Cloud (ODevC Yatra 2018, July, Pune and Mumbai)
1. and the Oracle
Platform -
Database,
WebLogic
& the Cloud
Intro to Docker Containers
Intro to Docker Containers for Oracle professionals1
Lucas Jellema, CTO of AMIS
ODevC Yatra, Pune, Friday 13th July 2018
2. Lucas Jellema
Architect / Developer
1994 started in IT at Oracle
2002 joined AMIS
Currently CTO & Solution Architect
Intro to Docker Containers for Oracle professionals 2
3. Presenting
• Oracle OpenWorld
• JavaOne
• Oracle Code
• Devoxx
• Java and Oracle User Group meetups
• JavaOne Rockstar (JavaOne 2015)
• ODevC Yatra 2018
Intro to Docker Containers for Oracle professionals 3
4. Writing
• Blogs at http://technology.amis.nl
• 1500 articles – from UI to Middle Tier, Database and Infrastructure
• Articles at Medium, DZone and Oracle Technology Network
• Books for McGraw Hill (Oracle Press)
• Oracle ACE Director & Developer Champion
Intro to Docker Containers for Oracle professionals 4
6. Setup for Oracle OpenWorld Demo
What I needed
• Local installation of a Kafka Cluster
• At least one Broker node and the Zookeeper
Kafka
Broker
Zookeeper
Demo Application
Intro to Docker Containers for Oracle professionals 6
7. Setup for Oracle OpenWorld Demo
What I received from Guido
• Simple text file – 140 lines
Intro to Docker Containers for Oracle professionals 7
Name of Docker
image to run
Hostname on internal network
between Docker containers
Environment variable
to pass to container
Dependency on other
container (to start first)
Container port to
expose externally
8. Setup for Oracle OpenWorld Demo
What I created in a few minutes
Intro to Docker Containers for Oracle professionals 8
Kafka
Broker
Zookeeper
Kafka
Rest ProxyKafka
Schema
Registry
Kafka
Connect
Kafka
Connect UI
Kafka
Schema
Registry UI
Kafka
Manager
9092
2181
9000
8084
80018083
8081
8002
9. Some Quick Conclusions
• Docker provides a great way to
• Build environments (application & platform)
(from simple, text based build files & public images)
• Share & Ship these environments
(either through build files or through ready-to-run images)
• Run environments making efficient use of physical resources
(that can be complex and have complex interdependencies)
• And Guido is a very nice guy
• And also:
• [Docker] Containers are pivotal in cloud native environments,
microservices architecture, DevOps and CD
• Any IT professional should know her or his way around containers
Intro to Docker Containers for Oracle professionals 9
10. Overview of today’s session
• Docker Container – what and why?
• Build, ship, run & operate
• Use in development, training, testing, delivery and production & operations
• Running custom containers on Oracle Container Cloud
• Microservices and the application platform of tomorrow
• Introducing Kubernetes and the upcoming Oracle Kubernetes Engine Cloud
• Building Containers with Oracle platform from Oracle GitHub repo
• Oracle Container Registry with prebaked images for Oracle platform
• Going forward…
Intro to Docker Containers for Oracle professionals 10
11. Linux essentials
• Applications share resources
Intro to Docker Containers for Oracle professionals 11
Disk Storage
Memory
CPUs
Application A
Application B
Application C
• Network interface
• IP address
• Ports
• Users & groups
• Environment
Variables
• Packages
• Services
12. Linux essentials: Control Groups and Namespaces
• Compartmentalize Resources
into isolated
units
Intro to Docker Containers for Oracle professionals 12
Disk
Storage
Memory
CPUs
• Network interface
• IP address
• Ports
• Users & groups
• Environment
Variables
• Packages
• Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
13. Linux essentials: Control Groups and Namespaces
• Expose units through
mapped network
ports
Intro to Docker Containers for Oracle professionals 13
Disk
Storage
Memory
CPUs
• Network interface
• IP address
• Ports
• Users & groups
• Environment
Variables
• Packages
• Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
14. Linux essentials: Each unit runs its own processes
• Units run their own
processes:
• OS (Linux)
• Platform
• Application
Intro to Docker Containers for Oracle professionals 14
Disk
Storage
Memory
CPUs
• Network interface
• IP address
• Ports
• Users & groups
• Environment
Variables
• Packages
• Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Application A Application B
Application C
15. This stuff is complex
• Core Linux features were hard to use
Intro to Docker Containers for Oracle professionals 15
Disk
Storage
Memory
CPUs• Network interface
• IP
address
• Ports
• Users & groups
• Environment Variables
• Packages
• Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Application A Application B
Application C
16. Docker has democratized Linux Containers
• Container Image – a serialized file from which we can instantiate a container
• Container Build script and workflow – to automate the creation of a container
(image) using straightforward vocabulary
• Engine – runtime platform for instantiating, running and managing containers,
volumes and networks (REST API and CLI)
• Docker Registry – Repository for Container Images
• And now also Docker Store
Intro to Docker Containers for Oracle professionals 16
Disk
Storage
Memory
CPUs• Network interface
• IP
add
res
s
• Por
ts
• Users & groups
• Environment Variables
• Packages
• Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Application A Application B
Application
C
17. Running Containers using Docker
• Create Container(s)
from Image plus:
• Port mapping
• Volume
• Environment
Variable
• (inter container)
Network
• Startup script
Intro to Docker Containers for Oracle professionals 17
Disk
Storage
Memory
CPUs
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Network interface
IP address
Ports
Users & groups
Environment Variables
Packages
Services
Application A Application B
Application C
Docker Hub
Docker Engine
Container
images
18. Running Containers using Docker on Windows
• Docker is a Linux mechanism
• In order to run on a Windows server,
we use a Linux VM
• VirtualBox
• Hyper-V
• …
• Docker Toolbox
• It is possible to run the Docker Engine inside a Docker Container
• Docker Container inside Docker Container [inside VM]
Disk
Storage
Memory
CPUs• Network interface
• IP
add
res
s
• Por
ts
• Users & groups
• Environment Variables
• Packages
• Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Network
interface
IP address
Ports
Users & groups
Environment
Variables
Packages
Services
Application A Application B
Application
C
Intro to Docker Containers for Oracle professionals 18
20. Running Containers using Docker
20
Application A
Docker Hub
Docker Engine
docker run
--name ApplicationA
amis/NodeAppRunnerImage:latest
/bin/bash
amis/NodeAppRunnerImage:1.4
ApplicationA
Intro to Docker Containers for Oracle professionals
21. Running Containers using Docker
Intro to Docker Containers for Oracle professionals 21
Application A
Docker Hub
Docker Engine
docker run
--name ApplicationA
-p 8010:8080 -p 8011:1521
--network=myBridgeNW
-e APP_HOME=/home/apps/applicationA
-e PARAM1=value1
amis/NodeAppRunnerImage:latest
/bin/bash
amis/NodeAppRunnerImage:1.4
8010
8011
8080
1521
ApplicationA
APP_HOME=
/home/apps/applicationA
PARAM1=
value1
22. Running Containers using Docker
Intro to Docker Containers for Oracle professionals 22
Disk
Storage
/host_files
/data
Application A
Docker Hub
Docker Engine
docker run
--name ApplicationA
-p 8010:8080 -p 8011:1521
--network=myBridgeNW
-v /hostworkdir
-v /tmp/files:/host_files
--volumes-from dataContainer
-e APP_HOME=/home/apps/applicationA
-e PARAM1=value1
amis/NodeAppRunnerImage:latest
/bin/bash
amis/NodeAppRunnerImage:1.4
8010
8011
8080
1521
dataContainer
ApplicationA
APP_HOME=
/home/apps/applicationA
PARAM1=
value1
23. Containers are ephemeral (*
Intro to Docker Containers for Oracle professionals 23
(* Candidate for IT word of the year 2018
24. Container state that needs to survive should be on an
externally mapped volume
Intro to Docker Containers for Oracle professionals 24
Host Disk Volume
-v /data:/u01/app/data
/u01/app/data
--mount source=/u01,target= /u01/app/data
/data
25. Implicit Docker Container Image Interface:
environment variables, ports, volumes
Intro to Docker Containers for Oracle professionals 25
Docker Hub
link mysql
Parameters:
WORDPRESS_DB_PASSWORD,
WORDPRESS_DB_USER, …
Volume
..:/var/lib
/mysql
Parameters:
MYSQL_DATABASE,
MYSQL_ROOT_PASSWORD
26. Running and Managing Containers
• Start | Pause | Stop | Delete | Export | Import containers
• Save | Load Images
• List containers | images | networks | …
• Inspect container
• Run multiple instances of an image
• Execute into running container
• Attach to (standard input | output | error stream of)
running container
• Get Container Logs
• Create Network
• Connect container to network
• Experimental feature: Snapshot (CRIU)
Intro to Docker Containers for Oracle professionals 26
29. Building container images
• Manual:
• Run a container
• Perform all installation and configuration
• Commit the container and tag as new Container Image
• Push Image to Registry to reuse
Docker Hub
Dockersig-trial:1.0
Intro to Docker Containers for Oracle professionals 29
30. Building container images
• Scripted
(automated & repeatable/evolvable):
• Create Docker Build file
• Select base image
• Gather files required during build
• Consider multistage build
• To purge intermediate artifacts
• Build and Commit Image
• Commit build file to Git
• Push Image to Registry
Docker Hub
Dockersig-trial:1.0
Intro to Docker Containers for Oracle professionals 30
31. Docker Build Files on GitHub
Intro to Docker Containers for Oracle professionals 31
Summer
2018
33. Do not share container image – share container build file!
34. Ship (Container Images)
• Package, Distribute, Share, Publish and Consume container images
• The frozen state of a container (committed after building and further manipulating)
• With everything needed to run the micro service: application and underlying platform &
OS, ready to run on any Docker Engine anywhere
• With an implicit interface (environment variables, ports, volume)
34Intro to Docker Containers for Oracle professionals
36. Shipping Container Images
• Containers can be Exported and Imported
• Via TAR-files
• Images can be Saved and Loaded
• Via TAR-files
Intro to Docker Containers for Oracle professionals 36
38. Container Use Cases for Oracle professionals
• R&D (aka Play) – try out technology
• Quickly, easily, cleanly
• Complex, multi-node configurations
• Leverage huge number of resources available out in the open
• Prepare and Share (running) environments for
• Playing, Training, Testing, Beta-testing,
• Deploy and Run application on generic cloud infrastructure
• Especially ephemeral (stateless) and dynamically scalable
• Streamlined CD across Development, Test and Production
• Prepare for Cloud (consolidate, lift & shift workloads)
• Analysis & What If Scenarios
• Clone an environment, spin up, investigate, tear down & quit
• Automated Testing
• Against rich dataset with minimum set up and tear down
• Microservices – implement, deploy and run
Intro to Docker Containers for Oracle professionals 38
39. Manage Test Data Set for (automated) tests
• Build a Container Image with:
• Oracle Database
• Application Database Objects from DDL
• Test Data Set (with all cases and relevant details)
• Commit and Tag
with (Sprint) Release
Intro to Docker Containers for Oracle professionals 39
Oracle Database
DDL
DML scripts or Export for
Test Data
AppTest:R17.49.1
40. Run (Automated) Test
• Run container image for designated release
• with –rm flag
• start database
• Execute test
• No set up, no tear down
• Stop container
Intro to Docker Containers for Oracle professionals 40
AppTest:R17.49.1
Oracle
Database
Test Data
Application
docker container run
-d -p 1521:1521
-rm AppTest:R17.49.1
41. Run (Automated) Test
• Run container image for designated release
• with –rm flag
• start database
• Execute test
• No set up, no tear down
• Stop container
• Next test – or even in parallel
Intro to Docker Containers for Oracle professionals 41
AppTest:R17.49.1
42. Manage Test Data Set for (automated) tests
After a new (Sprint) Release
• Run Container for Previous Release
• Apply DDL to Upgrade Application
• Manage Data Set – test cases to cater for new features
• Commit and Tag with new (Sprint) Release label
Intro to Docker Containers for Oracle professionals 42
DDL
R17.51.1
DML scripts or Export for Test
Data Updates R17.51.1
AppTest:R17.51.1
AppTest:R17.49.1
47. Once upon a time –
a container based microservice
Intro to Docker Containers for Oracle professionals 49
µ
http requests
48. Where is the container running?
• Any Docker Host – on premises or cloud based VM - or a Container Cloud Service
• For example: Oracle Container Cloud Service
Intro to Docker Containers for Oracle professionals 50
µ
49. How did the container start running in the runtime?
• Through a CI/CD Pipeline
• Build process
• Take a Basic runtime image – e.g. Linux plus Some Language VM
• Add application code
• Add runtime agents and tooling
• Add platform/runtime configuration
• Then Build the Image
• Test the Image
• Tag and Push Image to Registry
• Deploy the image plus applicable configuration to a specific destination
container runtime environment
• The CI/CD Process is managed manually or triggered by development
event
Intro to Docker Containers for Oracle professionals 51
µ
50. Wercker: Build, Test, Push and Deploy Pipelines for
Containers
Intro to Docker Containers for Oracle professionals 52
µ
µ
µ
52. Containers
• As vehicle for:
• Encapsulate
• Build
• Share & Ship
• Automated Tests
• Deploy
• Run
• Scale
• Relocate
• Standardize
Intro to Docker Containers for Oracle professionals 54
53. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 55
Compute
Node
54. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 56
Compute
Node
Compute Node
Compute
Node
55. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 57
Compute
Node
Compute Node
Compute
Node
56. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 58
Compute
Node
Compute Node
Compute
Node
57. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 59
Compute
Node
Compute Node
Compute
Node
58. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 60
Compute
Node
Compute Node
Compute
Node
Cloud
Storage
SAN
59. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 61
Compute
Node
Compute Node
Compute
Node
Cloud
Storage
SAN
Configuration
Map
Configuration
Map
60. Looking for a runtime platform for
Intro to Docker Containers for Oracle professionals 62
Compute
Node
Compute Node
Compute
Node
70. Example of Docker File
• Build a Docker Container
with Java 8 Runtime
Intro to Docker Containers for Oracle professionals 72
71. Build Docker Container for
Oracle Database 12.2.0.1 Enterprise Edition
• Download database
installation binaries
before building the
container
Intro to Docker Containers for Oracle professionals 73
72. Why run Oracle Database on Docker?
• Because we can…
• Automated testing
• Clone environments
• Similar to PDB cloning
• Quick provisioning of new environments
• R&D
• Production workloads?
• Automated Ops/DBA => Autonomous Database (?)
Intro to Docker Containers for Oracle professionals 74
73. Running WebLogic Server in Docker Containers -
when and why?
Intro to Docker Containers for Oracle professionals 75
78. Run Oracle Database from official Container Image
• docker run -d -it –-name ORA12201_1
–P container-registry.oracle.com/database/enterprise:12.2.0.1
Intro to Docker Containers for Oracle professionals 80
79. Run Oracle Database – from a container image
Intro to Docker Containers for Oracle professionals 81
81. Oracle Container Registry for Your Images
• After build and before run – container images need to be stored
• Secure (because runtime artefacts)
• Accessible (& low latency) to deployment engine and container runtime
• Scalable and Smart (no duplicate images and image layers)
Intro to Docker Containers for Oracle professionals 83
83. Going forward – what should be your moves?
• Start playing.
Intro to Docker Containers for Oracle professionals 85
https://www.katacoda.com/courses/docker
84. Going forward – what should be your moves?
• Learn about Docker
• Brush up on your Linux skills
• Install Docker and run some images
• Experiment with Port, Link, Volume, Environment Variables
• Create your own build file, build a container and commit as image
• Push your own image to a Docker Registry
• Using a trial on Oracle Cloud – run a container
from your image on the Container Cloud or on the OCI Kubernetes Engine Cloud
• Run containers based on the official Oracle Docker build files on GitHub
• Run containers based on the official Oracle Docker images on Oracle Container Registry
• Learn about Kubernetes (KataKoda is an excellent environment)
• Experiment with Kubernetes locally (on minikube)
• And on Oracle Kubernetes Engine Cloud
Intro to Docker Containers for Oracle professionals 86
85. Summary
• Docker is a great technology to
• Run
• Share, Ship & Deliver
• Build
encapsulated environments with run time platform
and application
• Containers are likely the core run time unit to manage:
deploy, configure, scale, monitor, interconnect, secure
• Kubernetes is the de facto distributed container
management platform for cloud and on premises
• Oracle does Docker and Kubernetes in anger
Session structure Introduce Containers - objectives, benefits, implementation Demo of Container build, package, ship and run
Discuss Container Management systems - run time Container platforms, such as Oracle Container Cloud
Demo of deploying and running a Container first locally
then on the Oracle Container Cloud
Discussion of CD, DevOps and microservices - and how the Orace platform components fit in
(including a discussion of multitenant architecture in DB and WLS)
Introduction of Oracle Docker Images
Demonstration of building containers based on Oracle Docker Images
Run multiple containers based on various Oracle Docker Images and have them interact with each other
Deploy and Run (Docker Containers)
Distributed infrastructure (scalable and available)
Hide infrastructure from DevOps teams
Auto-healing
Elastic Scale
Wire up the micros – connect dynamically (service discovery)
Load Balance
Provide Persistent storage
Rolling Upgrade
Configuration & Secret Management
Secure
Deploy and Run (Docker Containers)
Distributed infrastructure (scalable and available)
Hide infrastructure from DevOps teams
Auto-healing
Elastic Scale
Wire up the micros – connect dynamically (service discovery)
Load Balance
Provide Persistent storage
Rolling Upgrade
Configuration & Secret Management
Secure
Deploy and Run (Docker Containers)
Distributed infrastructure (scalable and available)
Hide infrastructure from DevOps teams
Auto-healing
Elastic Scale
Wire up the micros – connect dynamically (service discovery)
Load Balance
Provide Persistent storage
Rolling Upgrade
Configuration & Secret Management
Secure