Arkena's video-on-demand platform is used as backend by major european channels (TF1 / beIN SPORTS / Elisa) to propose a non-linear experience to their customers.
Previously hosted on Heroku, the number of our users is increasing constantly. In order to optimize resources we decided to move on a bare metal infrastructure powered by Kubernetes.
We'll share thoughts, feedbacks and technical details about this successful transition.
Sched Link:
This document discusses testing Kubernetes and OpenShift at scale. It describes installing large clusters of 1000+ nodes, using scalability test tools like the Kubernetes performance test repo and OpenShift SVT repo to load clusters and generate traffic. Sample results show loading clusters with thousands of pods and projects, and peaks in master node resource usage when loading and deleting hundreds of pods simultaneously.
Leveraging the Power of containerd Events - Evan HazlettDocker, Inc.
containerd provides the low-level functionality that enables the Docker Engine to run containers. containerd events provide a simple, yet powerful mechanism to integrate with virtually any other system with minimal effort. This talk will cover what containerd events are and how to use them for integration with systems ranging from monitoring and logging to container networking using CNI (Container Network Interface) plugins.
Using Docker EE to Scale Operational Intelligence at SplunkDocker, Inc.
With more than 14,000 customers in 110+ countries, Splunk is the market leader in analyzing machine data to deliver operational intelligence for security, IT and the business. Our rapid growth as a company meant that our Infrastructure Engineering Team, responsible for all the common tooling, build and test systems and frameworks utilized by the Splunk engineers, was bogged down with a sprawl of virtual machines and physical servers that were becoming incredibly difficult to manage. And as our customer’s demand for data has grown, testing at the scale of petabytes/day has become our new normal. We needed a reliable and scalable “Test Lab” for functional and performance testing.
With Docker Enterprise Edition, our engineers are able to create small test stacks on their laptop just as easily as creating multi-petabyte stacks in our Test Lab. Support for Windows, Role Based Access Control and having support for both the orchestration platform and the container engine were key in deciding to go with Docker over other solutions.
In this talk, we will cover the architecture, tooling, and frameworks we built to manage our workloads, which have grown to run on over 600 bare-metal servers, with tens of thousands of containers being created every day. We will share the lessons learned from running at scale. Lastly, we will demonstrate how we use Splunk to monitor and manage Docker Enterprise Edition.
1. Docker EE will include an unmodified Kubernetes distribution to provide orchestration capabilities alongside Docker Swarm.
2. When running mixed workloads across orchestrators, resource contention is a risk and it is recommended to separate workloads by orchestrator on each node for now.
3. Docker EE aims to address the shortcomings of running mixed workloads to better support this in the future.
LlinuxKit security, Security Scanning and NotaryDocker, Inc.
The document discusses security updates from the Moby Summit. It talks about LinuxKit and its focus on minimal and securely configured kernels. It mentions containerd for running services and dhcpcd for networking. There is also a discussion around secure build chains and signing of images. The document outlines the LinuxKit Security SIG which meets biweekly to discuss projects like Landlock LSM and WireGuard. It proposes standardizing scanning report formats and attaching them to images. Finally, it provides an update on Notary submitting to CNCF and its technical work towards TUF 1.0 specification.
Fabric8 - Being devOps doesn't suck anymoreHenryk Konsek
Fabric8 is a tool that aims to reduce the gap between development and operations by allowing developers to deploy and manage applications from development through production. It uses profiles and containers to deploy applications in a unified way across environments. Fabric8 supports various containers like Tomcat, Docker, and OpenShift. It includes integrated monitoring via Hawt.io and a shell to help with scripting and diagnostics. The goal is to make developers more self-sufficient by enabling them to take on more operations tasks in their own sandboxes.
Arkena's video-on-demand platform is used as backend by major european channels (TF1 / beIN SPORTS / Elisa) to propose a non-linear experience to their customers.
Previously hosted on Heroku, the number of our users is increasing constantly. In order to optimize resources we decided to move on a bare metal infrastructure powered by Kubernetes.
We'll share thoughts, feedbacks and technical details about this successful transition.
Sched Link:
This document discusses testing Kubernetes and OpenShift at scale. It describes installing large clusters of 1000+ nodes, using scalability test tools like the Kubernetes performance test repo and OpenShift SVT repo to load clusters and generate traffic. Sample results show loading clusters with thousands of pods and projects, and peaks in master node resource usage when loading and deleting hundreds of pods simultaneously.
Leveraging the Power of containerd Events - Evan HazlettDocker, Inc.
containerd provides the low-level functionality that enables the Docker Engine to run containers. containerd events provide a simple, yet powerful mechanism to integrate with virtually any other system with minimal effort. This talk will cover what containerd events are and how to use them for integration with systems ranging from monitoring and logging to container networking using CNI (Container Network Interface) plugins.
Using Docker EE to Scale Operational Intelligence at SplunkDocker, Inc.
With more than 14,000 customers in 110+ countries, Splunk is the market leader in analyzing machine data to deliver operational intelligence for security, IT and the business. Our rapid growth as a company meant that our Infrastructure Engineering Team, responsible for all the common tooling, build and test systems and frameworks utilized by the Splunk engineers, was bogged down with a sprawl of virtual machines and physical servers that were becoming incredibly difficult to manage. And as our customer’s demand for data has grown, testing at the scale of petabytes/day has become our new normal. We needed a reliable and scalable “Test Lab” for functional and performance testing.
With Docker Enterprise Edition, our engineers are able to create small test stacks on their laptop just as easily as creating multi-petabyte stacks in our Test Lab. Support for Windows, Role Based Access Control and having support for both the orchestration platform and the container engine were key in deciding to go with Docker over other solutions.
In this talk, we will cover the architecture, tooling, and frameworks we built to manage our workloads, which have grown to run on over 600 bare-metal servers, with tens of thousands of containers being created every day. We will share the lessons learned from running at scale. Lastly, we will demonstrate how we use Splunk to monitor and manage Docker Enterprise Edition.
1. Docker EE will include an unmodified Kubernetes distribution to provide orchestration capabilities alongside Docker Swarm.
2. When running mixed workloads across orchestrators, resource contention is a risk and it is recommended to separate workloads by orchestrator on each node for now.
3. Docker EE aims to address the shortcomings of running mixed workloads to better support this in the future.
LlinuxKit security, Security Scanning and NotaryDocker, Inc.
The document discusses security updates from the Moby Summit. It talks about LinuxKit and its focus on minimal and securely configured kernels. It mentions containerd for running services and dhcpcd for networking. There is also a discussion around secure build chains and signing of images. The document outlines the LinuxKit Security SIG which meets biweekly to discuss projects like Landlock LSM and WireGuard. It proposes standardizing scanning report formats and attaching them to images. Finally, it provides an update on Notary submitting to CNCF and its technical work towards TUF 1.0 specification.
Fabric8 - Being devOps doesn't suck anymoreHenryk Konsek
Fabric8 is a tool that aims to reduce the gap between development and operations by allowing developers to deploy and manage applications from development through production. It uses profiles and containers to deploy applications in a unified way across environments. Fabric8 supports various containers like Tomcat, Docker, and OpenShift. It includes integrated monitoring via Hawt.io and a shell to help with scripting and diagnostics. The goal is to make developers more self-sufficient by enabling them to take on more operations tasks in their own sandboxes.
Open shift enterprise 3.1 paas on kubernetesSamuel Terburg
The document provides an overview of Red Hat OpenShift Enterprise, which is Red Hat's container application platform. It discusses key OpenShift concepts like Docker containers, Kubernetes cluster management, projects for sandboxing environments, image builds, deployments, services and routes. It also summarizes added value of OpenShift over plain Docker and Kubernetes through features like templates, web console, multi-language support, automated builds and deployments, and enterprise capabilities.
The document discusses OpenShift security context constraints (SCCs) and how to configure them to allow running a WordPress container. It begins with an overview of SCCs and their purpose in OpenShift for controlling permissions for pods. It then describes issues running the WordPress container under the default "restricted" SCC due to permission errors. The document explores editing the "restricted" SCC and removing capabilities and user restrictions to address the errors. Alternatively, it notes the "anyuid" SCC can be used which is more permissive and standard for allowing the WordPress container to run successfully.
This presentation includes information on Kubernetes Architecture, Container Orchestration, Internal Routing, External Routing, Configuration Management, Credentials Management, Persistent Volumes, Rolling Out Updates, Autoscaling, Package Management, and a Hello World example using Helm.
Java in 2019 was predicted to be business as usual by many. We have seen new Java releases coming out as planned, AdoptOpenJDK became the main trustful source of binaries and Oracle fighting for the trademark again by preventing the use of javax as namespace. Everything looks like it would be a silent year for Java. But one thing seems obvious. Java's popularity is not gaining any more traction. New language features keep it up to date but people are getting more selective when it comes to implementation choices. Especially in the age of containers and cloud infrastructures. This talk walks you through the why and how of Java in containers. We will talk about image size and development and deployment processes.
Presentation delivered at LinuxCon China 2017.
Operating systems need to move faster without sacrificing stability. New hardware, new software features, and bugfixes are making it into distribution components every day. To maintain stability, packagers and distribution developers are looking toward lessons learned in the DevOps movement to implement Continuous Integration/Continuous Delivery (CI/CD) workflows that provide quicker test feedback to developers.
This talk highlights some of the coming trends in Fedora such as: streamlined base package sets, userspace applications delivered as containers, continuous validation of individual distro components and the distro as a whole, and collaboration with the CentOS Project.
Architectural caching patterns for kubernetesRafał Leszko
The document discusses various architectural caching patterns for Kubernetes, including embedded, embedded distributed, client-server, cloud, sidecar, reverse proxy, and reverse proxy sidecar caching. It provides examples of implementing each pattern using Hazelcast and discusses the pros and cons of each approach.
Kubernetes deep dive - - Huawei 2015-10Vishnu Kannan
Kubernetes is an open-source container orchestration system that automates deployment, scaling, and management of containerized applications. It was originally designed by Google based on years of experience running containers internally. Kubernetes runs containerized applications across multiple machines, dynamically allocating resources and balancing load. It supports both public and private cloud environments as well as bare metal servers. The system aims to simplify container operations while providing portability and scalability.
Platform Orchestration with Kubernetes and DockerJulian Strobl
Big companies like Google containerize theirs environments for easier maintaining, scaling, and reliability. This talk gives an introduction how to build such an environment and maintain applications written in distinct programming languages. The container orchestration is done with Kubernetes by Google and Docker containers. For mass deployment CoreOS is used.
KubeCon EU 2016: Multi-Tenant KubernetesKubeAcademy
Today Kubernetes is mostly employed in single tenant deployment, either private cloud, or as a COE on top of IaaS. By leveraging virtualized container like Hyper, Kubernetes will be the core of multi-tenant Container-as-a-Service. This talk will present Hypernetes, a secure Kubernetes distro focusing on the public container hosting service.
Sched Link: http://sched.co/6BYD
KubeCon EU 2016: Kubernetes and the Potential for Higher Level InterfacesKubeAcademy
Kubernetes provides rock-solid APIs for building and running your distributed systems. Pods, Services and ReplicationControllers provide trustworthy and scalable abstractions that make solving real-world infrastructure problems simpler. But that doesn’t mean interacting with those low-level primitives will be the only option for developers and operators.
Sched Link: http://sched.co/67dA
JavaLand 2022, März, Brühl, Mario-Leander Reimer (@LeanderReimer, Principal Software Architect bei QAware).
== Dokument bitte herunterladen, falls unscharf! Please download slides if blurred! ==
This session focuses on modern and efficient Inter Process Communication (IPC) for microservices. We start with a REST API, built using JAX-RS and Quarkus to briefly discuss the pros and cons of this approach. Then, we will extend the API with an efficient Protobuf payload representation in order to finally transform the API into a fully fledged high-performance gRPC interface definition. But that's not all! To put some extra icing on the cake, this talk will demonstrate how to consume the gRPC service from a JavaScript web client and also how to completely generate a matching REST API from an enhanced gRPC interface definition to ensure full interoperability in a microservice architecture.
Developer joy for distributed teams with CodeReady Workspaces | DevNation Tec...Red Hat Developers
Natale Vinto discusses CodeReady Workspaces, a developer environment tool that runs on OpenShift. It provides containerized developer workspaces that enable coding directly within Kubernetes clusters. Key features include the Eclipse Che IDE, compatibility with VSCode extensions, and use of "devfiles" to define and standardize reproducible developer environments. CodeReady Workspaces aims to improve productivity for remote and distributed teams by reducing setup times and enabling self-service access to development environments.
Using Docker Containers to Improve Reproducibility in Software and Web Engine...Vincenzo Ferme
The ability to replicate and reproduce scientific results has become an increasingly important topic for many academic disciplines. In computer science and, more specifically, software and Web engineering, contributions of scientific work rely on developed algorithms, tools and prototypes, quantitative evaluations, and other computational analyses. Published code and data come with many undocumented assumptions, dependencies, and configurations that are internal knowledge and make reproducibility hard to achieve. This tutorial presents how Docker containers can overcome these issues and aid the reproducibility of research artefacts in software engineering and discusses their applications in the field.
Cite us: http://link.springer.com/chapter/10.1007/978-3-319-38791-8_58
Docker Meetup - Melbourne 2015 - Kubernetes Deep DiveKen Thompson
This document provides an overview of Kubernetes networking and storage capabilities. It begins with an agenda that includes a deep dive on Kubernetes networking and persistent volumes, as well as live demos of persistent storage and another topic. The document then discusses Kubernetes networking at the host level using pods that share IP, IPC, and disk, as well as inter-host networking solutions like OpenShift SDN. It also covers Kubernetes persistent volume claims that allow administrators to provision storage and developers to request storage that is independent of the underlying devices. The document concludes with demos of storage and another topic.
CI/CD Pipeline mit Gitlab CI und Kubernetesinovex GmbH
Docker Meetup Mannheim, 20.06.2017
Speaker: Johannes M. Scheuermann
Mehr Tech-Vorträge auf: https://www.inovex.de/de/content-pool/vortraege/
oder in unserem Blog: https://www.inovex.de/blog/
Thanks to tools like vagrant, puppet/chef, and Platform as a Service services like Heroku, developers are extremely used to being able to spin up a development environment that is the same every time. What if we could go a step further and make sure our development environment is not only using the same software, but 100% configured and set up like production. Docker will let us do that, and so much more. We’ll look at what Docker is, why you should look into using it, and all of the features that developers can take advantage of.
Kubernetes is a container cluster manager that aims to provide a platform for automating deployment, scaling, and operations of application containers across clusters of machines. It uses pods as the basic building block, which are groups of application containers that share storage and networking resources. Kubernetes includes control planes for replication, scheduling, and services to expose applications. It supports deployment of multi-tier applications through replication controllers, services, labels, and pod templates.
Enabling Production Grade Containerized Applications through Policy Based Inf...Docker, Inc.
This session covers the solution addressing the needs of enabling product-grade containerized applications. You will learn how operations teams running containerized applications in a shared infrastructure can define and enforce policies to provide security, monitoring, and performance for network, storage, and computing. You will learn about Contiv and Mantl, open source projects that create a framework for cloud native application development and infrastructure with application intent and operational policies. Contiv integrates Cisco infrastructure (UCS, Nexus, and ACI) with Docker Datacenter to help enterprises adopt containers at a larger scale.
The document analyzes the contents pages of three music magazines: Kerrang!, Q, and NME. It examines the layout, use of text, images, representation of audiences, and how each addresses its target readership. The Kerrang! and Q contents pages use bold colors, graphics and large images to represent younger audiences. NME contains more text in a simpler layout with fewer images, potentially appealing more to older readers. In summary, the document evaluates how the design elements of different music magazine contents pages are tailored to their intended age demographics.
Este documento describe el e-learning y la herramienta de software AMBAR. El e-learning permite la educación en línea a través de varias tecnologías como Internet, intranets y CD-ROM. AMBAR es una herramienta basada en la web que permite a profesores y estudiantes crear y participar en procesos de enseñanza-aprendizaje basados en objetos de aprendizaje reutilizables.
Open shift enterprise 3.1 paas on kubernetesSamuel Terburg
The document provides an overview of Red Hat OpenShift Enterprise, which is Red Hat's container application platform. It discusses key OpenShift concepts like Docker containers, Kubernetes cluster management, projects for sandboxing environments, image builds, deployments, services and routes. It also summarizes added value of OpenShift over plain Docker and Kubernetes through features like templates, web console, multi-language support, automated builds and deployments, and enterprise capabilities.
The document discusses OpenShift security context constraints (SCCs) and how to configure them to allow running a WordPress container. It begins with an overview of SCCs and their purpose in OpenShift for controlling permissions for pods. It then describes issues running the WordPress container under the default "restricted" SCC due to permission errors. The document explores editing the "restricted" SCC and removing capabilities and user restrictions to address the errors. Alternatively, it notes the "anyuid" SCC can be used which is more permissive and standard for allowing the WordPress container to run successfully.
This presentation includes information on Kubernetes Architecture, Container Orchestration, Internal Routing, External Routing, Configuration Management, Credentials Management, Persistent Volumes, Rolling Out Updates, Autoscaling, Package Management, and a Hello World example using Helm.
Java in 2019 was predicted to be business as usual by many. We have seen new Java releases coming out as planned, AdoptOpenJDK became the main trustful source of binaries and Oracle fighting for the trademark again by preventing the use of javax as namespace. Everything looks like it would be a silent year for Java. But one thing seems obvious. Java's popularity is not gaining any more traction. New language features keep it up to date but people are getting more selective when it comes to implementation choices. Especially in the age of containers and cloud infrastructures. This talk walks you through the why and how of Java in containers. We will talk about image size and development and deployment processes.
Presentation delivered at LinuxCon China 2017.
Operating systems need to move faster without sacrificing stability. New hardware, new software features, and bugfixes are making it into distribution components every day. To maintain stability, packagers and distribution developers are looking toward lessons learned in the DevOps movement to implement Continuous Integration/Continuous Delivery (CI/CD) workflows that provide quicker test feedback to developers.
This talk highlights some of the coming trends in Fedora such as: streamlined base package sets, userspace applications delivered as containers, continuous validation of individual distro components and the distro as a whole, and collaboration with the CentOS Project.
Architectural caching patterns for kubernetesRafał Leszko
The document discusses various architectural caching patterns for Kubernetes, including embedded, embedded distributed, client-server, cloud, sidecar, reverse proxy, and reverse proxy sidecar caching. It provides examples of implementing each pattern using Hazelcast and discusses the pros and cons of each approach.
Kubernetes deep dive - - Huawei 2015-10Vishnu Kannan
Kubernetes is an open-source container orchestration system that automates deployment, scaling, and management of containerized applications. It was originally designed by Google based on years of experience running containers internally. Kubernetes runs containerized applications across multiple machines, dynamically allocating resources and balancing load. It supports both public and private cloud environments as well as bare metal servers. The system aims to simplify container operations while providing portability and scalability.
Platform Orchestration with Kubernetes and DockerJulian Strobl
Big companies like Google containerize theirs environments for easier maintaining, scaling, and reliability. This talk gives an introduction how to build such an environment and maintain applications written in distinct programming languages. The container orchestration is done with Kubernetes by Google and Docker containers. For mass deployment CoreOS is used.
KubeCon EU 2016: Multi-Tenant KubernetesKubeAcademy
Today Kubernetes is mostly employed in single tenant deployment, either private cloud, or as a COE on top of IaaS. By leveraging virtualized container like Hyper, Kubernetes will be the core of multi-tenant Container-as-a-Service. This talk will present Hypernetes, a secure Kubernetes distro focusing on the public container hosting service.
Sched Link: http://sched.co/6BYD
KubeCon EU 2016: Kubernetes and the Potential for Higher Level InterfacesKubeAcademy
Kubernetes provides rock-solid APIs for building and running your distributed systems. Pods, Services and ReplicationControllers provide trustworthy and scalable abstractions that make solving real-world infrastructure problems simpler. But that doesn’t mean interacting with those low-level primitives will be the only option for developers and operators.
Sched Link: http://sched.co/67dA
JavaLand 2022, März, Brühl, Mario-Leander Reimer (@LeanderReimer, Principal Software Architect bei QAware).
== Dokument bitte herunterladen, falls unscharf! Please download slides if blurred! ==
This session focuses on modern and efficient Inter Process Communication (IPC) for microservices. We start with a REST API, built using JAX-RS and Quarkus to briefly discuss the pros and cons of this approach. Then, we will extend the API with an efficient Protobuf payload representation in order to finally transform the API into a fully fledged high-performance gRPC interface definition. But that's not all! To put some extra icing on the cake, this talk will demonstrate how to consume the gRPC service from a JavaScript web client and also how to completely generate a matching REST API from an enhanced gRPC interface definition to ensure full interoperability in a microservice architecture.
Developer joy for distributed teams with CodeReady Workspaces | DevNation Tec...Red Hat Developers
Natale Vinto discusses CodeReady Workspaces, a developer environment tool that runs on OpenShift. It provides containerized developer workspaces that enable coding directly within Kubernetes clusters. Key features include the Eclipse Che IDE, compatibility with VSCode extensions, and use of "devfiles" to define and standardize reproducible developer environments. CodeReady Workspaces aims to improve productivity for remote and distributed teams by reducing setup times and enabling self-service access to development environments.
Using Docker Containers to Improve Reproducibility in Software and Web Engine...Vincenzo Ferme
The ability to replicate and reproduce scientific results has become an increasingly important topic for many academic disciplines. In computer science and, more specifically, software and Web engineering, contributions of scientific work rely on developed algorithms, tools and prototypes, quantitative evaluations, and other computational analyses. Published code and data come with many undocumented assumptions, dependencies, and configurations that are internal knowledge and make reproducibility hard to achieve. This tutorial presents how Docker containers can overcome these issues and aid the reproducibility of research artefacts in software engineering and discusses their applications in the field.
Cite us: http://link.springer.com/chapter/10.1007/978-3-319-38791-8_58
Docker Meetup - Melbourne 2015 - Kubernetes Deep DiveKen Thompson
This document provides an overview of Kubernetes networking and storage capabilities. It begins with an agenda that includes a deep dive on Kubernetes networking and persistent volumes, as well as live demos of persistent storage and another topic. The document then discusses Kubernetes networking at the host level using pods that share IP, IPC, and disk, as well as inter-host networking solutions like OpenShift SDN. It also covers Kubernetes persistent volume claims that allow administrators to provision storage and developers to request storage that is independent of the underlying devices. The document concludes with demos of storage and another topic.
CI/CD Pipeline mit Gitlab CI und Kubernetesinovex GmbH
Docker Meetup Mannheim, 20.06.2017
Speaker: Johannes M. Scheuermann
Mehr Tech-Vorträge auf: https://www.inovex.de/de/content-pool/vortraege/
oder in unserem Blog: https://www.inovex.de/blog/
Thanks to tools like vagrant, puppet/chef, and Platform as a Service services like Heroku, developers are extremely used to being able to spin up a development environment that is the same every time. What if we could go a step further and make sure our development environment is not only using the same software, but 100% configured and set up like production. Docker will let us do that, and so much more. We’ll look at what Docker is, why you should look into using it, and all of the features that developers can take advantage of.
Kubernetes is a container cluster manager that aims to provide a platform for automating deployment, scaling, and operations of application containers across clusters of machines. It uses pods as the basic building block, which are groups of application containers that share storage and networking resources. Kubernetes includes control planes for replication, scheduling, and services to expose applications. It supports deployment of multi-tier applications through replication controllers, services, labels, and pod templates.
Enabling Production Grade Containerized Applications through Policy Based Inf...Docker, Inc.
This session covers the solution addressing the needs of enabling product-grade containerized applications. You will learn how operations teams running containerized applications in a shared infrastructure can define and enforce policies to provide security, monitoring, and performance for network, storage, and computing. You will learn about Contiv and Mantl, open source projects that create a framework for cloud native application development and infrastructure with application intent and operational policies. Contiv integrates Cisco infrastructure (UCS, Nexus, and ACI) with Docker Datacenter to help enterprises adopt containers at a larger scale.
The document analyzes the contents pages of three music magazines: Kerrang!, Q, and NME. It examines the layout, use of text, images, representation of audiences, and how each addresses its target readership. The Kerrang! and Q contents pages use bold colors, graphics and large images to represent younger audiences. NME contains more text in a simpler layout with fewer images, potentially appealing more to older readers. In summary, the document evaluates how the design elements of different music magazine contents pages are tailored to their intended age demographics.
Este documento describe el e-learning y la herramienta de software AMBAR. El e-learning permite la educación en línea a través de varias tecnologías como Internet, intranets y CD-ROM. AMBAR es una herramienta basada en la web que permite a profesores y estudiantes crear y participar en procesos de enseñanza-aprendizaje basados en objetos de aprendizaje reutilizables.
This document lists several interests including celebrities, gadgets, jokes, horoscopes, shopping and fashion. It asks the reader what they are interested in and provides those topics as examples of possible interests.
The document describes Jessica Lonsdale's work rebranding and redesigning the website for The Exemplar Group, a Dallas-based executive advisory and consulting firm. She was tasked with writing all new copy for the website. To showcase Exemplar's problem-solving abilities, she created short but powerful descriptions of their services. She categorized Exemplar's consulting services based on stages of a project's life cycle and created separate pages for each with brief descriptions and examples of their consulting tools.
Este documento alienta al lector a confiar en Jesús y dejar sus preocupaciones y problemas en sus manos. Aconseja repetir frecuentemente la oración "Jesús, yo confío en ti" y abandonarse completamente a la voluntad de Dios sin tratar de imponer sus propias ideas. Promete que si el lector sigue confiando a pesar de las dificultades, verá grandes milagros.
This document provides specifications and information for servicing an engine. It begins with general service rules and model identification information. The bulk of the document then lists detailed specifications for various engine and motorcycle components in sections covering items like the lubrication, fuel, cooling and electrical systems as well as the engine, frame, and wheels. It concludes with standard torque values and diagrams showing cable and harness routing.
Trend Micro uses Hadoop for processing large volumes of web data to quickly identify and block malicious URLs. They have expanded their Hadoop cluster significantly over time to support growing data and job volumes. They developed Hadooppet to automate deployment and management of their large, customized Hadoop distribution across hundreds of nodes. Profiling tools like Nagios, Ganglia and Splunk help monitor and troubleshoot cluster performance issues.
Exploring Phantom Traffic Jams in Your Data Flows ScyllaDB
This document discusses phantom traffic jams that can occur in data flows even without true hardware bottlenecks. It explains how producer-consumer systems with jitter at the producer, consumer, or dispatcher can lead to self-limiting behavior and effective dispatch rates below capacity. The document advocates for good metrics and analyzing bottlenecks with an interposer that can measure queue lengths rather than assuming hardware limitations. It provides a simulator experiment to demonstrate these effects and conclusions about checking bottlenecks throughout systems instead of only at hardware.
Navigating Complex Database Performance Hurdles
Tackling your own database performance challenges is serious business. For a change of pace, let’s have some fun learning from other teams’ performance predicaments.
Join us for an interactive session where we dissect 4 specific database performance challenges faced by teams considering or using ScyllaDB. For each dilemma:
- The presenters will describe the context and technical requirements
- Together, we’ll talk about potential solutions and cover the pros and cons of each
- Finally, we’ll disclose what approach the team took, and how it worked out
Throughout the event, we’ll have opportunities to win ScyllaDB swag and prizes! Come prepared to engage in lively discussions and gain valuable insight into database performance strategies.
Implementing High Availability Caching with MemcachedGear6
Typical Memcached deployments do not comprehensively address web site requirements for high availability. Depending on your web architecture, a single failure can disable your web caches. This presentation offers real world solutions to solving <a>high availability</a> challenges common to large, dynamic websites with Memcached, specifically:
* Options and benefits for deploying high availability services within Memcached
* How companies are approaching high availability
* Considerations on building and deploying high availability
o Recommendations for a typical Memcached environment
o Open source tools available
o High level costs for deployment
Have you heard that all in-memory databases are equally fast but unreliable, inconsistent and expensive? This session highlights in-memory technology that busts all those myths.
Redis, the fastest database on the planet, is not a simply in-memory key-value data-store; but rather a rich in-memory data-structure engine that serves the world’s most popular apps. Redis Labs’ unique clustering technology enables Redis to be highly reliable, keeping every data byte intact despite hundreds of cloud instance failures and dozens of complete data-center outages. It delivers full CP system characteristics at high performance. And with the latest Redis on Flash technology, Redis Labs achieves close to in-memory performance at 70% lower operational costs. Learn about the best uses of in-memory computing to accelerate everyday applications such as high volume transactions, real time analytics, IoT data ingestion and more.
How Development Teams Cut Costs with ScyllaDB.pdfScyllaDB
Now that teams are increasingly being pressed to cut costs, the database can be a low-hanging fruit for sizable cost reduction – especially if you’re managing terabytes to petabytes of data with millions of read/write operations per second.
Join Tzach Livyatan, VP of Product at ScyllaDB, as he shares four ways that teams commonly cut database costs by rethinking their database strategy. We’ll cover topics including:
- Cutting admin costs by reducing node sprawl and reducing the need for tuning
- ScyllaDB as a better, compatible Amazon DynamoDB
- Options to increase price performance through new cloud instances
- Ways to safely add more workloads to your cluster without compromising the performance of your latency-sensitive workloads
Pilot Hadoop Towards 2500 Nodes and Cluster RedundancyStuart Pook
Hadoop has become a critical part of Criteo's operations. What started out as a proof of concept has turned into two in-house bare-metal clusters of over 2200 nodes. Hadoop contains the data required for billing and, perhaps even more importantly, the data used to create the machine learning models, computed every 6 hours by Hadoop, that participate in real time bidding for online advertising.
Two clusters do not necessarily mean a redundant system, so Criteo must plan for any of the disasters that can destroy a cluster.
This talk describes how Criteo built its second cluster in a new datacenter and how to do it better next time. How a small team is able to run and expand these clusters is explained. More importantly the talk describes how a redundant data and compute solution at this scale must function, what Criteo has already done to create this solution and what remains undone.
Critical Attributes for a High-Performance, Low-Latency DatabaseScyllaDB
This document discusses the attributes of a high-performance, low-latency database like ScyllaDB. It begins with introductions and an overview of ScyllaDB. It then summarizes how hardware has evolved over 20 years with more cores, memory, and faster disks. ScyllaDB was redesigned from first principles to take advantage of modern hardware, using an asynchronous, shared-nothing architecture with one shard per core. This allows it to achieve significantly higher performance than Cassandra. The document shows benchmark results demonstrating ScyllaDB's lower latencies and ability to scale to higher throughput. It also discusses how ScyllaDB uses workload prioritization to manage different types of workloads.
RedisConf17 - Redis Cluster at flickr and tripodRedis Labs
Flickr uses Redis extensively across many services for caching, queueing, and more. Some key uses of Redis include caching for the camera roll and activity feed, which each have multiple masters and slaves distributed across data centers. Notifications are also queued in Redis. In total, Flickr operates over 100 physical hosts running 500+ Redis instances to support over 600,000 operations per second across services like the camera roll, activity feed, notifications, and more.
Kubernetes Failure Stories - KubeCon Europe BarcelonaHenning Jacobs
Talk given on 2019-05-21 at KubeCon Barcelona: https://kccnceu19.sched.com/event/MPcM/kubernetes-failure-stories-and-how-to-crash-your-clusters-henning-jacobs-zalando-se
Bootstrapping a Kubernetes cluster is easy, rolling it out to nearly 200 engineering teams and operating it at scale is a challenge. In this talk, we are presenting our approach to Kubernetes provisioning on AWS, operations and developer experience for our growing Zalando developer base. We will walk you through our horror stories of operating 100+ clusters and share the insights we gained from incidents, failures, user reports and general observations. Our failure stories will be sourced from recent and past incidents, so the talk will be up-to-date with our latest experiences.
Most of our learnings apply to other Kubernetes infrastructures (EKS, GKE, ..) as well. This talk strives to reduce the audience's unknown unknowns about running Kubernetes in production.
What's new with enterprise Redis - Leena Joshi, Redis LabsRedis Labs
Redis Labs manages over 160k+ HA databases, 10k clustered databases, without data loss in spite of one node failure a day and one data center outage per month. Using Enterprise
Redis(RLEC), Redis Labs delivers seamless zero downtime scaling, true high availability with persistence, cross-rack/zone/
datacenter replication and instant automatic failover. Learn how. Join this session for a deep dive into how enterprise Redis makes for no-hassle Redis deployments and the roadmap for new Redis capabilities. Discover new cost savings with Redis on Flash for cost-effective high performance operations and analytics
Case Study with Answers.com on Scaling with Memcached and MySQLanswers
The document discusses Answers.com's experience scaling their website and databases. They used several techniques including adding solid state drives to MySQL servers, implementing a Memcached caching tier with Gear6 for redundancy and failover, using Lighttpd to monitor database servers, and hardware load balancers. Gear6 provided an alternative to developing a custom Memcached wrapper that mirrored data and improved reliability over the standard Memcached implementation. The presentation provides examples of how these techniques helped Answers.com scale their infrastructure to support over 50 million monthly users.
The document discusses Oracle RAC and Docker, including why Oracle would be used in containers, considerations for using Oracle RAC in containers, how containers and virtual networks work, preparing storage, images, and networking for Oracle RAC containers, and how to configure Oracle Grid Infrastructure in Docker containers. Key points include reducing resources and time through containers, challenges of shared-nothing architecture and privileged access in containers, and steps to configure storage, virtual networking, and Oracle software in images before deploying Oracle RAC containers.
The document discusses an agenda for a meetup about Redis Labs and Kubernetes operators. Key points:
- An introduction to Redis Enterprise architecture and Redis Labs products.
- A discussion of "double orchestration" using Kubernetes and PKS to manage Redis clusters for performance and resource management.
- An overview of Redis Labs' Kubernetes solution using StatefulSets, services, and a custom controller.
- An introduction to operators, how they provide lifecycle management and simplify deployments compared to static YAML files or Helm.
- Details on Redis Labs' operator development process and challenges in building idempotent APIs and handling state changes and validation in the reconciliation loop.
This document discusses microservices architecture compared to a monolithic architecture. A microservices architecture breaks an application into smaller, independent services that each perform discrete functions. This allows for more rapid development and improved scalability. However, a microservices architecture is also more complex to deploy and manage. The document provides an example of how a VoIP application could use a microservices approach by breaking components like billing, fraud detection, and call analytics into separate services. It also discusses using Docker containers and services to deploy and scale the microservices architecture.
Under The Hood Of A Shard-Per-Core Database ArchitectureScyllaDB
This document summarizes the key design decisions behind ScyllaDB's shard-per-core database architecture. It discusses how ScyllaDB addresses the challenges of scaling databases across hundreds of CPU cores by utilizing an asynchronous task model with one thread and one data shard per CPU core. This allows for linear scalability. It also overhauls the I/O scheduling to prioritize workloads and maximize throughput from SSDs under mixed read/write workloads. Benchmark results show ScyllaDB's architecture can handle petabyte-scale databases with high performance and low latency even on commodity hardware.
Learning Rust the Hard Way for a Production Kafka + ScyllaDB PipelineScyllaDB
🎥 Sign up for upcoming webinars or browse through our library of on-demand recordings here: https://www.scylladb.com/resources/webinars/
About this webinar:
Numberly operates business-critical data pipelines and applications where failure and latency means "lost money" in the best-case scenario. Most of their data pipelines and applications are deployed on Kubernetes and rely on Kafka and ScyllaDB, with Kafka acting as the message bus and ScyllaDB as the source of data for enrichment. The availability and latency of both systems are thus very important for data pipelines. While most of Numberly’s applications are developed using Python, they found a need to move high-performance applications to Rust in order to benefit from a lower-level programming language.
Learn the lessons from Numberly’s experience, including:
- Rationale in selecting a lower-level language
- Developing using a lower-level Rust code base
- Observability and analyzing latency impacts with Rust
- Tuning everything from Apache Avro to driver client settings
- How to build a mission-critical system combining Apache Kafka and ScyllaDB
- Half a year Rust in production feedback
Real-Time Machine Learning with Redis, Apache Spark, Tensor Flow, and more wi...Databricks
Predictive intelligence from machine learning has the potential to change everything in our day to day experiences, from education to entertainment, from travel to healthcare, from business to leisure and everything in between. Modern ML frameworks are batch by nature and cannot pivot on the fly to changing user data or situations. Many simple ML applications such as those that enhance the user experience, can benefit from real-time robust predictive models that adapt on the fly.
Join this session to learn how common practices in machine learning such as running a trained model in production can be substantially accelerated and radically simplified by using Redis modules that natively store and execute common models generated by Spark ML and Tensorflow algorithms. We will also discuss the implementation of simple, real-time feed-forward neural networks with Neural Redis and scenarios that can benefit from such efficient, accelerated artificial intelligence.
Real-life implementations of these new techniques at a large consumer credit company for fraud analytics, at an online e-commerce provider for user recommendations and at a large media company for targeting content will also be discussed.
Getting Started with MariaDB with DockerMariaDB plc
This document discusses deploying MariaDB databases with Docker from development to production. It recommends using Docker containers to encapsulate dependencies and isolate processes for easy deployment on-premise, in the cloud, or in hybrid environments. It highlights challenges like orchestration complexity and outlines requirements for data durability, self-discovery, self-healing, and application discovery of database clusters. It demonstrates building a Python/Flask app in Docker, deploying it to a Swarm cluster, and scaling the web tier behind HAProxy. It also shows deploying a 3-node Galera MariaDB cluster and 2-node MaxScale proxy for high availability.
Big Data in Container; Hadoop Spark in Docker and MesosHeiko Loewe
3 examples for Big Data analytics containerized:
1. The installation with Docker and Weave for small and medium,
2. Hadoop on Mesos w/ Appache Myriad
3. Spark on Mesos
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Letter and Document Automation for Bonterra Impact Management (fka Social Sol...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on automated letter generation for Bonterra Impact Management using Google Workspace or Microsoft 365.
Interested in deploying letter generation automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
9. 9
Docker Downsides
0) Immutable
1) requires Linux
2) only runs one process
3) tooling is awful (especially on not Linux)
4) troubleshooting is hard
5) need to rethink how you develop/deploy
15. 15
Better Solution from Hashicorp
- use puppet/chef/salt/ansible
- still can use shell scripts
- output a variety of formats
- AMI
- GCE
- Droplets
- Virtual box
- Openstack
- Docker
22. 22
Better Solution from Hashicorp (3)
So do you work for them or something??
- Job scheduler
- Docker
- Java
- Binary
- QEMU
- Cron scheduler
- Orchestration
- Service discovery
25. 25
Better Solution from Hashicorp (4)
- single binary
- Raft consensus / HA
- Gossip protocol
- Rest API
- DNS Interface
- Key/Value store
- Locking
34. 34
New Problem
- You can deploy too fast……
Which one of the 10 deploys caused the issue and
which one fixed it?
Two deployment windows a day 11:00 and 3:00
means everyone is on hand and able to jump in if
there's an issue
35. 35
Conclusion (1)
You are going to end up with an infrastructure that consists of
Docker-ized and other types of services
Accordingly, pick the tool set that is the most flexible! (my bias
is clear here...)
36. 36
Conclusion (2)
Docker is worth it for some things:
- application servers
- stateless applications
- development environment
Don't bother:
- data stores
- slow moving monolithic apps