The document appears to be a presentation about deploying a graph database on AWS. It discusses using Scala and Akka for building components, Neo4j as the graph database, Packer for creating machine images, and deploying the components on AWS using a t1.micro instance type to build Amazon machine images.
In this talk we review 3 different orchestration frameworks to run Docker application stacks on AWS:
- Docker for AWS
- Docker Cloud
- Docker Datacenter
Wieldy remote apis with Kekkonen - ClojureD 2016Metosin Oy
Today, most of the apps we build involve the browser and thus need an remoting layer between the ui and the backend. New Technology like Falcor, Relay and Om Next are leading the innovation in this space. Is REST dead? CQRS to the rescue? How should I build my next api?
Our 2 cents is Kekkonen, a fresh new multi-paradigm api library for Clojure.
Sparklyr: Recap, Updates, and Use Cases with Javier LuraschiDatabricks
This document provides a recap and updates on sparklyr, an R package that provides an interface to Apache Spark. It discusses the history and design principles of R and S, an overview of sparklyr functionality and architecture, and examples use cases for SQL queries, machine learning, graph analytics, and distributed execution using sparklyr. The document outlines recent updates in sparklyr versions 0.4, 0.5, and the new features planned for version 0.6, including distributed R workers.
Riot Games serves an international base of players that creates terabytes of data daily. With their data centers quickly reaching capacity, Riot Games migrated their entire warehouse to AWS in order to scale operations more effectively. Learn how Riot Games used DynamoDB and Elastic MapReduce to import millions of rows of customer metrics. Hear about the technical specifics involved, and the lessons learned along the way of migrating large amounts of data to AWS.
This document discusses zero-configuration provisioning of Kubernetes clusters on unmanaged infrastructure. It describes using immutable bootstrapping to provision operating systems and install Docker and Kubernetes (using Kubeadm) across nodes without requiring centralized orchestration or SSH access. The document also discusses potential future directions for the Kubernetes community regarding node admission controls and dynamic Kubelet configuration to further reduce external configuration requirements during cluster provisioning.
Schema tools-and-trics-and-quick-intro-to-clojure-spec-22.6.2016Metosin Oy
The document discusses Schema and Clojure.spec, two libraries for data validation and specification in Clojure. It provides an overview of Schema's capabilities for runtime validation, documentation, and data transformations. It also summarizes Clojure.spec's focus on specification through generative testing. The document compares the two approaches and argues that Schema is currently better for runtime validation in web applications, while Clojure.spec will be important for system-level specification.
In this talk we review 3 different orchestration frameworks to run Docker application stacks on AWS:
- Docker for AWS
- Docker Cloud
- Docker Datacenter
Wieldy remote apis with Kekkonen - ClojureD 2016Metosin Oy
Today, most of the apps we build involve the browser and thus need an remoting layer between the ui and the backend. New Technology like Falcor, Relay and Om Next are leading the innovation in this space. Is REST dead? CQRS to the rescue? How should I build my next api?
Our 2 cents is Kekkonen, a fresh new multi-paradigm api library for Clojure.
Sparklyr: Recap, Updates, and Use Cases with Javier LuraschiDatabricks
This document provides a recap and updates on sparklyr, an R package that provides an interface to Apache Spark. It discusses the history and design principles of R and S, an overview of sparklyr functionality and architecture, and examples use cases for SQL queries, machine learning, graph analytics, and distributed execution using sparklyr. The document outlines recent updates in sparklyr versions 0.4, 0.5, and the new features planned for version 0.6, including distributed R workers.
Riot Games serves an international base of players that creates terabytes of data daily. With their data centers quickly reaching capacity, Riot Games migrated their entire warehouse to AWS in order to scale operations more effectively. Learn how Riot Games used DynamoDB and Elastic MapReduce to import millions of rows of customer metrics. Hear about the technical specifics involved, and the lessons learned along the way of migrating large amounts of data to AWS.
This document discusses zero-configuration provisioning of Kubernetes clusters on unmanaged infrastructure. It describes using immutable bootstrapping to provision operating systems and install Docker and Kubernetes (using Kubeadm) across nodes without requiring centralized orchestration or SSH access. The document also discusses potential future directions for the Kubernetes community regarding node admission controls and dynamic Kubelet configuration to further reduce external configuration requirements during cluster provisioning.
Schema tools-and-trics-and-quick-intro-to-clojure-spec-22.6.2016Metosin Oy
The document discusses Schema and Clojure.spec, two libraries for data validation and specification in Clojure. It provides an overview of Schema's capabilities for runtime validation, documentation, and data transformations. It also summarizes Clojure.spec's focus on specification through generative testing. The document compares the two approaches and argues that Schema is currently better for runtime validation in web applications, while Clojure.spec will be important for system-level specification.
4.Building a Data Product using apache Zeppelin - Apache Kylin Meetup @ShanghaiLuke Han
The document discusses building a data product using Apache Zeppelin (incubating) that sends email notifications about new open source projects on GitHub. It outlines the steps to download data from the GitHub archive, explore and filter the data to focus on interesting companies, join additional data from the GitHub API, generate an HTML template to visualize the results, and schedule sending the email notifications.
ClojuTRE2015: Kekkonen - making your Clojure web APIs more awesomeMetosin Oy
Thanks to REST and Swagger, we can build beautiful apis to feed both our browser front-ends and external applications. But, wrapping your Clojure code into resources mostly for your ClojureScript front-end doesn't feel right? Just use RPC? Meet in the middle?
Kekkonen is a small library for managing your (web) apis as commands and queries. No magic, data-driven, un-restful and non-rpc. It's goals are to be small, explicit, extendable and to help enforce your business rules both on the server side and on the ClojureScript frontend. Besides Swagger, it provides run-time context-aware apidocs for Clojure(Script).
Apache Zeppelin is interactive data analytics environment for large scale data processing systems. It deeply integrates to Apache spark and many other frameworks, provides beautiful interactive web-based interface, data visualization, collaborative work environment and many other nice features to make your data science lifecycle more fun and enjoyable. Helium is a framework that manages pluggable components like Visualization, Spell inside of Zeppelin. Pluggable component extends Zeppelin's capability and particularly useful when Zeppelin is being used as a collaborative data science environment. Moon will demonstrate create custom visualization, publish to Helium online registry and use them in the notebook. Also talk about how Helium framework and Helium online registry works behind the scene and future roadmap as well. You'll see not only how easy creating and publishing Helium package is but also what possibility these pluggable modules gives to Zeppelin as a data science tool and business intelligence tool.
How Parse Built a Mobile Backend as a Service on AWS (MBL307) | AWS re:Invent...Amazon Web Services
Parse is a BaaS for mobile developers that is built entirely on AWS. With over 150,000 mobile apps hosted on Parse, the stability of the platform is our primary concern, but it coexists with rapid growth and a demanding release schedule. This session is a technical discussion of the current architecture and the design decisions that went in to scaling the platform rapidly and robustly over the past year and a half. We talk about some of the lessons learned managing and scaling MongoDB, Cassandra, Redis, and MySQL in the cloud. We also discuss how Parse went from launching individual instances using chef to managing clusters of hosts with Auto Scaling groups, with instance discovery and registry handled by ZooKeeper, thus enabling us to manage vastly larger sets of services with fewer human resources. This session is useful to anyone who is trying to scale up from startup to established platform without sacrificing agility.
This document discusses simplifying, standardizing, and automating application deployment processes before moving to the cloud. It recommends using central configuration repositories and automation tools like Chef to deploy identical environments for development, staging, and production. This allows using the same processes and tools across environments. AWS services like OpsWorks can then be used to deploy production using the same Chef configurations. The key is treating the cloud as a tool to deploy standardized, automated applications at scale.
The document discusses using Python with AWS Lambda. It introduces serverless frameworks like AWS Lambda, Google Cloud Functions, and Microsoft Azure Functions. It then focuses on using Python with AWS Lambda, including how to package dependencies, deploy code, and leverage the Serverless framework to simplify deployments. Logging to CloudWatch Logs and adding unit tests are also covered.
Apache Zeppelin on Kubernetes with Spark and Kafka - meetup @twitterApache Zeppelin
This document summarizes the agenda for a Zeppelin Meetup. The agenda includes demos of real-time streaming and running Zeppelin on Kubernetes, as well as discussing the Zeppelin roadmap and taking questions. The roadmap focuses on modernizing the front-end, adding collaboration features, running applications alongside notebooks on Kubernetes, and improving visualization capabilities.
Dan Russell, Director of Technical Operations at an unknown company, advocates for more transparency in business ownership so companies can no longer be used anonymously against the public good. He proposes igniting world opinion, changing laws, and launching a new era of openness. The document also discusses using Fastly's API to support all features in their web interface, colocating a CDN and code through a unified code repository for versioning, deploys, discoverability, and linking CDN features to drive automated tests in a recurring monitoring system that end-to-end colocates code, databases, tests, environments, configurations, and monitoring.
- Big data refers to large amounts of data that are constantly being generated and require advanced techniques and technologies to capture, manage and analyze. Examples include web logs, social media, sensor data, and financial transactions.
- Hadoop is an open-source software framework for distributed storage and processing of large datasets across clusters of commodity servers. It allows for the reliable, scalable and distributed processing of large data sets across clusters of commodity hardware.
- Key components of Hadoop include HDFS for storage, and MapReduce for distributed processing of large datasets in parallel on multiple servers. HDFS stores data across clusters of commodity servers as blocks, providing reliability even if individual servers fail.
The document discusses Parse's process for benchmarking MongoDB upgrades by replaying recorded production workloads on test servers. They found a 33-75% drop in throughput when upgrading from 2.4.10 to 2.6.3 due to query planner bugs. Working with MongoDB, they identified and helped fix several bugs, improving performance in 2.6.5 but still below 2.4.10 levels initially. Further optimization work increased throughput above 2.4.10 levels when testing with more workers and operations.
Serverless is a new framework that allows developers to easily harness AWS Lambda and Api Gateway to build and deploy full fledged API services without needing to deal with any ops level overhead or paying for servers when they're not in use. It's kinda like Heroku on-demand for single functions.
This document provides an overview of Kubernetes concepts including pods, replica sets, deployments, services, ingress, volumes, config maps and secrets. It discusses how pods are the basic scalable unit in Kubernetes and how replica sets and deployments help manage and update pods. It also describes how services provide discovery and load balancing for pods and how ingress manages external traffic. Finally, it summarizes how volumes, config maps and secrets can be used to provide storage and configuration to containers.
Big Data at Riot Games – Using Hadoop to Understand Player Experience - Stamp...StampedeCon
At the StampedeCon 2013 Big Data conference in St. Louis, Riot Games discussed Using Hadoop to Understand and Improve Player Experience. Riot Games aims to be the most player-focused game company in the world. To fulfill that mission, it’s vital we develop a deep, detailed understanding of players’ experiences. This is particularly challenging since our debut title, League of Legends, is one of the most played video games in the world, with more than 32 million active monthly players across the globe. In this presentation, we’ll discuss several use cases where we sought to understand and improve the player experience, the challenges we faced to solve those use cases, and the big data infrastructure that supports our capability to provide continued insight.
Fullstack DDD with ASP.NET Core and Anguar 2 - Ronald Harmsen, NForzaCodemotion Tel Aviv
This document discusses building fullstack applications using domain-driven design (DDD) principles with ASP.NET Core and Angular 2. It focuses on using a task-based user interface design with Angular 2 to present data from read models and allow changing data through commands, fitting well with DDD and command query responsibility segregation (CQRS). The backend is built with ASP.NET Core hosting a frontend web API with JavaScript services and using HAL to expose a REST API to the Angular 2 frontend.
Our challenge for Bulkload reliability improvementSatoshi Akama
This document summarizes a technical talk about improving reliability in bulk loading data using Embulk. It discusses developing Embulk plugins including adding retry logic to handle errors from external services, choosing between Java and JRuby, writing unit tests, and battling specifications of different cloud services. It also covers using Embulk at Treasure Data for bulk loading including architecture, APIs, and writing integration tests. Areas for future improvement discussed are handling timeouts, errors from services, and API limits in tests as well as implementing new validation endpoints.
This document discusses using Kubernetes as a cloud data platform and summarizes key services and considerations. It outlines using Presto and Spark on Kubernetes for data processing and analytics workloads. Example architectures are provided for Presto and Spark deployments on Kubernetes, including building Docker images. Operating and managing Kubernetes clusters on AWS is also addressed.
With more than 140 million users, KakaoTalk is the most popular mobile messaging platform in South Korea. The team at daumkakao has been using OpenStack with the intention for tranforming the current legacy infrastructure into scale out based cloud to build and offer new services for its users. In this session, we'd like to share our experiences with the OpenStack community, specifically in regards to meeting our needs for networking with Neutron.OpenStack Neutron offers a lot of methods to implement networking for VMs and containers. For production operations, VM migration can be a common activity to manage resources and improve uptime. It's not hard using shared storage like Ceph, but network settings, such as IP addresses need to be preserved. With a shared storage environment, an image can be attached anywhere inside of a data center, but a service IP for a virtual machine is different story. And when you don't use the floating IPs, keeping the same IP across a data center-wide set of VLANs is hard job.To maintain a virtual machine's IP settings and balance IPs between VLANS, we tried several options including overlay, SDN, and NFV technologies. In the end we came to use a route-only network for our virtual machine networks, leveraging technology like Quagga for RIP, OSPF BGP integrated with Neutron.
This document provides an overview of some techniques for enhancing iOS development using blocks and categories. It discusses enhancing NSArray with enumeration, filtering, transformation and other methods implemented using blocks. It also covers sizing and laying out views by implementing sizeThatFits and layoutSubviews in a paired way. Handling gestures with blocks is described as providing a simpler approach than selector-based actions. The document encourages using these techniques to simplify code and make it easier to maintain.
A presentation given to the Melbourne Scala User Group in August 2013 talking about how PlayUp is using scala, akka, spray and neo4j to drive it's next generation of mobile sports applications.
4.Building a Data Product using apache Zeppelin - Apache Kylin Meetup @ShanghaiLuke Han
The document discusses building a data product using Apache Zeppelin (incubating) that sends email notifications about new open source projects on GitHub. It outlines the steps to download data from the GitHub archive, explore and filter the data to focus on interesting companies, join additional data from the GitHub API, generate an HTML template to visualize the results, and schedule sending the email notifications.
ClojuTRE2015: Kekkonen - making your Clojure web APIs more awesomeMetosin Oy
Thanks to REST and Swagger, we can build beautiful apis to feed both our browser front-ends and external applications. But, wrapping your Clojure code into resources mostly for your ClojureScript front-end doesn't feel right? Just use RPC? Meet in the middle?
Kekkonen is a small library for managing your (web) apis as commands and queries. No magic, data-driven, un-restful and non-rpc. It's goals are to be small, explicit, extendable and to help enforce your business rules both on the server side and on the ClojureScript frontend. Besides Swagger, it provides run-time context-aware apidocs for Clojure(Script).
Apache Zeppelin is interactive data analytics environment for large scale data processing systems. It deeply integrates to Apache spark and many other frameworks, provides beautiful interactive web-based interface, data visualization, collaborative work environment and many other nice features to make your data science lifecycle more fun and enjoyable. Helium is a framework that manages pluggable components like Visualization, Spell inside of Zeppelin. Pluggable component extends Zeppelin's capability and particularly useful when Zeppelin is being used as a collaborative data science environment. Moon will demonstrate create custom visualization, publish to Helium online registry and use them in the notebook. Also talk about how Helium framework and Helium online registry works behind the scene and future roadmap as well. You'll see not only how easy creating and publishing Helium package is but also what possibility these pluggable modules gives to Zeppelin as a data science tool and business intelligence tool.
How Parse Built a Mobile Backend as a Service on AWS (MBL307) | AWS re:Invent...Amazon Web Services
Parse is a BaaS for mobile developers that is built entirely on AWS. With over 150,000 mobile apps hosted on Parse, the stability of the platform is our primary concern, but it coexists with rapid growth and a demanding release schedule. This session is a technical discussion of the current architecture and the design decisions that went in to scaling the platform rapidly and robustly over the past year and a half. We talk about some of the lessons learned managing and scaling MongoDB, Cassandra, Redis, and MySQL in the cloud. We also discuss how Parse went from launching individual instances using chef to managing clusters of hosts with Auto Scaling groups, with instance discovery and registry handled by ZooKeeper, thus enabling us to manage vastly larger sets of services with fewer human resources. This session is useful to anyone who is trying to scale up from startup to established platform without sacrificing agility.
This document discusses simplifying, standardizing, and automating application deployment processes before moving to the cloud. It recommends using central configuration repositories and automation tools like Chef to deploy identical environments for development, staging, and production. This allows using the same processes and tools across environments. AWS services like OpsWorks can then be used to deploy production using the same Chef configurations. The key is treating the cloud as a tool to deploy standardized, automated applications at scale.
The document discusses using Python with AWS Lambda. It introduces serverless frameworks like AWS Lambda, Google Cloud Functions, and Microsoft Azure Functions. It then focuses on using Python with AWS Lambda, including how to package dependencies, deploy code, and leverage the Serverless framework to simplify deployments. Logging to CloudWatch Logs and adding unit tests are also covered.
Apache Zeppelin on Kubernetes with Spark and Kafka - meetup @twitterApache Zeppelin
This document summarizes the agenda for a Zeppelin Meetup. The agenda includes demos of real-time streaming and running Zeppelin on Kubernetes, as well as discussing the Zeppelin roadmap and taking questions. The roadmap focuses on modernizing the front-end, adding collaboration features, running applications alongside notebooks on Kubernetes, and improving visualization capabilities.
Dan Russell, Director of Technical Operations at an unknown company, advocates for more transparency in business ownership so companies can no longer be used anonymously against the public good. He proposes igniting world opinion, changing laws, and launching a new era of openness. The document also discusses using Fastly's API to support all features in their web interface, colocating a CDN and code through a unified code repository for versioning, deploys, discoverability, and linking CDN features to drive automated tests in a recurring monitoring system that end-to-end colocates code, databases, tests, environments, configurations, and monitoring.
- Big data refers to large amounts of data that are constantly being generated and require advanced techniques and technologies to capture, manage and analyze. Examples include web logs, social media, sensor data, and financial transactions.
- Hadoop is an open-source software framework for distributed storage and processing of large datasets across clusters of commodity servers. It allows for the reliable, scalable and distributed processing of large data sets across clusters of commodity hardware.
- Key components of Hadoop include HDFS for storage, and MapReduce for distributed processing of large datasets in parallel on multiple servers. HDFS stores data across clusters of commodity servers as blocks, providing reliability even if individual servers fail.
The document discusses Parse's process for benchmarking MongoDB upgrades by replaying recorded production workloads on test servers. They found a 33-75% drop in throughput when upgrading from 2.4.10 to 2.6.3 due to query planner bugs. Working with MongoDB, they identified and helped fix several bugs, improving performance in 2.6.5 but still below 2.4.10 levels initially. Further optimization work increased throughput above 2.4.10 levels when testing with more workers and operations.
Serverless is a new framework that allows developers to easily harness AWS Lambda and Api Gateway to build and deploy full fledged API services without needing to deal with any ops level overhead or paying for servers when they're not in use. It's kinda like Heroku on-demand for single functions.
This document provides an overview of Kubernetes concepts including pods, replica sets, deployments, services, ingress, volumes, config maps and secrets. It discusses how pods are the basic scalable unit in Kubernetes and how replica sets and deployments help manage and update pods. It also describes how services provide discovery and load balancing for pods and how ingress manages external traffic. Finally, it summarizes how volumes, config maps and secrets can be used to provide storage and configuration to containers.
Big Data at Riot Games – Using Hadoop to Understand Player Experience - Stamp...StampedeCon
At the StampedeCon 2013 Big Data conference in St. Louis, Riot Games discussed Using Hadoop to Understand and Improve Player Experience. Riot Games aims to be the most player-focused game company in the world. To fulfill that mission, it’s vital we develop a deep, detailed understanding of players’ experiences. This is particularly challenging since our debut title, League of Legends, is one of the most played video games in the world, with more than 32 million active monthly players across the globe. In this presentation, we’ll discuss several use cases where we sought to understand and improve the player experience, the challenges we faced to solve those use cases, and the big data infrastructure that supports our capability to provide continued insight.
Fullstack DDD with ASP.NET Core and Anguar 2 - Ronald Harmsen, NForzaCodemotion Tel Aviv
This document discusses building fullstack applications using domain-driven design (DDD) principles with ASP.NET Core and Angular 2. It focuses on using a task-based user interface design with Angular 2 to present data from read models and allow changing data through commands, fitting well with DDD and command query responsibility segregation (CQRS). The backend is built with ASP.NET Core hosting a frontend web API with JavaScript services and using HAL to expose a REST API to the Angular 2 frontend.
Our challenge for Bulkload reliability improvementSatoshi Akama
This document summarizes a technical talk about improving reliability in bulk loading data using Embulk. It discusses developing Embulk plugins including adding retry logic to handle errors from external services, choosing between Java and JRuby, writing unit tests, and battling specifications of different cloud services. It also covers using Embulk at Treasure Data for bulk loading including architecture, APIs, and writing integration tests. Areas for future improvement discussed are handling timeouts, errors from services, and API limits in tests as well as implementing new validation endpoints.
This document discusses using Kubernetes as a cloud data platform and summarizes key services and considerations. It outlines using Presto and Spark on Kubernetes for data processing and analytics workloads. Example architectures are provided for Presto and Spark deployments on Kubernetes, including building Docker images. Operating and managing Kubernetes clusters on AWS is also addressed.
With more than 140 million users, KakaoTalk is the most popular mobile messaging platform in South Korea. The team at daumkakao has been using OpenStack with the intention for tranforming the current legacy infrastructure into scale out based cloud to build and offer new services for its users. In this session, we'd like to share our experiences with the OpenStack community, specifically in regards to meeting our needs for networking with Neutron.OpenStack Neutron offers a lot of methods to implement networking for VMs and containers. For production operations, VM migration can be a common activity to manage resources and improve uptime. It's not hard using shared storage like Ceph, but network settings, such as IP addresses need to be preserved. With a shared storage environment, an image can be attached anywhere inside of a data center, but a service IP for a virtual machine is different story. And when you don't use the floating IPs, keeping the same IP across a data center-wide set of VLANs is hard job.To maintain a virtual machine's IP settings and balance IPs between VLANS, we tried several options including overlay, SDN, and NFV technologies. In the end we came to use a route-only network for our virtual machine networks, leveraging technology like Quagga for RIP, OSPF BGP integrated with Neutron.
This document provides an overview of some techniques for enhancing iOS development using blocks and categories. It discusses enhancing NSArray with enumeration, filtering, transformation and other methods implemented using blocks. It also covers sizing and laying out views by implementing sizeThatFits and layoutSubviews in a paired way. Handling gestures with blocks is described as providing a simpler approach than selector-based actions. The document encourages using these techniques to simplify code and make it easier to maintain.
A presentation given to the Melbourne Scala User Group in August 2013 talking about how PlayUp is using scala, akka, spray and neo4j to drive it's next generation of mobile sports applications.
Hypermedia for the iOS developer - Swipe 2012Kevin O'Neill
The document discusses consuming hypermedia APIs for iOS development. It introduces hypermedia and RESTful principles like loose coupling, uniform interfaces, and using HTTP and links to guide clients. The presentation explains how to handle different media types and link relations, with examples showing embedded links for alternate representations and versioning. Errors are also handled through linked responses. Overall it provides an overview of designing and using hypermedia APIs on iOS with a focus on REST constraints.
Building Hypermedia API's - YOW! Night - March 2013Kevin O'Neill
Building a client API that allows you to scale and evolve is a delicate task. If your API is intended to be used with mobile devices the problem's of latency, bandwidth uncertain connection state, multiple client versions and capabilities compound the problem. See how PlayUp has used hypermedia (ReST) to power and evolve it's applications on iOS, Android, Windows Phone and Windows 8. See a general overview of the principles that guide development, how the client and the server work, the 'oh I wish we hadn't done that moments' and take a tour of the document exchanges that power PlayUp's ticketing service.
YOW Mobile Night 2011 - The realestate.com.au mobile storyKevin O'Neill
A brief summary of the path taken by the mobile development team at realestate.com.au.
It provides an overview of the project from inception to cancellation, to rebirth, multi-platform and tablet.
The presentation concludes with some observations, on the platform, technologies and the future.
- Server-side rendering techniques like pjax and client-side includes allow for partial page updates and better caching while enabling cross-team components and server-driven updates.
- These techniques provide benefits like simpler deployments and operations while addressing challenges like rate of change and device diversity.
- Examples include using pjax for partial page updates over AJAX and client-side includes for caching and reusable content across includes.
Elasticsearch is recommended to create an archive to search ACM/BPM case and process data that is up to 7 years old. Elasticsearch allows storing and searching large volumes of data quickly and in near real-time. It was tested by uploading over 40,000 documents from a use case involving tweets. This allowed full-text search of case data and searching within office documents. While Elasticsearch is schema-less and easy to evolve with Oracle releases, its limitations regarding transactions and an overview of case history would need to be considered.
Terraform is an open source tool that helps you control your infrastructure configuration through code. This talk will serve as a primer showing how to build a basic infrastructure in the Google Cloud and how we can re-use our code to construct multiple, identical environments.
Ensuring Quality in Data Lakes (D&D Meetup Feb 22)lakeFS
The document discusses improving data quality in a data lake. It describes three levels (L1-L3) of data lake maturity:
L1 involves storing data in an object store in a basic format like CSV files. This provides good performance, cost efficiency, and developer experience.
L2 adds optimized table formats like Delta Lake, Hudi and Iceberg that maintain metadata and transaction logs to enable features like schema enforcement, data versioning and isolation.
L3 adds data version control systems like lakeFS that extend the object store with Git-like source control operations. This allows instantly reverting bad data, developing data in isolation, and simplifying data reproducibility. LakeFS was demonstrated as an example solution
Blockchain Software for Hardware: The Canaan AvalonMiner Open Source Embedded...Mike Qin
The Canaan AvalonMiner 721 is hardware blockchain securing hardware technology built using Open Source software. The latest hardware AvalonMiner 721 contains 72 customized ASIC processors named Canaan A3212 which are harmonized together and then across potentially thousands of other AvalonMiners, to do SHA-256 hashing calculations.From the software toolchain all the way to OpenWrt used in Canaan's AvalonMiner Controller, this presentation describes the entire process from start to finish how the software is built, developed, launched and maintained. Of specific interest are the contributions by Canaan back to the Open Source community, including to CGMiner, OpenWRT and other projects.
This document discusses various topics related to building serverless applications on AWS, including API Gateway and Kinesis, authentication and authorization, testing, monitoring, logging, performance optimization, and CI/CD practices. It also provides information about DAZN, an online sports streaming service, and mentions that they are hiring.
This document summarizes a webinar about importing crime data from Chicago into Neo4j. It discusses loading a CSV file of crime data into Neo4j using LOAD CSV and creating nodes and relationships. It also describes using Spark to preprocess the CSV into multiple Neo4j-formatted files and bulk loading them using the Neo4j Import tool. The document then covers enriching the graph with additional crime data from JSON and updating the graph with new crimes.
This document provides an overview of Kubernetes concepts including pods, replica sets, deployments, stateful sets, and services. It explains how pods group containers, and how replica sets, deployments, and stateful sets manage pods. It also describes how services provide discovery and load balancing for sets of pods. The document uses diagrams to illustrate how these resources interact and are managed in a Kubernetes cluster.
Kube Your Enthusiasm - Paul CzarkowskiVMware Tanzu
This document provides an overview of container platforms and Kubernetes concepts. It discusses hardware platforms, infrastructure as a service (IaaS), container as a service (CaaS), platform as a service (PaaS), and function as a service (FaaS). It then covers Kubernetes architecture and resources like pods, services, volumes, replica sets, deployments, and stateful sets. Examples are given of using kubectl to deploy and manage applications on Kubernetes.
This document is a presentation from OpenStack Summit Sydney. It describes how to easily install OpenStack on Kubernetes. It explains Kubernetes and OpenStack-Helm.
The document discusses using Visual Studio Code for building, optimizing, and deploying Angular applications to Azure using Docker. It provides steps for scaffolding a new Angular app, building for development and production with optimizations, analyzing build outputs, debugging Dockerized apps in VS Code, and deploying built apps to Azure Web Apps from within VS Code using Docker. Key advantages highlighted include great debugging experiences for Node.js and Docker apps, useful Azure and Node.js extensions, and an integrated Docker development experience.
Spring is a large ecosystem of projects including Spring Boot for creating microservice applications, Spring Data for database access, Spring Cloud for distributed systems, and many others. It advocates convention over configuration and provides tools for integrating technologies like databases, messaging, web services, and more.
(ARC402) Deployment Automation: From Developers' Keyboards to End Users' Scre...Amazon Web Services
Some of the best businesses today are deploying their code dozens of times a day. How? By making heavy use of automation, smart tools, and repeatable patterns to get process out of the way and keep the workflow moving. Come to this session to learn how you can do this too, using services such as AWS OpsWorks, AWS CloudFormation, Amazon Simple Workflow Service, and other tools. We'll discuss a number of different deployment patterns, and what aspects you need to focus on when working toward deployment automation yourself.
The fundamental performance characteristics of Node.js make it ideal for building highly performant microservices for a number of workloads. Translating that into highly responsive, scalable solutions however is still far from easy. This session will not just discuss why Node.js is a natural fit for microservices, but will introduce you to the tools and best practices for creating, building, deploying, monitoring and tracing microservices that are both scalable and fault tolerant, and show through a live demo how do that with minimal effort.
Speakers:
Chris Bailey, Chief Architect, Cloud Native Runtimes, IBM
Beth Griggs, Node.js Developer, IBM
This document summarizes some modern Objective-C features including:
1. New object literal syntax that reduces verbosity when creating arrays, dictionaries, and numbers. Object subscripting allows using subscript syntax with objects.
2. The @import declaration provides a simpler way to import frameworks compared to #import, automatically linking frameworks.
3. The instancetype contextual keyword helps the compiler infer the correct return type for initializers and convenience constructors.
Cloud Native Night June 2019, Munich: Talk by Josef Fuchshuber (@fuchshuber, Principal Software Architect at QAware)
Join our Meetup: www.meetup.com/cloud-native-muc
Abstract: Kubernetes ist komplex geworden. Eigentlich so komplex, dass man sich als App Developer mit diesem Komplexitätsgrad auf Platform-Level nur selten befassen will. Was aber ist die richtige Abstraktionsebene für App Developer? In diesem Talk werden zwei mögliche Lösungen vorgestellt:
- Crossplane: Ein Multicloud Control Plane um Workload und Ressourcen unabhängig von Cloud Providern ausrollen zu können
- Knative: Plattform um Serverless Workloads zu Bauen, Deployen und zu Managen
Beide Tools definieren eine abstrahierte Sicht auf K8s, haben aber jeweils einen eigenen Lösungsansatz und unterschiedliche Einsatzszenarien. Diese werden in der Präsentation vorgestellt und mit Demos detaillierter erläutert.
Node Interactive: Node.js Performance and Highly Scalable Micro-ServicesChris Bailey
The fundamental performance characteristics of Node.js, along with the improvements driven through the community benchmarking workgroup, makes Node.js ideal for highly performing micro-service workloads. Translating that into highly responsive, scalable solutions however is still far from easy. This session will discuss why Node.js is right for micro-services, introduce the best practices for building scalable deployments, and show you how to monitor and profile your applications to identify and resolve performance bottlenecks.
Grokking Engineering - Data Analytics Infrastructure at Viki - Huy NguyenHuy Nguyen
This document outlines Viki's analytics infrastructure, including data collection, storage, processing, and visualization. It discusses collecting behavioral data from various sources and storing it in Hadoop. Data is centralized, cleaned, transformed, and loaded into a PostgreSQL data warehouse for analysis. Real-time data is processed using Apache Storm and visualized on dashboards and alerts. Technologies used include Ruby, Python, Java, Hadoop, Hive, and Amazon Redshift for analytics and PostgreSQL, MongoDB, and Redis for transactional data.
Ruby is just over 20 years old. It's no longer young or hip, and that’s a good thing! In the last decade, Ruby has matured as a web technology. It's being used in many successful companies out there such as Hulu, GitHub, and Bloomberg. The ecosystem is comprised of many stable libraries and tools to handle most common web tasks, allowing you to focus on adding features to improve your product and better serve your customers. We'll talk about how you can build scalable and reliable software, but still maintain fast development turnaround by leveraging the maturity and creativity of the Ruby community.
Tutorial 1: Your First Science App - Araport Developer WorkshopVivek Krishnakumar
Slide deck pertaining to Tutorial 1 of the Araport Developer Workshop conducted at TACC, Austin TX on November 5, 2014.
Presented by Vivek Krishnakumar
I would like to present our CI Provisioning with Openstack solution and how it improved our development. The CI provisioning is meant to replace your static CI env servers to a auto provisioned during your build stage for every commit.
Abstract:
1. CI Flow Quick view
2. Openstack CI integration maven plugin
3. Orchestration using facters
4. Openstack instance queue for faster provisioning
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
AI-Powered Food Delivery Transforming App Development in Saudi Arabia.pdfTechgropse Pvt.Ltd.
In this blog post, we'll delve into the intersection of AI and app development in Saudi Arabia, focusing on the food delivery sector. We'll explore how AI is revolutionizing the way Saudi consumers order food, how restaurants manage their operations, and how delivery partners navigate the bustling streets of cities like Riyadh, Jeddah, and Dammam. Through real-world case studies, we'll showcase how leading Saudi food delivery apps are leveraging AI to redefine convenience, personalization, and efficiency.
Things to Consider When Choosing a Website Developer for your Website | FODUUFODUU
Choosing the right website developer is crucial for your business. This article covers essential factors to consider, including experience, portfolio, technical skills, communication, pricing, reputation & reviews, cost and budget considerations and post-launch support. Make an informed decision to ensure your website meets your business goals.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
25. Neo4j
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
26. Neo4j
•
•
•
•
Graph Database
Vertices and Edges
Properties
Full Text Index
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
31. JVM
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
32. JVM
•
•
•
Graph - 8G of Heap
Others 128M - 1G
Parallel GC
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
33. The remaining memory
is reserved for file
system cache and
memory mapped buffers
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
38. Debian Packages
•
•
•
Places binaries in opt
Adds an upstart file for the
Permissions on data and log directories
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
40. Base configuration is for
production
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
41. Baking
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
42. Packer
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
43. Create machine images
for multiple platforms
from a single config
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
44. Images have lots of benefits
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
86. Talk to Me or Andrea
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
87. You miss 100 percent
of the the shots you
don't take
– Wayne Gretzky
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013
88. Questions?
Kevin O’Neill
CTO - PlayUp
@kevinoneill
kevin@playup.com
Kevin O’Neill - CTO PlayUp - @kevinoneill
Infrastructure Coders - Melbourne - November 2013