This document provides an overview and agenda for a presentation on Spring Data and in-memory data management using Apache Geode. The presentation will cover Apache Geode functionality, integrating Geode with Spring frameworks, and examples of caching, events, data access and improvements in Geode and related projects. It lists caching, scalability, availability and other capabilities of Geode. The roadmap discusses upcoming versions of Spring Data GemFire and Geode as well as integration with Spring Boot, Session and other projects.
Using the awesome power of Spring Boot with Spring Data Geode to build highly-scalable, distributed Spring/Java applications using Apache Geode or Pivotal GemFire.
This is an introduction to Apache Geode (incubating) that explains some of the key concepts of the project, including steps create a build from the source code, how to create a local cluster with and without Docker, and how to create Teeny, a simple but very scalable URL shortener application.
A Series of Fortunate Events: Building an Operator in JavaVMware Tanzu
SpringOne 2021:
Session Title: A Series of Fortunate Events: Building an Operator in Java
Speakers: Alberto C. Ríos, Staff Engineer at VMware; Bella Bai, Software Engineer at VMware
Staying Ahead of the Curve with Spring and Cassandra 4 (SpringOne 2020)Alexandre Dutra
Spring and Cassandra are two of the leading technologies for building cloud native applications. In this talk by the project leads for Spring Data and the Cassandra Java Driver, we’ll cover the recent improvements in the latest and greatest versions of Spring Boot, Spring Data Cassandra, Cassandra 4.0 and the Cassandra Java driver. Whether you’re a novice, intermediate, or expert developer, this content will help you get started or migrate your existing application to the latest innovations. We’ll illustrate these new concepts with code samples and snippets that you can find on GitHub to help you get things done faster with these tools.
6 Things You Need to Know to Safely Run KubernetesVMware Tanzu
Kubernetes has exploded in popularity among developers. But as operations teams prepare to support Kubernetes in production, they have more considerations—namely, how to operate a stable platform while maintaining security and compliance. How Kubernetes is configured and deployed has a marked impact on these attributes.
Attend this session with Pivotal’s Vice President of Technology, Cornelia Davis, to learn the following:
● How to isolate tenants in your Kubernetes environment.
● How to make upgrading Kubernetes clusters boring.
● What you should—and shouldn’t—let your developers do.
● What you need around your Kubernetes clusters to keep them safe.
Presenter :
Cornelia Davis, Author and Vice President of Technology, Pivotal
Using the awesome power of Spring Boot with Spring Data Geode to build highly-scalable, distributed Spring/Java applications using Apache Geode or Pivotal GemFire.
This is an introduction to Apache Geode (incubating) that explains some of the key concepts of the project, including steps create a build from the source code, how to create a local cluster with and without Docker, and how to create Teeny, a simple but very scalable URL shortener application.
A Series of Fortunate Events: Building an Operator in JavaVMware Tanzu
SpringOne 2021:
Session Title: A Series of Fortunate Events: Building an Operator in Java
Speakers: Alberto C. Ríos, Staff Engineer at VMware; Bella Bai, Software Engineer at VMware
Staying Ahead of the Curve with Spring and Cassandra 4 (SpringOne 2020)Alexandre Dutra
Spring and Cassandra are two of the leading technologies for building cloud native applications. In this talk by the project leads for Spring Data and the Cassandra Java Driver, we’ll cover the recent improvements in the latest and greatest versions of Spring Boot, Spring Data Cassandra, Cassandra 4.0 and the Cassandra Java driver. Whether you’re a novice, intermediate, or expert developer, this content will help you get started or migrate your existing application to the latest innovations. We’ll illustrate these new concepts with code samples and snippets that you can find on GitHub to help you get things done faster with these tools.
6 Things You Need to Know to Safely Run KubernetesVMware Tanzu
Kubernetes has exploded in popularity among developers. But as operations teams prepare to support Kubernetes in production, they have more considerations—namely, how to operate a stable platform while maintaining security and compliance. How Kubernetes is configured and deployed has a marked impact on these attributes.
Attend this session with Pivotal’s Vice President of Technology, Cornelia Davis, to learn the following:
● How to isolate tenants in your Kubernetes environment.
● How to make upgrading Kubernetes clusters boring.
● What you should—and shouldn’t—let your developers do.
● What you need around your Kubernetes clusters to keep them safe.
Presenter :
Cornelia Davis, Author and Vice President of Technology, Pivotal
Cloud Computing:
Cloud computing is the delivery of different services through the Internet. These resources include tools and applications like data storage, servers, databases, networking, and software.
Scaling Prometheus Metrics in Kubernetes with Telegraf | Chris Goller | Influ...InfluxData
Scaling Prometheus in Kubernetes seems easy with service-discovery, but quickly devolves into manual DevOps snowflake setup. Additionally, a single developer is able to overwhelm a federated Prometheus setup and impact the system as a whole without being able to self-service debug. In this talk, Chris will focus on a variety of architectures using Telegraf to scale scraping in Kubernetes and empower developers.
He’ll describe his experiences around scaling /metrics in the microservices of InfluxData’s Cloud 2.0 Kubernetes system…as he was the single developer that added just one more label…
Build bundles in the cloud - How Cloudyle PaaS+ helps creating OSGi applicati...mfrancis
OSGi Community Event 2014
Abstract:
Platform as a service (PaaS) provides a ready to use development or deployment environment and lets a developer quickly start coding instead of wasting time for setup and configuration. The PaaS manages the underlying O/S, provides database installations, language runtimes and application frameworks almost instantly. The developer can immediately focus on the data model and application logic, and therefore deliver a solution to the market in less time.
This talks gives an introduction to our open PaaS+ Cloud Platform for modular OSGi applications based on OpenShift [1]. PaaS+ is a modular and scalable Java PaaS that features a modular build and deployment mechanism and helps to speed up application development while also making it more robust. We will show how PaaS+ supports OSGi application developers in the complete application lifecycle from development to production to achieve a faster time to market. Because the PaaS environment used in development, testing and production is the same, there is less overhead or problems staging the application from development to QA to productive deployment. The platform comes with a build in Apache Karaf server runtime enhanced by OSGi enabled base services such as Authentication, Rules Engine, Business Process Engine, Polyglot Persistence, Search and Indexing and an integrated OBR. These services are available through a consistent API and are orchestrated by the integrated OSGi framework.
The OSGi PaaS+ platform is based on research results from the TRESOR - Trusted Ecosystem for Standardized and Open cloud-based Resources – project to provide an open ecosystem for cloud applications for the health care sector [2]. TRESOR is part of the Trusted Cloud initiative [3], funded by the German Federal Ministry of Economics and Technology.
Topics covered:
Benefits of Cloud and Platform as a Service
Overview of the OSGi PaaS+ Cloud
Benefits of OSGi for modular cloud platforms
Cloud deployment and distributed services
PaaS+ Enterprise API
Speaker Bios:
Alexander Grzesik
Alexander is the head of development of medisite Systemhaus GmbH and responsible for the development of the the PaaS+ cloud platform the clinical information system m.life and software architect for the TRESOR Project.
He has 15 years of work experience in medical Software development as team leader and software architect.
Expert for Software Architecture, OSGi, Java and Java EE.
Alexander has been a speaker at several conferences including EclipseCon Europe and the OpenShift Community Day.
Torsten Frank
Torsten Frank is an entrepreneur and healthcare IT expert with international experience and a strong background in product and business development. He is CEO and founder of medisite, a company that delivers highly specialized healthcare IT solutions to their customers for more than 10 years. medisite is a winner of the Federal Ministry for Economic Affairs and Energy (BMWi)
Simplifying Apache Geode with Spring DataVMware Tanzu
SpringOne Platform 2017
John Blum, Pivotal
Building effective Apache Geode applications quickly and easily requires a framework that provides the right level of abstraction. In this session we take Alan Kay's infamous quote "Simple things should be simple; Complex things should be possible" to a whole new level with Spring Data Geode using Spring Boot. I'll show you how the new Annotation-based configuration model, which builds on existing concepts like SD Repositories, Spring's Cache Abstraction and Apache Geode CQ, helps you rapidly build working Apache Geode client/server applications in minutes. We end the session with a quick look at the roadmap and what users can expect next. You won't want to miss this.
Microservices with Spring and Cloud FoundryAlain Sahli
Cloud Foundry allows developers to deploy easily and often their applications. This is a great thing but it is only a part of the whole picture. In order to smoothly scale an application some patterns and basic architecture concepts must be taken into consideration. With Spring Boot, applications can be split into smaller parts called Microservices, which can be scaled in a more fine-grained manner. In addition Spring Cloud provides tools to quickly build some of the common patterns in distributed systems (e.g. configuration management, service discovery, circuit breakers).
Gradle is an open-source build automation tool focused on flexibility, build reproducibility and performance. Over the years, this tool has evolved and introduced new concepts and features around dependency management, publication and other aspects on build and release of artifacts for the Java platform.
Keeping up to date with all these features across several projects can be challenging. How do you make sure that all your projects can be upgraded to the latest version of Gradle? What if you have thousands of projects and hundreds of engineers? How can you abstract common tasks for them and make sure that new releases work as expected?
At Netflix, we built Nebula, a collection of Gradle plugins that helps engineers remove boilerplate in Gradle build files, and makes building software the Netflix way easy. This reduces the cognitive load on developers, allowing them to focus on writing code.
In this talk, I’ll share with you our philosophy on how to build JVM artifacts and the pieces that help us boost the productivity of engineers at Netflix. I’ll talk about:
- What is Nebula
- What are the common problems we face and try to solve
- How we distribute it to every JVM engineer
- How we ensure that Nebula/Gradle changes do not break builds so we can ship new features with confidence at Netflix.
---
About Roberto: Roberto Perez Alcolea is a Senior Software Engineer at Netflix. He is a member of the Java Platform team providing the core language and framework components that enable the Java community at Netflix. He's an active maintainer of Netflix Nebula Plugins (https://nebula-plugins.github.io/) and passionate about Gradle. Prior to that, he spent several years building high performant APIs with Ratpack and web applications using Grails.
Jordi Mon Companys presents an overview of Weave GitOps Core for the Free GitOps Workshop on August 19, 2021.
Weave GitOps Core is a continuous delivery product to run apps in any Kubernetes. It is free and open source, and you can get started today!
https://www.weave.works/product/gitops-core/
Chat with us on our Slack channel! #weave-gitops http://bit.ly/WeaveGitOpsSlack
If you need to invite yourself to the Slack, visit https://slack.weave.works/
An Early Evaluation of Running Spark on KubernetesDataWorks Summit
Kubernetes is an open source system to deploy, scale, and manage containerized applications anywhere. It builds on 15 years of running Google's containerized workloads and the valuable contributions from the open source community. To shepherd Kubernetes' evolution with the open source community, Google helped form the Cloud Native Computing Foundation (CNCF) and donated Kubernetes as the founding project. Starting in Spark 2.3.0, Spark has an experimental option to run clusters managed by Kubernetes. This feature makes use of the native Kubernetes scheduler that has been added to Spark. In this talk, we will provide a baseline understanding of what Kubernetes is, why it is relevant for the Spark community and how it compares to YARN. We will then look under the hood of Spark managed by Kubernetes to better understand how this works. Finally, we provide an early evaluation of this feature as well as our thoughts on the future of running Spark on Kubernetes.
Kubernetes: від знайомства до використання у CI/CDStfalcon Meetups
Kubernetes: від знайомства до використання у CI/CD
Олександр Занічковський
Technical Lead у компанії SoftServe
14+ років досвіду розробки різноманітного програмного забезпечення, як для десктопа, так і для веб
Працював фріланс-програмістом та в команді
Цікавиться архітектурою ПЗ, автоматизацією процесів інтеграції та доставки нових версій продукту, хмарними технологіями
Віднедавна займається менторінгом майбутніх техлідів
У вільний від роботи час грає на гітарі і мріє про велику сцену
Олександр поділиться власним досвідом роботи з Kubernetes:
ознайомить з базовими поняттями та примітивами K8S
опише можливі сценарії використання Kubernetes для CI/CD на прикладі GitLab
покаже, як можна використовувати постійне сховище, збирати метрики контейнерів, використовувати Ingress для роутинга запитів за певними правилами
покаже, як можна самому встановити K8S для ознайомлення чи локальної роботи
Cloud Computing:
Cloud computing is the delivery of different services through the Internet. These resources include tools and applications like data storage, servers, databases, networking, and software.
Scaling Prometheus Metrics in Kubernetes with Telegraf | Chris Goller | Influ...InfluxData
Scaling Prometheus in Kubernetes seems easy with service-discovery, but quickly devolves into manual DevOps snowflake setup. Additionally, a single developer is able to overwhelm a federated Prometheus setup and impact the system as a whole without being able to self-service debug. In this talk, Chris will focus on a variety of architectures using Telegraf to scale scraping in Kubernetes and empower developers.
He’ll describe his experiences around scaling /metrics in the microservices of InfluxData’s Cloud 2.0 Kubernetes system…as he was the single developer that added just one more label…
Build bundles in the cloud - How Cloudyle PaaS+ helps creating OSGi applicati...mfrancis
OSGi Community Event 2014
Abstract:
Platform as a service (PaaS) provides a ready to use development or deployment environment and lets a developer quickly start coding instead of wasting time for setup and configuration. The PaaS manages the underlying O/S, provides database installations, language runtimes and application frameworks almost instantly. The developer can immediately focus on the data model and application logic, and therefore deliver a solution to the market in less time.
This talks gives an introduction to our open PaaS+ Cloud Platform for modular OSGi applications based on OpenShift [1]. PaaS+ is a modular and scalable Java PaaS that features a modular build and deployment mechanism and helps to speed up application development while also making it more robust. We will show how PaaS+ supports OSGi application developers in the complete application lifecycle from development to production to achieve a faster time to market. Because the PaaS environment used in development, testing and production is the same, there is less overhead or problems staging the application from development to QA to productive deployment. The platform comes with a build in Apache Karaf server runtime enhanced by OSGi enabled base services such as Authentication, Rules Engine, Business Process Engine, Polyglot Persistence, Search and Indexing and an integrated OBR. These services are available through a consistent API and are orchestrated by the integrated OSGi framework.
The OSGi PaaS+ platform is based on research results from the TRESOR - Trusted Ecosystem for Standardized and Open cloud-based Resources – project to provide an open ecosystem for cloud applications for the health care sector [2]. TRESOR is part of the Trusted Cloud initiative [3], funded by the German Federal Ministry of Economics and Technology.
Topics covered:
Benefits of Cloud and Platform as a Service
Overview of the OSGi PaaS+ Cloud
Benefits of OSGi for modular cloud platforms
Cloud deployment and distributed services
PaaS+ Enterprise API
Speaker Bios:
Alexander Grzesik
Alexander is the head of development of medisite Systemhaus GmbH and responsible for the development of the the PaaS+ cloud platform the clinical information system m.life and software architect for the TRESOR Project.
He has 15 years of work experience in medical Software development as team leader and software architect.
Expert for Software Architecture, OSGi, Java and Java EE.
Alexander has been a speaker at several conferences including EclipseCon Europe and the OpenShift Community Day.
Torsten Frank
Torsten Frank is an entrepreneur and healthcare IT expert with international experience and a strong background in product and business development. He is CEO and founder of medisite, a company that delivers highly specialized healthcare IT solutions to their customers for more than 10 years. medisite is a winner of the Federal Ministry for Economic Affairs and Energy (BMWi)
Simplifying Apache Geode with Spring DataVMware Tanzu
SpringOne Platform 2017
John Blum, Pivotal
Building effective Apache Geode applications quickly and easily requires a framework that provides the right level of abstraction. In this session we take Alan Kay's infamous quote "Simple things should be simple; Complex things should be possible" to a whole new level with Spring Data Geode using Spring Boot. I'll show you how the new Annotation-based configuration model, which builds on existing concepts like SD Repositories, Spring's Cache Abstraction and Apache Geode CQ, helps you rapidly build working Apache Geode client/server applications in minutes. We end the session with a quick look at the roadmap and what users can expect next. You won't want to miss this.
Microservices with Spring and Cloud FoundryAlain Sahli
Cloud Foundry allows developers to deploy easily and often their applications. This is a great thing but it is only a part of the whole picture. In order to smoothly scale an application some patterns and basic architecture concepts must be taken into consideration. With Spring Boot, applications can be split into smaller parts called Microservices, which can be scaled in a more fine-grained manner. In addition Spring Cloud provides tools to quickly build some of the common patterns in distributed systems (e.g. configuration management, service discovery, circuit breakers).
Gradle is an open-source build automation tool focused on flexibility, build reproducibility and performance. Over the years, this tool has evolved and introduced new concepts and features around dependency management, publication and other aspects on build and release of artifacts for the Java platform.
Keeping up to date with all these features across several projects can be challenging. How do you make sure that all your projects can be upgraded to the latest version of Gradle? What if you have thousands of projects and hundreds of engineers? How can you abstract common tasks for them and make sure that new releases work as expected?
At Netflix, we built Nebula, a collection of Gradle plugins that helps engineers remove boilerplate in Gradle build files, and makes building software the Netflix way easy. This reduces the cognitive load on developers, allowing them to focus on writing code.
In this talk, I’ll share with you our philosophy on how to build JVM artifacts and the pieces that help us boost the productivity of engineers at Netflix. I’ll talk about:
- What is Nebula
- What are the common problems we face and try to solve
- How we distribute it to every JVM engineer
- How we ensure that Nebula/Gradle changes do not break builds so we can ship new features with confidence at Netflix.
---
About Roberto: Roberto Perez Alcolea is a Senior Software Engineer at Netflix. He is a member of the Java Platform team providing the core language and framework components that enable the Java community at Netflix. He's an active maintainer of Netflix Nebula Plugins (https://nebula-plugins.github.io/) and passionate about Gradle. Prior to that, he spent several years building high performant APIs with Ratpack and web applications using Grails.
Jordi Mon Companys presents an overview of Weave GitOps Core for the Free GitOps Workshop on August 19, 2021.
Weave GitOps Core is a continuous delivery product to run apps in any Kubernetes. It is free and open source, and you can get started today!
https://www.weave.works/product/gitops-core/
Chat with us on our Slack channel! #weave-gitops http://bit.ly/WeaveGitOpsSlack
If you need to invite yourself to the Slack, visit https://slack.weave.works/
An Early Evaluation of Running Spark on KubernetesDataWorks Summit
Kubernetes is an open source system to deploy, scale, and manage containerized applications anywhere. It builds on 15 years of running Google's containerized workloads and the valuable contributions from the open source community. To shepherd Kubernetes' evolution with the open source community, Google helped form the Cloud Native Computing Foundation (CNCF) and donated Kubernetes as the founding project. Starting in Spark 2.3.0, Spark has an experimental option to run clusters managed by Kubernetes. This feature makes use of the native Kubernetes scheduler that has been added to Spark. In this talk, we will provide a baseline understanding of what Kubernetes is, why it is relevant for the Spark community and how it compares to YARN. We will then look under the hood of Spark managed by Kubernetes to better understand how this works. Finally, we provide an early evaluation of this feature as well as our thoughts on the future of running Spark on Kubernetes.
Kubernetes: від знайомства до використання у CI/CDStfalcon Meetups
Kubernetes: від знайомства до використання у CI/CD
Олександр Занічковський
Technical Lead у компанії SoftServe
14+ років досвіду розробки різноманітного програмного забезпечення, як для десктопа, так і для веб
Працював фріланс-програмістом та в команді
Цікавиться архітектурою ПЗ, автоматизацією процесів інтеграції та доставки нових версій продукту, хмарними технологіями
Віднедавна займається менторінгом майбутніх техлідів
У вільний від роботи час грає на гітарі і мріє про велику сцену
Олександр поділиться власним досвідом роботи з Kubernetes:
ознайомить з базовими поняттями та примітивами K8S
опише можливі сценарії використання Kubernetes для CI/CD на прикладі GitLab
покаже, як можна використовувати постійне сховище, збирати метрики контейнерів, використовувати Ingress для роутинга запитів за певними правилами
покаже, як можна самому встановити K8S для ознайомлення чи локальної роботи
This is my Spring 2015 studio project. The 2nd Year Foundation Studio focused on developing an existing parking lot for UVa's sports facilities into a mixed use student housing area. My project focused on creating spaces for interaction between students and fans at the center of game day activity.
Collecting quality clinical cases is a priotity to document more evidence in medicine, providing a wide data-base allowing research, tuition and high quality reported cases.
Basel III - Implications of ImplementationDavid Kyson
This report has been commissioned to give an investigative insight into the implementation of Basel III; the implications of implementing, previous accords and also the impact this has on various systems and activities. It will explore the previous shortcomings of the accords, aswel as the new requirements. There will be a brief description on each topic as well as a sound, but critical analysis of the impact upon each of these, caused by Basel III. Topics include: Basel III, previous accords, Global Bank Lending and the Bank System.
It incorporates a variety of information sources to gain a broader understanding of viewpoints and effects, but will focus largely on Bank Behaviour in Response to Basel III: A Cross-Country Analysis by Thomas F. Cosimano and Dalia S. Hakura (2011).
CodeCamp Iasi - Creating serverless data analytics system on GCP using BigQueryMárton Kodok
Teaser: provide developers a new way of understanding advanced analytics and choosing the right cloud architecture
The new buzzword is #serverless, as there are many great services that helps us abstract away the complexity associated with managing servers. In this session we will see how serverless helps on large data analytics backends.
We will see how to architect for Cloud and implement into an existing project components that will take us into the #serverless architecture that will ingest our streaming data, run advanced analytics on petabytes of data using BigQuery on Google Cloud Platform - all this next to an existing stack, without being forced to reengineer our app.
BigQuery enables super-fast, SQL/Javascript queries against petabytes of data using the processing power of Google’s infrastructure. We will cover its core features, SQL 2011 standard, working with streaming inserts, User Defined Functions written in Javascript, reference external JS libraries, and several use cases for everyday backend developer: funnel analytics, email heatmap, custom data processing, building dashboards, extracting data using JS functions, emitting rows based on business logic.
Gimel at Dataworks Summit San Jose 2018Romit Mehta
We presented Gimel at Dataworks Summit in San Jose.
Gimel is the open source unified data API which enables connectivity to any data store with a single API. Along with the API which works with Scala and Python, we are also surfacing a SQL interface to access any data store with just SQL.
Now data scientists and analysts can directly consume data from big data platforms like Kafka for real-time streaming data access or Elastic for search-related data all with SQL just like they can access Oracle or Teradata.
On the other hand, data engineers can relax now with this abstracted API since it isolates the ever-changing world of big data infrastructure from their code. No longer do they need to worry about API versions, connector versions, data store-specific semantics, or compute engine and version.
Gimel is also tightly integrated with Jupyter notebooks so all of the power is now available to anyone with a browser.
gimel.io
unifieddatacatalog.io
ppextensions.io
Ping me on LinkedIn for more info!
https://dataworkssummit.com/san-jose-2018/expo-theatre/gimel-paypals-analytics-data-platform/
At PayPal, data engineers, analysts, and data scientists work with a variety of data sources (Messaging, NoSQL, RDBMS, Documents, TSDB), compute engines (Spark, Flink, Beam, Hive), languages (Scala, Python, SQL) and execution models (stream, batch, interactive). Due to this complex matrix of technologies and thousands of datasets, engineers spend considerable time learning about different data sources, formats, programming models, APIs, optimizations, etc. which impacts time-to-market (TTM). To solve this problem and to make product development more effective, PayPal Data Platform developed “Gimel”, a unified analytics data platform which provides access to any storage through a single unified data API and SQL, that is powered by a centralized data catalog. In this session, we will introduce you to the various components of Gimel – Compute Platform, Data API, PCatalog, GSQL, and Notebooks. We will provide a demo depicting how Gimel reduces TTM by helping our engineers write a single line of code to access any storage without knowing the complexity behind the scenes.
Site | https://www.infoq.com/qconai2018/
Youtube | https://www.youtube.com/watch?v=2h0biIli2F4&t=19s
At PayPal, data engineers, analysts and data scientists work with a variety of datasources (Messaging, NoSQL, RDBMS, Documents, TSDB), compute engines (Spark, Flink, Beam, Hive), languages (Scala, Python, SQL) and execution models (stream, batch, interactive).
Due to this complex matrix of technologies and thousands of datasets, engineers spend considerable time learning about different data sources, formats, programming models, APIs, optimizations, etc. which impacts time-to-market (TTM). To solve this problem and to make product development more effective, PayPal Data Platform developed "Gimel", a unified analytics data platform which provides access to any storage through a single unified data API and SQL, that are powered by a centralized data catalog.
In this session, we will introduce you to the various components of Gimel - Compute Platform, Data API, PCatalog, GSQL and Notebooks. We will provide a demo depicting how Gimel reduces TTM by helping our engineers write a single line of code to access any storage without knowing the complexity behind the scenes.
Gimel and PayPal Notebooks @ TDWI Leadership Summit OrlandoRomit Mehta
This is my presentation at TDWI Leadership Summit. It talks about how products like Gimel, Unified Data Catalog and PayPal Notebooks help improve data scientist productivity and enable machine learning at scale at PayPal.
Event: Passcamp, 07.12.2017
Speaker: Stefan Kirner
Mehr Tech-Vorträge: https://www.inovex.de/de/content-pool/vortraege/
Mehr Tech-Artikel: https://www.inovex.de/blog
In this session we review the design of the current capabilities of the Spring Data GemFire API that supports Geode, and explore additional use cases and future direction that the Spring API and underlying Geode support might evolve.
#GeodeSummit - Large Scale Fraud Detection using GemFire Integrated with Gree...PivotalOpenSourceHub
In this session we explore a case study of a large-scale government fraud detection program that prevents billions of dollars in fraudulent payments each year leveraging the beta release of the GemFire+Greenplum Connector, which is planned for release in GemFire 9. Topics will include an overview of the system architecture and a review of the new GemFire+Greenplum Connector features that simplify use cases requiring a blend of massively parallel database capabilities and accelerated in-memory data processing.
Pivoting Spring XD to Spring Cloud Data Flow with Sabby AnandanPivotalOpenSourceHub
Pivoting Spring XD to Spring Cloud Data Flow: A microservice based architecture for stream processing
Microservice based architectures are not just for distributed web applications! They are also a powerful approach for creating distributed stream processing applications. Spring Cloud Data Flow enables you to create and orchestrate standalone executable applications that communicate over messaging middleware such as Kafka and RabbitMQ that when run together, form a distributed stream processing application. This allows you to scale, version and operationalize stream processing applications following microservice based patterns and practices on a variety of runtime platforms such as Cloud Foundry, Apache YARN and others.
About Sabby Anandan
Sabby Anandan is a Product Manager at Pivotal. Sabby is focused on building products that eliminate the barriers between application development, cloud, and big data.
SpringBoot and Spring Cloud Service for MSAOracle Korea
Cloud 환경에서 MSA를 하기 위해서 Service Discovery, Circuit Breaker 등을 사용하여 Application을 개발하는 방법과 SpringBoot 와 Spring Cloud Service 를 사용하는데, Cloud에서 Kubernetes를 위시한 Container 생태계가 어떻게 MSA에 영향을 미치는지 알아봅니다.
How a distributed graph analytics platform uses Apache Kafka for data ingesti...HostedbyConfluent
Using Kafka to stream data into TigerGraph, a distributed graph database, is a common pattern in our customers’ data architecture. In the TigerGraph database, Kafka Connect framework was used to build the native S3 data loader. In TigerGraph Cloud, we will be building native integration with many data sources such as Azure Blob Storage and Google Cloud Storage using Kafka as an integrated component for the Cloud Portal.
In this session, we will be discussing both architectures: 1. built-in Kafka Connect framework within TigerGraph database; 2. using Kafka cluster for cloud native integration with other popular data sources. Demo will be provided for both data streaming processes.
Best practices for developing your Magento Commerce on CloudOleg Posyniak
Properly implementing Magento Commerce Cloud is critical to the success of your online store. In this session, we’ll take a look under the hood and share how to maximize the value of your Cloud project through Docker-based local development, configurations to optimize deployments, and tools for performance monitoring (New Relic), and optimization (Blackfire).
Integrating Google Cloud Dataproc with Alluxio for faster performance in the ...Alluxio, Inc.
Alluxio Tech Talk
Dec 10, 2019
Chris Crosbie and Roderick Yao from the Google Dataproc team and Dipti Borkar of Alluxio will demo how to set up Google Cloud Dataproc with Alluxio so jobs can seamlessly read from and write to Cloud Storage. They’ll also show how to run Dataproc Spark against a remote HDFS cluster.
For more Alluxio events: https://www.alluxio.io/events/
PayPal datalake journey | teradata - edge of next | san diego | 2017 october ...Deepak Chandramouli
PayPal Data Lake Journey | 2017-Oct | San Diego | Teradata Edge of Next
Gimel [http://www.gimel.io] is a Big Data Processing Library, open sourced by PayPal.
https://www.youtube.com/watch?v=52PdNno_9cU&t=3s
Gimel empowers analysts, scientists, data engineers alike to access a variety of Big Data / Traditional Data Stores - with just SQL or a single line of code (Unified Data API).
This is possible via the Catalog of Technical properties abstracted from users, along with a rich collection of Data Store Connectors available in Gimel Library.
A Catalog provider can be Hive or User Supplied (runtime) or UDC.
In addition, PayPal recently open sourced UDC [Unified Data Catalog], which can host and serve the Technical Metatada of the Data Stores & Objects. Visit http://www.unifieddatacatalog.io to experience first hand.
Voxxed Days Cluj - Powering interactive data analysis with Google BigQueryMárton Kodok
Every company,
no matter how far from the tech they are,
is evolving into a software company,
and by extension a data company.
For a small company it’s important
to have access to modern BigData tools
without running a dedicated team for it.
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Quarkus Hidden and Forbidden ExtensionsMax Andersen
Quarkus has a vast extension ecosystem and is known for its subsonic and subatomic feature set. Some of these features are not as well known, and some extensions are less talked about, but that does not make them less interesting - quite the opposite.
Come join this talk to see some tips and tricks for using Quarkus and some of the lesser known features, extensions and development techniques.
Field Employee Tracking System| MiTrack App| Best Employee Tracking Solution|...informapgpstrackings
Keep tabs on your field staff effortlessly with Informap Technology Centre LLC. Real-time tracking, task assignment, and smart features for efficient management. Request a live demo today!
For more details, visit us : https://informapuae.com/field-staff-tracking/
Exploring Innovations in Data Repository Solutions - Insights from the U.S. G...Globus
The U.S. Geological Survey (USGS) has made substantial investments in meeting evolving scientific, technical, and policy driven demands on storing, managing, and delivering data. As these demands continue to grow in complexity and scale, the USGS must continue to explore innovative solutions to improve its management, curation, sharing, delivering, and preservation approaches for large-scale research data. Supporting these needs, the USGS has partnered with the University of Chicago-Globus to research and develop advanced repository components and workflows leveraging its current investment in Globus. The primary outcome of this partnership includes the development of a prototype enterprise repository, driven by USGS Data Release requirements, through exploration and implementation of the entire suite of the Globus platform offerings, including Globus Flow, Globus Auth, Globus Transfer, and Globus Search. This presentation will provide insights into this research partnership, introduce the unique requirements and challenges being addressed and provide relevant project progress.
Top Features to Include in Your Winzo Clone App for Business Growth (4).pptxrickgrimesss22
Discover the essential features to incorporate in your Winzo clone app to boost business growth, enhance user engagement, and drive revenue. Learn how to create a compelling gaming experience that stands out in the competitive market.
Software Engineering, Software Consulting, Tech Lead.
Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Security,
Spring Transaction, Spring MVC,
Log4j, REST/SOAP WEB-SERVICES.
Custom Healthcare Software for Managing Chronic Conditions and Remote Patient...Mind IT Systems
Healthcare providers often struggle with the complexities of chronic conditions and remote patient monitoring, as each patient requires personalized care and ongoing monitoring. Off-the-shelf solutions may not meet these diverse needs, leading to inefficiencies and gaps in care. It’s here, custom healthcare software offers a tailored solution, ensuring improved care and effectiveness.
Understanding Globus Data Transfers with NetSageGlobus
NetSage is an open privacy-aware network measurement, analysis, and visualization service designed to help end-users visualize and reason about large data transfers. NetSage traditionally has used a combination of passive measurements, including SNMP and flow data, as well as active measurements, mainly perfSONAR, to provide longitudinal network performance data visualization. It has been deployed by dozens of networks world wide, and is supported domestically by the Engagement and Performance Operations Center (EPOC), NSF #2328479. We have recently expanded the NetSage data sources to include logs for Globus data transfers, following the same privacy-preserving approach as for Flow data. Using the logs for the Texas Advanced Computing Center (TACC) as an example, this talk will walk through several different example use cases that NetSage can answer, including: Who is using Globus to share data with my institution, and what kind of performance are they able to achieve? How many transfers has Globus supported for us? Which sites are we sharing the most data with, and how is that changing over time? How is my site using Globus to move data internally, and what kind of performance do we see for those transfers? What percentage of data transfers at my institution used Globus, and how did the overall data transfer performance compare to the Globus users?
Paketo Buildpacks : la meilleure façon de construire des images OCI? DevopsDa...Anthony Dahanne
Les Buildpacks existent depuis plus de 10 ans ! D’abord, ils étaient utilisés pour détecter et construire une application avant de la déployer sur certains PaaS. Ensuite, nous avons pu créer des images Docker (OCI) avec leur dernière génération, les Cloud Native Buildpacks (CNCF en incubation). Sont-ils une bonne alternative au Dockerfile ? Que sont les buildpacks Paketo ? Quelles communautés les soutiennent et comment ?
Venez le découvrir lors de cette session ignite
Large Language Models and the End of ProgrammingMatt Welsh
Talk by Matt Welsh at Craft Conference 2024 on the impact that Large Language Models will have on the future of software development. In this talk, I discuss the ways in which LLMs will impact the software industry, from replacing human software developers with AI, to replacing conventional software with models that perform reasoning, computation, and problem-solving.
SOCRadar Research Team: Latest Activities of IntelBrokerSOCRadar
The European Union Agency for Law Enforcement Cooperation (Europol) has suffered an alleged data breach after a notorious threat actor claimed to have exfiltrated data from its systems. Infamous data leaker IntelBroker posted on the even more infamous BreachForums hacking forum, saying that Europol suffered a data breach this month.
The alleged breach affected Europol agencies CCSE, EC3, Europol Platform for Experts, Law Enforcement Forum, and SIRIUS. Infiltration of these entities can disrupt ongoing investigations and compromise sensitive intelligence shared among international law enforcement agencies.
However, this is neither the first nor the last activity of IntekBroker. We have compiled for you what happened in the last few days. To track such hacker activities on dark web sources like hacker forums, private Telegram channels, and other hidden platforms where cyber threats often originate, you can check SOCRadar’s Dark Web News.
Stay Informed on Threat Actors’ Activity on the Dark Web with SOCRadar!
Cyaniclab : Software Development Agency Portfolio.pdfCyanic lab
CyanicLab, an offshore custom software development company based in Sweden,India, Finland, is your go-to partner for startup development and innovative web design solutions. Our expert team specializes in crafting cutting-edge software tailored to meet the unique needs of startups and established enterprises alike. From conceptualization to execution, we offer comprehensive services including web and mobile app development, UI/UX design, and ongoing software maintenance. Ready to elevate your business? Contact CyanicLab today and let us propel your vision to success with our top-notch IT solutions.
Innovating Inference - Remote Triggering of Large Language Models on HPC Clus...Globus
Large Language Models (LLMs) are currently the center of attention in the tech world, particularly for their potential to advance research. In this presentation, we'll explore a straightforward and effective method for quickly initiating inference runs on supercomputers using the vLLM tool with Globus Compute, specifically on the Polaris system at ALCF. We'll begin by briefly discussing the popularity and applications of LLMs in various fields. Following this, we will introduce the vLLM tool, and explain how it integrates with Globus Compute to efficiently manage LLM operations on Polaris. Attendees will learn the practical aspects of setting up and remotely triggering LLMs from local machines, focusing on ease of use and efficiency. This talk is ideal for researchers and practitioners looking to leverage the power of LLMs in their work, offering a clear guide to harnessing supercomputing resources for quick and effective LLM inference.
TROUBLESHOOTING 9 TYPES OF OUTOFMEMORYERRORTier1 app
Even though at surface level ‘java.lang.OutOfMemoryError’ appears as one single error; underlyingly there are 9 types of OutOfMemoryError. Each type of OutOfMemoryError has different causes, diagnosis approaches and solutions. This session equips you with the knowledge, tools, and techniques needed to troubleshoot and conquer OutOfMemoryError in all its forms, ensuring smoother, more efficient Java applications.
Providing Globus Services to Users of JASMIN for Environmental Data AnalysisGlobus
JASMIN is the UK’s high-performance data analysis platform for environmental science, operated by STFC on behalf of the UK Natural Environment Research Council (NERC). In addition to its role in hosting the CEDA Archive (NERC’s long-term repository for climate, atmospheric science & Earth observation data in the UK), JASMIN provides a collaborative platform to a community of around 2,000 scientists in the UK and beyond, providing nearly 400 environmental science projects with working space, compute resources and tools to facilitate their work. High-performance data transfer into and out of JASMIN has always been a key feature, with many scientists bringing model outputs from supercomputers elsewhere in the UK, to analyse against observational or other model data in the CEDA Archive. A growing number of JASMIN users are now realising the benefits of using the Globus service to provide reliable and efficient data movement and other tasks in this and other contexts. Further use cases involve long-distance (intercontinental) transfers to and from JASMIN, and collecting results from a mobile atmospheric radar system, pushing data to JASMIN via a lightweight Globus deployment. We provide details of how Globus fits into our current infrastructure, our experience of the recent migration to GCSv5.4, and of our interest in developing use of the wider ecosystem of Globus services for the benefit of our user community.
In software engineering, the right architecture is essential for robust, scalable platforms. Wix has undergone a pivotal shift from event sourcing to a CRUD-based model for its microservices. This talk will chart the course of this pivotal journey.
Event sourcing, which records state changes as immutable events, provided robust auditing and "time travel" debugging for Wix Stores' microservices. Despite its benefits, the complexity it introduced in state management slowed development. Wix responded by adopting a simpler, unified CRUD model. This talk will explore the challenges of event sourcing and the advantages of Wix's new "CRUD on steroids" approach, which streamlines API integration and domain event management while preserving data integrity and system resilience.
Participants will gain valuable insights into Wix's strategies for ensuring atomicity in database updates and event production, as well as caching, materialization, and performance optimization techniques within a distributed system.
Join us to discover how Wix has mastered the art of balancing simplicity and extensibility, and learn how the re-adoption of the modest CRUD has turbocharged their development velocity, resilience, and scalability in a high-growth environment.
Spring Data and In-Memory Data Management in Action
1. Spring Data and In-Memory
Data Management in Action
John Blum • Luke Shannon
@john_blum • @lukewshannon
2. Agenda
• Brief Apache Geode Overview
• Apache Geode in Spring
• Spring with Apache Geode in Action:
• Caching in [Pivotal] CloudFoundry
• Events with Continuous Query
• Data Access with Repositories
• New Improvements
• Roadmap
• *Bonus Feature
• QA
3
5. Spring Data GemFire / Geode
7
1. Use the Spring Framework’s powerful, non-invasive programming model
and concepts to simplify the configuration and development of highly-
scalable, distributed Spring applications based on Pivotal GemFire or
Apache Geode.
1. Integration with the Spring portfolio…
Purpose
6. Spring Data GemFire / Geode
8
+ +
Using Pivotal GemFire or Apache Geode
With Spring’s Transaction Management and Spring Data GemFire/Geode
To support both Local (Cache) & Global (JTA) Transactions
Spring Framework
7. Spring Data GemFire / Geode
9
+ +
Using Pivotal GemFire or Apache Geode
With Spring’s Cache Abstraction and Spring Data GemFire/Geode
To serve as a JCache (JSR-107) caching provider
Spring Framework
8. Spring Data GemFire / Geode
10
Spring Data Commons
+
Using Pivotal GemFire or Apache Geode
With SD Commons Repository Infrastructure and SD GemFire/Geode
To rapidly build Data Access Objects (CRUD + Mapping + Querying)
|
9. Spring Data GemFire / Geode
11
Spring Data REST & Spring HATEOAS
+
Using Pivotal GemFire or Apache Geode
With SDC Repositories, Spring Data REST and Spring HATEAOS
To easily create mature, hypermedia-driven REST web services
| +
10. Spring Data GemFire
12
Spring Integration
+
Using Pivotal GemFire with Spring Integration
For Inbound (Cache Events & CQ) and Outbound Channel Adapters,
Message Store, Lock Registry, Metadata Store
11. Spring Data GemFire
13
Spring XD
+
Using Pivotal GemFire with Spring XD (CQ, source & sink)
To simplify Big Data applications
Performing Ingest, Analytics, Batch Jobs and Data Export
12. Spring Data GemFire
14
Spring Session
+
Using Pivotal GemFire with Spring Session (Data GemFire)
To simplify (HTTP) Session State Management
+
13. Spring Data GemFire
15
Spring Boot
+
Using Pivotal GemFire with Spring Boot
To get up and running as quickly as possible
+
14. Spring Data GemFire
16
Spring Cloud
+
Use Spring Cloud (Connectors)
To build Cloud Native Applications and Microservices
Using Pivotal GemFire
16. Spring Data GemFire / Geode
18
Spring Data GemFire
<dependency>
<groupId>org.springframework.data</groupId>
<artifactId>spring-data-gemfire</artifactId>
<version>1.8.2.RELEASE</version>
</dependency>
Spring Data Geode
<dependency>
<groupId>org.springframework.data</groupId>
<artifactId>spring-data-geode</artifactId>
<version>1.0.0.APACHE-GEODE-INCUBATING-M2</version>
</dependency>
Pivotal
GemFire
Maven Artifacts
http://search.maven.org/#artifactdetails%7Corg.springframework.data%7Cspring-data-
gemfire%7C1.8.2.RELEASE%7Cjar
http://search.maven.org/#artifactdetails%7Corg.springframework.data%7Cspring-data-
geode%7C1.0.0.APACHE-GEODE-INCUBATING-M2%7Cjar
17. Spring Data GemFire / Geode
19
Spring Data GemFire
<dependency>
<groupId>org.springframework.boot</groupId>
<artifactId>spring-boot-starter-data-gemfire</artifactId>
<version>1.4.0.RELEASE</version>
</dependency>
Pivotal
GemFire
Spring Boot Maven Artifact
http://search.maven.org/#artifactdetails%7Corg.springframework.data%7Cspring-data-
gemfire%7C1.8.2.RELEASE%7Cjar
18. Roadmap – Annotation Support for Callbacks
20
@Component
public class ExampleApplicationCallbacks {
@CacheListener(region = “EventSource”)
public void afterCreate(EntryEvent<K, V> event) {
..
}
@CacheLoader(region = “DataSink”)
public void load(LoaderHelper<K, V> helper) {
..
}
@CacheWriter(region = “DataSource”)
public void beforeUpdate(EntryEvent<K, V> event) {
..
}
}
19. Spring Data GemFire / Geode
Spring Data GemFire 2.0
Java 8
Spring Framework 5.0
GemFire 9.0
Projections; Java 8 Types (Optional<T>, Stream<T>); Reactive Support (?)
Codebase re-organization
Test Framework support
21
Just around the corner…
20. Spring Boot
22
PR #6224 – Auto-configuration support for SDG Repositories
interface CustomerRepository extends Repository<Customer, Long> {
}
@Region(“Customers”)
class Customer {
}
compile “org.springframework.boot:spring-boot-starter-data-gemfire:1.5.0.M1”
No longer necessary…
@SpringBootApplication
@EnableGemfireRepositories
class ExampleSpringBootApplication {
}
21. Spring Boot
23
PR #5445 – Spring Boot Starter Data Geode
<dependency>
<groupId>org.springframework.boot</groupId>
<artifactId>spring-boot-starter-data-geode</artifactId>
<version>???</version>
</dependency>`
22. Spring Session
24
PR #366 – Spring Session Data Geode
<dependency>
<groupId>org.springframework.session</groupId>
<artifactId>spring-session-data-geode</artifactId>
<version>???</version>
</dependency>`
23. References
25
Spring Data GemFire (SDG) project page
SDG GitHub source code
SDG Reference Guide
SDG API
SDG Wiki page
SDG Examples
SDG JIRA Issue Tracking
SDG StackOverflow spring-data-gemfire tag
spring-data Gitter IM Channel
24. Learn More. Stay Connected.
Oliver Gierke
Advanced Spring Data REST
11:00 – 12:15 PM
@springcentral
spring.io/blog
@pivotal
pivotal.io/blog
@pivotalcf
http://engineering.pivotal.io
27. Safe Harbor Statement
• The following is intended to outline the general direction of Pivotal's offerings. It
is intended for information purposes only and may not be incorporated into any
contract. Any information regarding pre-release of Pivotal offerings, future
updates or other planned modifications is subject to ongoing evaluation by
Pivotal and is subject to change. This information is provided without warranty
or any kind, express or implied, and is not a commitment to deliver any material,
code, or functionality, and should not be relied upon in making purchasing
decisions regarding Pivotal's offerings. These purchasing decisions should only
be based on features currently available. The development, release, and timing
of any features or functionality described for Pivotal's offerings in this
presentation remain at the sole discretion of Pivotal. Pivotal has no obligation to
update forward looking information in this presentation.
29