This document provides an overview of Riak, an open source distributed database. It discusses Riak's key features like fault tolerance, horizontal scalability, and high availability. It also summarizes Riak's data model, interfaces, client libraries, ways to store and query data including CRUD, MapReduce, secondary indexing, and search. The document outlines Riak's architecture including consistent hashing, virtual nodes, vector clocks, and append-only storage. It previews upcoming Riak 1.4 features and discusses commercial Riak products and the growing Riak community.
This slide deck is the part of the talk, generally centered around the topics and details of the Riak Architecture & related material. It currently doesn't have the Azure sample commands or other elements around that, as it is the live part of the presentation. I'll likely add these parts in the future though.
This document summarizes the agenda for the 3rd Erlang Distributed Systems Workshop in Tokyo, including presentation titles, presenters, and times. There will be 5 presentations on distributed systems topics like deploying Kai and understanding RabbitMQ internals. The agenda also includes a break and a lightning talk. The workshop will conclude with a party at 21:00.
This document summarizes Tom Wilkie's presentation on Acunu & OCaml. It discusses how Acunu has evolved from small databases in 1990 to distributed, shared-nothing databases today. It presents the architecture and components of Acunu's storage core, which is built using OCaml. Performance results are shown for Acunu's prototypes using different data structures like doubling arrays, demonstrating high insertion rates.
AMD's new Bobcat core architecture is a low power x86 core designed for small die area and optimized for cloud clients. It features dual x86 decoders, out-of-order execution, 32KB L1 caches, a 512KB L2 cache, and advanced power reduction techniques to target sub-one watt operation. The goal of Bobcat is to provide 90% of the performance of mainstream notebook CPUs while using half the die area.
This document provides an overview and configuration details for TarPersistenceManager (TarPM) and clustering with TarPM in Adobe Experience Manager. TarPM is a disk-based persistence manager that uses standard tar files for efficient append-only operations. The document discusses TarPM functionality, configuration, optimization, hot backup, and migration. It also covers clustering architecture with a master node, global data store, cluster setup and configuration for both the repository and workspace levels.
The document discusses caching and new features in Ehcache 2 and Hibernate caching. It describes reasons for caching like offloading resources, improving performance, and enabling scalability. It discusses how caching works by leveraging principles of locality of reference and Pareto distributions. It also covers challenges like data size, staleness, and maintaining coherency in a clustered environment.
This slide deck is the part of the talk, generally centered around the topics and details of the Riak Architecture & related material. It currently doesn't have the Azure sample commands or other elements around that, as it is the live part of the presentation. I'll likely add these parts in the future though.
This document summarizes the agenda for the 3rd Erlang Distributed Systems Workshop in Tokyo, including presentation titles, presenters, and times. There will be 5 presentations on distributed systems topics like deploying Kai and understanding RabbitMQ internals. The agenda also includes a break and a lightning talk. The workshop will conclude with a party at 21:00.
This document summarizes Tom Wilkie's presentation on Acunu & OCaml. It discusses how Acunu has evolved from small databases in 1990 to distributed, shared-nothing databases today. It presents the architecture and components of Acunu's storage core, which is built using OCaml. Performance results are shown for Acunu's prototypes using different data structures like doubling arrays, demonstrating high insertion rates.
AMD's new Bobcat core architecture is a low power x86 core designed for small die area and optimized for cloud clients. It features dual x86 decoders, out-of-order execution, 32KB L1 caches, a 512KB L2 cache, and advanced power reduction techniques to target sub-one watt operation. The goal of Bobcat is to provide 90% of the performance of mainstream notebook CPUs while using half the die area.
This document provides an overview and configuration details for TarPersistenceManager (TarPM) and clustering with TarPM in Adobe Experience Manager. TarPM is a disk-based persistence manager that uses standard tar files for efficient append-only operations. The document discusses TarPM functionality, configuration, optimization, hot backup, and migration. It also covers clustering architecture with a master node, global data store, cluster setup and configuration for both the repository and workspace levels.
The document discusses caching and new features in Ehcache 2 and Hibernate caching. It describes reasons for caching like offloading resources, improving performance, and enabling scalability. It discusses how caching works by leveraging principles of locality of reference and Pareto distributions. It also covers challenges like data size, staleness, and maintaining coherency in a clustered environment.
The document discusses network intrusion detection and anomaly detection from a research perspective. It describes using network processors to develop a device that can perform high-speed packet capturing, timestamping, and processing. The device is used to build a traffic measurements system that can analyze traffic at wire speed and online to accurately characterize network traffic.
This document summarizes new features in Java SE 7 and Java EE 6. For Java SE 7, it outlines language changes like annotations on types and Project Coin small changes, as well as core changes like modularity support and concurrency updates. It describes VM changes like compressed pointers and garbage collection improvements. For Java EE 6, it discusses goals like rightsizing and extensibility, and new technologies like CDI and Bean Validation. It provides overviews of updated technologies and how profiles target specific capabilities.
Slides of my "Rapid JCR applications development with Sling" at ApacheCon EU 2009. Starts like the US 2008 version but uses a different example for the second part.
Ubuntu in the cloud What's Coming - Nick Barcet, CanonicalChris Purrington
This presentation discusses upcoming features in Ubuntu for cloud computing, including Ubuntu Oneiric Ocelot (11.10) which will be based on OpenStack and support KVM, Xen, and LXC hypervisors. It introduces Ubuntu Orchestra for bare metal provisioning and Ensemble for service-focused infrastructure management through formulas that define dependencies between components. The presentation encourages participation in the open development of these projects on IRC, Launchpad, and through the project websites.
Notes on a High-Performance JSON ProtocolDaniel Austin
This is my presentation from JSConf 2011. I am proposing a new Web protocol to improve performance across the Internet. It's based on a dual-band protocol layered over TCP/IP and UDP and is backward compatible with existing HTTP-based systems.
The document discusses network packet capture in the Linux kernel. It provides an overview of the Linux network stack and packet ingress flow. It describes important data structures like net devices and sk buffers. It also explains methods for capturing packets, including capturing packets during the packet ingress flow through the network stack layers and protocols.
Cloudera Impala - HUG Karlsruhe, July 04, 2013Alexander Alten
Low latency data processing with Impala
Impala provides fast, interactive SQL queries directly on your Apache Hadoop data stored in HDFS or HBase. In addition to using the same unified storage platform, Impala also uses the same metadata, SQL syntax (Hive SQL), JDBC driver and user interface (Hue Beeswax) as Apache Hive. This provides a familiar and unified platform for batch-oriented or real-time queries.
The document describes a presentation on Oracle Performance Measurement and Tuning with Solaris DTrace. It discusses using DTrace and BTrace to dynamically instrument applications like WebLogic Server. A key benefit is reducing the "observer effect" by using low overhead tools that allow probes to be enabled and disabled dynamically.
CloudFoundry and MongoDb, a marriage made in heavenPatrick Chanezon
This talk will provide an overview of the PaaS (Platform as a Service) landscape, and will describe the Cloud Foundry open source PaaS, with its multi-framework, multi-service, multi-cloud model. Cloud Foundry allows developers to provision apps in Java/Spring, Ruby/Rails, Ruby/Sinatra, Javascript/Node, and leverage services like MySQL, MongoDB, Reddis, Postgres and RabbitMQ. It can be used as a public PaaS on CloudFoundry.com and other service providers (ActiveState, AppFog), to create your own private cloud, or on your laptop using the Micro Cloud Foundry VM. Micro Cloud Foundry is a very easy way for developers to start working on their application using their framework of choice and MongoDB, without the need to setup a development environment, and your app is one command line away (vmc push) from deployment to cloudfoundry.com.
This document summarizes benchmarks for accessing data stored in different formats in a data lake. Scanning 1.2GB of text data stored in JSON, CSV, or Parquet formats was fastest when stored as Parquet, at 8.6 seconds. Retrieving a single row was also fastest from Parquet, at 1.75 seconds, because the Parquet format allows limiting the data read through predicate pushdown. The document concludes that Parquet is the best format even for row-oriented data when accessed through a distributed system like a data lake.
The presenter studied computer science and has over 5 years of experience with .NET and C#. He currently works for Infusion Development and has contributed to dotNetRDF for 1.5 years.
DotNetRDF is an open source .NET library for working with RDF data. It was started in 2009 as part of a PhD project and provides APIs for loading, querying, and updating RDF graphs from triple stores and serializations. The core concepts include nodes, triples, and graphs. It also supports SPARQL queries and triple store access.
The presenter discussed ongoing work to improve standards compliance and add features like OWL reasoning and SPIN. He has contributed APIs for fluent S
Symfony2 and MongoDB - MidwestPHP 2013 Pablo Godel
In this talk we will see how to use MongoDB in Symfony2 projects to speed up the development of web applications. We will give an introduction of MongoDB as a NoSQL database server and look at the options on how to work with it from Symfony2 and PHP applications.
Breaking The Clustering Limits @ AlphaCSP JavaEdge 2007Baruch Sadogursky
The document discusses the evolution of clustering in Java applications. It describes how early clustering solutions focused on replicating state between nodes for scalability and failover. More modern approaches use data grids, computational grids, and clustered virtual machines to distribute data and processing across nodes for improved performance and resource utilization. A variety of open source and commercial implementations are presented, including EHCache, GlassFish Shoal, Oracle Coherence, JBoss POJO Cache, GigaSpaces, and Terracotta.
This document discusses NoSQL databases and provides an overview of several popular NoSQL database types. It explains that NoSQL databases offer flexible schemas, join-less querying, and horizontal scalability. It also discusses the CAP theorem and how NoSQL databases achieve different levels of consistency and availability. Finally, it provides brief descriptions of several NoSQL databases like MongoDB, CouchDB, Redis, and Neo4j and their typical use cases.
Heroku is a platform as a service that allows developers to build, run, and operate applications entirely in the cloud. It supports Ruby, Java, Node.js, and other languages and frameworks. Heroku builds applications on Amazon Web Services infrastructure and provides automatic scaling of dynos (the lightweight Linux containers that run applications). Developers deploy code to Heroku using Git and the platform automatically distributes and runs the application across dynos.
This document provides an overview and update on Amazon Aurora, Amazon's relational database service. It discusses new performance enhancements including improved read performance through caching, NUMA-aware scheduling, and lock compression to reduce contention. New availability features are also summarized, such as automatic repair and replacement of failed database nodes and storage volumes that can grow to 64TB. The document outlines Aurora's architecture advantages over traditional databases for scaling in the cloud through its distributed, self-healing design.
It’s been an exciting year for Amazon Aurora, the MySQL-compatible relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. In this deep dive session, we’ll discuss best practices and explore new features, include high availability options and new integrations with AWS services. We’ll also discuss the recently-announced Aurora with PostgreSQL compatibility.
NoSQL refers to non-relational databases that were developed to handle large volumes of data across many servers. NoSQL databases such as Redis, Memcached, MongoDB and CouchDB do not use SQL for queries and are more scalable and flexible than traditional relational databases. They often sacrifice consistency, complex queries, and transaction support in favor of availability and flexibility.
Riak Use Cases : Dissecting The Solutions To Hard ProblemsAndy Gross
This document summarizes a presentation on using Riak to solve hard distributed systems problems. It discusses Riak's key features and tradeoffs in consistency and data model. It then provides examples of how companies like Comcast, Yammer, and Github use Riak for high availability and low latency applications like user metadata storage, notifications, and as a document store. The presentation emphasizes that choosing the right database requires understanding your problem and tradeoffs in consistency, availability and data model. It also highlights how Riak and Erlang help enable high availability, low latency and operational simplicity through features like Bitcask storage, fault tolerance and lightweight threading.
The document discusses considerations for building a private cloud using OpenStack Folsom. It covers topics such as the definition of a private cloud, sizing instances and flavors, network architecture including multiple networks, image storage and performance, and architecture examples for different sizes of private clouds. The document provides guidance on capacity planning, performance bottlenecks, and best practices for building a private cloud with OpenStack.
Ingesting Over Four Million Rows Per Second With QuestDB Timeseries Database ...javier ramirez
How would you build a database to support sustained ingestion of several hundreds of thousands rows per second while running near real-time queries on top?
In this session I will go over some of the technical decisions and trade-offs we applied when building QuestDB, an open source time-series database developed mainly in JAVA, and how we can achieve over four million row writes per second on a single instance without blocking or slowing down the reads. There will be code and demos, of course.
We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
The document discusses network intrusion detection and anomaly detection from a research perspective. It describes using network processors to develop a device that can perform high-speed packet capturing, timestamping, and processing. The device is used to build a traffic measurements system that can analyze traffic at wire speed and online to accurately characterize network traffic.
This document summarizes new features in Java SE 7 and Java EE 6. For Java SE 7, it outlines language changes like annotations on types and Project Coin small changes, as well as core changes like modularity support and concurrency updates. It describes VM changes like compressed pointers and garbage collection improvements. For Java EE 6, it discusses goals like rightsizing and extensibility, and new technologies like CDI and Bean Validation. It provides overviews of updated technologies and how profiles target specific capabilities.
Slides of my "Rapid JCR applications development with Sling" at ApacheCon EU 2009. Starts like the US 2008 version but uses a different example for the second part.
Ubuntu in the cloud What's Coming - Nick Barcet, CanonicalChris Purrington
This presentation discusses upcoming features in Ubuntu for cloud computing, including Ubuntu Oneiric Ocelot (11.10) which will be based on OpenStack and support KVM, Xen, and LXC hypervisors. It introduces Ubuntu Orchestra for bare metal provisioning and Ensemble for service-focused infrastructure management through formulas that define dependencies between components. The presentation encourages participation in the open development of these projects on IRC, Launchpad, and through the project websites.
Notes on a High-Performance JSON ProtocolDaniel Austin
This is my presentation from JSConf 2011. I am proposing a new Web protocol to improve performance across the Internet. It's based on a dual-band protocol layered over TCP/IP and UDP and is backward compatible with existing HTTP-based systems.
The document discusses network packet capture in the Linux kernel. It provides an overview of the Linux network stack and packet ingress flow. It describes important data structures like net devices and sk buffers. It also explains methods for capturing packets, including capturing packets during the packet ingress flow through the network stack layers and protocols.
Cloudera Impala - HUG Karlsruhe, July 04, 2013Alexander Alten
Low latency data processing with Impala
Impala provides fast, interactive SQL queries directly on your Apache Hadoop data stored in HDFS or HBase. In addition to using the same unified storage platform, Impala also uses the same metadata, SQL syntax (Hive SQL), JDBC driver and user interface (Hue Beeswax) as Apache Hive. This provides a familiar and unified platform for batch-oriented or real-time queries.
The document describes a presentation on Oracle Performance Measurement and Tuning with Solaris DTrace. It discusses using DTrace and BTrace to dynamically instrument applications like WebLogic Server. A key benefit is reducing the "observer effect" by using low overhead tools that allow probes to be enabled and disabled dynamically.
CloudFoundry and MongoDb, a marriage made in heavenPatrick Chanezon
This talk will provide an overview of the PaaS (Platform as a Service) landscape, and will describe the Cloud Foundry open source PaaS, with its multi-framework, multi-service, multi-cloud model. Cloud Foundry allows developers to provision apps in Java/Spring, Ruby/Rails, Ruby/Sinatra, Javascript/Node, and leverage services like MySQL, MongoDB, Reddis, Postgres and RabbitMQ. It can be used as a public PaaS on CloudFoundry.com and other service providers (ActiveState, AppFog), to create your own private cloud, or on your laptop using the Micro Cloud Foundry VM. Micro Cloud Foundry is a very easy way for developers to start working on their application using their framework of choice and MongoDB, without the need to setup a development environment, and your app is one command line away (vmc push) from deployment to cloudfoundry.com.
This document summarizes benchmarks for accessing data stored in different formats in a data lake. Scanning 1.2GB of text data stored in JSON, CSV, or Parquet formats was fastest when stored as Parquet, at 8.6 seconds. Retrieving a single row was also fastest from Parquet, at 1.75 seconds, because the Parquet format allows limiting the data read through predicate pushdown. The document concludes that Parquet is the best format even for row-oriented data when accessed through a distributed system like a data lake.
The presenter studied computer science and has over 5 years of experience with .NET and C#. He currently works for Infusion Development and has contributed to dotNetRDF for 1.5 years.
DotNetRDF is an open source .NET library for working with RDF data. It was started in 2009 as part of a PhD project and provides APIs for loading, querying, and updating RDF graphs from triple stores and serializations. The core concepts include nodes, triples, and graphs. It also supports SPARQL queries and triple store access.
The presenter discussed ongoing work to improve standards compliance and add features like OWL reasoning and SPIN. He has contributed APIs for fluent S
Symfony2 and MongoDB - MidwestPHP 2013 Pablo Godel
In this talk we will see how to use MongoDB in Symfony2 projects to speed up the development of web applications. We will give an introduction of MongoDB as a NoSQL database server and look at the options on how to work with it from Symfony2 and PHP applications.
Breaking The Clustering Limits @ AlphaCSP JavaEdge 2007Baruch Sadogursky
The document discusses the evolution of clustering in Java applications. It describes how early clustering solutions focused on replicating state between nodes for scalability and failover. More modern approaches use data grids, computational grids, and clustered virtual machines to distribute data and processing across nodes for improved performance and resource utilization. A variety of open source and commercial implementations are presented, including EHCache, GlassFish Shoal, Oracle Coherence, JBoss POJO Cache, GigaSpaces, and Terracotta.
This document discusses NoSQL databases and provides an overview of several popular NoSQL database types. It explains that NoSQL databases offer flexible schemas, join-less querying, and horizontal scalability. It also discusses the CAP theorem and how NoSQL databases achieve different levels of consistency and availability. Finally, it provides brief descriptions of several NoSQL databases like MongoDB, CouchDB, Redis, and Neo4j and their typical use cases.
Heroku is a platform as a service that allows developers to build, run, and operate applications entirely in the cloud. It supports Ruby, Java, Node.js, and other languages and frameworks. Heroku builds applications on Amazon Web Services infrastructure and provides automatic scaling of dynos (the lightweight Linux containers that run applications). Developers deploy code to Heroku using Git and the platform automatically distributes and runs the application across dynos.
This document provides an overview and update on Amazon Aurora, Amazon's relational database service. It discusses new performance enhancements including improved read performance through caching, NUMA-aware scheduling, and lock compression to reduce contention. New availability features are also summarized, such as automatic repair and replacement of failed database nodes and storage volumes that can grow to 64TB. The document outlines Aurora's architecture advantages over traditional databases for scaling in the cloud through its distributed, self-healing design.
It’s been an exciting year for Amazon Aurora, the MySQL-compatible relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. In this deep dive session, we’ll discuss best practices and explore new features, include high availability options and new integrations with AWS services. We’ll also discuss the recently-announced Aurora with PostgreSQL compatibility.
NoSQL refers to non-relational databases that were developed to handle large volumes of data across many servers. NoSQL databases such as Redis, Memcached, MongoDB and CouchDB do not use SQL for queries and are more scalable and flexible than traditional relational databases. They often sacrifice consistency, complex queries, and transaction support in favor of availability and flexibility.
Riak Use Cases : Dissecting The Solutions To Hard ProblemsAndy Gross
This document summarizes a presentation on using Riak to solve hard distributed systems problems. It discusses Riak's key features and tradeoffs in consistency and data model. It then provides examples of how companies like Comcast, Yammer, and Github use Riak for high availability and low latency applications like user metadata storage, notifications, and as a document store. The presentation emphasizes that choosing the right database requires understanding your problem and tradeoffs in consistency, availability and data model. It also highlights how Riak and Erlang help enable high availability, low latency and operational simplicity through features like Bitcask storage, fault tolerance and lightweight threading.
The document discusses considerations for building a private cloud using OpenStack Folsom. It covers topics such as the definition of a private cloud, sizing instances and flavors, network architecture including multiple networks, image storage and performance, and architecture examples for different sizes of private clouds. The document provides guidance on capacity planning, performance bottlenecks, and best practices for building a private cloud with OpenStack.
Ingesting Over Four Million Rows Per Second With QuestDB Timeseries Database ...javier ramirez
How would you build a database to support sustained ingestion of several hundreds of thousands rows per second while running near real-time queries on top?
In this session I will go over some of the technical decisions and trade-offs we applied when building QuestDB, an open source time-series database developed mainly in JAVA, and how we can achieve over four million row writes per second on a single instance without blocking or slowing down the reads. There will be code and demos, of course.
We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
Aurora is Amazon's cloud database that provides enterprise-grade capabilities at lower costs than traditional databases. It offers speed and availability through a distributed, fault-tolerant storage system and automatic scaling of storage and compute resources. Aurora provides cross-region replication for high availability and data locality. Engineering Aurora requires experience in databases, storage systems, and distributed systems.
Improving Your Heroku App Performance with Asset CDN and UnicornSimon Bagreev
This document summarizes tips for optimizing the performance of Rails applications using asset CDNs and the Unicorn web server. It discusses using Amazon S3 and CloudFront for caching and delivering assets to improve load times. It also explains how to configure Unicorn to handle requests concurrently across worker processes to better utilize dyno resources on Heroku. Benchmark tests show these approaches reduced load times and increased the number of concurrent requests applications can handle.
Porting a Streaming Pipeline from Scala to RustEvan Chan
How we at Conviva ported a streaming data pipeline in months from Scala to Rust. What are the important human and technical factors in our port, and what did we learn?
The document discusses priorities for continuous delivery including individuals and interactions over processes and tools, working software over documentation, and customer collaboration over contract negotiation. It emphasizes building, integrating, and maintaining speed through iterations. While tests and tools are mentioned, responding to change and communication are prioritized. The document concludes by returning to practicing continuous integration and delivery using tools like PaaS platforms and Kanban boards.
The document discusses cloud computing concepts like Platform as a Service (PaaS) and Infrastructure as a Service (IaaS). It mentions Amazon Web Services and Heroku as examples and references principles like continuous deployment, integration, and ideals like being agile and lean. The author discusses deploying to the cloud, building processes, and coding against the cloud. Continuous delivery is mentioned along with horizontally scaling and moving past limitations of relational databases and NoSQL.
Node.js rulz! JavaScript takes over the full StackAdron Hall
Node.js is a web server written in C/C++ that uses an event-driven and non-blocking I/O model. It allows for fast and easily scalable development using JavaScript and packages from NPM. The document discusses Node.js capabilities and recommends using Express.js, MongoDB, testing frameworks like QUnit and Nodeunit, and deploying applications to a Platform as a Service like Tier 3's Web Fabric for fast prototyping.
The document discusses challenges around deployments, testing, and prototyping for Visual Studio projects. It asks questions about the needs for deployments, how to speed up prototyping, and how to help with testing and quality assurance. Prerequisites and the use of ASP.NET MVC with the Razor view engine are also mentioned.
AWS Toolkit for Visual Studio - Version 1Adron Hall
The document discusses where to obtain various software components needed to develop .NET applications for Amazon Web Services (AWS). It outlines how to get the .NET Framework, Visual Studio, AWS SDK for .NET, and AWS deployment tools. It also provides basic verification steps to ensure the AWS toolkit is installed correctly and ready to use.
This document discusses cloud computing platforms and options for .NET development. It explores who created clouds, what they provide, and why developers care about them. It examines challenges with early cloud platforms and recommends AppHarbor as a good option for .NET developers that provides fast instance startup, testability, code quality practices and integration with CI/CD pipelines.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Letter and Document Automation for Bonterra Impact Management (fka Social Sol...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on automated letter generation for Bonterra Impact Management using Google Workspace or Microsoft 365.
Interested in deploying letter generation automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
This presentation provides valuable insights into effective cost-saving techniques on AWS. Learn how to optimize your AWS resources by rightsizing, increasing elasticity, picking the right storage class, and choosing the best pricing model. Additionally, discover essential governance mechanisms to ensure continuous cost efficiency. Whether you are new to AWS or an experienced user, this presentation provides clear and practical tips to help you reduce your cloud costs and get the most out of your budget.
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
Nunit vs XUnit vs MSTest Differences Between These Unit Testing Frameworks.pdfflufftailshop
When it comes to unit testing in the .NET ecosystem, developers have a wide range of options available. Among the most popular choices are NUnit, XUnit, and MSTest. These unit testing frameworks provide essential tools and features to help ensure the quality and reliability of code. However, understanding the differences between these frameworks is crucial for selecting the most suitable one for your projects.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
6. Use, when & what to use Riak for...
Users/Profiles Logging Systems
Metadata Sensor Data
Session Storage Notification Systems
Object Storage Record Systems
Wednesday, March 13, 13
9. {“Key”:“Value”}
• Values are stored against keys
• Key/Value + Metadata = Object
• Fundamental Unit of Replication
• Any Datatype will work
• Record to disk in binary format
Wednesday, March 13, 13
10. <<BUCKET>>/<<KEY>>
• Virtual Namespace
• Bucket + Keys = Object Address
• Buckets have properties
• Objects in bucket inherit properties
• No relationships between buckets
Wednesday, March 13, 13
12. INTERFACES
HTTP API - Via a little piece of magic called Webmachine
Largely-faithful REST implementation
Protocol Buffers API - Thanks, Google!
Compact, binary protocol
Wednesday, March 13, 13
13. CLIENT LIBS
Python Java Clojure
Erlang PHP Haskell
...and more via Basho
Ruby C/C++ OCaml
or our community.
Perl .NET Scala
Dart Go Node.js
Wednesday, March 13, 13
14. RIAK GIVES YOU
[FOUR] WAYS TO STORE,
RETRIEVE, AND QUERY
DATA
Wednesday, March 13, 13
15. CRUD
// GET
1
2
3
4 GET /buckets/bucket/keys/key
5
6
// PUT
7
8
9
10 POST /buckets/bucket/keys/key // Riak-‐defined key
11
PUT /buckets/bucket/keys/key // User-‐defined key
12
13
// DELETE
14
15
16
17 DELETE /buckets/bucket/keys/key
18
Wednesday, March 13, 13
16. MapReduce
Distributed processing system using Riak Pipe
Efficient for targeted queries over known key range
Write jobs in Erlang or JS. (Erlang more performant)
Wednesday, March 13, 13
17. Secondary Indexing (2i)
riak_object X-Riak-Index-email_bin
“mark@basho.com”
riak_object X-Riak-Index-value_int
“42”
Tag objects with custom metadata on PUT...
Exact match and range queries...
No multi-index queries yet...
Pagination is on its way...
Wednesday, March 13, 13
18. Riak Search
Store and index documents (JSON, text, XML, etc)
Current Riak Search supports subset of Solr API
Next iteration (Yokozuna; in beta)will implement distributed
Solr on Riak. It will be sexy.
Looking for beta testers to help harden Yokozuna
Wednesday, March 13, 13
19. ARCHITECTURE
Consistent Hashing
The scaleability and
Virtual Nodes
ease of operation
goals inform Handoff/Rebalancing
architectural decisions.
Vector Clocks
These come with tradeoffs.
Append-only storage
Active Anti-Entropy*
Wednesday, March 13, 13
20. Consistent Hashing
Location of data in the Riak ring is determined based on hash of
bucket + key.
Provides even distribution of storage and query load
Trades off advantages gained from locality
- e.g. Range queries and aggregates
Wednesday, March 13, 13
22. Virtual Nodes
Unit of addressing and concurrency in Riak
Each physical host manage many vnodes
Partition count / physical machines = vnodes/machine*
Decouples physical assets from data distribution. This provides:
- simplicity in cluster sizing
- failure isolation
Wednesday, March 13, 13
23. Handoff/Rebalancing
Mechanisms for data rebalancing
When nodes join/leave cluster, handoff and rebalancing
manage the date shuffling dynamically
Trades off speed of convergence vs. effects on cluster
performance
- causes disk & network load
Wednesday, March 13, 13
24. Vector Clocks
VCs used to rectify object consistency at READ time.
Lots of knobs to turn; well-documented
Trades off space, speed, and complexity for safety
- will store all sibling objects until resolved
- can lead to object size issues
Wednesday, March 13, 13
25. Append-Only Storage
Riak provides a pluggable backend interface. (Write your own;
we’ll probably hire you...)
Bitcask, LevelDB are most-heavily used. Both are
append - only
Provides crash safety and speed.
Trade off: periodic compaction/merge ops
Wednesday, March 13, 13
26. RIAK 1.3
(AKA “new hotness”)
Active Anti Entropy Riaknostic included by default
MapReduce Improvements Riak Control improvements
IPv6 Support Much more
Full release notes: https://github.com/basho/riak/blob/1.3/RELEASE-NOTES.md
Wednesday, March 13, 13
27. FUTURE WORK*
(1.4 and beyond)
Dynamic Ring Size Consistency
Yokozuna 2i Improvements
CRDTs/Data Types Riak Pipe work
Riak Object Much more
(* all code subject to ship early, late, or not at all)
Wednesday, March 13, 13
28. Multi-tenant cloud storage software for
public and private clouds.
Designed to provide simple, available, distributed
cloud storage at any scale.
S3-API compatible and supports per-tenant
reporting for billing and metering use cases.
Additional APIs on the way.
Stores files of arbitrary size. Under the hood
stores 1MB chunks along side a manifest.
Stateless proxy (CS) does chunking. Riak does
distribution, storage, etc.
Wednesday, March 13, 13
29. Extends Riak's capabilities with:
- multi-datacenter replication
- SNMP Configuration
- JMX-Monitoring
- 24x7 support from Basho Engineers
One cluster acts as a "source cluster". The source
cluster replicates its data to one or more
"sink clusters" using either real-time or full sync.
Data transfer is unidirectional (source -> sink).
Bidirectional synchronization can be
achieved by configuring a pair of connections
between clusters.
Wednesday, March 13, 13
30. RIAK COMMUNITY
Mailing List - 1300 developers
IRC - 200+ people every day yelling about software
GitHub - 1000s of watchers; 200+ contributors to all projects
Meetups - 10 Countries, 23 Cities, 3700+ Members & growing fast!
Deployments - 1000s in production.
Wednesday, March 13, 13
31. ricon.io/east.html
May 13-14th in New York City
Talks, hacking, parties
Dedicated to the future of Riak and
distributed systems in production
REGISTER https://ricon-east-2013.eventbrite.com/?discount=lovevnodes
NOW!
Wednesday, March 13, 13
32. GETTING STARTED
Downloads - http://docs.basho.com/riak/latest/downloads/
Docs - http://docs.basho.com
Riak Source Code - github.com/basho/riak
All Basho source Code - github.com/basho/
Riak Mailing List - http://bit.ly/riak-list
Email or Tweet me @adron or adron@basho.com
Wednesday, March 13, 13
33. Let’s Talk UI & CLI - Demo Things
Wednesday, March 13, 13
34. #WHOIS
Adron Hall | @adron | Coder, Messenger, Recon
Wednesday, March 13, 13