Making an application horizontally scalable in 30 minutes. This presentation describes how a linear processing application (mail merge) can be converted into a horizontally scalable using Redis and provides some context why a multi-process approach is preferable to a multi-threaded approach.
Martin Tepper presented on using MongoDB as a queryable cache for Travel IQ, a meta search engine for flights and hotels. Travel IQ was experiencing slow API response times due to complex queries on its normalized SQL database. MongoDB was implemented as a caching layer to store denormalized offer data to allow for faster querying. This improved response times but also led to some headaches around data consistency and segmentation faults that were later addressed.
The workshop will present how to combine tools to quickly query, transform and model data using command line tools.
The goal is to show that command line tools are efficient at handling reasonable sizes of data and can accelerate the data science
process. We will show that in many instances, command line processing ends up being much faster than ‘big-data’ solutions. The content
of the workshop is derived from the book of the same name (http://datascienceatthecommandline.com/). In addition, we will cover
vowpal-wabbit (https://github.com/JohnLangford/vowpal_wabbit) as a versatile command line tool for modeling large datasets.
Build your first MongoDB App in Ruby @ StrangeLoop 2013Steven Francia
This document provides an agenda and introduction for building a first MongoDB application. It discusses MongoDB fundamentals like documents, schemas, and queries. It then walks through setting up a Sinatra application scaffolded with folders, connecting it to a MongoDB database, and pre-populating the database with venue data. The document demonstrates basic CRUD operations in the MongoDB shell and indexes the venue location data geographically. Finally, it discusses starting the Sinatra server and opening the initial application in the browser.
Redis is an in-memory database that supports key-value, hash, list, and set data structures. It can backup data to disk using RDB snapshots or AOF file appends. Redis supports master-slave replication where the master dumps data to slaves. It has high performance for common use cases like caches but has higher memory usage than databases. It can be used for caching, queues, and more.
The document discusses implementing a hybrid database solution using both MongoDB and MySQL. It describes storing less frequently changing and reference data like users and products in MongoDB for flexibility, while storing transactional data like orders and inventory counts in MySQL for ACID compliance. The system keeps the data in sync between the two databases using listeners that update MySQL whenever related data is created or changed in MongoDB.
DaNode is a simple and small web server written in D that allows running websites using any programming language as CGI scripts. It has fewer than 2000 lines of code, runs on Windows, Linux, and ARM, and allows serving multiple secure domains from a single server using SSL and SNI. While simple and useful for quick tasks, it does not fully implement web standards and its security has limitations, but it serves as a basis for further development and adding features from other frameworks like vibe.d.
This document discusses using Logstash to collect, parse, and store logs from multiple sources in Elasticsearch. It describes Logstash's three main components - inputs, filters, and outputs. Examples are provided for using Logstash with Lumberjack to ship logs, parsing logs with grok filters, and outputting to Elasticsearch. Instructions are included for installing, configuring, and running Logstash, Elasticsearch, Kibana, and Lumberjack to build a log management pipeline.
Martin Tepper presented on using MongoDB as a queryable cache for Travel IQ, a meta search engine for flights and hotels. Travel IQ was experiencing slow API response times due to complex queries on its normalized SQL database. MongoDB was implemented as a caching layer to store denormalized offer data to allow for faster querying. This improved response times but also led to some headaches around data consistency and segmentation faults that were later addressed.
The workshop will present how to combine tools to quickly query, transform and model data using command line tools.
The goal is to show that command line tools are efficient at handling reasonable sizes of data and can accelerate the data science
process. We will show that in many instances, command line processing ends up being much faster than ‘big-data’ solutions. The content
of the workshop is derived from the book of the same name (http://datascienceatthecommandline.com/). In addition, we will cover
vowpal-wabbit (https://github.com/JohnLangford/vowpal_wabbit) as a versatile command line tool for modeling large datasets.
Build your first MongoDB App in Ruby @ StrangeLoop 2013Steven Francia
This document provides an agenda and introduction for building a first MongoDB application. It discusses MongoDB fundamentals like documents, schemas, and queries. It then walks through setting up a Sinatra application scaffolded with folders, connecting it to a MongoDB database, and pre-populating the database with venue data. The document demonstrates basic CRUD operations in the MongoDB shell and indexes the venue location data geographically. Finally, it discusses starting the Sinatra server and opening the initial application in the browser.
Redis is an in-memory database that supports key-value, hash, list, and set data structures. It can backup data to disk using RDB snapshots or AOF file appends. Redis supports master-slave replication where the master dumps data to slaves. It has high performance for common use cases like caches but has higher memory usage than databases. It can be used for caching, queues, and more.
The document discusses implementing a hybrid database solution using both MongoDB and MySQL. It describes storing less frequently changing and reference data like users and products in MongoDB for flexibility, while storing transactional data like orders and inventory counts in MySQL for ACID compliance. The system keeps the data in sync between the two databases using listeners that update MySQL whenever related data is created or changed in MongoDB.
DaNode is a simple and small web server written in D that allows running websites using any programming language as CGI scripts. It has fewer than 2000 lines of code, runs on Windows, Linux, and ARM, and allows serving multiple secure domains from a single server using SSL and SNI. While simple and useful for quick tasks, it does not fully implement web standards and its security has limitations, but it serves as a basis for further development and adding features from other frameworks like vibe.d.
This document discusses using Logstash to collect, parse, and store logs from multiple sources in Elasticsearch. It describes Logstash's three main components - inputs, filters, and outputs. Examples are provided for using Logstash with Lumberjack to ship logs, parsing logs with grok filters, and outputting to Elasticsearch. Instructions are included for installing, configuring, and running Logstash, Elasticsearch, Kibana, and Lumberjack to build a log management pipeline.
Este documento describe la historia, representaciones y rituales de la diosa Hécate en la mitología griega. Explica que Hécate era una diosa de la luna, la magia y las encrucijadas que se representaba con tres cabezas. Su ritual más conocido es el de los fuegos sagrados, que involucra encender 21 velas en un círculo e invocar a Hécate para pedir su sabiduría y guía.
Antrian adalah struktur data yang menyimpan elemen sesuai urutan masuk (FIFO). Terdapat beberapa metode utama pada antrian seperti enqueue untuk menambahkan elemen, dequeue untuk mengambil elemen pertama, dan peek untuk melihat elemen pertama tanpa menghapusnya. Antrian dapat diimplementasikan menggunakan array dengan menyimpan indeks elemen terakhir.
Algoritma dan pengetahuan terkait (menghitung, konversi, dll) Fazar Ikhwan Guntara
Dokumen tersebut membahas cara mengkonversi waktu dalam detik menjadi jam, menit dan detik dengan menggunakan konsep pembagian dan sisa bagi. Langkah-langkah perhitungannya dijelaskan secara rinci beserta ilustrasinya dalam bentuk skema alir.
1. Bab 5 membahas struktur data queue (antrian) dan implementasinya dalam bahasa pemrograman.
2. Queue adalah struktur data linear dimana penambahan elemen hanya bisa dilakukan di satu ujung dan penghapusan di ujung lain.
3. Queue dapat diimplementasikan menggunakan array linear atau linked list dengan operasi enqueue dan dequeue.
Makalah ini membahas tentang implementasi queue dengan bahasa pemrograman Pascal. Queue merupakan struktur data yang mengimplementasikan prinsip antrian First In First Out (FIFO). Makalah ini menjelaskan definisi dan gambaran umum queue, macam-macam queue, representasi queue secara statis menggunakan array dan representasi secara dinamis menggunakan linked list tunggal dan ganda. Juga dibahas queue berprioritas beserta contoh kode program untuk masing-masing implementasi queue.
Queue adalah linear list dimana data dimasukkan melalui rear dan dihapus dari front, mengikuti prinsip first in first out. Queue memiliki operasi seperti enqueue untuk memasukkan data ke rear, dequeue untuk menghapus data di front, serta mengakses elemen di front dan rear tanpa menghapusnya.
Laporan ini membahas tentang sistem antrian di Bread Talk Bakery di Surabaya. Studi kasus menganalisis model sistem antrian, rata-rata waktu antar kedatangan pelanggan dan pelayanan kasir, serta karakteristik lainnya seperti utilitas sistem dan rata-rata jumlah pelanggan dalam sistem."
Queue is a first-in first-out (FIFO) data structure where elements can only be added to the rear of the queue and removed from the front of the queue. It has two pointers - a front pointer pointing to the front element and a rear pointer pointing to the rear element. Queues can be implemented using arrays or linked lists. Common queue operations include initialization, checking if empty/full, enqueue to add an element, and dequeue to remove an element. The document then describes how these operations work for queues implemented using arrays, linked lists, and circular arrays. It concludes by providing exercises to implement specific queue tasks.
The document discusses different types of queues, including their definitions, properties, and implementations. It defines a queue as a linear data structure with two ends - one for adding elements and one for removing them, following a FIFO (first-in, first-out) approach. Key points covered include common queue operations like insertion and removal; circular queues which wrap elements around to avoid overflow; priority queues which order elements by priority; and deques which allow additions and removals from both ends.
Conexión de MongoDB con Hadoop - Luis Alberto Giménez - CAPSiDE #DevOSSAzureDaysCAPSiDE
This document discusses using MongoDB and Hadoop together. It provides an overview of MongoDB and Hadoop, describes how the MongoDB Hadoop Connector allows them to interoperate, and gives an example of building a graph of email sender-recipient relationships from Enron email data stored in MongoDB using Hadoop MapReduce, streaming, Pig, and Hive. The connector allows parallel processing of MongoDB data using Hadoop and integration with the Hadoop ecosystem.
Barcelona MUG MongoDB + Hadoop PresentationNorberto Leite
- The document discusses MongoDB and Hadoop, two popular big data platforms, and the MongoDB + Hadoop Connector which allows interoperation between the two.
- It provides an overview of MongoDB and Hadoop's key features for scalability, availability and processing large datasets.
- The connector allows processing data across MongoDB and Hadoop through MapReduce jobs without needing custom exports/imports.
- Examples show building a graph of email sender/recipient relationships from an Enron dataset stored in MongoDB using Hadoop Streaming, Pig and Hive.
Midwest php 2013 deploying php on paas- why & howdotCloud
Deploying PHP applications to Platform as a Service (PaaS) can provide several benefits over traditional hosting methods. PaaS allows developers to quickly deploy new environments for testing code changes. It also handles tasks like optimizing stacks, upgrading software, and providing comprehensive routing. PaaS aims to make deployment as simple as uploading code and eliminates the need to manually configure servers. While there is an initial learning curve to using PaaS tools and reworking some applications, it can improve the development to production workflow and allow applications to easily scale on demand.
Grunt is a JavaScript task runner that can automate front-end development tasks like linting, compiling Sass files to CSS, running tests and watching for file changes. It provides plugins for common tasks and allows developers to define custom workflows. The document discusses how Grunt can be used to integrate and run various tasks like linting, Sass compilation and watching files. It also explains how to define tasks, configure Grunt and run tasks from the command line. Custom tasks can be created and shared as plugins. The document provides an example of using Grunt to convert Markdown files to the Leanpub format and sync them to Dropbox.
The document summarizes PHP deployment on Platform as a Service (PaaS) and why it can help. It discusses how PHP deployment worked in the past with manual setups that were error-prone. PaaS allows fast deployment of new environments and leveraging of version control. It handles optimizations, upgrades, routing, and scaling automatically. The document provides an example of deploying a Symfony application to dotCloud PaaS, including configuring databases. Potential drawbacks are an initial learning curve and apps requiring reworking, but PaaS offers benefits like reduced costs and improved reliability.
Ready to leverage the power of a graph database to bring your application to the next level, but all the data is still stuck in a legacy relational database?
Fortunately, Neo4j offers several ways to quickly and efficiently import relational data into a suitable graph model. It's as simple as exporting the subset of the data you want to import and ingest it either with an initial loader in seconds or minutes or apply Cypher's power to put your relational data transactionally in the right places of your graph model.
In this webinar, Michael will also demonstrate a simple tool that can load relational data directly into Neo4j, automatically transforming it into a graph representation of your normalized entity-relationship model.
Automated Reports with Rstudio Server
Automated KPI reporting with Shiny Server
Process Validation Documentation with Jupyter Notebook
Automated Machine Learning with Dataiku
This document provides a summary of Hadoop and its ecosystem components. It begins with introducing the speaker, Andrew Brust, and his background. It then provides an overview of the key Hadoop components like MapReduce, HDFS, HBase, Hive, Pig, Sqoop and Flume. The document demonstrates several of these components and discusses how to use Microsoft Hadoop on Azure, Amazon EMR and Cloudera CDH virtual machines. It also summarizes the Mahout machine learning library and recommends a commercial product for visualizing and applying Mahout outputs.
This document provides an introduction to Node.js. It discusses that Node.js is an event-driven, non-blocking I/O platform for building scalable network applications using JavaScript. It was created to address issues with traditional blocking I/O by using asynchronous programming. The document outlines benefits of Node.js like using JavaScript for server-side applications, non-blocking I/O, a large module ecosystem, and an active community. It also provides examples of core modules, writing simple modules, and creating an HTTP server in Node.js.
introduction to data processing using Hadoop and PigRicardo Varela
In this talk we make an introduction to data processing with big data and review the basic concepts in MapReduce programming with Hadoop. We also comment about the use of Pig to simplify the development of data processing applications
YDN Tuesdays are geek meetups organized the first Tuesday of each month by YDN in London
Este documento describe la historia, representaciones y rituales de la diosa Hécate en la mitología griega. Explica que Hécate era una diosa de la luna, la magia y las encrucijadas que se representaba con tres cabezas. Su ritual más conocido es el de los fuegos sagrados, que involucra encender 21 velas en un círculo e invocar a Hécate para pedir su sabiduría y guía.
Antrian adalah struktur data yang menyimpan elemen sesuai urutan masuk (FIFO). Terdapat beberapa metode utama pada antrian seperti enqueue untuk menambahkan elemen, dequeue untuk mengambil elemen pertama, dan peek untuk melihat elemen pertama tanpa menghapusnya. Antrian dapat diimplementasikan menggunakan array dengan menyimpan indeks elemen terakhir.
Algoritma dan pengetahuan terkait (menghitung, konversi, dll) Fazar Ikhwan Guntara
Dokumen tersebut membahas cara mengkonversi waktu dalam detik menjadi jam, menit dan detik dengan menggunakan konsep pembagian dan sisa bagi. Langkah-langkah perhitungannya dijelaskan secara rinci beserta ilustrasinya dalam bentuk skema alir.
1. Bab 5 membahas struktur data queue (antrian) dan implementasinya dalam bahasa pemrograman.
2. Queue adalah struktur data linear dimana penambahan elemen hanya bisa dilakukan di satu ujung dan penghapusan di ujung lain.
3. Queue dapat diimplementasikan menggunakan array linear atau linked list dengan operasi enqueue dan dequeue.
Makalah ini membahas tentang implementasi queue dengan bahasa pemrograman Pascal. Queue merupakan struktur data yang mengimplementasikan prinsip antrian First In First Out (FIFO). Makalah ini menjelaskan definisi dan gambaran umum queue, macam-macam queue, representasi queue secara statis menggunakan array dan representasi secara dinamis menggunakan linked list tunggal dan ganda. Juga dibahas queue berprioritas beserta contoh kode program untuk masing-masing implementasi queue.
Queue adalah linear list dimana data dimasukkan melalui rear dan dihapus dari front, mengikuti prinsip first in first out. Queue memiliki operasi seperti enqueue untuk memasukkan data ke rear, dequeue untuk menghapus data di front, serta mengakses elemen di front dan rear tanpa menghapusnya.
Laporan ini membahas tentang sistem antrian di Bread Talk Bakery di Surabaya. Studi kasus menganalisis model sistem antrian, rata-rata waktu antar kedatangan pelanggan dan pelayanan kasir, serta karakteristik lainnya seperti utilitas sistem dan rata-rata jumlah pelanggan dalam sistem."
Queue is a first-in first-out (FIFO) data structure where elements can only be added to the rear of the queue and removed from the front of the queue. It has two pointers - a front pointer pointing to the front element and a rear pointer pointing to the rear element. Queues can be implemented using arrays or linked lists. Common queue operations include initialization, checking if empty/full, enqueue to add an element, and dequeue to remove an element. The document then describes how these operations work for queues implemented using arrays, linked lists, and circular arrays. It concludes by providing exercises to implement specific queue tasks.
The document discusses different types of queues, including their definitions, properties, and implementations. It defines a queue as a linear data structure with two ends - one for adding elements and one for removing them, following a FIFO (first-in, first-out) approach. Key points covered include common queue operations like insertion and removal; circular queues which wrap elements around to avoid overflow; priority queues which order elements by priority; and deques which allow additions and removals from both ends.
Conexión de MongoDB con Hadoop - Luis Alberto Giménez - CAPSiDE #DevOSSAzureDaysCAPSiDE
This document discusses using MongoDB and Hadoop together. It provides an overview of MongoDB and Hadoop, describes how the MongoDB Hadoop Connector allows them to interoperate, and gives an example of building a graph of email sender-recipient relationships from Enron email data stored in MongoDB using Hadoop MapReduce, streaming, Pig, and Hive. The connector allows parallel processing of MongoDB data using Hadoop and integration with the Hadoop ecosystem.
Barcelona MUG MongoDB + Hadoop PresentationNorberto Leite
- The document discusses MongoDB and Hadoop, two popular big data platforms, and the MongoDB + Hadoop Connector which allows interoperation between the two.
- It provides an overview of MongoDB and Hadoop's key features for scalability, availability and processing large datasets.
- The connector allows processing data across MongoDB and Hadoop through MapReduce jobs without needing custom exports/imports.
- Examples show building a graph of email sender/recipient relationships from an Enron dataset stored in MongoDB using Hadoop Streaming, Pig and Hive.
Midwest php 2013 deploying php on paas- why & howdotCloud
Deploying PHP applications to Platform as a Service (PaaS) can provide several benefits over traditional hosting methods. PaaS allows developers to quickly deploy new environments for testing code changes. It also handles tasks like optimizing stacks, upgrading software, and providing comprehensive routing. PaaS aims to make deployment as simple as uploading code and eliminates the need to manually configure servers. While there is an initial learning curve to using PaaS tools and reworking some applications, it can improve the development to production workflow and allow applications to easily scale on demand.
Grunt is a JavaScript task runner that can automate front-end development tasks like linting, compiling Sass files to CSS, running tests and watching for file changes. It provides plugins for common tasks and allows developers to define custom workflows. The document discusses how Grunt can be used to integrate and run various tasks like linting, Sass compilation and watching files. It also explains how to define tasks, configure Grunt and run tasks from the command line. Custom tasks can be created and shared as plugins. The document provides an example of using Grunt to convert Markdown files to the Leanpub format and sync them to Dropbox.
The document summarizes PHP deployment on Platform as a Service (PaaS) and why it can help. It discusses how PHP deployment worked in the past with manual setups that were error-prone. PaaS allows fast deployment of new environments and leveraging of version control. It handles optimizations, upgrades, routing, and scaling automatically. The document provides an example of deploying a Symfony application to dotCloud PaaS, including configuring databases. Potential drawbacks are an initial learning curve and apps requiring reworking, but PaaS offers benefits like reduced costs and improved reliability.
Ready to leverage the power of a graph database to bring your application to the next level, but all the data is still stuck in a legacy relational database?
Fortunately, Neo4j offers several ways to quickly and efficiently import relational data into a suitable graph model. It's as simple as exporting the subset of the data you want to import and ingest it either with an initial loader in seconds or minutes or apply Cypher's power to put your relational data transactionally in the right places of your graph model.
In this webinar, Michael will also demonstrate a simple tool that can load relational data directly into Neo4j, automatically transforming it into a graph representation of your normalized entity-relationship model.
Automated Reports with Rstudio Server
Automated KPI reporting with Shiny Server
Process Validation Documentation with Jupyter Notebook
Automated Machine Learning with Dataiku
This document provides a summary of Hadoop and its ecosystem components. It begins with introducing the speaker, Andrew Brust, and his background. It then provides an overview of the key Hadoop components like MapReduce, HDFS, HBase, Hive, Pig, Sqoop and Flume. The document demonstrates several of these components and discusses how to use Microsoft Hadoop on Azure, Amazon EMR and Cloudera CDH virtual machines. It also summarizes the Mahout machine learning library and recommends a commercial product for visualizing and applying Mahout outputs.
This document provides an introduction to Node.js. It discusses that Node.js is an event-driven, non-blocking I/O platform for building scalable network applications using JavaScript. It was created to address issues with traditional blocking I/O by using asynchronous programming. The document outlines benefits of Node.js like using JavaScript for server-side applications, non-blocking I/O, a large module ecosystem, and an active community. It also provides examples of core modules, writing simple modules, and creating an HTTP server in Node.js.
introduction to data processing using Hadoop and PigRicardo Varela
In this talk we make an introduction to data processing with big data and review the basic concepts in MapReduce programming with Hadoop. We also comment about the use of Pig to simplify the development of data processing applications
YDN Tuesdays are geek meetups organized the first Tuesday of each month by YDN in London
This document provides an introduction to streams and their uses for data processing and analysis. Streams allow processing large datasets in a manageable way by handling data as a sequential flow rather than loading entire files into memory at once. The document discusses readable and writable streams that sources and sink data, as well as transform streams that manipulate data. It provides examples of using streams for tasks like scraping websites, normalizing data, and performing map-reduce operations. The programming benefits of streams like separation of concerns and a functional programming style are also outlined.
Hadoop is an open-source framework for storing and processing large datasets in a distributed computing environment. It allows for the storage and analysis of datasets that are too large for single servers. The document discusses several key Hadoop components including HDFS for storage, MapReduce for processing, HBase for column-oriented storage, Hive for SQL-like queries, Pig for data flows, and Sqoop for data transfer between Hadoop and relational databases. It provides examples of how each component can be used and notes that Hadoop is well-suited for large-scale batch processing of data.
PyconJP: Building a data preparation pipeline with Pandas and AWS LambdaFabian Dubois
Building a data preparation pipeline with Pandas and AWS Lambda
What is data preparation and why it is required.
How to prepare data with pandas.
How to set up a pipeline with AWS Lambda
https://youtu.be/pc0Xn0uAm34?t=9m15s
This document provides an introduction and overview of using Amazon's Elastic MapReduce (EMR) service for data intensive computing. It discusses uploading data to S3 storage, writing mappers and reducers in various languages like Python and streaming utilities, and executing a MapReduce job on EMR to process the data in parallel across a cluster of Amazon EC2 instances. The key steps involve loading input data to S3, defining the mapper and reducer processing logic, and downloading outputs from S3 upon job completion.
Why and How Powershell will rule the Command Line - Barcamp LA 4Ilya Haykinson
PowerShell is a command shell for Windows that treats commands as objects that interact through pipes and objects. It provides a fully-fledged programming language where commands manipulate objects and share a common naming convention. PowerShell holds that commands should do one thing well and interact through a consistent environment, addressing issues with text parsing between traditional command line programs.
This document provides a high-level overview of MapReduce and Hadoop. It begins with an introduction to MapReduce, describing it as a distributed computing framework that decomposes work into parallelized map and reduce tasks. Key concepts like mappers, reducers, and job tracking are defined. The structure of a MapReduce job is then outlined, showing how input is divided and processed by mappers, then shuffled and sorted before being combined by reducers. Example map and reduce functions for a word counting problem are presented to demonstrate how a full MapReduce job works.
iText 7 provides major improvements over iText 5 for creating and manipulating PDF documents, including a complete rewrite of the font and layout engines. It introduces a modular approach and improves support for advanced typography and Indic scripts. Key features include form filling and flattening, as well as efficient merging of multiple filled forms into a single PDF.
This document discusses Fluentd, an open source data collector. It provides an overview of Fluentd's architecture and components including input plugins, parser plugins, buffer plugins, output plugins, and formatter plugins. It also outlines Fluentd's roadmap, including plans to add filtering capabilities and improve the plugin API. Examples are given throughout to illustrate how Fluentd works and can be configured for use cases like log collection.
Spark is a fast and general engine for large-scale data processing. It runs on Hadoop clusters through YARN and Mesos, and can also run standalone. Spark is up to 100x faster than Hadoop for certain applications because it keeps data in memory rather than disk, and it supports iterative algorithms through its Resilient Distributed Dataset (RDD) abstraction. The presenter provides a demo of Spark's word count algorithm in Scala, Java, and Python to illustrate how easy it is to use Spark across languages.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
2. Target
Create a Zip file of PDF’s
based on a CSV data file
‣ Linear version
‣ Making it scale with Redis
parse csv
create pdf
create pdf
...
create pdf
zip
4. Simple Templating with String Interpolation
invoice.html
<<Q
<div class="title">
INVOICE #{invoice_nr}
‣ Merge data into HTML
•
template =
File.new('invoice.html').
read
•
html =
eval("<<QQQn#{template}
nQQQ”)
</div>
<div class="address">
#{name}</br>
#{street}</br>
#{zip} #{city}</br>
</div>
Q
5. Step 1: linear
‣ Create PDF
• prince xml using princely gem
• http://www.princexml.com
• p = Princely.new
p.add_style_sheets('invoice.css')
p.pdf_from_string(html)
6. Step 1: linear
‣ Create ZIP
• Zip::ZipOutputstream.
open(zipfile_name)do |zos|
files.each do |file, content|
zos.new_entry(file)
zos.puts content
end
end
7. Full Code
require 'csv'!
require 'princely'!
require 'zip/zip’!
!
DATA_FILE = ARGV[0]!
DATA_FILE_BASE_NAME = File.basename(DATA_FILE, ".csv”)!
!
# create a pdf document from a csv line!
def create_pdf(invoice_nr, name, street, zip, city)!
template = File.new('../resources/invoice.html').read!
html = eval("<<WTFMFn#{template}nWTFMF")!
p = Princely.new!
p.add_style_sheets('../resources/invoice.css')!
p.pdf_from_string(html)!
end!
!
# zip files from hash !
def create_zip(files_h)!
zipfile_name = "../out/#{DATA_FILE_BASE_NAME}.#{Time.now.to_s}.zip"!
Zip::ZipOutputStream.open(zipfile_name) do |zos|!
files_h.each do |name, content|!
zos.put_next_entry "#{name}.pdf"!
zos.puts content!
end!
end!
zipfile_name!
end!
!
# load data from csv!
docs = CSV.read(DATA_FILE) # array of arrays!
!
# create a pdf for each line in the csv !
# and put it in a hash!
files_h = docs.inject({}) do |files_h, doc|!
files_h[doc[0]] = create_pdf(*doc)!
files_h!
end!
!
# zip all pfd's from the hash !
create_zip files_h!
!
9. Step 2: from linear ...
parse csv
create pdf
create pdf
...
create pdf
zip
10. Step 2: ...to parallel
parse csv
create pdf
create pdf
zip
Threads
?
create pdf
11. Multi Threaded
‣ Advantage
• Lightweight (minimal overhead)
‣ Challenges (or why is it hard)
• Hard to code: most data structures are not thread safe by default, they
need synchronized access
• Hard to test: different execution paths , timings
• Hard to maintain
‣ Limitation
• single machine - not a solution for horizontal scalability
beyond the multi core cpu
12. Step 2: ...to parallel
parse csv
?
create pdf
create pdf
zip
create pdf
13. Multi Process
• scale across machines
• advanced support for debugging and monitoring at the
OS level
• simpler (code, testing, debugging, ...)
• slightly more overhead
BUT
14. But
all this assumes
“shared state across processes”
MemCached
parse csv
SQL?
shared state
create pdf
create pdf
create pdf
shared state
File System
zip
… OR …
Terra Cotta
15. Hello Redis
‣ Shared Memory Key Value Store with
High Level Data Structure support
• String (String, Int, Float)
• Hash (Map, Dictionary)
• List (Queue)
• Set
• ZSet (ordered by member or score)
16. About Redis
• Single threaded : 1 thread to serve them all
• (fit) Everything in memory
•
“Transactions” (multi exec)
•
Expiring keys
•
LUA Scripting
•
Publisher-Subscriber
•
Auto Create and Destroy
•
Pipelining
•
But … full clustering (master-master) is not available (yet)
17. Hello Redis
‣ redis-cli
•
•
•
•
set name “pascal” =
“pascal”
incr counter = 1
incr counter = 2
hset pascal name
“pascal”
•
hset pascal address
“merelbeke”
•
•
sadd persons pascal
smembers persons =
[pascal]
•
•
•
•
•
•
•
keys *
type pascal = hash
lpush todo “read” = 1
lpush todo “eat” = 2
lpop todo = “eat”
rpoplpush todo done =
“read”
lrange done 0 -1 =
“read”
19. Spread the Work
parse csv
process
1
zip
counter
Queue with data
create pdf
process
create pdf
process
...
20. Ruby on Redis
‣
Put PDF Create Input data on a Queue and do the counter
bookkeeping
!
docs.each do |doc|!
data = YAML::dump(doc)!
!r.lpush 'pdf:queue’, data!
r.incr ctr” # bookkeeping!
end!
22. Ruby on Redis
‣
Read PDF input data from Queue and do the counter bookkeeping
and put each created PDF in a Redis hash and signal if ready
while (true)!
_, msg = r.brpop 'pdf:queue’!
!doc = YAML::load(msg)!
#name of hash, key=docname, value=pdf!
r.hset(‘pdf:pdfs’, doc[0], create_pdf(*doc))
!
ctr = r.decr ‘ctr’
!
r.rpush ready, done if ctr == 0!
end!
23. Zip When Done
parse csv
process
ready
zip
3
Hash with pdfs
create pdf
process
create pdf
process
...
24. Ruby on Redis
‣
Wait for the ready signal
Fetch all pdf ’s
And zip them
!
r.brpop ready“ # wait for signal!
pdfs = r.hgetall ‘pdf:pdfs‘ # fetch hash!
create_zip pdfs # zip it
25. More Parallelism
parse csv
zip
ready
ready
ready
counter
counter
counter
hash
hash Pdfs
Hash with
Queue with data
create pdf
create pdf
...
26. Ruby on Redis
‣
Put PDF Create Input data on a Queue and do the counter
bookkeeping
# unique id for this input file!
UUID = SecureRandom.uuid!
docs.each do |doc|!
data = YAML::dump([UUID, doc])!
!r.lpush 'pdf:queue’, data!
r.incr ctr:#{UUID}” # bookkeeping!
end!
27. Ruby on Redis
‣
Read PDF input data from Queue and do the counter bookkeeping and
put each created PDF in a Redis hash
while (true)!
_, msg = r.brpop 'pdf:queue’!
uuid, doc = YAML::load(msg)!
r.hset(uuid, doc[0], create_pdf(*doc))!
ctr = r.decr ctr:#{uuid}
!
r.rpush ready:#{uuid}, done if ctr == 0
end!
!
28. Ruby on Redis
‣
Wait for the ready signal
Fetch all pdf ’s
And zip them
!
r.brpop ready:#{UUID}“ # wait for signal!
pdfs = r.hgetall(‘pdf:pdfs‘) # fetch hash!
create_zip(pdfs) # zip it
29. Full Code
require 'csv'!
require 'princely'!
require 'zip/zip’!
!
DATA_FILE = ARGV[0]!
DATA_FILE_BASE_NAME = File.basename(DATA_FILE, .csv”)!
!
# create a pdf document from a csv line!
def create_pdf(invoice_nr, name, street, zip, city)!
template = File.new('../resources/invoice.html').read!
html = eval(WTFMFn#{template}nWTFMF)!
p = Princely.new!
p.add_style_sheets('../resources/invoice.css')!
p.pdf_from_string(html)!
end!
!
# zip files from hash !
def create_zip(files_h)!
zipfile_name = ../out/#{DATA_FILE_BASE_NAME}.#{Time.now.to_s}.zip!
Zip::ZipOutputStream.open(zipfile_name) do |zos|!
files_h.each do |name, content|!
zos.put_next_entry #{name}.pdf!
zos.puts content!
end!
end!
zipfile_name!
end!
!
# load data from csv!
docs = CSV.read(DATA_FILE) # array of arrays!
!
# create a pdf for each line in the csv !
# and put it in a hash!
files_h = docs.inject({}) do |files_h, doc|!
files_h[doc[0]] = create_pdf(*doc)!
files_h!
end!
!
# zip all pfd's from the hash !
create_zip files_h!
!
LINEAR
require 'csv’!
require 'zip/zip'!
require 'redis'!
require 'yaml'!
require 'securerandom'!
!
# zip files from hash !
def create_zip(files_h)!
zipfile_name = ../out/#{DATA_FILE_BASE_NAME}.#{Time.now.to_s}.zip!
Zip::ZipOutputStream.open(zipfile_name) do |zos|!
files_h.each do |name, content|!
zos.put_next_entry #{name}.pdf!
zos.puts content!
end!
end!
zipfile_name!
end!
!
DATA_FILE = ARGV[0]!
DATA_FILE_BASE_NAME = File.basename(DATA_FILE, .csv)!
UUID = SecureRandom.uuid!
!
r = Redis.new!
my_counter = ctr:#{UUID}!
!
# load data from csv!
docs = CSV.read(DATA_FILE) # array of arrays!
!
docs.each do |doc| # distribute!!
r.lpush 'pdf:queue' , YAML::dump([UUID, doc])!
r.incr my_counter!
end!
!
r.brpop ready:#{UUID} #collect!!
create_zip(r.hgetall(UUID)) !
!
# clean up!
r.del my_counter!
r.del UUID !
puts All done!”!
MAIN
require 'redis'!
require 'princely'!
require 'yaml’!
!
# create a pdf document from a csv line!
def create_pdf(invoice_nr, name, street, zip, city)!
template = File.new('../resources/invoice.html').read!
html = eval(WTFMFn#{template}nWTFMF)!
p = Princely.new!
p.add_style_sheets('../resources/invoice.css')!
p.pdf_from_string(html)!
end!
!
r = Redis.new!
while (true)!
_, msg = r.brpop 'pdf:queue'!
uuid, doc = YAML::load(msg)!
r.hset(uuid , doc[0] , create_pdf(*doc))!
ctr = r.decr ctr:#{uuid} !
r.rpush ready:#{uuid}, done if ctr == 0!
end!
WORKER
Key functions (create pdf and create zip)
remain unchanged.
Distribution code highlighted
31. Multi Language Participants
parse csv
zip
counter
counter
counter
Queue with data
create pdf
hash
hash pdfs
Hash with
create pdf
...
32. Conclusions
From Linear To Multi Process Distributed
Is easy with
Redis Shared Memory High Level Data Structures
Atomic Counter for bookkeeping
Queue for work distribution
Queue as Signal
Hash for result sets