This document compares the performance of using XML-RPC versus Psycopg2 to access and update data in an OpenERP database. It describes programs written in Python using each method and measures the execution times. XML-RPC was slower, taking 3.27 seconds to run compared to 0.25 seconds for Psycopg2. Psycopg2 provides better performance by allowing direct SQL queries rather than going through the ORM. The document concludes that while XML-RPC provides security and compatibility with the OpenERP ORM, Psycopg2 is faster and allows standard SQL queries for better portability.
MongoDB Indexing Constraints and Creative SchemasMongoDB
This document summarizes Chris Winslett's talk on MongoDB indexing and improving performance. It discusses:
1) Common performance issues in MongoDB like inefficient queries causing high CPU from table scans, and how to address them through indexing and schema design.
2) Examples of performance problems like social networking queries and messages, and how indexing on viewable data rather than relationships improved queries.
3) Tools for debugging like mongostat, logs, and currentOp to find slow queries and operations.
4) The importance of measuring performance, finding slow queries, and refactoring schemas and queries to optimize for desired queries through indexing rather than always adding more indexes.
Optimizing Slow Queries with Indexes and CreativityMongoDB
This document discusses optimizing slow queries in MongoDB through the use of indexes and schema design. It begins with examples of common performance issues like inefficient reads from table scans and writes causing locks. Each example shows the slow query, explains the problem, and proposes a solution like adding an index. It emphasizes that new indexes only scale up to a point and sometimes the best approach is to redesign the schema or add hardware. The talk provides guidance on measuring performance, finding slow queries, determining if they can be indexed, and rephrasing problems in terms of desired queries.
One of the strongest points for using a NoSQL database is their focus on distribution — both for replication and sharding. This talks takes a short look at what replication is, why you should use it, and what is so difficult about it. We then take a look at MongoDB’s implementation in general and finally focus on what can go wrong. In a practical demo you see how to find the right balance between performance versus data safety and how to use it in your Java application.
Nyc open data project ii -- predict where to get and return my citibikeVivian S. Zhang
NYC Data Science Academy, NYC Open Data Meetup, Big Data, Data Science, NYC, Vivian Zhang, SupStat Inc,NYC, GBM, Machine learning, Time Series, Citibike usage prodiction, advanced R
Database replication involves keeping identical copies of data on different servers to provide redundancy and minimize downtime. Replication is recommended for databases in production from the start. A MongoDB replica set consists of a primary server that handles client requests and secondary servers that copy the primary's data. Replica sets can include up to 50 members with 7 voting members and use an oplog to replicate operations from the primary to secondaries. For elections and writes to succeed, a majority of voting members must be reachable.
PgCenter is a tool for monitoring and troubleshooting PostgreSQL. It provides a graphical interface to view key performance metrics and statuses. Some of its main features include displaying server health, load, memory and disk usage, statement performance, replication status and more. It aims to help PostgreSQL administrators quickly check the health of their databases and identify potential problems.
This document summarizes MySQL's init_connect feature which allows SQL statements to be executed for each client connection. It provides examples of setting init_connect to log client connections to a table and discusses how to address issues like preventing the logs from being written to binary logs. The document also estimates storage needed for connection logs and provides an example of periodically deleting old log entries.
This document compares the performance of using XML-RPC versus Psycopg2 to access and update data in an OpenERP database. It describes programs written in Python using each method and measures the execution times. XML-RPC was slower, taking 3.27 seconds to run compared to 0.25 seconds for Psycopg2. Psycopg2 provides better performance by allowing direct SQL queries rather than going through the ORM. The document concludes that while XML-RPC provides security and compatibility with the OpenERP ORM, Psycopg2 is faster and allows standard SQL queries for better portability.
MongoDB Indexing Constraints and Creative SchemasMongoDB
This document summarizes Chris Winslett's talk on MongoDB indexing and improving performance. It discusses:
1) Common performance issues in MongoDB like inefficient queries causing high CPU from table scans, and how to address them through indexing and schema design.
2) Examples of performance problems like social networking queries and messages, and how indexing on viewable data rather than relationships improved queries.
3) Tools for debugging like mongostat, logs, and currentOp to find slow queries and operations.
4) The importance of measuring performance, finding slow queries, and refactoring schemas and queries to optimize for desired queries through indexing rather than always adding more indexes.
Optimizing Slow Queries with Indexes and CreativityMongoDB
This document discusses optimizing slow queries in MongoDB through the use of indexes and schema design. It begins with examples of common performance issues like inefficient reads from table scans and writes causing locks. Each example shows the slow query, explains the problem, and proposes a solution like adding an index. It emphasizes that new indexes only scale up to a point and sometimes the best approach is to redesign the schema or add hardware. The talk provides guidance on measuring performance, finding slow queries, determining if they can be indexed, and rephrasing problems in terms of desired queries.
One of the strongest points for using a NoSQL database is their focus on distribution — both for replication and sharding. This talks takes a short look at what replication is, why you should use it, and what is so difficult about it. We then take a look at MongoDB’s implementation in general and finally focus on what can go wrong. In a practical demo you see how to find the right balance between performance versus data safety and how to use it in your Java application.
Nyc open data project ii -- predict where to get and return my citibikeVivian S. Zhang
NYC Data Science Academy, NYC Open Data Meetup, Big Data, Data Science, NYC, Vivian Zhang, SupStat Inc,NYC, GBM, Machine learning, Time Series, Citibike usage prodiction, advanced R
Database replication involves keeping identical copies of data on different servers to provide redundancy and minimize downtime. Replication is recommended for databases in production from the start. A MongoDB replica set consists of a primary server that handles client requests and secondary servers that copy the primary's data. Replica sets can include up to 50 members with 7 voting members and use an oplog to replicate operations from the primary to secondaries. For elections and writes to succeed, a majority of voting members must be reachable.
PgCenter is a tool for monitoring and troubleshooting PostgreSQL. It provides a graphical interface to view key performance metrics and statuses. Some of its main features include displaying server health, load, memory and disk usage, statement performance, replication status and more. It aims to help PostgreSQL administrators quickly check the health of their databases and identify potential problems.
This document summarizes MySQL's init_connect feature which allows SQL statements to be executed for each client connection. It provides examples of setting init_connect to log client connections to a table and discusses how to address issues like preventing the logs from being written to binary logs. The document also estimates storage needed for connection logs and provides an example of periodically deleting old log entries.
Talking about Neo4j after 1 year of using it production. This presentation covering db structure(internals), cypher queries, extensions development, db tuning & settings.
Phil Roth presented the results of a malware classification model bakeoff between several machine learning algorithms. The models evaluated were k-nearest neighbors, logistic regression, support vector machines, naive Bayes, random forests, gradient boosted decision trees, and deep learning. Based on the performance, size, and query time metrics, gradient boosted decision trees had the best overall results. However, the presenter noted that deep learning approaches deserve more research due to their potential to learn directly from file content. The conclusions were that gradient boosted decision trees could be deployed to endpoints while larger deep learning models could be used in the cloud after further development.
This document provides information about Olivier Duchenne and his experience and qualifications. It summarizes his educational background which includes a Ph.D in Computer Science from ENS Paris/INRIA and a postdoctoral fellowship at Carnegie Mellon University. It also lists his professional experience which includes positions at NEC Labs, Intel, and as a co-founder of Solidware. The document then provides guidelines for machine learning and discusses challenges such as having enough and changing data. It explores the history and reasons for increased use of machine learning in computer vision.
Streaming Way to Webscale: How We Scale Bitly via StreamingAll Things Open
All Things Open 2014 - Day 2
Thursday, October 23rd, 2014
Peter Herndon
Senior Application Engineer for Bitly
DevOps
Streaming Way to Webscale: How We Scale Bitly via Streaming
This document discusses the development of Apache Pig on Tez, an execution engine for Pig jobs. Pig on Tez allows Pig workflows to be executed as directed acyclic graphs (DAGs) using Tez, improving performance over the default MapReduce execution. Key benefits of Tez include eliminating intermediate data writes, reducing job launch overhead, and allowing more flexible data flows. However, challenges remain around automatically determining optimal parallelism and integrating Tez with user interface and monitoring tools. Future work is needed to address these issues.
This document discusses building price models using data mining techniques. It describes creating a wine price dataset based on wine rating and age, with price determined by a wineprice function. The dataset is then used to test k-nearest neighbors (k-NN) algorithms and weighted k-NN algorithms for price estimation. Cross-validation and handling of non-homogeneous variables like bottle size and aisle location are also covered. Optimization techniques like hill climbing, simulated annealing, and genetic algorithms are applied to find optimal weight values for variables in the weighted k-NN algorithm.
FØCAL Boston AiR - Computer Vision Tracing and Hardware SimulationFØCAL
FØCAL presentation at Boston AIR (AI, Autonomy & Robotics) meetup on 28 Mar 2019.
Challenges of getting computer vision applications from prototype to production.
Image processing and computer vision pipeline tracing and edge device hardware selection platform.
Exploring OpenFaaS autoscalability on Kubernetes with the Chaos ToolkitSylvain Hellegouarch
A report generated by the Chaos Toolkit after a Chaos Engineering experiment against OpenFaaS on Kubernetes.
View the run of the experiment at https://asciinema.org/a/dv4cNOMC5k1oWDhWDe97d5eij
1. The document discusses using Deeplearning4j and Kafka together for machine learning workflows. It describes how Deeplearning4j can be used to build, train, and deploy neural networks on JVM and Spark, while Kafka can be used to stream data for training and inference.
2. An example application is described that performs anomaly detection on log files from a CDN by aggregating the data to reduce the number of data points. This allows the model to run efficiently on available GPU hardware.
3. The document provides a link to a GitHub repository with a code example that uses Kafka to stream data, Keras to train a model, and Deeplearning4j to perform inference in Java and deploy the model.
This document discusses measuring and optimizing data processing performance on CPUs and GPUs. It provides examples using Cupy, NumPy, Apache Ignite, and ROCm to perform operations and measure throughput on different hardware. Specific examples include sorting random data on CPU and GPU to compare performance, running Apache Ignite benchmarks on different CPUs, and using ROCm and PyOpenCL to leverage an AMD GPU for general-purpose computing. The document emphasizes that tightly coupling fast memory and computation is important for workloads with large data processing requirements.
This package loads data from flat files into SQL Server tables, performs transformations and aggregations, and writes output to multiple destinations. It uses the following tasks:
1. Execute SQL tasks to truncate tables and log package events and errors.
2. A Foreach loop container to iterate through multiple flat files.
3. A data flow task containing sources, destinations, and transformations to extract, transform, and load the data. Transformations include aggregation, data type conversion, creating new columns, and conditional splitting.
4. Additional tasks to extract a name from a table, display it, and count row numbers.
PyCon India 2009 Presentation Python tools for Network Security
The document discusses various Python tools for network security including Pypcap, Dpkt and Scapy. It provides an overview of packet capture and inspection capabilities of these tools and code examples to capture and analyze network packets. Specific examples demonstrated include an HTTP protocol sniffer, host scanning and DNS queries using Scapy.
Machine learning is teaching the computer how to learn by itself. It is far easier to be done, especially when you have small data set and a good level of expertise in your field. Classifying objects, predicting who will buy, where to park your car or which test will fail may be achieved with grassy algorithm like neural networks, genetic algorithms or ant herding. PHP is in good position to make use of such teachings, and take advantages of related technologies like R. By the end of the session, you’ll know where you want to try it.
This study sequenced and analyzed the genome of Chlamydia trachomatis. Quality control was performed on sequencing reads. The reads were then assembled using SPAdes and evaluated for quality. The assembled contigs were annotated using Prokka. Reads were aligned to the assembly using Bowtie2 and counted using HTSeq to quantify gene expression. Differential expression analysis was performed using DESeq2 to compare gene expression at different time points of infection. Various plots were generated to visualize the data.
Easing the Complex with SPBench frameworkadriano1mg
SPBench is a framework for stream processing benchmarks in C++. It provides real-world stream processing applications in a high-level and reusable abstraction. This allows users to easily build custom benchmarks for evaluating different stream processing approaches. SPBench hides low-level implementation details and automatically handles data encapsulation and metrics collection. It includes several common stream processing benchmarks and supports parallel implementations. The SPBench API and command line interface provide tools for managing benchmarks, configuring execution, and collecting performance metrics.
The document discusses Jugaad, a proof-of-concept toolkit that demonstrates code injection on Linux systems similar to CreateRemoteThread on Windows. It does this using the ptrace system call to attach to a process, read/write its memory, and inject shellcode that allocates memory and creates a thread to execute arbitrary code within the target process context. First, it explains how ptrace can be used to manipulate another process. Then it describes how Jugaad uses these ptrace capabilities to meet the requirements of allocating memory, creating a thread, and executing payload code inside the target process.
Nagios Conference 2014 - Rodrigo Faria - Developing your PluginNagios
Rodrigo Faria's presentation on Developing your Plugin.
The presentation was given during the Nagios World Conference North America held Oct 13th - Oct 16th, 2014 in Saint Paul, MN. For more information on the conference (including photos and videos), visit: http://go.nagios.com/conference
Using open source tools for network device dataplane testing.
Our experiences from redGuardian DDoS mitigation scrubber testing.
Presented at PLNOG 20 (2018).
PLNOG20 - Paweł Małachowski - Stress your DUT–wykorzystanie narzędzi open sou...PROIDEA
Wybór docelowej platformy sieciowej (np. routera, firewalla, scrubbera DDoS) jest często poprzedzony jej testami. Jednym z celów testów jest sprawdzenie, czy parametry wydajnościowe deklarowane przez producenta odpowiadają rzeczywistości. Zespół rozwijający redGuardian Anty DDoS testuje rozwiązanie regresyjnie i wydajnościowo w sposób zautomatyzowany od początku jego istnienia. W czasie prezentacji przeanalizujemy aspekty, na które warto zwrócić uwagę w czasie testów wydajnościowych urządzeń IP oraz przyjrzymy się narzędziom open source pomocnym w realizacji tego zadania.
PgQ Generic high-performance queue for PostgreSQLelliando dias
PgQ is a generic high-performance queue for PostgreSQL. It allows for processing of data created in transactions at a later time in a high-throughput, low-latency, and robust manner. This is achieved by exporting PostgreSQL's MVCC visibility information which allows querying events that occurred between snapshots in a efficient manner without locking. PgQ provides APIs for easily inserting and consuming events in batches between snapshots.
This document provides an overview of Node.js application performance analysis and optimization as well as distributed system design. It discusses analyzing and optimizing CPU, memory, file I/O and network I/O usage. It also covers profiling Node.js applications using tools like Linux profiling tools, Node.js libraries, and V8 profiling tools. Lastly it discusses designing distributed systems using single machine and cluster approaches.
Talking about Neo4j after 1 year of using it production. This presentation covering db structure(internals), cypher queries, extensions development, db tuning & settings.
Phil Roth presented the results of a malware classification model bakeoff between several machine learning algorithms. The models evaluated were k-nearest neighbors, logistic regression, support vector machines, naive Bayes, random forests, gradient boosted decision trees, and deep learning. Based on the performance, size, and query time metrics, gradient boosted decision trees had the best overall results. However, the presenter noted that deep learning approaches deserve more research due to their potential to learn directly from file content. The conclusions were that gradient boosted decision trees could be deployed to endpoints while larger deep learning models could be used in the cloud after further development.
This document provides information about Olivier Duchenne and his experience and qualifications. It summarizes his educational background which includes a Ph.D in Computer Science from ENS Paris/INRIA and a postdoctoral fellowship at Carnegie Mellon University. It also lists his professional experience which includes positions at NEC Labs, Intel, and as a co-founder of Solidware. The document then provides guidelines for machine learning and discusses challenges such as having enough and changing data. It explores the history and reasons for increased use of machine learning in computer vision.
Streaming Way to Webscale: How We Scale Bitly via StreamingAll Things Open
All Things Open 2014 - Day 2
Thursday, October 23rd, 2014
Peter Herndon
Senior Application Engineer for Bitly
DevOps
Streaming Way to Webscale: How We Scale Bitly via Streaming
This document discusses the development of Apache Pig on Tez, an execution engine for Pig jobs. Pig on Tez allows Pig workflows to be executed as directed acyclic graphs (DAGs) using Tez, improving performance over the default MapReduce execution. Key benefits of Tez include eliminating intermediate data writes, reducing job launch overhead, and allowing more flexible data flows. However, challenges remain around automatically determining optimal parallelism and integrating Tez with user interface and monitoring tools. Future work is needed to address these issues.
This document discusses building price models using data mining techniques. It describes creating a wine price dataset based on wine rating and age, with price determined by a wineprice function. The dataset is then used to test k-nearest neighbors (k-NN) algorithms and weighted k-NN algorithms for price estimation. Cross-validation and handling of non-homogeneous variables like bottle size and aisle location are also covered. Optimization techniques like hill climbing, simulated annealing, and genetic algorithms are applied to find optimal weight values for variables in the weighted k-NN algorithm.
FØCAL Boston AiR - Computer Vision Tracing and Hardware SimulationFØCAL
FØCAL presentation at Boston AIR (AI, Autonomy & Robotics) meetup on 28 Mar 2019.
Challenges of getting computer vision applications from prototype to production.
Image processing and computer vision pipeline tracing and edge device hardware selection platform.
Exploring OpenFaaS autoscalability on Kubernetes with the Chaos ToolkitSylvain Hellegouarch
A report generated by the Chaos Toolkit after a Chaos Engineering experiment against OpenFaaS on Kubernetes.
View the run of the experiment at https://asciinema.org/a/dv4cNOMC5k1oWDhWDe97d5eij
1. The document discusses using Deeplearning4j and Kafka together for machine learning workflows. It describes how Deeplearning4j can be used to build, train, and deploy neural networks on JVM and Spark, while Kafka can be used to stream data for training and inference.
2. An example application is described that performs anomaly detection on log files from a CDN by aggregating the data to reduce the number of data points. This allows the model to run efficiently on available GPU hardware.
3. The document provides a link to a GitHub repository with a code example that uses Kafka to stream data, Keras to train a model, and Deeplearning4j to perform inference in Java and deploy the model.
This document discusses measuring and optimizing data processing performance on CPUs and GPUs. It provides examples using Cupy, NumPy, Apache Ignite, and ROCm to perform operations and measure throughput on different hardware. Specific examples include sorting random data on CPU and GPU to compare performance, running Apache Ignite benchmarks on different CPUs, and using ROCm and PyOpenCL to leverage an AMD GPU for general-purpose computing. The document emphasizes that tightly coupling fast memory and computation is important for workloads with large data processing requirements.
This package loads data from flat files into SQL Server tables, performs transformations and aggregations, and writes output to multiple destinations. It uses the following tasks:
1. Execute SQL tasks to truncate tables and log package events and errors.
2. A Foreach loop container to iterate through multiple flat files.
3. A data flow task containing sources, destinations, and transformations to extract, transform, and load the data. Transformations include aggregation, data type conversion, creating new columns, and conditional splitting.
4. Additional tasks to extract a name from a table, display it, and count row numbers.
PyCon India 2009 Presentation Python tools for Network Security
The document discusses various Python tools for network security including Pypcap, Dpkt and Scapy. It provides an overview of packet capture and inspection capabilities of these tools and code examples to capture and analyze network packets. Specific examples demonstrated include an HTTP protocol sniffer, host scanning and DNS queries using Scapy.
Machine learning is teaching the computer how to learn by itself. It is far easier to be done, especially when you have small data set and a good level of expertise in your field. Classifying objects, predicting who will buy, where to park your car or which test will fail may be achieved with grassy algorithm like neural networks, genetic algorithms or ant herding. PHP is in good position to make use of such teachings, and take advantages of related technologies like R. By the end of the session, you’ll know where you want to try it.
This study sequenced and analyzed the genome of Chlamydia trachomatis. Quality control was performed on sequencing reads. The reads were then assembled using SPAdes and evaluated for quality. The assembled contigs were annotated using Prokka. Reads were aligned to the assembly using Bowtie2 and counted using HTSeq to quantify gene expression. Differential expression analysis was performed using DESeq2 to compare gene expression at different time points of infection. Various plots were generated to visualize the data.
Easing the Complex with SPBench frameworkadriano1mg
SPBench is a framework for stream processing benchmarks in C++. It provides real-world stream processing applications in a high-level and reusable abstraction. This allows users to easily build custom benchmarks for evaluating different stream processing approaches. SPBench hides low-level implementation details and automatically handles data encapsulation and metrics collection. It includes several common stream processing benchmarks and supports parallel implementations. The SPBench API and command line interface provide tools for managing benchmarks, configuring execution, and collecting performance metrics.
The document discusses Jugaad, a proof-of-concept toolkit that demonstrates code injection on Linux systems similar to CreateRemoteThread on Windows. It does this using the ptrace system call to attach to a process, read/write its memory, and inject shellcode that allocates memory and creates a thread to execute arbitrary code within the target process context. First, it explains how ptrace can be used to manipulate another process. Then it describes how Jugaad uses these ptrace capabilities to meet the requirements of allocating memory, creating a thread, and executing payload code inside the target process.
Nagios Conference 2014 - Rodrigo Faria - Developing your PluginNagios
Rodrigo Faria's presentation on Developing your Plugin.
The presentation was given during the Nagios World Conference North America held Oct 13th - Oct 16th, 2014 in Saint Paul, MN. For more information on the conference (including photos and videos), visit: http://go.nagios.com/conference
Using open source tools for network device dataplane testing.
Our experiences from redGuardian DDoS mitigation scrubber testing.
Presented at PLNOG 20 (2018).
PLNOG20 - Paweł Małachowski - Stress your DUT–wykorzystanie narzędzi open sou...PROIDEA
Wybór docelowej platformy sieciowej (np. routera, firewalla, scrubbera DDoS) jest często poprzedzony jej testami. Jednym z celów testów jest sprawdzenie, czy parametry wydajnościowe deklarowane przez producenta odpowiadają rzeczywistości. Zespół rozwijający redGuardian Anty DDoS testuje rozwiązanie regresyjnie i wydajnościowo w sposób zautomatyzowany od początku jego istnienia. W czasie prezentacji przeanalizujemy aspekty, na które warto zwrócić uwagę w czasie testów wydajnościowych urządzeń IP oraz przyjrzymy się narzędziom open source pomocnym w realizacji tego zadania.
PgQ Generic high-performance queue for PostgreSQLelliando dias
PgQ is a generic high-performance queue for PostgreSQL. It allows for processing of data created in transactions at a later time in a high-throughput, low-latency, and robust manner. This is achieved by exporting PostgreSQL's MVCC visibility information which allows querying events that occurred between snapshots in a efficient manner without locking. PgQ provides APIs for easily inserting and consuming events in batches between snapshots.
This document provides an overview of Node.js application performance analysis and optimization as well as distributed system design. It discusses analyzing and optimizing CPU, memory, file I/O and network I/O usage. It also covers profiling Node.js applications using tools like Linux profiling tools, Node.js libraries, and V8 profiling tools. Lastly it discusses designing distributed systems using single machine and cluster approaches.
How Automated Vulnerability Analysis Discovered Hundreds of Android 0-daysPriyanka Aash
Death from a million bugs. Android has become one of the world’s most deployed operating systems. Recently researchers have been focused on uncovering vulnerabilities in the Android smartphone ecosystem. This session will present newly developed automated vulnerability analysis techniques that resulted in the discovery of hundreds of previously unknown vulnerabilities.
Learning Objectives:
1: Learn how to use automated vulnerability analysis to ID security bugs at scale.
2: Learn about state-of-the-art and novel techniques for automated vulnerability analysis.
3: Learn proven techniques to find vulnerabilities in bootloaders, kernel drives and apps.
(Source: RSA Conference USA 2018)
The document discusses the potential for offering Memcached as a cloud service. It notes that simplicity drives separation of concerns in cloud architectures. The document presents results from ping tests showing significantly faster response times for some cloud DNS and CDN providers than traditional DNS servers. It argues that anything can potentially be offered as a cloud service and that people will pay for easy-to-use services. It advocates making services as easy to use as possible then making them even easier to use. It also suggests that services should work with other cloud offerings and don't need to do everything themselves.
Design of a lightweight set of data pipelines to scrub PII information.
Scrubbing PII information from data brings ease of sharing data.
It also helps organisations to confidently push data outside organisation for large scale analytics on the cloud.
The Hidden Face of Cost-Based Optimizer: PL/SQL Specific StatisticsMichael Rosenblum
Database statistics are not limited to tables, columns, and indexes. PL/SQL functions also have a number of associated statistics, namely costs (CPU, I/O, network), selectivity, and cardinality (for functions that return collections). These statistics have default values that only somewhat represent reality. However, these values are always used by Oracle's cost-based optimizer to build execution plans. This session uses real-life examples to illustrate how properly managed PL/SQL statistics can significantly improve executions plans. It also demonstrates that Oracle's extensible optimizer is flexible enough to support packaged functions.
Similar to Image classification with Deeplearning4j (20)
This document outlines steps in a research process and poses several questions. It first lists asking a question and collecting data as initial steps. It then humorously suggests "profit" as a final step. The following questions could be topics of research and include whether black chocolate reduces cancer risk, if supplements increase vitamin D, whether replacing meat affects weight, and if outdoor evening time impacts sleep.
The poem describes an entity that devours all living and non-living things such as birds, beasts, trees, flowers, iron and steel. It also grinds hard stones, slays kings, ruins towns, and beats mountains down. The entity represents time, which gradually destroys all things that exist.
Slides for a talk at the Seattle Java User Group about building a workflow management application for a biomedical lab on top of the OSGi module system and the Eclipse Rich Client Platform.
The document discusses 5 lessons learned from developing the beta.uniprot.org website. The first lesson is not to fall in love with a particular technology. The second lesson is to observe how users actually interact with a system instead of relying only on their feedback. The third lesson is that users cannot be simply classified as beginners or experts. The fourth lesson is to ensure support from someone with political influence. The fifth and final lesson is to watch the time allotted for presentations.
UniProt is a large life sciences database that uses several ontologies like the Gene Ontology, keywords, taxonomy, and pathways to provide consistent navigation and aggregation of its data on protein sequences, functions, and features. The document discusses how ontologies help organize UniProt's large amount of data and provide benefits like auto-completion, set-oriented views, and potential for automation. It also seeks feedback on which additional data from UniProt could be organized through more detailed ontologies.
artificial intelligence and data science contents.pptxGauravCar
What is artificial intelligence? Artificial intelligence is the ability of a computer or computer-controlled robot to perform tasks that are commonly associated with the intellectual processes characteristic of humans, such as the ability to reason.
› ...
Artificial intelligence (AI) | Definitio
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...shadow0702a
This document serves as a comprehensive step-by-step guide on how to effectively use PyCharm for remote debugging of the Windows Subsystem for Linux (WSL) on a local Windows machine. It meticulously outlines several critical steps in the process, starting with the crucial task of enabling permissions, followed by the installation and configuration of WSL.
The guide then proceeds to explain how to set up the SSH service within the WSL environment, an integral part of the process. Alongside this, it also provides detailed instructions on how to modify the inbound rules of the Windows firewall to facilitate the process, ensuring that there are no connectivity issues that could potentially hinder the debugging process.
The document further emphasizes on the importance of checking the connection between the Windows and WSL environments, providing instructions on how to ensure that the connection is optimal and ready for remote debugging.
It also offers an in-depth guide on how to configure the WSL interpreter and files within the PyCharm environment. This is essential for ensuring that the debugging process is set up correctly and that the program can be run effectively within the WSL terminal.
Additionally, the document provides guidance on how to set up breakpoints for debugging, a fundamental aspect of the debugging process which allows the developer to stop the execution of their code at certain points and inspect their program at those stages.
Finally, the document concludes by providing a link to a reference blog. This blog offers additional information and guidance on configuring the remote Python interpreter in PyCharm, providing the reader with a well-rounded understanding of the process.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
17. log.info("Loading data from {}...", imagePath);
PathLabelGenerator labelMaker = new ImageDescriptionLabelGenerator("Trail");
BalancedPathFilter pathFilter = new BalancedPathFilter(rand, NativeImageLoader.ALLOWED_FORMATS, labelMaker);
FileSplit fileSplit = new FileSplit(imagePath, NativeImageLoader.ALLOWED_FORMATS, rand);
InputSplit[] inputSplit = fileSplit.sample(pathFilter, splitTrainTest, 1 - splitTrainTest);
InputSplit trainingData = inputSplit[0];
InputSplit testingData = inputSplit[1];
log.info("Building model...");
AlexNet model = new AlexNet(numLabels, seed, iterations);
model.setInputShape(new int[][] {{ channels, width, height }});
MultiLayerNetwork network = model.init();
log.info("Training model using {} samples...", trainingData.length());
ImageRecordReader recordReader = new ImageRecordReader(height, width, channels, labelMaker);
recordReader.initialize(trainingData, null);
DataSetIterator dataIter = new RecordReaderDataSetIterator(recordReader, batchSize, 1, numLabels);
scaler.fit(dataIter);
DataNormalization scaler = new ImagePreProcessingScaler(0, 1);
dataIter.setPreProcessor(scaler);
MultipleEpochsIterator trainIter = new MultipleEpochsIterator(epochs, dataIter);
network.fit(trainIter);
log.info("Evaluating model using {} samples...", testingData.length());
recordReader.initialize(testingData);
dataIter = new RecordReaderDataSetIterator(recordReader, batchSize, 1, numLabels);
scaler.fit(dataIter);
dataIter.setPreProcessor(scaler);
Evaluation eval = network.evaluate(dataIter);
log.info(eval.stats(true));
log.info("Saving model to {}...", modelPath);
ModelSerializer.writeModel(network, modelPath, true);
18. 14:57:44.074 INFO ImageModelBuilder - Loading data from imagesphotos-1000...
14:57:57.518 INFO ImageModelBuilder - Building model...
14:57:57.550 INFO org.nd4j.linalg.factory.Nd4jBackend - Loaded [CpuBackend] backend
14:57:58.398 INFO org.nd4j.nativeblas.NativeOpsHolder - Number of threads used for NativeOps: 2
14:57:58.724 INFO org.nd4j.nativeblas.Nd4jBlas - Number of threads used for BLAS: 2
14:57:58.726 INFO o.n.l.a.o.e.DefaultOpExecutioner - Backend used: [CPU]; OS: [Windows 10]
14:57:58.726 INFO o.n.l.a.o.e.DefaultOpExecutioner - Cores: [4]; Memory: [10.7GB];
14:57:58.726 INFO o.n.l.a.o.e.DefaultOpExecutioner - Blas vendor: [OPENBLAS]
14:58:03.013 INFO o.d.nn.multilayer.MultiLayerNetwork - Starting MultiLayerNetwork with WorkspaceModes
set to [training: SINGLE; inference: SINGLE]
14:58:04.717 INFO ImageModelBuilder - Training model using 354 samples...
14:58:14.191 INFO o.d.d.i.MultipleEpochsIterator - Epoch 1, number of batches completed 2
14:58:18.025 INFO o.d.o.l.ScoreIterationListener - Score at iteration 0 is 0.8973041230682596
14:58:20.517 INFO o.d.d.i.MultipleEpochsIterator - Epoch 2, number of batches completed 2
14:58:20.937 INFO o.d.o.l.ScoreIterationListener - Score at iteration 1 is 1.3137047898462706
14:58:24.958 INFO o.d.o.l.ScoreIterationListener - Score at iteration 2 is 0.8076716838671805
14:58:26.688 INFO o.d.o.l.ScoreIterationListener - Score at iteration 3 is 1.592903741321036
14:58:26.719 INFO o.d.d.i.MultipleEpochsIterator - Epoch 3, number of batches completed 2
14:58:30.798 INFO o.d.o.l.ScoreIterationListener - Score at iteration 4 is 0.8828689747742253
14:58:32.682 INFO o.d.o.l.ScoreIterationListener - Score at iteration 5 is 1.2827695432226482
14:58:33.259 INFO o.d.d.i.MultipleEpochsIterator - Epoch 4, number of batches completed 2
14:58:36.885 INFO o.d.o.l.ScoreIterationListener - Score at iteration 6 is 0.8240061319768378
14:58:38.641 INFO o.d.o.l.ScoreIterationListener - Score at iteration 7 is 1.4599294093575166
14:58:39.405 INFO o.d.d.i.MultipleEpochsIterator - Epoch 5, number of batches completed 2
14:58:42.013 INFO o.d.o.l.ScoreIterationListener - Score at iteration 8 is 0.8207273637625443
14:58:43.364 INFO o.d.o.l.ScoreIterationListener - Score at iteration 9 is 1.3621771943792143
14:59:21.842 INFO ImageModelBuilder - Evaluating model using 88 samples...
14:59:24.084 INFO ImageModelBuilder -
Examples labeled as not_trail classified by model as not_trail: 44 times
Examples labeled as trail classified by model as not_trail: 44 times
==========================Scores========================================
# of classes: 2
Accuracy: 0.5000
Precision: 0.5000
Recall: 0.5000
F1 Score: 0.0000
========================================================================
14:59:24.084 INFO ImageModelBuilder - Saving model to modelsalexnet-1000-dev.net...