This document provides instructions to install Nginx 0.8.x and PHP 5.2.10 using the FastCGI mode on a Linux system. It involves installing prerequisite packages and libraries, compiling and configuring MySQL, PHP, and additional PHP modules like memcache and imagick. The PHP installation is configured to use FastCGI mode and optimized using eAccelerator. Configuration files are modified and shell scripts are created to manage the MySQL service.
1. The document provides installation steps for Nginx 0.9.x, PHP 5.3.4, and related components on a Linux server using the yum package manager and compiling from source code.
2. It involves installing dependencies, compiling and configuring MySQL, PHP, and extensions like memcache and imagick.
3. The PHP configuration is optimized for performance using eAccelerator caching and the Zend Guard Loader. Nginx, PHP-FPM, and MySQL systemd service scripts are also configured.
The document provides guidelines for deploying an L.N.M.P environment on a 64-bit server. It specifies directory locations for source code, installed software, scripts and logs. It also outlines steps to update the system, install and configure MySQL, Nginx, PHP and other packages, including compiling Nginx with specific modules and options, setting Nginx as a service, and enabling syntax highlighting for Nginx configuration files.
The document provides instructions for installing, configuring, and uninstalling the Apache HTTP Server on Linux systems. It discusses downloading and extracting the Apache files, using yum to install dependencies, configuring modules, and stopping or removing Apache using various commands. Graphical tools for configuring Apache like system-config-httpd are also mentioned.
The document provides instructions on Docker practice including prerequisites, basic Docker commands, running containers from images, committing container changes to new images, logging into Docker Hub and pushing images.
It begins with prerequisites of having Ubuntu 18.04 or higher and installing the latest Docker engine and Docker compose. It then explains that Docker runs processes in isolated containers and uses layered images.
The document demonstrates basic commands like docker version, docker images, docker pull, docker search, docker run, docker ps, docker stop, docker rm and docker rmi. It also shows how to commit container changes to a new image with docker commit, tag and push images to Docker Hub. Other topics covered include docker exec, docker save/load, docker
This document provides instructions and examples for properly configuring and using .htaccess files. It discusses Apache and PHP configuration as well as settings for performance, security, redirects, and more. Sections cover topics like error documents, rewriting URLs, compression, caching, and protecting files. The goal is to help optimize a WordPress site using the .htaccess file.
Virtualization and automation of library software/machines + PuppetOmar Reygaert
The document discusses virtualization, automation, and Puppet. It begins with an introduction to virtualization and hands-on labs. It then covers automation through kickstart files and preseeding to automate operating system installation. Hands-on labs are also provided for automation. Finally, it discusses Puppet for configuration management, including node definitions, modules, and resources to manipulate files, packages, users and more. Hands-on labs are presented for implementing SFX configuration with Puppet.
WordPress Home Server with Raspberry PiYuriko IKEDA
This document provides instructions for setting up a WordPress site on a Raspberry Pi server using a static IP address. It details steps such as installing necessary software like Apache, MySQL, PHP and WordPress using apt-get. It also covers configuring PHP, securing phpMyAdmin with HTTP authentication, creating a MySQL database and user, and generating security keys for WordPress. The document concludes with suggestions for additional configuration tasks and references for more information.
1. The document provides installation steps for Nginx 0.9.x, PHP 5.3.4, and related components on a Linux server using the yum package manager and compiling from source code.
2. It involves installing dependencies, compiling and configuring MySQL, PHP, and extensions like memcache and imagick.
3. The PHP configuration is optimized for performance using eAccelerator caching and the Zend Guard Loader. Nginx, PHP-FPM, and MySQL systemd service scripts are also configured.
The document provides guidelines for deploying an L.N.M.P environment on a 64-bit server. It specifies directory locations for source code, installed software, scripts and logs. It also outlines steps to update the system, install and configure MySQL, Nginx, PHP and other packages, including compiling Nginx with specific modules and options, setting Nginx as a service, and enabling syntax highlighting for Nginx configuration files.
The document provides instructions for installing, configuring, and uninstalling the Apache HTTP Server on Linux systems. It discusses downloading and extracting the Apache files, using yum to install dependencies, configuring modules, and stopping or removing Apache using various commands. Graphical tools for configuring Apache like system-config-httpd are also mentioned.
The document provides instructions on Docker practice including prerequisites, basic Docker commands, running containers from images, committing container changes to new images, logging into Docker Hub and pushing images.
It begins with prerequisites of having Ubuntu 18.04 or higher and installing the latest Docker engine and Docker compose. It then explains that Docker runs processes in isolated containers and uses layered images.
The document demonstrates basic commands like docker version, docker images, docker pull, docker search, docker run, docker ps, docker stop, docker rm and docker rmi. It also shows how to commit container changes to a new image with docker commit, tag and push images to Docker Hub. Other topics covered include docker exec, docker save/load, docker
This document provides instructions and examples for properly configuring and using .htaccess files. It discusses Apache and PHP configuration as well as settings for performance, security, redirects, and more. Sections cover topics like error documents, rewriting URLs, compression, caching, and protecting files. The goal is to help optimize a WordPress site using the .htaccess file.
Virtualization and automation of library software/machines + PuppetOmar Reygaert
The document discusses virtualization, automation, and Puppet. It begins with an introduction to virtualization and hands-on labs. It then covers automation through kickstart files and preseeding to automate operating system installation. Hands-on labs are also provided for automation. Finally, it discusses Puppet for configuration management, including node definitions, modules, and resources to manipulate files, packages, users and more. Hands-on labs are presented for implementing SFX configuration with Puppet.
WordPress Home Server with Raspberry PiYuriko IKEDA
This document provides instructions for setting up a WordPress site on a Raspberry Pi server using a static IP address. It details steps such as installing necessary software like Apache, MySQL, PHP and WordPress using apt-get. It also covers configuring PHP, securing phpMyAdmin with HTTP authentication, creating a MySQL database and user, and generating security keys for WordPress. The document concludes with suggestions for additional configuration tasks and references for more information.
This document discusses the setup and architecture of a MySQL Spider configuration. It describes installing multiple MySQL instances on a single server, including 4 MariaDB 10.1 nodes and 6 MySQL 5.7 nodes. It then configures the Spider storage engine to partition tables across these nodes for both data sharding and high availability. Examples are provided of creating servers, tables, and inserting/selecting data to demonstrate the federated and sharded architecture.
The document discusses Nouka, an open source inventory management tool for Linux. Nouka consists of three parts - Nouka data collector, Naya data store, and Yaoya data converter. Nouka data collector runs commands periodically on Linux machines and sends the results to Naya data store. Naya uses Fluentd and MongoDB to store the collected data. Yaoya then converts and outputs the data in various formats like JSON, CSV for analysis. Overall, Nouka provides an automatic and periodic way to collect and centralize inventory data from Linux machines.
The document provides instructions for updating the FreeBSD 7.2 ports tree on a WebServer. The following steps are outlined:
1. Log in as the user "sermpan" and su to root.
2. Extract a backup of the ports tree files from /backups/distfiles72.tar.
3. Install and clean the cvsup port to update the ports tree files.
This document provides instructions for setting up a local repository, installing packages, configuring services like MySQL, PHP, Tomcat, and JavaBridge. It also covers setting up an FTP server, adding virtual hosts, optimizing PHP settings, installing VPN, and taking database backups with Percona XtraBackup. Key steps include disabling SELinux and firewall, creating the local yum repository, installing packages like PHP, MySQL, Tomcat from the repository, and configuring the services and virtual hosts.
This document provides guidance on deploying and upgrading a MongoDB sharded cluster. It discusses the components of a sharded cluster including config servers, shards, and mongos processes. It recommends a production deployment has at least 3 config servers, 3 nodes per shard replica set, and multiple mongos instances. The document outlines steps for deploying each component, including initializing replica sets and adding shards. It also provides a checklist for upgrading between minor and major versions, such as changes to configuration options, deprecated operations, and connectivity changes.
This document provides instructions for installing and configuring Cacti on openSUSE Leap 42.1. It describes getting root access, adding repositories, installing required packages like MySQL and Cacti, configuring services, securing MySQL, configuring Apache and file permissions, creating the Cacti database, configuring Cacti, and completing the installation process in a web browser.
This document discusses using LVS (Linux Virtual Server) and Keepalived to implement high performance and high availability load balancing. It provides instructions on installing LVS and Keepalived, configuring LVS for load balancing, and using Keepalived to realize load balancing and high availability. Key aspects covered include using LVS-DR and real server scripts, and configuring Keepalived on the master and backup servers.
Hadoop Admin role & Hive Data Warehouse supportmdcdwh
This document provides configuration details for a Hadoop cluster running various components including HDFS, YARN, Hive, HBase, Spark, and Sqoop. It outlines the Linux OS, Java, and Hadoop versions in use as well as configuration properties for core Hadoop systems. It also describes commands for common Hadoop administration tasks and how to configure high availability for HDFS, YARN, and HBase. Kerberos configuration is covered along with details on using Sqoop, Spark, Hive, and Cloudera Manager. The document concludes with sections on data analysis and visualization using Power BI.
MySQL is a widely used open-source relational database management system. The presentation covered how to install, configure, start, stop, and connect to MySQL. It also discussed how to load and view data, backup databases, set up user authentication, and where to go for additional training resources. Common MySQL commands and tools were demonstrated.
Installation of Subversion on Ubuntu,...wensheng wei
The document provides instructions for installing Subversion on Ubuntu with Apache, SSL, and BasicAuth to allow hosting SVN repositories on a web server, including installing necessary packages, configuring Apache with a SSL certificate and virtual host, creating repositories under /var/svn, setting up authentication using htpasswd, and enabling WebDAV and SVN support in Apache.
Ansible is a tool for configuration management and application deployment. It works by connecting to target systems using SSH and executing commands. Ansible has simple definitions for describing system configurations and states using YAML files and modules written in Python. Modules allow Ansible to assess system states, make changes idempotently to ensure systems match the defined states. Ansible is highly modular and has many contributors due to its architecture, examples, documentation and an active community.
This document discusses the Puppet configuration management tool. It provides an overview of Puppet including its open source nature, supported platforms, file structure, and types of resources it can manage like files, packages, services. It also discusses Facter for collecting system facts. Several examples are shown of how to configure files, packages, services. Finally Amazon EC2 is mentioned as a way to deploy Puppet in a scalable environment.
The document provides instructions for installing and configuring a full-featured Linux server for hosting websites, email, and other services. It includes steps to install and configure an ISPConfig control panel along with associated software like Apache, PHP, MySQL, Postfix, Dovecot, PureFTPd, BIND, Roundcube webmail, and more. The overall process allows you to set up an full-fledged web hosting server on a Linux machine.
agri inventory - nouka data collector / yaoya data convertorToshiaki Baba
This document provides instructions for setting up and using an agri inventory system called nouka and yaoya. Nouka collects data from servers using commands and sends it to the naya data store, which uses fluentd and MongoDB. It explains the components, data formats, and provides steps to get the required software and set up the system.
This document provides step-by-step instructions for configuring a LAMP stack on an Ubuntu 10.04 server including: installing Apache, MySQL, PHP; enabling modules; installing Drush and Drush Make; setting up Virtualmin for virtual hosting; and configuring security such as firewall rules and remote database access. It also includes links to additional resources for Drupal, Ubuntu, and LAMP configuration.
Site Performance - From Pinto to FerrariJoseph Scott
This document discusses ways to improve website performance from slow "Pinto" levels to faster "Ferrari" levels. It recommends using an opcode cache like APC to speed up PHP, optimizing databases with technologies like Memcached, using caching plugins in WordPress, and considering architectures with load balancing and replication. The goal is to identify bottlenecks and apply techniques that reduce page load times through an understanding of how websites and underlying technologies work.
This document provides instructions for installing Redmine, an open source project management tool, on a Debian Wheezy system. It describes how to install Ruby, Rails, gems, and the Passenger application server. It then covers downloading and configuring Redmine, creating a MySQL database, and configuring Apache virtual hosts to serve Redmine at a sub-URI. Testing the installation and troubleshooting tips are also included.
Ansible : what's ansible & use case by REXSaewoong Lee
The document discusses using Ansible to upgrade MySQL from version 3.10 to 3.12 across 1000 servers. It provides steps to create backups and run upgrade scripts on each server using Ansible playbooks and bash scripts in a loop. It also asks how to make the deployment easier, safer and more comfortable. Later sections explain Ansible concepts like installation, modules, playbooks, tags, inventory, variables and demonstrate usage through examples.
The document summarizes the steps taken to set up a Django project called "he" on Ubuntu. It shows commands used to install Python, virtualenv, Django and other dependencies. Database setup with PostgreSQL is also demonstrated. An app called "board" is created, with a Post model defined and admin configured. Templates are added and the development server is run. Authentication and registration are implemented along with forms to add new posts. The project is developed iteratively through multiple versions.
Using reader responses to make revisionswritRHET -
The document discusses using reader responses to strengthen arguments in a paper. It advises not ignoring counterarguments but instead addressing them to increase credibility with skeptical readers, make the main argument more complex, and develop a sturdier paper structure. The document then provides an example of a student argument for the DREAM Act and responses raising questions about costs to states and impact on domestic jobs. It shows how the student strengthened their argument by addressing these counterpoints and building a revised structure that transitions between economic and ethical considerations.
This document discusses the setup and architecture of a MySQL Spider configuration. It describes installing multiple MySQL instances on a single server, including 4 MariaDB 10.1 nodes and 6 MySQL 5.7 nodes. It then configures the Spider storage engine to partition tables across these nodes for both data sharding and high availability. Examples are provided of creating servers, tables, and inserting/selecting data to demonstrate the federated and sharded architecture.
The document discusses Nouka, an open source inventory management tool for Linux. Nouka consists of three parts - Nouka data collector, Naya data store, and Yaoya data converter. Nouka data collector runs commands periodically on Linux machines and sends the results to Naya data store. Naya uses Fluentd and MongoDB to store the collected data. Yaoya then converts and outputs the data in various formats like JSON, CSV for analysis. Overall, Nouka provides an automatic and periodic way to collect and centralize inventory data from Linux machines.
The document provides instructions for updating the FreeBSD 7.2 ports tree on a WebServer. The following steps are outlined:
1. Log in as the user "sermpan" and su to root.
2. Extract a backup of the ports tree files from /backups/distfiles72.tar.
3. Install and clean the cvsup port to update the ports tree files.
This document provides instructions for setting up a local repository, installing packages, configuring services like MySQL, PHP, Tomcat, and JavaBridge. It also covers setting up an FTP server, adding virtual hosts, optimizing PHP settings, installing VPN, and taking database backups with Percona XtraBackup. Key steps include disabling SELinux and firewall, creating the local yum repository, installing packages like PHP, MySQL, Tomcat from the repository, and configuring the services and virtual hosts.
This document provides guidance on deploying and upgrading a MongoDB sharded cluster. It discusses the components of a sharded cluster including config servers, shards, and mongos processes. It recommends a production deployment has at least 3 config servers, 3 nodes per shard replica set, and multiple mongos instances. The document outlines steps for deploying each component, including initializing replica sets and adding shards. It also provides a checklist for upgrading between minor and major versions, such as changes to configuration options, deprecated operations, and connectivity changes.
This document provides instructions for installing and configuring Cacti on openSUSE Leap 42.1. It describes getting root access, adding repositories, installing required packages like MySQL and Cacti, configuring services, securing MySQL, configuring Apache and file permissions, creating the Cacti database, configuring Cacti, and completing the installation process in a web browser.
This document discusses using LVS (Linux Virtual Server) and Keepalived to implement high performance and high availability load balancing. It provides instructions on installing LVS and Keepalived, configuring LVS for load balancing, and using Keepalived to realize load balancing and high availability. Key aspects covered include using LVS-DR and real server scripts, and configuring Keepalived on the master and backup servers.
Hadoop Admin role & Hive Data Warehouse supportmdcdwh
This document provides configuration details for a Hadoop cluster running various components including HDFS, YARN, Hive, HBase, Spark, and Sqoop. It outlines the Linux OS, Java, and Hadoop versions in use as well as configuration properties for core Hadoop systems. It also describes commands for common Hadoop administration tasks and how to configure high availability for HDFS, YARN, and HBase. Kerberos configuration is covered along with details on using Sqoop, Spark, Hive, and Cloudera Manager. The document concludes with sections on data analysis and visualization using Power BI.
MySQL is a widely used open-source relational database management system. The presentation covered how to install, configure, start, stop, and connect to MySQL. It also discussed how to load and view data, backup databases, set up user authentication, and where to go for additional training resources. Common MySQL commands and tools were demonstrated.
Installation of Subversion on Ubuntu,...wensheng wei
The document provides instructions for installing Subversion on Ubuntu with Apache, SSL, and BasicAuth to allow hosting SVN repositories on a web server, including installing necessary packages, configuring Apache with a SSL certificate and virtual host, creating repositories under /var/svn, setting up authentication using htpasswd, and enabling WebDAV and SVN support in Apache.
Ansible is a tool for configuration management and application deployment. It works by connecting to target systems using SSH and executing commands. Ansible has simple definitions for describing system configurations and states using YAML files and modules written in Python. Modules allow Ansible to assess system states, make changes idempotently to ensure systems match the defined states. Ansible is highly modular and has many contributors due to its architecture, examples, documentation and an active community.
This document discusses the Puppet configuration management tool. It provides an overview of Puppet including its open source nature, supported platforms, file structure, and types of resources it can manage like files, packages, services. It also discusses Facter for collecting system facts. Several examples are shown of how to configure files, packages, services. Finally Amazon EC2 is mentioned as a way to deploy Puppet in a scalable environment.
The document provides instructions for installing and configuring a full-featured Linux server for hosting websites, email, and other services. It includes steps to install and configure an ISPConfig control panel along with associated software like Apache, PHP, MySQL, Postfix, Dovecot, PureFTPd, BIND, Roundcube webmail, and more. The overall process allows you to set up an full-fledged web hosting server on a Linux machine.
agri inventory - nouka data collector / yaoya data convertorToshiaki Baba
This document provides instructions for setting up and using an agri inventory system called nouka and yaoya. Nouka collects data from servers using commands and sends it to the naya data store, which uses fluentd and MongoDB. It explains the components, data formats, and provides steps to get the required software and set up the system.
This document provides step-by-step instructions for configuring a LAMP stack on an Ubuntu 10.04 server including: installing Apache, MySQL, PHP; enabling modules; installing Drush and Drush Make; setting up Virtualmin for virtual hosting; and configuring security such as firewall rules and remote database access. It also includes links to additional resources for Drupal, Ubuntu, and LAMP configuration.
Site Performance - From Pinto to FerrariJoseph Scott
This document discusses ways to improve website performance from slow "Pinto" levels to faster "Ferrari" levels. It recommends using an opcode cache like APC to speed up PHP, optimizing databases with technologies like Memcached, using caching plugins in WordPress, and considering architectures with load balancing and replication. The goal is to identify bottlenecks and apply techniques that reduce page load times through an understanding of how websites and underlying technologies work.
This document provides instructions for installing Redmine, an open source project management tool, on a Debian Wheezy system. It describes how to install Ruby, Rails, gems, and the Passenger application server. It then covers downloading and configuring Redmine, creating a MySQL database, and configuring Apache virtual hosts to serve Redmine at a sub-URI. Testing the installation and troubleshooting tips are also included.
Ansible : what's ansible & use case by REXSaewoong Lee
The document discusses using Ansible to upgrade MySQL from version 3.10 to 3.12 across 1000 servers. It provides steps to create backups and run upgrade scripts on each server using Ansible playbooks and bash scripts in a loop. It also asks how to make the deployment easier, safer and more comfortable. Later sections explain Ansible concepts like installation, modules, playbooks, tags, inventory, variables and demonstrate usage through examples.
The document summarizes the steps taken to set up a Django project called "he" on Ubuntu. It shows commands used to install Python, virtualenv, Django and other dependencies. Database setup with PostgreSQL is also demonstrated. An app called "board" is created, with a Post model defined and admin configured. Templates are added and the development server is run. Authentication and registration are implemented along with forms to add new posts. The project is developed iteratively through multiple versions.
Using reader responses to make revisionswritRHET -
The document discusses using reader responses to strengthen arguments in a paper. It advises not ignoring counterarguments but instead addressing them to increase credibility with skeptical readers, make the main argument more complex, and develop a sturdier paper structure. The document then provides an example of a student argument for the DREAM Act and responses raising questions about costs to states and impact on domestic jobs. It shows how the student strengthened their argument by addressing these counterpoints and building a revised structure that transitions between economic and ethical considerations.
BuilHigh Performance Weibo Platform-Qcon2011Yiwei Ma
The document discusses building a high performance Weibo platform. It covers using MySQL with user_timeline indexing initially, and then integrating caching with Memcached and Redis. It also discusses using NoSQL databases like MongoDB and HBase. The key topics are data structures like JSON, binary formats, and case studies of architectures with one or multiple data centers and regions for timelines. The overall goal is to build scalable timelines and optimize performance.
The document provides steps to install and configure various monitoring tools including Cacti, MySQL, SNMP, RRDtool, and plugins on a Linux server. It includes instructions for downloading packages, compiling source code, configuring services, and testing the monitoring installations.
The document discusses Netflix's cloud architecture on Amazon Web Services (AWS). It aims to be faster, scalable, available and allow developers to work more productively. Some key points are moving from a central SQL database to distributed NoSQL stores, replacing sticky in-memory sessions with a shared cache, and optimizing for latency tolerance over chatty protocols. The architecture also focuses on layered service interfaces over tangled code and instrumenting services rather than code.
The document provides instructions on installing, configuring, and using the MySQL database on Linux. It explains how to download and install MySQL, set the root password, secure the installation, and start the MySQL daemon. Advanced SQL topics like unions, subqueries, and ranking functions are also briefly mentioned.
Drupaljam 2017 - Deploying Drupal 8 onto Hosted Kubernetes in Google CloudDropsolid
In this presentation I explain using video examples how kubernetes works and how this can be used to host your Drupal 7 or 8 site. There are obviously also gotcha's and I'd like to warn you to not use this in production until you've verified it
The following article is the best simplified steps that will help you install and configure LEMP stack. its written by one of the genius engineers or Rootgate.com
Hands on Docker - Launch your own LEMP or LAMP stack - SunshinePHPDana Luther
In this tutorial we will go over setting up a standard LEMP stack for development use and learn how to modify it to mimic your production/pre-production environments as closely as possible. We will go over how to switch from Nginx to Apache, upgrade PHP versions and introduce additional storage engines such as Redis to the equation. We'll also step through how to run both unit and acceptance suites using headless Selenium images in the stack. Leave here fully confident in knowing that whatever environment you get thrown into, you can replicate it and work in it comfortably.
The document discusses various stages in architecting an application for the cloud as it grows in scale and complexity.
Stage 1 involves a simple architecture suitable for startups with low overhead. Stage 2 adds redundancy as the business grows. Stage 3 requires the addition of load balancers and more servers as publicity increases load. Stage 4 requires database replication and partitioning as single databases can no longer handle the load. Later stages involve rearchitecting the application and databases to support further scaling through techniques like data partitioning, database clustering, and optimizing code and resources.
This document provides an overview of Ansible, an open source tool for configuration management and application deployment. It discusses how Ansible aims to simplify infrastructure automation tasks through a model-driven approach without requiring developers to learn DevOps tools. Key points:
- Ansible uses YAML playbooks to declaratively define server configurations and deployments in an idempotent and scalable way.
- It provides ad-hoc command execution and setup facts gathering via SSH. Playbooks can target groups of servers to orchestrate complex multi-server tasks.
- Variables, templates, conditionals allow playbooks to customize configurations for different environments. Plugins support integration with cloud, monitoring, messaging tools.
- Ansible aims to reduce complexity compared
This document provides instructions for installing and configuring Cacti on openSUSE 13.2. It involves installing required packages, creating a MySQL database for Cacti, importing the Cacti database, configuring Cacti settings, changing file permissions, adding a cron job, enabling and restarting services, and completing the Cacti installation wizard. Finally, the Cacti dashboard is shown and additional documentation is provided.
Caching and tuning fun for high scalabilityWim Godden
Caching has been a 'hot' topic for a few years. But caching takes more than merely taking data and putting it in a cache : the right caching techniques can improve performance and reduce load significantly. But we'll also look at some major pitfalls, showing that caching the wrong way can bring down your site. If you're looking for a clear explanation about various caching techniques and tools like Memcached, Nginx and Varnish, as well as ways to deploy them in an efficient way, this talk is for you.
This document discusses the use of Chef, an open source configuration management tool, for server management. It notes that Chef allows for repeatable system provisioning and ease of scaling servers without vendor lock-in. Chef manages over 120 servers across 10 environments for the company discussed. Chef uses Ruby code and resources like packages, templates and services to configure and maintain server configurations. It works both on single servers via chef-solo and with a centralized chef-server for cluster management. Common resources, attributes, definitions and recipes are discussed as the basic building blocks for automation with Chef. Gotchas around idempotency, package sources and attribute abuse are also covered.
The document describes how to configure and run multiple MySQL database instances on a single server. It involves:
1. Creating separate data directories and configuration files for each instance, configured to use different ports;
2. Installing the database tables into each data directory;
3. Starting each MySQL instance using its customized configuration file.
This allows running two instances on ports 3307 and 3308, each with its own isolated data and configuration. The server processes are started and the ports are verified to confirm the instances are running independently.
Walter gives an introduction to compiling PHP from source. Some key points covered include:
- Reasons for compiling PHP yourself include supporting old PHP versions, testing multiple versions, and using proprietary extensions.
- The build process involves running ./configure, make, and make install commands. Various configuration options can be specified.
- Popular PHP extensions like APC, Xdebug, PostgreSQL can be installed via PECL. Extensions are also compiled from source using phpize and make.
- The compiled PHP is typically installed to /usr/local/ and configurations made in the php.ini file.
This document discusses the setup and architecture of a MySQL Spider configuration. It describes installing multiple MySQL instances on a single server, including 4 MariaDB 10.1 nodes and 6 MySQL 5.7 nodes. It then configures the Spider storage engine to partition tables across these nodes for both data sharding and high availability. Key steps include defining servers, creating federated tables, and testing data is replicated correctly across nodes. The architecture provides both data sharding and high availability through table partitioning and failure recovery mechanisms.
Converting Your Dev Environment to a Docker Stack - php[world]Dana Luther
Heard a lot about docker but not sure where to start? Frustrated maintaining development VMs? In this presentation we will go over the simplest ways to convert your development environment over to a docker stack, including support for full acceptance testing with Selenium. We’ll then go over how to modify the stack to mimic your production/pre-production environment(s) as closely as possible, and demystify working with the containers in the stack.
The document discusses using Fabric for deployment and system administration tasks across multiple servers. It provides examples of Fabric configuration, defining roles for servers, writing tasks to run commands on servers, and how to structure tasks for a full deployment workflow. Fabric allows running commands remotely via SSH and provides tools for task composition and failure handling.
This document summarizes the key topics covered in Day 2 of a Docker and container technology introduction and hands-on course, including:
1) An overview of Docker Hub and how it relates to GitHub for automatically building images
2) Basic Git commands
3) Configuring automatic builds on Docker Hub by linking a GitHub repository
4) Docker network and volume commands, and exercises using these commands
5) Using Docker Compose to run multiple connected containers defined in a compose file
6) A demonstration of running TensorFlow using Docker
OpenStack Tokyo Meeup - Gluster Storage DayDan Radez
November 2012 Tokyo OpenStack meetup was dedicated to using Gluster storage. This presentation showed the fuse mount method to integrating gluster into OpenStack. There are new drivers that have been developed that make mounting gluster volumes to instances more efficient. This presentation doesn't show how to use them.
Converting Your Dev Environment to a Docker Stack - CascadiaDana Luther
The document discusses converting a development environment to a Docker stack. It describes some of the benefits of using Docker for consolidating different environments that may have varying PHP and MySQL versions. It provides an example docker-compose.yml file that defines services for Nginx, MySQL, and PHP-FPM containers along with networking and volume configurations. It also includes some sidebars on Docker concepts like the hierarchy of images, containers, services and stacks, as well as common Docker commands.
Code testing and Continuous Integration are just the first step in a source code to production process. Combined with infrastructure-as-code tools such as Puppet the whole process can be automated, and tested!
The document discusses different topics related to software development processes and tools. It provides information about Scrum methodology and roles like Product Manager and development teams. It also talks about version control tools like SVN and continuous integration tools like Hudson. Various software development concepts are explained like trunk-based development, feature flags, and deploying features to production in phases. Overall workflows involving coding, code reviews, testing and deploying software updates are described.
This document summarizes new features in Spring 3 and 3.1 for component-based application design. Spring 3 focuses on annotation-based components while also supporting concise XML configurations. Key features include stereotypes, factory methods, expression language support, standardized annotations, validation, formatting, scheduling, and REST support. Spring 3.1 enhances environments with profiles for bean definitions, enables Java-based configuration, adds a "c:" namespace for XML, and introduces declarative caching capabilities.
This document discusses Google's infrastructure and data centers. It describes Google's use of large data centers containing thousands of servers and petabytes of storage. It also summarizes Google's development of technologies like GFS, MapReduce, and BigTable to handle massive amounts of data across their infrastructure. Key details are provided on hardware specifications, network switches, reliability targets, and the engineers involved in developing Google's data-handling systems.
Netflix uses cloud computing to address challenges in scaling its infrastructure to support unpredictable growth. It has transitioned its website to be nearly 100% cloud-based using Amazon Web Services (AWS) to gain the scale, availability and agility needed. AWS provides tools and features like auto-scaling that allow Netflix to easily expand capacity as its subscriber base grows by over 50% per year. By leveraging AWS' mature cloud platform, Netflix can focus on its core video business rather than managing data centers.
This document summarizes a talk about Facebook's use of HBase for messaging data. It discusses how Facebook migrated data from MySQL to HBase to store metadata, search indexes, and small messages in HBase for improved scalability. It also outlines performance improvements made to HBase, such as for compactions and reads, and future plans such as cross-datacenter replication and running HBase in a multi-tenant environment.
The document discusses domain-driven design and modeling complex domains. It provides an example of modeling a shipping domain to understand cargo routing. Entities in the domain include Cargo, Itinerary, and Leg. A Cargo has an origin and destination. An Itinerary is generated by a Routing Service and consists of a series of Legs, where each Leg specifies a load and unload location for the Cargo. Modeling these concepts helps address routing needs like booking or rerouting shipments.
This document discusses how to create "big agility" by focusing on goals and outcomes rather than processes. It advocates questioning assumptions and continuously learning through experiments. Key points discussed include developing personas and story maps to understand users' needs, planning iterations to balance discovery, delivery and learning, and measuring real value delivered rather than effort spent. Cross-team collaboration and creating a shared understanding of what success means for stakeholders is also emphasized. The document provides examples of tools and practices for building agility within, across, and outside of teams.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
[OReilly Superstream] Occupy the Space: A grassroots guide to engineering (an...Jason Yip
The typical problem in product engineering is not bad strategy, so much as “no strategy”. This leads to confusion, lack of motivation, and incoherent action. The next time you look for a strategy and find an empty space, instead of waiting for it to be filled, I will show you how to fill it in yourself. If you’re wrong, it forces a correction. If you’re right, it helps create focus. I’ll share how I’ve approached this in the past, both what works and lessons for what didn’t work so well.
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
Discover top-tier mobile app development services, offering innovative solutions for iOS and Android. Enhance your business with custom, user-friendly mobile applications.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/how-axelera-ai-uses-digital-compute-in-memory-to-deliver-fast-and-energy-efficient-computer-vision-a-presentation-from-axelera-ai/
Bram Verhoef, Head of Machine Learning at Axelera AI, presents the “How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-efficient Computer Vision” tutorial at the May 2024 Embedded Vision Summit.
As artificial intelligence inference transitions from cloud environments to edge locations, computer vision applications achieve heightened responsiveness, reliability and privacy. This migration, however, introduces the challenge of operating within the stringent confines of resource constraints typical at the edge, including small form factors, low energy budgets and diminished memory and computational capacities. Axelera AI addresses these challenges through an innovative approach of performing digital computations within memory itself. This technique facilitates the realization of high-performance, energy-efficient and cost-effective computer vision capabilities at the thin and thick edge, extending the frontier of what is achievable with current technologies.
In this presentation, Verhoef unveils his company’s pioneering chip technology and demonstrates its capacity to deliver exceptional frames-per-second performance across a range of standard computer vision networks typical of applications in security, surveillance and the industrial sector. This shows that advanced computer vision can be accessible and efficient, even at the very edge of our technological ecosystem.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/temporal-event-neural-networks-a-more-efficient-alternative-to-the-transformer-a-presentation-from-brainchip/
Chris Jones, Director of Product Management at BrainChip , presents the “Temporal Event Neural Networks: A More Efficient Alternative to the Transformer” tutorial at the May 2024 Embedded Vision Summit.
The expansion of AI services necessitates enhanced computational capabilities on edge devices. Temporal Event Neural Networks (TENNs), developed by BrainChip, represent a novel and highly efficient state-space network. TENNs demonstrate exceptional proficiency in handling multi-dimensional streaming data, facilitating advancements in object detection, action recognition, speech enhancement and language model/sequence generation. Through the utilization of polynomial-based continuous convolutions, TENNs streamline models, expedite training processes and significantly diminish memory requirements, achieving notable reductions of up to 50x in parameters and 5,000x in energy consumption compared to prevailing methodologies like transformers.
Integration with BrainChip’s Akida neuromorphic hardware IP further enhances TENNs’ capabilities, enabling the realization of highly capable, portable and passively cooled edge devices. This presentation delves into the technical innovations underlying TENNs, presents real-world benchmarks, and elucidates how this cutting-edge approach is positioned to revolutionize edge AI across diverse applications.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.