This document discusses Linux text stream filters and provides examples of common Unix commands used to process and modify text streams. These commands include cat, head, tail, cut, and split. Cat prints the contents of files, head prints the first few lines, tail prints the last few lines, cut extracts parts of each line, and split divides files into smaller parts. The document also covers input/output redirection and how it can be used with text stream filters.
This document discusses various UNIX commands for processing text streams and filtering text, including cat, cut, head, tail, and split. It provides examples of using each command to select, modify, or restructure the output. The commands can be used to select parts of files like lines (head/tail) or columns (cut), and to split files into multiple parts (split). Pipelines and redirection are also covered.
This document provides an introduction and overview of Linux commands and Perl basics. It discusses key Linux commands for system information, user management, files/directories, permissions, processes, networking and more. It also covers Perl data types, variables, input/output, strings, arithmetic, comparisons, functions and file handling. The document aims to teach Linux commands and Perl programming basics.
This document provides an overview of Linux commands for redirecting standard input, output, and error streams. Some key points covered include:
- Redirecting input (<), output (> and >>), and pipes (|) to send output as input to another command
- Common redirection operators like tee to send output to both a file and stdout, and xargs to expand input to command line arguments
- Redirecting standard error (2>) independently from standard output using 2>, 2>>, 2>&1
- Using command substitution (`command` and $(command)) to capture output and use as arguments
- Chaining commands together in pipelines (|) to filter and transform text streams
- Examples of
This document provides an overview of common Linux commands used to process text streams and filter output, including cat, cut, head, tail, and split. It discusses how these commands can be used to select, sort, reformat, and summarize data by printing certain parts of files like columns, lines, or characters. Redirection is also covered as a way to modify command input and output. The goal is to explain the key knowledge areas and objectives for the Junior Level Linux Certification exam related to GNU and Unix commands.
This document provides an overview of various networking tools in Linux, including commands for network configuration (ifconfig, route), connectivity testing (ping, traceroute), name resolution (host, nslookup), port and protocol inspection (netstat, tcpdump), and secure remote access (SSH, PuTTY). It also covers tools for firewall management (ufw), network mapping (Nmap), raw socket programming (netcat), link status (ethtool), and more. Examples are given for common tasks like viewing routing tables, capturing packets, remotely controlling systems, and accessing services over Telnet versus SSH. A references section at the end provides additional learning resources.
This document provides summaries of Linux commands that can be used to select and manipulate parts of files. It discusses the cat, head, tail, cut, split, sort, tac, uniq, and tr commands. Cat concatenates or displays files, head displays the first few lines of a file, and tail displays the last few lines. Cut extracts columns from a file, split divides a file into smaller parts, and sort sorts lines alphabetically or numerically. Tac displays the lines of a file in reverse order and uniq removes duplicate lines from a sorted file. Tr translates characters within a file, such as converting uppercase to lowercase.
This document provides information about monitoring, controlling, and killing processes on Linux systems. It discusses running processes in the foreground and background, disconnecting processes from terminals using nohup and setsid, and sending signals to processes using commands like kill, jobs, fg, and bg. Monitoring tools like ps and top are also covered, along with common signals like SIGTERM. The overall objective is to teach examinees how to create, monitor and kill processes on Linux systems.
The document discusses Linux input/output redirection and piping. It explains how the shell creates new processes to handle redirection and piping. Standard streams like stdout and stderr can be redirected to files or other programs using operators like >, >>, |, and 2>&1. Pipes allow the output of one command to serve as input to another command. The tee command diverts a copy of the input stream to both a file and stdout.
This document discusses various UNIX commands for processing text streams and filtering text, including cat, cut, head, tail, and split. It provides examples of using each command to select, modify, or restructure the output. The commands can be used to select parts of files like lines (head/tail) or columns (cut), and to split files into multiple parts (split). Pipelines and redirection are also covered.
This document provides an introduction and overview of Linux commands and Perl basics. It discusses key Linux commands for system information, user management, files/directories, permissions, processes, networking and more. It also covers Perl data types, variables, input/output, strings, arithmetic, comparisons, functions and file handling. The document aims to teach Linux commands and Perl programming basics.
This document provides an overview of Linux commands for redirecting standard input, output, and error streams. Some key points covered include:
- Redirecting input (<), output (> and >>), and pipes (|) to send output as input to another command
- Common redirection operators like tee to send output to both a file and stdout, and xargs to expand input to command line arguments
- Redirecting standard error (2>) independently from standard output using 2>, 2>>, 2>&1
- Using command substitution (`command` and $(command)) to capture output and use as arguments
- Chaining commands together in pipelines (|) to filter and transform text streams
- Examples of
This document provides an overview of common Linux commands used to process text streams and filter output, including cat, cut, head, tail, and split. It discusses how these commands can be used to select, sort, reformat, and summarize data by printing certain parts of files like columns, lines, or characters. Redirection is also covered as a way to modify command input and output. The goal is to explain the key knowledge areas and objectives for the Junior Level Linux Certification exam related to GNU and Unix commands.
This document provides an overview of various networking tools in Linux, including commands for network configuration (ifconfig, route), connectivity testing (ping, traceroute), name resolution (host, nslookup), port and protocol inspection (netstat, tcpdump), and secure remote access (SSH, PuTTY). It also covers tools for firewall management (ufw), network mapping (Nmap), raw socket programming (netcat), link status (ethtool), and more. Examples are given for common tasks like viewing routing tables, capturing packets, remotely controlling systems, and accessing services over Telnet versus SSH. A references section at the end provides additional learning resources.
This document provides summaries of Linux commands that can be used to select and manipulate parts of files. It discusses the cat, head, tail, cut, split, sort, tac, uniq, and tr commands. Cat concatenates or displays files, head displays the first few lines of a file, and tail displays the last few lines. Cut extracts columns from a file, split divides a file into smaller parts, and sort sorts lines alphabetically or numerically. Tac displays the lines of a file in reverse order and uniq removes duplicate lines from a sorted file. Tr translates characters within a file, such as converting uppercase to lowercase.
This document provides information about monitoring, controlling, and killing processes on Linux systems. It discusses running processes in the foreground and background, disconnecting processes from terminals using nohup and setsid, and sending signals to processes using commands like kill, jobs, fg, and bg. Monitoring tools like ps and top are also covered, along with common signals like SIGTERM. The overall objective is to teach examinees how to create, monitor and kill processes on Linux systems.
The document discusses Linux input/output redirection and piping. It explains how the shell creates new processes to handle redirection and piping. Standard streams like stdout and stderr can be redirected to files or other programs using operators like >, >>, |, and 2>&1. Pipes allow the output of one command to serve as input to another command. The tee command diverts a copy of the input stream to both a file and stdout.
The document discusses network layering models and TCP/IP fundamentals. It describes:
1. Networking problems are divided into layers for easier understanding and standardization, with the two main models being OSI and TCP/IP.
2. The TCP/IP model has four or five layers - process, host-to-host transport, internet, network access, and sometimes physical.
3. Packets are encapsulated as they leave a machine and decapsulated on the receiving host, with each layer adding headers.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, and ls. The examples cover a wide range of tasks from compressing/extracting files to searching/editing text to managing processes and permissions.
This document provides 50 examples of common Linux/Unix commands organized by command name. It begins with tar, grep, find, ssh, sed, awk, vim, diff, sort, and export examples. The document is intended as a quick reference for users to learn practical uses of fundamental Linux commands.
The document provides an overview of Linux operating system concepts including:
- Linux is an open source operating system that interacts with hardware and allocates resources.
- It supports multi-tasking and multi-user environments. Common types include Debian, Ubuntu, and Redhat.
- Key components include the kernel, shell programs, file management commands, text editors, browsers, and programming tools.
The document discusses Linux commands for redirecting standard input, output, and error streams. It covers commands like tee, xargs, <, >, >>, |, and backticks. Pipes and redirects allow commands to be chained together and output sent to files or other programs. Specific redirection examples are provided like redirecting just errors, appending output, and piping output between commands.
Linux is an open-source operating system that can run on various hardware. The document discusses various Linux commands and concepts related to directories, files, permissions, users, groups, text editors like vi and vim, process management, disk partitioning and more. It also covers Linux installation, package management, shell scripting and configuring network and services like SSH, web servers and more. Exercises are included to help understand concepts like mount points, journaling and file attributes.
The document discusses various data structures and functions related to network packet processing in the Linux kernel socket layer. It describes the sk_buff structure that is used to pass packets between layers. It also explains the net_device structure that represents a network interface in the kernel. When a packet is received, the interrupt handler will raise a soft IRQ for processing. The packet will then traverse various protocol layers like IP and TCP to be eventually delivered to a socket and read by a userspace application.
Networking in Linux discusses DNS related commands in Linux. It begins by listing DNS concepts like zones and records. It then demonstrates commands like nslookup, host and dig to query DNS records like A, MX, NS, SOA records and perform operations like reverse lookups. It shows how to use specific nameservers, change ports and timeouts. The document provides examples of using these tools to troubleshoot DNS issues like propagation.
This document provides information about creating partitions and filesystems in Linux. It discusses various Linux filesystem types like ext2, ext3, xfs, reiserfs v3, and vfat. It covers the commands and tools used to create partitions (fdisk, mkfs), filesystems (mkfs), and swap spaces (mkswap, swapon). It also discusses viewing filesystem information, mounting filesystems, and the Filesystem Hierarchy Standard for directory structure in Linux.
This document provides an overview of basic Linux commands for navigation, listing directories, reading and manipulating files. It explains commands like pwd, cd, ls, cat, cp, mv, rm, mkdir to change directories, list files, read files, copy, move and delete files/directories. It also introduces the vi editor for creating new files and mentions some other miscellaneous commands like date, chmod, user management tools.
This document discusses Linux input/output redirection and piping commands. It covers redirecting standard input, output, and error to files or other programs. Specific redirection operators like >, >>, |, and < are explained along with examples. Commands like tee and xargs that are used to manage input/output streams are also overviewed. The key techniques covered include redirecting streams, piping outputs between commands, and using filters to process text streams.
This lecture discusses a group of techniques to use commands output/Input to feed into other commands or into files. It also covers argument expansion and quoting
Check the other Lectures and courses in
http://Linux4EnbeddedSystems.com
or Follow our Facebook Group at
- Facebook: @LinuxforEmbeddedSystems
Lecturer Profile:
- https://www.linkedin.com/in/ahmedelarabawy
The document logs the startup process of a messaging client. It records the loading of configuration files and plugins, the establishment of network connections, and the initial authentication process with the messaging service. Errors are encountered when loading some plugins due to missing dependencies. The client then successfully connects to the messaging service and authenticates the user account.
This document provides an overview of common Linux commands organized into categories such as file handling, process management, disk usage, searching, editing, and remote access. It describes the purpose and basic usage of commands like vi, ls, grep, find, scp, and more. Examples are given for piping output, redirection, appending, and searching/replacing text within files. Links to additional online resources on Linux basics and tips are also included.
This document provides an overview and introduction to using the command line interface and submitting jobs to the NIAID High Performance Computing (HPC) Cluster. The objectives are to learn basic Unix commands, practice file manipulation from the command line, and submit a job to the HPC cluster. The document covers topics such as the anatomy of the terminal, navigating directories, common commands, tips for using the command line more efficiently, accessing and mounting drives on the HPC cluster, and an overview of the cluster queue system.
Efficient System Monitoring in Cloud Native EnvironmentsGergely Szabó
This document discusses efficient system monitoring in cloud native environments using eBPF. It provides an overview of eBPF and how it can be used for monitoring applications like Prometheus. Specific topics covered include BPF, Linux kernel tracing using kprobes and tracepoints, eBPF maps and programs, and an example Prometheus exporter that leverages eBPF to export metrics.
Server is a machine configured to accept requests from clients and respond accordingly. Linux is commonly used for servers, with distributions like Ubuntu, Redhat, and Debian. Key principles of Linux include treating everything as a file, storing configuration data in text files, and using pipes to connect programs. Common server files include /etc/group, /etc/passwd, and /etc/shadow which contain user and group information.
This document provides commands for basic file management and system utilities in Linux/Unix systems. It includes commands for listing, moving, copying, deleting and changing permissions of files and directories. It also includes commands for editing files, finding files, archiving files, printing files, managing processes, debugging programs, I/O redirection and setting environment variables.
Linux is a prominent example of free and open source software. It can be installed on a wide variety of devices from embedded systems to supercomputers. Linux is commonly used for servers, with estimates that it powers around 60% of web servers. Linux distributions package the Linux kernel with other software like utilities, libraries and desktop environments. Programming languages and build tools like GCC are supported. Embedded Linux is often used in devices due to its low cost and ease of modification.
This document provides an overview of shell scripting, cron, and atd for automating tasks in Linux. It discusses using shell scripts to execute a series of commands from a file, as well as topics like shebangs, variables, and conditionals. Cron allows scheduling commands to run on a time-based schedule using crontab files, while atd can run one-off commands at a specific time. The document provides examples and explanations of common cron and atd utilities.
This document provides an overview of shell scripting, cron, and atd for scheduling tasks on Linux systems. It discusses using shell scripts to automate commands, the cron daemon for executing commands on a schedule, and atd for running one-time jobs. Key points covered include using the shebang to directly execute shell scripts, crontab syntax for defining schedules, and commands like crond, at, atq and atrm for managing cron and at jobs. The document is intended as a learning resource for system administration tasks using scheduling utilities in Linux.
The document discusses network layering models and TCP/IP fundamentals. It describes:
1. Networking problems are divided into layers for easier understanding and standardization, with the two main models being OSI and TCP/IP.
2. The TCP/IP model has four or five layers - process, host-to-host transport, internet, network access, and sometimes physical.
3. Packets are encapsulated as they leave a machine and decapsulated on the receiving host, with each layer adding headers.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, and ls. The examples cover a wide range of tasks from compressing/extracting files to searching/editing text to managing processes and permissions.
This document provides 50 examples of common Linux/Unix commands organized by command name. It begins with tar, grep, find, ssh, sed, awk, vim, diff, sort, and export examples. The document is intended as a quick reference for users to learn practical uses of fundamental Linux commands.
The document provides an overview of Linux operating system concepts including:
- Linux is an open source operating system that interacts with hardware and allocates resources.
- It supports multi-tasking and multi-user environments. Common types include Debian, Ubuntu, and Redhat.
- Key components include the kernel, shell programs, file management commands, text editors, browsers, and programming tools.
The document discusses Linux commands for redirecting standard input, output, and error streams. It covers commands like tee, xargs, <, >, >>, |, and backticks. Pipes and redirects allow commands to be chained together and output sent to files or other programs. Specific redirection examples are provided like redirecting just errors, appending output, and piping output between commands.
Linux is an open-source operating system that can run on various hardware. The document discusses various Linux commands and concepts related to directories, files, permissions, users, groups, text editors like vi and vim, process management, disk partitioning and more. It also covers Linux installation, package management, shell scripting and configuring network and services like SSH, web servers and more. Exercises are included to help understand concepts like mount points, journaling and file attributes.
The document discusses various data structures and functions related to network packet processing in the Linux kernel socket layer. It describes the sk_buff structure that is used to pass packets between layers. It also explains the net_device structure that represents a network interface in the kernel. When a packet is received, the interrupt handler will raise a soft IRQ for processing. The packet will then traverse various protocol layers like IP and TCP to be eventually delivered to a socket and read by a userspace application.
Networking in Linux discusses DNS related commands in Linux. It begins by listing DNS concepts like zones and records. It then demonstrates commands like nslookup, host and dig to query DNS records like A, MX, NS, SOA records and perform operations like reverse lookups. It shows how to use specific nameservers, change ports and timeouts. The document provides examples of using these tools to troubleshoot DNS issues like propagation.
This document provides information about creating partitions and filesystems in Linux. It discusses various Linux filesystem types like ext2, ext3, xfs, reiserfs v3, and vfat. It covers the commands and tools used to create partitions (fdisk, mkfs), filesystems (mkfs), and swap spaces (mkswap, swapon). It also discusses viewing filesystem information, mounting filesystems, and the Filesystem Hierarchy Standard for directory structure in Linux.
This document provides an overview of basic Linux commands for navigation, listing directories, reading and manipulating files. It explains commands like pwd, cd, ls, cat, cp, mv, rm, mkdir to change directories, list files, read files, copy, move and delete files/directories. It also introduces the vi editor for creating new files and mentions some other miscellaneous commands like date, chmod, user management tools.
This document discusses Linux input/output redirection and piping commands. It covers redirecting standard input, output, and error to files or other programs. Specific redirection operators like >, >>, |, and < are explained along with examples. Commands like tee and xargs that are used to manage input/output streams are also overviewed. The key techniques covered include redirecting streams, piping outputs between commands, and using filters to process text streams.
This lecture discusses a group of techniques to use commands output/Input to feed into other commands or into files. It also covers argument expansion and quoting
Check the other Lectures and courses in
http://Linux4EnbeddedSystems.com
or Follow our Facebook Group at
- Facebook: @LinuxforEmbeddedSystems
Lecturer Profile:
- https://www.linkedin.com/in/ahmedelarabawy
The document logs the startup process of a messaging client. It records the loading of configuration files and plugins, the establishment of network connections, and the initial authentication process with the messaging service. Errors are encountered when loading some plugins due to missing dependencies. The client then successfully connects to the messaging service and authenticates the user account.
This document provides an overview of common Linux commands organized into categories such as file handling, process management, disk usage, searching, editing, and remote access. It describes the purpose and basic usage of commands like vi, ls, grep, find, scp, and more. Examples are given for piping output, redirection, appending, and searching/replacing text within files. Links to additional online resources on Linux basics and tips are also included.
This document provides an overview and introduction to using the command line interface and submitting jobs to the NIAID High Performance Computing (HPC) Cluster. The objectives are to learn basic Unix commands, practice file manipulation from the command line, and submit a job to the HPC cluster. The document covers topics such as the anatomy of the terminal, navigating directories, common commands, tips for using the command line more efficiently, accessing and mounting drives on the HPC cluster, and an overview of the cluster queue system.
Efficient System Monitoring in Cloud Native EnvironmentsGergely Szabó
This document discusses efficient system monitoring in cloud native environments using eBPF. It provides an overview of eBPF and how it can be used for monitoring applications like Prometheus. Specific topics covered include BPF, Linux kernel tracing using kprobes and tracepoints, eBPF maps and programs, and an example Prometheus exporter that leverages eBPF to export metrics.
Server is a machine configured to accept requests from clients and respond accordingly. Linux is commonly used for servers, with distributions like Ubuntu, Redhat, and Debian. Key principles of Linux include treating everything as a file, storing configuration data in text files, and using pipes to connect programs. Common server files include /etc/group, /etc/passwd, and /etc/shadow which contain user and group information.
This document provides commands for basic file management and system utilities in Linux/Unix systems. It includes commands for listing, moving, copying, deleting and changing permissions of files and directories. It also includes commands for editing files, finding files, archiving files, printing files, managing processes, debugging programs, I/O redirection and setting environment variables.
Linux is a prominent example of free and open source software. It can be installed on a wide variety of devices from embedded systems to supercomputers. Linux is commonly used for servers, with estimates that it powers around 60% of web servers. Linux distributions package the Linux kernel with other software like utilities, libraries and desktop environments. Programming languages and build tools like GCC are supported. Embedded Linux is often used in devices due to its low cost and ease of modification.
This document provides an overview of shell scripting, cron, and atd for automating tasks in Linux. It discusses using shell scripts to execute a series of commands from a file, as well as topics like shebangs, variables, and conditionals. Cron allows scheduling commands to run on a time-based schedule using crontab files, while atd can run one-off commands at a specific time. The document provides examples and explanations of common cron and atd utilities.
This document provides an overview of shell scripting, cron, and atd for scheduling tasks on Linux systems. It discusses using shell scripts to automate commands, the cron daemon for executing commands on a schedule, and atd for running one-time jobs. Key points covered include using the shebang to directly execute shell scripts, crontab syntax for defining schedules, and commands like crond, at, atq and atrm for managing cron and at jobs. The document is intended as a learning resource for system administration tasks using scheduling utilities in Linux.
The document provides an overview of Unix and shell scripting. It discusses the history and architecture of Unix operating systems. It then covers various Unix commands and utilities for file management, processes, communication, and system administration. Finally, it describes the basics of shell scripting including variables, conditional statements, loops, and here documents.
Linux is an open-source alternative to Microsoft Windows that is freely available and reliable. It has a command line interface and features like virtual memory, networking capabilities, multiple users, and graphical user interfaces. Common Linux commands include ls, cd, mkdir, rmdir, cat, cp, and editors like vi, emacs, nano are used to create and edit files. The Linux directory structure is hierarchical with key directories being /, /home, /usr, /var, and editors allow editing files in different modes like insert and command modes.
This document provides a cheat sheet of common Linux commands and their usage. It covers basic file operations like copying, moving, deleting files and directories. It also includes commands for viewing files, compressing/decompressing files, finding files, remote access, and getting system information. The commands are explained over 3 pages with examples of proper syntax and usage for each one.
This document provides instructions for 27 common Linux commands: mkdir, rmdir, ls, cd, cat, touch, wc, who, pwd, rm, mv, chmod, cp, grep, cal, date, vi, tput, ps, export, type, tail, sudo, head, man, clear, and adduser. For each command, it lists the syntax and provides 1-3 examples of common uses. The document is presented over 28 pages with the commands organized topic-by-topic and includes formatting like headings and indentation to aid readability.
Linux is an open-source operating system that can be used as a server or standalone OS. It offers advantages like stability, robustness, security, and high performance. Linux directories include / for the root directory, /bin for essential programs, /home for user files, and /var for variable data. Common Linux commands include ls to list files, cd to change directories, grep to search files, and vi/emacs for text editing. Cron jobs allow scheduling commands to run periodically using a crontab file.
This document provides an introduction to Linux, including:
- What Linux is and where it came from, as an open source operating system created by Linus Torvalds in 1991.
- How to get Linux through downloading a distribution like Red Hat and the packages it includes.
- Why Linux has become significant due to its growing popularity, ability to run on multiple hardware platforms, and being free.
- An overview of Linux commands and utilities for file management, processes, users and permissions.
A brief talk on systems performance for the July 2013 meetup "A Midsummer Night's System", video: http://www.youtube.com/watch?v=P3SGzykDE4Q. This summarizes how systems performance has changed from the 1990's to today. This was the reason for writing a new book on systems performance, to provide a reference that is up to date, covering new tools, technologies, and methodologies.
You have a system with an advanced programmatic tracer: do you know what to do with it? Brendan has used numerous tracers in production environments, and has published hundreds of tracing-based tools. In this talk he will share tips and know-how for creating CLI tracing tools and GUI visualizations, to solve real problems effectively. Programmatic tracing is an amazing superpower, and this talk will show you how to wield it!
Linux is a free, open-source operating system that provides functionality similar to Unix systems. The document then lists and describes several common Linux commands used to navigate directories, view files and system information, manage processes, and edit text files. It also includes a diagram of the standard Linux directory structure and brief descriptions of the purpose of each top-level directory.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, export, xargs, ls, pwd, cd, gzip, bzip2, unzip, shutdown, ftp, crontab, service, ps, top, df, kill, rm, cp, mv, cat, mount, chmod, chown, passwd, mkdir, ifconfig, and uname. The document is intended to give readers a quick start on frequently used commands.
This document provides descriptions of common Linux commands organized into categories including file manipulation, text processing, documentation, editors, file and directory management, file display and printing, programming tools, communications, process control, status information, and web commands. It describes what each command is used for and provides examples of some commonly used commands and their functions.
Linux Survival Kit for Proof of Concept & Proof of TechnologyNugroho Gito
The document provides a survival guide and cheat sheets for working efficiently in a Linux/UNIX environment while doing proof of concept or prototype work, covering topics such as working in the shell, storage and disk management using tools like LVM, text processing utilities, and configuration files that are important to understand. It also includes shortcuts for common tasks like adding a disk without restarting or creating an offline yum package repository for dependency resolution.
An operating system acts as an interface between hardware and software, managing resources and presenting an easier programming interface than the underlying hardware. The UNIX system is organized into three levels - the kernel which manages tasks and storage, the shell which connects programs and executes commands, and tools/applications which provide additional functionality. PuTTY is a free and open-source terminal emulator and SSH client most commonly used to access UNIX/Linux systems remotely from Windows. The course objectives are to learn UNIX commands, shell programming, process management, memory management, file organization, and implementing related algorithms.
This document discusses Linux file systems and creating partitions and filesystems in Linux. It covers the following key points:
1. Linux supports various filesystems like ext2, ext3, xfs, and ReiserFS that can be created using mkfs. Swap spaces are created with mkswap.
2. Partitions and filesystems can be created using tools like fdisk, cfdisk, and gpart. Filesystem types include ext3, xfs, FAT, etc.
3. The Filesystem Hierarchy Standard defines the directory structure and recommended layout of files on Linux systems with directories like /bin, /etc, /home, /usr, /var, etc.
This document summarizes options for running the Odoo server from the command line, including:
- Specifying database, modules, and module paths for installation/updating
- Configuring workers for multiprocessing and limiting resources per worker
- Setting logging options like the log file, database, or specific log levels
- Advanced options like auto-reloading or enabling a proxy
Similar to 101 3.2 process text streams using filters (20)
This document discusses the differences between vulnerability scanning and penetration testing. Vulnerability scanning uses automated tools to identify vulnerabilities while penetration testing involves manual techniques to simulate real-world attacks. The document also notes that vulnerability scanning and penetration testing have different levels of testing and are covered in Topic 24 of the Security+ Study Guide for 10 minutes.
This document discusses application security controls and techniques. It covers secure coding concepts, other security controls and techniques, and provides a 10 minute overview of application security controls and techniques as outlined in the CompTIA Security+ Study Guide. The document was prepared by Acácio Oliveira in 2021.
This document discusses types of application attacks and is divided into two sections. The first section defines application attacks and covers common types. The second section explains the goals of application attacks and how they can divulge weaknesses in some applications. The entire document is part of a Security+ study guide and covers types of application attacks over a period of 20 minutes.
Security+ Lesson 01 Topic 19 - Summary of Social Engineering Attacks.pptxAcácio Oliveira
This document provides a summary of social engineering attacks. Social engineering attacks are effective because they target human nature. There are different types of social engineering attacks, such as pretexting where attackers pretend to be someone trustworthy to get victims to reveal sensitive information. The document is part of a study guide for the CompTIA Security+ certification and focuses on summarizing social engineering attacks.
This document discusses security assessment tools, including types of security assessments and specific assessment tools. It is part of a CompTIA Security+ study guide prepared in 2021, focusing on an overview of security assessment tools in about 10 minutes.
This document provides a 10 minute summary of wireless attacks. It discusses common types of wireless attacks and attacks on wireless encryption. The document is part of a study guide for the CompTIA Security+ certification and is prepared by Acácio Oliveira in 2021.
This document discusses security enhancement techniques for network security. It covers the difference between detection controls and prevention controls. The topic is part of the Security+ Study Guide Concepts prepared by Acácio Oliveira in 2021 and focuses on network security enhancement techniques for 10 minutes as part of the study guide's Topic 22 on security enhancement techniques.
This document discusses risk management best practices for the CompTIA Security+ exam, covering business continuity concepts, fault tolerance, and is part of a study guide prepared by Acácio Oliveira in 2021 with an estimated reading time of 10 minutes.
This document discusses physical security and environmental controls, which are important concepts for the CompTIA Security+ exam. It covers control types for physical security and the environment, such as restricting physical access and monitoring for issues like fires, floods or power outages. The document is from a study guide prepared in 2021 to help study for the Security+ exam.
This document discusses disaster recovery concepts for the CompTIA Security+ exam, including maintaining disaster recovery sites, performing regular data backups, and ensuring the ability to recover data and systems in the event of a disaster within 10 minutes of scheduled content.
This document discusses wireless security considerations for the CompTIA Security+ exam. It covers the unique security challenges of wireless networks compared to wired networks. Specific topics include security for wireless networks and an index of study guide topics related to wireless security considerations with an allocated time of 10 minutes.
Security+ Lesson 01 Topic 04 - Secure Network Design Elements and Components....Acácio Oliveira
This document discusses secure network design elements and components. It covers defense in depth as a strategy as well as elements and components of network design such as firewalls, intrusion detection systems, and virtual private networks. The document is from Acácio Oliveira's 2021 CompTIA Security+ Study Guide and focuses on secure network design for 10 minutes.
This document provides a 3-topic outline for a CompTIA Security+ study guide prepared by Acácio Oliveira in 2021. Topic 1 discusses secure network administration concepts like spam filters and network devices. Topic 2 is titled "Secure Network Administration Concepts" and is allocated 10 minutes of study time. The document also includes a topics index for the study guide.
This document provides an overview of the concepts to be covered in Lesson 01 on introduction to network devices from the CompTIA Security+ Study Guide. The lesson will be divided into three parts covering OSI models, basic network devices and layer security concepts in part 1; network devices in part 2; and spam filters and additional network devices in part 3, for a total time of 30 minutes.
Security+ Lesson 01 Topic 08 - Integrating Data and Systems with Third Partie...Acácio Oliveira
This document discusses integrating data and systems with third parties. It examines evaluating risks with integration and considerations for integration. The document is part of a Security+ study guide prepared by Acácio Oliveira in 2021 on the topic of integrating data and systems with third parties.
Security+ Lesson 01 Topic 07 - Risk Related Concepts.pptxAcácio Oliveira
This document outlines concepts related to risk for the CompTIA Security+ certification. It covers three topics on risk-related concepts over 30 minutes: policies for controlling and reducing risk, qualitative versus quantitative risk assessments, and treating risks such as those involving cloud computing and virtualization. The document was prepared by Acácio Oliveira for Security+ study.
Security+ Lesson 01 Topic 05 - Common Network Protocols.pptxAcácio Oliveira
This document provides an overview of the topics to be covered in a 30-minute Common Network Protocols lesson for CompTIA Security+ certification preparation. The lesson will be divided into three parts covering IPv4 and IPv6, network storage protocols, the difference between ports and protocols, common protocols, and end-to-end security.
This document provides a 10 minute overview of incident response concepts for Security+ certification preparation. It discusses first responder responsibilities during security incidents and outlines incident response procedures and concepts covered in the Security+ study guide. The document is titled "Incident Response Concepts" and was prepared by Acácio Oliveira in 2021 as part of a Security+ study guide.
Security+ Lesson 01 Topic 12 - Security Related Awareness and Training.pptxAcácio Oliveira
This document discusses security related awareness and training. It covers the security policy, security awareness, and is part of a study guide for the CompTIA Security+ exam. The topic is allocated 10 minutes as part of the study guide index.
Security+ Lesson 01 Topic 17 - Types of Malware.pptxAcácio Oliveira
Malware is defined and common types are discussed including viruses, worms, Trojans, ransomware, and spyware. The document provides an overview of different types of malware for 10 minutes as part of a CompTIA Security+ study guide on the topic of types of malware.
Dive into the realm of operating systems (OS) with Pravash Chandra Das, a seasoned Digital Forensic Analyst, as your guide. 🚀 This comprehensive presentation illuminates the core concepts, types, and evolution of OS, essential for understanding modern computing landscapes.
Beginning with the foundational definition, Das clarifies the pivotal role of OS as system software orchestrating hardware resources, software applications, and user interactions. Through succinct descriptions, he delineates the diverse types of OS, from single-user, single-task environments like early MS-DOS iterations, to multi-user, multi-tasking systems exemplified by modern Linux distributions.
Crucial components like the kernel and shell are dissected, highlighting their indispensable functions in resource management and user interface interaction. Das elucidates how the kernel acts as the central nervous system, orchestrating process scheduling, memory allocation, and device management. Meanwhile, the shell serves as the gateway for user commands, bridging the gap between human input and machine execution. 💻
The narrative then shifts to a captivating exploration of prominent desktop OSs, Windows, macOS, and Linux. Windows, with its globally ubiquitous presence and user-friendly interface, emerges as a cornerstone in personal computing history. macOS, lauded for its sleek design and seamless integration with Apple's ecosystem, stands as a beacon of stability and creativity. Linux, an open-source marvel, offers unparalleled flexibility and security, revolutionizing the computing landscape. 🖥️
Moving to the realm of mobile devices, Das unravels the dominance of Android and iOS. Android's open-source ethos fosters a vibrant ecosystem of customization and innovation, while iOS boasts a seamless user experience and robust security infrastructure. Meanwhile, discontinued platforms like Symbian and Palm OS evoke nostalgia for their pioneering roles in the smartphone revolution.
The journey concludes with a reflection on the ever-evolving landscape of OS, underscored by the emergence of real-time operating systems (RTOS) and the persistent quest for innovation and efficiency. As technology continues to shape our world, understanding the foundations and evolution of operating systems remains paramount. Join Pravash Chandra Das on this illuminating journey through the heart of computing. 🌟
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
leewayhertz.com-AI in predictive maintenance Use cases technologies benefits ...alexjohnson7307
Predictive maintenance is a proactive approach that anticipates equipment failures before they happen. At the forefront of this innovative strategy is Artificial Intelligence (AI), which brings unprecedented precision and efficiency. AI in predictive maintenance is transforming industries by reducing downtime, minimizing costs, and enhancing productivity.
Skybuffer AI: Advanced Conversational and Generative AI Solution on SAP Busin...Tatiana Kojar
Skybuffer AI, built on the robust SAP Business Technology Platform (SAP BTP), is the latest and most advanced version of our AI development, reaffirming our commitment to delivering top-tier AI solutions. Skybuffer AI harnesses all the innovative capabilities of the SAP BTP in the AI domain, from Conversational AI to cutting-edge Generative AI and Retrieval-Augmented Generation (RAG). It also helps SAP customers safeguard their investments into SAP Conversational AI and ensure a seamless, one-click transition to SAP Business AI.
With Skybuffer AI, various AI models can be integrated into a single communication channel such as Microsoft Teams. This integration empowers business users with insights drawn from SAP backend systems, enterprise documents, and the expansive knowledge of Generative AI. And the best part of it is that it is all managed through our intuitive no-code Action Server interface, requiring no extensive coding knowledge and making the advanced AI accessible to more users.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
4. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Rundown of commands for Process text steams using filters:
commands Overview
4
• cat – concatenate files (or just show a single file without alteration)
• cut – cut chosen text out of each line of a file and display it.
• expand – expand tabs into spaces
• fmt – reformat a text file with a consistent right margin
• head – show the first few (10) lines of a file
• join – join lines of two files on a common field
• nl – print the file with numbered lines
• od – octal dump of a file (or hexadecimal).
• paste – print a number of files side by side
• pr – format for printing (split into pages or columns and add headers)
• sed – stream editor (search and replace, append, cut, delete and more)
• sort – sort in alphabetical order (and numerical order too)
• split – split a single input into multiple files
• tac – print the lines of a file from back to front (backwards cat)
• tail – print the last few lines of a file
• tr – character translation (e.g. upper case to lower case).
• unexpand – convert spaces to tabs (unlike expand).
• uniq – remove duplicate lines from a sorted file
• wc – word count (and line count, and byte count)
5. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Bash makes it possible to redirect the input and output of a command.
•Input - comes from the keyboard (ended by pressing Ctrl+D),
•Output and any errors - are displayed on the screen.
Redirection can change the input of a process, its output and the destination of the errors.
Input and output redirection
5
Redirection Effect of redirection
cmd < file Command reads input from a file
cmd > file Output of command goes to file
cmd 2> file Errors from the command go to a file
cmd >> file Output of a command is added to a file
cmd > file 2>&1 Output and Errors go to a file
cmd >& file
cmd &> file
cmd1 | cmd2 Output from command1 is input for command2
9. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
head – print out the first lines
Selecting parts of a file (Filters that print out various parts of the input they receive)
9
foo:~ # head /var/log/boot.log
Apr 7 08:28:22 foo allrc: syslogd startup succeeded
Apr 7 08:28:22 foo allrc: klogd startup succeeded
Apr 7 08:28:23 foo allrc: portmap startup succeeded
Apr 7 08:27:56 foo rc.sysinit: Mounting proc filesystem: succeeded
Apr 7 08:27:56 foo rc.sysinit: Unmounting initrd: succeeded
Apr 7 08:27:56 foo sysctl: net.ipv4.ip_forward = 0
Apr 7 08:27:56 foo sysctl: net.ipv4.conf.default.rp_filter = 1
Apr 7 08:27:56 foo sysctl: kernel.sysrq = 0
Apr 7 08:28:26 foo lpd: execvp: No such file or directory
Apr 7 08:27:56 foo sysctl: kernel.core_uses_pid = 1
Ex: By default head prints out the first 10 lines of a file.
10. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Selecting parts of a file (Filters that print out various parts of the input they receive)
10
foo:~ $ ls -l / | head -n 6
total 232
drwxr-xr-x 2 root root 4096 Feb 21 15:49 bin
drwxr-xr-x 3 root root 4096 Jan 7 10:25 boot
drwxr-xr-x 5 root root 20480 Jan 10 11:35 data
drwxr-xr-x 21 root root 118784 Apr 7 08:28 dev
drwxr-xr-x 64 root root 8192 Apr 7 08:28 etc
Ex: head can print out a specific number of lines from a file or a stream.
use head to extract an exact number of bytes from an input stream (rather than lines).
Here's how to get a copy of the partition sector of a disk (be careful with that redirection).
foo:~ # head -c 512 < /dev/hda > mbr
foo:~ # ls -la mbr
-rw-r--r-- 1 root root 512 Apr 7 10:27 mbr
11. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
tail – show the end of a file
Selecting parts of a file (Filters that print out various parts of the input they receive)
11
root@foo:root # tail /var/log/messages
Apr 7 11:19:34 foo dhcpd: Wrote 9 leases to leases file.
Apr 7 11:19:34 foo dhcpd: DHCPREQUEST for 10.0.0.169 from
00:80:ad:02:65:7c via eth0
Apr 7 11:19:35 foo dhcpd: DHCPACK on 10.0.0.169 to
00:80:ad:02:65:7c via eth0
Apr 7 11:20:01 foo kdm[1151]: Cannot convert Internet address
10.0.0.168 to host name
Apr 7 11:26:46 foo ipop3d[22026]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:55 foo ipop3d[22028]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:58 foo ipop3d[22035]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:27:01 foo ipop3d[22036]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:29:31 foo kdm[21954]: pam_unix2: session started for user
joe, service xdm
Apr 7 11:32:41 foo sshd[22316]: Accepted publickey for root from
10.0.0.143 port 1250 ssh2
Ex: tail is just like head, but it shows the tail end of the file.
tail can be used to watch a file as it grows.
Run the command tail –f /var/log/messages on one console and then log in on another virtual console.
tail –n 20 file or tail -20 file will show last 20 lines of file. tail –c 20 file will show last 20 characters of a file.
12. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
cut – pull out columns
Selecting parts of a file (Filters that print out various parts of the input they receive)
12
root@foo:root # tail /var/log/messages
Apr 7 11:19:34 foo dhcpd: Wrote 9 leases to leases file.
Apr 7 11:19:34 foo dhcpd: DHCPREQUEST for 10.0.0.169 from
00:80:ad:02:65:7c via eth0
Apr 7 11:19:35 foo dhcpd: DHCPACK on 10.0.0.169 to
00:80:ad:02:65:7c via eth0
Apr 7 11:20:01 foo kdm[1151]: Cannot convert Internet address
10.0.0.168 to host name
Apr 7 11:26:46 foo ipop3d[22026]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:55 foo ipop3d[22028]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:58 foo ipop3d[22035]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:27:01 foo ipop3d[22036]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:29:31 foo kdm[21954]: pam_unix2: session started for user
joe, service xdm
Apr 7 11:32:41 foo sshd[22316]: Accepted publickey for root from
10.0.0.143 port 1250 ssh2
Ex:
Cut can be used to select certain columns of the input stream.
Columns can be defined by either their position, or by being separated by field separators.
15. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Split can split files into more manageable parts, e.g. for FTP uploads. The parts can be recombined with cat.
Selecting parts of a file (Filters that print out various parts of the input they receive)
15
foo:~/download $ ls -la anomy-sanitizer-1.56.tar.gz
-rw-rw-r-- 1 georgem georgem 124356 Oct 22 18:37 anomysanitizer-1.56.tar.gz
foo:~/download $ split -b 32k anomy-sanitizer-1.56.tar.gz
foo:~/download $ ls -la x*
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xaa
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xab
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xac
-rw-rw-r-- 1 georgem georgem 26052 Apr 7 11:48 xad
Ex:
Here's how to use cat to recombine parts (using md5sum to check whether the whole is equal to the sum of
the parts).
foo:~/download $ split -b 32k anomy-sanitizer-1.56.tar.gz part-
foo:~/download $ ls -la part-*
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-aa
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-ab
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-ac
-rw-rw-r-- 1 georgem georgem 26052 Apr 7 11:49 part-ad
foo:~/download $ cat part-* > newfile
foo:~/download $ md5sum newfile anomy-sanitizer-1.56.tar.gz
1a977bad964b0ede863272114bfc2482 newfile
1a977bad964b0ede863272114bfc2482 anomy-sanitizer-1.56.tar.gz
18. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Sorting
18
uniq – discard duplicate lines - uniq is usually used with sort to discard duplicates
foo:~ $ cat /etc/passwd | cut -d : -f 4 | sort -n | fmt
0 0 0 0 0 0 1 2 4 7 12 13 14 25 26 28 29 30 32 37 38 42 43 47 48 50 51
69 74 77 80 89 99 100 500 501 503 504 505 506 507 509 511 512 65534
Ex: Ex: cutting fourth field out of the password file (the group ID) and sorting in numerical order.
fmt is used to make the results display on a single line.
Ex: Ex: same command pipeline, but are removing duplicates with uniq before formatting.
foo:~ $ cat /etc/passwd | cut -d : -f 4 | sort -n | uniq | fmt
0 1 2 4 7 12 13 14 25 26 28 29 30 32 37 38 42 43 47 48 50 51 69 74 77
80 89 99 100 500 501 503 504 505 506 507 509 511 512 65534
19. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
19
tr – character set translation - usually used for converting upper case to lower case.
Can also do other character translations. tr -d can remove specific characters from a stream.
foo:~ $ man -P cat man | tr A-Z a-z | less
Ex: Ex: Translating from UPPER CASE to lower case. Asking man to use cat as its pager, instead of
using less, and opening up the man page for man itself.
Translating from lower case to UPPER CASE:
foo:~ $ man -P cat man | tr a-z A-Z | less
Convert file names to lowercase.
foo:/windows/C $ for FILE in * ; do
mv "$FILE" $( echo "$FILE" | tr A-Z a-z ) ; done
Using tr -d to delete the Carriage Returns (r) from a file created with Windows Notepad.
foo:/windows/C $ tr -d 'r' notepad.dos.txt > notepad.unix.txt
Ex:
Ex:
Ex:
21. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
21
Joining /etc/passwd and /etc/shadow based on their first field (user name).
Since /etc/passwd and /etc/shadow use a colon to separate fields, it is necessary to use -t: option.
foo:~ # join -t : /etc/passwd /etc/shadow | head
root:x:0:0:root:/root:/bin/bash:$1$LHNUbu7U$oiuhqwd1oiuhqhAdiuHvA0:1
2146:0:99999:7:::
bin:x:1:1:bin:/bin:/sbin/nologin:*:11974:0:99999:7:::
daemon:x:2:2:daemon:/sbin:/sbin/nologin:*:11974:0:99999:7:::
adm:x:3:4:adm:/var/adm:/sbin/nologin:*:11974:0:99999:7:::
lp:x:4:7:lp:/var/spool/lpd:/sbin/nologin:*:11974:0:99999:7:::
sync:x:5:0:sync:/sbin:/bin/sync:*:11974:0:99999:7:::
shutdown:x:6:0:shutdown:/sbin:/sbin/shutdown:*:11974:0:99999:7:::
halt:x:7:0:halt:/sbin:/sbin/halt:*:11974:0:99999:7:::
mail:x:8:12:mail:/var/spool/mail:/sbin/nologin:*:11974:0:99999:7:::
Ex:
join allows to specify which particular field to join on, and also which particular fields should appear in the
output (similar to cut)
23. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
23
sed – stream editor - sed does text transformations on an input stream.
foo:~ $ echo "Hi Fred, how are you" | sed 's/Fred/Joe/'
Hi Joe, how are you
foo:~ $ echo "Hi Fred, how is Fred?" | sed 's/Fred/Joe/'
Hi Joe, how is Fred?
foo:~ $ echo "Hi Fred, how is Fred?" | sed 's/Fred/Joe/g‘
Hi Joe, how is Joe?
Ex:
•sed works by making only one pass over the inputs.
•A sed program consists of one or more sed commands which are applied to each line of the input.
•A command may be prefixed by an address range indicating lines for which the command performes.
sed commands: s/PATTERN/REPLACEMENT/g
search and replace. If you add g at end - search and replace applies as many times as possible to single line.
You can also use i at the end of the s command to make the search case insensitive
24. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
24
foo:~ $ ls /bin/ | fmt -30 | nl | sed '4,15d'
1 arch ash ash.static awk
2 basename bash cat chgrp
3 chmod chown chvt cp cpio csh
16 true umount uname usleep vi
17 vim zcat zsh
foo:~ $ ls /bin | fmt -40 | nl | sed '/e/ d'
2 bash cat chgrp chmod chown chvt cp
13 vi vim zcat zsh
Ex:
sed commands: d – delete the line.
You need to select the lines as explained in the next paragraph.
The commands are most useful if you specify a range of lines to which the command applies.
Here's how to specify specific lines for the d command:
•/PATTERN/d – delete all lines contains the pattern
•4d – delete line 4
•4,10d – delete lines 4 to 10
•6,$d – delete from line 6 to the last line
29. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
29
paste – paste two files together
foo:~/tmp $ cat file2
cabbage green leaves
cat white cat-hair
piano brown wood
foo:~/tmp $ cat file1
cat animal
cabbage vegetable
piano mineral
coal mineral
foo:~/tmp $ paste file2 file1 | expand -t 22
cabbage green leaves cat animal
cat white cat-hair cabbage vegetable
piano brown wood piano mineral
coal mineral
Ex:
Using paste is like taking two printouts of two files and sticking the right margin of one to the left margin of
the other. The glue between the files is a tab stop.
You can specify a delimiter between the files which is not a tab space with option -d
30. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Formatting
30
fmt – format nicely - format text into a neatly word wrapped structure. Default right margin is 75
foo:~ $ ls /bin | fmt
arch ash ash.static aumix-minimal awk basename bash bash2 bsh cat
chgrp
chmod chown cp cpio csh cut date dd df dmesg dnsdomainname doexec
domainname dumpkeys echo ed egrep env ex false fgrep gawk gettext
grep gtar gunzip gzip hostname igawk ipcalc kbd_mode kill link ln
loadkeys login ls mail mkdir mknod mktemp more mount mt mv netstat
nice
nisdomainname pgawk ping ps pwd red rm rmdir rpm rvi rview sed
setfont
setserial sh sleep sort stty su sync tar tcsh touch true umount
uname
unicode_start unicode_stop unlink usleep vi vi.minimal view
ypdomainname
Ex:
foo:~ $ ls /bin | fmt -40 | head
arch ash ash.static aumix-minimal
awk basename bash bash2 bsh cat chgrp
chmod chown cp cpio csh cut date dd df
dmesg dnsdomainname doexec domainname
dumpkeys echo ed egrep env ex false
fgrep gawk gettext grep gtar gunzip
gzip hostname igawk ipcalc kbd_mode
kill link ln loadkeys login ls mail
mkdir mknod mktemp more mount mt mv
netstat nice nisdomainname pgawk ping
32. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Formatting
32
od – octal dump (and other formats)
Ex:
foo:/tmp $ echo "Hello World" > hello
foo:/tmp $ cat hello
Hello World
od, oddly enough, does not just print out in octal, but in other formats.
foo:/tmp $ od hello
0000000 062510 066154 020157 067527 066162 005144
0000014
od's behaviour is rather odd when it is used without any arguments. It prints out the octal
value of two pairs of bytes in the file. Sometimes this is useful, but usually it isnot.
Using -t switch tells od to use a specific format type (the default is od –t o2).
od –t c means character format. You will notice that file ends in a newline character (n).
foo:/tmp $ od -t c hello
0000000 H e l l o W o r l d n
0000014
od –t d1 specifies decimal format, with a width of one. character encoding is ASCII.
Ex:
Ex:
foo:/tmp $ od -t d1 hello
0000000 72 101 108 108 111 32 87 111 114 108 100 10
0000014