This document discusses various UNIX commands for processing text streams and filtering text, including cat, cut, head, tail, and split. It provides examples of using each command to select, modify, or restructure the output. The commands can be used to select parts of files like lines (head/tail) or columns (cut), and to split files into multiple parts (split). Pipelines and redirection are also covered.
This document discusses Linux text stream filters and provides examples of common Unix commands used to process and modify text streams. These commands include cat, head, tail, cut, and split. Cat prints the contents of files, head prints the first few lines, tail prints the last few lines, cut extracts parts of each line, and split divides files into smaller parts. The document also covers input/output redirection and how it can be used with filters to modify command output and send it to files.
This document discusses Linux text stream filters and provides examples of common Unix commands used to process and modify text streams. These commands include cat, head, tail, cut, and split. Cat prints the contents of files, head prints the first few lines, tail prints the last few lines, cut extracts parts of each line, and split divides files into smaller parts. The document also covers input/output redirection and how it can be used with text stream filters.
This document provides an overview of Linux commands for redirecting standard input, output, and error streams. Some key points covered include:
- Redirecting input (<), output (> and >>), and pipes (|) to send output as input to another command
- Common redirection operators like tee to send output to both a file and stdout, and xargs to expand input to command line arguments
- Redirecting standard error (2>) independently from standard output using 2>, 2>>, 2>&1
- Using command substitution (`command` and $(command)) to capture output and use as arguments
- Chaining commands together in pipelines (|) to filter and transform text streams
- Examples of
This document provides an overview of common Linux commands used to process text streams and filter output, including cat, cut, head, tail, and split. It discusses how these commands can be used to select, sort, reformat, and summarize data by printing certain parts of files like columns, lines, or characters. Redirection is also covered as a way to modify command input and output. The goal is to explain the key knowledge areas and objectives for the Junior Level Linux Certification exam related to GNU and Unix commands.
This document provides an introduction and overview of Linux commands and Perl basics. It discusses key Linux commands for system information, user management, files/directories, permissions, processes, networking and more. It also covers Perl data types, variables, input/output, strings, arithmetic, comparisons, functions and file handling. The document aims to teach Linux commands and Perl programming basics.
This document provides information about monitoring, controlling, and killing processes on Linux systems. It discusses running processes in the foreground and background, disconnecting processes from terminals using nohup and setsid, and sending signals to processes using commands like kill, jobs, fg, and bg. Monitoring tools like ps and top are also covered, along with common signals like SIGTERM. The overall objective is to teach examinees how to create, monitor and kill processes on Linux systems.
This document provides an overview of various networking tools in Linux, including commands for network configuration (ifconfig, route), connectivity testing (ping, traceroute), name resolution (host, nslookup), port and protocol inspection (netstat, tcpdump), and secure remote access (SSH, PuTTY). It also covers tools for firewall management (ufw), network mapping (Nmap), raw socket programming (netcat), link status (ethtool), and more. Examples are given for common tasks like viewing routing tables, capturing packets, remotely controlling systems, and accessing services over Telnet versus SSH. A references section at the end provides additional learning resources.
This document provides summaries of Linux commands that can be used to select and manipulate parts of files. It discusses the cat, head, tail, cut, split, sort, tac, uniq, and tr commands. Cat concatenates or displays files, head displays the first few lines of a file, and tail displays the last few lines. Cut extracts columns from a file, split divides a file into smaller parts, and sort sorts lines alphabetically or numerically. Tac displays the lines of a file in reverse order and uniq removes duplicate lines from a sorted file. Tr translates characters within a file, such as converting uppercase to lowercase.
This document discusses Linux text stream filters and provides examples of common Unix commands used to process and modify text streams. These commands include cat, head, tail, cut, and split. Cat prints the contents of files, head prints the first few lines, tail prints the last few lines, cut extracts parts of each line, and split divides files into smaller parts. The document also covers input/output redirection and how it can be used with filters to modify command output and send it to files.
This document discusses Linux text stream filters and provides examples of common Unix commands used to process and modify text streams. These commands include cat, head, tail, cut, and split. Cat prints the contents of files, head prints the first few lines, tail prints the last few lines, cut extracts parts of each line, and split divides files into smaller parts. The document also covers input/output redirection and how it can be used with text stream filters.
This document provides an overview of Linux commands for redirecting standard input, output, and error streams. Some key points covered include:
- Redirecting input (<), output (> and >>), and pipes (|) to send output as input to another command
- Common redirection operators like tee to send output to both a file and stdout, and xargs to expand input to command line arguments
- Redirecting standard error (2>) independently from standard output using 2>, 2>>, 2>&1
- Using command substitution (`command` and $(command)) to capture output and use as arguments
- Chaining commands together in pipelines (|) to filter and transform text streams
- Examples of
This document provides an overview of common Linux commands used to process text streams and filter output, including cat, cut, head, tail, and split. It discusses how these commands can be used to select, sort, reformat, and summarize data by printing certain parts of files like columns, lines, or characters. Redirection is also covered as a way to modify command input and output. The goal is to explain the key knowledge areas and objectives for the Junior Level Linux Certification exam related to GNU and Unix commands.
This document provides an introduction and overview of Linux commands and Perl basics. It discusses key Linux commands for system information, user management, files/directories, permissions, processes, networking and more. It also covers Perl data types, variables, input/output, strings, arithmetic, comparisons, functions and file handling. The document aims to teach Linux commands and Perl programming basics.
This document provides information about monitoring, controlling, and killing processes on Linux systems. It discusses running processes in the foreground and background, disconnecting processes from terminals using nohup and setsid, and sending signals to processes using commands like kill, jobs, fg, and bg. Monitoring tools like ps and top are also covered, along with common signals like SIGTERM. The overall objective is to teach examinees how to create, monitor and kill processes on Linux systems.
This document provides an overview of various networking tools in Linux, including commands for network configuration (ifconfig, route), connectivity testing (ping, traceroute), name resolution (host, nslookup), port and protocol inspection (netstat, tcpdump), and secure remote access (SSH, PuTTY). It also covers tools for firewall management (ufw), network mapping (Nmap), raw socket programming (netcat), link status (ethtool), and more. Examples are given for common tasks like viewing routing tables, capturing packets, remotely controlling systems, and accessing services over Telnet versus SSH. A references section at the end provides additional learning resources.
This document provides summaries of Linux commands that can be used to select and manipulate parts of files. It discusses the cat, head, tail, cut, split, sort, tac, uniq, and tr commands. Cat concatenates or displays files, head displays the first few lines of a file, and tail displays the last few lines. Cut extracts columns from a file, split divides a file into smaller parts, and sort sorts lines alphabetically or numerically. Tac displays the lines of a file in reverse order and uniq removes duplicate lines from a sorted file. Tr translates characters within a file, such as converting uppercase to lowercase.
The document discusses Linux input/output redirection and piping. It explains how the shell creates new processes to handle redirection and piping. Standard streams like stdout and stderr can be redirected to files or other programs using operators like >, >>, |, and 2>&1. Pipes allow the output of one command to serve as input to another command. The tee command diverts a copy of the input stream to both a file and stdout.
This document provides an overview of common Linux networking commands such as ifconfig, route, traceroute, nslookup, arp, dig, and netstat that are used to configure network interfaces, display routing tables, trace network routes, lookup domain names, manage address resolution, query DNS servers, and view network statistics. It also discusses how to use ifconfig to assign IP addresses to interfaces, route to view routing tables, arp to manage the address resolution cache, and dig for more powerful DNS lookups than nslookup.
The document discusses Linux commands for redirecting standard input, output, and error streams. It covers commands like tee, xargs, <, >, >>, |, and backticks. Pipes and redirects allow commands to be chained together and output sent to files or other programs. Specific redirection examples are provided like redirecting just errors, appending output, and piping output between commands.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, and ls. The examples cover a wide range of tasks from compressing/extracting files to searching/editing text to managing processes and permissions.
The document discusses network layering models and TCP/IP fundamentals. It describes:
1. Networking problems are divided into layers for easier understanding and standardization, with the two main models being OSI and TCP/IP.
2. The TCP/IP model has four or five layers - process, host-to-host transport, internet, network access, and sometimes physical.
3. Packets are encapsulated as they leave a machine and decapsulated on the receiving host, with each layer adding headers.
This document provides 50 examples of common Linux/Unix commands organized by command name. It begins with tar, grep, find, ssh, sed, awk, vim, diff, sort, and export examples. The document is intended as a quick reference for users to learn practical uses of fundamental Linux commands.
introduction to linux kernel tcp/ip ptocotol stack monad bobo
This document provides an introduction and overview of the networking code in the Linux kernel source tree. It discusses the different layers including link (L2), network (L3), and transport (L4) layers. It describes the input and output processing, device interfaces, traffic directions, and major developers for each layer. Config and benchmark tools are also mentioned. Resources for further learning about the Linux kernel networking code are provided at the end.
The document provides an overview of Linux operating system concepts including:
- Linux is an open source operating system that interacts with hardware and allocates resources.
- It supports multi-tasking and multi-user environments. Common types include Debian, Ubuntu, and Redhat.
- Key components include the kernel, shell programs, file management commands, text editors, browsers, and programming tools.
Linux is an open-source operating system that can run on various hardware. The document discusses various Linux commands and concepts related to directories, files, permissions, users, groups, text editors like vi and vim, process management, disk partitioning and more. It also covers Linux installation, package management, shell scripting and configuring network and services like SSH, web servers and more. Exercises are included to help understand concepts like mount points, journaling and file attributes.
This document discusses Linux input/output redirection and piping commands. It covers redirecting standard input, output, and error to files or other programs. Specific redirection operators like >, >>, |, and < are explained along with examples. Commands like tee and xargs that are used to manage input/output streams are also overviewed. The key techniques covered include redirecting streams, piping outputs between commands, and using filters to process text streams.
The document discusses various data structures and functions related to network packet processing in the Linux kernel socket layer. It describes the sk_buff structure that is used to pass packets between layers. It also explains the net_device structure that represents a network interface in the kernel. When a packet is received, the interrupt handler will raise a soft IRQ for processing. The packet will then traverse various protocol layers like IP and TCP to be eventually delivered to a socket and read by a userspace application.
This document provides an overview of basic Linux commands for navigation, listing directories, reading and manipulating files. It explains commands like pwd, cd, ls, cat, cp, mv, rm, mkdir to change directories, list files, read files, copy, move and delete files/directories. It also introduces the vi editor for creating new files and mentions some other miscellaneous commands like date, chmod, user management tools.
The document logs the startup process of a messaging client. It records the loading of configuration files and plugins, the establishment of network connections, and the initial authentication process with the messaging service. Errors are encountered when loading some plugins due to missing dependencies. The client then successfully connects to the messaging service and authenticates the user account.
This document provides information about creating partitions and filesystems in Linux. It discusses various Linux filesystem types like ext2, ext3, xfs, reiserfs v3, and vfat. It covers the commands and tools used to create partitions (fdisk, mkfs), filesystems (mkfs), and swap spaces (mkswap, swapon). It also discusses viewing filesystem information, mounting filesystems, and the Filesystem Hierarchy Standard for directory structure in Linux.
This document provides an overview of common Linux commands organized into categories such as file handling, process management, disk usage, searching, editing, and remote access. It describes the purpose and basic usage of commands like vi, ls, grep, find, scp, and more. Examples are given for piping output, redirection, appending, and searching/replacing text within files. Links to additional online resources on Linux basics and tips are also included.
Networking in Linux discusses DNS related commands in Linux. It begins by listing DNS concepts like zones and records. It then demonstrates commands like nslookup, host and dig to query DNS records like A, MX, NS, SOA records and perform operations like reverse lookups. It shows how to use specific nameservers, change ports and timeouts. The document provides examples of using these tools to troubleshoot DNS issues like propagation.
This lecture discusses a group of techniques to use commands output/Input to feed into other commands or into files. It also covers argument expansion and quoting
Check the other Lectures and courses in
http://Linux4EnbeddedSystems.com
or Follow our Facebook Group at
- Facebook: @LinuxforEmbeddedSystems
Lecturer Profile:
- https://www.linkedin.com/in/ahmedelarabawy
Efficient System Monitoring in Cloud Native EnvironmentsGergely Szabó
This document discusses efficient system monitoring in cloud native environments using eBPF. It provides an overview of eBPF and how it can be used for monitoring applications like Prometheus. Specific topics covered include BPF, Linux kernel tracing using kprobes and tracepoints, eBPF maps and programs, and an example Prometheus exporter that leverages eBPF to export metrics.
This document discusses processes in Linux and tools for managing them. It begins by explaining that a process is a set of instructions loaded into memory that has a numeric PID for identification. It then covers commands like ps, pgrep, and pidof for listing and finding processes. The document also discusses signals for inter-process communication, scheduling priority, job control, and tools like top, at and crontab for interactive process management and scheduling jobs.
Linux is a prominent example of free and open source software. It can be installed on a wide variety of devices from embedded systems to supercomputers. Linux is commonly used for servers, with estimates that it powers around 60% of web servers. Linux distributions package the Linux kernel with other software like utilities, libraries and desktop environments. Programming languages and build tools like GCC are supported. Embedded Linux is often used in devices due to its low cost and ease of modification.
This document provides an overview of shell scripting, cron, and atd for automating tasks in Linux. It discusses using shell scripts to execute a series of commands from a file, as well as topics like shebangs, variables, and conditionals. Cron allows scheduling commands to run on a time-based schedule using crontab files, while atd can run one-off commands at a specific time. The document provides examples and explanations of common cron and atd utilities.
The document discusses Linux input/output redirection and piping. It explains how the shell creates new processes to handle redirection and piping. Standard streams like stdout and stderr can be redirected to files or other programs using operators like >, >>, |, and 2>&1. Pipes allow the output of one command to serve as input to another command. The tee command diverts a copy of the input stream to both a file and stdout.
This document provides an overview of common Linux networking commands such as ifconfig, route, traceroute, nslookup, arp, dig, and netstat that are used to configure network interfaces, display routing tables, trace network routes, lookup domain names, manage address resolution, query DNS servers, and view network statistics. It also discusses how to use ifconfig to assign IP addresses to interfaces, route to view routing tables, arp to manage the address resolution cache, and dig for more powerful DNS lookups than nslookup.
The document discusses Linux commands for redirecting standard input, output, and error streams. It covers commands like tee, xargs, <, >, >>, |, and backticks. Pipes and redirects allow commands to be chained together and output sent to files or other programs. Specific redirection examples are provided like redirecting just errors, appending output, and piping output between commands.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, and ls. The examples cover a wide range of tasks from compressing/extracting files to searching/editing text to managing processes and permissions.
The document discusses network layering models and TCP/IP fundamentals. It describes:
1. Networking problems are divided into layers for easier understanding and standardization, with the two main models being OSI and TCP/IP.
2. The TCP/IP model has four or five layers - process, host-to-host transport, internet, network access, and sometimes physical.
3. Packets are encapsulated as they leave a machine and decapsulated on the receiving host, with each layer adding headers.
This document provides 50 examples of common Linux/Unix commands organized by command name. It begins with tar, grep, find, ssh, sed, awk, vim, diff, sort, and export examples. The document is intended as a quick reference for users to learn practical uses of fundamental Linux commands.
introduction to linux kernel tcp/ip ptocotol stack monad bobo
This document provides an introduction and overview of the networking code in the Linux kernel source tree. It discusses the different layers including link (L2), network (L3), and transport (L4) layers. It describes the input and output processing, device interfaces, traffic directions, and major developers for each layer. Config and benchmark tools are also mentioned. Resources for further learning about the Linux kernel networking code are provided at the end.
The document provides an overview of Linux operating system concepts including:
- Linux is an open source operating system that interacts with hardware and allocates resources.
- It supports multi-tasking and multi-user environments. Common types include Debian, Ubuntu, and Redhat.
- Key components include the kernel, shell programs, file management commands, text editors, browsers, and programming tools.
Linux is an open-source operating system that can run on various hardware. The document discusses various Linux commands and concepts related to directories, files, permissions, users, groups, text editors like vi and vim, process management, disk partitioning and more. It also covers Linux installation, package management, shell scripting and configuring network and services like SSH, web servers and more. Exercises are included to help understand concepts like mount points, journaling and file attributes.
This document discusses Linux input/output redirection and piping commands. It covers redirecting standard input, output, and error to files or other programs. Specific redirection operators like >, >>, |, and < are explained along with examples. Commands like tee and xargs that are used to manage input/output streams are also overviewed. The key techniques covered include redirecting streams, piping outputs between commands, and using filters to process text streams.
The document discusses various data structures and functions related to network packet processing in the Linux kernel socket layer. It describes the sk_buff structure that is used to pass packets between layers. It also explains the net_device structure that represents a network interface in the kernel. When a packet is received, the interrupt handler will raise a soft IRQ for processing. The packet will then traverse various protocol layers like IP and TCP to be eventually delivered to a socket and read by a userspace application.
This document provides an overview of basic Linux commands for navigation, listing directories, reading and manipulating files. It explains commands like pwd, cd, ls, cat, cp, mv, rm, mkdir to change directories, list files, read files, copy, move and delete files/directories. It also introduces the vi editor for creating new files and mentions some other miscellaneous commands like date, chmod, user management tools.
The document logs the startup process of a messaging client. It records the loading of configuration files and plugins, the establishment of network connections, and the initial authentication process with the messaging service. Errors are encountered when loading some plugins due to missing dependencies. The client then successfully connects to the messaging service and authenticates the user account.
This document provides information about creating partitions and filesystems in Linux. It discusses various Linux filesystem types like ext2, ext3, xfs, reiserfs v3, and vfat. It covers the commands and tools used to create partitions (fdisk, mkfs), filesystems (mkfs), and swap spaces (mkswap, swapon). It also discusses viewing filesystem information, mounting filesystems, and the Filesystem Hierarchy Standard for directory structure in Linux.
This document provides an overview of common Linux commands organized into categories such as file handling, process management, disk usage, searching, editing, and remote access. It describes the purpose and basic usage of commands like vi, ls, grep, find, scp, and more. Examples are given for piping output, redirection, appending, and searching/replacing text within files. Links to additional online resources on Linux basics and tips are also included.
Networking in Linux discusses DNS related commands in Linux. It begins by listing DNS concepts like zones and records. It then demonstrates commands like nslookup, host and dig to query DNS records like A, MX, NS, SOA records and perform operations like reverse lookups. It shows how to use specific nameservers, change ports and timeouts. The document provides examples of using these tools to troubleshoot DNS issues like propagation.
This lecture discusses a group of techniques to use commands output/Input to feed into other commands or into files. It also covers argument expansion and quoting
Check the other Lectures and courses in
http://Linux4EnbeddedSystems.com
or Follow our Facebook Group at
- Facebook: @LinuxforEmbeddedSystems
Lecturer Profile:
- https://www.linkedin.com/in/ahmedelarabawy
Efficient System Monitoring in Cloud Native EnvironmentsGergely Szabó
This document discusses efficient system monitoring in cloud native environments using eBPF. It provides an overview of eBPF and how it can be used for monitoring applications like Prometheus. Specific topics covered include BPF, Linux kernel tracing using kprobes and tracepoints, eBPF maps and programs, and an example Prometheus exporter that leverages eBPF to export metrics.
This document discusses processes in Linux and tools for managing them. It begins by explaining that a process is a set of instructions loaded into memory that has a numeric PID for identification. It then covers commands like ps, pgrep, and pidof for listing and finding processes. The document also discusses signals for inter-process communication, scheduling priority, job control, and tools like top, at and crontab for interactive process management and scheduling jobs.
Linux is a prominent example of free and open source software. It can be installed on a wide variety of devices from embedded systems to supercomputers. Linux is commonly used for servers, with estimates that it powers around 60% of web servers. Linux distributions package the Linux kernel with other software like utilities, libraries and desktop environments. Programming languages and build tools like GCC are supported. Embedded Linux is often used in devices due to its low cost and ease of modification.
This document provides an overview of shell scripting, cron, and atd for automating tasks in Linux. It discusses using shell scripts to execute a series of commands from a file, as well as topics like shebangs, variables, and conditionals. Cron allows scheduling commands to run on a time-based schedule using crontab files, while atd can run one-off commands at a specific time. The document provides examples and explanations of common cron and atd utilities.
This document provides an overview of shell scripting, cron, and atd for scheduling tasks on Linux systems. It discusses using shell scripts to automate commands, the cron daemon for executing commands on a schedule, and atd for running one-time jobs. Key points covered include using the shebang to directly execute shell scripts, crontab syntax for defining schedules, and commands like crond, at, atq and atrm for managing cron and at jobs. The document is intended as a learning resource for system administration tasks using scheduling utilities in Linux.
Linux is an open-source alternative to Microsoft Windows that is freely available and reliable. It has a command line interface and features like virtual memory, networking capabilities, multiple users, and graphical user interfaces. Common Linux commands include ls, cd, mkdir, rmdir, cat, cp, and editors like vi, emacs, nano are used to create and edit files. The Linux directory structure is hierarchical with key directories being /, /home, /usr, /var, and editors allow editing files in different modes like insert and command modes.
This document provides instructions for 27 common Linux commands: mkdir, rmdir, ls, cd, cat, touch, wc, who, pwd, rm, mv, chmod, cp, grep, cal, date, vi, tput, ps, export, type, tail, sudo, head, man, clear, and adduser. For each command, it lists the syntax and provides 1-3 examples of common uses. The document is presented over 28 pages with the commands organized topic-by-topic and includes formatting like headings and indentation to aid readability.
The document provides an overview of Unix and shell scripting. It discusses the history and architecture of Unix operating systems. It then covers various Unix commands and utilities for file management, processes, communication, and system administration. Finally, it describes the basics of shell scripting including variables, conditional statements, loops, and here documents.
This document provides a cheat sheet of common Linux commands and their usage. It covers basic file operations like copying, moving, deleting files and directories. It also includes commands for viewing files, compressing/decompressing files, finding files, remote access, and getting system information. The commands are explained over 3 pages with examples of proper syntax and usage for each one.
Linux is an open-source operating system that can be used as a server or standalone OS. It offers advantages like stability, robustness, security, and high performance. Linux directories include / for the root directory, /bin for essential programs, /home for user files, and /var for variable data. Common Linux commands include ls to list files, cd to change directories, grep to search files, and vi/emacs for text editing. Cron jobs allow scheduling commands to run periodically using a crontab file.
Server is a machine configured to accept requests from clients and respond accordingly. Linux is commonly used for servers, with distributions like Ubuntu, Redhat, and Debian. Key principles of Linux include treating everything as a file, storing configuration data in text files, and using pipes to connect programs. Common server files include /etc/group, /etc/passwd, and /etc/shadow which contain user and group information.
You have a system with an advanced programmatic tracer: do you know what to do with it? Brendan has used numerous tracers in production environments, and has published hundreds of tracing-based tools. In this talk he will share tips and know-how for creating CLI tracing tools and GUI visualizations, to solve real problems effectively. Programmatic tracing is an amazing superpower, and this talk will show you how to wield it!
A brief talk on systems performance for the July 2013 meetup "A Midsummer Night's System", video: http://www.youtube.com/watch?v=P3SGzykDE4Q. This summarizes how systems performance has changed from the 1990's to today. This was the reason for writing a new book on systems performance, to provide a reference that is up to date, covering new tools, technologies, and methodologies.
This document provides 50 examples of common Linux/Unix commands along with brief explanations and usage examples for each command. Some of the commands highlighted include tar, grep, find, ssh, sed, awk, vim, diff, sort, export, xargs, ls, pwd, cd, gzip, bzip2, unzip, shutdown, ftp, crontab, service, ps, top, df, kill, rm, cp, mv, cat, mount, chmod, chown, passwd, mkdir, ifconfig, and uname. The document is intended to give readers a quick start on frequently used commands.
An operating system acts as an interface between hardware and software, managing resources and presenting an easier programming interface than the underlying hardware. The UNIX system is organized into three levels - the kernel which manages tasks and storage, the shell which connects programs and executes commands, and tools/applications which provide additional functionality. PuTTY is a free and open-source terminal emulator and SSH client most commonly used to access UNIX/Linux systems remotely from Windows. The course objectives are to learn UNIX commands, shell programming, process management, memory management, file organization, and implementing related algorithms.
This document provides an introduction to Linux, including:
- What Linux is and where it came from, as an open source operating system created by Linus Torvalds in 1991.
- How to get Linux through downloading a distribution like Red Hat and the packages it includes.
- Why Linux has become significant due to its growing popularity, ability to run on multiple hardware platforms, and being free.
- An overview of Linux commands and utilities for file management, processes, users and permissions.
This document summarizes options for running the Odoo server from the command line, including:
- Specifying database, modules, and module paths for installation/updating
- Configuring workers for multiprocessing and limiting resources per worker
- Setting logging options like the log file, database, or specific log levels
- Advanced options like auto-reloading or enabling a proxy
This document provides descriptions of common Linux commands organized into categories including file manipulation, text processing, documentation, editors, file and directory management, file display and printing, programming tools, communications, process control, status information, and web commands. It describes what each command is used for and provides examples of some commonly used commands and their functions.
Linux is a free, open-source operating system that provides functionality similar to Unix systems. The document then lists and describes several common Linux commands used to navigate directories, view files and system information, manage processes, and edit text files. It also includes a diagram of the standard Linux directory structure and brief descriptions of the purpose of each top-level directory.
This document discusses Linux file systems and creating partitions and filesystems in Linux. It covers the following key points:
1. Linux supports various filesystems like ext2, ext3, xfs, and ReiserFS that can be created using mkfs. Swap spaces are created with mkswap.
2. Partitions and filesystems can be created using tools like fdisk, cfdisk, and gpart. Filesystem types include ext3, xfs, FAT, etc.
3. The Filesystem Hierarchy Standard defines the directory structure and recommended layout of files on Linux systems with directories like /bin, /etc, /home, /usr, /var, etc.
Similar to 3.2 process text streams using filters (20)
This document discusses the differences between vulnerability scanning and penetration testing. Vulnerability scanning uses automated tools to identify vulnerabilities while penetration testing involves manual techniques to simulate real-world attacks. The document also notes that vulnerability scanning and penetration testing have different levels of testing and are covered in Topic 24 of the Security+ Study Guide for 10 minutes.
This document discusses application security controls and techniques. It covers secure coding concepts, other security controls and techniques, and provides a 10 minute overview of application security controls and techniques as outlined in the CompTIA Security+ Study Guide. The document was prepared by Acácio Oliveira in 2021.
This document discusses types of application attacks and is divided into two sections. The first section defines application attacks and covers common types. The second section explains the goals of application attacks and how they can divulge weaknesses in some applications. The entire document is part of a Security+ study guide and covers types of application attacks over a period of 20 minutes.
Security+ Lesson 01 Topic 19 - Summary of Social Engineering Attacks.pptxAcácio Oliveira
This document provides a summary of social engineering attacks. Social engineering attacks are effective because they target human nature. There are different types of social engineering attacks, such as pretexting where attackers pretend to be someone trustworthy to get victims to reveal sensitive information. The document is part of a study guide for the CompTIA Security+ certification and focuses on summarizing social engineering attacks.
This document discusses security assessment tools, including types of security assessments and specific assessment tools. It is part of a CompTIA Security+ study guide prepared in 2021, focusing on an overview of security assessment tools in about 10 minutes.
This document provides a 10 minute summary of wireless attacks. It discusses common types of wireless attacks and attacks on wireless encryption. The document is part of a study guide for the CompTIA Security+ certification and is prepared by Acácio Oliveira in 2021.
This document discusses security enhancement techniques for network security. It covers the difference between detection controls and prevention controls. The topic is part of the Security+ Study Guide Concepts prepared by Acácio Oliveira in 2021 and focuses on network security enhancement techniques for 10 minutes as part of the study guide's Topic 22 on security enhancement techniques.
This document discusses risk management best practices for the CompTIA Security+ exam, covering business continuity concepts, fault tolerance, and is part of a study guide prepared by Acácio Oliveira in 2021 with an estimated reading time of 10 minutes.
This document discusses physical security and environmental controls, which are important concepts for the CompTIA Security+ exam. It covers control types for physical security and the environment, such as restricting physical access and monitoring for issues like fires, floods or power outages. The document is from a study guide prepared in 2021 to help study for the Security+ exam.
This document discusses disaster recovery concepts for the CompTIA Security+ exam, including maintaining disaster recovery sites, performing regular data backups, and ensuring the ability to recover data and systems in the event of a disaster within 10 minutes of scheduled content.
This document discusses wireless security considerations for the CompTIA Security+ exam. It covers the unique security challenges of wireless networks compared to wired networks. Specific topics include security for wireless networks and an index of study guide topics related to wireless security considerations with an allocated time of 10 minutes.
Security+ Lesson 01 Topic 04 - Secure Network Design Elements and Components....Acácio Oliveira
This document discusses secure network design elements and components. It covers defense in depth as a strategy as well as elements and components of network design such as firewalls, intrusion detection systems, and virtual private networks. The document is from Acácio Oliveira's 2021 CompTIA Security+ Study Guide and focuses on secure network design for 10 minutes.
This document provides a 3-topic outline for a CompTIA Security+ study guide prepared by Acácio Oliveira in 2021. Topic 1 discusses secure network administration concepts like spam filters and network devices. Topic 2 is titled "Secure Network Administration Concepts" and is allocated 10 minutes of study time. The document also includes a topics index for the study guide.
This document provides an overview of the concepts to be covered in Lesson 01 on introduction to network devices from the CompTIA Security+ Study Guide. The lesson will be divided into three parts covering OSI models, basic network devices and layer security concepts in part 1; network devices in part 2; and spam filters and additional network devices in part 3, for a total time of 30 minutes.
Security+ Lesson 01 Topic 08 - Integrating Data and Systems with Third Partie...Acácio Oliveira
This document discusses integrating data and systems with third parties. It examines evaluating risks with integration and considerations for integration. The document is part of a Security+ study guide prepared by Acácio Oliveira in 2021 on the topic of integrating data and systems with third parties.
Security+ Lesson 01 Topic 07 - Risk Related Concepts.pptxAcácio Oliveira
This document outlines concepts related to risk for the CompTIA Security+ certification. It covers three topics on risk-related concepts over 30 minutes: policies for controlling and reducing risk, qualitative versus quantitative risk assessments, and treating risks such as those involving cloud computing and virtualization. The document was prepared by Acácio Oliveira for Security+ study.
Security+ Lesson 01 Topic 05 - Common Network Protocols.pptxAcácio Oliveira
This document provides an overview of the topics to be covered in a 30-minute Common Network Protocols lesson for CompTIA Security+ certification preparation. The lesson will be divided into three parts covering IPv4 and IPv6, network storage protocols, the difference between ports and protocols, common protocols, and end-to-end security.
This document provides a 10 minute overview of incident response concepts for Security+ certification preparation. It discusses first responder responsibilities during security incidents and outlines incident response procedures and concepts covered in the Security+ study guide. The document is titled "Incident Response Concepts" and was prepared by Acácio Oliveira in 2021 as part of a Security+ study guide.
Security+ Lesson 01 Topic 12 - Security Related Awareness and Training.pptxAcácio Oliveira
This document discusses security related awareness and training. It covers the security policy, security awareness, and is part of a study guide for the CompTIA Security+ exam. The topic is allocated 10 minutes as part of the study guide index.
Security+ Lesson 01 Topic 17 - Types of Malware.pptxAcácio Oliveira
Malware is defined and common types are discussed including viruses, worms, Trojans, ransomware, and spyware. The document provides an overview of different types of malware for 10 minutes as part of a CompTIA Security+ study guide on the topic of types of malware.
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
Digital Banking in the Cloud: How Citizens Bank Unlocked Their MainframePrecisely
Inconsistent user experience and siloed data, high costs, and changing customer expectations – Citizens Bank was experiencing these challenges while it was attempting to deliver a superior digital banking experience for its clients. Its core banking applications run on the mainframe and Citizens was using legacy utilities to get the critical mainframe data to feed customer-facing channels, like call centers, web, and mobile. Ultimately, this led to higher operating costs (MIPS), delayed response times, and longer time to market.
Ever-changing customer expectations demand more modern digital experiences, and the bank needed to find a solution that could provide real-time data to its customer channels with low latency and operating costs. Join this session to learn how Citizens is leveraging Precisely to replicate mainframe data to its customer channels and deliver on their “modern digital bank” experiences.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
This presentation provides valuable insights into effective cost-saving techniques on AWS. Learn how to optimize your AWS resources by rightsizing, increasing elasticity, picking the right storage class, and choosing the best pricing model. Additionally, discover essential governance mechanisms to ensure continuous cost efficiency. Whether you are new to AWS or an experienced user, this presentation provides clear and practical tips to help you reduce your cloud costs and get the most out of your budget.
4. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Rundown of commands for Process text steams using filters:
commands Overview
4
• cat – concatenate files (or just show a single file without alteration)
• cut – cut chosen text out of each line of a file and display it.
• expand – expand tabs into spaces
• fmt – reformat a text file with a consistent right margin
• head – show the first few (10) lines of a file
• join – join lines of two files on a common field
• nl – print the file with numbered lines
• od – octal dump of a file (or hexadecimal).
• paste – print a number of files side by side
• pr – format for printing (split into pages or columns and add headers)
• sed – stream editor (search and replace, append, cut, delete and more)
• sort – sort in alphabetical order (and numerical order too)
• split – split a single input into multiple files
• tac – print the lines of a file from back to front (backwards cat)
• tail – print the last few lines of a file
• tr – character translation (e.g. upper case to lower case).
• unexpand – convert spaces to tabs (unlike expand).
• uniq – remove duplicate lines from a sorted file
• wc – word count (and line count, and byte count)
5. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Bash makes it possible to redirect the input and output of a command.
•Input - comes from the keyboard (ended by pressing Ctrl+D),
•Output and any errors - are displayed on the screen.
Redirection can change the input of a process, its output and the destination of the errors.
Input and output redirection
5
Redirection Effect of redirection
cmd < file Command reads input from a file
cmd > file Output of command goes to file
cmd 2> file Errors from the command go to a file
cmd >> file Output of a command is added to a file
cmd > file 2>&1 Output and Errors go to a file
cmd >& file
cmd &> file
cmd1 | cmd2 Output from command1 is input for command2
9. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
head – print out the first lines
Selecting parts of a file (Filters that print out various parts of the input they receive)
9
foo:~ # head /var/log/boot.log
Apr 7 08:28:22 foo allrc: syslogd startup succeeded
Apr 7 08:28:22 foo allrc: klogd startup succeeded
Apr 7 08:28:23 foo allrc: portmap startup succeeded
Apr 7 08:27:56 foo rc.sysinit: Mounting proc filesystem: succeeded
Apr 7 08:27:56 foo rc.sysinit: Unmounting initrd: succeeded
Apr 7 08:27:56 foo sysctl: net.ipv4.ip_forward = 0
Apr 7 08:27:56 foo sysctl: net.ipv4.conf.default.rp_filter = 1
Apr 7 08:27:56 foo sysctl: kernel.sysrq = 0
Apr 7 08:28:26 foo lpd: execvp: No such file or directory
Apr 7 08:27:56 foo sysctl: kernel.core_uses_pid = 1
Ex: By default head prints out the first 10 lines of a file.
10. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Selecting parts of a file (Filters that print out various parts of the input they receive)
10
foo:~ $ ls -l / | head -n 6
total 232
drwxr-xr-x 2 root root 4096 Feb 21 15:49 bin
drwxr-xr-x 3 root root 4096 Jan 7 10:25 boot
drwxr-xr-x 5 root root 20480 Jan 10 11:35 data
drwxr-xr-x 21 root root 118784 Apr 7 08:28 dev
drwxr-xr-x 64 root root 8192 Apr 7 08:28 etc
Ex: head can print out a specific number of lines from a file or a stream.
use head to extract an exact number of bytes from an input stream (rather than lines).
Here's how to get a copy of the partition sector of a disk (be careful with that redirection).
foo:~ # head -c 512 < /dev/hda > mbr
foo:~ # ls -la mbr
-rw-r--r-- 1 root root 512 Apr 7 10:27 mbr
11. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
tail – show the end of a file
Selecting parts of a file (Filters that print out various parts of the input they receive)
11
root@foo:root # tail /var/log/messages
Apr 7 11:19:34 foo dhcpd: Wrote 9 leases to leases file.
Apr 7 11:19:34 foo dhcpd: DHCPREQUEST for 10.0.0.169 from
00:80:ad:02:65:7c via eth0
Apr 7 11:19:35 foo dhcpd: DHCPACK on 10.0.0.169 to
00:80:ad:02:65:7c via eth0
Apr 7 11:20:01 foo kdm[1151]: Cannot convert Internet address
10.0.0.168 to host name
Apr 7 11:26:46 foo ipop3d[22026]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:55 foo ipop3d[22028]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:58 foo ipop3d[22035]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:27:01 foo ipop3d[22036]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:29:31 foo kdm[21954]: pam_unix2: session started for user
joe, service xdm
Apr 7 11:32:41 foo sshd[22316]: Accepted publickey for root from
10.0.0.143 port 1250 ssh2
Ex: tail is just like head, but it shows the tail end of the file.
tail can be used to watch a file as it grows.
Run the command tail –f /var/log/messages on one console and then log in on another virtual console.
tail –n 20 file or tail -20 file will show last 20 lines of file. tail –c 20 file will show last 20 characters of a file.
12. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
cut – pull out columns
Selecting parts of a file (Filters that print out various parts of the input they receive)
12
root@foo:root # tail /var/log/messages
Apr 7 11:19:34 foo dhcpd: Wrote 9 leases to leases file.
Apr 7 11:19:34 foo dhcpd: DHCPREQUEST for 10.0.0.169 from
00:80:ad:02:65:7c via eth0
Apr 7 11:19:35 foo dhcpd: DHCPACK on 10.0.0.169 to
00:80:ad:02:65:7c via eth0
Apr 7 11:20:01 foo kdm[1151]: Cannot convert Internet address
10.0.0.168 to host name
Apr 7 11:26:46 foo ipop3d[22026]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:55 foo ipop3d[22028]: connect from 10.0.0.10
(10.0.0.10)
Apr 7 11:26:58 foo ipop3d[22035]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:27:01 foo ipop3d[22036]: connect from 10.0.0.3 (10.0.0.3)
Apr 7 11:29:31 foo kdm[21954]: pam_unix2: session started for user
joe, service xdm
Apr 7 11:32:41 foo sshd[22316]: Accepted publickey for root from
10.0.0.143 port 1250 ssh2
Ex:
Cut can be used to select certain columns of the input stream.
Columns can be defined by either their position, or by being separated by field separators.
15. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Split can split files into more manageable parts, e.g. for FTP uploads. The parts can be recombined with cat.
Selecting parts of a file (Filters that print out various parts of the input they receive)
15
foo:~/download $ ls -la anomy-sanitizer-1.56.tar.gz
-rw-rw-r-- 1 georgem georgem 124356 Oct 22 18:37 anomysanitizer-1.56.tar.gz
foo:~/download $ split -b 32k anomy-sanitizer-1.56.tar.gz
foo:~/download $ ls -la x*
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xaa
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xab
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:48 xac
-rw-rw-r-- 1 georgem georgem 26052 Apr 7 11:48 xad
Ex:
Here's how to use cat to recombine parts (using md5sum to check whether the whole is equal to the sum of
the parts).
foo:~/download $ split -b 32k anomy-sanitizer-1.56.tar.gz part-
foo:~/download $ ls -la part-*
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-aa
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-ab
-rw-rw-r-- 1 georgem georgem 32768 Apr 7 11:49 part-ac
-rw-rw-r-- 1 georgem georgem 26052 Apr 7 11:49 part-ad
foo:~/download $ cat part-* > newfile
foo:~/download $ md5sum newfile anomy-sanitizer-1.56.tar.gz
1a977bad964b0ede863272114bfc2482 newfile
1a977bad964b0ede863272114bfc2482 anomy-sanitizer-1.56.tar.gz
18. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Sorting
18
uniq – discard duplicate lines - uniq is usually used with sort to discard duplicates
foo:~ $ cat /etc/passwd | cut -d : -f 4 | sort -n | fmt
0 0 0 0 0 0 1 2 4 7 12 13 14 25 26 28 29 30 32 37 38 42 43 47 48 50 51
69 74 77 80 89 99 100 500 501 503 504 505 506 507 509 511 512 65534
Ex: Ex: cutting fourth field out of the password file (the group ID) and sorting in numerical order.
fmt is used to make the results display on a single line.
Ex: Ex: same command pipeline, but are removing duplicates with uniq before formatting.
foo:~ $ cat /etc/passwd | cut -d : -f 4 | sort -n | uniq | fmt
0 1 2 4 7 12 13 14 25 26 28 29 30 32 37 38 42 43 47 48 50 51 69 74 77
80 89 99 100 500 501 503 504 505 506 507 509 511 512 65534
19. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
19
tr – character set translation - usually used for converting upper case to lower case.
Can also do other character translations. tr -d can remove specific characters from a stream.
foo:~ $ man -P cat man | tr A-Z a-z | less
Ex: Ex: Translating from UPPER CASE to lower case. Asking man to use cat as its pager, instead of
using less, and opening up the man page for man itself.
Translating from lower case to UPPER CASE:
foo:~ $ man -P cat man | tr a-z A-Z | less
Convert file names to lowercase.
foo:/windows/C $ for FILE in * ; do
mv "$FILE" $( echo "$FILE" | tr A-Z a-z ) ; done
Using tr -d to delete the Carriage Returns (r) from a file created with Windows Notepad.
foo:/windows/C $ tr -d 'r' notepad.dos.txt > notepad.unix.txt
Ex:
Ex:
Ex:
21. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
21
Joining /etc/passwd and /etc/shadow based on their first field (user name).
Since /etc/passwd and /etc/shadow use a colon to separate fields, it is necessary to use -t: option.
foo:~ # join -t : /etc/passwd /etc/shadow | head
root:x:0:0:root:/root:/bin/bash:$1$LHNUbu7U$oiuhqwd1oiuhqhAdiuHvA0:1
2146:0:99999:7:::
bin:x:1:1:bin:/bin:/sbin/nologin:*:11974:0:99999:7:::
daemon:x:2:2:daemon:/sbin:/sbin/nologin:*:11974:0:99999:7:::
adm:x:3:4:adm:/var/adm:/sbin/nologin:*:11974:0:99999:7:::
lp:x:4:7:lp:/var/spool/lpd:/sbin/nologin:*:11974:0:99999:7:::
sync:x:5:0:sync:/sbin:/bin/sync:*:11974:0:99999:7:::
shutdown:x:6:0:shutdown:/sbin:/sbin/shutdown:*:11974:0:99999:7:::
halt:x:7:0:halt:/sbin:/sbin/halt:*:11974:0:99999:7:::
mail:x:8:12:mail:/var/spool/mail:/sbin/nologin:*:11974:0:99999:7:::
Ex:
join allows to specify which particular field to join on, and also which particular fields should appear in the
output (similar to cut)
23. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
23
sed – stream editor - sed does text transformations on an input stream.
foo:~ $ echo "Hi Fred, how are you" | sed 's/Fred/Joe/'
Hi Joe, how are you
foo:~ $ echo "Hi Fred, how is Fred?" | sed 's/Fred/Joe/'
Hi Joe, how is Fred?
foo:~ $ echo "Hi Fred, how is Fred?" | sed 's/Fred/Joe/g‘
Hi Joe, how is Joe?
Ex:
•sed works by making only one pass over the inputs.
•A sed program consists of one or more sed commands which are applied to each line of the input.
•A command may be prefixed by an address range indicating lines for which the command performes.
sed commands: s/PATTERN/REPLACEMENT/g
search and replace. If you add g at end - search and replace applies as many times as possible to single line.
You can also use i at the end of the s command to make the search case insensitive
24. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
24
foo:~ $ ls /bin/ | fmt -30 | nl | sed '4,15d'
1 arch ash ash.static awk
2 basename bash cat chgrp
3 chmod chown chvt cp cpio csh
16 true umount uname usleep vi
17 vim zcat zsh
foo:~ $ ls /bin | fmt -40 | nl | sed '/e/ d'
2 bash cat chgrp chmod chown chvt cp
13 vi vim zcat zsh
Ex:
sed commands: d – delete the line.
You need to select the lines as explained in the next paragraph.
The commands are most useful if you specify a range of lines to which the command applies.
Here's how to specify specific lines for the d command:
•/PATTERN/d – delete all lines contains the pattern
•4d – delete line 4
•4,10d – delete lines 4 to 10
•6,$d – delete from line 6 to the last line
29. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Manipulation
29
paste – paste two files together
foo:~/tmp $ cat file2
cabbage green leaves
cat white cat-hair
piano brown wood
foo:~/tmp $ cat file1
cat animal
cabbage vegetable
piano mineral
coal mineral
foo:~/tmp $ paste file2 file1 | expand -t 22
cabbage green leaves cat animal
cat white cat-hair cabbage vegetable
piano brown wood piano mineral
coal mineral
Ex:
Using paste is like taking two printouts of two files and sticking the right margin of one to the left margin of
the other. The glue between the files is a tab stop.
You can specify a delimiter between the files which is not a tab space with option -d
30. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Formatting
30
fmt – format nicely - format text into a neatly word wrapped structure. Default right margin is 75
foo:~ $ ls /bin | fmt
arch ash ash.static aumix-minimal awk basename bash bash2 bsh cat
chgrp
chmod chown cp cpio csh cut date dd df dmesg dnsdomainname doexec
domainname dumpkeys echo ed egrep env ex false fgrep gawk gettext
grep gtar gunzip gzip hostname igawk ipcalc kbd_mode kill link ln
loadkeys login ls mail mkdir mknod mktemp more mount mt mv netstat
nice
nisdomainname pgawk ping ps pwd red rm rmdir rpm rvi rview sed
setfont
setserial sh sleep sort stty su sync tar tcsh touch true umount
uname
unicode_start unicode_stop unlink usleep vi vi.minimal view
ypdomainname
Ex:
foo:~ $ ls /bin | fmt -40 | head
arch ash ash.static aumix-minimal
awk basename bash bash2 bsh cat chgrp
chmod chown cp cpio csh cut date dd df
dmesg dnsdomainname doexec domainname
dumpkeys echo ed egrep env ex false
fgrep gawk gettext grep gtar gunzip
gzip hostname igawk ipcalc kbd_mode
kill link ln loadkeys login ls mail
mkdir mknod mktemp more mount mt mv
netstat nice nisdomainname pgawk ping
32. CoreLinuxforRedHatandFedoralearningunderGNUFreeDocumentationLicense-Copyleft(c)AcácioOliveira2012
Everyoneispermittedtocopyanddistributeverbatimcopiesofthislicensedocument,changingisallowed
Process text streams using filters
Formatting
32
od – octal dump (and other formats)
Ex:
foo:/tmp $ echo "Hello World" > hello
foo:/tmp $ cat hello
Hello World
od, oddly enough, does not just print out in octal, but in other formats.
foo:/tmp $ od hello
0000000 062510 066154 020157 067527 066162 005144
0000014
od's behaviour is rather odd when it is used without any arguments. It prints out the octal
value of two pairs of bytes in the file. Sometimes this is useful, but usually it isnot.
Using -t switch tells od to use a specific format type (the default is od –t o2).
od –t c means character format. You will notice that file ends in a newline character (n).
foo:/tmp $ od -t c hello
0000000 H e l l o W o r l d n
0000014
od –t d1 specifies decimal format, with a width of one. character encoding is ASCII.
Ex:
Ex:
foo:/tmp $ od -t d1 hello
0000000 72 101 108 108 111 32 87 111 114 108 100 10
0000014