Embedded Recipes 2018 - Yoctoception: Containers in the embedded world - Jéré...Anne Nicolas
Containerisation is a new player in the embedded world. Provisionning and rapid deployment doesn’t really make sense for embedded devices, but the extra security that container partitionning brings to the table is quickly becoming a “must have” for every embedded device.
However, the embedded world has its own constraints. Generic distributions like debian or Fedora are rarely used and tools like dockers can’t build software with custom build-chains the way Yocto or buildroot does it.
This talk will quickly review what are linux containers, why they are usefull in the embedde world, and then will study ways to generate container images using the yocto infrastructure and integrate them in another yocto image without breaking the package-based update system that Yocto provides.
There is no open source business model - only people selling complementary goods, and using an open source project as a development and distribution model.
What do you do when disaster strikes? In part 9 of our DB2 Support Nightmare series we look at another DB2 disaster scenario and how it was resolved by the experts at Triton Consulting.
Design patterns and plan for developing high available azure applicationsHimanshu Sahu
1. Design Patterns High Availability of Azure Applications
2. Practical Demo on points to take care for High Availability from Infrastructure point of view(the points we discussed in last seminar)
3. Different Patterns for High Availability
3.1 Health Endpoint Monitoring Pattern
3.2 Queue-based Load Leveling Pattern
3.2 Throttling Pattern
3.3 Retry Pattern
3.4 Multiple Datacenter Deployment Guidance
4. Architecture for High Availability of Azure Applications
5. best practices for developing High Available Azure Applications
Embedded Recipes 2018 - Yoctoception: Containers in the embedded world - Jéré...Anne Nicolas
Containerisation is a new player in the embedded world. Provisionning and rapid deployment doesn’t really make sense for embedded devices, but the extra security that container partitionning brings to the table is quickly becoming a “must have” for every embedded device.
However, the embedded world has its own constraints. Generic distributions like debian or Fedora are rarely used and tools like dockers can’t build software with custom build-chains the way Yocto or buildroot does it.
This talk will quickly review what are linux containers, why they are usefull in the embedde world, and then will study ways to generate container images using the yocto infrastructure and integrate them in another yocto image without breaking the package-based update system that Yocto provides.
There is no open source business model - only people selling complementary goods, and using an open source project as a development and distribution model.
What do you do when disaster strikes? In part 9 of our DB2 Support Nightmare series we look at another DB2 disaster scenario and how it was resolved by the experts at Triton Consulting.
Design patterns and plan for developing high available azure applicationsHimanshu Sahu
1. Design Patterns High Availability of Azure Applications
2. Practical Demo on points to take care for High Availability from Infrastructure point of view(the points we discussed in last seminar)
3. Different Patterns for High Availability
3.1 Health Endpoint Monitoring Pattern
3.2 Queue-based Load Leveling Pattern
3.2 Throttling Pattern
3.3 Retry Pattern
3.4 Multiple Datacenter Deployment Guidance
4. Architecture for High Availability of Azure Applications
5. best practices for developing High Available Azure Applications
High Availability Options for Modern Oracle InfrastructuresSimon Haslam
Today's enterprise architect has a bewildering array of choices when it comes to building a highly available infrastructure to run Oracle. This presentation considers approaches using the Oracle technology layer, resilient virtualisation (Oracle and other vendors), hardware clustering and storage replication. It covers the core Oracle Database and Fusion Middleware products and, based on practical experience, aims to give attendees a broad picture of alternatives with their pros and cons.
Delivered on 5 December 2011 at UKOUG 2011 by Simon Haslam and Julian Dyke.
SQLSaturday Bulgaria : HA & DR with SQL Server AlwaysOn Availability Groupsturgaysahtiyan
The AlwaysOn Availability Groups feature is a high-availability and disaster-recovery solution that provides an enterprise-level alternative to database mirroring. Introduced in SQL Server 2012, AlwaysOn Availability Groups maximizes the availability of a set of user databases for an enterprise. In this session we will talk about what’s coming with Always On, and how does it help to improve high availability and disaster recovery solutions.
A Step-By-Step Disaster Recovery Blueprint & Best Practices for Your NetBacku...Symantec
In this technical session we will share a few customer tested blueprints for implementing DR strategies with NetBackup appliances showing support for onsite and offsite disaster recovery. This includes the architecture design with Symantec best practices, down to execution of the wizards and command lines needed to implement the solution.
Watch the recording of this Google+ Hangout: http://bit.ly/13oTjvp
I gave this talk at Krakow/Poland DevOPS meetup. It was a lightning talk covering subject of High Availability solutions, architecture, planning and deploying.
SharePoint Backup And Disaster Recovery with Joel OlesonJoel Oleson
This walks through the various options around backup and restore with SharePoint. This deck was presented at Tech Ed South East Asia 2008 by Joel Oleson
AWS provides a platform that is ideally suited for building highly available systems, enabling you to build reliable, affordable, fault-tolerant systems that operate with a minimal amount of human interaction. This session covers many of the high-availability and fault-tolerance concepts and features of the various services that you can use to build highly reliable and highly available applications in the AWS Cloud: architectures involving multiple Availability Zones, including EC2 best practices and RDS Multi-AZ deployments; loosely coupled and self-healing systems involving SQS and Auto Scaling; networking best practices for high availability, including Elastic IP addresses, load balancing, and DNS; leveraging services that inherently are built with high-availability and fault tolerance in mind, including S3, Elastic Beanstalk and more.
Best Practices in Disaster Recovery Planning and TestingAxcient
Axcient and industry expert Paul Kirvan have put together this presentation on avoiding common disaster recovery mistakes and leveraging industry best practices to create a technology disaster recovery plan that works best for you.
This presentation gives you the many elements necessary of a well-executed disaster recovery plan, including:
- Guidelines for creating your own Disaster Recovery plan
- A checklist of key items to consider based on your business objectives
- The common mistakes and pitfalls to avoid
- Technology considerations for Disaster Recovery
- Tips for planning and executing a successful Disaster Recovery test
Whether you're in the process of creating a disaster recovery plan or you already have one in place, this presentation will guide you through the steps you need to follow to help ensure your plan is complete.
AWS re:Invent 2016: Disaster Recovery and Business Continuity for Systemicall...Amazon Web Services
Modern financial services organizations rely heavily on technology and automated systems to run business-as-usual. However, if this technology were interrupted by natural disasters or other events, there could be a devastating impact on investors and market participants, and in turn your reputational brand. In this session, we provide a step-by-step disaster recovery solution employed by a major exchange. This solution leverages Amazon EC2 Container Service to provide Docker containers, Weave Net to support a multicast overlay network that enables high volume multicast feeds in a cloud environment, and AWS CloudFormation for the ability to easily create and manage AWS assets. The session also covers the importance of redundancy (not just operationally, but for SEC compliance reasons as well) and how financial services organizations can increase geographical diversification of their primary and disaster recovery data centers. We dive deep into each major component of the solution.
High Availability can be a curiously nebulous term, and most people probably don't care about it until they can't access their online banking service, or their plane crashes.
This presentation examines some of the considerations necessary when building highly available computer systems, then focuses on the HA infrastructure software currently available from the Corosync/OpenAIS, Linux-HA and Pacemaker projects.
Originally presented at Linux Users Victoria in April 2010 (http://luv.asn.au/2010/04/06)
The A to Z Guide to Business Continuity and Disaster RecoverySirius
Companies often face challenges during business continuity and disaster recovery (BC/DR) planning. One of the key challenges is to reach consensus to ensure everyone at the company is on the same page. Therefore, it is important for the business and IT to have a comprehensive discussion about its current capabilities, needs, procedures and expectations for BC/DR.
To help with these conversations, we have developed an alphabetical guide and identified 26 important terms. This list is not meant to be exhaustive, but rather a good starting point for this discussion.
Business continuity and disaster recovery are not the same but complement each other. Planning on BCP and DRP is necessary for all business. This slide contains information on how to achieve and maintain them.
Relax and Recover (ReaR) is an open source bare metal disaster recovery solution for Linux (http://rear.sf.net). This session will introduce you to advanced Linux disaster recovery concepts and will feature a live demonstration on how to automatically recover a failed system with ReaR. Finally, this session will cover common best practice usage scenarios of ReaR and introduce you to basic setup and configuration for ReaR.
High Availability Options for Modern Oracle InfrastructuresSimon Haslam
Today's enterprise architect has a bewildering array of choices when it comes to building a highly available infrastructure to run Oracle. This presentation considers approaches using the Oracle technology layer, resilient virtualisation (Oracle and other vendors), hardware clustering and storage replication. It covers the core Oracle Database and Fusion Middleware products and, based on practical experience, aims to give attendees a broad picture of alternatives with their pros and cons.
Delivered on 5 December 2011 at UKOUG 2011 by Simon Haslam and Julian Dyke.
SQLSaturday Bulgaria : HA & DR with SQL Server AlwaysOn Availability Groupsturgaysahtiyan
The AlwaysOn Availability Groups feature is a high-availability and disaster-recovery solution that provides an enterprise-level alternative to database mirroring. Introduced in SQL Server 2012, AlwaysOn Availability Groups maximizes the availability of a set of user databases for an enterprise. In this session we will talk about what’s coming with Always On, and how does it help to improve high availability and disaster recovery solutions.
A Step-By-Step Disaster Recovery Blueprint & Best Practices for Your NetBacku...Symantec
In this technical session we will share a few customer tested blueprints for implementing DR strategies with NetBackup appliances showing support for onsite and offsite disaster recovery. This includes the architecture design with Symantec best practices, down to execution of the wizards and command lines needed to implement the solution.
Watch the recording of this Google+ Hangout: http://bit.ly/13oTjvp
I gave this talk at Krakow/Poland DevOPS meetup. It was a lightning talk covering subject of High Availability solutions, architecture, planning and deploying.
SharePoint Backup And Disaster Recovery with Joel OlesonJoel Oleson
This walks through the various options around backup and restore with SharePoint. This deck was presented at Tech Ed South East Asia 2008 by Joel Oleson
AWS provides a platform that is ideally suited for building highly available systems, enabling you to build reliable, affordable, fault-tolerant systems that operate with a minimal amount of human interaction. This session covers many of the high-availability and fault-tolerance concepts and features of the various services that you can use to build highly reliable and highly available applications in the AWS Cloud: architectures involving multiple Availability Zones, including EC2 best practices and RDS Multi-AZ deployments; loosely coupled and self-healing systems involving SQS and Auto Scaling; networking best practices for high availability, including Elastic IP addresses, load balancing, and DNS; leveraging services that inherently are built with high-availability and fault tolerance in mind, including S3, Elastic Beanstalk and more.
Best Practices in Disaster Recovery Planning and TestingAxcient
Axcient and industry expert Paul Kirvan have put together this presentation on avoiding common disaster recovery mistakes and leveraging industry best practices to create a technology disaster recovery plan that works best for you.
This presentation gives you the many elements necessary of a well-executed disaster recovery plan, including:
- Guidelines for creating your own Disaster Recovery plan
- A checklist of key items to consider based on your business objectives
- The common mistakes and pitfalls to avoid
- Technology considerations for Disaster Recovery
- Tips for planning and executing a successful Disaster Recovery test
Whether you're in the process of creating a disaster recovery plan or you already have one in place, this presentation will guide you through the steps you need to follow to help ensure your plan is complete.
AWS re:Invent 2016: Disaster Recovery and Business Continuity for Systemicall...Amazon Web Services
Modern financial services organizations rely heavily on technology and automated systems to run business-as-usual. However, if this technology were interrupted by natural disasters or other events, there could be a devastating impact on investors and market participants, and in turn your reputational brand. In this session, we provide a step-by-step disaster recovery solution employed by a major exchange. This solution leverages Amazon EC2 Container Service to provide Docker containers, Weave Net to support a multicast overlay network that enables high volume multicast feeds in a cloud environment, and AWS CloudFormation for the ability to easily create and manage AWS assets. The session also covers the importance of redundancy (not just operationally, but for SEC compliance reasons as well) and how financial services organizations can increase geographical diversification of their primary and disaster recovery data centers. We dive deep into each major component of the solution.
High Availability can be a curiously nebulous term, and most people probably don't care about it until they can't access their online banking service, or their plane crashes.
This presentation examines some of the considerations necessary when building highly available computer systems, then focuses on the HA infrastructure software currently available from the Corosync/OpenAIS, Linux-HA and Pacemaker projects.
Originally presented at Linux Users Victoria in April 2010 (http://luv.asn.au/2010/04/06)
The A to Z Guide to Business Continuity and Disaster RecoverySirius
Companies often face challenges during business continuity and disaster recovery (BC/DR) planning. One of the key challenges is to reach consensus to ensure everyone at the company is on the same page. Therefore, it is important for the business and IT to have a comprehensive discussion about its current capabilities, needs, procedures and expectations for BC/DR.
To help with these conversations, we have developed an alphabetical guide and identified 26 important terms. This list is not meant to be exhaustive, but rather a good starting point for this discussion.
Business continuity and disaster recovery are not the same but complement each other. Planning on BCP and DRP is necessary for all business. This slide contains information on how to achieve and maintain them.
Relax and Recover (ReaR) is an open source bare metal disaster recovery solution for Linux (http://rear.sf.net). This session will introduce you to advanced Linux disaster recovery concepts and will feature a live demonstration on how to automatically recover a failed system with ReaR. Finally, this session will cover common best practice usage scenarios of ReaR and introduce you to basic setup and configuration for ReaR.
I have described all about linux OS starting from basics.
I guess this PPT will really be very very helpful for you guys.
This was one of the most appreciable PPT in my time when i presented it in my class.
This presentation will provide the information about the Linux Root File systems and its hierarchy. So any technocrate who is willing to gain info about root files of Linux can easily understand . preffered for Embedded system design Students who are pursuing diploma courses in various CDAC centers.
Open Source Backup Conference 2014: Rear, by Ralf DannertNETWAYS
ReaR(Relax and Recover) is delivered as part of the SUSE Linux High Availability Extension.
We show -by way of example- how corporations integrate ReaR during Preparation, Testing and Recovery as buildingblock of their disaster recovery strategy.In the technical part we will highlight the AutoYaST/YaST integration with rear-suse.
We will also investigate some of the adaptations, that had to be done to make ReaR work with upcoming SLES12, that will include systemd and grub2 to be able to automatically recover btrfs subvolumes.
Chapter 8 Common Forensic ToolsOverviewIn this chapter, youl.docxchristinemaritza
Chapter 8: Common Forensic Tools
Overview
In this chapter, you'll learn more about:
· Explore disk imaging tools, forensic software tool sets, and miscellaneous software tools
· Understand computer forensic hardware
· Assemble your forensic tool kit
The first steps in any investigation nearly always involve old-fashioned detective work. As a forensic investigator, you need to observe and record your observations first. Once you start examining media contents, you'll need some tools to help you find and make sense of stored data.
Forensic investigators and computer examiners need several different types of tools to identify and acquire computer evidence. Some evidence is hidden from the casual observer and requires specialized tools to find and access. In this chapter, we'll examine a sampling of some common and popular tools available to carry out computer forensic tasks.
Disk Imaging and Validation Tools
After identifying the physical media that they suspect contains evidence, forensic investigators must make sure media is preserved before any further steps are taken. Preserving the media is necessary to provide assurance the evidence acquired is valid.
Chapter 3, "Computer Evidence," and Chapter 4, "Common Tasks," both emphasize the importance of copying all media first and then analyzing the copy. It's usually best to create an exact image of the media and verify that it matches the original before continuing the investigation. It's rare to examine the original evidence for any investigation that might end up in court. For other types of investigations, however, forensic investigators might perform a targeted examination on the original evidence. For example, assume the job is to examine a user's home folder on a server for suspected inappropriate material. It might be impossible or extremely difficult to create a mirror image of the disk drive, but the disk can be scanned for existing or deleted files while it is in use. Although examining media while in use might not always be the best practice, informal investigations use this technique frequently.
To Copy or Not to Copy?
Whenever possible, create a duplicate of the original evidence, verify the copy, and then examine the copy. Always invest the time and effort to copy original media for any investigation that might end up in a court of law. If you are sure your investigation will not end up in court, you might decide to analyze the original evidence directly. This is possible and desirable in cases where copying media would cause service interruptions.
Your choice of tools to use depends on several factors, including:
· Operating system(s) supported
Operating system(s) in which the tool runs
File systems the tool supports
· Price
· Functionality
· Personal preference
The following sections list some tools used to create and verify media copies. Some products appear in two places in the chapter. That's because several products play multiple roles. This section lists several products ...
Business Continuity Planning with Bareos and rear (Loadays 2015)Gratien D'haese
Business Continuity Planning explained in details with examples and going a bit deeper with the best Open Source Linux Disaster Recovery framework "Relax-and-Recover" integrated with the Open Source backup solution BAREOS.
Upgrade-UX is an open source framework developed to assist in patching and/or updating Unix Operating Systems in a consistent and repeatable way. Especially in the industry it is forbidden just to run yum update (on Linux) to update your Linux system, therefore, upgrade-ux may proof to be a handy tool to guide you through the patching and/or update process as it follows a track you control (evidence gathering, pre/post executing of scripts, logging, and so on).
Presenting adhocr (abbreviation for Ad-hoc copy and run) as a simple, but powerful UNIX administrator tool. If you like to retrieve data or execute commands on lots of systems simultaneously then this tool is your friend. There is no need to exchange your ssh keys as the power behind adhocr is the expect tool (language). For example, it is plain easy to use adhocr to distribute your public ssh key to all your systems. The real power of adhocr is the central point of logging, which is perfect for \'grep\'ing into stuff you\'re looking for.
You also have the ability to execute commands via the \'sudo su -\' way, which is a blessing in environments where root is not permitted to login directly.
You can even use it monitoring your systems thanks to the powerful error catching.
cd /tmp/mindi-busybox-1.7.3 make oldconfig make busybox make CONFIG_PREFIX=/usr/local/lib/mindi/rootfs install ---------------------------------------------------------------------------- cd /tmp/mindi-2.0.0 ./install.sh mindi 2.0.0-r1883 will be installed under /usr/local Creating target directories ... Copying files ... # mindi Mindi Linux mini-distro generator v2.0.0-r1883 Latest Mindi is available from http://www.mondorescue.org BusyBox sources are available from http://www.busybox.net ------------------------------------------------------------------------------ Mindi-BusyBox v1.7.3-r1873 (2008-02-26 12:19:11 CET) multi-call binary Do you want to use your own kernel to build the boot disk ([y]/n) ? find: /boot/efi: No such file or directory No kernel matches exactly. Are there any duff kernels? Sorry, no duff kernels either Could not find your kernel. Analyzing dependency requirements Done. Making complete dependency list Done. Analyzing your keyboard's configuration. Searching for rc.config ...Unknown config detected. Default keyboard map will be used. Assembling dependency files.................................... Done. Your mountlist will look like this: 8 directories. lvmdiskscan not found. Won't handle LVM. DEVICE MOUNTPOINT FORMAT SIZE (MB) LABEL/UUID /dev/sdb2 / ext3 66891 /dev/sdb3 swap swap 1953 Tarring and zipping the groups........................... Done. Creating data disk #1...#2...#3...#4...#5...#6...#7...#8... Done. FATAL ERROR. PBDI - cannot find 57344 kernel Please e-mail a copy of /var/log/mindi.log to the mailing list. See http://www.mondorescue.org for more information. WE CANNOT HELP unless you enclose that file. # mkdir -m 755 /boot/efi # mount -t vfat /dev/sdb1 /boot/efi # mindi Mindi Linux mini-distro generator v2.0.0-r1883 Latest Mindi is available from http://www.mondorescue.org BusyBox sources are available from http://www.busybox.net ------------------------------------------------------------------------------ Mindi-BusyBox v1.7.3-r1873 (2008-02-26 12:19:11 CET) multi-call binary Do you want to use your own kernel to build the boot disk ([y]/n) ? Analyzing dependency requirements Done. Making complete dependency list Done. Analyzing your keyboard's configuration. Searching for rc.config ...Unknown config detected. Default keyboard map will be used. Assembling dependency files.....................................Done. Your mountlist will look like this: 8 directories. lvmdiskscan not found. Won't handle LVM. DEVICE MOUNTPOINT FORMAT SIZE (MB) LABEL/UUID /dev/sdb2 / ext3 66891 /dev/sdb3 swap swap 1953 Tarring and zipping the groups........................... Done. Creating data disk #1...#2...#3...#4...#5...#6...#7...#8... Done. Making 16384KB boot disk...............udev device manager found ERROR: No product name found for Hardware support Hardware Information found and saved ... ......22934 blocks ......OK, you don't have a /boot/boot.b file, which is odd because most _good_ Linux distributions come with one, even if it's only a softlink grep: /etc/lilo.conf: No such file or directory Nor can I find it from your /etc/lilo.conf file. This is very odd. I'm going to use '' CBBF -- warning -- cannot find your boot.b file. That's it, I quit... (j/k) cp: `/tmp/mindi.BzNyJ29838/mountpoint.29836/elilo.conf' and `/tmp/mindi.BzNyJ29838/mountpoint.29836/elilo.conf' are the same file In the directory '/var/cache/mindi' you will find the images:- Done. mindi-bootroot.16384.img mindi-data-1.img mindi-data-2.img mindi-data-3.img mindi-data-4.img mindi-data-5.img mind i-data-6.img mindi-data-7.img mindi-data-8.img Would you like to create boot+data floppy disks now (y/[n]) ? Shall I make a bootable USB image ? (y/[n]) Shall I make a bootable CD image? (y/[n]) y Created bootable ISO image at /var/cache/mindi/mindi.iso Finished. Boot and data disk images were created. hpx189:/tmp/mondo-2.2.5# ls -l /var/cache/mindi/mindi.iso -rw-r--r-- 1 root root 54169600 Feb 26 14:05 /var/cache/mindi/mindi.iso ----------------------------------------------------------------------------- # cd mondo-2.2.5 # ./configure # make # make install
On Debian/IA64 # mondoarchive W E L C O M E T O M O N D O R E S C U E Making catalog of files to be backed up Done. Dividing filelist into sets Done. Copying Mondo's core files to the scratch directory Done. Calling MINDI to cr+-----¦ Generating boot+data disks +-----+ ¦ ¦ ¦ Generating boot+data disks ¦ ¦ ¦ ¦ 1% ¦ ¦ ¦ ¦ ¦ ¦ Working.................... / ¦ +----------------------------------------+ DEVICE MOUNTPOINT FORMAT SIZE (MB) LABEL/U /dev/sda2 lvm lvm 0 /dev/sdb2 / ext3 66891 /dev/sdb3 swap swap 1953 Tarring and zipping the groups................................. Generating boot+data disks
Problem on Debian/IA64 via LAN Console => no reaction on keyboard because of newt and dialog
Only problem with mkCDrec v0.9.8 was the parted version. On RedHat Enterprise the layout of parted looks like: Disk geometry for /dev/sdb: 0.000-34732.890 megabytes Disk label type: gpt Minor Start End Filesystem Name Flags 1 0.017 100.002 fat16 boot 2 100.002 32740.452 ext3 32740.452 34732.873 linux-swap And on Debian/IA64 it looks: Disk /dev/sdb: 73.4GB Sector size (logical/physical): 512B/512B Partition Table: gpt Number Start End Size File system Name Flags 1 17.4kB 100MB 100MB fat16 boot 2 100MB 71.4GB 71.3GB ext3 3 71.4GB 73.4GB 2049MB linux-swap Needed to change rd-base.sh script to get a good partitioning file. Next release (v0.9.9) will have the updated code. For the rest no difference with RedHat (for booting and recovering).
# rear -s mkrescue Relax & Recover Version 1.6 / 2007-12-05 Simulation mode activated, ReaR base directory: /usr/share/rear Source prep/default/01_progress_start.sh Source prep/ISO/default/30_check_iso_dir.sh Source prep/ISO/default/32_check_cdrom_size.sh Source prep/ISO/Linux-ia64/33_find_elilo_efi.sh Source prep/ISO/Linux-ia64/34_define_console_ia64.sh Source prep/default/99_progress_stop.sh Source dr/default/01_mk_config_dir_recovery.sh Source dr/GNU/Linux/10_describe_physical_devices.sh Source dr/GNU/Linux/11_describe_mountpoint_device.sh Source dr/GNU/Linux/12_describe_filesystems.sh Source dr/GNU/Linux/13_describe_swap.sh Source dr/GNU/Linux/15_copy_proc_partitions.sh Source dr/GNU/Linux/21_describe_md.sh Source dr/GNU/Linux/23_describe_lvm2.sh Source dr/GNU/Linux/29_find_required_devices.sh Source dr/Linux-ia64/30_mk_partitions_with_parted.sh Source dr/Linux-ia64/31_describe_device_properties.sh Source dr/GNU/Linux/80_copy_fstab_file.sh Source rescue/default/00_set_recovery_config.sh Source rescue/default/01_merge_skeletons.sh Source rescue/default/10_hostname.sh Source rescue/default/20_etc_issue.sh Source rescue/GNU/Linux/30_dns.sh Source rescue/GNU/Linux/31_network_devices.sh Source rescue/GNU/Linux/35_routing.sh Source rescue/GNU/Linux/40_kernel_modules.sh Source rescue/default/43_prepare_timesync.sh Source rescue/GNU/Linux/50_clone_keyboard_mappings.sh Source rescue/default/50_ssh.sh Source build/GNU/Linux/20_copy_as_is.sh Source build/GNU/Linux/39_copy_binaries_libraries.sh Source build/GNU/Linux/40_copy_modules.sh Source build/default/50_patch_sshd_config.sh Source pack/GNU/Linux/00_create_symlinks.sh Source pack/GNU/Linux/90_create_initramfs.sh Source output/ISO/Linux-ia64/20_mount_bootimg.sh Source output/ISO/Linux-ia64/30_create_bootimg.sh Source output/ISO/Linux-ia64/40_create_local_efi_dir.sh Source output/ISO/Linux-ia64/70_umount_bootimg.sh Source output/ISO/Linux-ia64/80_create_isofs.sh Source output/default/95_email_result_files.sh Source cleanup/default/01_progress_start.sh Source cleanup/default/99_progress_stop.sh Finished in 0 seconds. The only problem we had was with agetty which doesn’t exist on Debian (getty), but forgot to change inittab file. Solution is quite simple: /usr/share/rear/skel/Linux-ia64# ls -l sbin total 0 lrwxrwxrwx 1 root root 11 Mar 13 16:09 agetty -> /sbin/getty Looking through the /tmp/rear.log file we can see the following: 2008-03-13 16:15:06 Including rescue/default/01_merge_skeletons.sh 2008-03-13 16:15:06 Adding 'default' 2008-03-13 16:15:06 Adding 'Linux-ia64' 2008-03-13 16:15:06 Adding 'GNU/Linux' 2008-03-13 16:15:06 No 'Debian/default' or 'Debian/default.tar.gz' found 2008-03-13 16:15:06 No ' Debian/ia64 ' or 'Debian/ia64.tar.gz' found 2008-03-13 16:15:06 No 'Debian/4.0' or 'Debian/4.0.tar.gz' found 2008-03-13 16:15:06 No 'REQUESTRESTORE' or 'REQUESTRESTORE.tar.gz' found