Menu

Serious performance issues with disks and partitions

2022-01-05
2022-01-20
  • Bernard Michaud

    Bernard Michaud - 2022-01-05

    The problem started with stable version clonezilla-live-20211116-impish-amd64. The previous stable version, clonezilla-live-20210817-hirsute-amd64, worked fairly well considering the number of partitions with and without logical volumes that I have to back up.

    The machine is a quite old but extremely reliable Hewlett-Packard Compaq 6200 Pro SFF PC with a 64-bit Core Intel Quad Core i5-2400 and 8GB of RAM. It has two internal and three external hard disk drives, two of the latter in a dual enclosure connected through USB3, and the other in a single-drive enclosure connected through eSATA.

    As a retired IT professional, I have the peculiar hobby of “collecting” Linux distributions, and the HP has fourteen of them installed in a three partition layout each, all but one using logical volumes (lvm2). The system disks are as follows:

    ata1: SATA link up 6.0 Gbps
    
    [sda] Seagate Constellation ES.3 (ST6000NM0044) 6TB 128MB Cache 7200RPM SATA
          6.0Gb/s 3.5" Internal HDD
    
    ata2: SATA link up 3.0 Gbps
    
    [sdb] Seagate Barracuda 7200.12 (ST3500413AS) 500GB 7200RPM 16MB Cache SATA 6.0Gb/s 3.5" Internal HDD
    
    ata5: SATA link up 3.0 Gbps
    
        StarTech.com S3510BMU33ET USB 3.0/eSATA Trayless 3.5" SATA III HDD External     Enclosure with UASP
        |__ [sde] Seagate Barracuda 7200.12 (ST31000528AS) 1TB 7200RPM SATA 3GB/s 
                    32MB Cache 3.5" Internal HDD
    
    Bus 04.Port 1: Dev 1, Class=root_hub, Driver=xhci_hcd/4p, 5000M
        ID 1d6b:0003 Linux Foundation 3.0 root hub
        |__ Port 3: Dev 2, If 0, Class=Mass Storage, Driver=uas, 5000M
            ID 152d:9561 JMicron Technology Corp. / JMicron USA Technology Corp.
            |__ StarTech.com S3520BU33ER USB 3.0/eSATA Trayless Dual 3.5" SATA III HDD
                External RAID Enclosure with UASP
                |__ [sdc] Seagate Barracuda XT (ST33000651AS) 3TB 7200RPM 64MB Cache 
                          SATA 6.0Gb/s 3.5" Internal HDD
                |__ [sdd] Seagate Barracuda 7200.12 (ST31000528AS) 1TB 7200RPM SATA 
                          3GB/s 32MB Cache 3.5" Internal HDD
    

    This is the backup disk, connected through USB3:

    Bus 04.Port 1: Dev 1, Class=root_hub, Driver=xhci_hcd/4p, 5000M
        ID 1d6b:0003 Linux Foundation 3.0 root hub
        |__ Port 2: Dev 3, If 0, Class=Mass Storage, Driver=uas, 5000M
            ID 0bc2:ab21 Seagate RSS LLC Backup Plus Slim
            |__ [sdf] Seagate Backup Plus Slim 2GB External HDD
    

    The disks are not especially fast but have proven to be dependable. Backing up all those Linux distributions and several independent filesystems are an all-day affair, particularly with maximum compression, but it was manageable.

    The partitions and filesystems are as follows:

    NAME                   FSTYPE        SIZE PARTLABEL
    sda                                  5.5T 
    ├─sda1                 vfat           60M EFI system partition
    ├─sda2                 ext3          200M Linux filesystem
    ├─sda3                 LVM2_member   200G Linux LVM
    │ ├─RHEL6vg-lv00       swap            8G 
    │ ├─RHEL6vg-lv01       ext4            4G 
    │ ├─RHEL6vg-lv02       ext4           20G 
    │ ├─RHEL6vg-lv03       ext4           10G 
    │ ├─RHEL6vg-lv04       ext4           10G 
    │ ├─RHEL6vg-lv05       ext4           20G 
    │ ├─RHEL6vg-lv06       ext4           10G 
    │ ├─RHEL6vg-lv07       ext4           20G 
    │ ├─RHEL6vg-lv08       ext4           10G 
    │ └─RHEL6vg-lv09       ext4           10G 
    ├─sda4                 vfat           50M EFI system partition
    ├─sda5                 ext4          450M Linux filesystem
    ├─sda6                 LVM2_member   200G Linux LVM
    │ ├─fedoravg-lv00      swap            8G 
    │ ├─fedoravg-lv01      ext4           40G 
    │ ├─fedoravg-lv02      ext4           10G 
    │ ├─fedoravg-lv03      ext4           20G 
    │ ├─fedoravg-lv04      ext4           20G 
    │ ├─fedoravg-lv05      ext4           10G 
    │ ├─fedoravg-lv06      ext4           10G 
    │ ├─fedoravg-lv07      ext4           10G 
    │ └─fedoravg-lv08      ext4           10G 
    ├─sda7                 vfat           50M EFI system partition
    ├─sda8                 ext4          450M Linux filesystem
    ├─sda9                 LVM2_member   200G Linux LVM
    │ ├─mintvg-lv00        swap            8G 
    │ ├─mintvg-lv01        ext4           40G 
    │ ├─mintvg-lv02        ext4           10G 
    │ ├─mintvg-lv03        ext4           20G 
    │ ├─mintvg-lv04        ext4           20G 
    │ ├─mintvg-lv05        ext4           10G 
    │ ├─mintvg-lv06        ext4           10G 
    │ ├─mintvg-lv07        ext4           10G 
    │ └─mintvg-lv08        ext4           10G 
    ├─sda10                vfat           50M EFI system partition
    ├─sda11                ext4          450M Linux filesystem
    ├─sda12                LVM2_member   200G Linux LVM
    │ ├─ubuntuvg-lv00      swap            8G 
    │ ├─ubuntuvg-lv01      ext4           40G 
    │ ├─ubuntuvg-lv02      ext4           10G 
    │ ├─ubuntuvg-lv03      ext4           20G 
    │ ├─ubuntuvg-lv04      ext4           20G 
    │ ├─ubuntuvg-lv05      ext4           10G 
    │ ├─ubuntuvg-lv06      ext4           10G 
    │ ├─ubuntuvg-lv07      ext4           10G 
    │ └─ubuntuvg-lv08      ext4           10G 
    ├─sda13                vfat           50M EFI system partition
    ├─sda14                ext4          450M Linux filesystem
    ├─sda15                LVM2_member   200G Linux LVM
    │ ├─debianvg-lv00      swap            8G 
    │ ├─debianvg-lv01      xfs            40G 
    │ ├─debianvg-lv02      xfs            10G 
    │ ├─debianvg-lv03      xfs            20G 
    │ ├─debianvg-lv04      xfs            20G 
    │ ├─debianvg-lv05      xfs            10G 
    │ ├─debianvg-lv06      xfs            10G 
    │ ├─debianvg-lv07      xfs            10G 
    │ └─debianvg-lv08      xfs            10G 
    ├─sda16                vfat           50M EFI system partition
    ├─sda17                ext4          250M Linux filesystem
    ├─sda18                LVM2_member   200G Linux LVM
    │ ├─archvg-lv00        swap            8G 
    │ ├─archvg-lv01        ext4           40G 
    │ ├─archvg-lv02        ext4           10G 
    │ ├─archvg-lv03        ext4           20G 
    │ ├─archvg-lv04        ext4           20G 
    │ ├─archvg-lv05        ext4           10G 
    │ ├─archvg-lv06        ext4           10G 
    │ ├─archvg-lv07        ext4           20G 
    │ └─archvg-lv08        ext4           20G 
    ├─sda19                vfat           50M EFI system partition
    ├─sda20                ext4          450M Linux filesystem
    ├─sda21                LVM2_member   200G Linux LVM
    │ ├─mageiavg-lv00      swap            8G 
    │ ├─mageiavg-lv01      ext4           40G 
    │ ├─mageiavg-lv02      ext4           10G 
    │ ├─mageiavg-lv03      ext4           20G 
    │ ├─mageiavg-lv04      ext4           20G 
    │ ├─mageiavg-lv05      ext4           10G 
    │ ├─mageiavg-lv06      ext4           10G 
    │ ├─mageiavg-lv07      ext4           10G 
    │ └─mageiavg-lv08      ext4           10G 
    ├─sda22                vfat           50M EFI system partition
    ├─sda23                ext4          450M Linux filesystem
    ├─sda24                LVM2_member   200G Linux LVM
    │ ├─tumbleweedvg-lv00  swap            8G 
    │ ├─tumbleweedvg-lv01  ext4           40G 
    │ ├─tumbleweedvg-lv02  ext4           10G 
    │ ├─tumbleweedvg-lv03  ext4           20G 
    │ ├─tumbleweedvg-lv04  ext4           20G 
    │ ├─tumbleweedvg-lv05  ext4           10G 
    │ ├─tumbleweedvg-lv06  ext4           10G 
    │ ├─tumbleweedvg-lv07  ext4           10G 
    │ └─tumbleweedvg-lv08  ext4           10G 
    ├─sda25                vfat           50M EFI system partition
    ├─sda26                ext4          250M Linux filesystem
    ├─sda27                LVM2_member   200G Linux LVM
    │ ├─antergosvg-lv00    swap            8G 
    │ ├─antergosvg-lv01    ext4           40G 
    │ ├─antergosvg-lv02    ext4           10G 
    │ ├─antergosvg-lv03    ext4           20G 
    │ ├─antergosvg-lv04    ext4           40G 
    │ ├─antergosvg-lv05    ext4           10G 
    │ ├─antergosvg-lv06    ext4           10G 
    │ ├─antergosvg-lv07    ext4           10G 
    │ └─antergosvg-lv08    ext4           10G 
    ├─sda28                vfat          300M EFI system partition
    ├─sda29                swap            8G Linux swap
    ├─sda30                ext4           60G Linux filesystem
    ├─sda31                vfat           50M EFI System Partition
    ├─sda32                xfs           850M Linux filesystem
    ├─sda33                LVM2_member   200G Linux LVM
    │ ├─almavg-lv00        swap            8G 
    │ ├─almavg-lv01        xfs            40G 
    │ ├─almavg-lv02        xfs            10G 
    │ ├─almavg-lv03        xfs            20G 
    │ ├─almavg-lv04        xfs            20G 
    │ ├─almavg-lv05        xfs            10G 
    │ ├─almavg-lv06        xfs            10G 
    │ ├─almavg-lv07        xfs            10G 
    │ └─almavg-lv08        xfs            10G 
    ├─sda34                vfat           50M EFI system partition
    ├─sda35                xfs           850M Linux filesystem
    ├─sda36                LVM2_member   200G Linux LVM
    │ ├─rockyvg-lv00       swap            8G 
    │ ├─rockyvg-lv01       xfs            40G 
    │ ├─rockyvg-lv02       xfs            10G 
    │ ├─rockyvg-lv03       xfs            20G 
    │ ├─rockyvg-lv04       xfs            20G 
    │ ├─rockyvg-lv05       xfs            10G 
    │ ├─rockyvg-lv06       xfs            10G 
    │ ├─rockyvg-lv07       xfs            10G 
    │ └─rockyvg-lv08       xfs            10G 
    ├─sda37                vfat           50M EFI system partition
    ├─sda38                xfs           850M Linux filesystem
    ├─sda39                LVM2_member   200G Linux LVM
    │ ├─oraclelinuxvg-lv00 swap            8G 
    │ ├─oraclelinuxvg-lv01 xfs            40G 
    │ ├─oraclelinuxvg-lv02 xfs            10G 
    │ ├─oraclelinuxvg-lv03 xfs            20G 
    │ ├─oraclelinuxvg-lv04 xfs            20G 
    │ ├─oraclelinuxvg-lv05 xfs            10G 
    │ ├─oraclelinuxvg-lv06 xfs            10G 
    │ ├─oraclelinuxvg-lv07 xfs            10G 
    │ └─oraclelinuxvg-lv08 xfs            10G 
    ├─sda40                               50M EFI system partition
    ├─sda41                              850M Linux filesystem
    ├─sda42                              200G Linux LVM
    ├─sda43                vfat           50M EFI system partition
    ├─sda44                ext4          350M Linux filesystem
    └─sda45                LVM2_member   200G Linux LVM
      ├─sabayonvg-lv00     swap            8G 
      ├─sabayonvg-lv01     ext4           40G 
      ├─sabayonvg-lv02     ext4           10G 
      ├─sabayonvg-lv03     ext4           20G 
      ├─sabayonvg-lv04     ext4           20G 
      ├─sabayonvg-lv05     ext4           10G 
      ├─sabayonvg-lv06     ext4           10G 
      ├─sabayonvg-lv07     ext4           10G 
      └─sabayonvg-lv08     ext4           10G 
    
    sdb                                465.8G 
    ├─sdb1                 ntfs          100M 
    ├─sdb2                 ntfs        456.1G 
    └─sdb3                 ntfs          9.5G 
    
    sdc                                  2.7T 
    ├─sdc1                 ntfs          100G Microsoft basic data
    ├─sdc2                 ext4          100G Linux filesystem
    ├─sdc3                 ext4          300G Linux filesystem
    └─sdc4                 crypto_LUKS   320G Linux filesystem
    
    sdd                                931.5G 
    ├─sdd1                 ntfs         87.9G 
    └─sdd2                 ntfs        843.6G 
    
    sde                                931.5G 
    └─sde1                 ext4          400G Linux filesystem
    

    The problem I have is that partition and disk detection has become extremely slow starting with stable version clonezilla-live-20211116-impish-amd64. I have also tested with clonezilla-live-20220103-impish-amd64 with the same results. At issue is the part were the partitions to be backed up have to be selected:

    Finding all disks and partitions..
    Excluding busy partition..
    Excluding linux raid member partitions..
    Finding partitions..
    Partition number: 56

    With version clonezilla-live-20210817-hirsute-amd64, this would take barely one minutes; since version clonezilla-live-20211116-impish-amd64, this takes on average between seven and nine minutes:

    2022-01-05 12:52:00.051918200 -0500 ocs-cache/disk_list.cache
    2022-01-05 12:52:00.055918200 -0500 ocs-cache/pttable_for_disk.txt
    2022-01-05 12:54:59.319922200 -0500 ocs-cache/dev_fs_size_type.cache
    2022-01-05 12:55:15.015922500 -0500 ocs-cache/pttable_blkid_for_dev.txt
    2022-01-05 12:58:51.935927300 -0500 ocs-cache/pttable_for_part.txt
    2022-01-05 12:58:51.935927300 -0500 ocs-cache/part_list.cache

    This means that the waiting time has gone from roughly fourteen minutes (14x1 min.) to an hour and thirty-eight minutes (14x7) to back up all fourteen Linux distributions, not counting additional partitions.

    For now I am sticking with clonezilla-live-20210817-hirsute-amd64. I apologise for the slightly excessive amount of data I’ve provided, but I am hoping it will help in diagnosing this issue.

     

    Last edit: Bernard Michaud 2022-01-05
  • Steven Shiau

    Steven Shiau - 2022-01-06

    So could you please give testing Clonezilla live a try, e.g. 2.8.1-12 or 20220103-*:
    https://clonezilla.org/downloads.php
    We have made some improvements about this issue.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-06

    Hello Steven,

    I have should have mentioned that the last test was in fact performed with test version clonezilla-live-20220103-impish-amd64. This is why I wrote in my initial submission that the issue started with clonezilla-live-20211116-impish-amd64.

     
  • Steven Shiau

    Steven Shiau - 2022-01-06

    Not sure if it's related to Linux kernel. If so, please give 2.8.1-12 or 20220103-jammy a try?
    They come with newer Linux kernel so the results might be different.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-06

    I tried clonezilla-live-20220103-jammy-amd64 with the 5.13.0-19 kernel, and if anything it was even worse. First, the devices were out of sequence, as if udisks was not properly correlating the device entries to their corresponding AHCI port numbers. I had to restart Clonezilla three times until they came out in their proper sequence. I can’t explain why that was, but I seem to recall that this would sometimes happen on Ubuntu. Then, the backup drive selection took over ten minutes, worse than clonezilla-live-20220103-impish-amd64. Finally, tried an image backup, and this was again much worse than before:

    2022-01-05 23:00:07.504062200 -0500 ocs-cache-jammy/pttable_for_disk.txt
    2022-01-05 23:00:07.504062200 -0500 ocs-cache-jammy/disk_list.cache
    2022-01-05 23:12:23.720085800 -0500 ocs-cache-jammy/dev_fs_size_type.cache
    2022-01-05 23:12:43.728086400 -0500 ocs-cache-jammy/pttable_blkid_for_dev.txt
    2022-01-05 23:19:41.924099800 -0500 ocs-cache-jammy/part_list.cache
    2022-01-05 23:19:41.928099800 -0500 ocs-cache-jammy/pttable_for_part.txt

    The dev/partition/filesystem processing took over nineteen minutes. This seems utterly broken, and as such all recent versions of Clonezilla are essentially unusable on this machine. It runs fairly fast on my HP Spectre laptop which has a 1TB nvme SSD, which admittedly is a much more recent machine, so this might be an issue with SATA/SCSI hard drives.

     

    Last edit: Bernard Michaud 2022-01-06
  • Steven Shiau

    Steven Shiau - 2022-01-06

    Since you mentioned it's a old machine, could you please give Clonezilla live 2.1.8-12 (Debian-based) a try?

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-06

    I just tried it. Under Debian, the disk devices were in the proper sequence and the backup disk selection took about the same time as under Ubuntu, but the disk and partition processing for an image backup ran even slower than for Ubuntu, from 2022-01-06 00:57:33 to 2022-01-06 02:06:41, an unbelievable hour and nine minutes.

    I must reiterate that version clonezilla-live-20210817-hirsute-amd64 runs fast, devices, partitions and filesystems being processed in around a minute, while every version since 2021116 does not, running eight to ten times slower. It stands to reason therefore that this must be due to changes in the way Clonezilla now processes devices, partitions and filesystems. As others have mentioned, the disk access indicators show constant activity, but with little actual work being seemingly accomplished. It appears that there is excessive looping within the code.

     
  • Steven Shiau

    Steven Shiau - 2022-01-06

    Is that possible I can access your machine via ssh after booting Clonezilla live?
    If so, please email me at steven at clonezilla org.
    With that, it's easier for us to debug and find why.
    Thanks.

    Steven

     

    Last edit: Steven Shiau 2022-01-06
  • Steven Shiau

    Steven Shiau - 2022-01-06

    Or you can post the files which you have saved in the image dir, including the files:
    *.sf
    lvm_*.*
    I can try to use those files to create a virtual machine, then try to reproduce this issue.

    Steven

     
  • Steven Shiau

    Steven Shiau - 2022-01-06

    Another way is you to create a virtual machine which you can reproduce this issue, and share that VM with me.
    If the issue can be reproduced, then it can almost be fixed.
    Thanks.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-08

    Thank you Steven for taking the time to address this problem, as well as for offering several helpful proposals. Unfortunately, opening an Internet-facing ssh port is far too risky for my security infrastructure. Such ports are detected within minutes through constant aggressive scanning and are relentlessly subjected to brute-force attacks almost immediately. There have been sufficient vulnerabilities reported in ssh and ssl implementations in the last two years alone not to mention weaknesses in the underlying physical infrastructure to make this a hazardous undertaking.

    As for the second proposal, I sincerely doubt that the issue I’m having with Clonezilla is data-related. The slowest part is concentrated in the disk and partition handling, not the actual data backup through partclone. In my opinion, installing the data elsewhere would probably not adequately reproduce the unique combination of the five SATA and USB-attached disks with their 56 partitions that is peculiar to this machine.

    The thing to remember is that the 20210817-hirsute-amd64 release of Clonezilla works perfectly, and is still the one I am currently using. The problem started with the 20211116-impish-amd64 release, and continues with the recent 20220103-impish-amd64 release. Whatever changes were made to disk and partition handling between the 20210817 and 20211116 releases is responsible for the issue I’m having.

    I also doubt the underlying operating systems are to blame, in this case Ubuntu, as I have the 21.10 release with the 5.13.0-23 kernel installed on that very machine and there have been no noticeable disk or partition issues with it. In fact, the machine successfully handles kernels ranging from 5.13 to the very latest 5.15.13 with nary a difficulty in sight, and several Red Hat versions to openSUSE Tumbleweed on the way to Arch Linux with no significant problems.

    Believe me, I aim to assist in identifying the source of this problem and hopefully contribute to correcting it, but the machine is a critical part of my infrastructure and I have limited opportunities for extended downtime for testing purposes.

     
  • Steven Shiau

    Steven Shiau - 2022-01-09

    OK, we have pushed Clonezilla live 2.8.1-12 and 20220103-impish as the stable release, but did not address the issue you have. We will keep trying to reproduce this issue and fix it.
    Therefore, could you please boot Clonezilla live 20220103-impish on the machine, enter command line prompt, then
    run:
    sudo time ocs-prep-cache
    to see how long did it run.
    If you are familiar with screen or tmux, please enter it, and run it again with:
    sudo bash -x ocs-prep-cache
    Then cop & paste the output on the screen to a file, then post it.
    Thanks.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-12

    My apologies for the tardy response, I had a hectic last few days, but I’ve freed some time from a busy schedule to try and move this forward.

    I did notice that Clonezilla live 20220109-impish is available; would you rather I use that instead of 20220103-impish? Either is fine with me. Also, do I run “sudo time ocs-prep-cache” before or after the backup disk selection dialogue?

    Finally, it’s been at least two decades that I’ve had to use the screen command; it was my favourite for running tasks on remote servers to protect my processes from inadvertent session disconnections. I can’t quite remember if it automatically saves the output from commands being run within; this would be practical in order the retain all output from ocs-prep-cache.

    Thanks.

     
  • Steven Shiau

    Steven Shiau - 2022-01-12

    Yes, please use 20220109-impish to test that.
    Please boot it and enter command line prompt, before doing anything, just run:
    sudo time ocs-prep-cache
    Actually in this release, we add a boot parameter to disable the devices list info cache, i.e., you can use
    use_dev_list_cache=no
    to make that. For more info, please check changelog.
    Thank you for debugging this issue with us. Please let us know the results if you test both of them.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-12

    I’ll use the Clonezilla live 20220109-impish then. As well, sorry for asking what may seen obvious, but how to I specify the boot parameter use_dev_list_cache=no — do I add it to /tftpboot/nbi_img/pxelinux.cfg/default as mentioned in the FAQ?

    Thanks again.

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-13

    I ended up modifying my usual Clonezilla entry in boot/grub/grub.cfg, adding the use_dev_list_cache=no entry at the end of the command line. Upon booting, a cursory check confirmed the parameter was enabled.

    As well, the time command is apparently not installed by default, which is surprising. I had to download the time_1.9-0.1_amd64.deb package and install it with dpkg to get that to run.

    Here’s the output from sudo time ocs-prep-cache:

    use_dev_list_cache is set as no. Do not prepar device name cache files in /tmp/ocs-cache//...
    0.08user 0.03system 0:00.11elapsed 102%CPU (0avgtext+0avgdata 15596maxresident)k
    320inputs+0outputs (0major+7053minor)pagefaults 0swaps
    

    It ran extremely quickly, which is not at all what I expected. This does not appear to be the same disk/partition/filesystem processing that takes several minutes for each partition. Should I still run sudo bash -x ocs-prep-cache? If so, what is the best way to save the output? Running as a live system, there is no local file storage available.

     
  • Steven Shiau

    Steven Shiau - 2022-01-13

    If you use use_dev_list_cache=no in the boot parameter, then no need to run "time ocs-prep-cache" since it will be skipped.
    To test ocs-prep-cache, you have to remove "use_dev_list_cache=no" from the boot parameter.
    What I suggest is:

    1. Add "use_dev_list_cache=no" in the boot parameter, then just do a regular saving. To see if any difference between this version and the previous one.
    2. Do not add "use_dev_list_cache=no", then run: "time ocs-prep-cache".

    BTW, time is also builtin in bash, so I should correct my description. Just switch to root by "sudo -i", then run:
    time ocs-prep-cache

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-13

    All right, this was very interesting. I had to run a test image backup three times, all for different reasons. All three were run using Clonezilla live 20220109-impish with boot parameter use_dev_list_cache=no.

    • First run, I had forgotten to comment out os-prober in /usr/share/drbl/sbin/ocs-functions. Usually, every time a new stable release comes out, I go to the trouble of un-squashing the filesystem.squashfs, editing ocs-functions to add a ‘#’ to the os-prober line, and re-squashing the filesystem. This is because os-prober takes forever to run on that machine. I also disable GRUB’s os-prober in all of my many Linuxes.
    • Second run, I followed the instructions in the split file dialogue and left the default of 0 for no splitting. The image backup appeared to run normally, but the image verification failed with a strange ‘invalid header’ error, and then froze when checking the lvm image files.
    • Third and last run, I entered ‘1000000’ in the split file dialogue, which was the recommendation for the older releases of Clonezilla, 20210817-hirsute included, if file splitting was not required. This one worked.

    So, the conclusion of all this is that Clonezilla live 20220109-impish with boot parameter use_dev_list_cache=no works on this machine much the same as it did with 20210817-hirsute, which is good news. Would you still want me to run “time ocs-prep-cache” without the use_dev_list_cache=no parameter?

    As well, as an enhancement request, it would be nice if os-prober could be disabled, probably through another parameter.

     

    Last edit: Bernard Michaud 2022-01-13
  • Steven Shiau

    Steven Shiau - 2022-01-13

    "Would you still want me to run “time ocs-prep-cache” without the use_dev_list_cache=no parameter?" -> Yes.
    " it would be nice if os-prober could be disabled, probably through another parameter.: -> Which files are related to this? Please list them so that we know how to deal with.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-13

    I ran “time ocs-prep-cache” from the command line without the use_dev_list_cache=no parameter, and it ran relatively fast:

    Start preparing device name cache files in /tmp/ocs-cache//...
    Finding all disk(s)...
    Finding all partition(s)...
    Generating dev filesystem/size/type info cache files...
    24.18user 8.79system 1:05.78elapsed 50%CPU (0avgtext+0avgdata 21812maxresident)k
    865987inputs+0outputs (11major+2443705minor)pagefaults 0swaps
    

    However, when running Clonezilla live 20220109-impish in the usual way without the use_dev_list_cache=no parameter, the backup disk selection dialogue took as long as earlier attempts, disk activity never stopping. I halted it after ten minutes by rebooting. I can’t explain why running ocs-prep-cache manually from the command line runs faster than when running Clonezilla live as usual. It seems obvious that the lengthy device/partition/filesystem processing is more than just ocs-prep-cache.

    As for disabling os-prober, a change must be made in squashfs-root/usr/share/drbl/sbin/ocs-functions from:

    if type os-prober &>/dev/null; then
      echo "Saving OS info from the device..."
      echo "This OS-related info was saved from this machine with os-prober at $save_time:" > $img_dir/Info-OS-prober.txt
      LC_ALL=C os-prober >> $img_dir/Info-OS-prober.txt
    fi
    

    to:

    if type os-prober &>/dev/null; then
      echo "Saving OS info from the device..."
      echo "This OS-related info was saved from this machine with os-prober at $save_time:" > $img_dir/Info-OS-prober.txt
    # LC_ALL=C os-prober >> $img_dir/Info-OS-prober.txt
    fi
    

    Commenting out os-prober prevents it from running.

     

    Last edit: Bernard Michaud 2022-01-13
  • Bernard Michaud

    Bernard Michaud - 2022-01-17

    Well, it’s not all bad; at least I now know to specify the use_dev_list_cache=no parameter in Clonezilla releases 20220109-impish and beyond and get back the same processing speed I had before. I guess it beats standing still and continue to use release 20210817-hirsute, which works fine on my machine but will no longer evolve.

     
  • Steven Shiau

    Steven Shiau - 2022-01-18

    OK, we have another release: 20220118-* or 2.8.2-5 which you can give it a try:
    https://clonezilla.org/downloads.php

    In this release, a boot parameter was added to disable os-prober, as you have requested:
    use_os_prober=no
    For more info, please check changelog.

    BTW, when you mentioned:
    "Second run, I followed the instructions in the split file dialogue and left the default of 0 for no splitting. The image backup appeared to run normally, but the image verification failed with a strange ‘invalid header’ error, and then froze when checking the lvm image files." -> which parameters did you choose? Is that in expert mode and you choose to use "-z5p" (pixz) or "-z5" (xz)? If so, a bug related to this has been fixed in the same version of Clonezilla live.

    Thank you for debugging this issue with us. Please let us know the results if you test both of them.

    Steven

     
  • Bernard Michaud

    Bernard Michaud - 2022-01-20

    I gave the new test release 20220118-impish a try. Beforehand, I edited /boot/grub/grub.cfg and added parameters use_dev_list_cache=no and use_os_prober=no to my customary boot entries. I can report that performance is well within the realm of 20210817-hirsute, my reference release. So, all things considered, I am satisfied that I can continue to use Clonezilla on this machine by disabling the dev list cache. As well, I appreciate the parameter to disable os_prober; the workaround was not overly complicated, but it was tedious to remember to apply it.

    As for the other issue, it wasn’t with the compression option but the split image file value. My old release 20210817-hirsute had instructions to enter a “big number” such as ‘1000000’ so as not to split the image file, and had 4096 as the default value. New releases since 20211116-impish instead instruct to use ‘0’ to not split the image, and this is also the default value. Allowing ‘0’ to be used as the split image value causes the image verification to fail. Running it again but using ‘1000000’ as the split image value works properly. I have no idea why this is so.

     
  • Steven Shiau

    Steven Shiau - 2022-01-20

    OK, got it.
    If the"split" issue is always reproducible when you use "0", please let us know.

    Steven

     

Log in to post a comment.

MongoDB Logo MongoDB