====== Notizzettel vom brot ====== ===== Reset HP OfficeJet Pro 8600 Plus ===== My printer is a N911n, and I have not found a consistent way to reset it. I did get the Home, Back, Home, Home to work possibly one time a couple months ago, so not sure if the Oct 31st firmware changed something, as I have not been able to do any kind of reset. (tried the Back x 4-5 times as well at boot up, but nothing happens, after power off/unplugging for 1-10 mins). https://h30434.www3.hp.com/t5/Printer-Setup-Software-Drivers/HP-Officejet-Pro-8600-N911g-Firmware-Update-Fails/td-p/7876436 ===== NAS Einführung btrfs ===== Stand siehe Unten - tl;dr: 3x 3TB (500mb efi, 500mb raid1 /boot, 50G raid5 /, 2,7TB /mnt/raid_daten) Einbau neuer 2x 5TB wd red ==== Anlegen der neuen Partitionstabellen ==== nas brot # gdisk /dev/sde GPT fdisk (gdisk) version 0.8.10 Caution: invalid backup GPT header, but valid main header; regenerating backup header from main header. Warning! Main and backup partition tables differ! Use the 'c' and 'e' options on the recovery & transformation menu to examine the two tables. Warning! One or more CRCs don't match. You should repair the disk! Partition table scan: MBR: protective BSD: not present APM: not present GPT: damaged **************************************************************************** Caution: Found protective or hybrid MBR and corrupt GPT. Using GPT, but disk verification and recovery are STRONGLY recommended. **************************************************************************** Command (? for help): o This option deletes all partitions and creates a new protective MBR. Proceed? (Y/N): y Command (? for help): n Partition number (1-128, default 1): First sector (34-9767541134, default = 2048) or {+-}size{KMGTP}: Last sector (2048-9767541134, default = 9767541134) or {+-}size{KMGTP}: 1026047 Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): ef00 Changed type of partition to 'EFI System' Command (? for help): n Partition number (2-128, default 2): First sector (34-9767541134, default = 1026048) or {+-}size{KMGTP}: Last sector (1026048-9767541134, default = 9767541134) or {+-}size{KMGTP}: 2050047 Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): fd00 Changed type of partition to 'Linux RAID' Command (? for help): n Partition number (3-128, default 3): First sector (34-9767541134, default = 2050048) or {+-}size{KMGTP}: Last sector (2050048-9767541134, default = 9767541134) or {+-}size{KMGTP}: Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): Changed type of partition to 'Linux filesystem' Command (? for help): p Disk /dev/sde: 9767541168 sectors, 4.5 TiB Logical sector size: 512 bytes Disk identifier (GUID): D69BC615-B1AC-4A38-BA71-791CD94C2154 Partition table holds up to 128 entries First usable sector is 34, last usable sector is 9767541134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 9767541134 4.5 TiB 8300 Linux filesystem Command (? for help): w Final checks complete. About to write GPT data. THIS WILL OVERWRITE EXISTING PARTITIONS!! Do you want to proceed? (Y/N): y OK; writing new GUID partition table (GPT) to /dev/sde. The operation has completed successfully. ==== Kopieren der EFI-Partition ==== nas brot # umount /boot/efi nas brot # cp /dev/sda1 /dev/sdd1 nas brot # cp /dev/sda1 /dev/sde1 ==== Hinzufügen der /boot-Partitionen zum bestehenden RAID1 ==== nas brot # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 3 Persistence : Superblock is persistent Update Time : Sun Nov 16 19:15:17 2014 State : clean Active Devices : 2 Working Devices : 3 Failed Devices : 0 Spare Devices : 1 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 248 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 4 8 33 1 active sync /dev/sdc1 3 8 17 - spare /dev/sdb1 nas brot # mdadm --verbose --manage --add /dev/md0 /dev/sdd2 mdadm: added /dev/sdd2 nas brot # mdadm --verbose --manage --add /dev/md0 /dev/sde2 mdadm: added /dev/sde2 nas brot # cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] md2 : active raid5 sdb3[6] sdc3[5] sda3[0] 52430848 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/3] [UUU] md4 : active raid5 sdb4[2] sdc4[1] sda4[3] 5753362432 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/3] [UUU] bitmap: 0/22 pages [0KB], 65536KB chunk md0 : active raid1 sde2[5](S) sdd2[2](S) sdb1[3](S) sdc1[4] sda2[0] 511988 blocks super 1.0 [2/2] [UU] unused devices: nas brot # mdadm /dev/md0 --fail /dev/sdc1 mdadm: set /dev/sdc1 faulty in /dev/md0 nas brot # mdadm /dev/md0 --fail /dev/sdb1 mdadm: set /dev/sdb1 faulty in /dev/md0 nas brot # cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] md2 : active raid5 sdb3[6] sdc3[5] sda3[0] 52430848 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/3] [UUU] md4 : active raid5 sdb4[2] sdc4[1] sda4[3] 5753362432 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/3] [UUU] bitmap: 0/22 pages [0KB], 65536KB chunk md0 : active raid1 sde2[5] sdd2[2](S) sdb1[3](S) sdc1[4](F) sda2[0] 511988 blocks super 1.0 [2/1] [U_] [==========>..........] recovery = 52.9% (270912/511988) finish=0.0min speed=67728K/sec nas brot # mdadm /dev/md0 --remove /dev/sdb1 /dev/sdc1 mdadm: hot removed /dev/sdb1 from /dev/md0 mdadm: hot removed /dev/sdc1 from /dev/md0 nas brot # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 3 Persistence : Superblock is persistent Update Time : Sun Nov 16 19:41:39 2014 State : clean Active Devices : 2 Working Devices : 3 Failed Devices : 0 Spare Devices : 1 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 277 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 5 8 66 1 active sync /dev/sde2 2 8 50 - spare /dev/sdd2 nas brot # mdadm --grow /dev/md0 --raid-devices=3 raid_disks for /dev/md0 set to 3 unfreeze nas brot # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Update Time : Sun Nov 16 19:45:49 2014 State : clean, degraded, recovering Active Devices : 2 Working Devices : 3 Failed Devices : 0 Spare Devices : 1 Rebuild Status : 81% complete Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 296 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 5 8 66 1 active sync /dev/sde2 2 8 50 2 spare rebuilding /dev/sdd2 ==== btrfs Filesystem erzeugen und Subvolumes anlegen ==== nas brot # mkfs.btrfs -d raid1 -m raid1 /dev/sdd3 /dev/sde3 nas brot # btrfs filesystem show /dev/sdd3 Label: none uuid: c356873f-4034-43cc-8052-b2a4bf68cb26 Total devices 2 FS bytes used 112.00KiB devid 1 size 4.55TiB used 2.03GiB path /dev/sdd3 devid 2 size 4.55TiB used 2.01GiB path /dev/sde3 Btrfs v3.17 nas brot # mount /dev/sdd3 /mnt/btrfs_rootsubvol/ nas brot # btrfs subvolume create /mnt/btrfs_rootsubvol/root-subvol Create subvolume '/mnt/btrfs_rootsubvol/root-subvol' nas brot # btrfs subvolume create /mnt/btrfs_rootsubvol/data-subvol Create subvolume '/mnt/btrfs_rootsubvol/data-subvol' nas brot # btrfs subvolume create /mnt/btrfs_rootsubvol/backup-subvol Create subvolume '/mnt/btrfs_rootsubvol/backup-subvol' nas brot # btrfs subvolume create /mnt/btrfs_rootsubvol/brot-backup-subvol Create subvolume '/mnt/btrfs_rootsubvol/brot-backup-subvol' nas brot # btrfs fi df /mnt/btrfs_rootsubvol/ Data, RAID1: total=1.00GiB, used=512.00KiB Data, single: total=8.00MiB, used=0.00B System, RAID1: total=8.00MiB, used=16.00KiB System, single: total=4.00MiB, used=0.00B Metadata, RAID1: total=1.00GiB, used=176.00KiB Metadata, single: total=8.00MiB, used=0.00B GlobalReserve, single: total=16.00MiB, used=0.00B nas ~ # btrfs subvolume list /mnt/btrfs_rootsubvol/ ID 258 gen 155 top level 5 path root-subvol ID 259 gen 8 top level 5 path data-subvol ID 260 gen 9 top level 5 path backup-subvol ID 261 gen 10 top level 5 path brot-backup-subvol nas ~ # btrfs subvolume set-default 258 /mnt/btrfs_rootsubvol/ (setzen vom rootfs als std-subvol) ==== Umzug Rootfs ==== * Boot von USB-Stick (grml) * einbinden der alten Dateisysteme (/dev/md2 als /, md1 als /boot und EFI) * einbinden des neuen BTRFS-Volumes ( auf die richtige Subvolume zuordnung achten) * cp -a /mnt/old_rootfs /mnt/btrfs-rootsubvol * grml-chroot /mnt/btrfs-rootsubvol /bin/bash * grub-install /dev/sdd * grub-install /dev/sde * fstab anpassen - nur die wichtigen einträge aktiv lassen um bootfehler zu vermeiden * initrd (u kernel) neubauen mit dracut /dev/md0 /boot ext4 noatime 1 2 UUID=c356873f-4034-43cc-8052-b2a4bf68cb26 / btrfs noatime 0 1 UUID=4549-7F03 /boot/efi vfat defaults 0 2 ==== Kopieren der Daten ==== * Einbinden der Subvolumes (primär erstmal daten und backup) * mount -o ro /dev/md4 /mnt/raid_datenpartition * cp -av /mnt/raid_datenpartition/Daten /mnt/btrfs-daten-subvol/ | tee cp-log.tee * ==== Kopieren der Backups ==== Problem: btrfs-image-files auf den alten FS. Jetzt: subvolumes auf dem btrfs-fs Lösung: Übertragen aller Snapshots per btrfs send/receive nas ~ # for i in $( ls /mnt/current_backup_file/ | grep backup); do btrfs property set -t s /mnt/current_backup_file/$i ro true; done nas ~ # btrfs send -v /mnt/current_backup_file/backup-2011.11.07-22\:43/ | btrfs receive -v /mnt/btrfs-brot-backup-subvol/ At subvol /mnt/current_backup_file/backup-2011.11.07-22:43/ At subvol backup-2011.11.07-22:43 receiving subvol backup-2011.11.07-22:43 uuid=9c52dfc1-499f-9640-a4f9-0f8b89974a61, stransid=0 ==== Hinzufügen der alten Platten in den BTRFS-Pool ==== * Platte sdc soll dem BTRFS-RAID1 hinzugefügt werden (sda und sdb sind die neuen Platten) * Degradieren der RAID5s * Neupartitionieren der Platte nas brot # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Update Time : Mon Nov 17 23:42:11 2014 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10324 Number Major Minor RaidDevice State 0 8 51 0 active sync /dev/sdd3 6 8 67 1 active sync /dev/sde3 5 8 35 2 active sync /dev/sdc3 nas brot # mdadm -D /dev/md4 /dev/md4: Version : 1.2 Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 2876681216 (2743.42 GiB 2945.72 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Sun Nov 30 21:05:22 2014 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : nas:4 (local to host nas) UUID : f6924d42:7358df85:416f97f9:83aeac62 Events : 16217 Number Major Minor RaidDevice State 3 8 52 0 active sync /dev/sdd4 1 8 36 1 active sync /dev/sdc4 2 8 68 2 active sync /dev/sde4 nas brot # mdadm --manage --set-faulty /dev/md2 /dev/sdc3 mdadm: set /dev/sdc3 faulty in /dev/md2 nas brot # mdadm --manage --set-faulty /dev/md4 /dev/sdc4 mdadm: set /dev/sdc4 faulty in /dev/md4 nas brot # mdadm --manage /dev/md2 -r /dev/sdc3 mdadm: hot removed /dev/sdc3 from /dev/md2 nas brot # mdadm --manage /dev/md4 -r /dev/sdc4 mdadm: hot removed /dev/sdc4 from /dev/md4 nas brot # mdadm -D /dev/md4 /dev/md4: Version : 1.2 Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 2876681216 (2743.42 GiB 2945.72 GB) Raid Devices : 3 Total Devices : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Sun Dec 7 12:07:18 2014 State : clean, degraded Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : nas:4 (local to host nas) UUID : f6924d42:7358df85:416f97f9:83aeac62 Events : 16220 Number Major Minor RaidDevice State 3 8 52 0 active sync /dev/sdd4 2 0 0 2 removed 2 8 68 2 active sync /dev/sde4 nas brot # gdisk /dev/sdc GPT fdisk (gdisk) version 0.8.10 Partition table scan: MBR: protective BSD: not present APM: not present GPT: present Found valid GPT with protective MBR; using GPT. Command (? for help): p Disk /dev/sdc: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 106907647 50.0 GiB FD00 Linux RAID 4 106907648 5860533134 2.7 TiB FD00 Command (? for help): d Partition number (1-4): 4 Command (? for help): d Partition number (1-3): 3 Command (? for help): p Disk /dev/sdc: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 5858485101 sectors (2.7 TiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID Command (? for help): n Partition number (3-128, default 3): First sector (34-5860533134, default = 2050048) or {+-}size{KMGTP}: Last sector (2050048-5860533134, default = 5860533134) or {+-}size{KMGTP}: Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): Changed type of partition to 'Linux filesystem' Command (? for help): p Disk /dev/sdc: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 5860533134 2.7 TiB 8300 Linux filesystem Command (? for help): w Final checks complete. About to write GPT data. THIS WILL OVERWRITE EXISTING PARTITIONS!! Do you want to proceed? (Y/N): y OK; writing new GUID partition table (GPT) to /dev/sdc. The operation has completed successfully. nas brot # btrfs device add /dev/sdc3 / -f nas brot # btrfs balance start -v / Dumping filters: flags 0x7, state 0x0, force is off DATA (flags 0x0): balancing METADATA (flags 0x0): balancing SYSTEM (flags 0x0): balancing nas brot # watch -n 10 'btrfs balance status /' Balance on '/' is running 1 out of about 3680 chunks balanced (2 considered), 100% left nas brot # mdadm -D /dev/md4 /dev/md4: Version : 1.2 Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 2876681216 (2743.42 GiB 2945.72 GB) Raid Devices : 3 Total Devices : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Sat Dec 27 10:31:33 2014 State : clean, degraded Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : nas:4 (local to host nas) UUID : f6924d42:7358df85:416f97f9:83aeac62 Events : 18754 Number Major Minor RaidDevice State 3 8 52 0 active sync /dev/sdd4 2 0 0 2 removed 2 8 68 2 active sync /dev/sde4 nas brot # mdadm -S /dev/md4 mdadm: stopped /dev/md4 nas brot # mdadm -S /dev/md2 mdadm: stopped /dev/md2 nas brot # gdisk /dev/sdd GPT fdisk (gdisk) version 0.8.10 Partition table scan: MBR: protective BSD: not present APM: not present GPT: present Found valid GPT with protective MBR; using GPT. Command (? for help): p Disk /dev/sdd: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 106907647 50.0 GiB FD00 Linux RAID 4 106907648 5860533134 2.7 TiB FD00 Command (? for help): d Partition number (1-4): 4 Command (? for help): d Partition number (1-3): 3 Command (? for help): n Partition number (3-128, default 3): First sector (34-5860533134, default = 2050048) or {+-}size{KMGTP}: Last sector (2050048-5860533134, default = 5860533134) or {+-}size{KMGTP}: Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): Changed type of partition to 'Linux filesystem' Command (? for help): p Disk /dev/sdd: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 5860533134 2.7 TiB 8300 Linux filesystem Command (? for help): w Final checks complete. About to write GPT data. THIS WILL OVERWRITE EXISTING PARTITIONS!! Do you want to proceed? (Y/N): y OK; writing new GUID partition table (GPT) to /dev/sdd. Warning: The kernel is still using the old partition table. The new table will be used at the next reboot. The operation has completed successfully. nas brot # gdisk /dev/sde GPT fdisk (gdisk) version 0.8.10 Partition table scan: MBR: protective BSD: not present APM: not present GPT: present Found valid GPT with protective MBR; using GPT. Command (? for help): p Disk /dev/sde: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 106907647 50.0 GiB FD00 Linux RAID 4 106907648 5860533134 2.7 TiB FD00 Command (? for help): d Partition number (1-4): 4 Command (? for help): d Partition number (1-3): 3 Command (? for help): n Partition number (3-128, default 3): First sector (34-5860533134, default = 2050048) or {+-}size{KMGTP}: Last sector (2050048-5860533134, default = 5860533134) or {+-}size{KMGTP}: Current type is 'Linux filesystem' Hex code or GUID (L to show codes, Enter = 8300): Changed type of partition to 'Linux filesystem' Command (? for help): p Disk /dev/sde: 5860533168 sectors, 2.7 TiB Logical sector size: 512 bytes Disk identifier (GUID): 0C5623E2-2FF4-4787-9609-750745C161DA Partition table holds up to 128 entries First usable sector is 34, last usable sector is 5860533134 Partitions will be aligned on 2048-sector boundaries Total free space is 2014 sectors (1007.0 KiB) Number Start (sector) End (sector) Size Code Name 1 2048 1026047 500.0 MiB EF00 EFI System 2 1026048 2050047 500.0 MiB FD00 Linux RAID 3 2050048 5860533134 2.7 TiB 8300 Linux filesystem Command (? for help): w Final checks complete. About to write GPT data. THIS WILL OVERWRITE EXISTING PARTITIONS!! Do you want to proceed? (Y/N): y OK; writing new GUID partition table (GPT) to /dev/sde. Warning: The kernel is still using the old partition table. The new table will be used at the next reboot. The operation has completed successfully. nas brot # nas brot # sync nas brot # btrfs device add /dev/sdd3 / -f nas brot # btrfs device add /dev/sde3 / -f nas brot # btrfs balance start -v / Dumping filters: flags 0x7, state 0x0, force is off DATA (flags 0x0): balancing METADATA (flags 0x0): balancing SYSTEM (flags 0x0): balancing (anderer screen) nas brot # watch -n 10 'btrfs balance status /' Every 10,0s: btrfs balance status / Tue Dec 30 11:02:30 2014 Balance on '/' is running 1 out of about 4941 chunks balanced (2 considered), 100% left ===== NAS Festplattenausbau ===== http://neil.brown.name/blog/20090817000931 ==== Derzeitiger Status ==== === Partitionierung === nas ~ # fdisk -l Disk /dev/sda: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sda1 2048 1026047 500M EFI System /dev/sda2 1026048 2050047 500M Linux RAID /dev/sda3 2050048 106907647 50G Linux RAID /dev/sda4 106907648 2983721397 1.3T Linux RAID /dev/sda5 2983723008 5860533134 1.3T Linux RAID Disk /dev/sdb: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: C767012E-AA8B-4ACE-AD4A-8ADCAB4300D5 Device Start End Size Type /dev/sdb1 2048 1026047 500M Linux RAID /dev/sdb2 1026048 53460000 25G Linux RAID /dev/sdb3 53460992 2930277134 1.3T Linux RAID Disk /dev/sdc: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: 4B660130-B8DD-4BA3-9D8C-615F929B0433 Device Start End Size Type /dev/sdc1 2048 1026047 500M Linux RAID /dev/sdc2 1026048 53460000 25G Linux RAID /dev/sdc3 53460992 2930277134 1.3T Linux RAID Disk /dev/sdd: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: D01BE9B6-18F2-4C99-B30B-544B0C9BFA62 Device Start End Size Type /dev/sdd1 2048 1026047 500M Linux RAID /dev/sdd2 1026048 53460000 25G Linux RAID /dev/sdd3 53460992 2930277134 1.3T Linux RAID Disk /dev/md2: 75 GiB, 80533782528 bytes, 157292544 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 524288 bytes / 1572864 bytes Disk /dev/md3: 4 TiB, 4418775810048 bytes, 8630421504 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 524288 bytes / 1572864 bytes Disk /dev/md0: 500 MiB, 524275712 bytes, 1023976 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes === mdstat === nas ~ # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 4 Persistence : Superblock is persistent Update Time : Mon Nov 25 21:35:39 2013 State : clean Active Devices : 2 Working Devices : 4 Failed Devices : 0 Spare Devices : 2 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 178 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 2 8 49 1 active sync /dev/sdd1 3 8 33 - spare /dev/sdc1 4 8 17 - spare /dev/sdb1 nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Mon Dec 2 21:01:28 2013 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 1041 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 50 1 active sync /dev/sdd2 2 8 34 2 active sync /dev/sdc2 4 8 18 3 active sync /dev/sdb2 nas ~ # mdadm -D /dev/md3 /dev/md3: Version : 1.2 Creation Time : Fri Aug 24 04:23:04 2012 Raid Level : raid5 Array Size : 4315210752 (4115.31 GiB 4418.78 GB) Used Dev Size : 1438403584 (1371.77 GiB 1472.93 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Sun Dec 1 19:08:30 2013 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:3 UUID : cf205a21:16cfbf89:1b7442b4:130edcb3 Events : 42942 Number Major Minor RaidDevice State 0 8 4 0 active sync /dev/sda4 4 8 35 1 active sync /dev/sdc3 3 8 51 2 active sync /dev/sdd3 5 8 19 3 active sync /dev/sdb3 ===== Umzug auf größere Platten ===== - Neue Platten einhängen [x] - Partitionen kopieren [x] - Neues 3 Device RAID-5 mit 1 missing Device anlegen [x] - Daten kopieren [x] - Große Platte aus altem RAID umhängen in das neue [ X ] - Systempartitionen auf die neuen Platten holen [ X ] === Partitionierung nach Anpassung === nas ~ # fdisk -l Disk /dev/sda: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sda1 2048 1026047 500M EFI System /dev/sda2 1026048 2050047 500M Linux RAID /dev/sda3 2050048 106907647 50G Linux RAID /dev/sda4 106907648 2983721397 1.3T Linux RAID /dev/sda5 2983723008 5860533134 1.3T Linux RAID Disk /dev/sdb: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: C767012E-AA8B-4ACE-AD4A-8ADCAB4300D5 Device Start End Size Type /dev/sdb1 2048 1026047 500M Linux RAID /dev/sdb2 1026048 53460000 25G Linux RAID /dev/sdb3 53460992 2930277134 1.3T Linux RAID Disk /dev/sdc: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: 4B660130-B8DD-4BA3-9D8C-615F929B0433 Device Start End Size Type /dev/sdc1 2048 1026047 500M Linux RAID /dev/sdc2 1026048 53460000 25G Linux RAID /dev/sdc3 53460992 2930277134 1.3T Linux RAID Disk /dev/sdd: 1.4 TiB, 1500301910016 bytes, 2930277168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: D01BE9B6-18F2-4C99-B30B-544B0C9BFA62 Device Start End Size Type /dev/sdd1 2048 1026047 500M Linux RAID /dev/sdd2 1026048 53460000 25G Linux RAID /dev/sdd3 53460992 2930277134 1.3T Linux RAID Disk /dev/sde: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sde1 2048 1026047 500M EFI System /dev/sde2 1026048 2050047 500M Linux RAID /dev/sde3 2050048 106907647 50G Linux RAID /dev/sde4 106907648 5860533134 2.7T Linux RAID Disk /dev/sdf: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sdf1 2048 1026047 500M EFI System /dev/sdf2 1026048 2050047 500M Linux RAID /dev/sdf3 2050048 106907647 50G Linux RAID /dev/sdf4 106907648 5860533134 2.7T Linux RAID Disk /dev/md2: 75 GiB, 80533782528 bytes, 157292544 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 524288 bytes / 1572864 bytes Disk /dev/md0: 500 MiB, 524275712 bytes, 1023976 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk /dev/md3: 4 TiB, 4418775810048 bytes, 8630421504 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 524288 bytes / 1572864 bytes === Anlegen des neuen RAID-Devices === nas ~ # mdadm --create --verbose /dev/md4 --level=5 --raid-devices=3 missing /dev/sde4 /dev/sdf4 mdadm: layout defaults to left-symmetric mdadm: layout defaults to left-symmetric mdadm: chunk size defaults to 512K mdadm: size set to 2876681216K mdadm: automatically enabling write-intent bitmap on large array mdadm: Defaulting to version 1.2 metadata mdadm: array /dev/md4 started. nas ~ # cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] md4 : active raid5 sdf4[2] sde4[1] 5753362432 blocks super 1.2 level 5, 512k chunk, algorithm 2 [3/2] [_UU] bitmap: 0/22 pages [0KB], 65536KB chunk === Kopieren der Daten === nas newraid # rsync -av /mnt/raid_datenpartition/* /mnt/newraid/ === Umhängen der großen Platte im alten RAID === nas ~ # mdadm --verbose --manage --set-faulty /dev/md3 /dev/sda4 mdadm: set /dev/sda4 faulty in /dev/md3 nas ~ # mdadm --verbose --manage /dev/md3 --remove /dev/sda4 mdadm: hot removed /dev/sda4 from /dev/md3 nas ~ # fdisk /dev/sda Welcome to fdisk (util-linux 2.24). Changes will remain in memory only, until you decide to write them. Be careful before using the write command.Command (m for help): p Disk /dev/sda: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sda1 2048 1026047 500M EFI System /dev/sda2 1026048 2050047 500M Linux RAID /dev/sda3 2050048 106907647 50G Linux RAID /dev/sda4 106907648 2983721397 1.3T Linux RAID /dev/sda5 2983723008 5860533134 1.3T Linux RAID Command (m for help): d Partition number (1-5, default 5): Partition 5 has been deleted. Command (m for help): d Partition number (1-4, default 4): Partition 4 has been deleted. Command (m for help): n Partition number (4-128, default 4): First sector (34-5860533134, default 106907648): Last sector, +sectors or +size{K,M,G,T,P} (106907648-5860533134, default 5860533134): Created a new partition 4 of type 'Linux filesystem' and of size 2.7 TiB. Command (m for help): t Partition number (1-4, default 4): Partition type (type L to list all types): 14 Changed type of partition 'Linux filesystem' to 'Linux RAID'. Command (m for help): p Disk /dev/sda: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: 0C5623E2-2FF4-4787-9609-750745C161DA Device Start End Size Type /dev/sda1 2048 1026047 500M EFI System /dev/sda2 1026048 2050047 500M Linux RAID /dev/sda3 2050048 106907647 50G Linux RAID /dev/sda4 106907648 5860533134 2.7T Linux RAID Command (m for help): w The partition table has been altered. Calling ioctl() to re-read partition table. Re-reading the partition table failed.: Device or resource busy The kernel still uses the old table. The new table will be used at the next reboot or after you run partprobe(8) or kpartx(8). nas ~ # partprobe nas ~ # mdadm --verbose --manage --add /dev/md4 /dev/sda4 mdadm: added /dev/sda4 nas ~ # mdadm -D /dev/md4 /dev/md4: Version : 1.2 Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 2876681216 (2743.42 GiB 2945.72 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Tue Dec 10 19:38:06 2013 State : active, degraded, recovering Active Devices : 2 Working Devices : 3 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Rebuild Status : 0% complete Name : nas:4 (local to host nas) UUID : f6924d42:7358df85:416f97f9:83aeac62 Events : 5393 Number Major Minor RaidDevice State 3 8 4 0 spare rebuilding /dev/sda4 1 8 68 1 active sync /dev/sde4 2 8 84 2 active sync /dev/sdf4 === Systempartitionen auf die neuen Platten holen === nas ~ # mdadm --verbose --manage --add /dev/md2 /dev/sde3 mdadm: added /dev/sde3 nas ~ # mdadm --verbose --manage --add /dev/md2 /dev/sdf3 mdadm: added /dev/sdf3 nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 6 Persistence : Superblock is persistent Update Time : Tue Dec 10 19:44:51 2013 State : clean Active Devices : 4 Working Devices : 6 Failed Devices : 0 Spare Devices : 2 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 1043 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 50 1 active sync /dev/sdd2 2 8 34 2 active sync /dev/sdc2 4 8 18 3 active sync /dev/sdb2 5 8 67 - spare /dev/sde3 6 8 83 - spare /dev/sdf3 nas ~ # mdadm --verbose --manage --set-faulty /dev/md2 /dev/sdd2 mdadm: set /dev/sdd2 faulty in /dev/md2 nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 6 Persistence : Superblock is persistent Update Time : Tue Dec 10 19:46:24 2013 State : clean, degraded, resyncing (DELAYED) Active Devices : 3 Working Devices : 5 Failed Devices : 1 Spare Devices : 2 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 1045 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 spare rebuilding /dev/sdf3 2 8 34 2 active sync /dev/sdc2 4 8 18 3 active sync /dev/sdb2 1 8 50 - faulty /dev/sdd2 5 8 67 - spare /dev/sde3 :!::!::!: Auf rebuild warten :!::!::!: nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 6 Persistence : Superblock is persistent Update Time : Wed Dec 11 18:58:28 2013 State : clean Active Devices : 4 Working Devices : 5 Failed Devices : 1 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10124 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 2 8 34 2 active sync /dev/sdc2 4 8 18 3 active sync /dev/sdb2 1 8 50 - faulty /dev/sdd2 5 8 67 - spare /dev/sde3 nas ~ # mdadm --verbose --manage --set-faulty /dev/md2 /dev/sdc2 mdadm: set /dev/sdc2 faulty in /dev/md2 nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 6 Persistence : Superblock is persistent Update Time : Wed Dec 11 19:02:46 2013 State : active, degraded, recovering Active Devices : 3 Working Devices : 4 Failed Devices : 2 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Rebuild Status : 4% complete Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10129 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 spare rebuilding /dev/sde3 4 8 18 3 active sync /dev/sdb2 1 8 50 - faulty /dev/sdd2 2 8 34 - faulty /dev/sdc2 nas ~ # mdadm --verbose --manage /dev/md2 --remove /dev/sdd2 /dev/sdc2 mdadm: hot removed /dev/sdd2 from /dev/md2 mdadm: hot removed /dev/sdc2 from /dev/md2 nas ~ # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 4 Persistence : Superblock is persistent Update Time : Tue Dec 10 19:35:24 2013 State : clean Active Devices : 2 Working Devices : 4 Failed Devices : 0 Spare Devices : 2 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 178 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 2 8 49 1 active sync /dev/sdd1 3 8 33 - spare /dev/sdc1 4 8 17 - spare /dev/sdb1 nas ~ # mdadm --verbose --manage /dev/md0 --remove /dev/sdb1 /dev/sdc1 mdadm: hot removed /dev/sdb1 from /dev/md0 mdadm: hot removed /dev/sdc1 from /dev/md0 nas ~ # mdadm --verbose --manage /dev/md0 --add /dev/sdf1 /dev/sde1 mdadm: added /dev/sdf1 mdadm: added /dev/sde1 nas ~ # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 19:53:51 2013 State : clean Active Devices : 2 Working Devices : 4 Failed Devices : 0 Spare Devices : 2 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 182 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 2 8 49 1 active sync /dev/sdd1 3 8 81 - spare /dev/sdf1 4 8 65 - spare /dev/sde1 nas ~ # mdadm --verbose --manage --set-faulty /dev/md0 /dev/sdd1 mdadm: set /dev/sdd1 faulty in /dev/md0 nas ~ # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 19:55:33 2013 State : clean Active Devices : 2 Working Devices : 3 Failed Devices : 1 Spare Devices : 1 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 201 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 4 8 65 1 active sync /dev/sde1 2 8 49 - faulty /dev/sdd1 3 8 81 - spare /dev/sdf1 nas ~ # mdadm --verbose --manage /dev/md0 --remove /dev/sdd1 mdadm: hot removed /dev/sdd1 from /dev/md0 ARRAY /dev/md2 metadata=1.2 name=localhost.localdomain:2 UUID=a9d908f3:f8632fcf:59a69085:83d5f6aa ARRAY /dev/md0 metadata=1.0 spares=1 name=localhost.localdomain:0 UUID=262c8e68:c7c4fa6d:19af27e8:0cfc68a1 ARRAY /dev/md4 metadata=1.2 name=nas:4 UUID=f6924d42:7358df85:416f97f9:83aeac62 nas ~ # mdadm -D /dev/md0 /dev/md0: Version : 1.0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Array Size : 511988 (500.07 MiB 524.28 MB) Used Dev Size : 511988 (500.07 MiB 524.28 MB) Raid Devices : 2 Total Devices : 3 Persistence : Superblock is persistent Update Time : Wed Dec 11 19:56:04 2013 State : clean Active Devices : 2 Working Devices : 3 Failed Devices : 0 Spare Devices : 1 Name : localhost.localdomain:0 UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Events : 202 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 4 8 65 1 active sync /dev/sde1 3 8 81 - spare /dev/sdf1 nas ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 20:04:01 2013 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10193 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 4 8 18 3 active sync /dev/sdb2 nas ~ # mdadm -D /dev/md4 /dev/md4: Version : 1.2 Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 2876681216 (2743.42 GiB 2945.72 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Wed Dec 11 09:35:53 2013 State : active Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : nas:4 (local to host nas) UUID : f6924d42:7358df85:416f97f9:83aeac62 Events : 15769 Number Major Minor RaidDevice State 3 8 4 0 active sync /dev/sda4 1 8 68 1 active sync /dev/sde4 2 8 84 2 active sync /dev/sdf4 === tatsächlicher Umzug der / Partition === * grml von USB-Booten * resize2fs root@grml ~ # mdadm -D /dev/md/2 /dev/md/2: Version : 1.2 Creation Time : Tue Aug 28 02:45:52 2012 Raid Level : raid5 Array Size : 78646272 (75.00 GiB 80.53 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 19:34:29 2013 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10193 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 4 8 18 3 active sync /dev/sdb2 root@grml ~ # resize2fs -P /dev/md2 resize2fs 1.42.8 (20-Jun-2013) Estimated minimum size of the filesystem: 4376945 root@grml ~ # e2fsck -fv /dev/md2 :( e2fsck 1.42.8 (20-Jun-2013) Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information 607662 inodes used (12.34%, out of 4923392) 1443 non-contiguous files (0.2%) 229 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 600410/124 4580213 blocks used (23.30%, out of 19661568) 0 bad blocks 1 large file 545388 regular files 54723 directories 180 character device files 97 block device files 2 fifos 597 links 7262 symbolic links (6840 fast symbolic links) 1 socket ------------ 608250 files root@grml ~ # resize2fs -p /dev/md2 6076945 resize2fs 1.42.8 (20-Jun-2013) Resizing the filesystem on /dev/md2 to 6076945 (4k) blocks. Begin pass 2 (max = 2204238) Relocating blocks XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX Begin pass 3 (max = 601) Scanning inode table XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX Begin pass 4 (max = 56313) Updating inode references XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX The filesystem on /dev/md2 is now 6076945 blocks long. resize2fs -p /dev/md2 6076945 10.80s user 24.18s system 2% cpu 25:40.55 total root@grml ~ # mdadm --verbose --grow /dev/md2 --array-size 52430848 :( root@grml ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 02:45:52 2012 Raid Level : raid5 Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 20:15:58 2013 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10195 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 4 8 18 3 active sync /dev/sdb2 root@grml ~ # mdadm --verbose --grow /dev/md2 -n 3 --backup-file=/root/md2.backup :( mdadm: Need to backup 3072K of critical section.. root@grml ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 02:45:52 2012 Raid Level : raid5 Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 3 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 20:29:46 2013 State : clean, reshaping Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Reshape Status : 0% complete Delta Devices : -1, (4->3) Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10199 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 4 8 18 3 active sync /dev/sdb2 root@grml ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 02:45:52 2012 Raid Level : raid5 Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 3 Total Devices : 4 Persistence : Superblock is persistent Update Time : Wed Dec 11 20:39:03 2013 State : clean Active Devices : 3 Working Devices : 4 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10273 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 4 8 18 - spare /dev/sdb2root@grml ~ # mdadm --verbose --manage /dev/md2 --remove /dev/sdb2 :( mdadm: hot removed /dev/sdb2 from /dev/md2 root@grml ~ # mdadm -D /dev/md2 /dev/md2: Version : 1.2 Creation Time : Tue Aug 28 02:45:52 2012 Raid Level : raid5 Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 26215424 (25.00 GiB 26.84 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Update Time : Wed Dec 11 20:44:24 2013 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : localhost.localdomain:2 UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Events : 10274 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 6 8 83 1 active sync /dev/sdf3 5 8 67 2 active sync /dev/sde3 root@grml ~ # e2fsck -nvf /dev/md2 :( e2fsck 1.42.8 (20-Jun-2013) Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information 607662 inodes used (39.88%, out of 1523712) 2606 non-contiguous files (0.4%) 226 non-contiguous directories (0.0%) # of inodes with ind/dind/tind blocks: 0/0/0 Extent depth histogram: 599732/802 4365533 blocks used (71.84%, out of 6076945) 0 bad blocks 1 large file 545388 regular files 54723 directories 180 character device files 97 block device files 2 fifos 597 links 7262 symbolic links (6840 fast symbolic links) 1 socket ------------ 608250 files root@grml ~ # resize2fs -p /dev/md2 resize2fs 1.42.8 (20-Jun-2013) Resizing the filesystem on /dev/md2 to 13107712 (4k) blocks. The filesystem on /dev/md2 is now 13107200 blocks long. -> 78646272 / 3 (aktive devs) = 26215424; 26215424 * 2 (zukünftig aktive devs) = 52430848 === Detaillierte mdadm Informationen nach dem Umzug === Hilfreich wenn es Probleme mit mdadm gibt: nas ~ # mdadm -Evvvvs mdadm: No md superblock detected on /dev/md4. mdadm: No md superblock detected on /dev/md0. mdadm: No md superblock detected on /dev/md2. /dev/sdc4: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : f6924d42:7358df85:416f97f9:83aeac62 Name : nas:4 (local to host nas) Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 5753363343 (2743.42 GiB 2945.72 GB) Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 5753362432 (2743.42 GiB 2945.72 GB) Data Offset : 262144 sectors Super Offset : 8 sectors Unused Space : before=262056 sectors, after=911 sectors State : clean Device UUID : 571c6397:7414485c:bc2f56b7:d96e1278 Internal Bitmap : 8 sectors from superblock Update Time : Sun Dec 15 17:01:47 2013 Bad Block Log : 512 entries available at offset 72 sectors Checksum : 7e228231 - correct Events : 15769 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 1 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) /dev/sdc3: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Name : localhost.localdomain:2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 104855552 (50.00 GiB 53.69 GB) Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 52430848 (25.00 GiB 26.84 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1960 sectors, after=52424704 sectors State : active Device UUID : cdfa2219:81384094:5f4cab42:ca201111 Update Time : Mon Dec 16 22:04:43 2013 Bad Block Log : 512 entries available at offset 72 sectors Checksum : 5adf8d9c - correct Events : 10275 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 2 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) mdadm: No md superblock detected on /dev/sdc2. /dev/sdc1: Magic : a92b4efc Version : 1.0 Feature Map : 0x0 Array UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Name : localhost.localdomain:0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Raid Devices : 2 Avail Dev Size : 1023976 (500.07 MiB 524.28 MB) Array Size : 511988 (500.07 MiB 524.28 MB) Super Offset : 1023984 sectors State : clean Device UUID : d5716487:f5940dab:47a8c91d:5add272a Update Time : Wed Dec 11 22:27:12 2013 Bad Block Log : 512 entries available at offset -8 sectors Checksum : 383b258a - correct Events : 209 Device Role : Active device 1 Array State : AA ('A' == active, '.' == missing, 'R' == replacing) /dev/sdc: MBR Magic : aa55 Partition[0] : 4294967295 sectors at 1 (type ee) /dev/sdb4: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : f6924d42:7358df85:416f97f9:83aeac62 Name : nas:4 (local to host nas) Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 5753363343 (2743.42 GiB 2945.72 GB) Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 5753362432 (2743.42 GiB 2945.72 GB) Data Offset : 262144 sectors Super Offset : 8 sectors Unused Space : before=262056 sectors, after=911 sectors State : clean Device UUID : 4b7b8103:30e8cb92:a7820f4d:1ea29fad Internal Bitmap : 8 sectors from superblock Update Time : Sun Dec 15 17:01:47 2013 Bad Block Log : 512 entries available at offset 72 sectors Checksum : ec0b3b11 - correct Events : 15769 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 2 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) /dev/sdb3: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Name : localhost.localdomain:2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 104855552 (50.00 GiB 53.69 GB) Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 52430848 (25.00 GiB 26.84 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1960 sectors, after=52424704 sectors State : active Device UUID : ce0a470b:360249bf:136e7cfa:b7ca891b Update Time : Mon Dec 16 22:04:43 2013 Bad Block Log : 512 entries available at offset 72 sectors Checksum : 3a5632f5 - correct Events : 10275 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 1 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) mdadm: No md superblock detected on /dev/sdb2. /dev/sdb1: Magic : a92b4efc Version : 1.0 Feature Map : 0x0 Array UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Name : localhost.localdomain:0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Raid Devices : 2 Avail Dev Size : 1023976 (500.07 MiB 524.28 MB) Array Size : 511988 (500.07 MiB 524.28 MB) Super Offset : 1023984 sectors State : clean Device UUID : 5c8b5072:c1efebaa:21248884:16eb4c6e Update Time : Wed Dec 11 22:27:12 2013 Bad Block Log : 512 entries available at offset -8 sectors Checksum : cde92372 - correct Events : 209 Device Role : spare Array State : AA ('A' == active, '.' == missing, 'R' == replacing) /dev/sdb: MBR Magic : aa55 Partition[0] : 4294967295 sectors at 1 (type ee) /dev/sda4: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : f6924d42:7358df85:416f97f9:83aeac62 Name : nas:4 (local to host nas) Creation Time : Mon Dec 9 22:06:12 2013 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 5753363343 (2743.42 GiB 2945.72 GB) Array Size : 5753362432 (5486.83 GiB 5891.44 GB) Used Dev Size : 5753362432 (2743.42 GiB 2945.72 GB) Data Offset : 262144 sectors Super Offset : 8 sectors Unused Space : before=262056 sectors, after=911 sectors State : clean Device UUID : 23e771d6:94004c21:4b1d5ef6:de3a33dc Internal Bitmap : 8 sectors from superblock Update Time : Sun Dec 15 17:01:47 2013 Bad Block Log : 512 entries available at offset 72 sectors Checksum : 255df2b4 - correct Events : 15769 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 0 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) /dev/sda3: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : a9d908f3:f8632fcf:59a69085:83d5f6aa Name : localhost.localdomain:2 Creation Time : Tue Aug 28 04:45:52 2012 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 104855552 (50.00 GiB 53.69 GB) Array Size : 52430848 (50.00 GiB 53.69 GB) Used Dev Size : 52430848 (25.00 GiB 26.84 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1968 sectors, after=52424704 sectors State : clean Device UUID : 11727d2d:55248f2f:c6bfd477:6ee8c689 Update Time : Mon Dec 16 22:04:43 2013 Checksum : b8602b71 - correct Events : 10274 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 0 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) /dev/sda2: Magic : a92b4efc Version : 1.0 Feature Map : 0x0 Array UUID : 262c8e68:c7c4fa6d:19af27e8:0cfc68a1 Name : localhost.localdomain:0 Creation Time : Fri Aug 24 04:18:31 2012 Raid Level : raid1 Raid Devices : 2 Avail Dev Size : 1023976 (500.07 MiB 524.28 MB) Array Size : 511988 (500.07 MiB 524.28 MB) Super Offset : 1023984 sectors Unused Space : before=0 sectors, after=8 sectors State : clean Device UUID : d004dd2b:281bdf01:1ab30378:b08f7be0 Update Time : Wed Dec 11 22:27:12 2013 Checksum : 440afbe4 - correct Events : 209 Device Role : Active device 0 Array State : AA ('A' == active, '.' == missing, 'R' == replacing) /dev/sda1: MBR Magic : aa55 /dev/sda: MBR Magic : aa55 Partition[0] : 4294967295 sectors at 1 (type ee) ===== Virtualisierung ===== Die aktuellen Windows virtio-Treiber gibt es unter. http://alt.fedoraproject.org/pub/alt/virtio-win/latest/images/bin/ ===== web.py ===== ==== Postgres Datenbanken ==== === Serial === web.py: get auto-increment value after db.insert This is probably too simple, so no one ever bothered to write about it. (Or I couldn't come up with the proper keywords for Google). The situation is simple. I have a table where the primary key is some auto-incrementing ID number (in the case of postgresql, with the primary key called nid, the definition is nid serial primary key). After inserting a record into the table, I want to know what ID it has got. Turns out it's very simple. I need only to pass a custom seqname to db.insert(). nid = db.insert("tablename", seqname="tablename_nid_seq",...) For the serial column, postgresql creates a sequence automatically and assigns it a particularly constructed name. More details at FAQ: Using Sequences in PostgreSQL. This is probably not the best style (I'm hard-coding the seqname in the source code), and I'm not even sure if it's the right way. But at least I can proceed with my prototyping and hacking now :) http://blog.zhangsen.org/2011/07/webpy-get-auto-increment-value-after.html