Ik heb (of misschien had) hier een raid5 systeem van 4 disks van 250GB elk. Omdat mijn systeem soms vreemd bootte heb ik 2 van de 4 disks voorzien van een andere ide kabel.
Blijkbaar had ik deze kabel niet goed aangedrukt (DOMDOMDOM!) en is deze vermoedelijk lostgeraakt. Het gevolg was dat er dus 2 disken weg waren uit de raid5. Dit is dus dodelijk voor raid5..
Oct 22 16:46:26 server01 kernel: [16139.192000] hdc: drive_cmd: status=0x51 { DriveReady SeekComplete Error }
Oct 22 16:46:26 server01 kernel: [16139.192000] hdc: drive_cmd: error=0x04 { DriveStatusError }
Oct 22 16:46:26 server01 kernel: [16139.192000] ide: failed opcode was: 0xea
Oct 22 16:46:26 server01 kernel: [16139.192000] md: super_written gets error=-5, uptodate=0
Oct 22 16:46:26 server01 kernel: [16139.192000] hdd: drive_cmd: status=0x2a { DeviceFault DataRequest Index }
Oct 22 16:46:26 server01 kernel: [16139.192000] ide: failed opcode was: 0xea
Oct 22 16:46:46 server01 kernel: [16159.192000] hdd: dma_timer_expiry: dma status == 0x61
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: DMA timeout error
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: dma timeout error: status=0x35 { DeviceFault SeekComplete CorrectedError Error }
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: dma timeout error: error=0x35 { DriveStatusError SectorIdNotFound AddrMarkNotFound }, LBAsect=8830590465333, high=526344, low=3487029, sector=488391871
Oct 22 16:46:56 server01 kernel: [16169.192000] ide: failed opcode was: unknown
Oct 22 16:46:56 server01 kernel: [16169.192000] hdc: DMA disabled
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: DMA disabled
Oct 22 16:46:56 server01 kernel: [16169.240000] ide1: reset: master: error (0x50?)
Oct 22 16:46:56 server01 kernel: [16169.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:46:56 server01 kernel: [16169.240000] ide: failed opcode was: unknown
Oct 22 16:46:56 server01 kernel: [16169.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:46:56 server01 kernel: [16169.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: lost interrupt
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:47:26 server01 kernel: [16199.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:47:26 server01 kernel: [16199.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.288000] ide1: reset: master: error (0x50?)
Oct 22 16:47:26 server01 kernel: [16199.288000] end_request: I/O error, dev hdd, sector 488391871
Oct 22 16:47:26 server01 kernel: [16199.288000] md: super_written gets error=-5, uptodate=0
Oct 22 16:47:26 server01 kernel: [16199.320000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.320000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 0, o:0, dev:hdc1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.320000] lost page write due to I/O error on dm-0
Oct 22 16:47:26 server01 last message repeated 9 times
Oct 22 16:47:26 server01 kernel: [16199.336000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.336000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.336000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.336000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.348000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.348000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.348000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.348000] disk 3, o:1, dev:hda1
Oct 22 16:49:15 server01 kernel: [16308.228000] ReiserFS: dm-0: warning: zam-7001: io error in reiserfs_find_entry
Oct 22 16:49:15 server01 last message repeated 2 times
Oct 22 16:49:17 server01 kernel: [16309.844000] ReiserFS: dm-0: warning: zam-7001: io error in reiserfs_find_entry
Oct 22 16:49:17 server01 last message repeated 2 times
SCHRIK!!!
Nu probeer ik de array te starten in forced mode, maar ook dit lukt niet meer.. Hij heeft hdd1 helemaal uit het array verwijderd (kicking non-fresh hdd1 from array!). Vervolgens geprobeerd het array op te starten zonder hdd, maar nog niks:
root@server01:/var/log# mdadm -A /dev/md0 /dev/hdc1 /dev/hdb1 /dev/hda1
mdadm: no recogniseable superblock on /dev/hdc1
mdadm: /dev/hdc1 has no superblock - assembly aborted
Ik durf niet zoveel meer te doen, ik hoop dat er ergens een raid guru rondloopt die me weer aan mijn data kan helpen........
Blijkbaar had ik deze kabel niet goed aangedrukt (DOMDOMDOM!) en is deze vermoedelijk lostgeraakt. Het gevolg was dat er dus 2 disken weg waren uit de raid5. Dit is dus dodelijk voor raid5..
Oct 22 16:46:26 server01 kernel: [16139.192000] hdc: drive_cmd: status=0x51 { DriveReady SeekComplete Error }
Oct 22 16:46:26 server01 kernel: [16139.192000] hdc: drive_cmd: error=0x04 { DriveStatusError }
Oct 22 16:46:26 server01 kernel: [16139.192000] ide: failed opcode was: 0xea
Oct 22 16:46:26 server01 kernel: [16139.192000] md: super_written gets error=-5, uptodate=0
Oct 22 16:46:26 server01 kernel: [16139.192000] hdd: drive_cmd: status=0x2a { DeviceFault DataRequest Index }
Oct 22 16:46:26 server01 kernel: [16139.192000] ide: failed opcode was: 0xea
Oct 22 16:46:46 server01 kernel: [16159.192000] hdd: dma_timer_expiry: dma status == 0x61
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: DMA timeout error
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: dma timeout error: status=0x35 { DeviceFault SeekComplete CorrectedError Error }
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: dma timeout error: error=0x35 { DriveStatusError SectorIdNotFound AddrMarkNotFound }, LBAsect=8830590465333, high=526344, low=3487029, sector=488391871
Oct 22 16:46:56 server01 kernel: [16169.192000] ide: failed opcode was: unknown
Oct 22 16:46:56 server01 kernel: [16169.192000] hdc: DMA disabled
Oct 22 16:46:56 server01 kernel: [16169.192000] hdd: DMA disabled
Oct 22 16:46:56 server01 kernel: [16169.240000] ide1: reset: master: error (0x50?)
Oct 22 16:46:56 server01 kernel: [16169.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:46:56 server01 kernel: [16169.240000] ide: failed opcode was: unknown
Oct 22 16:46:56 server01 kernel: [16169.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:46:56 server01 kernel: [16169.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: lost interrupt
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:47:26 server01 kernel: [16199.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.240000] hdd: status error: status=0x50 { DriveReady SeekComplete }
Oct 22 16:47:26 server01 kernel: [16199.240000] ide: failed opcode was: unknown
Oct 22 16:47:26 server01 kernel: [16199.288000] ide1: reset: master: error (0x50?)
Oct 22 16:47:26 server01 kernel: [16199.288000] end_request: I/O error, dev hdd, sector 488391871
Oct 22 16:47:26 server01 kernel: [16199.288000] md: super_written gets error=-5, uptodate=0
Oct 22 16:47:26 server01 kernel: [16199.320000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.320000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 0, o:0, dev:hdc1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.320000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.320000] lost page write due to I/O error on dm-0
Oct 22 16:47:26 server01 last message repeated 9 times
Oct 22 16:47:26 server01 kernel: [16199.336000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.336000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.336000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.336000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 2, o:0, dev:hdd1
Oct 22 16:47:26 server01 kernel: [16199.336000] disk 3, o:1, dev:hda1
Oct 22 16:47:26 server01 kernel: [16199.348000] RAID5 conf printout:
Oct 22 16:47:26 server01 kernel: [16199.348000] --- rd:4 wd:2
Oct 22 16:47:26 server01 kernel: [16199.348000] disk 1, o:1, dev:hdb1
Oct 22 16:47:26 server01 kernel: [16199.348000] disk 3, o:1, dev:hda1
Oct 22 16:49:15 server01 kernel: [16308.228000] ReiserFS: dm-0: warning: zam-7001: io error in reiserfs_find_entry
Oct 22 16:49:15 server01 last message repeated 2 times
Oct 22 16:49:17 server01 kernel: [16309.844000] ReiserFS: dm-0: warning: zam-7001: io error in reiserfs_find_entry
Oct 22 16:49:17 server01 last message repeated 2 times
SCHRIK!!!
Nu probeer ik de array te starten in forced mode, maar ook dit lukt niet meer.. Hij heeft hdd1 helemaal uit het array verwijderd (kicking non-fresh hdd1 from array!). Vervolgens geprobeerd het array op te starten zonder hdd, maar nog niks:
root@server01:/var/log# mdadm -A /dev/md0 /dev/hdc1 /dev/hdb1 /dev/hda1
mdadm: no recogniseable superblock on /dev/hdc1
mdadm: /dev/hdc1 has no superblock - assembly aborted
Ik durf niet zoveel meer te doen, ik hoop dat er ergens een raid guru rondloopt die me weer aan mijn data kan helpen........