Gentoo Forums
Gentoo Forums
Gentoo Forums
Quick Search: in
raidreconf RAID5 grow fail / lsraid strangeness...
View unanswered posts
View posts from last 24 hours

 
Reply to topic    Gentoo Forums Forum Index Other Things Gentoo
View previous topic :: View next topic  
Author Message
autom8on
n00b
n00b


Joined: 27 Apr 2005
Posts: 11

PostPosted: Tue Dec 06, 2005 9:44 am    Post subject: raidreconf RAID5 grow fail / lsraid strangeness... Reply with quote

All,

I created a 4 disk (4x120GB) RAID-5 setup which has been working fine (since I swapped out a broken IDE controller) - here's it's /etc/raidtab file:

Code:
raiddev /dev/md0
        raid-level      5
        nr-raid-disks   4
        nr-spare-disks  0
        persistent-superblock 1
        parity-algorithm left-symmetric
        chunk-size      128
        device          /dev/hda1
        raid-disk       0
        device          /dev/hdb1
        raid-disk       1
        device          /dev/hdc1
        raid-disk       2
        device          /dev/hdd1
        raid-disk       3


After copying some data around, I managed to free up a couple more partitions. /dev/hdf is the exact same make/model/size as the disks in the existing array (120GB), whilst /dev/hdh is a slightly bigger disk (200GB):

Code:
dmesg | grep sectors
hda: 234493056 sectors (120060 MB) w/8192KiB Cache, CHS=16383/255/63, UDMA(100)
hdb: 234493056 sectors (120060 MB) w/8192KiB Cache, CHS=16383/255/63, UDMA(100)
hdc: 234493056 sectors (120060 MB) w/8192KiB Cache, CHS=16383/255/63, UDMA(100)
hdd: 234493056 sectors (120060 MB) w/8192KiB Cache, CHS=16383/255/63, UDMA(100)
hde: 156301488 sectors (80026 MB) w/2048KiB Cache, CHS=65535/16/63, UDMA(100)
hdf: 234493056 sectors (120060 MB) w/8192KiB Cache, CHS=16383/255/63, UDMA(100)
hdh: 390721968 sectors (200049 MB) w/8192KiB Cache, CHS=24321/255/63, UDMA(100)


Here's the fdisk info from a disk in the existing array:

Code:
Disk /dev/hda: 120.0 GB, 120060444672 bytes
255 heads, 63 sectors/track, 14596 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hda1               1       14596   117242338+  fd  Linux raid autodetect


So I created a new partition at hdf1 to look the same:

Code:
Disk /dev/hdf: 120.0 GB, 120060444672 bytes
255 heads, 63 sectors/track, 14596 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hdf1               1       14596   117242338+  fd  Linux raid autodetect


Next I spllit the larger drive down into two partitions (is this valid? I'm assuming that I can use a partition rather than having to assign the entire drive? I'm hoping to eventually use the second partition to mirror my boot-disk (I'll get around to that once I've sorted this problem out)):

Code:
Disk /dev/hdh: 200.0 GB, 200049647616 bytes
255 heads, 63 sectors/track, 24321 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hdh1               1       14596   117242338+  fd  Linux raid autodetect
/dev/hdh2           14597       24321    78116062+  83  Linux


So - at this point I think I've created two new partitions of the exact same size as the old partitions. Then I created an /etc/raidtab.new file to include these new partitions:

Code:
raiddev /dev/md0
        raid-level      5
        nr-raid-disks   6
        nr-spare-disks  0
        persistent-superblock 1
        parity-algorithm left-symmetric
        chunk-size      128
        device          /dev/hda1
        raid-disk       0
        device          /dev/hdb1
        raid-disk       1
        device          /dev/hdc1
        raid-disk       2
        device          /dev/hdd1
        raid-disk       3
        device          /dev/hdf1
        raid-disk       4
        device          /dev/hdh1
        raid-disk       5


I unmounted the raid array - and ran raidstop /dev/md0. However, when I try and grow the array, I just get an error:

Code:
# raidreconf -o /etc/raidtab -n /etc/raidtab.new -m /dev/md0
Working with device /dev/md0
Parsing /etc/raidtab
Parsing /etc/raidtab.new
reconfiguration failed


I can't seem to get anything more verbose out of it than the "reconfiguration failed" message and there doesn't seem to be anything meaningful in any of the log files I've looked at.

Ideas, anyone?

Cheers,

Steve.


Last edited by autom8on on Wed Dec 07, 2005 2:32 pm; edited 2 times in total
Back to top
View user's profile Send private message
AxelTerizaki
n00b
n00b


Joined: 12 Apr 2003
Posts: 33

PostPosted: Tue Dec 06, 2005 1:02 pm    Post subject: Reply with quote

Consider yourself lucky to have had this error message.

<mylife>
On last friday I tried to add a 250 Gb to my already existing array of 5 x 250 Gb.

The problem was, one of my HDs was a slightly older model and was like 300 Mb shorter than the 4 others. So the RAID array was based on the size of this particular disk.

However, the raidreconf tool started resizing the RAID array, but failed at 99% (-_-) with a "aborting" error: it said the block was out of range on the slightly shorter disk, which isn't surprising if it tried to resize it TOO...

I guess this is a bug, but it made me lost 1 Tb of data.
(hooray for backups)
</mylife>
Back to top
View user's profile Send private message
autom8on
n00b
n00b


Joined: 27 Apr 2005
Posts: 11

PostPosted: Wed Dec 07, 2005 1:14 am    Post subject: Reply with quote

AxelTerizaki wrote:
Consider yourself lucky to have had this error message.

Well, I just found some more... I noticed someone talking about the lsraid tool in another thread - so I thought I'd have a play around with that and see if I could gain any enlightenment. Sadly. it's just confused me even more by giving me some odd messages.

Code:
# lsraid -A -a /dev/md0
lsraid: Unable to find superblock for "/dev/hdc1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdd1": Invalid argument
lsraid: Unable to find superblock for "/dev/hda1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdb1": Invalid argument
[dev   9,   0] /dev/md0         00000000.00000000.00000000.00000000 online
[dev   ?,   ?] (unknown)        00000000.00000000.00000000.00000000 missing
[dev   ?,   ?] (unknown)        00000000.00000000.00000000.00000000 missing
[dev   ?,   ?] (unknown)        00000000.00000000.00000000.00000000 missing
[dev   ?,   ?] (unknown)        00000000.00000000.00000000.00000000 missing


Code:
lsraid -D -a /dev/md0
lsraid: Unable to find superblock for "/dev/hdc1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdd1": Invalid argument
lsraid: Unable to find superblock for "/dev/hda1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdb1": Invalid argument
lsraid: md device [dev 9, 0] /dev/md0 is offline: Please specify a disk to query


Although it claims that the device is offline at the time I ran it ^^^ - it wasn't. It was live and mounted...

Code:
lsraid -R -a /dev/md0
lsraid: Unable to find superblock for "/dev/hdc1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdd1": Invalid argument
lsraid: Unable to find superblock for "/dev/hda1": Invalid argument
lsraid: Unable to find superblock for "/dev/hdb1": Invalid argument
# This raidtab was generated by lsraid version 0.7.0.
# It was created from a query on the following devices:
#       /dev/md0

# md device [dev 9, 0] /dev/md0 queried online
raiddev /dev/md0
        raid-level              5
        nr-raid-disks           4
        nr-spare-disks          0
        persistent-superblock   1
        chunk-size              128

        device          /dev/null
        failed-disk             0
        device          /dev/null
        failed-disk             1
        device          /dev/null
        failed-disk             2
        device          /dev/null
        failed-disk             3


Any more ideas, anyone? I notice that /dev/md0 is actually a link to /dev/md/0 - does it matter that I'm referencing it indirectly - or shouldn't that make a lot of difference?

I'm hoping that the "Invalid argument" errors are just a bug in the way the lsraid software is calling whatever lower-level tools it uses.

Cheers,

Steve.
Back to top
View user's profile Send private message
Display posts from previous:   
Reply to topic    Gentoo Forums Forum Index Other Things Gentoo All times are GMT
Page 1 of 1

 
Jump to:  
You cannot post new topics in this forum
You cannot reply to topics in this forum
You cannot edit your posts in this forum
You cannot delete your posts in this forum
You cannot vote in polls in this forum