Mailinglist Archive: opensuse (4547 mails)

< Previous Next >
software raid-1 issue
  • From: UGRON Balazs <balee@xxxxxxxxxxx>
  • Date: Mon, 24 May 2004 10:43:57 +0200 (CEST)
  • Message-id: <Pine.LNX.4.44.0405241042520.31248-100000@xxxxxxxxxxxxxxxxx>

Hey,

I have a Suse 9.0 pro, and the partitions work on raid-1. I have a strange
problem and now I don't have any ideas what to do (man, google, and
all my ideas failed).

So, the situation: hda3 and hdc3 are in the software raid-1 array md1,
which array _seems_ to be working well (if I monitor the winchester I/O,
then hda and hdc are both used with the same throughput while I reading
or writing md1), but I can see strange things with lsraid (see later).

So, what I see:

The concerned part of /var/log/boot.msg:
========================================

[...]
<6>md: considering hda3 ...
<6>md: adding hda3 ...
<6>md: adding hdc3 ...
<6>md: created md1
<6>md: bind<hdc3,1>
<6>md: bind<hda3,2>
<6>md: running: <hda3><hdc3>
<6>md: hda3's event counter: 000000ba
<6>md: hdc3's event counter: 000000ba
<6>md: RAID level 1 does not need chunksize! Continuing anyway.
<6>md1: max total readahead window set to 508k
<6>md1: 1 data-disks, max readahead per data-disk: 508k
<6>raid1: device hda3 operational as mirror 0
<6>raid1: device hdc3 operational as mirror 1
<6>raid1: raid set md1 active with 2 out of 2 mirrors
<6>md: updating md1 RAID superblock on device
<6>md: hda3 [events: 000000bb]
<6>(write) hda3's sb offset: 10490368
<6>md: hdc3 [events: 000000bb]
<6>(write) hdc3's sb offset: 10490368
<6> [events: 00000000]
<3>md: invalid raid superblock magic on md1
<4>md: md1 has invalid sb, not importing!
<4>md: no nested md device found
[...]

After this nothing special with md1, hda, hdc.
I don't really understand the last three lines, but they are there
at every other raid-1 array too, so I'm sure this is not that problem.
I think it's pretty good so far.

/var/log/messages
=================

There is no string "hdc" since the last boot.


By /proc/mdstat the raid is cool:
=================================

root@zuza:~# cat /proc/mdstat
Personalities : [raid1]
read_ahead 1024 sectors
[...]
md1 : active raid1 hdc3[1] hda3[0]
10490368 blocks [2/2] [UU]
[...]
unused devices: <none>

That's ok, too.

But lsraid says:
===============

root@zuza:~# lsraid -p
[...]
[dev 9, 1] /dev/md1 A17B3B97.AC450237.7C21B1E7.94EBAD5D online
[dev 3, 3] /dev/hda3 A17B3B97.AC450237.7C21B1E7.94EBAD5D good
[...]
[dev 22, 3] /dev/hdc3 A17B3B97.AC450237.7C21B1E7.94EBAD5D unbound
[...]

lsraid part II:
===============

root@zuza:~# lsraid -l -D -p
[dev 3, 3] /dev/hda3:
md version = 0.90.0
superblock uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
md minor number = 1
created = 1073584440 (Thu Jan 8 18:54:00 2004)
last updated = 1085065922 (Thu May 20 17:12:02 2004)
raid level = 1
chunk size = 4 KB
apparent disk size = 10490368 KB
disks in array = 1
required disks = 2
active disks = 2
working disks = 2
failed disks = -1
spare disks = 0
position in disk list = 0
position in md device = 0
state = good
[...]
[dev 22, 3] /dev/hdc3:
md version = 0.90.0
superblock uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
md minor number = 1
created = 1073584440 (Thu Jan 8 18:54:00 2004)
last updated = 1085059886 (Thu May 20 15:31:26 2004)
raid level = 1
chunk size = 4 KB
apparent disk size = 10490368 KB
disks in array = 1
required disks = 2
active disks = 2
working disks = 2
failed disks = -1
spare disks = 0
position in disk list = 1
position in md device = 1
state = unknown
[...]

It is strange for me, that "disks in array=1" "active disks=2".
Why says lsraid, that hdc3 isn't in the array?
And ok, it is not in, but then why is hdc disk active?

lsraid part III:
================

root@zuza:~# lsraid -D -p
[dev 3, 3] /dev/hda3:
md device = [dev 9, 1] /dev/md1
md uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
state = good

[dev 22, 3] /dev/hdc3:
old md device = [dev 9, 1]
old md uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
state = unknown

/etc/raidtab (I generated with lsraid):
=======================================

raiddev /dev/md1
raid-level 1
nr-raid-disks 2
nr-spare-disks 0
persistent-superblock 1
chunk-size 4

device /dev/hda3
raid-disk 0
device /dev/null
failed-disk 1

I tried everything I could imagine - my best idea was to remove hdc3 from
the array, clear the partition and then put it back to the array again.
After this, there is the same situation like before, moreover, I saw
strange things again after a removed hdc3:

Remove hdc3:
============

root@zuza:~# raidsetfaulty /dev/md1 /dev/hdc3
root@zuza:~# raidhotremove /dev/md1 /dev/hdc3

lsraid after:
=============

root@zuza:~# lsraid -p
[...]
[dev 9, 1] /dev/md1 00000000.00000000.00000000.00000000 online
[dev ?, ?] (unknown) 00000000.00000000.00000000.00000000 missing
[...]
[dev 22, 3] /dev/hdc3 A17B3B97.AC450237.7C21B1E7.94EBAD5D unbound
[...]

Hmmm... there is nothing in the raid array??? Moreover, hda3 disappeared
from the array, after I removed hdc3??? And there is no hdc3 even in the
unbound list.
And surprising (at least for me), that md1 works even now, I see the
file system, etc (only hda is used this time - by I/O monitoring).

The az mdstat this time:
========================

root@zuza:~# cat /proc/mdstat
Personalities : [raid1]
read_ahead 1024 sectors
[...]
md1 : active raid1 hda3[0]
10490368 blocks [2/1] [U_]
[...]
unused devices: <none>

I think it is ok again.

lsraid this time:
=================

root@zuza:~# lsraid -D -p
[...]
[dev 3, 3] /dev/hda3:
old md device = [dev 9, 1]
old md uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
state = unknown
[...]
[dev 22, 3] /dev/hdc3:
old md device = [dev 9, 1]
old md uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
state = unknown
[...]


lsraid this time II:
====================

root@zuza:~# lsraid -l -D -p
[...]
[dev 22, 3] /dev/hdc3:
md version = 0.90.0
superblock uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
md minor number = 1
created = 1073584440 (Thu Jan 8 18:54:00 2004)
last updated = 1085065922 (Thu May 20 17:12:02 2004)
raid level = 1
chunk size = 4 KB
apparent disk size = 10490368 KB
disks in array = 1
required disks = 2
active disks = 2
working disks = 2
failed disks = -1
spare disks = 0
position in disk list = 1
position in md device = 1
state = unknown
[...]
[dev 3, 3] /dev/hda3:
md version = 0.90.0
superblock uuid = A17B3B97.AC450237.7C21B1E7.94EBAD5D
md minor number = 1
created = 1073584440 (Thu Jan 8 18:54:00 2004)
last updated = 1085065922 (Thu May 20 17:12:02 2004)
raid level = 1
chunk size = 4 KB
apparent disk size = 10490368 KB
disks in array = 1
required disks = 2
active disks = 2
working disks = 2
failed disks = -1
spare disks = 0
position in disk list = 0
position in md device = 0
state = unknown
[...]

So, it seems for me, that everything works fine, only lsraid shows
strange things. Although, lsraid shows correctly the other raid-1 arrays,
so it is frustrating.

This situation is since I had to remove the disk hda for a while.

If anybody has a good idea what this is, and how I can fix it,
tell me please.

Thanks,
Balazs



< Previous Next >
This Thread
  • No further messages