Jump to content

Featured Replies

Posted

Hi,

 

It's been a long time since i needed any help with Linux but now I'm at the point of tearing my hair out.

 

I'm using Mint 13 with Mate desktop.

I've been running a mdadm Raid 6 system for quite some time.

It consists of 4 hard drives consisting of 2TB's each.

 

Yesterday I went to access my Raid 6 at md0 using the icon found in the Computer folder.

The icon was not present.

When i opened Disk Utility and checked the Raid 6 entry it showed it was up and running.

All drives show no errors and they are in the proper configuration.

 

I decided to stop mdadm and do a repair/resync.

That is when i noticed that when I stopped mdadm the icon I usually click on appeared in the Computer folder.

 

Today the repair/resync finished but I am still unable to access it.

 

I tried numerous commands to check the array and disks and nothing appears to be wrong.

My Raid 6 uses the following drives;

sde, sdf, sdg, sdi

 

If I try to mount md0 from the terminal using mount /dev/md0 I get an error that md0 is not found in fstab.

So I added an entry in fstab to mount it on boot.

It did not mount and generated no error messages.

 

I checked the kernel logs and they all have the same entries for the past week.

In fact i have another hard drive that has Mint on it and also ran my Raid 6. I replaced it about 3 months ago with a new drive because it was failing.

The kernel logs on that drive have the same entries for mdadm and md0 that my new drive has.

 

Here is a copy of today's kernel logs;

 

Code:

ar 23 10:11:20 DaLinux kernel: [ 5.551901] md: bind<sdf>

Mar 23 10:11:20 DaLinux kernel: [ 5.569025] md: bind<sde>

Mar 23 10:11:20 DaLinux kernel: [ 5.572212] md: bind<sdi>

Mar 23 10:11:20 DaLinux kernel: [ 5.601595] md: bind<sdg>

Mar 23 10:11:20 DaLinux kernel: [ 5.629825] bio: create slab <bio-1> at 1

Mar 23 10:11:20 DaLinux kernel: [ 5.629840] md/raid:md0: device sdg operational as raid disk 2

Mar 23 10:11:20 DaLinux kernel: [ 5.629842] md/raid:md0: device sdi operational as raid disk 3

Mar 23 10:11:20 DaLinux kernel: [ 5.629844] md/raid:md0: device sde operational as raid disk 0

Mar 23 10:11:20 DaLinux kernel: [ 5.629845] md/raid:md0: device sdf operational as raid disk 1

Mar 23 10:11:20 DaLinux kernel: [ 5.630175] md/raid:md0: allocated 4280kB

Mar 23 10:11:20 DaLinux kernel: [ 5.630222] md/raid:md0: raid level 6 active with 4 out of 4 devices, algorithm 2

Mar 23 10:11:20 DaLinux kernel: [ 5.630224] RAID conf printout:

Mar 23 10:11:20 DaLinux kernel: [ 5.630225] --- level:6 rd:4 wd:4

Mar 23 10:11:20 DaLinux kernel: [ 5.630226] disk 0, o:1, dev:sde

Mar 23 10:11:20 DaLinux kernel: [ 5.630227] disk 1, o:1, dev:sdf

Mar 23 10:11:20 DaLinux kernel: [ 5.630228] disk 2, o:1, dev:sdg

Mar 23 10:11:20 DaLinux kernel: [ 5.630229] disk 3, o:1, dev:sdi

Mar 23 10:11:20 DaLinux kernel: [ 5.630250] md0: detected capacity change from 0 to 4000528203776

 

 

Here is is the output of cat /proc/mdstat:

 

Code:

$ cat /proc/mdstat

Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]

md0 : active raid6 sde[0] sdi[3] sdg[2] sdf[1]

3906765824 blocks super 1.2 level 6, 512k chunk, algorithm 2 [4/4] [uUUU]

 

unused devices: <none>

I tried the following command and here is the output;

 

Code:

sudo mdadm --assemble --force /dev/md0 /dev/sd[efgi]

mdadm: /dev/md0 has been started with 4 drives.

I also ran this command and it shows that md0 is clean;

 

Code:

sudo mdadm --detail /dev/md0

 

/dev/md0:

Version : 1.2

Creation Time : Mon Jan 21 21:21:18 2013

Raid Level : raid6

Array Size : 3906765824 (3725.78 GiB 4000.53 GB)

Used Dev Size : 1953382912 (1862.89 GiB 2000.26 GB)

Raid Devices : 4

Total Devices : 4

Persistence : Superblock is persistent

 

Update Time : Sun Mar 23 16:10:38 2014

State : clean

Active Devices : 4

Working Devices : 4

Failed Devices : 0

Spare Devices : 0

 

Layout : left-symmetric

Chunk Size : 512K

 

Name : DaNixBox:0

UUID : 320fd47e:235e2eee:675ae146:6bb49a2e

Events : 46

 

Number Major Minor RaidDevice State

0 8 64 0 active sync /dev/sde

1 8 80 1 active sync /dev/sdf

2 8 96 2 active sync /dev/sdg

3 8 128 3 active sync /dev/sdi

 

Any help would be appreciated.

Thanks!

 

Continue reading...

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...