++ ATOT official NEF thread part IV ++

Page 2910 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.
Apr 20, 2008
10,161
984
126
reply.gif
 

Red Squirrel

No Lifer
May 24, 2003
67,196
12,025
126
www.anyf.ca
Added more disk space to one of my raid arrays, just because.

Code:
[root@isengard volumes]# mdadm --grow /dev/md0 --raid-devices=4
mdadm: Need to backup 24576K of critical section..
[root@isengard volumes]# 
[root@isengard volumes]# 
[root@isengard volumes]# mdadm --detail /dev/md0
/dev/md0:
        Version : 1.2
  Creation Time : Thu Jun 13 22:22:44 2013
     Raid Level : raid5
     Array Size : 5860270080 (5588.79 GiB 6000.92 GB)
  Used Dev Size : 2930135040 (2794.39 GiB 3000.46 GB)
   Raid Devices : 4
  Total Devices : 4
    Persistence : Superblock is persistent

    Update Time : Tue Sep  3 02:17:18 2013
          State : clean, reshaping 
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 4096K

 Reshape Status : 0% complete
  Delta Devices : 1, (3->4)

           Name : isengard.loc:0  (local to host isengard.loc)
           UUID : 88ce3b37:ef775547:b52541bd:0b2fce09
         Events : 73

    Number   Major   Minor   RaidDevice State
       0       8       32        0      active sync   /dev/sdc
       1       8       48        1      active sync   /dev/sdd
       3       8       64        2      active sync   /dev/sde
       4       8        0        3      active sync   /dev/sda
[root@isengard volumes]#

Actually was going to make that a raid 10. I had made it a raid 5 for now while I wait for a DOA drive to be RMAed but figured I'd just grow the raid 5 instead.

This could take a while.

Code:
[root@isengard volumes]# cat /proc/mdstat 
Personalities : [raid6] [raid5] [raid4] 
md0 : active raid5 sda[4] sde[3] sdd[1] sdc[0]
      5860270080 blocks super 1.2 level 5, 4096k chunk, algorithm 2 [4/4] [UUUU]
      [>....................]  reshape =  0.4% (13680392/2930135040) finish=802.0min speed=60606K/sec
      
unused devices: <none>
[root@isengard volumes]#


I don't actually have any plans for that raid array, it's empty. So I might change my mind and do 10. :p I will be taking all the drives from my existing server and putting them in this server, so this array is just spare space. Might use for extra backups or something. May also mirror Linux repos, because why not.

Goal is to setup Kerberos and LDAP and do a proper NFS setup.
 

Red Squirrel

No Lifer
May 24, 2003
67,196
12,025
126
www.anyf.ca
Added more disk space to one of my raid arrays, just because.

Code:
[root@isengard volumes]# mdadm --grow /dev/md0 --raid-devices=4
mdadm: Need to backup 24576K of critical section..
[root@isengard volumes]# 
[root@isengard volumes]# 
[root@isengard volumes]# mdadm --detail /dev/md0
/dev/md0:
        Version : 1.2
  Creation Time : Thu Jun 13 22:22:44 2013
     Raid Level : raid5
     Array Size : 5860270080 (5588.79 GiB 6000.92 GB)
  Used Dev Size : 2930135040 (2794.39 GiB 3000.46 GB)
   Raid Devices : 4
  Total Devices : 4
    Persistence : Superblock is persistent

    Update Time : Tue Sep  3 02:17:18 2013
          State : clean, reshaping 
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 4096K

 Reshape Status : 0% complete
  Delta Devices : 1, (3->4)

           Name : isengard.loc:0  (local to host isengard.loc)
           UUID : 88ce3b37:ef775547:b52541bd:0b2fce09
         Events : 73

    Number   Major   Minor   RaidDevice State
       0       8       32        0      active sync   /dev/sdc
       1       8       48        1      active sync   /dev/sdd
       3       8       64        2      active sync   /dev/sde
       4       8        0        3      active sync   /dev/sda
[root@isengard volumes]#

Actually was going to make that a raid 10. I had made it a raid 5 for now while I wait for a DOA drive to be RMAed but figured I'd just grow the raid 5 instead.

This could take a while.

Code:
[root@isengard volumes]# cat /proc/mdstat 
Personalities : [raid6] [raid5] [raid4] 
md0 : active raid5 sda[4] sde[3] sdd[1] sdc[0]
      5860270080 blocks super 1.2 level 5, 4096k chunk, algorithm 2 [4/4] [UUUU]
      [>....................]  reshape =  0.4% (13680392/2930135040) finish=802.0min speed=60606K/sec
      
unused devices: <none>
[root@isengard volumes]#


I don't actually have any plans for that raid array, it's empty. So I might change my mind and do 10. :p I will be taking all the drives from my existing server and putting them in this server, so this array is just spare space. Might use for extra backups or something. May also mirror Linux repos, because why not.

Goal is to setup Kerberos and LDAP and do a proper NFS setup.
 

mnewsham

Lifer
Oct 2, 2010
14,539
428
136
Added more disk space to one of my raid arrays, just because.

Code:
[root@isengard volumes]# mdadm --grow /dev/md0 --raid-devices=4
mdadm: Need to backup 24576K of critical section..
[root@isengard volumes]# 
[root@isengard volumes]# 
[root@isengard volumes]# mdadm --detail /dev/md0
/dev/md0:
        Version : 1.2
  Creation Time : Thu Jun 13 22:22:44 2013
     Raid Level : raid5
     Array Size : 5860270080 (5588.79 GiB 6000.92 GB)
  Used Dev Size : 2930135040 (2794.39 GiB 3000.46 GB)
   Raid Devices : 4
  Total Devices : 4
    Persistence : Superblock is persistent

    Update Time : Tue Sep  3 02:17:18 2013
          State : clean, reshaping 
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 4096K

 Reshape Status : 0% complete
  Delta Devices : 1, (3->4)

           Name : isengard.loc:0  (local to host isengard.loc)
           UUID : 88ce3b37:ef775547:b52541bd:0b2fce09
         Events : 73

    Number   Major   Minor   RaidDevice State
       0       8       32        0      active sync   /dev/sdc
       1       8       48        1      active sync   /dev/sdd
       3       8       64        2      active sync   /dev/sde
       4       8        0        3      active sync   /dev/sda
[root@isengard volumes]#
Actually was going to make that a raid 10. I had made it a raid 5 for now while I wait for a DOA drive to be RMAed but figured I'd just grow the raid 5 instead.

This could take a while.

Code:
[root@isengard volumes]# cat /proc/mdstat 
Personalities : [raid6] [raid5] [raid4] 
md0 : active raid5 sda[4] sde[3] sdd[1] sdc[0]
      5860270080 blocks super 1.2 level 5, 4096k chunk, algorithm 2 [4/4] [UUUU]
      [>....................]  reshape =  0.4% (13680392/2930135040) finish=802.0min speed=60606K/sec
      
unused devices: <none>
[root@isengard volumes]#
I don't actually have any plans for that raid array, it's empty. So I might change my mind and do 10. :p I will be taking all the drives from my existing server and putting them in this server, so this array is just spare space. Might use for extra backups or something. May also mirror Linux repos, because why not.

Goal is to setup Kerberos and LDAP and do a proper NFS setup.


even after my full explanation of how to not double post and you still do it? :|


I don't even know why I try... :\