«608 »
  • Post
  • Reply
Megaman
May 8, 2004
I didn't read the thread BUT...

kiwid posted:

Bigger drives, more drives, and going from raidz1 to raidz2. As far as I know, you still can't expand a vdev with more drives, right?

Nope, not as far as I know

titaniumone
Jun 10, 2001



Combat Pretzel posted:

The Solaris one has 600000+ load cycles, the Drobo only 4000+. I don't get it?

"only"

That's still a horrendous load cycle count.

eames
May 9, 2009



Combat Pretzel posted:

The Solaris one has 600000+ load cycles, the Drobo only 4000+. I don't get it?

that’s because I’m dumb and got the screenshots mixed up. its the other way around. edited...

Bob Morales
Aug 18, 2006

I love the succulent taste of cop boots

BackBlaze has a new blog post up about enterprise vs consumer drives

http://blog.backblaze.com/2013/12/0...ve-reliability/

quote:

So, Are Enterprise Drives Worth The Cost?
From a pure reliability perspective, the data we have says the answer is clear: No.

Enterprise drives do have one advantage: longer warranties. That’s a benefit only if the higher price you pay for the longer warranty is less than what you expect to spend on replacing the drive.

This leads to an obvious conclusion: If you’re OK with buying the replacements yourself after the warranty is up, then buy the cheaper consumer drives.

kiwid
Sep 30, 2013



so I just happened to look at that SMART values on my other NAS with older drives. I haven't been getting any SMART errors or warnings or anything (tests are scheduled weekly) but one of the drives looks like this:

code:
/dev/ada2
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x000b   096   096   016    Pre-fail  Always       -       458753
  2 Throughput_Performance  0x0005   131   131   054    Pre-fail  Offline      -       99
  3 Spin_Up_Time            0x0007   132   132   024    Pre-fail  Always       -       434 (Average 434)
  4 Start_Stop_Count        0x0012   100   100   000    Old_age   Always       -       31
  5 Reallocated_Sector_Ct   0x0033   100   100   005    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x000b   100   100   067    Pre-fail  Always       -       0
  8 Seek_Time_Performance   0x0005   135   135   020    Pre-fail  Offline      -       26
  9 Power_On_Hours          0x0012   098   098   000    Old_age   Always       -       14438
 10 Spin_Retry_Count        0x0013   100   100   060    Pre-fail  Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       31
192 Power-Off_Retract_Count 0x0032   100   100   000    Old_age   Always       -       588
193 Load_Cycle_Count        0x0012   100   100   000    Old_age   Always       -       588
194 Temperature_Celsius     0x0002   171   171   000    Old_age   Always       -       35 (Min/Max 22/41)
196 Reallocated_Event_Count 0x0032   100   100   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0022   100   100   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0008   100   100   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x000a   200   200   000    Old_age   Always       -       0
Is that Raw_Read_Error_Rate something I need to be concerned about? All the other drives have a value of 0.

Gwaihir
Dec 8, 2009



Hair Elf

I don't think so, IIRC the one you really want to look at is Reallocated sector count/current pending sector count. Those represent actual bad things developing and indicating impending general failures.

IOwnCalculus
Apr 2, 2003





I wouldn't sweat it yet, you still have no reallocated sectors and no uncorrectable errors.

I'm also glad I got my latest shipment of 3TB Reds when I did; I was playing musical drives to try and get my three lowest-hour 1.5TB drives in the same raidz1 vdev, when my single highest-hour one (with over 37.7k hours on it!) decided to shit the bed and use up what was my last remaining 1.5TB spare. I have to check now but I think the three Samsung 1.5s that died on me were all in a very short string of nearly-sequential serial numbers I expect to find that it too has died due to a head crash.

For the record, here's how Newegg shipped mine - the third drive easily had enough room to rock from angled to sitting against the end of the box.

IOwnCalculus fucked around with this message at 21:00 on Dec 5, 2013

Combat Pretzel
Jun 23, 2004

No, seriously... what kurds?!

titaniumone posted:

"only"

That's still a horrendous load cycle count.
Huh, what? 4000 cycles in over 8000 power on hours being horrendous? When it's rated 300000?! By that rate it'll get only to 13-15% of the rated count, if at all, before going tits up. (--edit: Where tits up would be around 80000 hours, which amounts to 9 years of power on time.)

Gwaihir
Dec 8, 2009



Hair Elf

IOwnCalculus posted:

I wouldn't sweat it yet, you still have no reallocated sectors and no uncorrectable errors.

I'm also glad I got my latest shipment of 3TB Reds when I did; I was playing musical drives to try and get my three lowest-hour 1.5TB drives in the same raidz1 vdev, when my single highest-hour one (with over 37.7k hours on it!) decided to shit the bed and use up what was my last remaining 1.5TB spare. I have to check now but I think the three Samsung 1.5s that died on me were all in a very short string of nearly-sequential serial numbers

For the record, here's how Newegg shipped mine - the third drive easily had enough room to rock from angled to sitting against the end of the box.


Regardless of the lack of actually securing the drives, I really do like those bubble case things. They're so much less annoying than the old layers and layers of regular bubble wrap taped together every which way.

IOwnCalculus
Apr 2, 2003





Completely agreed, I have at least some hope (and if not, well, there's always WD's RMA process).

kiwid
Sep 30, 2013



IOwnCalculus posted:


For the record, here's how Newegg shipped mine - the third drive easily had enough room to rock from angled to sitting against the end of the box.


I really did luck out then. That sucks.

Psimitry
Jun 3, 2003

Hostile negotiations since 1978

IOwnCalculus posted:

For the record, here's how Newegg shipped mine

Yup. That's how mine looked too.

SO. I think I'm going to give NAS4Free a try. It's a little unnerving to try to learn something completely from scratch, but I am hoping I can make it work. (that, and apparently I need to secure an extra 4GB of DDR3 memory if I'm going to run it on that AMD E-350, or DDR2 if I'm going to run it on my spare Athlon X2 platform)

Not really sure, however, what the OS actually installs to. It looks like it runs off of a USB drive. If so, that means I get to return the 32GB SSD I bought off of Amazon to run Windows on, but whatever.

As far as stress testing the array once I've got it built...how does one go about it?

IOwnCalculus
Apr 2, 2003





You can run it off of the SSD, and there are some theoretical benefits to doing so (you can give it some swap space, for example). However, when I looked into it most of the documentation strongly recommended against that type of installation due to increased difficulty of updates. You could still do the USB-type installation (no swap, the system loads from it once to boot and never touches it again except to update the config.xml as needed) on a non-USB storage device, but typically you can save some money and free up a SATA port by just using whatever cheap USB stick you can get your hands on.

Or, if you aren't concerned about the cost, you could use a USB stick to boot from, and add the SSD in as a cache device.

Stress testing? Copy as much data as you can to it and read it back. Write as many copies of the Ubuntu install ISO to it as you can, and run zpool scrub over and over again so that ZFS is having to constantly validate the data.

It's been my experience with NAS4Free that if you're doing any sort of heavy workload like this and you're going to see a problem, the problem drive is going to start chucking errors left and right into the system log; otherwise it will quietly do what you ask of it.

Psimitry
Jun 3, 2003

Hostile negotiations since 1978

So then the best option (especially considering my status as a total novice), is just to run it off a USB stick?

Edit: moved into its own post.

Psimitry fucked around with this message at 22:18 on Dec 5, 2013

IOwnCalculus
Apr 2, 2003





If I were starting from scratch, yeah, that's what I'd do. Either use the SSD for cache or return it and save the money / the SATA port.

Psimitry
Jun 3, 2003

Hostile negotiations since 1978

As far as the testing, is there an automated way to do this?

I've heard something about using HDtune as an automated tool to run it for however long you specify, writing zeroes, reading the entire drive, writing ones, reading, repeat.

Obviously that's not going to spot random read errors, but again I am very new to all of this.

Ninja Rope
Oct 22, 2005

Wee.


I dd if=/dev/zero of=/dev/thedrive bs=32M count=100 or 1000 or so, then do SMART short offline test, and the next day do a long offline test. If it survives that I use it.

Edit: Actually, I think I started filling the drive (leave off the count=blah part of dd) then doing the rest.

Ninja Rope fucked around with this message at 01:50 on Dec 6, 2013

McGlockenshire
Dec 16, 2005


Bob Morales posted:

BackBlaze has a new blog post up about enterprise vs consumer drives

http://blog.backblaze.com/2013/12/0...ve-reliability/

Be careful with those conclusions -- they haven't had the enterprise drives for more than two years, so they don't have any data on third year failures.

Even if they do have the same failure rate, that's where the five-year warranty on most enterprise drives would come into play in the bang/buck calculations. There are also other warranty considerations, including the manufacturers liking to not honor warranties on consumer drives used in RAIDs.

Psimitry
Jun 3, 2003

Hostile negotiations since 1978

Ninja Rope posted:

I dd if=/dev/zero of=/dev/thedrive bs=32M count=100 or 1000 or so, then do SMART short offline test, and the next day do a long offline test. If it survives that I use it.

Edit: Actually, I think I started filling the drive (leave off the count=blah part of dd) then doing the rest.

I don't really have any idea what that means - is this done through a command line from NAS4Free or some other Linux variant?

Ninja Rope
Oct 22, 2005

Wee.


Yes, sorry. That command fills up the drive with zeroes, touching (ideally) every sector. dd is kind of complicated and I can't find a tutorial right now, but in order to use dd you'll need to know the raw device for the drive you want to test. I don't know if FreeNAS/NAS4Free shows you in the UI but it should be something like "/dev/ada0" or just "ada0". You'll also see them listed in the "dmseg" command (along with a ton of other data) or, if you've created the zpool already, "zpool status" may list the drives as well (but without the /dev/ part).

If the drive actually is /dev/ada0, the dd command to fill the drive with zeroes would be:

code:
dd if=/dev/zero of=/dev/ada0 bs=32M
If you get the drive part wrong you'll fill the wrong drive with zeroes and wipe out whatever data is on there, so be careful.

The command will take forever and not show any progress and probably not report any errors, but checking the SMART data afterwards (again I don't know if there's a FreeNAS/NAS4Free UI for this or if you have to use smartmontools) or forcing a SMART offline test would report any errors dd exposed.

Prince John
Jun 20, 2006

Oh, poppycock! Female bandits?



Ninja Rope posted:

The command will take forever and not show any progress and probably not report any errors, but checking the SMART data afterwards (again I don't know if there's a FreeNAS/NAS4Free UI for this or if you have to use smartmontools) or forcing a SMART offline test would report any errors dd exposed.

Gnu ddrescue will let you monitor progress and stop and resume which can be useful for large drives.

The badblocks program also does something similar but has a non destructive option if you don't want to write to the drive. It can also show progress and be resumed.

evol262
Nov 30, 2010
#!/usr/bin/perl

Prince John posted:

Gnu ddrescue will let you monitor progress and stop and resume which can be useful for large drives.

The badblocks program also does something similar but has a non destructive option if you don't want to write to the drive. It can also show progress and be resumed.

kill -USR1 will also show progress from dd.

Combat Pretzel
Jun 23, 2004

No, seriously... what kurds?!

I'm glad for the student-run computer shops in the nearby city, which houses a fairly large tech-oriented university. They get their drives delivered in well-padded manufacturer boxes of 20 each and sell them over the counter. None of that shit packing that's killing drives. And they're competitive with online shops in their pricing, too.

Gwaihir
Dec 8, 2009



Hair Elf

McGlockenshire posted:

Be careful with those conclusions -- they haven't had the enterprise drives for more than two years, so they don't have any data on third year failures.

Even if they do have the same failure rate, that's where the five-year warranty on most enterprise drives would come into play in the bang/buck calculations. There are also other warranty considerations, including the manufacturers liking to not honor warranties on consumer drives used in RAIDs.

I've never had Seagate or WD give me flack about RMAs on drives used in RAID arrays, even Greens. I guess I've just been lucky so far

In more relevant to the thread posting, I'm surprisingly happy with this LSI 9211-8i. Rebuilding the 3 * 3tb set in to a 4 * 3tb set is chugging along at 40 MB/sec, even while playing a movie off of it. ~20 hours rebuild time was a lot less than I was expecting. I had an array of 3 * 3tb 7200 RPM drives + 5 * 2tb green drives, just running mdadm raid-5s. Ditched all the greens, because screw that. I ended up picking up 4 3tb seagate SAS drives because a buddy was selling them new/unopened for cheap (100$ for Constellation ESes was too good to pass up for feeding my storage addiction). So I'm going to end up with an array of 4 * 3tb sata, 4 * 3tb SAS (Both raid5), and then the last 4 3tb SATA get shoved in to my desktop, probably in a RAID-10 array for FRAPS (Really dxtory and/or shadowplay, but fraps just sorta became synonymous with "Recording games" to most people) recording duties.

I feel like Agreed with his videocard addiction, except with storage at home

Daylen Drazzi
Mar 10, 2007

Why do I root for Notre Dame? Because I like pain, and disappointment, and anguish. Notre Dame Football has destroyed more dreams than the Irish Potato Famine, and that is the kind of suffering I can get behind.

Yay! My FreeNAS server is up and running, although I was apparently not comprehending correctly when I installed 4x2TB drives in a RAIDZ2 configuration. I was expecting to have 6TB of usable space, but only have 4TB. I still could have arranged it in a less-optimal configuration to get the 6TB, but I decided the massive redundancy was the better way to go (really don't want to lose any data). I've also got a 4TB external drive coming that I can use as a backup solution, so I'm happy that I don't have to build another server to do backup. Of course, when I do feel the need to expand my storage space I'm going to have to figure out how I want to go about doing it - at present my Frankenstein machine has a paltry 4 SATA ports, so it's maxed out in the number of drives I can shove into it. Friend says I'm getting too far ahead of myself, as 4TB will be more than enough space for a long time to come for me. Even so, I might start pulling together a parts list to make a proper FreeNAS server, so that if/when the money comes available I can do it right.

In case anyone is interested I'm using a Core2Duo E8400 CPU, Gigabyte GA-EP35-DS3L Motherboard, 8GB of DDR2 800 (PC2 6400) RAM, an AMS 4-in-3 Trayless backplane, and a 4GB USB flash drive to run FreeNAS, with 4x2TB Seagate NAS drives. The drives were, without a doubt, the most expensive part of the server and made up about 80% of the cost.

Combat Pretzel
Jun 23, 2004

No, seriously... what kurds?!

The number behind the Z in RAIDZ defines the amount of parity disks. Two disks are used on parity, so that's why 4TB instead of 6TB.

Daylen Drazzi
Mar 10, 2007

Why do I root for Notre Dame? Because I like pain, and disappointment, and anguish. Notre Dame Football has destroyed more dreams than the Irish Potato Famine, and that is the kind of suffering I can get behind.

I got that after reading the user's guide half a dozen times, searching the web, and trying to figure out what the hell was going on. Frustration is a great teacher.

kiwid
Sep 30, 2013



If you have good backups, you might almost be better to do a RAID10 at 4 drives (stripe two nested raid1 groups) and have the same usable space but increased performance. You can technically lose two drives but there is a chance you can only lose one.

edit: but for home use, you're probably better sticking with the raidz1-3

kiwid fucked around with this message at 17:20 on Dec 6, 2013

IOwnCalculus
Apr 2, 2003





Ninja Rope posted:

The command will take forever and not show any progress and probably not report any errors, but checking the SMART data afterwards (again I don't know if there's a FreeNAS/NAS4Free UI for this or if you have to use smartmontools) or forcing a SMART offline test would report any errors dd exposed.

There's three ways you would see errors pop up on this. One is in NAS4Free's SMART screen at Diagnostics -> Information -> SMART, where each drive will have an entry like this.

code:
=== START OF INFORMATION SECTION ===
Model Family:     SAMSUNG SpinPoint F2 EG
Device Model:     SAMSUNG HD154UI
Serial Number:    S1Y6J90S704649
LU WWN Device Id: 5 0024e9 200cf7ad3
Firmware Version: 1AG01118
User Capacity:    1,500,300,828,160 bytes [1.50 TB]
Sector Size:      512 bytes logical/physical
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ATA/ATAPI-7, ATA8-ACS T13/1699-D revision 3b
Local Time is:    Fri Dec  6 08:58:13 2013 MST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x00)	Offline data collection activity
					was never started.
					Auto Offline Data Collection: Disabled.
Self-test execution status:      (   0)	The previous self-test routine completed
					without error or no self-test has ever
					been run.
Total time to complete Offline
data collection: 		(18346) seconds.
Offline data collection
capabilities: 			 (0x7b) SMART execute Offline immediate.
					Auto Offline data collection on/off support.
					Suspend Offline collection upon new
					command.
					Offline surface scan supported.
					Self-test supported.
					Conveyance Self-test supported.
					Selective Self-test supported.
SMART capabilities:            (0x0003)	Saves SMART data before entering
					power-saving mode.
					Supports SMART auto save timer.
Error logging capability:        (0x01)	Error logging supported.
					General Purpose Logging supported.
Short self-test routine
recommended polling time: 	 (   2) minutes.
Extended self-test routine
recommended polling time: 	 ( 307) minutes.
Conveyance self-test routine
recommended polling time: 	 (  32) minutes.
SCT capabilities: 	       (0x003f)	SCT Status supported.
					SCT Error Recovery Control supported.
					SCT Feature Control supported.
					SCT Data Table supported.

SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x000f   100   100   051    Pre-fail  Always       -       5
  3 Spin_Up_Time            0x0007   072   072   011    Pre-fail  Always       -       9310
  4 Start_Stop_Count        0x0032   100   100   000    Old_age   Always       -       358
  5 Reallocated_Sector_Ct   0x0033   100   100   010    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x000f   100   100   051    Pre-fail  Always       -       0
  8 Seek_Time_Performance   0x0025   100   100   015    Pre-fail  Offline      -       11922
  9 Power_On_Hours          0x0032   093   093   000    Old_age   Always       -       34600
 10 Spin_Retry_Count        0x0033   100   100   051    Pre-fail  Always       -       0
 11 Calibration_Retry_Count 0x0012   100   100   000    Old_age   Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       169
 13 Read_Soft_Error_Rate    0x000e   100   100   000    Old_age   Always       -       5
183 Runtime_Bad_Block       0x0032   100   100   000    Old_age   Always       -       0
184 End-to-End_Error        0x0033   100   100   000    Pre-fail  Always       -       0
187 Reported_Uncorrect      0x0032   100   100   000    Old_age   Always       -       8
188 Command_Timeout         0x0032   100   100   000    Old_age   Always       -       0
190 Airflow_Temperature_Cel 0x0022   085   023   000    Old_age   Always       -       15 (Min/Max 15/17)
194 Temperature_Celsius     0x0022   084   022   000    Old_age   Always       -       16 (82 240 18 15 0)
195 Hardware_ECC_Recovered  0x001a   100   100   000    Old_age   Always       -       194066559
196 Reallocated_Event_Count 0x0032   100   100   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0012   100   100   000    Old_age   Always       -       3
198 Offline_Uncorrectable   0x0030   100   100   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x003e   099   094   000    Old_age   Always       -       772
200 Multi_Zone_Error_Rate   0x000a   100   100   000    Old_age   Always       -       0
201 Soft_Read_Error_Rate    0x000a   099   099   000    Old_age   Always       -       1
Interestingly, it looks like this drive is on its way out soon. I'm not seeing any errors on this drive in the next two places so it hasn't died yet, but it is showing at least two points that it is starting to fail - attributes 187 and 197 should both be 0 on a healthy drive. This one has 8 uncorrectable errors on it and has 3 sectors that are waiting to be reallocated. Attribute 5 is the biggest one, since once the drive starts to actually reallocate sectors it's right about at the end. This drive is also very old and is probably failing in the exact same way the rest of my 4+ year old HD154UIs have failed - a head crash.

Second place is (if you use ZFS) in the ZFS screen at Disks -> ZFS -> Information. You'll see something like this:

code:
pool: aggregate
 state: ONLINE
status: One or more devices is currently being resilvered.  The pool will
	continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
  scan: resilver in progress since Fri Dec  6 06:20:42 2013
        8.57T scanned out of 11.7T at 915M/s, 0h59m to go
        407G resilvered, 73.28% done
config:

	NAME             STATE     READ WRITE CKSUM
	aggregate        ONLINE       0     0     0
	  raidz1-0       ONLINE       0     0     0
	    da3          ONLINE       0     0     0
	    da5          ONLINE       0     0     0
	    da1          ONLINE       0     0     0
	  raidz1-1       ONLINE       0     0     0
	    ada2         ONLINE       0     0     0
	    ada1         ONLINE       0     0     0
	    ada3         ONLINE       0     0     0
	  raidz1-2       ONLINE       0     0     0
	    ada4         ONLINE       0     0     0
	    replacing-1  ONLINE       0     0     0
	      da4        ONLINE       0     0     0
	      da7        ONLINE       0     0     0  (resilvering)
	    da2          ONLINE       0     0     0
	cache
	  ada0           ONLINE       0     0     0

errors: No known data errors
Notice that da1 is not showing any checksum errors here yet, so either ZFS hasn't had problems reading those pending failed blocks, or I'm particularly lucky and the failed blocks don't actually contain any data. You'll see the checksum column start climbing when ZFS has problems actually reading from the disk. The resilver in progress is me going through a drive replacement - that's the second WD Red 3TB going in to replace a 1.5TB drive.

Third place you would see errors is going to be the system log at Diagnostics -> Log. My log doesn't have any errors in it since I rebooted it after the last drive failure, but when a drive decides to really shit the bed, you'll see a bunch of lines showing up repeatedly at the end of the log referencing being unable to read that particular drive. It's at this point that ZFS will typically drop the drive from the array altogether.

Really, you should set it up, start writing some data to it, and configure email monitoring such that it emails you if ZFS ever shows a status other than "ONLINE". There was a guide somewhere on how to do it but I can't seem to find it. I stuck the two scripts on pastebin: http://pastebin.com/j6VZKDXC and http://pastebin.com/HKM0CevX. I put these on a folder inside my ZFS array since I didn't want to dick around with unionfs and just have a cronjob set to run zfs_errors.sh every 30 minutes. I also have another cronjob set to run 'zpool scrub yourpoolnamehere' on a weekly basis.

ShaneB
Oct 22, 2002



Is there a way to do an initial drive test on a Synology box?

Combat Pretzel
Jun 23, 2004

No, seriously... what kurds?!

kiwid posted:

If you have good backups, you might almost be better to do a RAID10 at 4 drives (stripe two nested raid1 groups) and have the same usable space but increased performance. You can technically lose two drives but there is a chance you can only lose one.

edit: but for home use, you're probably better sticking with the raidz1-3
Depends on the scenario. Some home server to stream a bunch of movies from, zprefetch will take care of performance (back in the day I used it, it prefetched movies in 150-200MB chunks). Anything causing random IO like fuck, RAID-Z is the worst choice, due to its full stripe design, because to read a filesystem block, it has to touch ALL disks (minus parity).

Pudgygiant
Apr 8, 2004

Garnet and black? More like gold and blue or whatever the fuck colors these are

Well fuck, it looks like a hardware issue with this ReadyNAS, and of course it's out of warranty. Seeing as how it's btrfs based, can I just drop these drives into something else and mdadm it back to life? Or am I most likely going to lose everything?

deimos
Nov 30, 2006

Forget it man this bat is whack, it's got poobrain!


Pudgygiant posted:

Well fuck, it looks like a hardware issue with this ReadyNAS, and of course it's out of warranty. Seeing as how it's btrfs based, can I just drop these drives into something else and mdadm it back to life? Or am I most likely going to lose everything?

As with everything, it depends on how much ReadyNAS does or doesn't do. It might actually be as simple as mounting a single drive via btrfs, it would pick up the rest of the drives and give you access to the cluster.

Pudgygiant
Apr 8, 2004

Garnet and black? More like gold and blue or whatever the fuck colors these are

They're all assembled into a raid mountpoint (md127) which is then mounted to /data. I guess we'll see. I'm close to pulling the trigger on all this (and Freenas on a thumb drive), anything I should change?

code:
LIAN LI PC-Q25B
$99.99
Antec EarthWatts Green EA-380D Green 380W
$44.99
G.SKILL Ripjaws Series 8GB (2 x 4GB) 240-Pin DDR3 SDRAM DDR3 1600
$49.99
ASRock B85M-ITX
$74.99
Intel Core i3-4130T
$139.99

D. Ebdrup
Mar 13, 2009



Combat Pretzel posted:

The number behind the Z in RAIDZ defines the amount of parity disks. Two disks are used on parity, so that's why 4TB instead of 6TB.
This is bugging me a bit, so just to clarify: those numbers aren't indicators of the amount of parity disks - rather, it's an indicator of how many disks you can lose and still be able to rebuild/resilver because of parity stored on each drive in the vdev/zpool.

EDIT: here is a look into how zpools do parity across disks from my bookmarks.

D. Ebdrup fucked around with this message at 17:58 on Dec 7, 2013

dotster
Aug 28, 2013



D. Ebdrup posted:

This is bugging me a bit, so just to clarify: those numbers aren't indicators of the amount of parity disks - rather, it's an indicator of how many disks you can lose and still be able to rebuild/resilver because of parity stored on each drive in the array/pool.

EDIT: here is a look into how zpools do parity across disks from my bookmarks.

Well and the same is true for parity with RAID5/6 as well, nothing uses a dedicated disk for parity (unless you are doing RAID 4 somehow). RAID 5 is roughly equivalent to RAIDZ in that they can tolerate a single disk loss and RAID 6 and RAIDZ2 can tolerate two disk failures in a pool. The parity information is striped across all the disks in the pool in all cases listed. Many people say RAID5/RAIDZ has one parity disk and RAID6/RAIDZ2 has two parity disks because that is the amount of space you lost to parity and they were taught wrong or at least a simplified version of what was really going on.

brainwrinkle
Oct 18, 2009

What's going on in here?


Buglord

Are there any decent cases for ITX/mATX custom NAS boxes? I've got an IBM ServeRAID M1015 controller laying around, so I figure I could go with an ITX board (or mATX failing that). I probably only need 4 drives, so something as small and slick as a Synology box would be nice. I've read the past couple of pages, but I haven't seen any recent posts on hardware choices.

kiwid
Sep 30, 2013



Did I set this up correctly in that it is a RAID10 setup?



I have a feeling it's just two groups of RAID1 and no striping.

kiwid fucked around with this message at 01:02 on Dec 8, 2013

Combat Pretzel
Jun 23, 2004

No, seriously... what kurds?!

kiwid posted:

I have a feeling it's just two groups of RAID1 and no striping.
ZFS load balances data between different vdevs based on IO load, among other things. Data's being spread, just not in an interleaved fashion you're expecting.

Not sure why you need an intent log, tho. Are you writing out that much data?

D. Ebdrup posted:

This is bugging me a bit, so just to clarify: those numbers aren't indicators of the amount of parity disks - rather, it's an indicator of how many disks you can lose and still be able to rebuild/resilver because of parity stored on each drive in the vdev/zpool.
Same difference to me. RAIDZ loses 1 disk worth of space to parity, RAIDZ2 loses 2 disks worth and RAIDZ3 loses 3 disks worth.

Combat Pretzel fucked around with this message at 01:15 on Dec 8, 2013

kiwid
Sep 30, 2013



Combat Pretzel posted:

ZFS load balances data between different vdevs based on IO load, among other things. Data's being spread, just not in an interleaved fashion you're expecting.

Not sure why you need an intent log, tho. Are you writing out that much data?


Thanks, I thought I set it up in some weird JBOD of mirrored groups or something, didn't know it load balances across all vdevs.

As for the intent log. This isn't my main NAS, this is just a thing I'm fucking with for a home lab with vmware.

  • 1
  • 2
  • 3
  • 4
  • 5
  • Post
  • Reply
«608 »