Register a SA Forums Account here!
JOINING THE SA FORUMS WILL REMOVE THIS BIG AD, THE ANNOYING UNDERLINED ADS, AND STUPID INTERSTITIAL ADS!!!

You can: log in, read the tech support FAQ, or request your lost password. This dumb message (and those ads) will appear on every screen until you register! Get rid of this crap by registering your own SA Forums Account and joining roughly 150,000 Goons, for the one-time price of $9.95! We charge money because it costs us money per month for bills, and since we don't believe in showing ads to our users, we try to make the money back through forum registrations.
 
  • Post
  • Reply
szlevi
Sep 10, 2010

[[ POKE 65535,0 ]]

Vanilla posted:

I recall you mentioning isilon at some point in the past. Is it the cost that rules them out or a technical reason?

Isilon never been cheap but now that it's carrying the EMC logo it would be even more expensive, not to mention the exodus of old hands from Isilon, the rash of issues in the the past year or so I heard of (from costumers and ex-eomployees) etc etc.

Adbot
ADBOT LOVES YOU

szlevi
Sep 10, 2010

[[ POKE 65535,0 ]]

NippleFloss posted:

I'd be absolutely shocked if they can get their stated 12GB/s out of only 24 SSD drives. Assuming the data is protected at all you're losing at least a couple of those to parity drives, which means each SSD is doing more than 500 MB/s. You might get that out of a consumer quality SSD, but enterprise class SSDs only hit around 300 to 350 MB/s. There just aren't enough disks to get you to those throughput numbers with reliable hardware. It's possible that they are using consumer grade drives, but that would worry me, especially with a new vendor with no real track record. They make a lot of claims that border on magic, like 10 years of 1.2PB writes a weak without any performance loss, which just doesn't fit with the characteristics of SSD drives as they exist right now, especially not consumer grade drives. Could be true, but will they be around in 10 years to verify?


I'd say they are probably using some sort of 8+2 raid sets, plus one or two hot spare... that gives you 2x10 drives so yes, 500MB/s - that's not unheard of and again, they claim that by having their own custom-designed (eg I know they have a small amount of memory built into every drive), custom manufactured SSDs they can gain enough extra bandwidth to get there (10Gb/s) with raid etc vs commercially available drives... who knows, could be true.

quote:

I know a E5460 box from NetApp can do about 3GB/s (this may be higher, theoretically, this is just the number I've seen when sizing for Lustre with a certain block size and stream count) in a 4u enclosure that includes redundant controllers and 60 7.2k 2/3TB NL-SAS drives. That'll get you around 80TB, give or take, with raid-6 and 2TB drives. I've got no idea on price though, since, as I said, I don't support these at all. Could be cheap of very expensive. It's probably less than the $8/GB raw that Nimbus gear lists at, but whether you need the extra capacity is another matter.

I have to be able to feed 7-8 10gig-enabled people, that's ~8GB/s and I have no overhead left... with disks it'd be ridiculously more expensive, not to mention power, maintenance etc. We are also a Windows shop although I'm open for a proper (redundancy/HA, transparent Windows security support etc etc) non-Windows solution...
Capacity does not matter, around 2-3TB is fine, 5-6TB would be downright future-proof, it's only for this purpose and I'd do my backend scripting/linking kungfu to hide it in our DFS hierarchy.

quote:

Anyway, my point wasn't that all flash arrays are bad, I'm just trying to understand who is using them and why. If your requirements are for a very high throughput low capacity solution in a small footprint then it might be the right move for you. SSD throughput is only about double spinning drive throughput, at best, but that might be enough difference to get to your magic number.

I share your skepticism hence my request for a demo unit...

quote:

My guess is that Isilon would way too expensive and require too much gear to the to the 5 GB/s number he mentioned. I'd guess he'd be looking at 10 nodes, at minimum, to get to that number.

Exactly.

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

szlevi posted:

...stuff...

Cool, thanks for giving me the background. I'm genuinely curious to see how this stuff is getting positioned. Obviously long term I think spinning platters go away and the vast majority of storage is flash based, but it seems like the big storage players are taking a wait-and-see approach to all flash arrays. I'm just wondering if it's truly a niche market, and will be for a while, or if they're going to get caught off guard with a massive technology shift.

Mierdaan
Sep 14, 2004

Pillbug

szlevi posted:

Just upgraded my NAS boxes (Dell NX3000s) to Server 2012, I'll test SMB3.0 with direct FIO shares again - I'm sure it's got better but I doubt it's got that much better...

Post back here when you do this please, I'm interested to see what they did with SMB3.0.

evil_bunnY
Apr 2, 2003

What's the budget, szlevi? Is an all-SSD Equallogic too pricey?

Nomex
Jul 17, 2002

Flame retarded.

szlevi posted:

Yeah, that 'moving blocks'/tiering approach never worked, never will for this type of thing, I can tell you that already. :)
As for being sequential - it's only per file but you have several users opening and closing plenty of different files etc so it's far from the usual video editing load.
Size can vary from few gigabytes to 50-100GB per file (think of 4k and higher, RGBA fp32 etc) - we've developed our own raw format, wrote plugins for Fusion, Max etc so we're actually pretty flexible if it comes down to that...

FWIW I have 2 Fuison-IO Duo cards, they were very fast when I bought them for $15k apiece, now they are just fast but the issue from day 1 is Windows CIFS: up to 2008 R2 (SMB2.1) CIFS is an utter piece of poo poo, it simply chops up everything into 1-3k size pieces so it pretty much destroys any chance of taking advantage of its bandwidth.
Just upgraded my NAS boxes (Dell NX3000s) to Server 2012, I'll test SMB3.0 with direct FIO shares again - I'm sure it's got better but I doubt it's got that much better...

Since going with an established big name would be very expensive (10GB/s!) as I see I have to choose between two approaches:
1. building my own Server 2012-based boxes eg plugging in 3-4 2GB/s or faster PCIe storage cards, most likely running the shebang as a file sharing cluster (2012 got a new active-active scale-out mode), hoping SMB3.0 got massively better
2. going with some new solution, coming from a new, small firm, hoping they will be around or bought up - and only, of course, after acquiring a demo unit to see real performance

I can also wait until Dell etc buys up a company/rolls out something new but who knows when they will have affordable 10GB/s...?

Sorry, I shouldn't have said move in relation to blocks. Flash pool is a caching system. It doesn't do any tiering. Reads and overwrites are cached, but the flash pool is consistent with the disks.

Just out of curiosity, why are you using CIFS? Why not mount a LUN instead? You can slap a dual 8 gig FC HBA in and pull way way higher throughput than using CIFS. How many clients are running at a time and what kind of budget do you have for this?

Amandyke
Nov 27, 2004

A wha?

szlevi posted:

Isilon never been cheap but now that it's carrying the EMC logo it would be even more expensive, not to mention the exodus of old hands from Isilon, the rash of issues in the the past year or so I heard of (from costumers and ex-eomployees) etc etc.

Rash of issues would be news to me. I know of several customers throwing serious capital at huge isilon clusters. In the 8 figure range.

Vulture Culture
Jul 14, 2003

I was never enjoying it. I only eat it for the nutrients.

Amandyke posted:

Rash of issues would be news to me. I know of several customers throwing serious capital at huge isilon clusters. In the 8 figure range.
We've actually had much better experiences in the past year or so than the several years prior, on top of that. Things were definitely rocky in the first few months of the acquisition, because EMC dumped a ton of money into scaling up operations, so they had sales guys signing contracts on these things before they were even produced, and they had to hire so many brand-new product engineers and support guys that 90% of the organization was brand new and nobody had any idea what was going on. As far as I can tell, that's basically over. I can't remember the last time we even heard a peep out of our 108NL cluster, and the last few issues we did have were resolved really quickly.

Xenomorph
Jun 13, 2001
Is anyone here hosting Windows 8 roaming profiles on a Samba-based share?
Are there any known issues that keeps this from working?

psydude
Apr 1, 2008

Where the hell does NetApp keep their MIBs? I tried downloading them from the website and it was about as confusing, if not moreso, than trying to download a driver from Dell. With Cisco I was able to just download one MIB for every device they have and never have to worry about it ever again.

oogs
Dec 6, 2011
szlevi, have you looked at Nimble? We use them in our mailstore environment, and it consistently outperforms our compellent. Additionally, the pricing is amazing - no "extras", it's all included. At the rate we're going, we'll have 10 in production by summer, and we're hungry for more.

Rhymenoserous
May 23, 2008
Another nimble buddy! Hey :buddy:!

EDIT: They have expansion cabinets now, and the pricing looks good.

oogs
Dec 6, 2011

Rhymenoserous posted:

Another nimble buddy! Hey :buddy:!

EDIT: They have expansion cabinets now, and the pricing looks good.

Hey!

Yes, they do! We're getting some for our main units. We're drooling at this stuff. We have separate head units depending on their purpose, project, etc. You know how that goes...

oogs fucked around with this message at 20:48 on Dec 3, 2012

Intrepid00
Nov 10, 2003

I'm tired of the PM’s asking if I actually poisoned kittens, instead look at these boobies.
HP's Lefthand boxes have version 10 released :toot: Finally, a patch to the VSS provider I've been waiting for.

adorai
Nov 2, 2002

10/27/04 Never forget
Grimey Drawer
can someone ballpark what 15-20TB of nimble storage would cost me, per HA cluster? My only requirements are 10GBe, NFS or iSCSI, deduplication, and mirroring.

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

psydude posted:

Where the hell does NetApp keep their MIBs? I tried downloading them from the website and it was about as confusing, if not moreso, than trying to download a driver from Dell. With Cisco I was able to just download one MIB for every device they have and never have to worry about it ever again.

MIBs are stored on the appliance in the /etc/mib directory. Just connect via CIFS or NFS and pull them off of the device.

oogs posted:

szlevi, have you looked at Nimble? We use them in our mailstore environment, and it consistently outperforms our compellent. Additionally, the pricing is amazing - no "extras", it's all included. At the rate we're going, we'll have 10 in production by summer, and we're hungry for more.

Nimble won't do the throughput he's looking for at the price/density he wants. Nimble's caching approach is great for random I/O where the working set is a small portion of the total data set (like e-mail and some OLTP). It's not very good for high throughput applications since the density per controller is pretty low and the your SSD cache layer does basically nothing and you're limited to the aggregated throughput of your SATA drives.

YOLOsubmarine fucked around with this message at 01:21 on Dec 4, 2012

madsushi
Apr 19, 2009

Baller.
#essereFerrari

NippleFloss posted:

Nimble won't do the throughput he's looking for at the price/density he wants. Nimble's caching approach is great for random I/O where the working set is a small portion of the total data set (like e-mail and some OLTP). It's not very good for high throughput applications since the density per controller is pretty low and the your SSD cache layer does basically nothing and you're limited to the aggregated throughput of your SATA drives.

At the end of the day, Nimble is really just a shelf of SATA disk. It's got your normal NVRAM for write-caching and SSD for read-caching, but your consistent writes are limited to your SATA disks. If your reads fit well into a read-cache situation (Nimble, NetApp's FlashCache or Flash Pool, etc) then the SSDs will help your reads, but otherwise it's still just a shelf of SATA disk.

It's the same reason I'm always wary about Compellent: when the rubber meets the road, do you really want all of your production data on a small number of slow SATA disks? SAS disks are going to give you 6x the IOPS/GB of SATA.

FISHMANPET
Mar 3, 2007

Sweet 'N Sour
Can't
Melt
Steel Beams

madsushi posted:

At the end of the day, Nimble is really just a shelf of SATA disk. It's got your normal NVRAM for write-caching and SSD for read-caching, but your consistent writes are limited to your SATA disks. If your reads fit well into a read-cache situation (Nimble, NetApp's FlashCache or Flash Pool, etc) then the SSDs will help your reads, but otherwise it's still just a shelf of SATA disk.

It's the same reason I'm always wary about Compellent: when the rubber meets the road, do you really want all of your production data on a small number of slow SATA disks? SAS disks are going to give you 6x the IOPS/GB of SATA.

But you can get Compellent with SAS :confused:

bort
Mar 13, 2003

And a large number of them. Maybe they're only nearline?

madsushi
Apr 19, 2009

Baller.
#essereFerrari

FISHMANPET posted:

But you can get Compellent with SAS :confused:

Many of the Compellent installs I see are with 6 SAS (15k) and 12 SAS 7.2k drives. My overall point was that even though there's a fast "tier" there, if you are doing anything substantial you are still limited by the 7.2k drives. That's why tiering is a dangerous game to play: because once you are outside of its capabilities, you are limited by the slower drives.

bort
Mar 13, 2003

Writes always go to the top tier, though. Data that's migrated to lower tiers only does so specifically because it's idle. You can also peg data to a particular tier of disk if you are paranoid. I don't think your fear is completely unfounded but I think practically it may be rather unlikely. If by doing something substantial, you mean filling the array and not growing it so that it can tier properly, then yes, that is indeed a dangerous game.

GrandMaster
Aug 15, 2004
laidback
We've been looking at nimble boxes too, but I was surprised at how expensive they were considering it's full of lovely SATA. We are looking at ~100TB of storage, and it came in more expensive than Compellent, VNX5500 & FAS3250 boxes with similar capacity - the other boxes take up more space but I've got much more confidence around the performance since they all have truckloads of 15K SAS & SSD caching.

I'm concerned about how some of the workloads would perform on a Nimble like some of our OLTP/OLAP etc apps. I'm sure VMware/VDI would run pretty quick though.

oogs
Dec 6, 2011

madsushi posted:

Many of the Compellent installs I see are with 6 SAS (15k) and 12 SAS 7.2k drives. My overall point was that even though there's a fast "tier" there, if you are doing anything substantial you are still limited by the 7.2k drives. That's why tiering is a dangerous game to play: because once you are outside of its capabilities, you are limited by the slower drives.

The tiers on the compellent are interesting - it's fun to watch the raid stripe on the "slow" side rebuild, wreaking havoc on other VMs that share those disks through other (random) stripes.


NippleFloss - yup, that's exactly why we got ours. We like the Nimble because we deal with a small set of hot data that is heavy on random I/O, and a large set of data (old emails) that is usually inert. Our compellent can keep up with the random I/O if there isn't anything big going on in the background, but as soon as there's a rebuild, a few vmotions, or some automatic optimization, we see the effects spread farther and wider than we'd like.

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

bort posted:

Writes always go to the top tier, though. Data that's migrated to lower tiers only does so specifically because it's idle. You can also peg data to a particular tier of disk if you are paranoid. I don't think your fear is completely unfounded but I think practically it may be rather unlikely. If by doing something substantial, you mean filling the array and not growing it so that it can tier properly, then yes, that is indeed a dangerous game.

The top tier in Compellant is basically a write-through cache, which I'm not a big fan of. You're acknowledging that you're going to do triple the work for any write and in the form of write, then later read and re-write to slower disk. This is based on the assumption that you've got relatively idle time to perform that I/O, which is often, but not always true.

Likewise, I think that tiering via physically moving data is generally kludgy. The page sizes are generally much larger than actual transaction sizes. Compellant's page size is relatively small compared to the competition (512k, I think?) but you're unlikely to see a transaction larger than 64k and for truly random work, which benefits most from the SSD tier, your transaction sizes are likely smaller. So you're wasting a lot of space and I/O moving data you don't actually need. It's also slower responding than read cache since it requires moving blocks rather than just reading them into memory. The eviction process to move data down a tier can be particularly problematic since at the same time your users are accessing data on the SATA tier, and pushing that up to the SSD layer, the SSD layer is trying to evict data to make room for the incoming data, and it's writing that out to the SATA layer. So the automated tiering creates conflict between a user workload and a process that is meant to improve the speed of the user workload.

I guess I just have a hard time seeing what it's really good for. High throughput workloads will be bound by the SATA layer since you'll over-run your SSD capacity pretty quickly with those (this is a general problem with write cache). For random workloads I'd worry about the performance variance depending on where my data was coming from. I'd rather have consistent 10ms responses than some 1ms responses and some 50ms responses, which is what you get if you're not sure what tier you're getting data from at any given time. If I'm going to pin a workload to a specific tier then I'm not really tiering at that point anyway.

Some of these objections are based corner cases, and I know there are many happy users of these automated data tiering systems. I just think it's an inelegant solution since it creates excess spinning disk I/O and the most expensive operation a disk array ever performs is I/O to spinning disk.

bort
Mar 13, 2003

Really good points. I'm not sure if this is consistent across the board, but at least on our Compellent, tiering only happens every 24 hours and runs as a batch job during the slow part of the day. The data progression shouldn't contend with production workloads, but there is the chance that an occasionally run job will have data on a higher latency tier when it shouldn't. The point is really to save money and not have idle data on expensive disk. It's been performing well for the way we use our data, but you've given me something to think about.

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

bort posted:

Really good points. I'm not sure if this is consistent across the board, but at least on our Compellent, tiering only happens every 24 hours and runs as a batch job during the slow part of the day. The data progression shouldn't contend with production workloads, but there is the chance that an occasionally run job will have data on a higher latency tier when it shouldn't. The point is really to save money and not have idle data on expensive disk. It's been performing well for the way we use our data, but you've given me something to think about.
If it performs the way you need and the price is good then that's really all that matters. I like to discuss the technical details on this stuff, but I'd never recommend that someone switch from a storage vendor that they are happy with. So much can go wrong with migrations, no need to
rock the boat for a theoretical benefit that you likely won't notice.

Rhymenoserous
May 23, 2008

GrandMaster posted:

We've been looking at nimble boxes too, but I was surprised at how expensive they were considering it's full of lovely SATA. We are looking at ~100TB of storage, and it came in more expensive than Compellent, VNX5500 & FAS3250 boxes with similar capacity - the other boxes take up more space but I've got much more confidence around the performance since they all have truckloads of 15K SAS & SSD caching.

I'm concerned about how some of the workloads would perform on a Nimble like some of our OLTP/OLAP etc apps. I'm sure VMware/VDI would run pretty quick though.

At the 100TB mark I'd be looking at a bigger vendor too. To me Nimble's products make sense at the medium business level i.e. I need 20tb of storage for VM's etc.

Bitch Stewie
Dec 17, 2011
Anyone have any 3PAR dealings? Their new 7x00 entry level arrays look like they're worth investigating.

psydude
Apr 1, 2008

Anyone had problems pulling performance statistics off of their NetApps in WhatsUp Gold? I copied the MIBs to the MIB store and have all of the right SNMP credentials, but the only thing I'm getting is interface statistics. I have nothing about memory, CPU, or disk utilization.

evil_bunnY
Apr 2, 2003

Apple SMB, ladies and gents:

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

psydude posted:

Anyone had problems pulling performance statistics off of their NetApps in WhatsUp Gold? I copied the MIBs to the MIB store and have all of the right SNMP credentials, but the only thing I'm getting is interface statistics. I have nothing about memory, CPU, or disk utilization.

I didn't have any issues getting SolarWinds to query NetApp devices using the custom MIBs, though it's been a while since I set them up. Since OnCommand Performance Advisor is free now my customer uses that. Have you tried just running an snmpwalk at the base of the tree and seeing what it returns?

Nomex
Jul 17, 2002

Flame retarded.
I can vouch for Oncommand. I honestly don't know how I used Netapp hardware without it now.

Beelzebubba9
Feb 24, 2004

oogs posted:

Hey!

Yes, they do! We're getting some for our main units. We're drooling at this stuff. We have separate head units depending on their purpose, project, etc. You know how that goes...

Wooo more Nimble users! I've got a pair of CS460s coming in and I feel like I'm more excited to have them in house then I should be.

Beelzebubba9
Feb 24, 2004

adorai posted:

can someone ballpark what 15-20TB of nimble storage would cost me, per HA cluster? My only requirements are 10GBe, NFS or iSCSI, deduplication, and mirroring.

Any specific IOPS requirements? Also, Nimble doesn't do de-dupe, just compression. Not sure if that matters to your usage case (or how much it matters at all), but it bears mentioning.

Our CS460s were pushing 6 figures each with 2.4TB of Flash and four years of 4 hour support. If you don't need 60K+ stated IOPS you could get a CS240 for probably well under half what we paid. Feel free to PM me if you have any more questions or if you want me to reach out to my sales rep for sound figures.

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

Beelzebubba9 posted:

Any specific IOPS requirements? Also, Nimble doesn't do de-dupe, just compression. Not sure if that matters to your usage case (or how much it matters at all), but it bears mentioning.

Our CS460s were pushing 6 figures each with 2.4TB of Flash and four years of 4 hour support. If you don't need 60K+ stated IOPS you could get a CS240 for probably well under half what we paid. Feel free to PM me if you have any more questions or if you want me to reach out to my sales rep for sound figures.

Talking about IOPs without discussion what *kind* of IOPs is completely meaningless. What block size, read or write, how random is the workload? Saying "how many IOPs do you need, this system can do 60k" doesn't really say anything. I could do 1 million IOPs on just about any system if I can fit everything in cache. Or if my block size is 1 byte.

I don't mean to single you out, this is just a pet peeve of mine with storage talk in general. Talking about IOPs without context is about as useful ask asking someone how fast they can run and them saying "72".

Beelzebubba9
Feb 24, 2004

NippleFloss posted:

I don't mean to single you out, this is just a pet peeve of mine with storage talk in general. Talking about IOPs without context is about as useful ask asking someone how fast they can run and them saying "72".

You're totally right, I was just crudely try to get at Adorai's workload since it seems to me there are better values for the money than Nimble's SANs if you aren't going to leverage the strengths of their design, one of those being random write performance. I'll phrase my posts better in the future to avoid that kind of unclear language.

adorai
Nov 2, 2002

10/27/04 Never forget
Grimey Drawer

Beelzebubba9 posted:

Any specific IOPS requirements?
Very hard to define, we are considering replacing the SAN already in place at an acquisition. We are still in the "their team, our team" mode, and in an effort to avoid anyone feeling butthurt I am trying to avoid this kind of query. Reality is that after 7 months of production usage, it will go into read only mode and really just suck up electricity in case the FDIC asks us for data.

The other alternative is to relocate an existing netapp to the location and buy more storage for it.

Which brings me to a nice question for nipplefloss: is it worth considering (or even possible at this point) to get a PAM for a 2050?

YOLOsubmarine
Oct 19, 2004

When asked which Pokemon he evolved into, Kamara pauses.

"Motherfucking, what's that big dragon shit? That orange motherfucker. Charizard."

adorai posted:

Which brings me to a nice question for nipplefloss: is it worth considering (or even possible at this point) to get a PAM for a 2050?

No flashcache or flashpool on a 2050, unfortunately. The 2050 just does not have nearly enough memory to support them. You could see if your rep can track down a CPOC system as those generally see some pretty substantial discounts. A 2240 would be worlds better than a 2050.

evil_bunnY
Apr 2, 2003

I have a 2240 and wuve it.

Adbot
ADBOT LOVES YOU

madsushi
Apr 19, 2009

Baller.
#essereFerrari

adorai posted:

Which brings me to a nice question for nipplefloss: is it worth considering (or even possible at this point) to get a PAM for a 2050?

2050 is a dead box unfortunately, no ONTAP updates anymore.

In general, the 2xxx series doesn't have any PCI-E slots for expansion slots. So the 2020/2040/2050/2220/2240 can't support FlashCache (PAM). I don't think it has to do with the memory, it has to do with the fact that they don't have the right slot. In addition, there was an issue with some of the older 3xxx series that prevented FlashCache from working after you upgraded to 8+.

A 2240 isn't going to get you FlashCache, but it is going to be way faster than a 2050 is (in addition to all of the nifty 8+ features). 50% more RAM, faster CPUs, etc. I actually have a 2240 on my bench right now with a 10Gb card and it's very fast.

  • 1
  • 2
  • 3
  • 4
  • 5
  • Post
  • Reply