This is the second of a two-part post about why storage arrays and appliances with SSD drives can be a good idea, here is link to the first post.
So again, why would putting drive form factors SSDs be a bad idea for existing storage systems, arrays and appliances?
Benefits of SSD drive in storage systems, arrays and appliances:
- Familiarity with customers who buy and use these devices
- Reduces time to market enabling customers to innovate via deployment
- Establish comfort and confidence with SSD technology for customers
- Investment protection of currently installed technology (hardware and software)
- Interoperability with existing interfaces, infrastructure, tools and policies
- Reliability, availability and serviceability (RAS) depending on vendor implementation
- Features and functionality (replicate, snapshot, policy, tiering, application integration)
- Known entity in terms of hardware, software, firmware and microcode (good or bad)
- Share SSD technology across more servers or accessing applications
- Good performance assuming no controller, hardware or software bottlenecks
- Wear leveling and other SSD flash management if implemented
- Can end performance bottlenecks if backend (drives) are a problem
- Coexist or complemented with server-based SSD caching
Note, the mere presence of SSD drives in a storage system, array or appliance will not guarantee or enable the above items to be enabled, nor to their full potential. Different vendors and products will implement to various degrees of extensibility SSD drive support, so look beyond the check box of feature, functionality. Dig in and understand how extensive and robust the SSD implementation is to meet your specific requirements.
Caveats of SSD drives in storage systems, arrays and appliances:
- May not use full performance potential of nand flash SLC technology
- Latency can be an issue for those who need extreme speed or performance
- May not be the most innovative newest technology on the block
- Fun for startup vendors, marketers and their fans to poke fun at
- Not all vendors add value or optimization for endurance of drive SSD
- Seen as not being technology advanced vs. legacy or mature systems
Note that different vendors will have various performance characteristics, some good for IOPs, others for bandwidth or throughput while others for latency or capacity. Look at different products to see how they will vary to meet your particular needs.
Cost comparisons are tricky. SSD in HDD form factors certainly cost more than raw flash dies, however PCIe cards and FTL (flash translation layer) controllers also cost more than flash chips by themselves. In other words, apples to apples comparisons are needed. In the future, ideally the baseboard or motherboard vendors will revise the layout to support nand flash (or its replacement) with DRAM DIMM type modules along with associated FTL and BIOS to handle the flash program/erase cycles (P/E) and wear leveling management, something that DRAM does not have to encounter. While that provides great location or locality of reference (figure 1), it is also a more complex approach that takes time and industry cooperation.
Figure 1: Locality of reference for memory and storage
Certainly, for best performance, just like realty location matters and thus locality of reference comes into play. That is put the data as close to the server as possible, however when sharing is needed, then a different approach or a companion technique is required.
Here are some general thoughts about SSD:
- Some customers and organizations get the value and role of SSD
- Some see where SSD can replace HDD, others see where it compliments
- Yet others are seeing the potential, however are moving cautiously
- For many environments better than current performance is good enough
- Environments with the need for speed need every bit of performance they can get
- Storage systems and arrays or appliances continue to evolve including the media they use
- Simply looking at how some storage arrays, systems and appliances have evolved, you can get an idea on how they might look in the future which could include not only SAS as a backend or target, also PCIe. After all, it was not that long ago where backend drive connections went from propriety to open parallel SCSI or SSA to Fibre Channel loop (or switched) to SAS.
- Engineers and marketers tend to gravitate to newer products nand technology, which is good, as we need continued innovation on that front.
- Customers and business people tend to gravitate towards deriving greatest value out of what is there for as long as possible.
- Of course, both of the latter two points are not always the case and can be flip flopped.
- Ultrahigh end environments and corner case applications will continue to push the limits and are target markets for some of the newer products and vendors.
- Likewise, enterprise, mid market and other mainstream environments (outside of their corner case scenarios) will continue to push known technology to its limits as long as they can derive some business benefit value.
While not perfect, SSD in a HDD form factor with a SAS or SATA interface properly integrated by vendors into storage systems (or arrays or appliances) are a good fit for many environments today. Likewise, for some environments, new from the ground up SSD based solutions that leverage flash DIMM or daughter cards or PCIe flash cards are a fit. So to are PCIe flash cards either as a target, or as cache to complement storage system (arrays and appliances). Certainly, drive slots in arrays take up space for SSD, however so to does occupying PCIe space particularly in high density servers that require every available socket and slot for compute and DRAM memory. Thus, there are pros and cons, features and benefits of various approaches and which is best will depend on your needs and perhaps preferences, which may or may not be binary.
I agree that for some applications and solutions, non drive form factor SSD make sense while in others, compatibility has its benefits. Yet in other situations nand flash such as SLC combined with HDD and DRAM tightly integrated such as in my Momentus XT HHDD is good for laptops, however probably not a good fit for enterprise yet. Thus, SSD options and placements are not binary, of course, sometimes opinions and perspectives will be.
For some situations PCIe, based cards in servers or appliances make sense, either as a target or as cache. Likewise for other scenarios drive format SSD make sense in servers and storage systems, appliances, arrays or other solutions. Thus while all of those approaches are used for storing binary digital data, the solutions of what to use when and where often will not be binary, that is unless your approach is to use one tool or technique for everything.
Here are some related links to learn more about SSD, where and when to use what:
Why SSD based arrays and storage appliances can be a good idea (Part I)
IT and storage economics 101, supply and demand
Researchers and marketers dont agree on future of nand flash SSD
Speaking of speeding up business with SSD storage
EMC VFCache respinning SSD and intelligent caching (Part I)
EMC VFCache respinning SSD and intelligent caching (Part II)
SSD options for Virtual (and Physical) Environments: Part I Spinning up to speed on SSD
SSD options for Virtual (and Physical) Environments, Part II: The call to duty, SSD endurance
SSD options for Virtual (and Physical) Environments Part III: What type of SSD is best for you?
Ok, nuff said for now.
Cheers gs
Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)
twitter @storageio
All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved