Data footprint reduction (Part 1): Life beyond dedupe and changing data lifecycles

Over the past couple of weeks there has been a flurry of IT industry activity around data footprint impact reduction with Dell buying Ocarina and IBM acquiring Storwize. For those who want the quick (compacted, reduced) synopsis of what Dell buying Ocarina as well as IBM acquiring Storwize means read this post here along with some of my comments here and here.

Now, before any Drs or Divas of Dedupe get concerned and feel the need to debate dedupes expanding role, success or applicability, relax, take a deep breath, then read on and take another breath before responding if so inclined.

The reason I mention this is that some may mistake this as a piece against or not in favor of dedupe as it talks about life beyond dedupe which could be mistaken as indicating dedupes diminished role which is not the case (read ahead and see figure 5 to see the bigger picture).

Likewise some might feel that since this piece talks about archiving for compliance and non regulatory situations along with compression, data management and other forms of data footprint reduction they may be compelled to defend dedupes honor and future role.

Again, relax, take a deep breath and read on, this is not about the death of dedupe.

Now for others, you might wonder why the dedupe tongue in check humor mentioned above (which is what it is) and the answer is quite simple. The industry in general is drunk on dedupe and in some cases thus having numbed its senses not to mention having blurred its vision of the even bigger opportunities for the business benefits of data footprint reduction beyond todays backup centric or vmware server virtualization dedupe discussions.

Likewise, it is time for the industry to wake (or sober) up and instead of trying to stuff everything under or into the narrowly focused dedupe bottle. Instead, realize that there is a broader umbrella called data footprint impact reduction which includes among other techniques, dedupe, archive, compression, data management, data deletion and thin provisioning across all types of data and applications. What this means is a broader opportunity or market than what exists or being discussed today leveraging different techniques, technologies and best practices.

Consequently this piece is about expanding the discussion to the larger opportunity for vendors or vars to extend their focus to the bigger world of overall data footprint impact reduction beyond where currently focused. Likewise, this is about IT customers realizing that there are more opportunities to address data and storage optimization across your entire organization using various techniques instead of just focusing on backup.

In other words, there is a very bright future for dedupe as well as other techniques and technologies that fall under the data footprint reduction umbrella including data stored online, offline, near line, primary, secondary, tertiary, virtual and in a public or private cloud..

Before going further however lets take a step back and look at some business along with IT issues, challenges and opportunities.

What is the business and IT issue or challenge?
Given that there is no such thing as a data or information recession shown in figure 1, IT organizations of all size are faced with the constant demand to store more data, including multiple copies of the same or similar data, for longer periods of time.


Figure 1: IT resource demand growth continues

The result is an expanding data footprint, increased IT expenses, both capital and operational, due to additional Infrastructure Resource Management (IRM) activities to sustain given levels of application Quality of Service (QoS) delivery shown in figure 2.

Some common IT costs associated with supporting an increased data footprint include among others:

  • Data storage hardware and management software tools acquisition
  • Associated networking or IO connectivity hardware, software and services
  • Recurring maintenance and software renewal fees
  • Facilities fees for floor space, power and cooling along with IT staffing
  • Physical and logical security for data and IT resources
  • Data protection for HA, BC or DR including backup, replication and archiving


Figure 2: IT Resources and cost balancing conflicts and opportunities

Figure 2 shows the result is that IT organizations of all size are faced with having to do more with what they have or with less including maximizing available resources. In addition, IT organizations often have to overcome common footprint constraints (available power, cooling, floor space, server, storage and networking resources, management, budgets, and IT staffing) while supporting business growth.

Figure 2 also shows that to support demand, more resources are needed (real or virtual) in a denser footprint, while maintaining or enhancing QoS plus lowering per unit resource cost. The trick is improving on available resources while maintaining QoS in a cost effective manner. By comparison, traditionally if costs are reduced, one of the other curves (amount of resources or QoS) are often negatively impacted and vice versa. Meanwhile in other situations the result can be moving problems around that later resurface elsewhere. Instead, find, identify, diagnose and prescribe the applicable treatment or form of data footprint reduction or other IT IRM technology, technique or best practices to cure the ailment.

What is driving the expanding data footprint?
Granted more data can be stored in the same or smaller physical footprint than in the past, thus requiring less power and cooling per Gbyte, Tbyte or PByte. Data growth rates necessary to sustain business activity, enhanced IT service delivery and enable new applications are placing continued demands to move, protect, preserve, store and serve data for longer periods of time.

The popularity of rich media and Internet based applications has resulted in explosive growth of unstructured file data requiring new and more scalable storage solutions. Unstructured data includes spreadsheets, Power Point, slide decks, Adobe PDF and word documents, web pages, video and audio JPEG, MP3 and MP4 files. This trend towards increasing data storage requirements does not appear to be slowing anytime soon for organizations of all sizes.

After all, there is no such thing as a data or information recession!

Changing data access lifecycles
Many strategies or marketing stories are built around the premise that shortly after data is created data is seldom, if ever accessed again. The traditional transactional model lends itself to what has become known as information lifecycle management (ILM) where data can and should be archived or moved to lower cost, lower performing, and high density storage or even deleted where possible.

Figure 3 shows as an example on the left side of the diagram the traditional transactional data lifecycle with data being created and then going dormant. The amount of dormant data will vary by the type and size of an organization along with application mix. 


Figure 3: Changing access and data lifecycle patterns

However, unlike the transactional data lifecycle models where data can be removed after a period of time, Web 2.0 and related data needs to remain online and readily accessible. Unlike traditional data lifecycles where data goes dormant after a period of time, on the right side of figure 3, data is created and then accessed on an intermittent basis with variable frequency. The frequency between periods of inactivity could be hours, days, weeks or months and, in some cases, there may be sustained periods of activity.

A common example is a video or some other content that gets created and posted to a web site or social networking site such as Face book, Linked in, or You Tube among others. Once the content is discussed, while it may not change, additional comment and collaborative data can be wrapped around the data as additional viewers discover and comment on the content. Solution approaches for the new category and data lifecycle model include low cost, relative good performing high capacity storage such as clustered bulk storage as well as leveraging different forms of data footprint reduction techniques.

Given that a large (and growing) percentage of new data is unstructured, NAS based storage solutions including clustered, bulk, cloud and managed service offerings with file based access are gaining in popularity. To reduce cost along with support increased business demands (figure 2), a growing trend is to utilize clustered, scale out and bulk NAS file systems that support NFS, CIFS for concurrent large and small IOs as well as optionally pNFS for large parallel access of files. These solutions are also increasingly being deployed with either built in or add on accessorized data footprint reduction techniques including archive, policy management, dedupe and compression among others.

What is your data footprint impact?
Your data footprint impact is the total data storage needed to support your various business application and information needs. Your data footprint may be larger than how much actual data storage you have as seen in figure 4. In Figure 4, an example is an organization that has 20TBytes of storage space allocated and being used for databases, email, home directories, shared documents, engineering documents, financial and other data in different formats (structured and unstructured) not to mention varying access patterns.


Figure 4: Expanding data footprint due to data proliferation and copies being retained

Of the 20TBytes of data allocated and used, it is very likely that the consumed storage space is not 100 percent used. Database tables may be sparsely (empty or not fully) allocated and there is likely duplicate data in email and other shared documents or folders. Additionally, of the 20TBytes, 10TBytes are duplicated to three different areas on a regular basis for application testing, training and business analysis and reporting purposes.

The overall data footprint is the total amount of data including all copies plus the additional storage required for supporting that data such as extra disks for Redundant Array of Independent Disks (RAID) protection or remote mirroring.

In this overly simplified example, the data footprint and subsequent storage requirement are several times that of the 20TBytes of data. Consequently, the larger the data footprint the more data storage capacity and performance bandwidth needed, not to mention being managed, protected and housed (powered, cooled, situated in a rack or cabinet on a floor somewhere).

Data footprint reduction techniques
While data storage capacity has become less expensive on a relative basis, as data footprint continue to expand in order to support business requirements, more IT resources will be needed to be made available in a cost effective, yet QoS satisfying manner (again, refer back to figure 2). What this means is that more IT resources including server, storage and networking capacity, management tools along with associated software licensing and IT staff time will be required to protect, preserve and serve information.

By more effectively managing the data footprint across different applications and tiers of storage, it is possible to enhance application service delivery and responsiveness as well as facilitate more timely data protection to meet compliance and business objectives. To realize the full benefits of data footprint reduction, look beyond backup and offline data improvements to include online and active data using various techniques such as those in table 1 among others.

There are several methods (shown in table 1) that can be used to address data footprint proliferation without compromising data protection or negatively impacting application and business service levels. These approaches include archiving of structured (database), semi structured (email) and unstructured (general files and documents), data compression (real time and offline) and data deduplication.

 

Archiving

Compression

Deduplication

When to use

Structured (database), email and unstructured

Online (database, email, file sharing), backup or archive

Backup or archiving or recurring and similar data

Characteristic

Software to identify and remove unused data from active storage devices

Reduce amount of data to be moved (transmitted) or stored on disk or tape.

Eliminate duplicate files or file content observed over a period of time to reduce data footprint

Examples

Database, email, unstructured file solutions with archive storage

Host software, disk or tape, (network routers) and compression appliances or software as well as appearing in some primary storage system solutions

Backup and archive target devices and Virtual Tape Libraries (VTLs), specialized appliances

Caveats

Time and knowledge to know what and when to archive and delete, data and application aware

Software based solutions require host CPU cycles impacting application performance

Works well in background mode for backup data to avoid performance impact during data ingestion

Table 1: Data footprint reduction approaches and techniques

Archiving for compliance and general data retention
Data archiving is often perceived as a solution for compliance, however, archiving can be used for many other non compliance purposes. These include general data footprint reduction, to boost performance and enhance routine data maintenance and data protection. Archiving can be applied to structured databases data, semi structured email data and attachments and unstructured file data.

A key to deploying an archiving solution is having insight into what data exists along with applicable rules and policies to determine what can be archived, for how long, how many copies and how data ultimately may be finally retired or deleted. Archiving requires a combination of hardware, software and people to implement business rules.

A challenge with archiving is having the time and tools available to identify what data should be archived and what data can be securely destroyed when no longer needed. Further complicating archiving is that knowledge of the data value is also needed; this may well include legal issues as to who is responsible for making decisions on what data to keep or discard.

If a business can invest in the time and software tools, as well as identify which data to archive to support an effective archive strategy, the returns can be very positive towards reducing the data footprint without limiting the amount of information available for use.

Data compression (real time and offline)
Data compression is a commonly used technique for reducing the size of data being stored or transmitted to improve network performance or reduce the amount of storage capacity needed for storing data. If you have used a traditional or TCP/IP based telephone or cell phone, watched either a DVD or HDTV, listened to an MP3, transferred data over the internet or used email you have most likely relied on some form of compression technology that is transparent to you. Some forms of compression are time delayed, such as using PKZIP to zip files, while others are real time or on the fly based such as when using a network, cell phone or listening to an MP3.

Two different approaches to data compression that vary in time delay or impact on application performance along with the amount of compression and loss of data are loss less (no data loss) and lossy (some data loss for higher compression ratio). In addition to these approaches, there are also different implementations of including real time for no performance impact to applications and time delayed where there is a performance impact to applications.

In contrast to traditional ZIP or offline, time delayed compression approaches that require complete decompression of data prior to modification, online compression allows for reading from, or writing to, any location within a compressed file without full file decompression and resulting application or time delay. Real time appliance or target based compression capabilities are well suited for supporting online applications including databases, OLTP, email, home directories, web sites and video streaming among others without consuming host server CPU or memory resources or degrading storage system performance.

Note that with the increase of CPU server processing performance along with multiple cores, server based compression running in applications such as database, email, file systems or operating systems can be a viable option for some environments.

A scenario for using real time data compression is for time sensitive applications that require large amounts of data such as online databases, video and audio media servers, web and analytic tools. For example, databases such as Oracle support NFS3 Direct IO (DIO) and Concurrent IO (CIO) capabilities to enable random and direct addressing of data within an NFS based file. This differs from traditional NFS operations where a file would be sequential read or written.

Another example of using real time compression is to combine a NAS file server configured with 300GB or 600GB high performance 15.5K Fibre Channel or SAS HDDs in addition to flash based SSDs to boost the effective storage capacity of active data without introducing a performance bottleneck associated with using larger capacity HDDs. Of course, compression would vary with the type of solution being deployed and type of data being stored just as dedupe ratios will differ depending on algorithm along with if text or video or object based among other factors.

Deduplication (Dedupe)
Data deduplication (also known as single instance storage, commonalty factoring, data difference or normalization) is a data footprint reduction technique that eliminates the occurrence of the same data. Deduplication works by normalizing the data being backed up or stored by eliminating recurring or duplicate copies of files or data blocks depending on the implementation.

Some data deduplication solutions boast spectacular ratios for data reduction given specific scenarios, such as backup of repetitive and similar files, while providing little value over a broader range of applications.

This is in contrast with traditional data compression approaches that provide lower, yet more predictable and consistent data reduction ratios over more types of data and application, including online and primary storage scenarios. For example, in environments where there is little to no common or repetitive data files, data deduplication will have little to no impact while data compression generally will yield some amount of data footprint reduction across almost all types of data.

Some data deduplication solution providers have either already added, or have announced plans to add, compression techniques to compliment and increase the data footprint effectiveness of their solutions across a broader range of applications and storage scenarios, attesting to the value and importance of data compression to reduce data footprint.

When looking at deduplication solutions, determine if the solution is designed to scale in terms of performance, capacity and availability over a large amount of data along with how restoration of data will be impacted by scaling for growth. Other items to consider include how data is reduplicated, such as real time using inline or some form of time delayed post processing, and the ability to select the mode of operation.

For example, a dedupe solution may be able to process data at a specific ingest rate inline until a certain threshold is hit and then processing reverts to post processing so as to not cause a performance degradation to the application writing data to the deduplication solution. The downside of post processing is that more storage is needed as a buffer. It can, however, also enable solutions to scale without becoming a bottleneck during data ingestion.

However, there is life beyond dedupe which is to in no way diminish dedupe or its very strong and bright future, one that Im increasingly convinced of having talked with hundreds of IT professionals (e.g. the customers) is that only the surface is being scratched for dedupe, not to mention larger data footprint impact opportunity seen in figure 5.


Figure 5: Dedupe adoption and deployment waves over time

While dedupe is a popular technology from a discussion standpoint and has good deployment traction, it is far from reaching mass customer adoption or even broad coverage in environments where it is being used. StorageIO research shows broadest adoption of dedupe centered around backup in smaller or SMB environments (dedupe deployment wave one in figure 5) with some deployment in Remote Office Branch Office (ROBO) work groups as well as departmental environments.

StorageIO research also shows that complete adoption in many of those SMB, ROBO, work group or smaller environments has yet to reach 100 percent. This means that there remains a large population that has yet to deploy dedupe as well as further opportunities to increase the level of dedupe deployment by those already doing so.

There has also been some early adoption in larger core IT environments where dedupe coexists with complimenting existing data protection and preservation practices. Another current deployment scenario for dedupe has been for supporting core edge deployments in larger environments that provide support for backup and data protection of ROBO, work group and departmental systems.

Note that figure 5 simply shows the general types of environments in which dedupe is being adopted and not any sort of indicators as to the degree of deployment by a given customer or IT environment.

What to do about your expanding data footprint impact?
Develop an overall data foot reduction strategy that leverages different techniques and technologies addressing online primary, secondary and offline data. Assess and discover what data exists and how it is used in order to effectively manage storage needs.

Determine policies and rules for retention and deletion of data combining archiving, compression (online and offline) and dedupe in a comprehensive data footprint strategy. The benefit of a broader, more holistic, data footprint reduction strategy is the ability to address the overall environment, including all applications that generate and use data as well as IRM or overhead functions that compound and impact the data footprint.

Data footprint reduction: life beyond (and complimenting) dedupe
The good news is that the Drs. and Divas of dedupe marketing (the ones who also are good at the disco dedupe dance debates) have targeted backup as an initial market sweet (and success) spot shown in figure 5 given the high degree of duplicate data.


Figure 6: Leverage multiple data footprint reduction techniques and technologies

However that same good news is bad news in that there is now a stigma that dedupe is only for backup, similar to how archive was hijacked by the compliance marketing folks in the post Y2K era. There are several techniques that can be used individually to address specific data footprint reduction issues or in combination as seen in figure 7 to implement a more cohesive and effective data footprint reduction strategy.


Figure 7: How various data footprint reduction techniques are complimentary

What this means is that both archive, dedupe as well as other forms of data footprint reduction can and should be used beyond where they have been target marketed using the applicable tool for the task at hand. For example, a common industry rule of thumb is that on average, ten percent of data changes per day (your mileage and rate of change will certainly vary given applications, environment and other factors).

Now assuming that you have 100TB (feel free to subtract a zero or two, or add as many as needed) of data (note I did not say storage capacity or percent utilized), ten percent change would be 10TB that needs to be backed up, replicated and so forth. Now with basic 2 to 1 streaming tape compression (2.5 to 1 in upcoming LTO enhancements) would reduce the daily backup footprint from 10TB to 5TB.

Using dedupe with 10 to 1 would get that from 10TB down to 1TB or about the size of a large capacity disk drive. With 20 to 1 that cuts the daily backup down to 500GB and so forth. The net effect is that more daily backups can be stored in the same footprint which in turn helps expedite individual file recover by having more options to choose from off of the disk based cache, buffer or storage pool.

On the other hand, if your objective is to reduce and eliminate storage capacity, then the same amount of backups can be stored on less disk freeing up resources. Now take the savings times the number of days in your backup retention and you should see the numbers start to add up.

Now what about the other 90 percent of the data that may not have changed, or, that did change and exists on higher performance storage?

Can its footprint impact be reduced?

The answer should be perhaps or it depends as well as prompts the question of what tool would be best. There is a popular thinking as is often the case with industry buzzwords or technologies to use it everywhere. After all goes the thinking, if it is a good thing why not use and deploy more of it everywhere?

Keep in mind that dedupe trades time to perform thinking and apply intelligence to further reduce data in exchange for space capacity. Thus trading time for space capacity can have a negative impact on applications that need lower response time, higher performance where the focus is on rates vs ratios. For example, the other 90 to 100 percent of the data in the above example may have to be on a mix of high and medium performance storage to meet QoS or service level agreement (SLA) objectives. While it would fun or perhaps cool to try and achieve a high data reduction ratio on the entire 100TB of active data with dedupe (e.g. trying to achieve primary dedupe), the performance impacts could have a negative impact.

The option is to apply a mix of different data footprint reduction techniques across the entire 100TB. That is, use dedupe where applicable and higher reduction ratios can be achieved while balancing performance, compression used for streaming data to tape for retention or archive as well as in databases or other applications software not to mention in networks. Likewise, use real time compression or what some refer to as primary dedupe for online active changing data along with online static read only data.

Deploy a comprehensive data footprint reduction strategy combining various techniques and technologies to address point solution needs as well as the overall environment, including online, near line for backup, and offline for archive data.

Lets not forget about archiving, thin provisioning, space saving snapshots, commonsense data management among other techniques across the entire environment. In other words, if your focus is just on dedupe for backup to
achieve an optimized and efficient storage environment, you are also missing

out on a larger opportunity. However, this also means having multiple tools or

technologies in your IT IRM toolbox as well as understanding what to use when, where and why.

Data transfer rates is a key metric for performance (time) optimization such as meeting backup or restore or other data protection windows. Data reduction ratios is a key metric for capacity (space) optimization where the focus is on storing as much data in a given footprint

Some additional take away points:

  • Develop a data footprint reduction strategy for online and offline data
  • Energy avoidance can be accomplished by powering down storage
  • Energy efficiency can be accomplished by using tiered storage to meet different needs
  • Measure and compare storage based on idle and active workload conditions
  • Storage efficiency metrics include IOPS or bandwidth per watt for active data
  • Storage capacity per watt per footprint and cost is a measure for in active data
  • Small percentage reductions on a large scale have big benefits
  • Align the applicable form of virtualization for the given task at hand

Some links for additional reading on the above and related topics

Wrap up (for now, read part II here)

For some applications reduction ratios are an important focus on the tools or modes of operations that achieve those results.

Likewise for other applications where the focus is on performance with some data reduction benefit, tools are optimized for performance first and reduction secondary.

Thus I expect messaging from some vendors to adjust (expand) to those capabilities that they have in their toolboxes (product portfolios) offerings

Consequently, IMHO some of the backup centric dedupe solutions may find themselves in niche roles in the future unless they can diversity. Vendors with multiple data footprint reduction tools will also do better than those with only a single function or focused tool.

However for those who only have a single or perhaps a couple of tools, well, guess what the approach and messaging will be.

After all, if all you have is a hammer everything looks like a nail, if all you have is a screw driver, well, you get the picture.

On the other hand, if you are still not clear on what all this means, send me a note, give a call, post a comment or a tweet and will be happy to discuss with you.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

Availability or lack there of: Lessons From Our Frail & Aging Infrastructure

I have a new blog post over at Enterprise Efficiency about aging infrastructures including those involved with IT, Telcom and related ones.

As a society, we face growing problems repairing and maintaining the vital infrastructure we once took for granted.

Most of these incidents involve aging, worn-out physical infrastructure desperately in need of repair or replacement. But infrastructure doesn’t have to be old or even physical to cause problems when it fails.

The IT systems and applications all around us form a digital infrastructure that most enterprises take for granted until it’s not there.

Bottom line, there really isn’t much choice.

You can either pay up front now to update aging infrastructures, or, wait and pay more later. Either way, there will be a price to pay and you can not realize a cost savings until you actually embark on that endeavor.

Here is the link to the full blog post over at Enterprise Efficiency.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

July 2010 Odds and Ends: Perspectives, Tips and Articles

Here are some items that have been added to the main StorageIO website news, tips and articles, video podcast related pages that pertain to a variety of topics ranging from data storage, IO, networking, data centers, virtualization, Green IT, performance, metrics and more.

These content items include various odds and end pieces such as industry or technology commentary, articles, tips, ATEs (See additional ask the expert tips here) or FAQs as well as some video and podcasts for your mid summer (if in the northern hemisphere) enjoyment.

The New Green IT: Productivity, supporting growth, doing more with what you have

Energy efficient and money saving Green IT or storage optimization are often associated to mean things like MAID, Intelligent Power Management (IPM) for servers and storage disk drive spin down or data deduplication. In other words, technologies and techniques to minimize or avoid power consumption as well as subsequent cooling requirements which for some data, applications or environments can be the case. However there is also shifting from energy avoidance to that of being efficient, effective, productive not to mention profitable as forms of optimization. Collectively these various techniques and technologies help address or close the Green Gap and can reduce the amount of Green IT confusion in the form of boosting productivity (same goes for servers or networks) in terms of more work, IOPS, bandwidth, data moved, frames or packets, transactions, videos or email processed per watt per second (or other unit of time).

Click here to read and listen to my comments about boosting IOPs per watt, or here to learn more about the many facets of energy efficient storage and here on different aspects of storage optimization. Want to read more about the next major wave of server, storage, desktop and networking virtualization? Then click here to read more about virtualization life beyond consolidation where the emphasis or focus expands to abstraction, transparency, enablement in addition to consolidation for servers, storage, networks. If you are interested in metrics and measurements, Storage Resource Management (SRM) not to mention discussion about various macro data center metrics including PUE among others, click on the preceding links.

NAS and Shared Storage, iSCSI, DAS, SAS and more

Shifting gears to general industry trends and commentary, here are some comments on consumer and SOHO storage sharing, the role and importance Value Added Resellers (VARs) serve for SMB environments, as well as the top storage technologies that are in use and remain relevant. Here are some comments on iSCSI which continues to gain in popularity as well as storage options for small businesses.

Are you looking to buy or upgrade a new server? Here are some vendor and technology neutral tips to help determine needs along with requirements to help be a more effective informed buyer. Interested or do you want to know more about Serial Attached SCSI (6Gb/s SAS) including for use as external shared direct attached storage (DAS) for Exchange, Sharepoint, Oracle, VMware or HyperV clusters among other usage scenarios, check out this FAQ as well as podcast. Here are some other items including a podcast about using storage partitions in your data storage infrastructure, an ATE about what type of 1.5TB centralized storage to support multiple locations, and a video on scaling with clustered storage.

That is all for now, hope all is well and enjoy the content.

Cheers gs

Greg Schulz – Author The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier)
twitter @storageio

A Storage I/O Momentus Moment

I recently asked for and received from Seagate (See recent post about them moving their paper head quarters to Ireland here) a Momentus XT 500GB 7200 RPM 2.5 Hybrid Hard Disk Drive (HHDD) to use in an upcoming project. That project is not to test a bunch of different Hard Disk Drives (HDDs), HHDDs, Removable HDD (RHDDs) or Solid State Devices (read more about SSDs here and here or storage optimization here) in order to produce results for someone for a fee or some other consideration.

Do not worry, I am not jumping on the bandwagon of calling my office collection of computers, storage, networks and software the StorageIO Independent hands on test lab. Instead, my objective is to actually use the Momentus XT in conjunction with other storage I/O devices ranging from notebook or laptop, desktop or server, NAS and cloud based storage in conjunction with regular projects that Im working on both in the office as well as while traveling to various out and about activities.

More often than not these days, common thinking or perception is that if anybody is talking about a product or technology it must be a paid for activity as why would anyone write or talk about something without getting or expecting something in exchange (granted there are some exceptions). Given this era of transparency talk, lets walk the talk and here is my disclosure which for those who have read my content before hopefully you will realize that disclosures should be simple, straight forward, easy, fun and common sense based instead of having to dance around or hide what may be being done.

Disclosure moment:
This is not a paid for or sponsored blog (read my disclosure statement here) and in fact is no way connected to in conjunction with, endorsed, sanctioned or approved by Seagate for that matter nor have they been and currently are not a client. I did however ask them for and they offered to send to me a single 500GB Momentus XT Hybrid Hard Disk Drive (HHDD) with no enclosure, accessories, adapter, cables, software or other packaging to be used for a project I am working on. However I did buy from Amazon.com a Seagate GoFlex USB 3.0 to SATA 3 connection cable kit that I had been eyeing for some other projects. Nuff said about that.

What am I doing with a Seagate Momentus XT
As to the project I am working on, it has nothing to do with Seagate or any other vendors or clients for that matter as it is a new book that I will tell you more about in future posts. What I can share with you for now is that it is a follow on to my most previous books ( The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier) ). The new book will also be published by CRC Taylor and Francis.

Now for those who are interested in why would I request a Momentus XT Hybrid Hard Disk Drive (HHDD) from Seagate while turning down others offers of free hardware, software, services, trips and the like it is many fold. First I already own some Momentus (as perhaps you do and may not realize it) HDDs thus thought it would be fun and relatively straight forward to make some general comparisons. I needed some additional storage and I/O improvements to compliment and coexist with what I already have.

Does this mean that the book is going to be about flash Solid State Devices (SSD) since I am using a Momentus XT HHDD? The short answer is NO, it will be much more broadly focused however certainly various types of storage I/O control, public and private clouds, management, gaining control, networking, virtualization as well as other hardware, software, services techniques and technologies will be discussed building on my two previous books.

In addition, I want to see how compatible and useful in every day activities the HHDDs are as opposed to running a couple of standard iometer or other so called lab bench tests. After all, when you buy storage or any IT solutions, do you buy them to be used in your lab to run tests, or, do you buy them to do actual day to day tasks?

I also have been a fan of the HHDD as well as flash and DRAM based SSDs for many years (make that decades for SSDs) and see the opportunity to increase how I am actually using HDDs, HHDDs, SSDs as well as Removable Hard Disk Drives (RHDD) in conjunction with NAS, DAS and other storage to support my book writing as well as other projects that I have bought in the past.

What is the Seagate Momentus XT
The Seagate Momentus series of HDDs are positioned as desktop, notebook and laptop devices that vary in rotational speed (RPM), physical form factor, storage capacity as well as price. The XT is a Hybrid Hard Disk Drive (HHDD) that is essentially a best of breed (hence Hybrid) type device incorporating the high capacity and low cost of a traditional 2.5 7200 RPM HDD with performance boost of flash SSD memory. For example some initial testing of working with very large files have found that the XT can in some instances be as fast as a SSD while holding 10x the capacity with a favorable price.

In other words, an effective balance of cost per GByte capacity, cost per IOP and energy efficiency per IOP. This does not mean however that an XT should be used everywhere or for a replacement to DRAM or flash SSD quite to the contrary as those devices are good tools for specific needs or applications. Instead, the XT provides a good balance of performance and capacity to bridge the gap between traditional spinning HDDs price per capacity and performance per cost of SSD. (For those interested, here is a link to what Seagate is doing with SSD e.g. Pulsar in addition to HHDD and HDD).

Value proposition and business (or consumer) benefits moment
What is the benefit, why not just go all flash?

Simple and that is price unless your specific needs fit into the capacity space of an SSD and you need both the higher performance and lower energy draw (with subsequent heat generation). Note that I did not say heat elimination as during a recent quick test of copying 6GB of data to a flash based SSD it was warm just as the XT device was, however also a bit cooler than a comparable 7200 RPM 2.5 drive. If you can afford the full SSD flash or dram based device as well as it fits your needs and compatibility, go for it. However also make sure that you will see the full expected benefit of adding a SSD to your specific solutions as not all implementations are the same (e.g. do your homework).

Why not just go all HDD?

Simple, economics and performance which is why as I said back in 2005 that HHDDs had a very bright future and will IMHO drive a wedge between the traditional HDD and emerging flash based SSD markets at least for non consumer devices on a near term basis given their compatibility capabilities.

In other words, you could think of it as a compromise, or as a best of breed. For example I can see where for compatible not to mention cost and customer comfort ability of a known entity HHDD will gain some popularity in desktops, laptops, notebooks as well as other devices where a performance boost is needed however not at the expense of throwing out capacity or tight economic budgets.

I can also see some interesting scenarios for hosting virtual machines (VMs) to support server Virtualization with VMware, HyperV or Xen based solutions among others. Another scenario is for bulk storage or archive and backup solutions where the HHDD with their extended cache in the form of flash can help to boost performance of read or write operations on VTLs and dedupe devices, archive platforms, backup or other similar functions. Sure the Momentus XT is positioned as a desktop, notebook type device however has that ever stopped vendors or solution providers from using those types of devices in different roles other than what they were designed for? I am just sayin.

Speeds, feeds and buzzword bingo moment
Seagate has many different types of disk drives that can be found here. In general, the Momentus XT is a 2.5 small form factor (SFF) Hybrid Hard Disk Drive (HHDD) available in 500GB, 320GB and 250GB capacity (I have the 500GB model ST95005620AS) with 4GB SLC NAND (flash) SSD memory, 32MB of drive level cache, an underlying 7200RPM disk drive with SATA 3Gb/s interface including as well as Native Command Queuing (NCQ). Now if you want to say that the XT implements tiered storage in a single device (DRAM, flash and HDD) go ahead. Following are a couple of links of where you can learn more.

Seagate Seatools disk drive diagnostic software (free here)

Seagate FreeAgent Goflex Upgrade Cable (USB 3.0 to SATA 3 STAE104) (Seagate site and Amazon)

Seagate Momentus XT site with general information, product overview and data sheets as well as on Amazon

What does a Momentus XT have to do with writing a book?
If you have ever written a book, or for that matter, done a large development project of any type then things should be a bit familiar. These types of projects include the needs to keep organized as well as protected multiple copies of documents (a dedupers dream) including text, graphics or figures, spreadsheets not to mention project tracking material among others. Likewise as is the case with other authors who work for a living, much of these books are written, edited, proofed or thought about while traveling to different vents, client sites, conferences, meetings or on vacation for that matter. Hence the need to have multiple copies of data on different devices to help guard against when something happens (note that I did not say if).

This is nothing new as each of my last two solo book projects as well as when I was a coauthor contributing content to other books including The Resilient Enterprise (Veritas/Symantec). Much of the content was created while traveling relying on portable storage and backup while on the road. Something someone pointed out to me recently is that this is an example of eating your own dog food or eliminating the shoe makers children syndrome (where the shoe maker creates shoes for others however not for his own children).

Initial moments and general observations
From time to time I will post some notes and observations about how the Momentus XT is performing or behaving which if all goes as planned and so far has, it should be very transparent coexisting with some of my Removable Hard Disk Drives (RHDD) such as the Imation Odyssey which I bought several years ago for offsite bulk removable storage of data that goes to a secure vault somewhere.

Initial deployment other than a stupid mistake on my part has been smooth. What was the stupid mistake you ask? Simple, when I attached the drive via a USB 3.0 cable to SATA 3 connector to one of my XP SP3 systems, Windows saw the device however it did not show up in the list of available devices. Ok, I know I know, it was late in the evening however that is no excuse for realizing that the disk had not yet been initialized let alone formatted. A quick check using Seatools (free here) showed all was well. I then launched Windows Disk Manager, did the initialize, followed by format and all was good from that point on. Wow, wonder how much credibility I will lose over that gaff with the techno elite (that is a joke and a bit of humor btw).

I have already done some initial familiarization and compatibility testing with some of my other drives including a 2.5 64GB SATA flash SSD as well as a 2.5 7200RPM HDD both that I use for bulk data movement activities. At some point I also plan on attaching the XT to my Iomega IX4 NAS to try various things as I have done with other external devices in the past.

Granted these were not ideal conditions as I was in hurry and wanted to get some quick info. Given the probably less than ideal configuration as the format after the HDD was first initialized took about an hour using a FAT32 plug and play configuration. With NTFS and other optimizations I assume it can be better however this was again just to get an initial glimpse of the device in use.

Given that it is a HHDD that uses flash as a big buffer with a 500GB HDD plus 32MB of cache as a backing store, it was interesting attaching it to the computer, then waiting a few minutes, then launching a file copy. Where a normal HDD would start slightly vibrating due to rotation, it was a few moments before any vibration or noise was detected on the Momentus XT which should be of no surprise as the flash was doing its job acting as a buffer until the HDD spun up for work.

I did some initial file copying back and forth between different computers while LAN and NAS were busy doing other things including backups to the Mozy cloud. No discrete time or performance benchmarks to talk about yet, however overall, the XT not surprisingly does seem to be a bit faster than another external 7200 RPM 2.5 drive I use for bulk data moves both on reads and writes. Likewise, given that it is a hybrid HDD leveraging flash as an extended cache with an underlying HDD plus 32MB of cache, it may not always be as fast as my external 2.5 64GB flash SSD, however that is also a common apples to oranges comparison mistake (more on that in a future post).

For example, copying over 6GBytes of data (5 large files of various size) from a 7200 RPM 2.5 160GB Momentus drive in a laptop to the HHDD XT and a flash SSD both took about 8 to 9 minutes where as the normal copy to a 2.5 5400 RPM HDD takes at least 14 to 15 minutes if not longer. Note that these are very rough and far from accurate or reflective comparisons rather a quick gauge of benefits (e.g. getting data moved faster). When I get around to it, will do some more accurate comparisons and put into a follow up post. However I can see already where the XT has the performance similar to the SSD however with almost 10x the capacity which means it could possibly have an interesting role in supporting disk to disk (D2D) backups which I will give a try.

Eventually I will be removing the USB connector kit and actually installing the Momentus into a computer or two (not at the same time) however I am currently walking before running. Im still up in the air as to if I would install the XT into a computer with Windows XP SP3, or simply do a new install of Windows 7 on it to which Im open to thoughts, comments, feedback or applicable suggestions (besides switching to a Macbook or iPad).

Wrap up and fun moment

In the above photo, there is the Seagate Momentus (ST95005620AS), a Goflex USB 3.0 to SATA conversion attachment cable (docking device), a fortune cookie, couple of US quarters and Canadian two dollar coins (See out and about update), paper clips and fishing bobber on a note pad. Why the coins to show relative size and diversity across different geographies as this device will be traveling (it missed out on recent European trip to Holland).

Why the paper clips? Simple, why not, you never know when you will need one for something such as a MacGyver moment, or for pushing the tiny reset button on a device among other activities.

How about the fortune cookie? For good luck and I might need a quick snack while having a cup of coffee not to mention Chinese as well as Asian in general is one of my favorites cuisines to prepare or cook not to mention eat.

Oh, what about the fishing bobber? Why not, it was just laying around and you could also that Im fishing for information to see how the device fits into normal use or that it is there for fun or to add color to the photo.

Oh, and the note pad? Hmm, well, if you cannot figure that one out besides being a back drop, lets just say that the Momentus line in general as well as XT specifically are targeted for notebook, desktop, laptop or other deployment scenarios. If you still dont see the connection, ok fine, feel free to post a comment and I will happily clarify it for you.

That is all for the moment, however I will be following up with more soon.

In the meantime, enjoy your summer if in the northern hemisphere (or winter if in the south).

Take lots of photos, videos and make audio recordings to fill up those USB flash thumb drives (consumer SSD), SD memory cards, computer hard drives, cloud and online web hosting sites so that have you something to remember your special out and about moments by.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

Gregs StorageIO Out and About Update: June 2010

With the 2010 summer solstice having occurred in the northern hemisphere that means it is time for a quick out and about update. It has been a busy winter and spring in the office, on the road as well as at home.

Some results of this recent activity have appeared in blog, on my web site as well as via other sites and venues. For example, activity or content ranges from Industry Trends and Perspectives white papers, reports, blogs, newsletter commentary, interviews, Internet TV, videos, web cast, pod casts (including several appearances on StorageMonkeys Infosmack as well as Rich Brambleys Virtumania), ask the expert (ATE) questions, twitter tweets, tips and columns. Then there were the many in person presentations, key note and seminar events, conferences, briefing sessions along with virtual conferencing and advisory consulting sessions (read and see more here).

Greg Schulz and StorageIO in the news

Regarding having new content appearing in different or new venues, Silicon Angle (including a video), Newstex and Enterprise Efficiencies join the long list of industry and vertical, traditional along with new world venues that my content as well as industry trends and perspective commentary appear in. Read more about events and activities here, content here or commentary here.

Speaking of books, there is also some news in that The Green and Virtual Data Center (CRC) is now available on Amazon Kindle (click on links below) as well as having been translated and published in China not to mention having undergone another round of printing keeping up with demand to make more copies available via global venues.

The Green and Virtual Data Center Chineese Edition: ISBN 978-7-115-21827-8

As for what am I seeing and hearing, check out the new series of Industry Trends and Perspective (ITP) short blog posts that compliment other posts as well as content found on the main web site. These ITP pieces capture what I am hearing and seeing (that is of those what I can talk about that are not under NDA of course) while out and about.

Some of the cities that I have been at while out and about doing keynote speaking and seminar events as well as for other meetings have included Minneapolis, Miami, San Diego, Beverly Hills, San Jose, San Diego (again), Hollywood (again), Austin, Miami (again), New York City, Reston, Minneapolis (again), Irvine, New York City (again), Boston, Toronto, Atlanta, Chicago, Columbus, Philadelphia, Mountain View, Mahtomedia (Minneapolis area), Boston (again) and Indianapolis, Calgary, Jasper (Alberta), Vancouver in Canada as well as Nijkerk (Netherlands) for a one day seminar covering Industry Trends and Perspectives in addition to changing planes in Atlanta, Detroit, Memphis and Las Vegas.

The Planes should be obvious, however what about automobiles you ask? How about the following taken from my rental car while driving north of LAX on the 405 after a January storm during my trip from San Diego after a morning event to Beverly Hills to do an evening keynote.

Rainbow seen from 405 north of LAX
Driving north of LAX on the 405 with a rainbow after rain storm

Another car trip a few weeks later after a different event in San Diego I had a driver from a service behind the wheel so that I could get some work done before an evening meeting. Also on the car front, after flying into Indianapolis there was a car ride to Indianapolis Motor Speedway (IMS) to do a keynote for a CDW sponsored event in gasoline alley a few days before the big race there. While we are on the topic of automobiles and technology, if you have not seen it, check out a post I did about what NAS, NASA and NASCAR have in common.

Gasoline Alley at Indy 500 Practice during a speaking eventIndy 500 Practice during a speaking event

What about trains you ask?

VIA Rail: The CanadianWaiting for morning Train at Nijkerk Station to take me to Amsterdam Airport

Besides the normal airport trams or trains, there was a fun Amtrak Acela ride from New York City Penn station after a morning event in the city up to Boston so as to be in place for a morning event the next day. Other train activity besides airport, subway or commuter light rail in the US and Europe (Holland), there was also an overnight trip on VIA Rail Canada the Canadian from Jasper Alberta to Vancouver (some business tied into a long weekend). If you have never been to the Canadian Rockies, let alone traveled via train, check this one, it was a blast and I highly recommend it.

Lake Louise Alberta CanadaBear family seen near Jasper Alberta
Lake Louise and Jasper area bear family in Alberta Canada

It just dawned on me, what about any out and about via boats?

Other than the Boston water taxi to Logan Airport from the convention center where EMCworld was held and that I did an Internet TV interview along with @Stu and @Scott_Lowe, boat activity has been so far relegated to relaxation.

However, as all work and no play could make for a dull boy (or girl), I can update you that the out and about via boat fishing and sightseeing activity has been very good so far this fall even with high (then low, then high) water on the scenic St. Croix river way.

Here are some scenes from out and about on the St. Croix river including an eagle in its nest tending to its young who can not be seen in this photo as well as fishing (and catching and releasing).

Greg and his Fish Guide: Out and About on St. Croix River Photos by Karen SchulzWaleye Fish: Out and About on St. Croix River Photos by Karen Schulz
This is Walter: Out and About on St. Croix River Photos by Karen SchulzOne of our Neighbors who had an addition to their family this year: Out and About on St. Croix River Photos by Karen Schulz

In between travels (as well as during on planes, trains and in hotel rooms) as well as relaxation breaks, I have been also working on several other projects. Some of these can be seen on the news or tips and articles as well as video and pod cast pages in addition to custom research as well as advisory consulting services. I have also been working on some other projects some of which will become visible over the next weeks and months, others not for a longer period of time yet and yet others that fall under the NDA category so that is all I have to say about that.

If you are not receiving or have seen them, the inaugural issue of the Server and StorageIO newsletter appeared in late February followed by the second edition (Spring 2010) this past week. Both can be found here and here as well as at www.storageio.com/newsletter or subscribing via newsletter@storageio.com.

StorageIO Newsletter

A question I often get asked is what am I hearing or seeing particularly with regards to IT customers as well as with vars during my travels. Here are some photos covering some of the things that I have seen so far this year while out and about.


Super TV or Visualization device at Texas Advanced Computing Center (TACC) in Austin
Note all of the dell servers side by side under the screens required to drive the image.


Taking a walk inside a supercomputer (left) and Texas Supercomputer (Note the horns)

View of MTC during one of stops part of a five city server virtualizaiton series I did
Microsoft Technology Center (MTC)

view from coach classFlight travel tools
View from the back of the plane (left), Airplane long haul essentials: water, food, ipod, coffee, eye shades

Dutch boats
Boats in Holland taken after dinner before recent seminar I did in Nijkerk

Dutch snack (yum yum) foodDutch Soccer or Pub Grub
Dutch Soccer (Pub) food and snacks being enjoyed after a recent seminar in Nijkerk

Waiting at AMS for flight to MSPAirplane food and maps
Airport waiting for planes in AMS (left), more airplane snacks and a map (right)

As to what am I seeing and hearing pertaining to IT, storage, networking and server trends or issues they include among others (see the newsletter):

Whats on deck and and that I am working on?

Having had a busy fun winter and spring Im going to get some relaxation time in during a couple of week period of no travel, however there is plenty to do and get ready for. The summer months will slow down a bit on the out and about travel events scene, however not to a complete stop. In between preparing for upcoming events, advisory and consulting activities as well as researching new material and topics not to mention working on some projects that you will see or hear more about in the weeks and months to come.

For example I will be a guest on a webcast sponsored by Viridity discussing the importance of data center metrics, measurement and insight for effective management to enable energy efficient and effective data centers on July 8th. In addition, I will also be doing another five city storage virtualization series in Stamford, Cleveland, Miami, Tampa and Louisville during mid to late July among other upcoming activities including VMworld in San Francisco.


Check out the events page for more details, specific dates and venues.

What about you?

What have you been doing or have planned for your summer?

Let me know what you are seeing or hearing as well as have been doing.

In the meantime however keep these hints and tips in mind:

  • Have plenty of reading material (real physical books or magazines) or virtual (Kindle or other) as well as via Internet or online to read while at the beach (make sure your computer or PDA is backed up), pool side, in the backyard or elsewhere
  • Remember your eye shades (sun glasses or eye wear), hat and sun screen and if applicable, inspect or bug repellant (e.g. RAID is still useful)
  • Drink plenty of liquid fluids while outside in the summer heat including non alcoholic ones that do not have umbrellas or other interesting garnish
  • Have a place to backup and protect all those summer photos, videos and audio clips that you record while on your out and about adventure. However, keep in mind privacy concerns when uploading them to various social mediums. After all, what happens in Vegas stays in Vegas and what happens on the web stays on the web!

Thanks to everyone involved in the recent events which can be seen here, as well for those who will be participating in upcoming ones I look forward to meeting and talking with you.

Until next time have a fun, safe and relaxing summer if you are in the northern hemisphere and for those down under, not to worry, spring is on the way soon for you as well.

Cheers gs

Greg Schulz – Author The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier)
twitter @storageio

Follow via Google Feedburner here or via email subscription here.

Supreme Court Rules Sarbox intact, Oversight Board Changes


Today the US Supreme Court ruled on a Nevada case involving constitutionality of the 2002 Sarbanes-Oxley (Sarbox) accounting regulations pertaining to appointments to the independent public company accounting oversight board.

The Supreme Court ruled that the Sarbox regulations or law remains intact, however the process or controls around the oversight board must change.

My interpretation and perspective from reading a few different reports is that Sarbox as you know and love (or hate) it is essentially still intact. However what has changed or will be is that individual board members can now be removed or at least in an easier manner. Instead of the request to strike down the Sarbox regulations, the Supreme Court instead appears to have left the regulations intact instead ruling that board members can be changed or removed.

What does this all mean?

Perhaps not much other than firms who have been making money on Sarbox now having something else to talk or consult about (Hmmm, a Sarbox stimulus?).

On the other hand, with the ability to have Sarbox board members more easily removed, perhaps we will see a new board installed that could influence the thinking and thus applicability of Sarbox activity.

Near term, I can see this as being non news for some, and for others, confusion and lets not forget that in chaos or confusion there is opportunity.

Here are some links to read more

  • US Supreme Court website and other news
  • Supreme Court to Hear Challenge to Accounting Board
  • Court Strikes Down Part of Sarbanes-Oxley
  • Nuff said about this for now, whats your take?

    Cheers gs

    Greg Schulz – Author The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    VMware vExpert 2010: Thank You, Im Honored to be named a Member

    This week while traveling I received an email note from John Troyer of VMware informing me that I have been nominated and selected as a VMware vExpert for 2010.


    To say that I was surprised and honored would be an understatement.

    Thus, I would like to thank all those involved in the nominations, evaluation and selection process for being named to this esteemed group.

    I would also like to say congratulations, best wishes and hello to all of the other 2010 vExperts. Im Looking forward to being involved and participating in the VMware vExpert community.

    Cheers gs

    Greg Schulz – Author The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    June 2010 StorageIO Newsletter

    StorageIO News Letter Image
    June 2010 Newsletter

    Welcome to the June 2010 edition of the Server and StorageIO Group (StorageIO) newsletter. This follows the Spring 2010 edition building on the great feedback received from recipients.
    Items that are new in this expanded edition include:

    • Out and About Update
    • Industry Trends and Perspectives (ITP)
    • Featured Article

    You can access this news letter via various social media venues (some are shown below) in addition to StorageIO web sites and subscriptions. Click on the following links to view the June 2010 edition as an HTML or PDF or, to go to the newsletter page to view previous editions.

    Follow via Goggle Feedburner here or via email subscription here.

    You can also subscribe to the news letter by simply sending an email to newsletter@storageio.com

    Enjoy this edition of the StorageIO newsletter, let me know your comments and feedback.

    Cheers gs

    Greg Schulz – Author The Green and Virtual Data Center (CRC) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    Industry Trends and Perspectives: Tiered Storage, Systems and Mediums

    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    Two years ago we read about how the magnetic disk drive would be dead in a couple of years at the hand of flash SSD. Guess what, it is a couple of years later and the magnetic disk drive is far from being dead. Granted high performance Fibre Channel disks will continue to be replaced by high performance, small form factor 2.5" SAS drives along with continued adoption of high capacity SAS and SATA devices.

    Likewise, SSD or flash drives continue to be deployed, however outside of iPhone, iPod and other consumer or low end devices, nowhere near the projected or perhaps hoped for level. Rest assured the trend Im seeing and hearing from IT customers is that some will continue to look for places to strategically deploy SSD where possible, practical and affordable, there will continue to be a roll for disk and even tape devices on a go forward basis.

    Also watch for more coverage and discussion around the emergence of the Hybrid Hard Disk Drive (HHDD) that was discussed about four to five years ago. The HHDD made an appearance and then quietly went away for some time, perhaps more R and D time in the labs while flash SSD garnered the spotlight.

    There could be a good opportunity for HHDD technology leveraging the best of both worlds that is continued pricing decreases for disk with larger capacity using smaller yet more affordable amounts of flash in a solution that is transparent to the server or storage controller making for easier integration.

    Related and companion material:
    Blog: ILM = Has It Losts its Meaning
    Blog: SSD and Storage System Performance
    Blog: Has SSD put Hard Disk Drives (HDDs) On Endangered Species List
    Blog: Optimize Data Storage for Performance and Capacity Efficiency

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Industry Trends and Perspectives: RAID Rebuild Rates

    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    There is continued concern about how long large capacity disk drives take to be rebuilt in RAID sets particularly as the continued shift from 1TB to 2TB occurs. It should not be a surprise that a disk with more capacity will take longer to rebuild or copy as well as with more drives; the likely hood of one failing statistically increases.

    Not to diminish the issue, however also to avoid saying the sky is falling, we have been here before! In the late 90s and early 2000s there was a similar concern with the then large 9GB, 18GB let alone emerging 36GB and 72GB drives. There have been improvements in RAID as well as rebuild algorithms along with other storage system software or firmware enhancements not to mention boost in processor or IO bus performance.

    However not all storage systems are equal even if they use the same underlying processors, IO busses, adapters or disk drives. Some vendors have made significant improvements in their rebuild times where each generation of software or firmware can reconstruct a failed drive faster. Yet for others, each subsequent iteration of larger capacity disk drives brings increased rebuild times.

    If disk drive rebuild times are a concern, ask your vendor or solution provider what they are doing as well as have done over the past several years to boost their performance. Look for signs of continued improvement in rebuild and reconstruction performance as well as decrease in error rates or false drive rebuilds.

    Related and companion material:
    Blog: RAID data protection remains relevant
    Blog: Optimize Data Storage for Performance and Capacity Efficiency

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Industry Trends and Perspectives: Converged Networking and IO Virtualization (IOV)

    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    The trends that I am seeing with converged networking and I/O fall into a couple of categories. One being converged networking including unified communications, FCoE/DCB along with InfiniBand based discussions while the other being around I/O virtualization (IOV) including PCIe server based multi root IO virtualization (MRIOV).

    As is often the case with new technologies the trend of some saying these are the next great things thus drop everything and adopt them now as they are working and ready for prime time mission critical deployment. Then there are those who say no, stop, do not waste your time on these as they are temporary, they will die and go away anyway. In between, there is reality which takes a bit of balancing the old with the new, look before you leap, do your homework, and do not be scared however have a strategy and a plan on how to achieve it.

    Thus is FCoE a temporal or temporary technology? Well, in the scope that all technologies are temporary however it is their temporal timeframe that should be of interest. Thus given that FCoE will probably have at least a ten to fifteen year temporal timeline, I would say in technology terms it has a relative long life for supporting coexistence on the continued road to convergence which appears to be Ethernet.

    Related and companion material:
    Video: Storage and Networking Convergence
    Blog: I/O Virtualization (IOV) Revisited
    Blog: I/O, I/O, Its off to Virtual Work and VMworld I Go (or went)
    Blog: EMC VPLEX: Virtual Storage Redefined or Respun?

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Industry Trends and Perspectives: Storage Virtualization and Virtual Storage

    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    The topic of this post is a trend server virtualization and recent EMC virtual storage announcements.

    Virtual storage or storage virtualization has been as a technology and topic around for some time now. Some would argue that storage virtualization is several years old while others would say many decades depending on your view or definition which will vary by preferences, product, vendor, open or closed, hardware, network, software not to mention feature and functionality.

    Consequently there are many different views and definitions of storage virtualization some tied to that of product specifications often leading to apples and oranges comparison.

    Back in the early to mid 2000s, there was plenty of talk around storage virtualization which then gave way to a relative quiet period before seeing adoption pickup in terms of deployment later in the decade (at least for block based).

    More recently there has a been a renewed flurry of storage virtualization activity with many vendors now shipping their latest versions of tools and functionality, EMC announcing VPLEX as well as the file virtualization vendors continuing to try and create a market for their wares (give it time, like block based, it will evolve).

    One of the trends around storage virtualization and part of the play on words EMC is using is to change the order of the words. That is where storage virtualization is often aligned with product implementation (e.g. software on an appliance or switch or in a storage system) used primarily for aggregation of heterogeneous storage, with VPLEX EMC is referring to it as virtual storage.

    What is interesting here is the play on life beyond consolidation a trend that is also occurring with servers or using virtualization for agility, flexibility and ease of management for upgrades, add, move and changes as opposed to simply pooling of LUNs and underlying storage devices. Stay tuned and watch for more in this space as well as read the blog post below about VPLEX for more on this topic.

    Related and companion material:
    Blog: EMC VPLEX: Virtual Storage Redefined or Respun?

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Industry Trends and Perspectives: Tape, Disk and Dedupe Coexistence

    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    The topic of this post is a trend that I am seeing and hearing about during discussions with IT professionals pertaining to how tape is still alive despite common industry FUD.

    Not only is tape still very much alive with recent enhancements including LTO5 with an extended range roadmap, it is also finding new roles. In addition to being deployed in new roles, tape is coexisting and complimenting dedupe or other disk based backup and data protection approaches and vice versa.

    Hearing tape is alive in the same sentence as dedupe deployments continuing may sound counter intuitive if you only listen to some vendor pitches.

    However if you talk with IT customers particularly those in larger environments or with VARs that provide complete solution offering focus you will hear a different tune than tape is dead and dedupe rules. Tape is still alive however its roll is changing. Watch for more on this and related topics.

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Industry Trends and Perspectives: Tiered Hypervisors and Microsoft Hyper-V

    Storage I/O trends
    This is part of an ongoing series of short industry trends and perspectives blog posts briefs.

    These short posts complement other longer posts along with traditional industry trends and perspective white papers, research reports, solution brief content found at www.storageioblog.com/reports.

    Multiple – Tiered Hypervisors for Server Virtualization

    The topic of this post is a trend that I am seeing and hearing about during discussions with IT professionals of the use of two or more server virtualization hypervisors or what is known as tiered Hypervisors.

    Server Virtualization Hypervisor Trends

    A trends tied to server virtualization that I am seeing more of are that IT organizations are increasingly deploying or using two or more different hypervisors (e.g. Citrix/Xen, Microsoft/Hyper-V, VMware vSphere) in their environment (on separate physical server or blades).

    Tiered hypervisors is a concept similar to what many IT organizations already have in terms of different types of servers for various use cases, multiple operating systems as well as several kinds of storage mediums or devices.

    What Im seeing is that IT pros are using different hypervisors to meet various cost, management and vendor control goals aligning the applicable technology to the business or application service category.

    Tiered Virtualization Hypervisor Management

    Of course this brings up the discussion of how to manage multiple hypervisors and thus the real battle is or will be not about hypervisors, rather that of End to End (E2E) management.

    A question that I often ask VARs and IT customers if they see Microsoft on the offensive or defensive with Hyper-V vs. VMware and vice versa, that is if VMware is on the defense or offense against Microsoft.

    Not surprisingly the VMware and Microsoft faithful will say that the other is clearly on the defensive.

    Meanwhile from other people, the feelings are rather mixed with many feeling that Microsoft is increasingly on the offensive with VMware being seen by some as playing a strong defense with a ferocious offense.

    Learn more

    Related and companion material:
    Video: Beyond Virtualization Basics (Free: May require registration)
    Blog: Server and Storage Virtualization: Life beyond Consolidation
    Blog: Should Everything Be Virtualized?

    That is all for now, hope you find this ongoing series of current and emerging Industry Trends and Perspectives interesting.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved