Application Data Access Lifecycle Patterns Everything Is Not The Same

Application Data Access Life cycle Patterns Everything Is Not The Same(Part V)

Application Data Access Life cycle Patterns Everything Is Not The Same

Application Data Access Life cycle Patterns Everything Is Not The Same

This is part five of a five-part mini-series looking at Application Data Value Characteristics everything is not the same as a companion excerpt from chapter 2 of my new book Software Defined Data Infrastructure Essentials – Cloud, Converged and Virtual Fundamental Server Storage I/O Tradecraft (CRC Press 2017). available at Amazon.com and other global venues. In this post, we look at various application and data lifecycle patterns as well as wrap up this series.

Application Data Value Software Defined Data Infrastructure Essentials Book SDDC

Active (Hot), Static (Warm and WORM), or Dormant (Cold) Data and Lifecycles

When it comes to Application Data Value, a common question I hear is why not keep all data?

If the data has value, and you have a large enough budget, why not? On the other hand, most organizations have a budget and other constraints that determine how much and what data to retain.

Another common question I get asked (or told) it isn’t the objective to keep less data to cut costs?

If the data has no value, then get rid of it. On the other hand, if data has value or unknown value, then find ways to remove the cost of keeping more data for longer periods of time so its value can be realized.

In general, the data life cycle (called by some cradle to grave, birth or creation to disposition) is created, save and store, perhaps update and read with changing access patterns over time, along with value. During that time, the data (which includes applications and their settings) will be protected with copies or some other technique, and eventually disposed of.

Between the time when data is created and when it is disposed of, there are many variations of what gets done and needs to be done. Considering static data for a moment, some applications and their data, or data and their applications, create data which is for a short period, then goes dormant, then is active again briefly before going cold (see the left side of the following figure). This is a classic application, data, and information life-cycle model (ILM), and tiering or data movement and migration that still applies for some scenarios.

Application Data Value
Changing data access patterns for different applications

However, a newer scenario over the past several years that continues to increase is shown on the right side of the above figure. In this scenario, data is initially active for updates, then goes cold or WORM (Write Once/Read Many); however, it warms back up as a static reference, on the web, as big data, and for other uses where it is used to create new data and information.

Data, in addition to its other attributes already mentioned, can be active (hot), residing in a memory cache, buffers inside a server, or on a fast storage appliance or caching appliance. Hot data means that it is actively being used for reads or writes (this is what the term Heat map pertains to in the context of the server, storage data, and applications. The heat map shows where the hot or active data is along with its other characteristics.

Context is important here, as there are also IT facilities heat maps, which refer to physical facilities including what servers are consuming power and generating heat. Note that some current and emerging data center infrastructure management (DCIM) tools can correlate the physical facilities power, cooling, and heat to actual work being done from an applications perspective. This correlated or converged management view enables more granular analysis and effective decision-making on how to best utilize data infrastructure resources.

In addition to being hot or active, data can be warm (not as heavily accessed) or cold (rarely if ever accessed), as well as online, near-line, or off-line. As their names imply, warm data may occasionally be used, either updated and written, or static and just being read. Some data also gets protected as WORM data using hardware or software technologies. WORM (immutable) data, not to be confused with warm data, is fixed or immutable (cannot be changed).

When looking at data (or storage), it is important to see when the data was created as well as when it was modified. However, you should avoid the mistake of looking only at when it was created or modified: Instead, also look to see when it was the last read, as well as how often it is read. You might find that some data has not been updated for several years, but it is still accessed several times an hour or minute. Also, keep in mind that the metadata about the actual data may be being updated, even while the data itself is static.

Also, look at your applications characteristics as well as how data gets used, to see if it is conducive to caching or automated tiering based on activity, events, or time. For example, there is a large amount of data for an energy or oil exploration project that normally sits on slower lower-cost storage, but that now and then some analysis needs to run on.

Using data and storage management tools, given notice or based on activity, which large or big data could be promoted to faster storage, or applications migrated to be closer to the data to speed up processing. Another example is weekly, monthly, quarterly, or year-end processing of financial, accounting, payroll, inventory, or enterprise resource planning (ERP) schedules. Knowing how and when the applications use the data, which is also understanding the data, automated tools, and policies, can be used to tier or cache data to speed up processing and thereby boost productivity.

All applications have performance, availability, capacity, economic (PACE) attributes, however:

  • PACE attributes vary by Application Data Value and usage
  • Some applications and their data are more active than others
  • PACE characteristics may vary within different parts of an application
  • PACE application and data characteristics along with value change over time

Read more about Application Data Value, PACE and application characteristics in Software Defined Data Infrastructure Essentials (CRC Press 2017).

Where to learn more

Learn more about Application Data Value, application characteristics, PACE along with data protection, software defined data center (SDDC), software defined data infrastructures (SDDI) and related topics via the following links:

SDDC Data Infrastructure

Additional learning experiences along with common questions (and answers), as well as tips can be found in Software Defined Data Infrastructure Essentials book.

Software Defined Data Infrastructure Essentials Book SDDC

What this all means and wrap-up

Keep in mind that Application Data Value everything is not the same across various organizations, data centers, data infrastructures, data and the applications that use them.

Also keep in mind that there is more data being created, the size of those data items, files, objects, entities, records are also increasing, as well as the speed at which they get created and accessed. The challenge is not just that there is more data, or data is bigger, or accessed faster, it’s all of those along with changing value as well as diverse applications to keep in perspective. With new Global Data Protection Regulations (GDPR) going into effect May 25, 2018, now is a good time to assess and gain insight into what data you have, its value, retention as well as disposition policies.

Remember, there are different data types, value, life-cycle, volume and velocity that change over time, and with Application Data Value Everything Is Not The Same, so why treat and manage everything the same?

Ok, nuff said, for now.

Gs

Greg Schulz – Microsoft MVP Cloud and Data Center Management, VMware vExpert 2010-2017 (vSAN and vCloud). Author of Software Defined Data Infrastructure Essentials (CRC Press), as well as Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio. Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO. All Rights Reserved. StorageIO is a registered Trade Mark (TM) of Server StorageIO.

Like Data They Protect For Now Quantum Revenues Continue To Grow

For Now Quantum Revenues Continue To Grow

server storage I/O data infrastructure trends

For Now Quantum Revenues Continue To Grow. The other day following their formal announced, I received an summary update from Quantum pertaining to their recent Q1 Results (show later below).

Data Infrastructures Protect Preserve Secure and Serve Information
Various IT and Cloud Infrastructure Layers including Data Infrastructures

Quantums Revenues Continue To Grow Like Data

One of the certainties in life is change and the other is continued growth in data that gets transformed into information via IT and other applications. Data Infrastructures fundamental role is to enable an environment for applications and data to be transformed into information and delivered as services. In other words, Data Infrastructures exist to protect, preserve, secure and serve information along with the applications and data they depend on. Quantums role is to provide solutions and technologies for enabling legacy and cloud or other software defined data infrastructures to protect, preserve, secure and serve data.

What caught my eye in Quantums announcements was that while not earth shattering growth numbers normally associated with a hot startup, being a legacy data infrasture and storage vendor, Quantum’s numbers are hanging in there.

At a time when some legacy as well as startups struggle with increased competition from others including cloud, Quantum appears for at least now to be hanging in there with some gains.

The other thing that caught my eye is that most of the growth not surprisingly is non tape related solutions, particular around their bulk scale out StorNext storage solutions, there is some growth in tape.

Here is the excerpt of what Quantum sent out:


Highlights for the quarter (all comparisons are to the same period a year ago):

•	Grew total revenue and generated profit for 5th consecutive quarter
•	Total revenue was up slightly to $117M, with 3% increase in branded revenue
•	Generated operating profit of $1M with earnings per share of 4 cents, up 2 cents
•	Grew scale-out tiered storage revenue 10% to $34M, with strong growth in video surveillance and technical workflows
o	Key surveillance wins included deals with an Asian government for surveillance at a presidential palace and other government facilities, with a major U.S. port and with four new police department customers
o	Established several new surveillance partnerships – one of top three resellers/integrators in China (Uniview) and two major U.S. integrators (Protection 1 and Kratos)
o	Won two surveillance awards for StorNext – Security Industry Association’s New Product Showcase award and Security Today magazine’s Platinum Govies Government Security award
o	Key technical workflow wins included deals at an international defense and aerospace company to expand StorNext archive environment, a leading biotechnology firm for 1 PB genomic sequencing archive, a top automaker involving autonomous driving research data and a U.S. technology institute involving high performance computing  
o	Announced StorNext 6, which adds new advanced data management features to StorNext’s industry-leading performance and is now shipping
o	Announced scale-out partnerships with Veritone on artificial intelligence and DataFrameworks on data visualization and management  
•	Tape automation, devices and media revenue increased 6% overall while branded revenue for this product category was up 14%
o	Strong sales of newest generation Scalar i3 and i6 tape libraries
•	Established new/enhanced data protection partnerships
o	Enhanced partnership with Veeam, making it easier for their customers to deploy 3-2-1 data protection best practices
o	Became Pure Storage alliance partner, providing our data protection and archive solutions for their customers through mutual channel partners

Where To Learn More

Learn more about related technology, trends, tools, techniques, and tips with the following links.

What This All Means

Keep in mind that Data Infrastructures fundamental role is to enable an environment for applications and data to be transformed into information and delivered as services. Data Infrastructures exist to protect, preserve, secure and serve information along with the applications and data they depend on. Quantum continues to evolve their business as they have for several years from one focused on tape and related technologies to one that includes tape as well as many other solutions for legacy as well as software defined, cloud and virtual environments. For now, quantum revenues continue to grow and diversify.

Ok, nuff said, for now.
Gs

Greg Schulz – Multi-year Microsoft MVP Cloud and Data Center Management, VMware vExpert (and vSAN). Author of Software Defined Data Infrastructure Essentials (CRC Press), as well as Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio.

Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2023 Server StorageIO(R) and UnlimitedIO. All Rights Reserved.

VMware vSAN 6.6 hyper-converged (HCI) software defined data infrastructure

server storage I/O trends

VMware vSAN 6.6 hyper-converged (HCI) software defined data infrastructure

In case you missed it, VMware announced vSAN v6.6 hyper-converged infrastructure (HCI) software defined data infrastructure solution. This is the first of a five-part series about VMware vSAN V6.6. Part II (just the speeds feeds please) is located here, part III (reducing cost and complexity) located here, part IV (scaling ROBO and data centers today) found here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

VMware vSAN 6.6
Image via VMware

For those who are not aware, vSAN is a VMware virtual Storage Area Network (e.g. vSAN) that is software-defined, part of being a software-defined data infrastructure (SDDI) and software-defined data center (SDDC). Besides being software-defined vSAN is HCI combining compute (server), I/O networking, storage (space and I/O) along with hypervisors, management, and other tools.

Software-defined data infrastructure

Excuse Me, What is vSAN and who is if for

Some might find it odd having to explain what vSAN is, on the other hand, not everybody is dialed into the VMware world ecosystem, so let’s give them some help, for everybody else, and feel free to jump ahead.

For those not familiar, VMware vSAN is an HCI software-defined storage solution that converges compute (hypervisors and server) with storage space capacity and I/O performance along with networking. Being HCI means that with vSAN as you scale compute, storage space capacity and I/O performance also increases in an aggregated fashion. Likewise, increase storage space capacity and server I/O performance you also get more compute capabilities (along with memory).

For VMware-centric environments looking to go CI or HCI, vSAN offers compelling value proposition leveraging known VMware tools and staff skills (knowledge, experience, tradecraft). Another benefit of vSAN is the ability to select your hardware platform from different vendors, a trend that other CI/HCI vendors have started to offer as well.

CI and HCI data infrastructure

Keep in mind that fast applications need a fast server, I/O and storage, as well as server storage I/O needs CPU along with memory to generate I/O operations (IOPs) or move data. What this all means is that HCI solutions such as VMware vSAN combine or converge the server compute, hypervisors, storage file system, storage devices, I/O and networking along with other functionality into an easy to deploy (and management) turnkey solution.

Learn more about CI and HCI along with who some other vendors are as well as considerations at www.storageio.com/converge. Also, visit VMware sites to find out more about vSphere ESXi hypervisors, vSAN, NSX (Software Defined Networking), vCenter, vRealize along with other tools for enabling SDDC and SDDI.

Give Me the Quick Elevator Pitch Summary

VMware has enhanced vSAN with version 6.6 (V6.6) enabling new functionality, supporting new hardware platforms along with partners, while reducing costs, improving scalability and resiliency for SDDC and SDDI environments. This includes from small medium business (SMB) to mid-market to small medium enterprise (SME) as well as workgroup, departmental along with Remote Office Branch Office (ROBO).

Being a HCI solution, management functions of the server, storage, I/O, networking, hypervisor, hardware, and software are converged to improve management productivity. Also, vSAN integrated with VMware vSphere among other tools enable modern, robust data infrastructure that serves, protect, preserve, secure and stores data along with their associated applications.

Where to Learn More

The following are additional resources to learn more about vSAN and related technologies.

What this all means

Overall a good set of enhancements as vSAN continues its evolution looking back just a few years ago, to where it is today and will be in the future. If you have not looked at vSAN recently, take some time beyond reading this piece to learn some more.

Continue reading more about VMware vSAN 6.6 in part II (just the speeds feeds please) is located here, part III (reducing cost and complexity) located here, part IV (scaling ROBO and data centers today) located here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

Ok, nuff said (for now…).

Cheers
Gs

Greg Schulz – Microsoft MVP Cloud and Data Center Management, VMware vExpert (and vSAN). Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio. Watch for the spring 2017 release of his new book “Software-Defined Data Infrastructure Essentials” (CRC Press).

Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2023 Server StorageIO(R) and UnlimitedIO. All Rights Reserved.

VMware vSAN V6.6 Part II (just the speeds feeds features please)

server storage I/O trends

VMware vSAN v6.6 Part II (just the speeds feeds features please)

In case you missed it, VMware announced vSAN v6.6 hyper-converged infrastructure (HCI) software defined data infrastructure solution. This is the second of a five-part series about VMware vSAN V6.6. View Part I here, part III (reducing cost and complexity) located here, part IV (scaling ROBO and data centers today) found here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

VMware vSAN 6.6
Image via VMware

For those who are not aware, vSAN is a VMware virtual Storage Area Network (e.g. vSAN) that is software-defined, part of being a software-defined data infrastructure (SDDI) and software-defined data center (SDDC). Besides being software-defined vSAN is HCI combining compute (server), I/O networking, storage (space and I/O) along with hypervisors, management, and other tools.

Just the Speeds and Feeds Please

For those who just want to see the list of what’s new with vSAN V6.6, here you go:

  • Native encryption for data-at-rest
  • Compliance certifications
  • Resilient management independent of vCenter
  • Degraded Disk Handling v2.0 (DDHv2)
  • Smart repairs and enhanced rebalancing
  • Intelligent rebuilds using partial repairs
  • Certified file service & data protection solutions
  • Stretched clusters with local failure protection
  • Site affinity for stretched clusters
  • 1-click witness change for Stretched Cluster
  • vSAN Management Pack for vRealize
  • Enhanced vSAN SDK and PowerCLI
  • Simple networking with Unicast
  • vSAN Cloud Analytics with real-time support notification and recommendations
  • vSAN ConfigAssist with 1-click hardware lifecycle management
  • Extended vSAN Health Services
  • vSAN Easy Install with 1-click fixes
  • Up to 50% greater IOPS for all-flash with optimized checksum and dedupe
  • Support for new next-gen workloads
  • vSAN for Photon in Photon Platform 1.1
  • Day 0 support for latest flash technologies
  • Expanded caching tier choice
  • Docker Volume Driver 1.1

What’s New and Value Proposition of vSAN 6.6

Let’s take a closer look beyond the bullet list of what’s new with vSAN 6.6, as well as perspectives of those features to address different needs. The VMware vSAN proposition is to evolve and enable modernizing data infrastructures with HCI powered by vSphere along with vSAN.

Three main themes or characteristics (and benefits) of vSAN 6.6 include addressing (or enabling):

  • Reducing risk while scaling
  • Reducing cost and complexity
  • Scaling for today and tomorrow

VMware vSAN 6.6 summary
Image via VMware

Reducing risk while scaling

Reducing (or removing) risk while evolving your data infrastructure with HCI including flexibility of choosing among five support hardware vendors along with native security. This includes native security, availability and resiliency enhancements (including intelligent rebuilds) without sacrificing storage efficiency (capacity) or effectiveness (performance productivity), management and choice.

VMware vSAN DaRE
Image via VMware

Dat level Data at Rest Encryption (DaRE) of all vSAN dat objects that are enabled at a cluster level. The new functionality supports hybrid along with all flash SSD as well as stretched clusters. The VMware vSAN DaRE implementation is an alternative to using self-encrypting drives (SEDs) reducing cost, complexity and management activity. All vSAN features including data footprint reduction (DFR) features such as compression and deduplication are supported. For security, vSAN DaRE integrations with compliance key management technologies including those from SafeNet, Hytrust, Thales and Vormetric among others.

VMware vSAN management
Image via VMware

ESXi HTML 5 based host client, along with CLI via ESXCLI for administering vSAN clusters as an alternative in case your vCenter server(s) are offline. Management capabilities include monitoring of critical health and status details along with configuration changes.

VMware vSAN health management
Image via VMware

Health monitoring enhancements include handling of degraded vSAN devices with intelligence proactively detecting impending device failures. As part of the functionality, if a replica of the failing (or possible soon to fail) device exists, vSAN can take action to maintain data availability.

Where to Learn More

The following are additional resources to find out more about vSAN and related technologies.

What this all means

With each new release, vSAN is increasing its feature, functionality, resiliency and extensiveness associated with traditional storage and non-CI or HCI solutions. Continue reading more about VMware vSAN 6.6 in Part I here, part III (reducing cost and complexity) located here, part IV (scaling ROBO and data centers today) found here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

Ok, nuff said (for now…).

Cheers
Gs

Greg Schulz – Microsoft MVP Cloud and Data Center Management, VMware vExpert (and vSAN). Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio. Watch for the Spring 2017 release of his new book “Software-Defined Data Infrastructure Essentials” (CRC Press).

Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2023 Server StorageIO(R) and UnlimitedIO. All Rights Reserved.

VMware vSAN V6.6 Part III (reducing costs complexity)

server storage I/O trends

VMware vSAN V6.6 Part III (Reducing costs complexity)

In case you missed it, VMware announced vSAN v6.6 hyper-converged infrastructure (HCI) software defined data infrastructure solution. This is the third of a five-part series about VMware vSAN V6.6. View Part I here, Part II (just the speeds feeds please) is located here, part IV (scaling ROBO and data centers today) found here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

VMware vSAN 6.6
Image via VMware

For those who are not aware, vSAN is a VMware virtual Storage Area Network (e.g. vSAN) that is software-defined, part of being a software-defined data infrastructure (SDDI) and software-defined data center (SDDC). Besides being software-defined vSAN is HCI combining compute (server), I/O networking, storage (space and I/O) along with hypervisors, management, and other tools.

Reducing cost and complexity

Reducing your total cost of ownership (TCO) including lower capital expenditures (CapEx) and operating (OPEX). VMware is claiming CapEx and OpEx reduced TCO of 50%. Keep in mind that solutions such as vSAN also can help drive return on investment (ROI) as well as return on innovation (the other ROI) via improved productivity, effectiveness, as well as efficiencies (savings). Another aspect of addressing TCO and ROI includes flexibility leveraging stretched clusters to address HA, BR, BC and DR Availability needs cost effectively. These enhancements include efficiency (and effectiveness e.g. productivity) at scale, proactive cloud analytics, and intelligent operations.

VMware vSAN stretch cluster
Image via VMware

Low cost (or cost-effective) Local, Remote Resiliency and Data Protection with Stretched Clusters across sites. Upon a site failure, vSAN maintains availability is leveraging surviving site redundancy. For performance and productivity effectiveness, I/O traffic is kept local where possible and practical, reducing cross-site network workload. Bear in mind that the best I/O is the one you do not have to do, the second is the one with the least impact.

This means if you can address I/Os as close to the application as possible (e.g. locality of reference), that is a better I/O. On the other hand, when data is not local, then the best I/O is the one involving a local or remote site with least overhead impact to applications, as well as server storage I/O (including networks) resources. Also keep in mind that with vSAN you can fine tune availability, resiliency and data protection to meet various needs by adjusting fault tolerant mode (FTM) to address a different number of failures to tolerate.

server storage I/O locality of reference

Network and cloud friendly Unicast Communication enhancements. To improve performance, availability, and capacity (CPU demand reduction) multicast communications are no longer used making for easier, simplified single site and stretched cluster configurations. When vSAN clusters upgrade to V6.6 unicast is enabled.

VMware vSAN unicast
Image via VMware

Gaining insight, awareness, adding intelligence to avoid flying blind, introducing vSAN Cloud Analytics and Proactive Guidance. Part of a VMware customer, experience improvement program, leverages cloud-based health checks for easy online known issue detection along with relevant knowledge bases pieces as well as other support notices. Whether you choose to refer to this feature as advanced analytics, artificial intelligence (AI), proactive rules enabled management problem isolation, solving resolution I will leave that up to you.

VMware vSAN cloud analytics
Image via VMware

Part of the new tools analytics capabilities and prescriptive problem resolution (hmm, some might call that AI or advanced analytics, just saying), health check issues are identified, notifications along with suggested remediation. Another feature is the ability to leverage continuous proactive updates for advance remediation vs. waiting for subsequent vSAN releases. Net result and benefit are reducing time, the complexity of troubleshooting converged data infrastructure issues spanning servers, storage, I/O networking, hardware, software, cloud, and configuration. In other words, enable you more time to be productive vs. finding and fixing problems leveraging informed awareness for smart decision-making.

Where to Learn More

The following are additional resources to find out more about vSAN and related technologies.

What this all means

Continue reading more about VMware vSAN 6.6 in part I here, part II (just the speeds feeds please) located here, part IV (scaling ROBO and data centers today) found here, as well as part V here (VMware vSAN evolution, where to learn more and summary).

Ok, nuff said (for now…).

Cheers
Gs

Greg Schulz – Microsoft MVP Cloud and Data Center Management, VMware vExpert (and vSAN). Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio. Watch for the spring 2017 release of his new book “Software-Defined Data Infrastructure Essentials” (CRC Press).

Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2023 Server StorageIO(R) and UnlimitedIO. All Rights Reserved.

Part II Until the focus expands to data protection – What to do about it

Storage I/O trends

Part II – Until the focus expands to data protection – What to do about it

This is the second of a three-part series (read part I here) about how vendors are keeping backup alive, however what they can and should do to shift and expand the conversation to data protection and related themes.

Modernizing data protection and what to do about it

Building off of what was mentioned in the first post, lets take a look at what can be done including expanding the conversation around data protection in support of business continuance (BC), disaster recovery (DR), high availability (HA), business resiliency (BR) not to mention helping backup to actually retire (someday). Now when I backup retire, I’m not necessarily talking about a technology such as hardware, software or a service including clouds, rather when, where, why and how data gets protected. What I mean by this is to step back from looking at the tools and technologies to how they are used and can be used in new and different ways moving forward.

People convergenceStorageIO people convergence
Converged people and technology teams

All to often I see where new technologies or tools get used in old ways which while providing some near-term relief, the full capabilities of what is being used may not be fully realized. This also ties into the theme of people not technologies can be a barrier to convergence and transformation that you can read more about here and here.

Whats your data protection strategy, business or technology focused?

expand focus beyond tools
Data protection strategy evolving beyond tools looking for a problem to solve

Part of modernizing data protection is getting back to the roots or fundamentals including revisiting business needs, requirements along with applicable threat risks to then align application tools, technologies and techniques. This means expanding focus from just the technology, however also more importantly how to use different tools for various scenarios. In other words having a tool-box and know how to use it vs. everything looking like a nail as all you have is a hammer. Check out various webinars, Google+ hangouts and other live events that I’m involved with on the StorageIO.com events page on data protection and related data infrastructure themes including BackupU (getting back to the basics and fundamentals).

data protection options

Everything is not the same, leverage different data protection approaches to different situations

Wrap up (for now)

Continue reading part three of this series here to see what can be done (taking action) about shifting the conversation about modernizing data protection. Also check out conversations about trends, themes, technologies, techniques perspectives in my ongoing data protection diaries discussions (e.g. www.storageioblog.com/data-protection-diaries-main/).

Ok, nuff said

Cheers
Gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

 

Garbage data in, garbage information out, big data or big garbage?

StorageIO industry trends cloud, virtualization and big data

Do you know the computer technology saying, garbage data in results in garbage information out?

In other words even with the best algorithms and hardware, bad, junk or garbage data put in results in garbage information delivered. Of course, you might have data analysis and cleaning software to look for, find and remove bad or garbage data, however that’s for a different post on another day.

If garbage data in results in garbage information out, does garbage big data in result in big garbage out?

I’m sure my sales and marketing friends or their surrogates will jump at the opportunity to tell me why and how big data is the solution to the decades old garbage data in problem.

Likewise they will probably tell me big data is the solution to problems that have not even occurred or been discovered yet, yeah right.

However garbage data does not discriminate or show preference towards big data or little data, in fact it can infiltrate all types of data and systems.

Lets shift gears from big and little data to how all of that information is protected, backed up, replicated, copied for HA, BC, DR, compliance, regulatory or other reasons. I wonder how much garbage data is really out there and many garbage backups, snapshots, replication or other copies of data exist? Sounds like a good reason to modernize data protection.

If we don’t know where the garbage data is, how can we know if there is a garbage copy of the data for protection on some other tape, disk or cloud. That also means plenty of garbage data to compact (e.g. compress and dedupe) to cut its data footprint impact particular with tough economic times.

Does this mean then that the cloud is the new destination for garbage data in different shapes or forms, from online primary to back up and archive?

Does that then make the cloud the new virtual garbage dump for big and little data?

Hmm, I think I need to empty my desktop trash bin and email deleted items among other digital house keeping chores now.

On the other hand, just had a thought about orphaned data and orphaned storage, however lets leave those sleeping dogs lay where they rest for now.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

More modernizing data protection, virtualization and clouds with certainty

This is a follow-up to a recent post about modernizing data protection and doing more than simply swapping out media or mediums like flat tires on a car as well as part of the Quantum protecting data with certainty event series.

As part of a recent 15 city event series sponsored by Quantum (that was a disclosure btw ;) ) titled Virtualization, Cloud and the New Realities for Data Protection that had a theme of strategies and technologies that will help you adapt to a changing IT environment I was asked to present a keynote at the events around Modernizing data protection for cloud, virtual and legacy environments (see earlier and related posts here and here).

Quantum data protection with certainty

Since late June (taking July and most of August off) and wrapping up last week, the event series has traveled to Boston, Chicago, Palo Alto, Houston, New York City, Cleveland, Raleigh, Atlanta, Washington DC, San Diego, Los Angeles, Mohegan Sun CT, St. Louis, Portland Oregon and King of Prussia (Philadelphia area).

The following are a series of posts via IT Knowledge Exchange (ITKE) that covered these events including commentary and perspectives from myself and others.

Data protection in the cloud, summary of the events
Practical solutions for data protection challenges
Big data’s new and old realities
Can you afford to gamble on data protection
Conversations in and around modernizing data protection
Can you afford not to use cloud based data protection

In addition to the themes in the above links, here are some more images, thoughts and perspectives from while being out and about at these and other events.

Datalink does your data center suck sign
While I was traveling saw this advertisement sign from Datalink (who is a Quantum partner that participated in some of the events) in a few different airports which is a variation of the Datadomain tape sucks attention getter. For those not familiar, that creature on the right is an oversized mosquito with the company logos on the lower left being Datalink, NetApp, Cisco and VMware.

goddess of data fertility
When in Atlanta for one of the events at the Morton’s in the Sun trust plaza, the above sculpture was in the lobby. Its real title is the goddess of fertility, however I’m going to refer to it as the goddess of data fertility, after all, there is no such thing as a data or information recession.

The world and storageio runs on dunkin donuts
Traveling while out and about is like a lot of things particular IT and data infrastructure related which is hurry up and wait. Not only does America Run on Dunkin, so to does StorageIO.

Use your imagination
When out and about, sometimes instead of looking up, or around, take a moment and look down and see what is under your feet, then let your imagination go for a moment about what it means. Ok, nuff of that, drink your coffee and let’s get back to things shall we.

Delta 757 and PW2037 or PW2040
Just like virtualization and clouds, airplanes need physical engines to power them which have to be energy-efficient and effective. This means being very reliable, good performance, fuel-efficient (e.g. a 757 on a 1,500 mile trip if full can be in the neighborhood of 65 plus miles per gallon per passenger with a low latency (e.g. fast trip). In this case, a Pratt and Whitney PW2037 (could be a PW2040 as Delta has a few of them) on a Delta 757 is seen powering this flight as it climbs out of LAX on a Friday morning after one of the event series session the evening before in LA.

Ambulance waiting at casino
Not sure what to make out of this image, however it was taken while walking into the Mohegan Sun casino where we did one of the dinner events at the Michael Jordan restaraunt

David Chapa of Quantum in bank vault
Here is an image from one of the events in this series which is a restaurant in Cleveland where the vault is a dinning room. No that is not a banker, well perhaps a data protection banker, it is the one and only (@davidchapa) David Chapa aka the Chief Technology Evangelist (CTE) of Quantum, check out his blog here.

Just before landing in portland
Nice view just before landing in Portland Oregon where that evenings topic was as you might have guessed, data protection modernization, clouds and virtualization. Don’t be scared, be ready, learn and find concerns to overcome them to have certainty with data protection in cloud, virtual and physical environments.
Teamwork
Cloud, virtualization and data protection modernization is a shared responsibility requiring team work and cooperation between service or solution provider and the user or consumer. If the customer or consumer of a service is using the right tools, technologies, best practices and having had done their homework for applicable levels of services with SLAs and SLOs, then a service provider with good capabilities should be in harmony with each other. Of course having the right technologies and tools for the task at hand is also important.
Underground hallway connecting LAX terminals, path to the clouds
Moving your data to the cloud or a virtualized environment should not feel like a walk down a long hallway, that is assuming you have done your homework, that the service is safe and secure, well taken care of, there should be less of concerns. Now if that is a dark, dirty, dingy, dilapidated dungeon like hallway, then you just might be on the highway to hell vs. stairway to heaven or clouds ;).

clouds along california coastline
There continues to be barriers to cloud adoption and deployment for data protection among other users.

Unlike the mountain ranges inland from the LA area coastline causing a barrier for the marine layer clouds rolling further inland, many IT related barriers can be overcome. The key to overcoming cloud concerns and barriers is identifying and understanding what they are so that resolutions, solutions, best practices, tools or work around’s can be developed or put into place.

The world and storageio runs on dunkin donuts
Hmm, breakfast of champions and road warriors, Dunkin Donuts aka DD, not to be confused with DDUP the former ticker symbol of Datadomain.

Tiered coffee
In the spirit of not treating everything the same, have different technology or tools to meet various needs or requirements, it only makes sense that there are various hot beverage options including hot water for tea, regular and decaffeinated coffee. Hmm, tiered hot beverages?


On the lighter side, things including technology of all type will and do break, even with maintenance, so having a standby plan, or support service to call can come in handy. In this case the vehicle on the right did not hit the garage door that came off of its tracks due to wear and tear as I was preparing to leave for one of the data protection events. Note to self, consider going from bi-annual garage door preventive maintenance to annual service check-up.

Some salesman talking on phone in a quiet zone

While not part of or pertaining to data protection, clouds, virtualization, storage or data infrastructure topics, the above photo was taken while in a quiet section of an airport lounge waiting for a flight to one of the events. This falls in the class of a picture is worth a thousand words category as the sign just to the left of the sales person talking loudly on his cell phone about his big successful customer call says Quiet Zone with symbol of no cell phone conversations.

How do I know the guy was not talking about clouds, virtualization, data infrastructure or storage related topics? Simple, his conversation was so loud me and everybody else in the lounge could hear the details of the customer conversation as it was being relayed back to sales management.

Note to those involved in sales or customer related topics, be careful of your conversations in public and pseudo public places including airports, airport lounges, airplanes, trains, planes, hotel lobbies and other places, you never know who you will be broadcasting to.

Here is a link to a summary of the events along with common questions, thoughts and perspectives.

Quantum data protection with certainty

Thanks to everyone who participated in the events including attendees, as well as Quantum and their partners for sponsoring this event series, look forward to see you while out and about at some future event or venue.

Ok, nuff said.

Cheers Gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

Practical Email optimization and archiving strategies

Email is a popular tool for messaging, calendaring, and managing contacts along with attachments in most organizations.

Email and messaging

Given the popularity of email and diverse ways that it is used for managing various forms of unstructured data attachments including photos, video, audio, spreadsheets, presentations and other document objects, there are corresponding back end challenges. Those back end challenges including managing the data storage repositories (e.g. file systems and storage systems) that are used for preserving and serving email documents along with enabling regulatory or compliance mandates.

Email archiving is an important enabler for regulatory compliance and e-discovery functions. However there is another important use for E-mail archiving which as a data footprint reduction (DFR) technique and technology enables storage optimization, being green and supporting growth while stretching budgets further. There is after all no such thing as a data or information recession and all one has to do to verify the trend is to look at your own email activity.

Industry Trend: Data growth and demand

There are however constraints on time, budgets and demands to do more while relying on more information and email has become a central tool for messaging including social media networking, handling of attachments and means to manage all of that data.

DFR enables more data to be stored, retained, managed and maintenance in a cost effective manner. This includes storing more data managed per person, where when the additional data being retained adds value to an organization. Also included is keeping more data readily accessible, not necessarily instantly accessible, however but within minutes instead of hours or days depending on service requirements.

Data footprint reduction (DFR) techniques and technologies

Here is a link to a recent article that I did presenting five tips and strategies for optimizing e-mail using archiving.

Hopefully many of you will find these to be common sense tips being implemented, however if not, now is the time to take action to stretch your resources further to do more.

In general email optimization tips include:

  • Set policies for retention and disposal
  • Establish filters and rules
  • Index and organize your inbox
  • Archive messages regularly
  • Perform routine cleanup and optimization
  • Leverage cloud data protection services and solutions

When it comes to archiving projects, walk before you run, establish success to build upon for broader deployment of E-mail archiving by finding and address low hanging fruit opportunities.

Instead of trying to do to much, find opportunities that can be addressed and leveraged as examples to build business cases to move forward.

By having some success stories and proof points, these can be used to help convince management to support additional steps not to mention getting them to back your polices to achieve success.

An effective way to convince management these days is to show them how by taking additional Email archiving steps you can support increased growth demand, reduce costs while enhancing productivity not to mention adding compliance and ediscovery capabilities as side benefits.

You can read more here and here, ok, nuff said for now.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2011 StorageIO and UnlimitedIO All Rights Reserved

Cloud and Virtual Data Storage Networking book VMworld 2011 debut

Following up from a previous preview post about my new book Cloud and Virtual Data Storage Networking (CRC Press) for those for those attending VMworld 2011 in Las Vegas Monday August 29 through Thursday September 1st 2011, you can pick up your copy at the VMworld book store.

Cloud and Virtual Data Storage Networking Book

Book signing at VMworld 2011

On Tuesday August 30 at 1PM local time, I will be at the VMworld store signing books. Stop by the book store and say hello, pickup your copy of Cloud and Virtual Data Storage Networking (CRC Press). Also check out the other new releases by fellow vExpert authors during the event. I have also heard rumors that some exhibitors among others will be doing drawings, so keep an eye out in the expo hall and go visit those showing copies of my new book.

The VMworld book store hours are:

Monday 8:30am to 7:30pm
Tuesday 8:30am to 6:00pm
Wednesday 8:30am to 8:00pm
Thursday 8:00am to 2:00pm

For those not attending VMworld 2011, you can order your copy from different venues including Amazon.com, Barnes and Noble, DigitalGuru and CRC Press among others.

Learn more about Cloud and Virtual Data Storage Networking (CRC Press) at https://storageioblog.com/book3

Look forward to seeing you at the various VMworld events in Las Vegas as well as at other upcoming venues.

Ok, nuff said for now.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2011 StorageIO and UnlimitedIO All Rights Reserved

Cloud storage: Dont be scared, however look before you leap

Here is a link to a web cast on BrightTalk I will be doing live on Thursday June 9, 2011 at 1PM Pacific, 3PM Central or 4PM Eastern time lasting about 45 minutes. The web cast is titled: Cloud storage: Dont be scared, however look before you leap.

This web cast session takes a look at the state of public, private and hybrid cloud storage solutions and services including what you need to know to be prepared for a successful deployment. Topics to be covered include best practices, management and data protection in addition to navigating the hype and FUD associated with cloud storage today.

Cloud storage: Dont be scared, however look before you leap and do your homework

Check out the web cast either live or the replay later.

Cheers Gs

Greg Schulz – Author The Green and Virtual Data Center (CRC), Resilient Storage Networks (Elsevier) and coming summer 2011 Cloud and Virtual Data Storage Networking (CRC)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2011 StorageIO and UnlimitedIO All Rights Reserved

More Data Footprint Reduction (DFR) Material

This is part of an ongoing series of short industry trends and perspectives (ITP) blog posts briefs based on what I am seeing and hearing in my conversations with IT professionals on a global basis.

These short posts compliment other longer posts along with traditional industry trends and perspective white papers, research reports, videos, podcasts, webcasts as well as solution brief content found a www.storageio.com/reports and www.storageio.com/articles.

If you recall from previous posts including here, here or here among others, Data Footprint Reduction (DFR) is a collection of tools, technologies and best practices for addressing growing data storage management and cost impacts.

DFR encompasses many different tools, techniques and technologies across various applications ranging from active or primary storage to secondary and inactive along with backup and archive.

Some of the technologies techniques and technologies include archiving, backup modernization, compression, data management, dedupe, space saving snapshots and thin provisioning among others.

Following are some links to various articles and commentary pertaining to DFR:

  • Using DFR including dedupe and compression to defry storage and management costs
  • Deduplicate, compress and defray costs of data storage management
  • Virtual tape libraries: Old backup technology holdover or gateway to the future?
  • As well as here, here or here

In the spirit of DFR, that is doing more with less, nuff said (for now).

Of course let me know what your thoughts and perspectives are on this and other related topics.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

March Metric Madness: Fun with Simple Math

Its March and besides being spring in north America, it also means tournament season including the NCAA basket ball series among others known as March Madness.

Given the office pools and other forms of playing with numbers tied to the tournaments and real or virtual money, here is a quick timeout looking at some fun with math.

The fun is showing how simple math can be used to show relative growth for IT resources such as data storage. For example, say that you have 10Tbytes of storage or data and that it is growing at only 10 percent per year, in five years with simple math yields 14.6Tbytes.

Now lets assume growth rate is 50 percent per year and in the course of five years, instead of having 10Tbytes, that now jumps to 50.6Tbytes. If you have 100Tbytes today and at 50 percent growth rate, that would yield 506.3 Tbytes or about half of a petabyte in 5 years. If by chance you have say 1Pbyte or 1,000Tbytes today, at 25% year of year growth you would have 2.44Pbytes in 5 years.
Basic Storage Forecast
Figure 1 Fun with simple math and projected growth rates

Granted this is simple math showing basic examples however the point is that depending on your growth rate and amount of either current data or storage, you might be surprised at the forecast or projected needs in only five years.

In a nutshell, these are examples of very basic primitive capacity forecasts that would vary by other factors including if the data is 10Tbytes and your policies is for 25 percent free space, that would require even more storage than the base amount. Go with a different RAID level, some extra space for replication, snapshots, disk to disk backups and replication not to mention test development and those numbers go up even higher.

Sure those amounts can be offset with thin provisioning, dedupe, archiving, compression and other forms of data footprint reduction, however the point here is to realize how simple math can portray a very basic forecast and picture of growth.

Read more about performance and capacity in Chapter 10 – Performance and capacity planning for storage networks – Resilient Storage Networks (Elsevier) as well as at www.cmg.org (Computer Measurement Group)..

And that is all I have to say about this for now, enjoy March madness and fun with numbers.

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

Is There a Data and I/O Activity Recession?

Storage I/O trends

With all the focus on both domestic and international economic woes and discussion of recessions and depressions and possible future rapid inflation, recent conversations with IT professionals from organizations of all size across different industry sectors and geographies prompted the question, is there also a data and I/O activity recession?

Here’s the premise, if you listen to current economic and financial reports as well as employment information, the immediate conclusion is that yes, there should also be an I recession in the form of contraction in the amount of data being processed, moved and stored which would also impact I/O (e.g. DAS,, LAN, SAN, FAN or NAS, MAN, WAN) networking activity as well. After all, the server, storage, I/O and networking vendors earnings are all being impacted right?

As is often the case, there is more to the story, certainly vendor earnings are down and some vendors are shipping less product than during corresponding periods from a year or more ago. Likewise, I continue to hear from both IT organizations, vars and vendors of lengthened sales cycles due to increased due diligence and more security of IT acquisitions meaning that sales and revenue forecasts continue to be very volatile with some vendors pulling back on their future financial guidance.

However, does that mean fewer servers, storage, I/O and networking components not to mention less software is being shipped? In some cases there is or has been a slow down. However in other cases, due to pricing pressures, increased performance and capacity density where more work can be done by fewer devices, consolidation, data footprint reduction, optimization, virtualization including VMware and other techniques, not to mention a decrease in some activity, there is less demand. On the other hand, while some retail vendors are seeing their business volume decrease, others such as Amazon are seeing continued heavy demand and activity.

Been on a trip lately through an airport? Granted the airlines have instituted capacity management (e.g. capacity planning) and fleet optimization to align the number of flights or frequency as well as aircraft type (tiering) to the demand. In some cases smaller planes, in other cases larger planes, for some more stops at a lower price (trade time for money) or in other cases shorter direct routes for a higher fee. The point being is that while there is an economic recession underway, and granted there are fewer flights, many if not most of those flights are full which means transactions and information to process by the airlines reservations and operational as well as customer relations and loyalty systems.

Mergers and acquisitions usually mean a reduction or consolidation of activity resulting in excess and surplus technologies, yet talking with some financial services organizations, over time some of their systems will be consolidated to achieve operating efficiency and synergies, near term, in some cases, there is the need for more IT resources to support the increased activity of supporting multiple applications, increased customer inquiry and conversion activity.

On a go forward basis, there is the need to support more applications and services that will generate more I/O activity to enable data to be moved, processed and stored. Not to mention, data being retained in multiple locations for longer periods of time to meet both compliance and non regulatory compliance requirements as well as for BC/DR and business intelligence (BI) or data mining for marketing and other purposes.

Speaking of the financial sector, while the economic value of most securities is depressed, and with the wild valuation swings in the stock markets, the result is more data to process, move and store on a daily basis, all of which continues to place more demand on IT infrastructure resources including servers, storage, I/O networking, software, facilities and the people to support them.

Dow Jones Trading Activity Volume
Dow Jones Trading Activity Volume (Courtesy of data360.org)

For example, the amount of Dow Jones trading activity is on a logarithmic upward trend curve in the example chart from data360.org which means more transactions selling and buying. The result of more transactions is that there are also an increase in the number of back-office functions for settlement, tracking, surveillance, customer inquiry and reporting among others activities. This means that more I/Os are generated with data to be moved, processed, replicated, backed-up with additional downstream activity and processing.

Shifting gears, same things with telephone and in particular cell phone traffic which indirectly relates on IT systems particular for support email and other messaging activity. Speaking of email, more and more emails are sent every day, granted many are spam, yet these all result in more activity as well as data.

What’s the point in all of this?

There is a common awareness among most IT professionals that there is more data generated and stored every year and that there is also an awareness of the increased threats and reliance upon data and information. However what’s either not as widely discussed is the increase in I/O and networking activity. That is, the space capacity often gets talked about, however, the I/O performance, response time, activity and data movement can be forgotten about or its importance to productivity diminished. So the point is, keep performance, response time, and latency in focus as well as IOPS and bandwidth when looking at, and planning IT infrastructure to avoid data center bottlenecks.

Finally for now, what’s your take, is there a data and/or I/O networking recession, or is it business and activity as usual?

Ok, nuff said.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved