Data Center Infrastructure Management (DCIM) and IRM

StorageIO industry trends cloud, virtualization and big data

There are many business drivers and technology reasons for adopting data center infrastructure management (DCIM) and infrastructure Resource Management (IRM) techniques, tools and best practices. Today’s agile data centers need updated management systems, tools, and best practices that allow organizations to plan, run at a low-cost, and analyze for workflow improvement. After all, there is no such thing as an information recession driving the need to move process and store more data. With budget and other constraints, organizations need to be able to stretch available resources further while reducing costs including for physical space and energy consumption.

The business value proposition of DCIM and IRM includes:

DCIM, Data Center, Cloud and storage management figure

Data Center Infrastructure Management or DCIM also known as IRM has as their names describe a focus around management resources in the data center or information factory. IT resources include physical floor and cabinet space, power and cooling, networks and cabling, physical (and virtual) servers and storage, other hardware and software management tools. For some organizations, DCIM will have a more facilities oriented view focusing on physical floor space, power and cooling. Other organizations will have a converged view crossing hardware, software, facilities along with how those are used to effectively deliver information services in a cost-effective way.

Common to all DCIM and IRM practices are metrics and measurements along with other related information of available resources for gaining situational awareness. Situational awareness enables visibility into what resources exist, how they are configured and being used, by what applications, their performance, availability, capacity and economic effectiveness (PACE) to deliver a given level of service. In other words, DCIM enabled with metrics and measurements that matter allow you to avoid flying blind to make prompt and effective decisions.

DCIM, Data Center and Cloud Metrics Figure

DCIM comprises the following:

  • Facilities, power (primary and standby, distribution), cooling, floor space
  • Resource planning, management, asset and resource tracking
  • Hardware (servers, storage, networking)
  • Software (virtualization, operating systems, applications, tools)
  • People, processes, policies and best practices for management operations
  • Metrics and measurements for analytics and insight (situational awareness)

The evolving DCIM model is around elasticity, multi-tenant, scalability, flexibility, and is metered and service-oriented. Service-oriented, means a combination of being able to rapidly give new services while keeping customer experience and satisfaction in mind. Also part of being focused on the customer is to enable organizations to be competitive with outside service offerings while focusing on being more productive and economic efficient.

DCIM, Data Center and Cloud E2E management figure

While specific technology domain areas or groups may be focused on their respective areas, interdependencies across IT resource areas are a matter of fact for efficient virtual data centers. For example, provisioning a virtual server relies on configuration and security of the virtual environment, physical servers, storage and networks along with associated software and facility related resources.

You can read more about DCIM, ITSM and IRM in this white paper that I did, as well as in my books Cloud and Virtual Data Storage Networking (CRC Press) and The Green and Virtual Data Center (CRC Press).

Ok, nuff said, for now.

Cheers gs

Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

twitter @storageio

All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

Dell is buying Quest software, not the phone company Qwest

Dell Storage Customer Advisory Panel (CAP)

For those not familiar with Quest, they are a software company not to be confused with the telephone communications company formerly known as Qwest (aka now known as centurylink).

Both Dell and Quest have been on software related acquisition initiatives that past few years with Quest having purchased vKernel, Vizoncore (vRanger virtualization backup), BakBone (who had acquire Alavarii and Asempra) for traditional backup and data protection among others. Not to be out done, as well as purchasing Quest, Dell has also more recently bought Appassure (Disclosure: StorageIOblog site sponsor) for data protection, Sonicwall and Wyse in addition to some other recent purchases (ASAP, Boomi, Compellent, Exanet, EqualLogic, Force10, InsightOne, KACE, Ocarina, Perot, RNA and Scalent among others).

What does this mean?
Dell is expanding the scope of their business with more products (hardware, software), solution bundles, services and channel partnering opportunities Some of the software tools and focus areas that Quest brings to the Dell table or portfolio include:

Database management (Oracle, SQLserver)
Data protection (virtual and physical backup, replication, bc, dr)
Performance monitoring (DCIM and IRM) of applications and infrastructure
User workspace management (application delivery)
Windows server management (migrate and manage, AD, exchange, sharepoint)
Identify and access management (security, compliance, privacy)

What does Dell get by spending over $2B USD on quest?

  • Additional software titles or product
  • More software developers for their Software group
  • Sales people to help promote, partner and sell software solutions
  • Create demand pull for other Dell products and services via software
  • Increase its partner reach via existing Quest VARs and business partners
  • Extend the size of the Dell software and intellectual property (IP) portfolio
  • New revenue streams that compliment existing products and lines of business
  • Potential for better rate of return on some of its $12B USD in cash or equivalence

    Is this a good move for Dell?
    Yes for the above reasons

  • Is there a warning to this for Dell?
    Yes, they need to execute, keep the Quest team focused along with their other teams on the respective partners, products and market opportunities while expanding into new areas. Dell needs to also leverage Quest to further its cause in creating trust, confidence and strategic relationships with channel partners to reach new markets in different geographies. In addition, Dell needs to articulate its strategy and positioning of the various solutions to avoid products being perceived as competing vs. complimenting each other.

    Additional Dell related links:
    Dell Storage Customer Advisory Panel (CAP)
    Dell Storage Forum 2011 revisited
    Dude, is Dell doing a disk deal again with Compellent?
    Data footprint reduction (Part 2): Dell, IBM, Ocarina and Storwize
    Post Holiday IT Shopping Bargains, Dell Buying Exanet?
    Dell Will Buy Someone, However Not Brocade (At least for now)

    Ok, nuff said for now

    Cheers Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Spring (May) 2012 StorageIO news letter

    StorageIO News Letter Image
    Spring (May) 2012 News letter

    Welcome to the Spring (May) 2012 edition of the Server and StorageIO Group (StorageIO) news letter. This follows the Fall (December) 2011 edition.

    You can get access to this news letter via various social media venues (some are shown below) in addition to StorageIO web sites and subscriptions.

    Click on the following links to view the Spring May 2012 edition as an HTML or PDF or, to go to the news letter page to view previous editions.

    You can subscribe to the news letter by clicking here.

    Enjoy this edition of the StorageIO newsletter, let me know your comments and feedback.

    Nuff said for now

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Part IV: PureSystems, something old, something new, something from big blue

    This is the fourth in a five-part series around the recent IBM PureSystems announcements. You can view the earlier post here, and the next post here.

    So what does this mean for IBM Business Partners (BPs) and ISVs?
    What could very well differentiate IBM PureSystems from those of other competitors is to take what their partner NetApp has done with FlexPods combing third-party applications from Microsoft and SAP among others and take it to the next level. Similar to what helped make EMC Centera a success (or at least sell a lot of them) was inclusion and leveraging third-party ISVs and BPs  to add value. Compared to other vendors with object based or content accessible storage (CAS) or online archive platforms that focused on the technology feature, function speeds and feeds, EMC realized the key was getting ISVs to support so that BPs and their own direct sales force could sell the solution.

    With PureSystems, IBM is revisiting what they have done in the past which if offer bundled solutions providing incentives for ISVs to support and BPs to sell the IBM brand solution. EMC took an early step with including VMware with their Vblock combing server, storage, networking and software with NetApp taking the next step adding SAP, Microsoft and other applications. Dell, HP, Oracle and others are following suit so it only makes sense that IBM returns to its roots leveraging its DNA to reach out and get their ISVs who are now, have been in the past, or are new opportunities to be on board.

    IBM is throwing its resources including their innovation centers for training around the world where business partners can get the knowledge and technical support they need. In other words, workshops or seminars on how to sell deploy and setting up of these systems, application and customer testing or proof of concepts and things one would expect out of IBM for such an initiative. In addition to technology and sales training along with marketing support, IBM is making their financing capabilities available to help customers as well as offer incentives to their business partners to simplify acquisitions.

    So what buzzword bingo topics and themes did IBM address with this announcement:
    IBM did a fantastic job in terms of knocking the ball out of the park with this announcement pertaining buzzword bingo and deserves an atta boy or atta girl!

    So what about how this will affect sales of Bladecenters  or other systems?
    If all IBM and their BPs do are, encroach on existing systems sales to circle the wagons and protect the installed base, which would be one thing. However if IBM and their BPs can use the new packaging and model approach to reestablish customers and partnerships, or open and expand into new adjacent markets, then the net differences should be more Bladecenters (excuse me, PureFlex) being sold.

    So what will this cost?
    IBM is citing entry PureSystems Express models starting at around $100,000 USD for base systems with others starting at around $200,000 and $300,000 expandable into larger configurations and budgets. Note that like airlines that advertise a low airfare and then you get to pay extra for peanuts, drinks, extra bag space, changes to reservations and so forth, look at these and related systems not just for the first starting price, also for expansion costs over different time periods. Contact IBM, your BP or ISV to find out what one of these systems will do for and cost you.

    So what about VARs and IBM business partners (BPs)?
    This could be a boon for those BPs and ISVs  that had previously sold their software solutions bundled with IBM hardware platforms who were being challenged by other converged solution stacks or were being forced to unbundled. This will also allow those business partners to compete on par with other converged solutions or continue selling the pieces of what they are familiar with however under a new umbrellas. Of course, pricing will be a focus and concern for some who will want to see what added value exists vs. acquiring the various components. This also means that IBM will have to make incentives available for their partners to make a living while also allowing their customers to afford solutions and maximize their return on innovation (the new ROI) and enablement.

    Click here to view the next post in this series, ok nuff said for now.

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Part V: PureSystems, something old, something new, something from big blue

    This is the fifth in a five-part series around the recent IBM PureSystems announcements. You can view the earlier post here.

    So what about vendor or technology lock in?
    So who is responsible for vendor or technology lock in? When I was working in IT organizations, (e.g. what vendors call the customer) the thinking was vendors are responsible for lock in. Later when I worked for different vendors (manufactures and VARs) the thinking was lock in is what was caused by the competition. More recently I’m of the mind set that vendor lock in is a shared responsibility issue and topic. I’m sure some marketing wiz or sales type will be happy to explain the subtle differences of how their solution does not cause lock in.

    Vendor lock in can be a shared responsibility. Generally speaking, lock in, stickiness and account control are essentially the same, or at least strive to get similar results. For example, vendor lock in too some has a negative stigma. However vendor stickiness may be a new term, perhaps even sounding cool thus it is not a concern. Remember the Mary Poppins song a spoon full of sugar makes the medicine go down? In other words, sometimes changing and using a different term such as sticky vs. vendor lock in helps make the situation taste better.

    So what should you do?
    Take a closer look if you are considering converged infrastructures, cloud or data centers in a box, turnkey application or information services deployment platforms. Likewise, if you are looking at specific technologies such as those from Cisco UCS, Dell vStart, EMC Vblock (or via VCE), HP, NetApp FlexPod or Oracle (ExaLogic, ExaData, etc) among others, also check out the IBM PureSystems (Flex and PureApplication). Compare and contrast these converged solutions with your traditional procurement and deployment modes including cost of acquiring hardware, software, ongoing maintenance or service fees along with value or benefit of bundled tools. There may be a higher cost for converged systems in some scenarios, however compare on the value and benefit derived vs. doing the integration yourself.

    Compare and contrast how converged solutions enable, however also consider what constraints exists in terms of flexibility to reconfigure in the future or make other changes. For example as part of integration, does a solution take a lowest common denominator approach to software and firmware revisions for compatibility that may lag behind what you can apply to standalone components. Also, compare and contrast various reference architectures with different solution bundles or packages.

    Most importantly compare and evaluate the solutions on their ability to meet and exceed your base requirements while adding value and enabling return on innovation while also being cost-effective. Do not be scared of these bundled solutions; however do your homework to make informed decisions including overcoming any concerns of lock in or future costs and fees. While these types of solutions are cool or interesting from a technology perspective and can streamline acquisition and deployment, make sure that there is a business benefit that can be addressed as well as enablement of new capabilities.

    So what does this all mean?
    Congratulations to IBM with their PureSystems for leveraging their DNA and roots bundling what had been unbundled before cloud and stacks were popular and trendy. IBM has done a good job of talking vision and strategy along lines of converged and dynamic, elastic and smart, clouds and other themes for past couple of years while selling the pieces as parts of solutions or ala carte or packaged by their ISVs and business partners.

    What will be interesting to see is if bladecenter customers shift to buying PureFlex, which should be an immediate boost to give proof points of adoption, while essentially up selling what was previously available. However, more interesting will be to see if net overall new customers and footprints are sold as opposed to simply selling a newer and enhanced version of previous components.

    In other words will IBM be able to keep up their focus and execution where they have sold the previous available components, while also holding onto current ISV and BP footprint sales and perhaps enabling those partners to recapture some hardware and solution sales that had been unbundled (e.g. ISV software sold separate of IBM platforms) and move into new adjacent markets.

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) – Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Ok, so what is next, lets see how this unfolds for IBM and their partners.

    Nuff said for now.

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Part III: PureSystems, something old, something new, something from big blue

    This is the third in a five-part series around the recent IBM PureSystems announcements. You can view the earlier post here, and the next post here.

    So what about the IBM Virtual Appliance Factory?
    Where PureFlex and PureApplication (PureSystems) are the platforms or vehicles for enabling your journey to efficient and effective information services delivery, and PureSystem centre (or center for those of you in the US) is the portal or information center, the IBM Virtual Appliance Factory (VAF) is a collection of tools, technologies, processes and methodologies. The VAF  helps developers or ISVs to prepackage applications or solutions for deployment into Kernel Virtual Machine (KVM) on Intel and IBM PowerVM  virtualized environments that are also supported by PureFlex and PureApplication  systems.

    VAF technologies include Distributed Management Task Force (DMTF) Open Virtual Alliance (OVA) Open Virtualization Format (OVF) along with other tools for combing operating systems (OS), middleware and solution software into a delivery package or a virtual appliance that can be deployed into cloud and virtualized environments. Benefits include reducing complexity of working logical partions (LPAR) and VM configuration, abstraction and portability for deployment or movement from private to public environments. Net result should be less complexity lowering costs while reducing mean time to install and deploy. Here is a link to learn more about VAF and its capabilities and how to get started.

    So what does cloud ready mean?
    IBM is touting cloud ready capability in the context of rapid out of the box, ease of deployment and use as well as easy to acquire. This is in line with what others are doing with converged server, storage, networking, hardware, software and hypervisor solutions. IBM is also touting that they are using the same public available products as what they use in their own public services SmartCloud offerings.

    So what is scale in vs. scale up, scale out or scale within?
    Traditional thinking is that scaling refers to increasing capacity. Scaling also means increasing performance, availability, functionality with stability. Scaling with stability means that as performance, availability, capacity or other features are increased problems are not introduced or complexity is not increased. For example, scaling with stability for performance should not result in loss of availability or capacity, capacity increase should not be at the cost of performance or availability, should not cost performance or capacity and management tools should work for you, instead of you working for them.

    Scaling up and scaling out have been used to describe scaling performance, availability, capacity and other attributes beyond the limits of a single system, box or cabinet. For example clustered, cloud, grid and other approaches refer to scaling out or horizontally across different physical resources. Scaling up or scaling vertically means scaling within in a system using faster, denser technologies doing more in the same footprint. HDS announced a while back what they refer to 3D scaling which embraces the above notions of scaling up, out and within across different dimensions. IBM is building on that by emphasizing scaling leveraging faster, denser components such as Power7 and Intel processors to scale within the box or system or node, which can also be scaled out using enhanced networking from IBM and their partners.

    So what about backup/restore, BC, DR and general data protection?
    I would expect IBM to step up and talk about how they can leverage their data protection and associated management toolsets, technologies and products. IBM has the components (hardware, software) already for backup/restore, BC, DR, data protection and security along with associated service offerings. One would expect IBM to not only come out with a backup, restore, BC, DR and archiving solution or version, as well as ones for archiving or data preservation, compliance appliance variants as well as related themes. We know that IBM has the pieces, people, process and practices, let us see if IBM has learned from their competitors who may have missed data protection messaging opportunities. Sometimes what is assumed to be understood does not get discussed, however often what is assumed and is not understood should be discussed, hence, let us see if IBM does more than say oh yes, we have those capabilities and products too.

    So what do these have compared to others who are doing similar things?
    Different vendors have taken various approaches for bringing converged products or solutions to the market place. Not surprising, storage centric vendors EMC and NetApp have partnered with Cisco for servers (compute). Where Cisco was known for networking having more recently moved into compute servers, EMC and NetApp are known for storage and moving into converged space with servers. Since EMC and NetApp often compete with storage solutions offerings from traditional server vendors Dell, HP, IBM and Oracle among others, and now Cisco is also competing with those same server vendors it has previously partnered with for networking thus it makes sense for Cisco, EMC and NetApp to partner.

    While EMC owns a large share of VMware, they do also support Microsoft and other partners including Citrix. NetApp followed EMC into the converged space partnering with Cisco for compute and networking adding their own storage along with supporting hypervisors from Citrix, Microsoft and VMware along with third-party ISVs including Microsoft and SAP among others. Dell has evolved from reference architectures to products called vStart that leverage their own technologies along with those of partners.

    A challenge for Dell however is that vStart  sounds more like a service offering as opposed to a product that they or their VARs and business partners can sell and add value around. HP is also in the converged game as is Oracle among others. With PureSystems IBM is building on what their competitors and in some cases partners are doing by adding and messaging more around the many ISVs and applications that are part of the PureSystems initiative. Rest assured, there is more to PureSystems than simply some new marketing, press releases, videos and talking about partners and ISVs. The following table provides a basic high level comparison of what different vendors are doing or working towards and is not intended to be a comprehensive review.

    Who

    What

    Server

    Storage

    Network

    Software

    Other comments

    Cisco

    UCS

    Cisco

    Partner

    Cisco

    Cisco and Partners

    Various hypervisors and OS

    Dell

    vStart

    Dell

    Dell

    Dell and Partners

    Dell and partners

    Various hypervisors, OS and bundles

    EMC
    VCE

    Vblock VSPEX

    Cisco

    EMC

    Cisco and partners

    EMC, Cisco and partners

    Various hypervisors, OS and bundles, VSPEX adds more partner solution bundles

    HP

    Converged

    HP

    HP

    HP and partners

    HP and partners

    Various hypervisors, OS and bundles

    IBM

    PureFlex

    IBM

    IBM

    IBM and partners

    IBM and partners

    Various hypervisors, OS and bundles adding more ISV partners

    NetApp

    FlexPod

    Cisco

    NetApp

    Cisco and partners

    NetApp, Cisco and partners

    Various hypervisors, OS and bundles for SAP, Microsoft among others

    Oracle

    ExaLogic (Exadata  database)

    Oracle

    Oracle

    Partners

    Oracle and partners

    Various Oracle software tools and technologies

    So what took IBM so long compared to others?
    Good question, what is the saying? Rome was not built-in a day!

    Click here to view the next post in this series, ok, nuff said for now.

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Part II: PureSystems, something old, something new, something from big blue

    This is the second in a five-part series around the recent IBM PureSystems announcements. You can view the earlier post here, and the next post here.

    So what are the speeds and feeds of a PureFlex system?
    The components that make up the PureFlex line include:

    • IBM management node (server with management software tools).
    • 10Gb Ethernet (LAN) switch, adapters and associated cabling.
    • IBM V7000 virtual storage (also see here and here).
    • Dual 8GFC (8Gb Fibre Channel) SAN switches and adapters.
    • Servers with either x86 xSeries using for example Intel Sandy Bridge EP 2.6 GHz 8 core processors, or IBMs Power7 based pSeries for AIX. Note that IBM with their blade center systems (now rebadged as part of being PureSystems) support various IO and networking interfaces include SAS, Ethernet, Fibre Channel (FC), Fibre Channel over Ethernet (FCoE), and InfiniBand using adapters and switches from various partners.
    • Virtual machine (VM) hypervisors such as Microsoft Hyper V and VMware vSphere/ESX among others. In addition to x86 based hypervisors or kernel virtual machines (KVM), IBM also supports its own virtual technology found in Power7 based systems. Check IBM support matrix for specific configurations and current offerings.
    • Optional middleware such as IBM WebSphere.

    Read more speeds and feeds at the various IBM sites including on Tony Pearson’s blog site.

    So what is IBM PureApplication System?
    This builds off and on PureFlex systems as a foundation for deploying various software stacks to deliver traditional IT applications or cloud Platform as a Service (PaaS) or Software as a Service (SaaS) and Application as a Service (AaaS) models. For example cloud or web stacks, java, database, analytics or other applications with buzzwords of elastic, scalable, repeatable, self-service, rapid provisioning, resilient, multi tenant and secure among others. Note that if are playing or into Buzzword bingo, go ahead and say Bingo when you are ready as IBM has a winner in this category.

    So what is the difference between PureFlex and PureApplication systems?
    PureApplication systems leverage PureFlex technologies adding extra tools and functionality for cloud like application functionality delivery.

    So what is IBM PureSystems Centre?
    It is a portal or central place where IBM and their business partner solutions pertaining to PureApplication and PureFlex systems can be accessed for including information for first installation support along with maintenance and upgrades. At launch, IBM is touting more than 150 solutions or applications that are available or qualified for deployment on PureApplication and PureFlex systems. In addition, IBM Patterns (aka templates) can also be accessed via this venue. Examples of application or independent software vendor (ISV) developed solutions for banking, education, financial, government, healthcare and insurance can be found at the PureSystems Centre portal (here, here and here).

    So what part of this is a service and what is a product?
    Other than the PureSystem center, which is a web portal for accessing information and technologies, PureFlex and PureApplication along with Virtual Appliance Factory are products or solutions that can be bought from IBM or their business partners. In addition, IBM business partners or third parties can also use these solutions housed in their own, a customer, or third-party facility for delivering managed service provided (MSP) capabilities, along with other PaaS and SaaS or AaaS type functionalities. In other words, these solutions can be bought or leased by IT and other organizations for their own use in a traditional IT deployment model, private, hybrid or public cloud model.

    Another option is for service providers to acquire these solutions for use in developing and delivering their own public and private or hybrid services. IBM is providing the hard product (hardware and software) that enables your return on innovation (the new ROI) to create and deliver your own soft product (services and experiences) consumed by those who use those capabilities. In addition to traditional financial quantitative return on investment (traditional ROI) and total cost of ownership (TCO), the new ROI complements those by adding a qualitative aspect. Your return on innovation will be dependent on what you are capable of doing that enables your customers or clients to be productive or creative. For example enabling your customers or clients to boost productivity, remove complexity and cost while maintaining or enhancing Quality of Service (QoS), service level objectives (SLOs) and service level agreements (SLAs) in addition to supporting growth by using a given set of hard products. Thus, your soft product is a function of your return on innovation and vise versa.

    Note that in this context, not to be confused with hardware and software, hard product are those technologies including hardware, software and services that are obtained and deployed as a soft product. A soft product in this context does not refer to software, rather the combination of hard products plus your own developed or separately obtained software and tools along with best practices and usage models. Thus, two organizations can use the same hard products and deliver separate soft products with different attributes and characteristics including cost, flexibility and customer experience.

    So what is a Pattern of Expertise?
    Combines operational know how experience and knowledge about common infrastructure resource management (IRM), data center infrastructure management (DCIM) and other commonly repeatable related process, practices and workflows including provisioning. Common patterns of activity and expertise for routine or other time-consuming tasks, which some might refer to as templates or workflows enable policy driven based automation. For example, IBM cites recurring time-consuming tasks that lend themselves to being automated such as provisioning, configuration, and upgrades and associated IRM, DCIM and data protection, storage and application management activities. Automation software tools are included as part of the PureSystems with patterns being downloadable as packages for common tasks and applications found at the IBM PureSystem center.

    At announcement, there are three types or categories of patterns:

    • IBM patterns: Factory created and supplied with the systems based on experiences IBM has derived from various managers, engineers and technologist for automating common tasks including configuration, deployment and application upgrades and maintenance. The aim is to cut the amount of time and intervention for deployment of applications and other common functions enabling IT staff to be more productive and address other needs.
    • ISV patterns: These leverage experience and knowledge from ISVs partnered with IBM, which at time of launch numbers over 125 vendors offering certified PureSystems Ready applications. The benefit and objective are to cut the time and complexity associated with procuring (e.g. purchasing), deploying and managing third-party ISV software. Downloadable patterns packages can be found at the IBM PureSystem center.
    • Customer patterns: Enables customers to collect and package their own knowledge, processes, rules, policies and best practices into patterns for automation. In addition to collecting knowledge for acquisition, configuration, day to day management and troubleshooting, these patterns can facility automation of tasks to ease on boarding of new staff employees or contractors. In addition, these patterns or templates capture workflows for automation enabling shorter deployment times of systems and applications into locations where skill sets do not exist.

    Here is a link to some additional information about patterns on the IBM developerWorks site.

    Click here to view the next post in this series, ok, nuff said for now.

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    Part I: PureSystems, something old, something new, something from big blue

    This is the first in a five-part series around the recent IBM PureSystems announcements. You can view the next post here.

    For a certain generation of IBM faithful or followers the recently announced PureFlex and PureApplication systems might give a sense of DejaVu perhaps even causing some to wonder if they just woke up from a long Rip Van Winkle type nap.

    Yet for another generation who may not yet be future IBM followers, fans, partners or customers, there could be a sense of something new and revolutionary with the PureFlex and PureApplication systems (twitter @ibmpuresystems).

    In between those two groups, exist others who are either scratching their heads or reinvigorated with enthusiasm to get out and be able to discuss opportunities around little data (traditional and transactional) and big data, servers, virtualized, converged infrastructure, dynamic data centers, private clouds, ITaaS, SaaS and AaaS, PaaS, IaaS and other related themes or buzzword bingo topics.

    Let us dig a little deeper and look at some So What types of questions and industry trends perspectives comments around what IBM has announced.

    So what did IBM announce?
    IBM announced PureSystems including:

    • PureFlex systems, products and technologies
    • PureApplication systems
    • PureSystems Centre

    You can think of IBM PureSystems and Flex Systems Products and technology as a:

    • Private cloud or turnkey solution bundle solution
    • Platform deploying public or hybrid clouds
    • Data center in a box or converged and dynamic system
    • ITaaS or SaaS/AaaS or PaaS or IaaS or Cloud in a box
    • Rackem stack and package them type solution

    So what is an IBM PureFlex System and what is IBM using?
    It is a factory integrated data and compute infrastructure in a cabinet combing cloud, virtualization, servers, data and storage networking capabilities. The IBM PureFlex system is comprised of various IBM and products and technologies (hardware, software and services) optimized with management across physical and virtual resources (servers, storage (V7000), networking, operating systems, hypervisors and tools).

    PureFlex includes automation and optimization technologies along with what IBM is referring to as patterns of expertise or what you might relate to as templates. Support for various hypervisors and management integration along with application and operating system support by leveraging IBM xSeries (x86 such as Intel) and pSeries (Power7) based processors for compute. Storage is the IBM V7000 (here and here) with networking and connectivity via IBM and their partners. The solution is capable of supporting traditional, virtual and cloud deployment models as well as platform for deploying Infrastructure as a Service (IaaS) on a public, managed service provider (MSP), hosting or private basis.

    Click here to view the next post in this series, ok nuff said for now.

    Here are some links to learn more:
    Various IBM Redbooks and related content
    The blame game: Does cloud storage result in data loss?
    What do you need when its time to buy a new server?
    2012 industry trends perspectives and commentary (predictions)
    Convergence: People, Processes, Policies and Products
    Buzzword Bingo and Acronym Update V2.011
    The function of XaaS(X) Pick a letter
    Hard product vs. soft product
    Buzzword Bingo and Acronym Update V2.011
    Part I: PureSystems, something old, something new, something from big blue
    Part II: PureSystems, something old, something new, something from big blue
    Part III: PureSystems, something old, something new, something from big blue
    Part IV: PureSystems, something old, something new, something from big blue
    Part V: PureSystems, something old, something new, something from big blue
    Cloud and Virtual Data Storage Networking

    Cheers
    Gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved

    NetApp and Akorri: An E2E cross technology domain SRA play

    The other day NetApp announced that it was planning on doing another acquisition following on their recent purchase of Bycast (policy based storage and management software).

    This time, NetApp is doing yet another software acquisition of Infrastructure Resource Management (IRM) as well as End to End (E2E) cross technology domain management and Storage or Systems Resource Analysis (SRA) startup Akorri which also builds on its past acquisition of SRA solution Onaro.

    Is this a good move by NetApp?

    Assuming they got a good price, yes, this has very potential for NetApp assuming they can assimilate the solution as well as articulate where it fits complimenting its other management tools including SANscreen (aka Onaro).

    Is Akorii a good product?

    Yes, most of the customers and var partners of Akorri that I talk to have great things to say and having looked into the technology, it has lots of good potential for NetApp. However, there is a common theme around Akorri that has been its high price, something that was also heard from Onaro customers before NetApp did that acquisition. If NetApp can leverage its direct as well as partner touch to reduce the cost of sale for Akorri as well as rationalize the pricing or at least better articulate the value proposition to make it a must have vs nice to have, they can do well.

    The importance of E2E awareness of IT resources across different technology domains (or focus areas) is that you can not effectively manage what you do not have timely access or visibility into. Hence the theme of session being You cannot effectively manage what you do not know about in a timely manner. I recently did a couple of Industry Trends and Perspectives webcast events around the topic and themes of End to End (E2E) awareness and cross domain (or technology) management insight for cloud, virtual and other abstracted as well as physical IT environments.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    E2E Awareness and insight for IT environments

    I recently did a couple of Industry Trends and Perspectives webcast events around the topic and themes of End to End (E2E) awareness and cross domain (or technology) management insight for cloud, virtual and other abstracted as well as physical IT environments.

    The importance of E2E awareness of IT resources across different technology domains (or focus areas) is that you can not effectively manage what you do not have timely access or visibility into. Hence the theme of session being You cannot effectively manage what you do not know about in a timely manner.

    Here is the abstract for the webcast:

    Virtualization, clouds and other forms of abstraction help IT organizations enable flexible and scalable services delivery. While abstraction of underlying resources simplifies services delivery from an IT customers perspective, additional layers of technology along with interdependencies still need to be tracked as well as managed.  A key enabler for IT organizations is having end to end (E2E) situational awareness of available resources and how they are being used. By having timely situational awareness across various technology domains, IT organizations gain insight into how resources can be more effectively deployed in an efficient manner.

    Join independent IT industry analyst, author and blogger Greg Schulz as he looks at common challenges as well as opportunities for leveraging E2E situational awareness to remove blind spots from efficient effective IT services delivery. Greg will look several scenarios including among others cost reduction, maximize resource usage, shrink migration and data consolidation times for cloud, virtual and traditional IT environments while maintaining or enhancing IT services delivery.

    If you are interested in IT Infrastructure Resource Management (IRM) of servers, storage, IO networking, virtualization, cloud, backup or restore, optimization as well as cloud or legacy environments and metrics, I invite you to view the following web cast.

    E2E cross domain awareness webcast

    Click on the above image to access the BrightTalk web cast from their recent Virtualization Summit series (may require registration)

    If you are interested, here is a link to a previous post I did on E2E management, SRA (systems or storage resource analysis) and management insight along with a recent related white paper sponsored by SANpulse that you can access here.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Who is responsible for vendor lockin?

    Who is responsible for vendor lockin?

    data infrastructure server storage I/O vendor lockin

    Updated 1/21/2018

    Who is responsible for vendor lockin?

    Is vendor lockin caused by vendors, their partners or by customers?

    In my opinion vendor lockin can be from any or all of the above.

    What is vendor lockin

    Vendor lockin is a situation where a customer becomes dependent or locked in by choice or other circumstances to a particular supplier or technology.

    What is the difference between vendor lockin, account control and stickiness?

    Im sure some marketing wiz or sales type will be happy to explain the subtle differences. Generally speaking, lockin, stickiness and account control are essentially the same, or at least strive to obtain similar results. For example, vendor lockin too some has a negative stigma. However vendor stickiness may be a new term, perhaps even sounding cool thus it is not a concern. Remember the Mary Poppins song a spoon full of sugar makes the medicine go down? In other words sometimes changing and using a different term such as sticky vs vendor lockin helps make the situation taste better.

    Is vendor lockin or stickiness a bad thing?

    No, not necessarily, particularly if you the customer are aware and still in control of your environment.

    I have had different views of vendor lockin over the years.

    These have varied from when I was a customer working in IT organizations or being a vendor and later as an advisory analyst consultant. Even as a customer, I had different views of lockin which varied depending upon the situation. In some cases lockin was a result of upper management having their favorite vendor which meant when a change occurred further up the ranks, sometimes vendor lockin would shift as well. On the other hand, I also worked in IT environments where we had multiple vendors for different technologies to maintain competition across suppliers.

    As a vendor, I was involved with customer sites that were best of breed while others were aligned around a single or few vendors. Some were aligned around technologies from the vendors I worked for and others were aligned with someone elses technology. In some cases as a vendor we were locked out of an account until there was a change of management or mandates at those sites. In other cases where lock out occurred, once our product was OEMd or resold by an incumbent vendor, the lockout ended.

    Some vendors do a better job of establishing lockin, account management, account control or stickiness than compared to others. Some vendors may try to lock a customer in and thus there is perception that vendors lock customers in. Likewise, there is a perception that vendor lockin only occurs with the largest vendors however I have seen this also occur with smaller or niche vendors who gain control of their customers keeping larger or other vendors out.

    Sweet, sticky Sue Bee Honey

    Vendor lockin or stickiness is not always the result of the vendor, var, consultant or service provider pushing a particular technology, product or service. Customers can allow or enable vendor lockin as well, either by intent via alliances to drive some business initiative or accidentally by giving up account control management. Consequently vendor lockin is not a bad thing if it brings mutual benefit to the suppler and consumer.

    On the other hand, if lockin causes hardship on the consumer while only benefiting the supplier, than it can be a bad thing for the customer.

    Do some technologies lend themselves more to vendor lockin vs others?

    Yes, some technologies lend themselves more to stickiness or lockin then others. For example, often big ticket or expensive hardware are seen as being vulnerable to vendor lockin along with other hardware items however software is where I have seen a lot of stickiness or lockin around.

    However what about virtualization solutions after all the golden rule of virtualization is whoever controls the virtualization (hardware, software or services) controls the gold. This means that vendor lockin could be around a particular hypervisor or associated management tools.

    How about bundled solutions or what are now called integrated vendor technology stacks including PODs (here or here) or vBlocks among others? How about databases, do they enable or facilitate vendor lockin? Perhaps, just like virtualization or operating systems or networking technology, storage system, data protection or other solutions, if you let the technology or vendor manage you, then you enable vendor lockin.

    Where can vendor lockin or stickiness occur?

    Application software, databases, data or information tools, messaging or collaboration, infrastructure resource management (IRM) tools ranging from security to backup to hypervisors and operating systems to email. Lets not forget about hardware which has become more interoperable from servers, storage and networks to integrated marketing or alliance stacks.

    Another opportunity for lockin or stickiness can be in the form of drivers, agents or software shims where you become hooked on a feature functionality that then drives future decisions. In other words, lockin can occur in different locations both in traditional IT as well as via managed services, virtualization or cloud environments if you let it occur.

     

    Keep these thoughts in mind:

    • Customers need to manage their resources and suppliers
    • Technology and their providers should work for you the customer, not the other way around
    • Technology providers conversely need to get closer to influence customer thinking
    • There can be cost with single vendor or technology sourcing due to loss of competition
    • There can be a cost associated with best of breed or functioning as your own integrator
    • There is a cost switching from vendors and or their technology to keep in mind
    • Managing your vendors or suppliers may be easier than managing your upper management
    • Vendors sales remove barriers so they can sell and setting barriers for others
    • Virtualization and cloud can be both a source for lockin as well as a tool to help prevent it
    • As a customer, if lockin provides benefits than it can be a good thing for all involved

    Additional learning experiences along with common questions (and answers), as well as tips can be found in Software Defined Data Infrastructure Essentials book.

    Software Defined Data Infrastructure Essentials Book SDDC

    What This All Means

    Ultimately, its up to the customer to manage their environment and thus have a say if they will allow vendor lockin. Granted, upper management may be the source of the lockin and not surprisingly is where some vendors will want to focus their attention directly, or via influence of high level management consultants.

    So while a vendors solution may appear to be a locked in solution, it does not become a lockin issue or problem until a customer lets or allows it to be a lockin or sticky situation.

    What is your take on vendor lockin? Cast your vote and see results in the following polls.

    Is vendor lockin a good or bad thing?

    Who is responsible for managing vendor lockin

    Where is most common form or concern of vendor lockin

    Ok, nuff said, for now.

    Gs

    Greg Schulz – Microsoft MVP Cloud and Data Center Management, VMware vExpert 2010-2017 (vSAN and vCloud). Author of Software Defined Data Infrastructure Essentials (CRC Press), as well as Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press), Resilient Storage Networks (Elsevier) and twitter @storageio. Courteous comments are welcome for consideration. First published on https://storageioblog.com any reproduction in whole, in part, with changes to content, without source attribution under title or without permission is forbidden.

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO. All Rights Reserved. StorageIO is a registered Trade Mark (TM) of Server StorageIO.

    End to End (E2E) Systems Resource Analysis (SRA) for Cloud and Virtual Environments

    A new StorageIO Industry Trends and Perspective (ITP) white paper titled “End to End (E2E) Systems Resource Analysis (SRA) for Cloud, Virtual and Abstracted Environments” is now available at www.storageio.com/reports compliments of SANpulse technologies.

    End to End (E2E) Systems Resource Analysis (SRA) for Virtual, Cloud and abstracted environments: Importance of Situational Awareness for Virtual and Abstracted Environments

    Abstract:
    Many organizations are in the planning phase or already executing initiatives moving their IT applications and data to abstracted, cloud (public or private) virtualized or other forms of efficient, effective dynamic operating environments. Others are in the process of exploring where, when, why and how to use various forms of abstraction techniques and technologies to address various issues. Issues include opportunities to leverage virtualization and abstraction techniques that enable IT agility, flexibility, resiliency and salability in a cost effective yet productive manner.

    An important need when moving to a cloud or virtualized dynamic environment is to have situational awareness of IT resources. This means having insight into how IT resources are being deployed to support business applications and to meet service objectives in a cost effective manner.

    Awareness of IT resource usage provides insight necessary for both tactical and strategic planning as well as decision making. Effective management requires insight into not only what resources are at hand but also how they are being used to decide where different applications and data should be placed to effectively meet business requirements.

    Learn more about the importance and opportunities associated with gaining situational awareness using E2E SRA for virtual, cloud and abstracted environments in this StorageIO Industry Trends and Perspective (ITP) white paper compliments of SANpulse technologies by clicking here.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    EMC VPLEX: Virtual Storage Redefined or Respun?

    In a flurry of announcements that coincide with EMCworld occurring in Boston this week of May 10 2010 EMC officially unveiled the Virtual Storage vision initiative (aka twitter hash tag of #emcvs) and initial VPLEX product. The Virtual Storage initiative was virtually previewed back in March (See my previous post here along with one from Stu Miniman (twitter @stu) of EMC here or here) and according to EMC the VPLEX product was made generally available (GA) back in April.

    The Virtual Storage vision and associated announcements consisted of:

    • Virtual Storage vision – Big picture  initiative view of what and how to enable private clouds
    • VPLEX architecture – Big picture view of federated data storage management and access
    • First VPLEX based product – Local and campus (Metro to about 100km) solutions
    • Glimpses of how the architecture will evolve with future products and enhancements


    Figure 1: EMC Virtual Storage and Virtual Server Vision and Big Pictures

    The Big Picture
    The EMC Virtual Storage vision (Figure 1) is the foundation of a private IT cloud which should enable characteristics including transparency, agility, flexibility, efficient, always on, resiliency, security, on demand and scalable. Think of it this way, EMC wants to enable and facilitate for storage what is being done by server virtualization hypervisor vendors including VMware (which happens to be owned by EMC), Microsoft HyperV and Citrix/Xen among others. That is, break down the physical barriers or constraints around storage similar to how virtual servers release applications and their operating systems from being tied to a physical server.

    While the current focus of desktop, server and storage virtualization has been focused on consolidation and cost avoidance, the next big wave or phase is life beyond consolidation where the emphasis expands to agility, flexibility, ease of use, transparency, and portability (Figure 2). In the next phase which puts an emphasis around enablement and doing more with what you have while enhancing business agility focus extends from how much can be consolidated or the number of virtual machines per physical machine to that of using virtualization for flexibility, transparency (read more here and here or watch here).


    Figure 2: Virtual Storage Big Picture

    That same trend will be happening with storage where the emphasis also expands from how much data can be squeezed or consolidated onto a given device to that of enabling flexibility and agility for load balancing, BC/DR, technology upgrades, maintenance and other routine Infrastructure Resource Management (IRM) tasks.

    For EMC, achieving this vision (both directly for storage, and indirectly for servers via their VMware subsidiary) is via local and distributed (metro and wide area) federation management of physical resources to support virtual data center operations. EMC building blocks for delivering this vision including VPLEX, data and storage management federation across EMC and third party products, FAST (fully automated storage tiering), SSD, data protection and data footprint reduction and data protection management products among others.

    Buzzword bingo aside (e.g. LAN, SAN, MAN, WAN, Pots and Pans) along with Automation, DWDM, Asynchronous, BC, BE or Back End, Cache coherency, Cache consistency, Chargeback, Cluster, db loss, DCB, Director, Distributed, DLM or Distributed Lock Management, DR, Foe or Fibre Channel over Ethernet, FE or Front End, Federated, FAST, Fibre Channel, Grid, HyperV, Hypervisor, IRM or Infrastructure Resource Management, I/O redirection, I/O shipping, Latency, Look aside, Metadata, Metrics, Public/Private Cloud, Read ahead, Replication, SAS, Shipping off to Boston, SRA, SRM, SSD, Stale Reads, Storage virtualization, Synchronization, Synchronous, Tiering, Virtual storage, VMware and Write through among many other possible candidates the big picture here is about enabling flexibility, agility, ease of deployment and management along with boosting resource usage effectiveness and presumably productivity on a local, metro and future global basis.


    Figure 3: EMC Storage Federation and Enabling Technology Big Picture

    The VPLEX Big Picture
    Some of the tenants of the VPLEX architecture (Figure 3) include a scale out cluster or grid design for local and distributed (metro and wide area) access where you can start small and evolve as needed in a predictable and deterministic manner.


    Figure 4: Generic Virtual Storage (Local SAN and MAN/WAN) and where VPLEX fits

    The VPLEX architecture is targeted towards enabling next generation data centers including private clouds where ease and transparency of data movement, access and agility are essential. VPLEX sits atop existing EMC and third party storage as a virtualization layer between physical or virtual servers and in theory, other storage systems that rely on underlying block storage. For example in theory a NAS (NFS, CIFS, and AFS) gateway, CAS content archiving or Object based storage system or purpose specific database machine could sit between actual application servers and VPLEX enabling multiple layers of flexibility and agility for larger environments.

    At the heart of the architecture is an engine running a highly distributed data caching algorithm that uses an approach where a minimal amount of data is sent to other nodes or members in the VPLEX environment to reduce overhead and latency (in theory boosting performance). For data consistency and integrity, a distributed cache coherency model is employed to protect against stale reads and writes along with load balancing, resource sharing and failover for high availability. A VPLEX environment consists of a federated management view across multiple VPLEX clusters including the ability to create a stretch volume that is accessible across multiple VPLEX clusters (Figure 5).


    Figure 5: EMC VPLEX Big Picture


    Figure 6: EMC VPLEX Local with 1 to 4 Engines

    Each VPLEX local cluster (Figure 6) is made up of 1 to 4 engines (Figure 7) per rack with each engine consisting of two directors each having 64GByte of cache, localized compute Intel processors, 16 Front End (FE) and 16 Back End (BE) Fibre Channel ports configured in a high availability (HA). Communications between the directors and engines is Fibre Channel based. Meta data is moved between the directors and engines in 4K blocks to maintain consistency and coherency. Components are fully redundant and include phone home support.


    Figure 7: EMC VPLEX Engine with redundant directors

    VPLEX initially host servers supported include VMware, Cisco UCS, Windows, Solaris, IBM AIX, HPUX and Linux along with EMC PowerPath and Windows multipath management drivers. Local server clusters supported include Symantec VCS, Microsoft MSCS and Oracle RAC along with various volume mangers. SAN fabric connectivity supported includes Brocade and Cisco as well as Legacy McData based products.

    VPLEX also supports cache (Figure 8 ) write thru to preserve underlying array based functionality and performance with 8,000 total virtualized LUNs per system. Note that underlying LUNs can be aggregated or simply passed through the VPLEX. Storage that attaches to the BE Fibre Channel ports include EMC Symmetrix VMAX and DMX along with CLARiiON CX and CX4. Third party storage supported includes HDS9000 and USPV/VM along with IBM DS8000 and others to be added as they are certified. In theory given that the VPLEX presents block based storage to hosts; one would also expect that NAS, CAS or other object based gateways and servers that rely on underlying block storage to also be supported in the future.


    Figure 8: VPLEX Architecture and Distributed Cache Overview

    Functionality that can be performed between the cluster nodes and engines with VPLEX include data migration and workload movement across different physical storage systems or sites along with shared access with read caching on a local and distributed basis. LUNS can also be pooled across different vendors underlying storage solutions that also retain their native feature functionality via VPLEX write thru caching.

    Reads from various servers can be resolved by any node or engine that checks their cache tables (Figure 8 ) to determine where to resolve the actual I/O operation from. Data integrity checks are also maintained to prevent stale reads or write operations from occurring. Actual meta data communications between nodes is very small to enable state fullness while reducing overhead and maximizing performance. When a change to cache data occurs, meta information is sent to other nodes to maintain the distributed cache management index schema. Note that only pointers to where data and fresh cache entries reside are what is stored and communicated in the meta data via the distributed caching algorithm.


    Figure 9: EMC VPLEX Metro Today

    For metro deployments, two clusters (Figure 9) are utilized with distances supported up to about 100km or about 5ms of latency in a synchronous manner utilizing long distance Fibre Channel optics and transceivers including Dense Wave Division Multiplexing (DWDM) technologies (See Chapter 6: Metropolitan and Wide Area Storage Networking in Resilient Storage Networking (Elsevier) for additional details on LAN, MAN and WAN topics).

    Initially EMC is supporting local or Metro including Campus based VPLEX deployments requiring synchronous communications however asynchronous (WAN) Geo and Global based solutions are planned for the future (Figure 10).


    Figure 10: EMC VPLEX Future Wide Area and Global

    Online Workload Migration across Systems and Sites
    Online workload or data movement and migration across storage systems or sites is not new with solutions available from different vendors including Brocade, Cisco, Datacore, EMC, Fujitsu, HDS, HP, IBM, LSI and NetApp among others.

    For synchronization and data mobility operations such as a VMware Vmotion or Microsoft HyperV Live migration over distance, information is written to separate LUNs in different locations across what are known as stretch volumes to enable non disruptive workload relocation across different storage systems (arrays) from various vendors. Once synchronization is completed, the original source can be disconnected or taken offline for maintenance or other common IRM tasks. Note that at least two LUNs are required, or put another way, for every stretch volume, two LUNs are subtracted from the total number of available LUNs similar to how RAID 1 mirroring requires at least two disk drives.

    Unlike other approaches that for coherency and performance rely on either no cached data, or, extensive amounts of cached data along with subsequent overhead for maintaining state fullness (consistency and coherency) including avoiding stale reads or writes, VPLEX relies on a combination of distributed cache lookup tables along with pass thru access to underlying storage when or where needed. Consequently large amounts of data does not need to be cached as well as shipped between VPLEX devices to maintain data consistency, coherency or performance which should also help to keep costs affordable.

    Approach is not unique, it is the implementation
    Some storage virtualization solutions that have been software based running on an appliance or network switch as well as hardware system based have had a focus of emulating or providing competing capabilities with those of mid to high end storage systems. The premise has been to use lower cost, less feature enabled storage systems aggregated behind the appliance, switch or hardware based system to provide advanced data and storage management capabilities found in traditional higher end storage products.

    VPLEX while like any tool or technology could be and probably will be made to do other things than what it is intended for is really focused on, flexibility, transparency and agility as opposed to being used as a means of replacing underlying storage system functionality. What this means is that while there is data movement and migration capabilities including ability to synchronize data across sites or locations, VPLEX by itself is not a replacement for the underlying functionality present in both EMC and third party (e.g. HDS, HP, IBM, NetApp, Oracle/Sun or others) storage systems.

    This will make for some interesting discussions, debates and applies to oranges comparisons in particular with those vendors whose products are focused around replacing or providing functionality not found in underlying storage system products.

    In a nut shell summary, VPLEX and the Virtual Storage story (vision) is about enabling agility, resiliency, flexibility, data and resource mobility to simply IT Infrastructure Resource Management (IRM). One of the key themes of global storage federation is anywhere access on a local, metro, wide area and global basis across both EMC and heterogeneous third party vendor hardware.

    Lets Put it Together: When and Where to use a VPLEX
    While many storage virtualization solutions are focused around consolidation or pooling, similar to first wave server and desktop virtualization, the next general broad wave of virtualization is life beyond consolidation. That means expanding the focus of virtualization from consolidation, pooling or LUN aggregation to that of enabling transparency for agility, flexibility, data or system movement, technology refresh and other common time consuming IRM tasks.

    Some applications or usage scenarios in the future should include in addition to VMware Vmotion, Microsoft HypverV and Microsoft Clustering along with other host server closuring solutions.


    Figure 11: EMC VPLEX Usage Scenarios

    Thoughts and Industry Trends Perspectives:

    The following are various thoughts, comments, perspectives and questions pertaining to this and storage, virtualization and IT in general.

    Is this truly unique as is being claimed?

    Interestingly, the message Im hearing out of EMC is not the claim that this is unique, revolutionary or the industries first as is so often the case by vendors, rather that it is their implementation and ability to deploy on a broad perspective basis that is unique. Now granted you will probably hear as is often the case with any vendor or fan boy/fan girl spins of it being unique and Im sure this will also serve up plenty of fodder for mudslinging in the blogsphere, YouTube galleries, twitter land and beyond.

    What is the DejaVu factor here?

    For some it will be nonexistent, yet for others there is certainly a DejaVu depending on your experience or what you have seen and heard in the past. In some ways this is the manifestation of many vision and initiatives from the late 90s and early 2000s when storage virtualization or virtual storage in an open context jumped into the limelight coinciding with SAN activity. There have been products rolled out along with proof of concept technology demonstrators, some of which are still in the market, others including companies have fallen by the way side for a variety of reasons.

    Consequently if you were part of or read or listened to any of the discussions and initiatives from Brocade (Rhapsody), Cisco (SVC, VxVM and others), INRANGE (Tempest) or its successor CNT UMD not to mention IBM SVC, StorAge (now LSI), Incipient (now part of Texas Memory) or Troika among others you should have some DejaVu.

    I guess that also begs the question of what is VPLEX, in band, out of band or hybrid fast path control path? From what I have seen it appears to be a fast path approach combined with distributed caching as opposed to a cache centric inband approaches such as IBM SVC (either on a server or as was tried on the Cisco special service blade) among others.

    Likewise if you are familiar with IBM Mainframe GDPS or even EMC GDDR as well as OpenVMS Local and Metro clusters with distributed lock management you should also have DejaVu. Similarly if you had looked at or are familiar with any of the YottaYotta products or presentations, this should also be familiar as EMC acquired the assets of that now defunct company.

    Is this a way for EMC to sell more hardware along with software products?

    By removing barriers enabling IT staffs to support more data on more storage in a denser and more agile footprint the answer should be yes, something that we may see other vendors emulate, or, make noise about what they can or have been doing already.

    How is this virtual storage spin different from the storage virtualization story?

    That all depends on your view or definition as well as belief systems and preferences for what is or what is not virtual storage vs. storage virtualization. For some who believe that storage virtualization is only virtualization if and only if it involves software running on some hardware appliance or vendors storage system for aggregation and common functionality than you probably wont see this as virtual storage let alone storage virtualization. However for others, it will be confusing hence EMC introducing terms such as federation and avoiding terms including grid to minimize confusion yet play off of cloud crowd commotion.

    Is VPLEX a replacement for storage system based tiering and replication?

    I do not believe so and even though some vendors are making claims that tiered storage is dead, just like some vendors declared a couple of years ago that disk drives were going to be dead this year at the hands of SSD, neither has come to life so to speak pun intended. What this means for VPLEX is that it leverages underlying automated or manual tiering found in storage systems such as EMC FAST enabled or similar policy and manual functions in third party products.

    What VPLEX brings to the table is the ability to transparently present a LUN or volume locally or over distance with shared access while maintaining cache and data coherency. This means that if a LUN or volume moves the applications or file system or volume managers expecting to access that storage will not be surprised, panic or encounter failover problems. Of course there will be plenty of details to be dug into and seen how it all actually works as is the case with any new technology.

    Who is this for?

    I see this as for environments that need flexibility and agility across multiple storage systems either from one or multiple vendors on a local or metro or wide area basis. This is for those environments that need ability to move workloads, applications and data between different storage systems and sites for maintenance, upgrades, technology refresh, BC/DR, load balancing or other IRM functions similar to how they would use virtual server migration such as VMotion or Live migration among others.

    Do VPLEX and Virtual Storage eliminate need for Storage System functionality?

    I see some storage virtualization solutions or appliances that have a focus of replacing underlying storage system functionality instead of coexisting or complementing. A way to test for this approach is to listen or read if the vendor or provider says anything along the lines of eliminating vendor lock in or control of the underlying storage system. That can be a sign of the golden rule of virtualization of whoever controls the virtualization functionality (at the server hypervisor or storage) controls the gold! This is why on the server side of things we are starting to see tiered hypervisors similar to tiered servers and storage where mixed hypervisors are being used for different purposes. Will we see tiered storage hypervisors or virtual storage solutions the answer could be perhaps or it depends.

    Was Invista a failure not going into production and this a second attempt at virtualization?

    There is a popular myth in the industry that Invista never saw the light of day outside of trade show expo or other demos however the reality is that there are actual customer deployments. Invista unlike other storage virtualization products had a different focus which was that around enabling agility and flexibility for common IRM tasks, similar the expanded focus of VPLEX. Consequently Invista has often been in apples to oranges comparison with other virtualization appliances that have as focus pooling along with other functions or in some cases serving as an appliance based storage system.

    The focus around Invista and usage by those customers who have deployed it that I have talked with is around enabling agility for maintenance, facilitating upgrades, moves or reconfiguration and other common IRM tasks vs using it for pooling of storage for consolidation purposes. Thus I see VPLEX extending on the vision of Invista in a role of complimenting and leveraging underlying storage system functionality instead of trying to replace those capabilities with that of the storage virtualizer.

    Is this a replacement for EMC Invista?

    According to EMC the answer is no and that customers using Invista (Yes, there are customers that I have actually talked to) will continue to be supported. However I suspect that over time Invista will either become a low end entry for VPLEX, or, an entry level VPLEX solution will appear sometime in the future.

    How does this stack up or compare with what others are doing?

    If you are looking to compare to cache centric platforms such as IBMs SVC that adds extensive functionality and capabilities within the storage virtualization framework this is an apples to oranges comparison. VPLEX is providing cache pointers on a local and global basis functioning in a compliment to underlying storage system model where SVC caches at the specific cluster basis and enhancing functionality of underlying storage system. Rest assured there will be other apples to oranges comparisons made between these platforms.

    How will this be priced?

    When I asked EMC about pricing, they would not commit to a specific price prior to the announcement other than indicating that there will be options for on demand or consumption (e.g. cloud pricing) as well as pricing per engine capacity as well as subscription models (pay as you go).

    What is the overhead of VPLEX?

    While EMC runs various workload simulations (including benchmarks) internally as well as some publicly (e.g. Microsoft ESRP among others) they have been opposed to some storage simulation benchmarks such as SPC. The EMC opposition to simulations such as SPC have been varied however this could be a good and interesting opportunity for them to silence the industry (including myself) who continue ask them (along with a couple of other vendors including IBM and their XIV) when they will release public results.

    What the interesting opportunity I think is for EMC is that they do not even have to benchmark one of their own storage systems such as a CLARiiON or VMAX, instead simply show the performance of some third party product that already is tested on the SPC website and then a submission with that product running attached to a VPLEX.

    If the performance or low latency forecasts are as good as they have been described, EMC can accomplish a couple of things by:

    • Demonstrating the low latency and minimal to no overhead of VPLEX
    • Show VPLEX with a third party product comparing latency before and after
    • Provide a comparison to other virtualization platforms including IBM SVC

    As for EMC submitting a VMAX or CLARiiON SPC test in general, Im not going to hold my breath for that, instead, will continue to look at the other public workload tests such as ESRP.

    Additional related reading material and links:

    Resilient Storage Networks: Designing Flexible Scalable Data Infrastructures (Elsevier)
    Chapter 3: Networking Your Storage
    Chapter 4: Storage and IO Networking
    Chapter 6: Metropolitan and Wide Area Storage Networking
    Chapter 11: Storage Management
    Chapter 16: Metropolitan and Wide Area Examples

    The Green and Virtual Data Center (CRC)
    Chapter 3: (see also here) What Defines a Next-Generation and Virtual Data Center
    Chapter 4: IT Infrastructure Resource Management (IRM)
    Chapter 5: Measurement, Metrics, and Management of IT Resources
    Chapter 7: Server: Physical, Virtual, and Software
    Chapter 9: Networking with your Servers and Storage

    Also see these:

    Virtual Storage and Social Media: What did EMC not Announce?
    Server and Storage Virtualization – Life beyond Consolidation
    Should Everything Be Virtualized?
    Was today the proverbial day that he!! Froze over?
    Moving Beyond the Benchmark Brouhaha

    Closing comments (For now):
    As with any new vision, initiative, architecture and initial product there will be plenty of questions to ask, items to investigate, early adopter customers or users to talk with and determine what is real, what is future, what is usable and practical along with what is nice to have. Likewise there will be plenty of mud ball throwing and slinging between competitors, fans and foes which for those who enjoy watching or reading those you should be well entertained.

    In general, the EMC vision and story builds on and presumably delivers on past industry hype, buzz and vision with solutions that can be put into environments as productivity tool that works for the customer, instead of the customer working for the tool.

    Remember the golden rule of virtualization which is in play here is that whoever controls the virtualization or associated management controls the gold. Likewise keep in mind that aggregation can cause aggravation. So do not be scared, however look before you leap meaning do your homework and due diligence with appropriate levels of expectations, aligning applicable technology to the task at hand.

    Also, if you have seen or experienced something in the past, you are more likely to have DejaVu as opposed to seeing things as revolutionary. However it is also important to leverage lessons learned for future success. YottaYotta was a lot of NaddaNadda, lets see if EMC can leverage their past experiences to make this a LottaLotta.

    Ok, nuff said.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press), The Green and Virtual Data Center (CRC Press) and Resilient Storage Networks (Elsevier)
    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2024 Server StorageIO and UnlimitedIO LLC All Rights Reserved

    Spring 2010 StorageIO Newsletter

    Welcome to the spring 2010 edition of the Server and StorageIO (StorageIO) news letter.

    This edition follows the inaugural issue (Winter 2010) incorporating feedback and suggestions as well as building on the fantastic responses received from recipients.

    A couple of enhancements included in this issue (marked as New!) include a Featured Related Site along with Some Interesting Industry Links. Another enhancement based on feedback is to include additional comment that in upcoming issues will expand to include a column article along with industry trends and perspectives.

    StorageIO News Letter Image
    Spring 2010 Newsletter

    You can access this news letter via various social media venues (some are shown below) in addition to StorageIO web sites and subscriptions. Click on the following links to view the spring 2010 newsletter as HTML or PDF or, to go to the newsletter page.

    Follow via Goggle Feedburner here or via email subscription here.

    You can also subscribe to the news letter by simply sending an email to newsletter@storageio.com

    Enjoy this edition of the StorageIO newsletter, let me know your comments and feedback.

    Also, a very big thank you to everyone who has helped make StorageIO a success!.

    Cheers gs

    Greg Schulz – Author Cloud and Virtual Data Storage Networking (CRC Press, 2011), The Green and Virtual Data Center (CRC Press, 2009), and Resilient Storage Networks (Elsevier, 2004)

    twitter @storageio

    All Comments, (C) and (TM) belong to their owners/posters, Other content (C) Copyright 2006-2012 StorageIO and UnlimitedIO All Rights Reserved