Welcome to another EMC Ask the Expert event.
The value of software-defined storage is being able to deliver cloud-scale storage capabilities in software that can leverage low-cost commodity off the shelf components (COTS). Software-defined means customers have more choices but also more questions:
Join this event to interact with our experts and others to learn the pros and cons of each approach and how to obtain the cost benefits of commodity without sacrificing enterprise-grade manageability, scalability, and service levels.
Meet Your Experts:
|Product Marketing, ViPR, ECS - EMC |
George is a Senior Manager responsible for EMC ECS Appliance and Centera and Atmos object storage platform product marketing. He has worked in the technology industry for nearly 20 years as a product marketing manager, industry analyst, and research director. As an analyst, George covered cloud computing and services, IT infrastructure, and IT management software. George has worked for small, pre-IPO firms such as LogMeIn, boutique advisory firms like Yankee Group as well as established technology vendors EMC, CA and Sybase.
|Distinguished Engineer - EMC|
Mark is an EMC Distinguished Engineer and currently works within the Emerging Technologies Division at EMC Corporation. Among the products Mark covers is ViPR, a next general object storage system with support for heterogeneous storage and APIs, EMC Atmos, the industry’s leading multi-petabyte information management solution, and EMC Centera, the industry’s leading compliance and archiving offering. Mark has worked within EMC’s object and cloud storage group since 2001. Mark has driven the technology and business of cloud storage by creating numerous talks explaining cloud storage technologies, by architecting new features such as the GeoDrive cloud storage product, by maintaining an active blog about cloud storage trends and technologies, and by serving on numerous internal committees to formulate cloud storage strategy.
Senior Manager Engineering - EMCMike has been with EMC for 15 years. During this time he has had various roles, and currently works in EMC ViPR Data Services qualifying customer solutions. Prior to EMC, Mike owned and operated a trio of restaurants in the Washington, DC area with his brother-in-law.
Moderator: Mark Prahl
This discussion takes place from Jan. 26th - Feb. 6th. Get ready by bookmarking this page or signing up for e-mail notifications.
Share this event on Twitter or LinkedIn:
Welcome to our newest ATE session on ECN. We'd like to open this discussion for replies. Let's make of this an opportunity to respectfully engage our experts with your questions, comments or ideas. Have a great experience!
Hello and welcome to our Ask the Expert event about the critical considerations that go into that very important decision about whether to build or purchase an object storage cloud platform. During the course of the next two weeks, our object storage experts will engage in a lively discussion about the merits and pitfalls of open source and commodity, and commercial vendor offerings.
Let’s get started.
Experts: Starting with some basics. What is object storage and why is it the popular choice for web-scale storage?
Object storage is a method in which to logically store data; primarily unstructured data. Objects are discrete units of storage, characterized by their extended metadata. Each object is assigned a unique identifier which allows a server or end user to retrieve the object without needing to know the physical location of the data. Like files, objects contain data, but unlike files, objects are not organized in a hierarchy. Every object exists at the same level in a flat address space called a storage pool and one object cannot be placed inside another object.
Object storage systems feature a single, flat global namespace, regardless of the number of nodes or locations. This results in location-independence without requiring a file system to govern data placement. This also means it can scale linearly and efficiently to 10s and 100s of petabytes (even exabytes!). Multiple nodes and sites appear as one, logical storage system. Replication and geo-distribution of data is driven by policies rather than dedicated replication and backup infrastructure. Object storage provides geo-distribution with multiple sites in an active-active architecture. This means any site can deliver content in the event of a site failure.
Object is used widely for Web and mobile appications because of its simple archiecture and http/https based access. Access to data is via a REST based API and http/https. Examples of object storage APIs include Amazon S3, OpenStack Swift and EMC Atmos. Developers can write applications to one of the industry lead object storage APIs without having to worry about the infrastructure. Changes to the infrastructure don't require application code changes.
For more info, I recommend viewing the webcast, “How REST & Object Storage Make Next Generation Application Development Simple” to get an in-depth look at object architecture and writing apps to REST based APIs.
Continuing along the same line. What is considered commodity storage and components? Are these commodity components as reliable as a traditional SAN?
Commodity storage is generally considered low cost, high capacity SATA disk drives, off the shelf controllers, and standard X86 based servers. The intelligence and management is software-defined. This allows these commodity “off the shelf” components to perform reliably. With regard to whether a commodity platform is as reliable as traditional SAN, that depends on a number of factors.
Even a traditional array is comprised of primarily commodity components. But the benefit is there is nothing to build and the software and hardware are optimized to work together. The potential drawback is that it is purpose-built for a specific set of tasks and data. Software-defined storage, as its name implies, is hardware-independent and provides theability to use standard, COTS components. This can dramaticaly lower storge costs. But as far as reliability, commodity components are built to fail and do so frequently. Getting entrprise-grade reliability requires the right scale, the right software and operational skills.
OpenStack is often a consideration these days. What are some things an IT department should be doing to make sure they’re thoroughly “kicking the tires” and getting a relevant view of what OpenStack would be like in production?
It is important to evaluate the entire lifecycle of the environment to avoid surprises. Beyond testing application access to the system during normal operating conditions, things to test would include:
When a company evaluates the decision whether to build an on-prem vs. using a poublic cloud service, there are a couple general questions they ask.
For smaller organizatons that don't hve the skills or resources to build and operate their own infrastructure, public cloud is a great way to get the technology capabilities they need and get to market faster. For larger businesses, they have more of a decision to make. The good news is that software-defined storage solutions and commodity components have made public cloud economics more accessible to more companies. The next decision is whether to build an evironment from the ground up or use an intregrated vendor platform.