VMworld 2014/vSphere Storage Best Practices: Next-Gen Storage Technologies

From Omnia
Jump to navigation Jump to search

vSphere Storage Best Practices: Next-Gen Storage Technologies

STO2496 vSphere Storage Best Practices: Next-Gen Storage Technologies

"This VMware Technical Communities Session will present a technical best practices with emerging storage technologies for vSphere. The storage industry is experiencing a high level of innovation that will influence your next datacenter refresh. Storage industry experts present this session in a vendor neutral perspective on best practices with storage infrastructure technologies spanning host-based acceleration, all-flash, and hyper-converged. Vaughn and Chad focused on delivering a deep technical session and have invited Rawlinson Rivera of VMware to join and expand the knowledge transfer. This session will present best practices that span connectivity, performance, availability and failure domains, data protection and automation."

  • Rawlinson Rivera - Sr. Technical Marketing Architect, VMware, Inc
  • Chad Sakac - SVP, Global Systems Engineering, EMC
  • Vaughn Stewart - Chief Technical Evangelist, Pure Storage

Pure Storage

Flash Array by Pure Storage | Enterprise Flash Array - http://www.purestorage.com/

"Pure Storage released a flash memory product called FlashArray" (Pure Storage - Wikipedia - http://en.wikipedia.org/wiki/Pure_Storage)

Simplicity

Keep it simple, it will save you headaches

Best Storage Practice Documents

Read them. Each vendor may have different methods and policies, so don't apply the same everywhere.

Undersand Cloud I/O

aka Shared Virtual Infrastructures

The "I/O Blender"

Know your workloads and average I/O sizes

Hybrid Storage

All flash storage arrays differ

Benchmarking

What is your objective? Know what you are testing for.

Absurd Testing - just silly

data sets have an active i/o band and a cold band

  • spikes and repeat data are common

Principles:

  • don't let vendors steer you too much
  • talk to other customers and system integrators
  • benchmark over time
  • lots of different work loads - NOT A SINGLE GUEST, SINGLE DATASTORE!
  • benchmark mixed loads
  • use SLOB or IOmeter - still artificial workloads
  • hard to generate sufficient IO from a host for modern flash storage
  • absolute performance is not the only design consideration
    • if it doesn't meet performance needs - it's a boat anchor

Architecture Matters:

  • Always benchmark performance
  • always benchmark resiliency, availability and data management features
  • recommend testing with actual data

Best Practices

Vendor best practices seem to break each other

Always separate guest VM traffic from storage and Vmkernel network

Jumbo Frames

Avoid Jumbo Frames ("just my personal recommendation") - let the flames begin

Thin Provisioning

Thin Provisioning is not a data reduction technology!

It is just an on demand feature

Deduplication and compression are

Inline vs post processing - inline happens during transfer, post processing is like a garbage collector, and happens afterwards

UNMAP

T10 UNMAP is still not here with vSphere 5.5 - in the way people expect

still a manual process

5.1:

vmkfstools -k

5.5:

esxcli storage vmfs UNMAP

only cleans up deleted VMs, not within guests

Storage QoS

Ensures priority of service for some applications when storage IOPs are exhausted

Operation complexity is why most people don't do it today

Inconsistent capabilities limit broad adoption

There is no QoS all the way down to the guest

VMware Virtual SAN Integration

(see image)

Features:

enterprise features:

  • network i/o control
  • vmotion
  • svmotion
  • drs
  • ha

data protection:

  • linked clones
  • snapshots
  • vdp advanced

disaster recovery:

  • vsphere replication
  • vcenter site recovery manager

cloud ops and automation

  • vcenter operations manager
  • vcloud automation center

VMware Virtual SAN Best Practices

(see picture)

seek simplicity

network connectivity:

  • 10 Gbe preferred
  • leverage distributed switch (NIOC)
  • L2 multicast

storage controller queue depth:

  • supports depth 256 or higher
  • higher storage controller queue depth will help performance, resync, rebuild
  • pass-thru mode preferred

Disk and disk groups:

  • don't mix disk types in a cluster, for predictable performance
  • more disk groups are better than one

storage and flash sizing:

  • size of cache is equal to performance
  • start with recommendation of 10% of SSD
  • maintain utilization below 80%

cluster design

  • wider is better
  • include host failures scenarios as part of design

policies:

  • consider use of Object space reservation to sustain levels of performance
  • more disk groups are better than one

Go Wide!

Automation

Automate everything you can, infrastructure must be programmable

Automation - should never do anything more than once

Every array worth their salt has Restful API - use it

Don't hard-code to any vendor API - abstract via something open

Don't get hung up on Puppet vs Chef vs Ansible vs ... - Just pick one and start

To make it easy, start with EMC ViPR controller (http://www.emc.com/getvipr) - free with community support