Vm13 vnx mixed workloads

Embed Size (px)

DESCRIPTION

 

Text of Vm13 vnx mixed workloads

  • 1. Mixed Workloads onEMC VNX Storage Arrays

2. Tony Pittman @pittmantony TPittman@Varrow.comMartin Valencia @ubergiek MValencia@Varrow.com 3. Goals For This SessionDiscuss: How VNX storage pools work How common workloads compare Which workloads are compatible How to monitor performance How to mitigate performance problems 4. Goals For This Session Also check out this session at 2:55EMC Session: VNX PerformanceOptimization and Tuning - David Gadwah,EMC 5. VNX Basics VNX shines at mixed workloads IOPS - Mixed WorkloadsCX4VNX Series with Rotating drivesVNX Series with Flash drives# of Users VNX5100VNX5300 VNX5500 VNX5700 VNX7500CX4-120 CX4-240 CX4-480 CX4-960Platform 6. VNX Basics VNX is EMCs mid-tier unified storagearray FC, iSCSI or FCoE block connectivity Multiple SAS buses backend NFS and CIFS file connectivity Built for flash 7. VNX Architecture Object: Application servers Exchange serversClientsVirtual servers Oracle servers Atmos VESAN LAN VNX Unified Storage 10Gb 10GbFC iSCSI FCo Enet FC iSCSI FCo Enet EEVNX X-BladeFailoverVNX X-Blade VNX X-BladeVNX X-Blade VNX X-BladeVNX X-Blade VNX X-BladeVNX X-BladeVNX OE FILE VNX SP Failover VNX SP VNX OE BLOCK Power Supply Power Supply SPSSPS LCCLCC Flash drives Near-Line SAS drivesSAS drives 8. VNX Architecture Two Storage Processors with DRAMcache, frontend ports (FC, iSCSI,FCOE) and backend ports (6 Gb SAS) Each LUN owned by one SP, andaccessible by both Both SPs have active connections 9. VNX Architecture FAST Cache Second layer of read/write cache, housedon solid state drives Operates in 64 KB chunks Reactive in nature Great for random I/O Dont use it for sequential I/O 10. VNX Architecture Storage Pools Based on RAID RAID 5, RAID 1/0, RAID 6 FAST VP: Fully Automated Storage Tiering Pools with multiple drive types: EFD, SAS, NL-SAS Sub-LUN tiering Operates at 1 GB chunks Adjusts over time, not immediately FASTCache is more immediate 11. VNX ArchitectureWhen should I use traditional RAID Groups? As theexception: Very specific performance tuning (MetaLUNs) Internal array features (write intent logs, clone privateLUNs) Maybe Recoverpoint journals Supportability (Im looking at you, Meditech)Remember the limitations: Maximum of 16 drives Expand via metaLUNs No tiering 12. VNX Architecture IOPS per drive type (for sizing)3500 IOPS - EFD 180 IOPS - 15k rpm drive 140 IOPS - 10k rpm drive90 IOPS - 7200 rpm driveEffects of RAID Parity calculations (RAID 5 and RAID 6) Effect on response times Write penalty RAID 1/0 = 2x RAID 5 = 4x RAID 6 = 6x 13. VNX Architecture Real-world effect of write penalty: 10x 600 GB 15k SAS drives = 1800 readIOPS With RAID 1/0, capable of 900 write IOPS With RAID 5, capable of 450 write IOPS 1 write operation takes 4 I/O operations With RAID 6, capable of 300 write IOPS 1 write operation takes 6 I/O operations 14. WorkloadsCommon workloads seen in the field.Virtual Disks/VMFS (RAID5)DB Data files (RAID5)DB Transaction files (RAID 10)Unstructured Data, Backups (RAID6) 15. Real World WorkloadsStandard Performance Evaluation CorporationBenchmarking Real World PerformanceNon-profitUses generic applications rather thanspecific applicationsSPEC benchmarks rely on a mix of I/O to simulate ageneric applicationThis balances the need for real world performance andconsistency over time 16. Ideal Scenario Array with single application No budget constraints Separate storage pools for differentsub-workloads 17. Ideal Scenario The ideal SQL: PCIe flash and XtremeSW on the host FAST Cache in the array tempDB: Data files on separate RAID 5 storage pool User DBs: Each has tlogs on separate RAID 1/0 storage pool Each has data files on one or more RAID 5 storage pools, with the appropriate drive configuration (EFD+FAST) Backups / Dump files: Separate RAID 6 storage pool, maybe a separate array 18. Reality Cant isolate every workloadCost prohibitive, and do we have to? Business Critical Application maybe Management & Lower-tier application probably not 19. Basic Storage PoolLayout One or Two RAID 5 pools(ex: Gold & Silver) FAST with EFDs, SAS, NL-SAS accordingto skew or 5/20/75 rule RAID 1/0 pool for transaction logs. 15k SAS drives RAID 6 pool for backup files andunstructured data 7.2k NL-SAS drives 20. RAID 5 Pools VMFS DB Data Files Good for random read/write mix Use FASTCacheExample: Gold Pool: 5x EFD, 15x SAS, 16x NL-SAS Silver Pool: 15x SAS, 16x NL-SAS 21. Drive Composition: Skew 22. RAID 1/0 Pool Transaction Logs for many applications Specifically for small sequential writes Do Not Use FAST Cache Itll be wasted Itll hurt performanceExample: 8x 15k SAS drives 23. RAID 6 Pool Unstructured data Office Files (.doc, .xls, .etc) Images Backup files Split into separate pool if necessary Low I/O & high capacity Good for long sequential writes Do Not Use FAST Cache Itll be wastedExample: 16x 7.2k rpm NL-SAS drives 24. Pool Layout 25. Monitoring and TroubleshootingThere is no Set it and forget itWorkloads change over time Users get added Transaction load increases Requirements changeOften no one tells us 26. Problem identificationProactive performance review Admins wear too many hats Low priorityReactive to user impact (Too late) Crisis management 27. Troubleshooting MetricsWhere do we start? What do we look at? Cache Utilization Exceeding a high water marks, need to flush cache to disk Forced Flushes SP performance Balance the SP load Pool LUN migration (metadata) Online LUN migration 28. The ToolboxUnisphere Analyzer (On array) Proactively gathers data for review Data logging must be enabled on the array 29. The ToolboxVNX Monitoring and Reporting (Off array) Historical Data Collection Streamlined application based on Watch4net 30. The ToolboxEMC miTrend Leverages NAR (Navisphere analyzer data) that can beretrieved from the array Need EMC or partner (us) to perform the analysis 31. Troubleshooting / Problem MitigationSeveral options for mitigating a performance problem: Add drives OE 32 required to rebalance existing data Pre OE 32, must increase pool by originating drive count, existing data will not be rebalanced Migrate to a different pool Live migration avoids the need for an outage Performance Throttling minimizes performance impact 32. Troubleshooting / Problem Mitigation Rebalance at the application layer Storage vMotion Host-based data migration (Open Migrator, etc) Migrate data between arrays SANCopy Replication (Mirroring/RecoverPoint) Reduce workload Reschedule for off-hours (backups for example) Decommission non-critical workloads 33. Thank You!Questions