Re: Orion, NetApp, and iops

  • From: Henry Poras <henry@xxxxxxxxxxxxxxx>
  • To: Jakub Wartak <vnulllists@xxxxxxxxxxxx>
  • Date: Tue, 04 May 2010 10:18:53 -0400

Sure.
The OS is Linux (2.6.18) with FC. Orion is bundled with an 11.2 install. The NetApp filer is a 3050.

I'm not sure about the detailed NetApp settings and would have to find where to look. But even if various optimizations are turned on, I don't understandthese results. It might be read ahead, but if Orion is really random reads across the disk I wouldn't have expected this to show up.

I will try the "stats show disk" command as soon as I can. That should help.

I also believe that RAID-DP shouldn't lead to a major deviation from 48*180. Definitely not an increase of iops close to an order of magnitude.

Thanks for the suggestions.

Henry


Jakub Wartak wrote:

We need more data, bascially OS and storage (FC, iSCSI?), versions (OS, Data ONTAP), LVM configuration and NetApp model/configuration (sysconfig command from SSH on NetApp too), I/O scheduler used if this is Linux, vmstat & iostat outputs during the run, etc. NetApp has many optimization settings (like no_update_atime on vols), FlexScale & PAM & PAM-II flash accelerators (not to mention NVRAM write acceleration...), was deduplication turned on, etc.

There are some shortcut commands that you should launch on filer that could solve the mystery, these are "lun show -o -i 1 /vol/path/to/lun" and "stats show disk:*:disk_busy". Second is going to show you how much really disks are utilised.

BTW: LUN can be only be inside WAFL volume, volume can be only on single aggregate. So my understanding is that LUN is in single aggregate that has 3 raid group sets each consisting of 16 disks, correct? As far as I remember the calculation of IOPS in RAID-DP is not as simple as 48*180.

--
//www.freelists.org/webpage/oracle-l


Other related posts: