Tuesday, June 16, 2009
How to check part number of installed adapter in Ontap?
Sunday, June 7, 2009
How to configure ontap to send specific event type on your pager or handheld?
To send notification to an e-mail address, cell phone or text pager e-mail address, type:
options autosupport.noteto [pager1 mail address],[pager2 mail address],...
Ontap doesn’t sends all the messages in the specific event category through email to configured mail addresses however all autosupport type events are sent to NetApp, If you want to see what all the events trigger autosupport mail notification you can use this link from NetApp to check. By default ontap sends only critical and error messages types to autosupport.to and autosupport.noteto addresses however if you want to configure the ontap to trigger email notifications on your defined severity of alert you can use “autosupport.notify_threshold” option.
The “autosupport.notify_threshold” can be set to the following values:
Setting | Result |
"" | (empty string) No messages will be sent to customer e-mail addresses. |
"critical" | Only critical messages will be sent to customer e-mail addresses. |
"error" | Only critical and error messages are sent (default). |
"warning" | Warning messages and above are sent. |
"notice" | Notice messages and above are sent. |
"info" | Info messages and above. |
"debug" | All messages will be sent to customer addresses. |
Thursday, June 4, 2009
PCS - Predictive Cache Statistics for PAM cards
PCS: Determining If PAM Will Improve Performance
To begin using PCS, you enable the feature with the command:
options flexscale.enable pcs
options flexscale.enable off
With PCS enabled, you can find out what's happening using the following command:
stats show -p flexscale-pcs
Sample output is shown below
Use the following guidelines to help you interpret the data:
· If the hit/(invalidate+evict) ratio is small, then a lot of data is being discarded before it is used. The instance (ec0, ec1, ec2) may be too small.
· If the (hit+miss)/invalidate ratio is too small, it might indicate a workload with a large amount of updates; switch to metadata mode and check the hit% again.
· If the usage is stable and there are a small number of invalidates and evictions, then the working set fits well.
· The KB/s served by the cache is approximately equal to the hit/s × 4KB per block.
Note that the three caches simulated in PCS are cascading caches. In the example above, ec0 represents the first cache of size 8GB, ec1 represents the second cache of size 8GB, and ec3 represents the third cache of size 16GB. The hit per second for a 32GB cache is the sum of all the hits per second for all three caches. The key advantage of cascading caches is that in the process of measuring an accurate hit rate for a 32GB cache, we also obtain hit rate estimates of both 8GB and 16GB caches. This gives us three points on the hit rate curve and the ability to estimate hit rates for intermediate cache sizes.
PAM and FlexShare
FlexShare™ is a Data ONTAP option that allows you to set priorities for system resources (processors, memory, and I/O) at a volume level, thereby allocating more resources to workloads on particular volumes when the controller is under significant load. FlexShare is fully compatible with PAM, and settings made in FlexShare apply to the data kept in the PAM cache. With FlexShare, finer-grained control can be applied on top of the global policies you implement with PAM. For example, if an individual volume is given a higher priority with FlexShare, data from that volume will receive a higher priority in the cache.
Wednesday, June 3, 2009
PAM (Performance Accelerator Module) in NetApp
Notes: This is a multi part post and I have collected all the materials from various atricles published in Netapp and internet.
The PAM offers a way to optimize the performance of a NetApp storage system by improving Throughput and Latency while reducing the number of disk spindles/shelves required as well as power, cooling and rack space requirements.
It is a an array controller resident, intelligent 3/4 length PCIe card with 16GB of DDR2 SDRAM that is used as a read cache and is integrated with DataONTAP via FlexScale which is software that provides various tuning options and modes of operation.
Why the PAM?
In every disk array, the processors can access data from two areas: from disk or the Cache. If the data is not in the Cache, this results in a Cache miss which means the data will have to be fetched from disk. The time (latency) to fetch data directly from the disk is orders of magnitude greater than from cache. For read intensive random I/O applications that are latency sensitive this requires the configuration of a high number disk in order to provide user and application acceptable latencies. Most Microsoft applications fall in that category as well as Server virtualization platforms and File services which tend to generate, primarily, these types of workloads.
So the idea with deploying one or more PAMs is that a hit to the PAM's cache will considerably reduce the time it takes to fetch the data as compared to the same process occurring from disk. The additional benefit here is that it will also allow for a more reasonable amount of disks and shelves to be deployed to support the same application thus reducing costs.
In the simplest terms, the Performance Acceleration Module is a second-layer cache: a cache used to hold blocks evicted from the WAFL buffer cache. In a system without PAM, any attempt to read data that is not in system memory results in a disk read. With PAM, the storage system first checks to see whether a requested read has been cached in one of its installed modules before issuing a disk read. Data ONTAP maintains a set of cache tags in system memory and can determine whether or not a block resides in PAM without accessing the card. This reduces access latency because only one DMA operation is required on a cache hit. As with any cache, the key to success lies in the algorithms used to decide what goes into the cache. We’ll have more to say about that in the following section.
Hardware Architecture
PAM is a combination of both hardware and software (the PAM software is known as FlexScale.) A license is required to enable the hardware. The PAM hardware module is implemented as a ¾-length PCIe card offering dual-channel DMA access to 16GB of DDR2 memory per card and a custom-coded field-programmable gate array (FPGA) that provides the onboard intelligence necessary to accelerate caching tasks.
· PCI Express X4
· 16GB of RAM per module
· System interaction and cache managed by custom-coded FPGA
Power: 18W normal, 25W maximum at system initialization for approximately 10 seconds (For comparison, current disk shelves draw 340W continuous.) Operating environment: 10ºC to 40ºC ambient temperature Airflow: 150 LFM minimum Heat: 90 BTU/hr Form factor: ¾ length PCIe
Compatibility and Support
FAS/V-Series | Max # of Modules | Extended Cache Memory |
FAS6080 / V6080 FAS6070 / V6070 SA600 | 5 | 80GB |
FAS6040 / V6040 FAS6030 / V6030 FAS3170 / V3170 | 4 | 64GB |
FAS3070 / V3070 FAS3140 / V3140 SA300 | 2 | 32GB |
FAS3040 / V3040 | 1 | 16GB |
Types of caching in PAM
PAM supports three modes of operation:
· Default mode, in which data and metadata are cached
· Metadata mode, in which only metadata is cached
· Low-priority mode, which enables caching of sequential reads and other low-priority data
flexscale.enable on
flexscale.lopri_blocks off
flexscale.normal_data_blocks on
flexscale.enable on
flexscale.lopri_blocks off
flexscale.normal_data_blocks off
flexscale.enable on
flexscale.lopri_blocks on
flexscale.normal_data_blocks on
How usable space in NetApp array LUN is calculated
The usable space of an array LUN is reduced by the elements shown in the following table.
If you are using block checksums | If you are using zoned checksums |
10%–WAFL reserve | 10%–WAFL reserve |
1%– Core dump | 1%– Core dump |
20%–Volume-level Snapshot copy (default setting; optional) | 20%–Volume-level Snapshot copy (default setting; optional) |
5%–Aggregate-level Snapshot copy (default setting; optional) | 5%–Aggregate-level Snapshot copy (default setting; optional) |
12.5%–Checksum | |
Please note that the NetApp definition of a gigabyte (GB) is 1000 x 1024 x 1024 bytes.
Now you must be thinking why shouldn’t I go for zoned checksums to save 12.5% space? So what I would say is, if you are not much concerned about performance then you an easily go ahead and use zoned checksum over block checksum (specially in random-read intensive workloads) however please understand the difference between them before you make-up your mind to select one over another. There are a lot of references available on the internet to help you understand the concept even I will also try to cover them if I have some time.
Here’s couple of formulas from NetApp to calculate the space
Usable capacity is the actual space available to store the data and desired capacity is the exact array LUN size required to achieve a usable capacity.
Now if we denote usable capacity with Y and desired capacity with Z so we get the below formula here I have shown an example of 10GB in desired and usable space with Snapshot reserve set at default.
Block | Zoned |
Y = Z x {0.875 x 0.9 x 0.99 x Snapshot reserve} | Y = Z x {0.9 x 0.99 x Snapshot reserve} |
10 x {0.875 x 0.9 x 0.99 x 0.8} = 6.2 GB | 10 x {0.9 x 0.99 x 0.8} = 7.1 GB |
Z = Y / {0.875 x 0.9 x 0.99 x Snapshot reserve} | Z = Y / {0.9 x 0.99 x Snapshot reserve} |
10 / {0.875 x 0.9 x 0.99 x 0.8} = 16 GB | 10 / {0.9 x 0.99 x 0.8} = 14 GB |
Please note that I have taken the disk size as identified by system not the marketing size given by manufacturers.