Preface
Overview
Many data centers use iSCSI block storage for a variety of applications and workloads. As more data is stored on iSCSI, efficient network designs are under evaluation to minimize latency, maximize use of interswitch links between switches and to automate IP network configuration changes. For these reasons, Brocade’s VCS Fabric with Brocade VDX Switches is increasingly being deployed when new iSCSI storate arrays are added to the storage pool.
Purpose of This Document
The goal of this document is to validate Brocade VCS Fabric with the EMC VNX. The testing includes test cases that demonstrate VCS Fabric feature support, high availability and resiliency and automatic network reconfiguration when links are added or removed. This document provides validation test results for typical use cases of iSCSI with Brocade’s VCS Fabric and EMC VNX iSCSI storage arrays.
Audience
The content in this document is written for a technical audience, including solution architects, system engineers, and technical development representatives.
Objectives
- Validate interoperability with redundant and non-redundant network topologies with Brocade VCS and EMC VNX.
- Identify any unique configuration details necessary to establish end-to-end connectivity.
- Simulate real-world customer environment perturbations and validate end-to-end behavior and recovery.
Test Conclusions
- Brocade VCS Fabric and EMC VNX smoothly interoperate with both redundant and non-redundant configurations. No exceptions were noted.
- The configuration procedures followed Brocade VCS Fabric and EMC VNX standard documentation guides and found no issues with these procedures.
- All simulated changes in network and storage (add, remove, failover) resulted in expected behaviors and I/O recovery.
Related Documents
Document History
Date Version Description
10-8-2014 1.0 Initial version
Key Contributors
The content in this guide was provided by the following key contributors.
- Test Architects: Dustin Maiers, Patrick Stander
- Test Engineer: Robert Batesole
About Brocade
Brocade networking solutions help the world’s leading organizations transition smoothly to a world where applications and information reside anywhere. This vision is realized through the Brocade One™ strategy, which is designed to deliver key business benefits such as unmatched simplicity, non-stop networking, application optimization, and investment protection.
Innovative Ethernet and storage networking solutions for data center, campus, and service provider networks help reduce complexity and cost while enabling virtualization and cloud computing to increase business agility.
To help ensure a complete solution, Brocade partners with world-class IT companies and provides comprehensive education, support, and professional services offerings.
To learn more, visit (www.brocade.com)
About EMC
EMC is a global leader in enabling businesses and service providers to transform their operations and deliver information technology as a service (ITaaS). Fundamental to this transformation is cloud computing. Through innovative products and services, EMC accelerates the journey to cloud computing, helping IT departments to store, manage, protect and analyze their most valuable asset — information — in a more agile, trusted and cost-efficient way.
Test Plan
The storage array is connected to one VDX Ethernet fabric with multiple Windows and Red Hat server hosts to drive I/O.
Scope
Testing will be performed with GA versions of Brocade’s Network Operating System (NOS) in a heterogeneous environment. Test beds will include Brocade switches configured in routed and non-routed fabric configurations.
Testing is centered on interoperability and optimal configuration. Performance is observed within the context of best practice fabric configuration; however absolute maximum benchmark reporting of storage performance is beyond the scope of this test.
Test Configuration

Test Topology
The following diagram shows VCS Fabric switch port connections for each server and the EMC VNX. For example, “1/0/1” means the device port connects to Rbridge# = 1, Card# = 0 and Port# = 1. For fixed top of rack switches such as the VDX 6720, 6730 and 6740, there are no port cards, indicated by “0” for the Card#.

Device to VCS Fabric Switch Port Connections
DUT Descriptions
The following tables provide details about the devices under test (DUT).
Storage Array
| Model |
Vendor |
Description |
| VNX 5300 |
EMC |
Based on the powerful new family of Intel Xeon ES-25600 (Sandy Bridge) processores, the EMC VNX implements a modular architecture that integrates hardware components for block, file and object storage. Block storage support includes iSCSI leveraging 6 TB SAS disk drive technology with 1 GE and 10 GE connectivity options. |
Switches
| Model |
Vendor |
Description |
| BR-VDX6740 |
Brocade |
The Brocade VDX 6740 and 6740T are Ethernet fabric Top of Rack (ToR) switches featuring 10 GbE ports with 40 GbE uplinks. The new Brocade VDX 6740T-1G Switch offers dual-speed functionality. It can be deployed with 1000BASET for existing 1 GbE server connectivity and upgraded via software to 10GBASE-T for future bandwidth growth. Together with Brocade VCS Fabric technology, these switches deliver the high performance and low latency needed to support demanding virtualized data center environments. |
| BR-VDX8770-4 |
Brocade |
The Brocade VDX 8770 Switch is a highly scalable, low-latency, 1/10/40/100 Gigabit Ethernet (GbE) modular switch. Designed to easily scale out Brocade VCS fabrics, the Brocade VDX 8770 Switch brings new levels of performance to VCS fabric deployments. |
DUT Specifications
| iSCSI Array |
Version |
| EMC VNX 5300 |
Block Software 05.32.000.5.206, Unisphere 1.2.28.1.0110 |
| Brocade Switch |
Device ID |
Version |
| VDX 6720 |
C24-VCS20-1 |
NOS 5.0.0 |
| VDX 6730 |
C24F-VCS20-2 |
NOS 5.0.0 |
| VDX 6740 |
CASTOR-VCS20-12 |
NOS 5.0.0 |
| VDX 8770 |
M4-VCS20-11 |
NOS 5.0.0 |
| Operating System |
Adapter |
Driver Version |
| Red Hat 6.5 |
Brocade 1860 |
3.2.4.0 |
| Red Hat 6.5 |
Emulex OCe10102 |
10.0.803.31 |
| Windows Server 2008 R2 |
Brocade 1860 |
3.2.4.0 |
| Windows Server 2012 R2 |
Emulex OCe14102 |
10.0.803.37 |
| Server |
Memory |
Processor |
| IBM x3650 M3 |
8 GB |
E5506 @ 2.13 GHz (x4) |
| IBM x3650 M3 |
8 GB |
E5506 @ 2.13 GHz (x4) |
| HP ProLiant DL380 G7 |
8 GB |
E5506 @ 2.13 GHz (x4) |
| HP ProLiant DL380 G7 |
8 GB |
E5506 @ 2.13 GHz (x4) |
| Test Equipment |
Version |
| Medusa Labs Test Tools |
6.0.1.148039 |
| Multi-Path Software |
Operating System |
Version |
| EMC PowerPath |
Windows Server 2008 R2, 2012 R2 |
5.7 SP 4 (build 607) |
| EMC PowerPath |
Red Hat 6.5 |
5.7 SP 5 (build 2) |
| Device |
Interface |
IP Address |
| EMC VNX |
SP A, A1, Port 1 |
9.78.50.1 |
| EMC VNX |
SP B, B1, Port 1 |
9.79.50.2 |
| Windows 2012 Server R2 |
Ethernet 11 |
9.78.100.1 |
| Windows 2012 Server R2 |
Ethernet 12 |
9.79.100.2 |
| Windows 2008 Server R2 |
Local Area Connection 6 |
9.78.100.11 |
| Windows 2008 Server R2 |
Local Area Connection 7 |
9.79.100.12 |
| Red Hat 6.5 Server |
eth6 |
9.78.100.21 |
| Red Hat 6.5 Server |
eth7 |
9.79.100.22 |
| Red Hat 6.5 Server |
eth0 |
9.78.100.31 |
| Red Hat 6.5 Server |
eth1 |
9.79.100.32 |
Configure DUT and Test Equipment
This section describes the initial configuration of DUT and test equipment. In some test cases, additional configuration changes are made.
Task 1. Brocade VDX Switch Configuration – Logical Chassis Mode
On the VCS cluster, configure VLAN 978 and 979 interfaces on all cluster members. Configure the interfaces connecting to Windows and Red Hat hosts as well as the iSCSI storage array.
< ========== >
interface Vlan 978
interface Vlan 979
!
rbridge-id 1
interface Ve 978
ip address 9.78.1.1/16
no shutdown
!
interface Ve 979
ip address 9.79.1.1/16
no shutdown
!
!
rbridge-id 2
interface Ve 978
ip address 9.78.1.2/16
no shutdown
!
interface Ve 979
ip address 9.79.1.2/16
no shutdown
!
!
rbridge-id 11
interface Ve 978
ip address 9.78.1.11/16
no shutdown
!
interface Ve 979
ip address 9.79.1.11/16
no shutdown
!
!
rbridge-id 12
interface Ve 978
ip address 9.78.1.12/16
no shutdown
!
interface Ve 979
ip address 9.79.1.12/16
no shutdown
!
!
interface TenGigabitEthernet 1/0/1
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 978
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 1/0/2
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 978
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 1/0/3
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 978
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 1/0/4
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 978
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 2/0/1
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 2/0/2
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 2/0/3
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 2/0/4
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 11/1/11
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 978
spanning-tree shutdown
no shutdown
!
interface TenGigabitEthernet 12/0/11
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
spanning-tree shutdown
no shutdown
!
< ========== >
Task 2. EMC VNX Configuration
a. Configure Network Interface
Step 1. From the EMC Unisphere Dashboard, click on System.
EMC VNX Network Interface Configuration – Screen 1
Step 2. Click on Hardware.

EMC VNX Network Interface Configuration – Screen 2
Step 3. Click on Storage Hardware.
EMC VNX Network Interface Configuration – Screen 3
Step 4. Expand SPs -> SP A -> IO Modules -> desired slot and port. Right-click on the desired port and select Properties.
EMC VNX Network Interface Configuration – Screen 4
Step 5. Under Virtual Port Properties, click Add.

EMC VNX Network Interface Configuration – Screen 5
Step 6. In the IP Address field, enter the desired IP address. In the Gateway field, enter the desired default gateway. In the Subnet Mask field, enter the desired subnet mask. Click OK.

EMC VNX Network Interface Configuration – Screen 6
Step 7. Verify virtual port configuration under Virtual Port Properties. Note the port IQN identified under Port Identification – > IQN field. Click OK.

EMC VNX Network Interface Configuration – Screen 7
Step 8. Repeat steps 4-7 for an interface on VNX SP B.
b. Configure LUN
Step 1. From the EMC Unisphere Dashboard, click on Storage.
EMC VNX LUN Configuration – Screen 1
Step 2. Click on LUNs.
EMC VNX LUN Configuration – Screen 2
Step 3. Under the LUN tab, click on Create.
EMC VNX LUN Configuration – Screen 3
Step 4. In the Storage Pool for new LUN dropdown, select the desired storage pool with sufficient available space. Under LUN Properties, enter the desired capacity in the User Capacity field. The LUN ID field will automatically populate with the next available LUN ID value. Click Apply.

EMC VNX LUN Configuration – Screen 4
EMC Unisphere will display a LUN creation verification screen.

EMC VNX LUN Configuration – Screen 5
The previously created LUNs will now be displayed in the LUN table.
EMC VNX LUN Configuration – Screen 6
c. Configure Initiator
Step 1. Gather IQNs from each of the iSCSI initiators.
Windows 2012: Open the iSCSI Initiator administration tool. Click on the Configuration tab. Note the IQN identified under the Initiator Name field.

Windows 2012 Server iSCSI Initiator IQN Collection
Windows 2008: Open the iSCSI Initiator administration tool. Click on the Configuration tab. Note the IQN identified under the Initiator Name field.

Windows 2008 Server iSCSI Initiator IQN Collection
Red Hat 6.5:
< ========== >
[root@ilo65205 ~]# cat /etc/iscsi/initiatorname.iscsi
InitiatorName=iqn.1994-05.com.redhat:7dfed6681f9d
[root@rhel65066109 Desktop]# cat /etc/iscsi/initiatorname.iscsi
InitiatorName=iqn.1994-05.com.redhat:89f0253f16
< ========== >
Step 2. From the EMC Unisphere Dashboard, click on Hosts.
EMC VNX Initiator Configuration – Screen 1
Step 3. Click on Initiators.

EMC VNX Initiator Configuration – Screen 2
Step 4. Click on Create at the bottom, left.

EMC VNX Initiator Configuration – Screen 3
Step 5. Under the Initiator Information section and in the WWN/IQN field, enter the IQN from the iSCSI initiator recorded from previous steps. In the SP – port dropdown, select the iSCSI port that corresponds to the SP port previously configured. By default, EMC identifies the SP port in the last portion of the EMC-assigned IQN. In this example, the EMC VNX SP A port IQN of iqn.1992-04.com.emc:cx.apm00121201182.a7 corresponds to SP port “A-7v0 (iSCSI)”.
Under the Host Agent Information section and in the Host Name field, enter a desired host name. In the IP Address field, enter the iSCSI initiator’s IP address. Click OK.

EMC VNX Initiator Configuration – Screen 4
Step 6. Repeat steps 4-5, using EMC VNX SP B. In this example, the EMC VNX SP B port IQN of iqn.1992-04.com.emc:cx.apm00121201182.b7 corresponds to SP port “B-7v0 (iSCSI)”. Verify the IP address used in the IP Address field matches the entry used in step 5.

EMC VNX Initiator Configuration – Screen 5

EMC VNX Initiator Verification
d. Configure Storage Group
Step 1. From the EMC Unisphere Dashboard, click on Hosts.

EMC VNX Storage Group Configuration – Screen 1
Step 2. Click on Storage Groups.

EMC VNX Storage Group Configuration – Screen 2
Step 3. Click on Create at the bottom, left.

EMC VNX Storage Group Configuration – Screen 3
Step 4. In the Storage Group Name, enter the host name previously configured for the iSCSI initiator during initiator record creation. Click OK.

EMC VNX Storage Group Configuration – Screen 4
Step 5. When prompted to add LUNs or connect hosts, click Yes.

EMC VNX Storage Group Configuration – Screen 5
Step 6. Under the LUNs tab, find the desired LUN to assign to the storage group and click Add. The LUN may be listed under either SP. (This does not correspond to the SP configured with the network IP address from previous steps.)

EMC VNX Storage Group Configuration – Screen 6
Step 7. Under the Hosts tab, select the desired host to add to the storage group and click the right arrow. Click OK.

EMC VNX Storage Group Configuration – Screen 7

EMC VNX Storage Group Verification

EMC VNX Initiator in Storage Group Verification
Task 3. Windows Server Host Configuration
a. Configure iSCSI Initiator
———-
Note: The following screenshots are applicable to both Windows 2012 R2 and Windows 2008 R2. Only Windows 2012 R2 screenshots are shown.
———-
Step 1. In Windows, open the iSCSI Initiator administration tool. Navigate to the Discovery tab and click on Discover Portal.

Windows Server iSCSI Initiator Target Discovery – Screen 1
Step 2. In the IP Address or DNS name field, enter the IP address of the storage target. Click Advanced.

Windows Server iSCSI Initiator Target Discovery – Screen 2
Step 3. In the Local adapter field, select Microsoft iSCSI Initiator. In the Initiator IP field, select the IP address of the first desired NIC IP address. Click OK and click OK again.

Windows Server iSCSI Initiator Target Discovery – Screen 3
Step 4. In the iSCSI Initiator Properties window, navigate to the Targets tab. Highlight the first target IQN and click Connect.

Windows Server iSCSI Initiator Target Discovery – Screen 4
Step 5. Check the Enable multi-path box. Click Advanced.

Windows Server iSCSI Initiator Target Discovery – Screen 5
Step 6. In the Local adapter field, select Microsoft iSCSI Initiator. In the Initiator IP field, select the IP address of the first desired NIC IP address. In the Target portal IP, select the appropriate target IP address. Click OK and click OK again.

Windows Server iSCSI Initiator Target Discovery – Screen 6
Step 7. Repeat steps 1-6 for the second target IQN with the second desired NIC IP address.
b. Verify EMC PowerPath Multi-Path
Step 1. Install EMC PowerPath. Refer to EMC’s website for details on installation and management.
Step 2. Open PowerPath Administrator. Expand EMC PowerPath Admin – > Path Management – > Disks and click on the desired disk. Verify each of the paths to the disk listed as:
Status: Optimal
State: Alive
Mode: active

EMC PowerPath Verification
c. Initialize Disk and File System
Step 1. Within Windows Server, open the Computer Management administration tool and navigate to Disk Management.

Windows Disk Initialization – Screen 1
Step 2. Bring the disk online by right-clicking on the disk label and click Online.

Windows Disk Initialization – Screen 2
Step 3. Initialize the disk by right-clicking on the disk label and click Initialize Disk.

Windows Disk Initialization – Screen 3
Step 4. Select the desired disk(s) to bring online. Click OK.

Windows Disk Initialization – Screen 4

Windows Disk Initialization Verification
Step 5. Right-click within the unallocated space on the desired disk and click New Simple Volume. Then, click Next.

Windows File System Configuration – Screen 1
Step 6. With the Simple volume size in MB field, enter the desired volume size. Click Next.

Windows File System Configuration – Screen 2
Step 7. Select the desired drive letter for the new volume. Click Next.

Windows File System Configuration – Screen 3
Step 8. Within the Volume label field, enter the desired volume name. Click Next. Click Finish.

Windows File System Configuration – Screen 4

Windows File System Verification
Task 4. Configure Red Hat Linux Host
a. Configure iSCSI Initiator
Step 1. Start the iscsid and iscsi daemons and enable automatic start on boot.
< ========== >
[root@rhel65079141 ~]# chkconfig iscsid on
[root@rhel65079141 ~]# service iscsid start
[root@rhel65079141 ~]# chkconfig iscsi on
[root@rhel65079141 ~]# service iscsi start
< ========== >
Step 2. Discover the iSCSI targets.
< ========== >
[root@rhel65079141 ~]# iscsiadm -m discovery -t st -p
9.78.50.178.50.1:3260,5 iqn.1992-04.com.emc:cx.apm00121201182.a7 < ==
192.168.224.10:3260,3 iqn.1992-04.com.emc:cx.apm00121201182.a6
192.168.224.11:3260,4 iqn.1992-04.com.emc:cx.apm00121201182.b6
9.79.50.2:3260,6 iqn.1992-04.com.emc:cx.apm00121201182.b7 < ==
[root@rhel65079141 ~]# iscsiadm -m discovery -t st -p 9.79.50.2
9.79.50.2:3260,6 iqn.1992-04.com.emc:cx.apm00121201182.b7 < ==
192.168.224.11:3260,4 iqn.1992-04.com.emc:cx.apm00121201182.b6
192.168.224.10:3260,3 iqn.1992-04.com.emc:cx.apm00121201182.a6
9.78.50.1:3260,5 iqn.1992-04.com.emc:cx.apm00121201182.a7 < ==
< ========== >
Step 3. Log in to the target using the iSCSI target IP address.
< ========== >
[root@rhel65079141 ~]# iscsiadm -m node -T iqn.1992-04.com.emc:cx.apm00121201182.a7 -p 9.78.50.1 -l
[root@rhel65079141 ~]# iscsiadm -m node -T iqn.1992-04.com.emc:cx.apm00121201182.b7 -p 9.79.50.2 –l
< ========== >
Step 4. Configure automatic iSCSI login on boot.
< ========== >
[root@rhel65079141 ~]# iscsiadm -m node -T iqn.1992-04.com.emc:cx.apm00121201182.a7 -p 9.78.50.1 –op update -n node.startup -v automatic
[root@rhel65079141 ~]# iscsiadm -m node -T iqn.1992-04.com.emc:cx.apm00121201182.b7 -p 9.79.50.2 –op update -n node.startup -v automatic
< ========== >
Step 5. Reboot the RHEL host.
Step 6. Verify iSCSI connectivity.
< ========== >
[root@rhel65079141 ~]# iscsiadm -m session
tcp: [1] 9.79.50.2:3260,6 iqn.1992-04.com.emc:cx.apm00121201182.b7 < ==
tcp: [2] 9.78.50.1:3260,5 iqn.1992-04.com.emc:cx.apm00121201182.a7 < ==
< ========== >
b. Initalize Disk and Partition
Step 1. Verify iSCSI LUN exist and contain partitions. If partitions do not exist, refer to OS documentation for partition creation.
< ========== >
[root@rhel65079141 ~]# fdisk -l
…
Disk /dev/sdb: 21.5 GB, 21474836480 bytes < ==
64 heads, 32 sectors/track, 20480 cylinders
Units = cylinders of 2048 * 512 = 1048576 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/sdc: 21.5 GB, 21474836480 bytes < ==
64 heads, 32 sectors/track, 20480 cylinders
Units = cylinders of 2048 * 512 = 1048576 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
< ========== >
c. Verify EMC PowerPath Multi-Path
Step 1. Install EMC PowerPath. Refer to EMC’s website for details on installation and management. Reboot OS if required.
Step 2. From a command prompt, instruct PowerPath to detect and configure any new devices.
< ========== >
[root@rhel65079141 ~]# powermt config
< ========== >
Step 3. Verify each of the paths to the disk listed as:
Mode: active
State: alive
< ========== >
[root@rhel65079141 ~]# powermt display dev=all
Pseudo name=emcpowera
VNX ID=APM00121201182 [ec079141i]
Logical device ID=6006016046A230005EDA485D123DE411 [LUN 202]
state=alive; policy=CLAROpt; queued-IOs=0
Owner: default=SP B, current=SP B Array failover mode: 4
==============================================================================
————— Host ————— – Stor – — I/O Path — — Stats —
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
7 iSCSI Initiator over TCP/I sdb SP B7 active alive 0 0 < ==
8 iSCSI Initiator over TCP/I sdc SP A7 active alive 0 0 < ==
< ========== >
Step 4. Save PowerPath configuration.
< ========== >
[root@rhel65079141 ~]# powermt save
< ========== >
d. Initalize File System
Step 1. Create a file system on the PowerPath pseudo-device.
< ========== >
[root@rhel65079141 dev]# mkfs -t ext3 /dev/emcpowera
mke2fs 1.41.12 (17-May-2010)
Filesystem label=
OS type: Linux
Block size=4096 (log=2)
Fragment size=4096 (log=2)
Stride=0 blocks, Stripe width=0 blocks
1310720 inodes, 5242880 blocks
262144 blocks (5.00%) reserved for the super user
First data block=0
Maximum filesystem blocks=4294967296
160 block groups
32768 blocks per group, 32768 fragments per group
8192 inodes per group
Superblock backups stored on blocks:
32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208,
4096000
Writing inode tables: done
Creating journal (32768 blocks): done
Writing superblocks and filesystem accounting information: done
This filesystem will be automatically checked every 32 mounts or
180 days, whichever comes first. Use tune2fs -c or -i to override.
< ========== >
Step 2. Create a mount point for the PowerPath pseudo-device in a desired location.
< ========== >
[root@rhel65079141 mnt]# pwd
/mnt
[root@rhel65079141 mnt]# mkdir vnx5300
< ========== >
Step 3. Mount the file system to the mount point previously created.
< ========== >
[root@rhel65079141 mnt]# mount -t ext3 /dev/emcpowera /mnt/vnx5300
[root@rhel65079141 mnt]# mount
…
/dev/emcpowera on /mnt/vnx5300 type ext3 (rw) < ==
[root@rhel65079141 mnt]#
< ========== >
Task 5. Medusa I/O Generation Configuration
The following shows the configuration of the Medusa IO data patterns used in the test cases.
a. Standard I/O with 64KB Blocks
For test scenarios using standard I/O (64 KB testing size), configure Medusa for the following command line result:
< =========== >
pain 2GB -b64KB -v180 -H1 -l35 -Y1 -M180
< =========== >
b. Maximum I/O with 1 MB Blocks
For test scenarios using maximum bandwidth I/O (1 MB testing size), configure Medusa for the following command line result:
< =========== >
maim 2GB -b1MB -Q8 -n -u -q5 -w -o -Y1 -M180 –perf-mode
< =========== >
c. Microsoft Exchange Server I/O
For test scenarios using Microsoft Exchange Server Simulation (Medusa pre-configured Storage Simulation test), configure Medusa for the following command line result:
< =========== >
maim 2GB -%r50@4KB -%w25@64KB -%w25@4KB -Q1 -n -u -q5 -o -l69 -Y1 -M180
< =========== >
d. File Server I/O
For test scenarios using File Server Simulation (Medusa pre-configured Storage Simulation test), configure Medusa for the following command line result:
< =========== >
maim 2GB -%10:r80@512B,w20@512B -%5:r80@1KB,w20@1KB -%5:r80@2KB,w20@2KB -%60:r80@4KB,w20@4KB -%2:r50@8KB,w50@8KB -%4:r75@16KB,w25@16KB -%4:r75@32KB,w25@32KB -%10:r80@64KB,w20@64KB -Q1 -n -u -o -l69 -Y1 -M180
< =========== >
Test Cases
These test cases are designed to verify both basic and advanced functionality features between the Brocade VCS cluster and both host and storage devices, as well as stress all devices and confirm successful error recovery.
| 1.1 |
Ethernet Storage – Base Functionality |
| 1.1.1 |
Storage Device – Physical Connection and Speed Negotiation |
| 1.1.2 |
iSCSI Connectivity |
| 1.1.3 |
vLAG Configuration |
| 1.2 |
Ethernet Storage – Advanced Functionality |
| 1.2.1 |
Storage Device – Jumbo Frame/MTU Size Validation |
| 1.2.2 |
iSCSI Bandwidth Validation |
| 1.2.3 |
Storage Device – w/Congested Fabric |
| 1.2.4 |
Storage Device – iSCSI Protocol Jammer CRC Corruption on ISL Port |
| 1.2.5 |
Storage Device – iSCSI Protocol Jammer Packet Corruption on ISL Port |
| 1.2.6 |
Storage Device – iSCSI Protocol Jammer Packet Loss on ISL Port |
| 1.2.7 |
Storage Device – iSCSI Protocol Jammer CRC Corruption on Storage Port |
| 1.2.8 |
Storage Device – iSCSI Protocol Jammer Packet Corruption on Storage Port |
| 1.2.9 |
Storage Device – iSCSI Protocol Jammer Packet Loss on Storage Port |
| 1.3 |
Stress and Error Recovery |
| 1.3.1 |
Storage Device Fabric I/O Integrity – Congested Fabric |
| 1.3.2 |
Storage Device Integrity – Device Recovery from Port Toggle – Manual Cable Pull |
| 1.3.3 |
Storage Device Integrity – Device Recovery from ISL Port Relocation |
| 1.3.4 |
Storage Device Stress – Device Recovery from Device Port Toggle – Extended Run |
| 1.3.5 |
Storage Device Recovery – ISL Port Toggle – Extended Run |
| 1.3.6 |
Storage Device Recovery – All-ISL Port Toggle |
| 1.3.7 |
Storage Device Recovery – VDX8770 Line Card Maintenance |
| 1.3.8 |
Storage Device Recovery – Switch Offline |
| 1.3.9 |
Workload Simulation Test Suite – Microsoft Exchange Server |
| 1.3.10 |
Workload Simulation Test Suite – File Server Simulation |
Test Case Descriptions
Lines marked with ‘< ==’ indicate important output used to validate the test results.
1.1 Ethernet Storage – Base Functionality
1.1.1 Storage Device – Physical Connection and Speed Negotiation
Test Objective
Verify device connectivity to VDX switch with all supported speed settings. Configure storage port for iSCSI connectivity. Validate base connectivity.
Results
PASS. All interfaces up and operational at expected speed settings. Brocade VDX ‘show interface’ commands reflect expected interface speed.
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected) < ==
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit < ==
LineSpeed Configured : Auto, Duplex: Full
…
CASTOR-VCS20-2# show interface tengigabitethernet 12/0/11
TenGigabitEthernet 12/0/11 is up, line protocol is up (connected) < ==
Hardware is Ethernet, address is 0005.3365.2971
Current address is 0005.3365.2971
Pluggable media present
Interface index (ifindex) is 51942621194
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit < ==
LineSpeed Configured : Auto, Duplex: Full
…
< =========== >
1.1.2 iSCSI Connectivity
Test Objective
Verify Initiator-to-Target/LUN connectivity with iSCSI.
Results
PASS. All iSCSI targets on Windows and Red Hat hosts are successfully connected.
< =========== >
PS C:\Users\Administrator> iscsicli SessionList
Microsoft iSCSI Initiator Version 6.3 Build 9600
Total of 2 sessions
Session Id : ffffe001561ee020-4000013700000001
Initiator Node Name : iqn.1991-05.com.microsoft:brcd068114.englab.brocade.com
Target Node Name : (null)
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.a7
ISID : 40 00 01 37 00 01
TSID : 29 0d
Number Connections : 1
Connections:
Connection Id : ffffe001561ee020-0
Initiator Portal : 9.78.100.1/192
Target Portal : 9.78.50.1/3260
CID : 01 00
Devices:
Device Type : Disk
Device Number : 1
Storage Device Type : 7
Partition Number : 0
Friendly Name : DGC VRAID SCSI Disk Device
Device Description : Disk drive
Reported Mappings : Port 13, Bus 0, Target Id 0, LUN 0
Location : Bus Number 0, Target Id 0, LUN 0
Initiator Name : ROOT\ISCSIPRT\0000_0
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.a7
Device Interface Name : \\?\scsi#disk&ven_emc&prod_power&#{4a54205a-c920-4e28-88c5-9a6296a74b0b}&emcp&power3#{53f56307-b6bf-11d0-94f2-00a0c91efb8b}
Legacy Device Name : \\.\PhysicalDrive1
Device Instance : 0x2
Volume Path Names :
N:\
Session Id : ffffe001561ee020-4000013700000002
Initiator Node Name : iqn.1991-05.com.microsoft:brcd068114.englab.brocade.com
Target Node Name : (null)
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.b7
ISID : 40 00 01 37 00 01
TSID : 00 0e
Number Connections : 1
Connections:
Connection Id : ffffe001561ee020-1
Initiator Portal : 9.79.100.2/448
Target Portal : 9.79.50.2/3260
CID : 01 00
Devices:
Device Type : Disk
Device Number : 1
Storage Device Type : 7
Partition Number : 0
Friendly Name : DGC VRAID SCSI Disk Device
Device Description : Disk drive
Reported Mappings : Port 13, Bus 0, Target Id 1, LUN 0
Location : Bus Number 0, Target Id 1, LUN 0
Initiator Name : ROOT\ISCSIPRT\0000_0
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.b7
Device Interface Name : \\?\scsi#disk&ven_emc&prod_power&#{4a54205a-c920-4e28-88c5-9a6296a74b0b}&emcp&power3#{53f56307-b6bf-11d0-94f2-00a0c91efb8b}
Legacy Device Name : \\.\PhysicalDrive1
Device Instance : 0x3
Volume Path Names :
N:\
The operation completed successfully.
< =========== >
< =========== >
C:\Users\Administrator>iscsicli SessionList
Microsoft iSCSI Initiator Version 6.1 Build 7601
Total of 2 sessions
Session Id : fffffa800ae57018-4000013700000001
Initiator Node Name : iqn.1991-05.com.microsoft:brcd068128.englab.brocade.com
Target Node Name : (null)
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.a7
ISID : 40 00 01 37 00 01
TSID : d8 0e
Number Connections : 1
Connections:
Connection Id : fffffa800ae57018-0
Initiator Portal : 9.78.100.11/192
Target Portal : 9.78.50.1/3260
CID : 01 00
Devices:
Device Type : Disk
Device Number : 1
Storage Device Type : 7
Partition Number : 0
Friendly Name : PowerPath Devices
Device Description : Disk drive
Reported Mappings : Port 5, Bus 0, Target Id 0, LUN 0
Location : Bus Number 0, Target Id 0, LUN 0
Initiator Name : Root\ISCSIPRT\0000_0
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.a7
Device Interface Name : \\?\scsi#disk&ven_emc&prod_power&#{4a54205a-c92
0-4e28-88c5-9a6296a74b0b}&emcp&power5#{53f56307-b6bf-11d0-94f2-00a0c91efb8b}
Legacy Device Name : \\.\PhysicalDrive1
Device Instance : 0x82c
Volume Path Names :
N:\
Session Id : fffffa800ae57018-4000013700000002
Initiator Node Name : iqn.1991-05.com.microsoft:brcd068128.englab.brocade.com
Target Node Name : (null)
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.b7
ISID : 40 00 01 37 00 01
TSID : 94 0d
Number Connections : 1
Connections:
Connection Id : fffffa800ae57018-1
Initiator Portal : 9.79.100.12/448
Target Portal : 9.79.50.2/3260
CID : 01 00
Devices:
Device Type : Disk
Device Number : 1
Storage Device Type : 7
Partition Number : 0
Friendly Name : PowerPath Devices
Device Description : Disk drive
Reported Mappings : Port 5, Bus 0, Target Id 1, LUN 0
Location : Bus Number 0, Target Id 1, LUN 0
Initiator Name : Root\ISCSIPRT\0000_0
Target Name : iqn.1992-04.com.emc:cx.apm00121201182.b7
Device Interface Name : \\?\scsi#disk&ven_emc&prod_power&#{4a54205a-c92
0-4e28-88c5-9a6296a74b0b}&emcp&power5#{53f56307-b6bf-11d0-94f2-00a0c91efb8b}
Legacy Device Name : \\.\PhysicalDrive1
Device Instance : 0x894
Volume Path Names :
N:\
The operation completed successfully.
< =========== >
< =========== >
[root@rhel65079141 ~]# iscsiadm -m session
tcp: [1] 9.78.50.1:3260,5 iqn.1992-04.com.emc:cx.apm00121201182.a7
tcp: [2] 9.79.50.2:3260,6 iqn.1992-04.com.emc:cx.apm00121201182.b7
< =========== >
< =========== >
[root@rhel65066109 ~]# iscsiadm -m session
tcp: [1] 9.78.50.1:3260,5 iqn.1992-04.com.emc:cx.apm00121201182.a7
tcp: [2] 9.79.50.2:3260,6 iqn.1992-04.com.emc:cx.apm00121201182.b7
< =========== >
1.1.3 Multipath Connectivity
Test Objective
Verify multipath connectivity from initiator ports through VCS cluster to target LUNs.
Results
PASS. All port-channel interfaces are operational and in sync.
< =========== >
PS C:\Users\Administrator> powermt display dev=all
Pseudo name=harddisk1
VNX ID=APM00121201182 [ec068114i]
Logical device ID=6006016046A2300002D22C900A3DE411 [LUN 200]
state=alive; policy=CLAROpt; queued-IOs=0
Owner: default=SP B, current=SP B Array failover mode: 4
==============================================================================
————— Host ————— – Stor – — I/O Path — — Stats —
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
13 port13\path0\tgt1\lun0 c13t1d0 SP B7 active alive 0 0 < ==
13 port13\path0\tgt0\lun0 c13t0d0 SP A7 active alive 0 0 < ==
< =========== >
< =========== >
C:\Users\Administrator>powermt display dev=all
Pseudo name=harddisk1
VNX ID=APM00121201182 [ec068128i]
Logical device ID=6006016046A2300004D22C900A3DE411 [LUN 201]
state=alive; policy=CLAROpt; queued-IOs=0
Owner: default=SP B, current=SP B Array failover mode: 4
==============================================================================
————— Host ————— – Stor – — I/O Path — — Stats —
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
5 port5\path0\tgt1\lun0 c5t1d0 SP B7 active alive 0 0 < ==
5 port5\path0\tgt0\lun0 c5t0d0 SP A7 active alive 0 0 < ==
< =========== >
< =========== >
[root@rhel65079141 scripts]# powermt display dev=all
Pseudo name=emcpowera
VNX ID=APM00121201182 [ec079141i]
Logical device ID=6006016046A230005EDA485D123DE411 [LUN 202]
state=alive; policy=CLAROpt; queued-IOs=0
Owner: default=SP B, current=SP B Array failover mode: 4
==============================================================================
————— Host ————— – Stor – — I/O Path — — Stats —
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
5 iSCSI Initiator over TCP/I sdb SP A7 active alive 0 0 < ==
6 iSCSI Initiator over TCP/I sdc SP B7 active alive 0 0 < ==
< =========== >
< =========== >
[root@rhel65079141 scripts]# powermt display dev=all
Pseudo name=emcpowera
VNX ID=APM00121201182 [ec079141i]
Logical device ID=6006016046A230005EDA485D123DE411 [LUN 202]
state=alive; policy=CLAROpt; queued-IOs=0
Owner: default=SP B, current=SP B Array failover mode: 4
==============================================================================
————— Host ————— – Stor – — I/O Path — — Stats —
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
5 iSCSI Initiator over TCP/I sdb SP A7 active alive 0 0 < ==
6 iSCSI Initiator over TCP/I sdc SP B7 active alive 0 0 < ==
< =========== >
1.2 Ethernet Storage – Advanced Functionality
1.2.1 Storage Device – Jumbo Frame/MTU Size Validation
Test Objective
Perform I/O validation testing using maximum MTU possible supported by all devices in the I/O path.
Test Configuration
1. On EMC VNX 5300, reconfigure iSCSI port to increase MTU to 9000.

EMC VNX 5300 Port MTU Reconfiguration
2. For Red Hat hosts, add “MTU=9000” to /etc/sysconfig/network-scripts/ifcfg-ethX and then execute ‘service network restart’.
< =========== >
[root@rhel65079141 ~]# more /etc/sysconfig/network-scripts/ifcfg-eth0
DEVICE=eth0
IPADDR=9.78.100.21
NETMASK=255.255.0.0
BOOTPROTO=none
MTU=9000
ONBOOT=yes
USERCTL=no
NM_CONTROLLED=no
< =========== >
3. For Windows 2012 Server host, reconfigure NIC properties for maximum MTU:

Windows 2012 Server Emulex NIC MTU Configuration
4. For Windows 2008 Server host, reconfigure NIC properties for maximum MTU:

Windows 2008 Server Brocade NIC MTU Configuration
5. On VDX devices, configure as follows:
< =========== >
rbridge-id 1
interface ve 978
ip mtu 9018
rbridge-id 2
interface ve 978
ip mtu 9018
rbridge-id 11
interface ve 978
ip mtu 9018
rbridge-id 12
interface ve 978
ip mtu 9018
interface TenGigabitEthernet 1/0/1
mtu 9216
interface TenGigabitEthernet 1/0/2
mtu 9216
interface TenGigabitEthernet 1/0/3
mtu 9216
interface TenGigabitEthernet 1/0/4
mtu 9216
interface TenGigabitEthernet 11/1/11
mtu 9216
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All initiator and target interfaces pass jumbo frames successfully.
< =========== >
PS C:\Users\Administrator> ping 9.78.50.1 -l 8972 -f
Pinging 9.78.50.1 with 8972 bytes of data:
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Ping statistics for 9.78.50.1:
Packets: Sent = 4, Received = 4, Lost = 0 (0% loss), < ==
Approximate round trip times in milli-seconds:
Minimum = 0ms, Maximum = 0ms, Average = 0ms
< =========== >
< =========== >
C:\Users\Administrator>ping 9.78.50.1 -l 8972 -f
Pinging 9.78.50.1 with 8972 bytes of data:
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Reply from 9.78.50.1: bytes=8972 time<1ms TTL=128
Ping statistics for 9.78.50.1:
Packets: Sent = 4, Received = 4, Lost = 0 (0% loss), < ==
Approximate round trip times in milli-seconds:
Minimum = 0ms, Maximum = 0ms, Average = 0ms
< =========== >
< =========== >
[root@rhel65079141 ~]# ping -c 4 -M do -s 8972 9.78.50.1
PING 9.78.50.1 (9.78.50.1) 8972(9000) bytes of data.
8980 bytes from 9.78.50.1: icmp_seq=1 ttl=128 time=0.469 ms
8980 bytes from 9.78.50.1: icmp_seq=2 ttl=128 time=0.297 ms
8980 bytes from 9.78.50.1: icmp_seq=3 ttl=128 time=0.299 ms
8980 bytes from 9.78.50.1: icmp_seq=4 ttl=128 time=0.256 ms
— 9.78.50.1 ping statistics —
4 packets transmitted, 4 received, 0% packet loss, time 3000ms < ==
rtt min/avg/max/mdev = 0.256/0.330/0.469/0.082 ms
< =========== >
< =========== >
[root@rhel65066109 ~]# ping -c 4 -M do -s 8972 9.78.50.1
PING 9.78.50.1 (9.78.50.1) 8972(9000) bytes of data.
8980 bytes from 9.78.50.1: icmp_seq=1 ttl=128 time=0.304 ms
8980 bytes from 9.78.50.1: icmp_seq=2 ttl=128 time=0.266 ms
8980 bytes from 9.78.50.1: icmp_seq=3 ttl=128 time=0.281 ms
8980 bytes from 9.78.50.1: icmp_seq=4 ttl=128 time=0.273 ms
— 9.78.50.1 ping statistics —
4 packets transmitted, 4 received, 0% packet loss, time 2999ms < ==
rtt min/avg/max/mdev = 0.266/0.281/0.304/0.014 ms
< =========== >

Jumbo Frame MTU Size Medusa Results
< =========== >
M4-VCS20-11# show interface TenGigabitEthernet 1/0/1
TenGigabitEthernet 1/0/1 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1945
Current address is 50eb.1a22.1945
Pluggable media present
Interface index (ifindex) is 4697653248
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:03
Queueing strategy: fifo
Receive Statistics:
5425636 packets, 28969383340 bytes
Unicasts: 5423214, Multicasts: 2350, Broadcasts: 73
64-byte pkts: 1073483, Over 64-byte pkts: 862364, Over 127-byte pkts: 83
Over 255-byte pkts: 0, Over 511-byte pkts: 195, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 3489515 < ==
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
9191210 packets, 28581154883 bytes
Unicasts: 9191082, Multicasts: 44, Broadcasts: 84
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 530.027552 Mbits/sec, 9016 packets/sec, 5.30% of line-rate
Output 6.971712 Mbits/sec, 12301 packets/sec, 0.07% of line-rate
Time since last interface status change: 18:33:17
M4-VCS20-11# show interface TenGigabitEthernet 1/0/2
TenGigabitEthernet 1/0/2 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1946
Current address is 50eb.1a22.1946
Pluggable media present
Interface index (ifindex) is 4697686017
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:04
Queueing strategy: fifo
Receive Statistics:
4837967 packets, 25659123695 bytes
Unicasts: 4837852, Multicasts: 43, Broadcasts: 72
64-byte pkts: 1000167, Over 64-byte pkts: 746962, Over 127-byte pkts: 15
Over 255-byte pkts: 0, Over 511-byte pkts: 199, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 3090624 < ==
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
7987176 packets, 24169400648 bytes
Unicasts: 7987034, Multicasts: 57, Broadcasts: 85
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 493.244224 Mbits/sec, 8388 packets/sec, 4.93% of line-rate
Output 6.463616 Mbits/sec, 11401 packets/sec, 0.06% of line-rate
Time since last interface status change: 00:54:15
M4-VCS20-11# show interface TenGigabitEthernet 1/0/3
TenGigabitEthernet 1/0/3 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1947
Current address is 50eb.1a22.1947
Pluggable media present
Interface index (ifindex) is 4697718786
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:04
Queueing strategy: fifo
Receive Statistics:
6557275 packets, 29104828732 bytes
Unicasts: 6557240, Multicasts: 29, Broadcasts: 6
64-byte pkts: 69, Over 64-byte pkts: 3061460, Over 127-byte pkts: 0
Over 255-byte pkts: 2148, Over 511-byte pkts: 254, Over 1023-byte pkts: 2161
Over 1518-byte pkts(Jumbo): 3491183 < ==
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
25884577 packets, 29792715033 bytes
Unicasts: 25884355, Multicasts: 71, Broadcasts: 151
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 528.847672 Mbits/sec, 10030 packets/sec, 5.29% of line-rate
Output 8.251168 Mbits/sec, 13346 packets/sec, 0.08% of line-rate
Time since last interface status change: 00:58:01
M4-VCS20-11# show interface TenGigabitEthernet 1/0/4
TenGigabitEthernet 1/0/4 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1948
Current address is 50eb.1a22.1948
Pluggable media present
Interface index (ifindex) is 4697751555
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:04
Queueing strategy: fifo
Receive Statistics:
6719414 packets, 29077120204 bytes
Unicasts: 6719379, Multicasts: 30, Broadcasts: 6
64-byte pkts: 66, Over 64-byte pkts: 3223100, Over 127-byte pkts: 30
Over 255-byte pkts: 4932, Over 511-byte pkts: 254, Over 1023-byte pkts: 2678
Over 1518-byte pkts(Jumbo): 3488355 < ==
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
25828624 packets, 29788735513 bytes
Unicasts: 25828402, Multicasts: 71, Broadcasts: 151
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 532.040584 Mbits/sec, 10105 packets/sec, 5.32% of line-rate
Output 8.236296 Mbits/sec, 13309 packets/sec, 0.08% of line-rate
Time since last interface status change: 18:33:18
M4-VCS20-11# show interface TenGigabitEthernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:04
Queueing strategy: fifo
Receive Statistics:
68843506 packets, 112327785368 bytes
Unicasts: 68843517, Multicasts: 0, Broadcasts: 0
64-byte pkts: 9216784, Over 64-byte pkts: 14138699, Over 127-byte pkts: 702
Over 255-byte pkts: 838508, Over 511-byte pkts: 13637, Over 1023-byte pkts: 38338599
Over 1518-byte pkts(Jumbo): 6296577 < ==
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
23502202 packets, 112571124807 bytes
Unicasts: 23501982, Multicasts: 71, Broadcasts: 157
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 29.687392 Mbits/sec, 49946 packets/sec, 0.30% of line-rate
Output 2068.803536 Mbits/sec, 37267 packets/sec, 20.69% of line-rate
Time since last interface status change: 01:03:28
< =========== >
1.2.2 iSCSI Bandwidth Validation
Test Objective
- Validate maximum sustained bandwidth to storage port via iSCSI.
- After 15 minutes, verify I/O completes error-free.
Results
- PASS. All I/O completes error-free.
- PASS. All storage LUNs show utilization at maximum capacity.
< =========== >
CASTOR-VCS20-2# show interface TenGigabitEthernet 1/0/1
TenGigabitEthernet 1/0/1 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1945
Current address is 50eb.1a22.1945
Pluggable media present
Interface index (ifindex) is 4697653248
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
3175 packets, 347397 bytes
Unicasts: 611, Multicasts: 2477, Broadcasts: 87
64-byte pkts: 2580, Over 64-byte pkts: 349, Over 127-byte pkts: 48
Over 255-byte pkts: 0, Over 511-byte pkts: 198, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
1255 packets, 344640 bytes
Unicasts: 1106, Multicasts: 47, Broadcasts: 102
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 0.002312 Mbits/sec, 3 packets/sec, 0.00% of line-rate
Output 0.000680 Mbits/sec, 1 packets/sec, 0.00% of line-rate
Time since last interface status change: 2d19h08m
CASTOR-VCS20-2# show interface TenGigabitEthernet 2/0/1
TenGigabitEthernet 2/0/1 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a20.d610
Current address is 50eb.1a20.d610
Pluggable media present
Interface index (ifindex) is 8992620544
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:43
Queueing strategy: fifo
Receive Statistics:
76856641 packets, 116122234237 bytes
Unicasts: 76854084, Multicasts: 2474, Broadcasts: 83
64-byte pkts: 60229, Over 64-byte pkts: 222736, Over 127-byte pkts: 61344
Over 255-byte pkts: 6536, Over 511-byte pkts: 67426, Over 1023-byte pkts: 76438370
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
61847361 packets, 3994317176 bytes
Unicasts: 61846772, Multicasts: 95, Broadcasts: 494
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 973.596032 Mbits/sec, 80515 packets/sec, 9.74% of line-rate < ==
Output 33.914432 Mbits/sec, 65604 packets/sec, 0.34% of line-rate
Time since last interface status change: 2d19h08m
CASTOR-VCS20-2# show interface TenGigabitEthernet 1/0/2
TenGigabitEthernet 1/0/2 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1946
Current address is 50eb.1a22.1946
Pluggable media present
Interface index (ifindex) is 4697686017
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
751 packets, 183875 bytes
Unicasts: 616, Multicasts: 47, Broadcasts: 88
64-byte pkts: 181, Over 64-byte pkts: 350, Over 127-byte pkts: 17
Over 255-byte pkts: 0, Over 511-byte pkts: 203, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
1294 packets, 354656 bytes
Unicasts: 1111, Multicasts: 82, Broadcasts: 101
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 0.000416 Mbits/sec, 0 packets/sec, 0.00% of line-rate
Output 0.000000 Mbits/sec, 0 packets/sec, 0.00% of line-rate
Time since last interface status change: 01:48:54
CASTOR-VCS20-2# show interface TenGigabitEthernet 2/0/2
TenGigabitEthernet 2/0/2 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a20.d611
Current address is 50eb.1a20.d611
Pluggable media present
Interface index (ifindex) is 8992653313
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:43
Queueing strategy: fifo
Receive Statistics:
95662630 packets, 144592644147 bytes
Unicasts: 95662505, Multicasts: 47, Broadcasts: 78
64-byte pkts: 206408, Over 64-byte pkts: 54959, Over 127-byte pkts: 143044
Over 255-byte pkts: 11616, Over 511-byte pkts: 34744, Over 1023-byte pkts: 95211859
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
73075648 packets, 4721710360 bytes
Unicasts: 73075019, Multicasts: 130, Broadcasts: 499
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 1190.706552 Mbits/sec, 98480 packets/sec, 11.91% of line-rate < ==
Output 38.916168 Mbits/sec, 75292 packets/sec, 0.39% of line-rate
Time since last interface status change: 01:48:55
CASTOR-VCS20-2# show interface TenGigabitEthernet 1/0/3
TenGigabitEthernet 1/0/3 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1947
Current address is 50eb.1a22.1947
Pluggable media present
Interface index (ifindex) is 4697718786
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
1308 packets, 250352 bytes
Unicasts: 1270, Multicasts: 30, Broadcasts: 8
64-byte pkts: 76, Over 64-byte pkts: 964, Over 127-byte pkts: 0
Over 255-byte pkts: 0, Over 511-byte pkts: 268, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
1588 packets, 268041 bytes
Unicasts: 1309, Multicasts: 98, Broadcasts: 181
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 0.000752 Mbits/sec, 1 packets/sec, 0.00% of line-rate
Output 0.000752 Mbits/sec, 1 packets/sec, 0.00% of line-rate
Time since last interface status change: 02:40:03
CASTOR-VCS20-2# show interface TenGigabitEthernet 2/0/3
TenGigabitEthernet 2/0/3 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a20.d612
Current address is 50eb.1a20.d612
Pluggable media present
Interface index (ifindex) is 8992686082
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:43
Queueing strategy: fifo
Receive Statistics:
51227727 packets, 76068405962 bytes
Unicasts: 51227689, Multicasts: 30, Broadcasts: 8
64-byte pkts: 76, Over 64-byte pkts: 795395, Over 127-byte pkts: 41268
Over 255-byte pkts: 177733, Over 511-byte pkts: 347662, Over 1023-byte pkts: 49865593
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
39992986 packets, 2824233873 bytes
Unicasts: 39992271, Multicasts: 146, Broadcasts: 569
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 634.908280 Mbits/sec, 54058 packets/sec, 6.35% of line-rate < ==
Output 24.087632 Mbits/sec, 42643 packets/sec, 0.24% of line-rate
Time since last interface status change: 02:40:01
CASTOR-VCS20-2# show interface TenGigabitEthernet 1/0/4
TenGigabitEthernet 1/0/4 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a22.1948
Current address is 50eb.1a22.1948
Pluggable media present
Interface index (ifindex) is 4697751555
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
1318 packets, 253134 bytes
Unicasts: 1280, Multicasts: 32, Broadcasts: 6
64-byte pkts: 72, Over 64-byte pkts: 944, Over 127-byte pkts: 32
Over 255-byte pkts: 0, Over 511-byte pkts: 270, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
1585 packets, 262382 bytes
Unicasts: 1303, Multicasts: 99, Broadcasts: 183
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 0.000000 Mbits/sec, 0 packets/sec, 0.00% of line-rate
Output 0.000000 Mbits/sec, 0 packets/sec, 0.00% of line-rate
Time since last interface status change: 02:42:30
CASTOR-VCS20-2# show interface TenGigabitEthernet 2/0/4
TenGigabitEthernet 2/0/4 is up, line protocol is up (connected)
Hardware is Ethernet, address is 50eb.1a20.d613
Current address is 50eb.1a20.d613
Pluggable media present
Interface index (ifindex) is 8992718851
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
52133497 packets, 76336928028 bytes
Unicasts: 52133464, Multicasts: 32, Broadcasts: 6
64-byte pkts: 65, Over 64-byte pkts: 1091062, Over 127-byte pkts: 59575
Over 255-byte pkts: 269708, Over 511-byte pkts: 880391, Over 1023-byte pkts: 49832703
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
42581507 packets, 3005492438 bytes
Unicasts: 42580789, Multicasts: 147, Broadcasts: 571
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 662.249480 Mbits/sec, 56417 packets/sec, 6.62% of line-rate < ==
Output 25.869592 Mbits/sec, 45816 packets/sec, 0.26% of line-rate
Time since last interface status change: 02:42:29
CASTOR-VCS20-2# show interface TenGigabitEthernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:45
Queueing strategy: fifo
Receive Statistics:
3486 packets, 383388 bytes
Unicasts: 3486, Multicasts: 0, Broadcasts: 0
64-byte pkts: 634, Over 64-byte pkts: 2040, Over 127-byte pkts: 656
Over 255-byte pkts: 156, Over 511-byte pkts: 0, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
2722 packets, 352208 bytes
Unicasts: 2434, Multicasts: 99, Broadcasts: 189
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 0.000752 Mbits/sec, 1 packets/sec, 0.00% of line-rate
Output 0.001236 Mbits/sec, 1 packets/sec, 0.00% of line-rate
Time since last interface status change: 2d19h11m
CASTOR-VCS20-2# show interface TenGigabitEthernet 12/0/11
TenGigabitEthernet 12/0/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0005.3365.2971
Current address is 0005.3365.2971
Pluggable media present
Interface index (ifindex) is 51942621194
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:15:44
Queueing strategy: fifo
Receive Statistics:
217619557 packets, 14552796656 bytes
Unicasts: 217619557, Multicasts: 0, Broadcasts: 0
64-byte pkts: 133135797, Over 64-byte pkts: 84483273, Over 127-byte pkts: 376
Over 255-byte pkts: 94, Over 511-byte pkts: 17, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0
Transmit Statistics:
276006890 packets, 413316026559 bytes
Unicasts: 276006167, Multicasts: 146, Broadcasts: 577
Underruns: 0
Errors: 0, Discards: 26998
Rate info:
Input 124.261656 Mbits/sec, 231999 packets/sec, 1.24% of line-rate
Output 3513.364136 Mbits/sec, 293849 packets/sec, 35.13% of line-rate < ==
Time since last interface status change: 2d19h11m
< =========== >
EMC VNX 5300 LUN utilization is at maximum capacity.

EMC VNX 5300 LUN Utilization Graph
All Medusa I/O traffic completed without errors.

iSCSI Bandwidth Medusa Results
1.2.3 Storage Device – w/Congested Fabric
Test Objective
- Create network bottleneck through a single Ethernet Fabric ISL.
- Configure VDX switch for AUTONAS.
- Configure multiple ‘iSCSI to host’ data streams sufficient to saturate ISLs’ available bandwidth for 30 minutes.
- Verify file I/O completes error-free. Add L3 FCoE and L2 Ethernet traffic to fill available bandwidth.
- File I/O compares are disabled.
Test Configuration
1. Configure AUTONAS on VDX as follows:
< =========== >
nas auto-qos
nas server-ip 9.78.50.1/32 vlan 978
< =========== >
2. Add IXIA connections:

VCS Fabric Ports for IXIA Traffic Generator
4. Reconfigure VDX as follows:
< =========== >
interface TenGigabitEthernet 11/1/31
cee default
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
no shutdown
!
interface TenGigabitEthernet 12/0/31
cee default
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
no shutdown
!
interface TenGigabitEthernet 11/1/32
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
fcoeport default
no shutdown
!
interface TenGigabitEthernet 12/0/32
no fabric isl enable
no fabric trunk enable
switchport
switchport mode access
switchport access vlan 979
fcoeport default
no shutdown
!
interface TenGigabitEthernet 11/1/43
shutdown
!
interface TenGigabitEthernet 11/1/44
shutdown
!
interface TenGigabitEthernet 11/1/45
shutdown
!
interface TenGigabitEthernet 11/1/46
shutdown
!
interface TenGigabitEthernet 11/1/47
shutdown
!
interface TenGigabitEthernet 12/0/11
shutdown
!
< =========== >
Results
PASS. iSCSI I/O completes error-free through bandwidth congestion.

Storage Device Congested Fabric Medusa Results
< =========== >
M4-VCS20-11# show fabric isl
Rbridge-id: 11 #ISLs: 1
Src Src Nbr Nbr
Index Interface Index Interface Nbr-WWN BW Trunk Nbr-Name
———————————————————————————————-
59 Te 11/1/48 111 Te 12/0/48 10:00:00:05:33:65:29:60 10G Yes “CASTOR-VCS20-12” < ==
M4-VCS20-11# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:30:12
Queueing strategy: fifo
Receive Statistics:
1455598665 packets, 2224640103434 bytes
Unicasts: 1455592180, Multicasts: 6684, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 830168, Over 127-byte pkts: 410544
Over 255-byte pkts: 102470, Over 511-byte pkts: 393668, Over 1023-byte pkts: 753040
Over 1518-byte pkts(Jumbo): 1453108799
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 2446
Transmit Statistics:
1565112044 packets, 2068204116199 bytes
Unicasts: 1565111491, Multicasts: 640, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 2446
Rate info:
Input 9818.416716 Mbits/sec, 803102 packets/sec, 98.18% of line-rate < ==
Output 9129.502965 Mbits/sec, 865978 packets/sec, 91.30% of line-rate
Time since last interface status change: 3d22h34m
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/11
Interface TenGigabitEthernet 11/1/11
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 56 3760 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 0 0 3 0 0
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 223834227 15047992872 untag 288943073 436246113624 < ==
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/31
Interface TenGigabitEthernet 11/1/31
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 0 0 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 0 0 3 0 0
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 670912060 1006368102000 untag 582848400 874263395888 < ==
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/32
Interface TenGigabitEthernet 11/1/32
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 0 0 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 670951723 1006427423304 3 584342735 876513661544 < ==
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 60 5936 untag 6571 636784
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/48
Interface TenGigabitEthernet 11/1/48
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 582916980 890688245176 0 670991987 1025275257432 < ==
1 0 0 1 0 0
2 288994572 444416470856 2 223868409 21318601072 < ==
3 584376567 890590027648 3 670991489 1022591222784 < ==
4 0 0 4 0 0
5 0 0 5 0 0
6 12 888 6 3 216
7 2506 2959208 7 2506 2953664
untag 0 0 untag 0 0
M4-VCS20-11#
< =========== >
1.2.4 Storage Device – iSCSI Protocol Jammer CRC Corruption on ISL Port
Test Objective
- Perform iSCSI jammer testing using CRC corruption with a burst of 10 CRC errors on ISL link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
- Shutdown all ISL interfaces on M4-VCS20-11 with the exception of one ISL to CASTOR-VCS20-12.
- Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
- Connect Ethernet jammer hardware in-line with remaining ISL between M4-VCS20-11 and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 11/1/43
shutdown
interface TenGigabitEthernet 11/1/44
shutdown
interface TenGigabitEthernet 11/1/45
shutdown
interface TenGigabitEthernet 11/1/46
shutdown
interface TenGigabitEthernet 11/1/47
shutdown
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer CRC corruption on ISL port.
NTFS, CRC ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 12/0/48
TenGigabitEthernet 12/0/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0005.3365.2996
Current address is 0005.3365.2996
Pluggable media present
Interface index (ifindex) is 51943833653
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:31
Queueing strategy: fifo
Receive Statistics:
1163231 packets, 112100255 bytes
Unicasts: 1163207, Multicasts: 14, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 1098719, Over 127-byte pkts: 62831
Over 255-byte pkts: 7, Over 511-byte pkts: 0, Over 1023-byte pkts: 37
Over 1518-byte pkts(Jumbo): 1637
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 23
Transmit Statistics:
1443845 packets, 2183737378 bytes
Unicasts: 1443726, Multicasts: 119, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 19
Rate info:
Input 34.898688 Mbits/sec, 46270 packets/sec, 0.35% of line-rate
Output 695.900480 Mbits/sec, 57511 packets/sec, 6.96% of line-rate
Time since last interface status change: 01:03:26
< =========== >

iSCSI Protocol Jammer CRC Corruption on ISL Port Medusa Result 1
EXT3, CRC ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 12/0/48
TenGigabitEthernet 12/0/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0005.3365.2996
Current address is 0005.3365.2996
Pluggable media present
Interface index (ifindex) is 51943833653
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:30
Queueing strategy: fifo
Receive Statistics:
1168999 packets, 117405043 bytes
Unicasts: 1168986, Multicasts: 3, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 1110280, Over 127-byte pkts: 58704
Over 255-byte pkts: 6, Over 511-byte pkts: 0, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 9
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 22
Transmit Statistics:
1394664 packets, 2030034819 bytes
Unicasts: 1394566, Multicasts: 98, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 19
Rate info:
Input 39.370531 Mbits/sec, 49003 packets/sec, 0.39% of line-rate
Output 679.710602 Mbits/sec, 58279 packets/sec, 6.80% of line-rate
Time since last interface status change: 01:04:58
< =========== >

iSCSI Protocol Jammer CRC Corruption on ISL Port Medusa Result 2
NTFS, CRC ERROR BURST ON WRITE TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:32
Queueing strategy: fifo
Receive Statistics:
1513037 packets, 2252182834 bytes
Unicasts: 1512908, Multicasts: 119, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 19902, Over 127-byte pkts: 38448
Over 255-byte pkts: 0, Over 511-byte pkts: 8, Over 1023-byte pkts: 32326
Over 1518-byte pkts(Jumbo): 1422352
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 25
Transmit Statistics:
1473300 packets, 530896300 bytes
Unicasts: 1473280, Multicasts: 20, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 25
Rate info:
Input 3.560024 Mbits/sec, 4366 packets/sec, 0.04% of line-rate
Output 569.206952 Mbits/sec, 47107 packets/sec, 5.69% of line-rate
Time since last interface status change: 00:57:02
< =========== >

iSCSI Protocol Jammer CRC Corruption on ISL Port Medusa Result 3
EXT3, CRC ERROR BURST ON WRITE TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:34
Queueing strategy: fifo
Receive Statistics:
1580994 packets, 2259437931 bytes
Unicasts: 1580872, Multicasts: 111, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 27144, Over 127-byte pkts: 71810
Over 255-byte pkts: 6360, Over 511-byte pkts: 30111, Over 1023-byte pkts: 44728
Over 1518-byte pkts(Jumbo): 1400837
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 30
Transmit Statistics:
1588070 packets, 582746936 bytes
Unicasts: 1588062, Multicasts: 3, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 30
Rate info:
Input 5.575024 Mbits/sec, 6467 packets/sec, 0.06% of line-rate
Output 713.548756 Mbits/sec, 59783 packets/sec, 7.14% of line-rate
Time since last interface status change: 00:58:58
< =========== >

iSCSI Protocol Jammer CRC Corruption on ISL Port Medusa Result 4
1.2.5 Storage Device – iSCSI Protocol Jammer Packet Corruption on ISL Port
Test Objective
- Perform iSCSI jammer testing using packet corruption with a burst of 10 packet errors on ISL link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
- Shutdown all ISL interfaces on M4-VCS20-11 with the exception of one ISL to CASTOR-VCS20-12.
- Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
- Connect Ethernet jammer hardware in-line with remaining ISL between M4-VCS20-11 and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 11/1/43
shutdown
interface TenGigabitEthernet 11/1/44
shutdown
interface TenGigabitEthernet 11/1/45
shutdown
interface TenGigabitEthernet 11/1/46
shutdown
interface TenGigabitEthernet 11/1/47
shutdown
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer packet corruption on ISL port.
NTFS, PACKET ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 12/0/48
TenGigabitEthernet 12/0/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0005.3365.2996
Current address is 0005.3365.2996
Pluggable media present
Interface index (ifindex) is 51943833653
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:31
Queueing strategy: fifo
Receive Statistics:
1339042 packets, 368123753 bytes
Unicasts: 1339022, Multicasts: 10, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 1108919, Over 127-byte pkts: 63200
Over 255-byte pkts: 9, Over 511-byte pkts: 0, Over 1023-byte pkts: 3709
Over 1518-byte pkts(Jumbo): 163205
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 23
Transmit Statistics:
1468047 packets, 2198208552 bytes
Unicasts: 1467907, Multicasts: 140, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 21
Rate info:
Input 486.650088 Mbits/sec, 54185 packets/sec, 4.87% of line-rate
Output 226.341452 Mbits/sec, 22243 packets/sec, 2.26% of line-rate
Time since last interface status change: 01:11:37
< =========== >

iSCSI Protocol Jammer Packet Corruption on ISL Port Medusa Result 1
EXT3, PACKET ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 12/0/48
TenGigabitEthernet 12/0/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0005.3365.2996
Current address is 0005.3365.2996
Pluggable media present
Interface index (ifindex) is 51943833653
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:31
Queueing strategy: fifo
Receive Statistics:
1504838 packets, 430168280 bytes
Unicasts: 1504808, Multicasts: 20, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 1244511, Over 127-byte pkts: 64213
Over 255-byte pkts: 8, Over 511-byte pkts: 4263, Over 1023-byte pkts: 69
Over 1518-byte pkts(Jumbo): 191774
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 24
Transmit Statistics:
1552226 packets, 2222961797 bytes
Unicasts: 1552090, Multicasts: 136, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 20
Rate info:
Input 449.654936 Mbits/sec, 38060 packets/sec, 4.50% of line-rate
Output 9.197280 Mbits/sec, 4636 packets/sec, 0.09% of line-rate
Time since last interface status change: 01:43:06
< =========== >

iSCSI Protocol Jammer Packet Corruption on ISL Port Medusa Result 2
NTFS, PACKET ERROR BURST ON WRITE TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:33
Queueing strategy: fifo
Receive Statistics:
1511685 packets, 2258149683 bytes
Unicasts: 1511569, Multicasts: 106, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 15588, Over 127-byte pkts: 37178
Over 255-byte pkts: 0, Over 511-byte pkts: 8, Over 1023-byte pkts: 32420
Over 1518-byte pkts(Jumbo): 1426491
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 31
Transmit Statistics:
1416739 packets, 440661271 bytes
Unicasts: 1416735, Multicasts: 4, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 29
Rate info:
Input 3.331040 Mbits/sec, 4056 packets/sec, 0.03% of line-rate
Output 556.403596 Mbits/sec, 46035 packets/sec, 5.56% of line-rate
Time since last interface status change: 01:44:49
< =========== >

iSCSI Protocol Jammer Packet Corruption on ISL Port Medusa Result 3
EXT3, PACKET ERROR BURST ON WRITE TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:33
Queueing strategy: fifo
Receive Statistics:
1572456 packets, 2247430039 bytes
Unicasts: 1572323, Multicasts: 123, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 21607, Over 127-byte pkts: 70058
Over 255-byte pkts: 8050, Over 511-byte pkts: 36983, Over 1023-byte pkts: 52588
Over 1518-byte pkts(Jumbo): 1383170
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0, TrillportCtrlFrames: 28
Transmit Statistics:
1554104 packets, 487059124 bytes
Unicasts: 1554088, Multicasts: 16, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 28
Rate info:
Input 4.824016 Mbits/sec, 5610 packets/sec, 0.05% of line-rate
Output 610.986008 Mbits/sec, 51193 packets/sec, 6.11% of line-rate
Time since last interface status change: 01:46:27
< =========== >

iSCSI Protocol Jammer Packet Corruption on ISL Port Medusa Result 4
1.2.6 Storage Device – iSCSI Protocol Jammer Packet Loss on ISL Port
Test Objective
- Perform iSCSI jammer testing using packet loss with a burst of 10 packet drops on ISL link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
- Shutdown all ISL interfaces on M4-VCS20-11 with the exception of one ISL to CASTOR-VCS20-12.
- Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
- Connect Ethernet jammer hardware in-line with remaining ISL between M4-VCS20-11 and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 11/1/43
shutdown
interface TenGigabitEthernet 11/1/44
shutdown
interface TenGigabitEthernet 11/1/45
shutdown
interface TenGigabitEthernet 11/1/46
shutdown
interface TenGigabitEthernet 11/1/47
shutdown
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer packet loss on ISL port.
NTFS, PACKET DROP BURST ON READ TRAFFIC

iSCSI Protocol Jammer Packet Loss on ISL Port Medusa Result 1
EXT3, PACKET DROP BURST ON READ TRAFFIC

iSCSI Protocol Jammer Packet Loss on ISL Port Medusa Result 2
NTFS, PACKET DROP BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Loss on ISL Port Medusa Result 3
EXT3, PACKET DROP BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Loss on ISL Port Medusa Result 4
1.2.7 Storage Device – iSCSI Protocol Jammer CRC Corruption on Storage Port
Test Objective
- Perform iSCSI jammer testing using CRC corruption with a burst of 10 CRC errors on storage port link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
1. Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer packet corruption on storage port.
NTFS, CRC ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:33
Queueing strategy: fifo
Receive Statistics:
1130251 packets, 74930728 bytes
Unicasts: 1130241, Multicasts: 0, Broadcasts: 0
64-byte pkts: 1068513, Over 64-byte pkts: 61704, Over 127-byte pkts: 27
Over 255-byte pkts: 7, Over 511-byte pkts: 0, Over 1023-byte pkts: 0
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0
Transmit Statistics:
1417267 packets, 2104198745 bytes
Unicasts: 1417248, Multicasts: 13, Broadcasts: 5
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 22.342528 Mbits/sec, 42131 packets/sec, 0.22% of line-rate
Output 627.267200 Mbits/sec, 52808 packets/sec, 6.27% of line-rate
Time since last interface status change: 00:24:39
< =========== >

iSCSI Protocol Jammer CRC Corruption on Storage Port Medusa Result 1
EXT3, CRC ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:33
Queueing strategy: fifo
Receive Statistics:
1375559 packets, 225443018 bytes
Unicasts: 1375594, Multicasts: 0, Broadcasts: 1
64-byte pkts: 20, Over 64-byte pkts: 1287145, Over 127-byte pkts: 22
Over 255-byte pkts: 1895, Over 511-byte pkts: 32, Over 1023-byte pkts: 86445
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0
Transmit Statistics:
1554838 packets, 2188272227 bytes
Unicasts: 1554830, Multicasts: 1, Broadcasts: 8
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 538.782920 Mbits/sec, 61155 packets/sec, 5.39% of line-rate
Output 220.998048 Mbits/sec, 24077 packets/sec, 2.21% of line-rate
Time since last interface status change: 00:26:10
< =========== >

iSCSI Protocol Jammer CRC Corruption on Storage Port Medusa Result 2
NTFS, CRC ERROR BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer CRC Corruption on Storage Port Medusa Result 3
EXT3, CRC ERROR BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer CRC Corruption on Storage Port Medusa Result 4
1.2.8 Storage Device – iSCSI Protocol Jammer Packet Corruption on Storage Port
Test Objective
- Perform SMB/NFS jammer testing using packet corruption with a burst of 10 packet errors on storage port link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
1. Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer packet corruption on storage port.
NTFS, PACKET ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:32
Queueing strategy: fifo
Receive Statistics:
1323829 packets, 306929458 bytes
Unicasts: 1323819, Multicasts: 0, Broadcasts: 0
64-byte pkts: 1108726, Over 64-byte pkts: 63822, Over 127-byte pkts: 27
Over 255-byte pkts: 9, Over 511-byte pkts: 0, Over 1023-byte pkts: 151245
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0
Transmit Statistics:
1480454 packets, 2177465139 bytes
Unicasts: 1480449, Multicasts: 1, Broadcasts: 4
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 909.438560 Mbits/sec, 76666 packets/sec, 9.09% of line-rate
Output 4.255736 Mbits/sec, 7219 packets/sec, 0.04% of line-rate
Time since last interface status change: 00:54:08
< =========== >

iSCSI Protocol Jammer Packet Corruption on Storage Port Medusa Result 1
EXT3, PACKET ERROR BURST ON READ TRAFFIC
< =========== >
CASTOR-VCS20-2# show interface tengigabitethernet 11/1/11
TenGigabitEthernet 11/1/11 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71d8
Current address is 0027.f81c.71d8
Pluggable media present
Interface index (ifindex) is 47651848323
MTU 2500 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 00:00:33
Queueing strategy: fifo
Receive Statistics:
1330637 packets, 158313668 bytes
Unicasts: 1330634, Multicasts: 0, Broadcasts: 0
64-byte pkts: 16, Over 64-byte pkts: 1287066, Over 127-byte pkts: 31
Over 255-byte pkts: 939, Over 511-byte pkts: 15, Over 1023-byte pkts: 42570
Over 1518-byte pkts(Jumbo): 0
Runts: 0, Jabbers: 0, CRC: 10, Overruns: 0 < ==
Errors: 0, Discards: 0
Transmit Statistics:
1552786 packets, 2192793711 bytes
Unicasts: 1552775, Multicasts: 4, Broadcasts: 7
Underruns: 0
Errors: 0, Discards: 0
Rate info:
Input 278.033496 Mbits/sec, 52159 packets/sec, 2.78% of line-rate
Output 408.616744 Mbits/sec, 38335 packets/sec, 4.09% of line-rate
Time since last interface status change: 00:55:52
< =========== >

iSCSI Protocol Jammer Packet Corruption on Storage Port Medusa Result 2
NTFS, PACKET ERROR BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Corruption on Storage Port Medusa Result 3
EXT3, PACKET ERROR BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Corruption on Storage Port Medusa Result 4
1.2.9 Storage Device – iSCSI Protocol Jammer Packet Loss on Storage Port
Test Objective
- Perform iSCSI jammer testing using packet loss with a burst of 10 lost packets on ISL link, in each direction sequentially, while I/O is running from a single host.
- Verify I/O recovers successfully.
Test Configuration
1. Shutdown multi-path connection between iSCSI target and CASTOR-VCS20-12.
< =========== >
interface TenGigabitEthernet 12/0/11
shutdown
< =========== >
Results
PASS. All I/O completes error-free through Ethernet jammer packet loss on storage port.
NTFS, PACKET DROP BURST ON READ TRAFFIC

iSCSI Protocol Jammer Packet Loss on Storage Port Medusa Result 1
EXT3, PACKET DROP BURST ON READ TRAFFIC

iSCSI Protocol Jammer Packet Loss on Storage Port Medusa Result 2
NTFS, PACKET DROP BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Loss on Storage Port Medusa Result 3
EXT3, PACKET DROP BURST ON WRITE TRAFFIC

iSCSI Protocol Jammer Packet Loss on Storage Port Medusa Result 4
1.3 Stress and Error Recovery
3.1 Storage Device Fabric I/O Integrity – Congested Fabric
Test Objective
- Configure fabric and devices for maximum link and device saturation.
- Configure VDX switch for AUTONAS.
- Add L3 FCoE traffic (COS 3), SMB/NFS traffic (COS 2) and L2 Ethernet traffic (COS 0) to fill all available bandwidth.
- From all iSCSI initiators, start a mixture of READ/WRITE/VERIFY traffic with random data patterns continuously to all targets for 12 hours.
- Verify no unexpected change in I/O throughput occurs.
- File I/O compares are enabled.
Test Configuration
1. See test configuration for test 1.2.3.
Results
PASS. All I/O completes error-free through VCS fabric congestion.

Storage Device Fabric IO Integrity – Congested Fabric Medusa Result
< =========== >
M4-VCS20-11# show fabric isl
Rbridge-id: 11 #ISLs: 1
Src Src Nbr Nbr
Index Interface Index Interface Nbr-WWN BW Trunk Nbr-Name
———————————————————————————————-
59 Te 11/1/48 111 Te 12/0/48 10:00:00:05:33:65:29:60 10G Yes “CASTOR-VCS20-12” < ==
M4-VCS20-11# show interface tengigabitethernet 11/1/48
TenGigabitEthernet 11/1/48 is up, line protocol is up (connected)
Hardware is Ethernet, address is 0027.f81c.71fd
Current address is 0027.f81c.71fd
Pluggable media present
Interface index (ifindex) is 47653061254
MTU 9216 bytes
LineSpeed Actual : 10000 Mbit
LineSpeed Configured : Auto, Duplex: Full
Priority Tag disable
IPv6 RA Guard disable
Last clearing of show interface counters: 15:12:05
Queueing strategy: fifo
Receive Statistics:
44266823255 packets, 66374901967317 bytes
Unicasts: 44266621540, Multicasts: 201915, Broadcasts: 0
64-byte pkts: 0, Over 64-byte pkts: 633692499, Over 127-byte pkts: 201663416
Over 255-byte pkts: 19376256, Over 511-byte pkts: 23172044, Over 1023-byte pkts: 112120662
Over 1518-byte pkts(Jumbo): 43276798402
Runts: 0, Jabbers: 0, CRC: 0, Overruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 72544
Transmit Statistics:
47692776797 packets, 65985548976009 bytes
Unicasts: 47692757607, Multicasts: 19317, Broadcasts: 0
Underruns: 0
Errors: 0, Discards: 0, TrillportCtrlFrames: 72544
Rate info:
Input 9820.074000 Mbits/sec, 808128 packets/sec, 98.20% of line-rate < ==
Output 9163.368744 Mbits/sec, 840052 packets/sec, 91.63% of line-rate
Time since last interface status change: 4d16h17m
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/11
Interface TenGigabitEthernet 11/1/11
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 112 7528 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 0 0 3 0 0
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 8378973992 5757884193032 untag 4461248300 5506748708312 < ==
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/31
Interface TenGigabitEthernet 11/1/31
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 0 0 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 0 0 3 0 0
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 20251566661 30377349996840 untag 19954853705 29932002520672 < ==
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/32
Interface TenGigabitEthernet 11/1/32
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 0 0 0 0 0
1 0 0 1 0 0
2 0 0 2 0 0
3 20251614843 30377411765456 3 19852364986 29778528042696 < ==
4 0 0 4 0 0
5 0 0 5 0 0
6 0 0 6 0 0
7 0 0 7 0 0
untag 1823 180472 untag 198187 19257968
M4-VCS20-11# show qos queue interface tengigabitethernet 11/1/48
Interface TenGigabitEthernet 11/1/48
RX RX TX TX
CoS Packets Bytes CoS Packets Bytes
———————————— ————————————
0 19954928641 30490858358640 0 19646392089 30019666169808 < ==
1 0 0 1 0 0
2 4461270821 5631712926752 2 8378960910 5992436323824 < ==
3 19852388023 30255039519264 3 19669261975 29975955443144 < ==
4 0 0 4 0 0
5 0 0 5 0 0
6 433 32400 6 26 2064
7 72604 83625696 7 72604 83662512
untag 0 0 untag 0 0
M4-VCS20-11#
< =========== >
1.3.2 Storage Device Integrity – Device Recovery from Port Toggle – Manual Cable Pull
Test Objective
- Perform sequentially for each Storage Device and Adapter port.
- With I/O running, perform a quick port toggle every Storage Device and Adapter port.
- Verify host I/O will recover.
Results
PASS. All I/O recovers through port disconnections and reconnections.
1.3.3 Storage Device Integrity – Device Recovery from ISL Port Relocation
Test Objective
1. With I/O running, manually disconnect and reconnect ISL port to different port in same switch.
2. Verify fabric integrity and file I/O recovery.
3. Repeat test for all switch types.
Results
PASS. All I/O completes error-free through ISL port disconnections and reconnections.
1.3.4 Storage Device Stress – Device Recovery from Device Port Toggle – Extended Run
Test Objective
- Sequentially toggle each Initiator and Target port in fabric.
- Verify host I/O will recover to alternate path and toggled path will recover.
- Run for 12 hours.
Results
PASS. All I/O completes error-free through initiator and target port disable and enable commands.

Device Port Toggle Extended Run Medusa Result
1.3.5 Storage Device Recovery – ISL Port Toggle – Extended Run
Test Objective
- Verify fabric ISL path redundancy between hosts and storage devices.
- Sequentially toggle each ISL path on all switches.
- Host I/O may pause but should recover.
- Verify host I/O throughout test.
- Run for 12 hours.
Results
PASS. All I/O completes error-free through ISL port disable and enable commands.

ISL Port Toggle Extended Run Medusa Result
1.3.6 Storage Device Recovery – All-ISL Port Toggle
Test Objective
- Verify fabric switch path redundancy between hosts and storage devices. Sequentially, and for all switches, disable all ISLs on the switch under test.
- Verify I/O pauses and resume after switch fabric convergence.
Results
PASS. All I/O completes error-free through ISL port disable and enable commands.

VDX8770 All-ISL Port Toggle Medusa Result
1.3.7 Storage Device Recovery – VDX8770 Line Card Maintenance
Test Objective
- For extended time, verify device connectivity to VDX 8770 blades.
- Sequentially power off and on each Mercury line card.
- Verify host I/O will failover to alternate path and toggled path will recover.
Results
PASS. All I/O completes error-free through line card power off and on commands.

VDX8770 Line Card Maintenance Medusa Result
1.3.8 Storage Device Recovery – Switch Offline
Test Objective
1. Toggle each switch in sequential order with a mix of
– chassis disable,
– reboot, and
– power cycles.
2. For switches containing device under test, I/O will pause and then resume after switch comes online.
Results
1. Chassis Disable/Enable: PASS. I/O successfully resumes after chassis enable.
2. Reboot: PASS. I/O successfully resumes after chassis reboot.
3. Power Cycle: PASS. I/O successfully resumes after chassis power cycle.
1.3.9 Workload Simulation Test Suite – Microsoft Exchange Server
Test Objective
- Validate Storage/Fabric behavior while running a Medusa Microsoft Exchange Server workload simulation test suite for 30 minutes.
- Verify file I/O completes error-free. File I/O compares are disabled.
Results
PASS. All I/O completes error-free.

Microsoft Exchange Server Medusa Result
1.3.10 Workload Simulation Test Suite – File Server Simulation
Test Objective
- Validate Storage/Fabric behavior while running a Medusa File Server Simulation workload simulation test suite for 30 minutes.
- Verify file I/O completes error-free. File I/O compares are disabled.
Results
PASS. All I/O completes error-free.

File Server Simulation Medusa Result