Emulex has announced a second generation Ethernet Fibre Channel adapter and added encryption to its Fibre Channel adapter range using RSA technology. Emulex supplies Fibre Channel Host Bus Adapters (HBAs) with virtualisation capabilities so that physical servers can access Fibre Channel SANs, and virtual machines share their …
iSCSI is not supported by this card
The article says that iSCSI stack is offloaded, but that seems incorrect - Emulex cards cannot offload iSCSI, they even cannot offload TCP in statefull mode (full offload). Only ordinary partial IP optimizations are supported like large send offload but these things are supported even on cheap network adapters for many years already. Without TCP offload I found very little difference between this and ordinary fibre channel card.
There are two approaches to connecting storage:
1. FCoE (over modified Ethernet, special switches required)
2. iSCSI (over ordinary IP network)
1. FCoE is not yet standardized and the biggest problem with it is that it tries to add Fibre Channel features to the Ethernet, and that is trying to merge too completely different worlds. This will make switches more complex and might make them more expensive.
2. iSCSI is much more simple storage protocol running over very ordinary TCP/IP transport protocol. Connections are made as simple IP connections. It does not require special Ethernet switches nor any Ethernet modifications. Standards are in place, but there are problems also.
First problem with iSCSI is Broadcom company (long product support delays) practices. Majority of servers (Hewlett Packard Proliant, Dell servers and others) have Broadcom chips with TCP offload, and there was unacceptable delays to release TCP offload drivers for these Broadcom chips. Only a week ago drivers appeared for 1Gbps Broadcom chips that fully support all the offload features (currently IPv4 only), and for the Broadcom flagship 10Gbps iSCSI chip 57711 (currently part of HP BL495c servers) still there is no driver available that support iSCSI offload. 57711 10Gbps universal network chips (IP, RDMA, iSCSI) look like ideal solution on datasheets, but still there is no drivers for them that completely support the features.
Common, Broadcom, get the drivers done, you are killing the iSCSI market and your market share for IP storage networks!
Another problem is lack of disk arrays with 10Gbps iSCSI interfaces. Majority of disk arrays have 1Gbps connections. The good part is that many have several 1Gbps interfaces and they can be combined into one faster connection.
Third problem is the iSCSI stability under heavy loads. For example, iSCSI does not work stable in default Windows 2008 configuration under heavy loads. I was able to achieve acceptable stability only after two non-public Windows 2008 patches were applied and disk firmwares upgraded to the latest versions.
For iSCSI to become as a technology of choice the following must be done:
1. Broadcom must release working iSCSI offload drivers for their 10G iSCSI chip.
2. Windows 2008 SP2 must contain all necessary patches for iSCSI stability and all timeout parameters must be OK in their default configurations. iSCSI must work out of the box.
3. Disk arrays with native 10Gbps iSCSI interfaces must be released, and tested to run stable under maximum possible heavy loads.
If these little remaining things will be fixed/implemented soon (I hope this will happen if proper Broadcom and HP executives are reading this), I see no need for FCoE.
Personally I like more iSCSI because of its simplicity and that it does not require to modify Ethernet, as it is IP based it can run over very long distances etc. And any server with network adapter can access and mount the iSCSI disk. Performances may differ depending on IP network speeds/latencies, but iSCSI (TCP) connectivity is universal.
This is a new card from Emulex, not the old one.
This is a new card, and apparently uses new Ethernet silicon (not the Intel Oplin Ethernet ASIC on the current CNA). The new Emulex NIC is not yet shipping, so it is very possible it does support full offload (TOE), and the associated benefits for iSCSI and iWARP/RDMA protocols.
That said, partial offload may be adequate for iSCSI, depending on the workload.
The benefit of FCoE is it works with existing FC storage, without the need for special gateways. While it requires enhanced Ethernet (CEE/DCE), it only requires that enhanced link until the storage traffic is separated and put onto a standard FC link, which typically will only be from the server NIC to the access layer DCE switch.
From the Horse’s Mouth: What OneConnect Supports
I’m a director of product marketing at Emulex and as Magellen posted on 2/18, Emulex OneConnect referenced in the article is a new platform, not to be confused with the current shipping Emulex LP21000 FCoE CNA. The new Emulex OneConnect platform will enable a new class of products called universal CNAs that provide universal connectivity for the data center and replace a number of point products such as NICs, iSCSI HBAs and FCoE CNAs used today. With full offload and hardware acceleration for a number of protocols including FCoE and iSCSI, the technology delivers the efficiency necessary to extract maximum value from deployed servers, irrespective of the customer’s choice of protocols. Emulex has a reputation of bringing storage products with enterprise class scalability and reliability to the market, and will be delivering this value again as we introduce products based on the OneConnect platform later this year. Stay tuned for more details on the OneConnect family.
- Facebook offshores HUGE WAD OF CASH to Caymans - via Ireland
- Review Best budget Android smartphone there is? Must be the Moto G
- NSFW Confessions of a porn site boss: How the net porn industry flopped
- World's OLDEST human DNA found in leg bone – but that's not the only boning going on...
- OHM MY GOD! Move over graphene, here comes '100% PERFECT' stanene