Lenovo YUM Repository
Release and Change History
Bundle Release Information: |
Targeted Server Family (Machine Type): Flex System x280/x480/x880 X6 (7196, 4258) |
Targeted Operating System: RHEL6.9 |
YUM Repository Build Date: 2019_04_23 |
Device: |
IBM Flex System CN4054R 10Gb Virtual Fabric Adapter |
![]() Device Product Guide |
Part Number: 00Y3306 Feature Code: A4R6 |
|||
---|---|---|---|---|---|
PCIe Sub Vendor ID: 0x10DF PCIe Sub Device ID: 0xE731 |
Release History: |
Change History Documents |
Firmware | Firmware Name | Version | |
---|---|---|---|
Emulex OCe11xxx UCNA Firmware Update for Linux - 12.0.1169.20 - Release 18B | 18b-oc11-12.0.1169.20-2 | ||
Drivers | Driver Name | Discreate Package Download | Version |
brcmfcoe-vector-map enablement code |
brcmfcoe-vector-map- 1-1.rhel6u9.x86_64.rpm | 1 | |
brcm-brcmfcoe kernel driver |
kmod-brcm-brcmfcoe- 12.0.1169.32-1.rhel6u9.x86_64.rpm | 12.0.1169.32 | |
be2iscsi kernel driver |
kmod-be2iscsi-12.0.1169.32- 1.rhel6u9.x86_64.rpm | 12.0.1169.32 | |
be2net kernel driver |
kmod-be2net-12.0.1169.32- 1.rhel6u9.x86_64.rpm | 12.0.1169.32 |
Firmware | Firmware Name | Version | |
---|---|---|---|
Emulex OCe11xxx UCNA Firmware Update for Linux - 12.0.1169.7 - Release 18B | 18b-oc11-12.0.1169.7-1 | ||
Drivers | Driver Name | Discreate Package Download | Version |
brcmfcoe-vector-map enablement code |
brcmfcoe-vector-map- 1-1.rhel6u9.x86_64.rpm | 1 | |
brcm-brcmfcoe kernel driver |
kmod-brcm-brcmfcoe- 12.0.1154.3-1.rhel6u9.x86_64.rpm | 12.0.1154.3 | |
be2iscsi kernel driver |
kmod-be2iscsi-12.0.1154.2- 1.rhel6u9.x86_64.rpm | 12.0.1154.2 | |
be2net kernel driver |
kmod-be2net-12.0.1154.1- 1.rhel6u9.x86_64.rpm | 12.0.1154.1 |
Firmware | Firmware Name | Version | |
---|---|---|---|
Emulex OCe11xxx UCNA Firmware Update for Linux - 12.0.1138.0 - Release 18B | 18b-oc11-12.0.1138.0-2 | ||
Drivers | Driver Name | Discreate Package Download | Version |
brcmfcoe-vector-map enablement code |
brcmfcoe-vector-map- 1-1.rhel6u9.x86_64.rpm | 1 | |
brcm-brcmfcoe kernel driver |
kmod-brcm-brcmfcoe- 12.0.1141.6-1.rhel6u9.x86_64.rpm | 12.0.1141.6 | |
be2iscsi kernel driver |
kmod-be2iscsi-12.0.1141.10- 1.rhel6u9.x86_64.rpm | 12.0.1141.10 | |
be2net kernel driver |
kmod-be2net-12.0.1141.7- 1.rhel6u9.x86_64.rpm | 12.0.1141.7 |
Release 19A Change History Documents |
=============================================================================== Emulex OCe11xxx UCNA Firmware Package =============================================================================== Firmware Version: 12.0.1169.20 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Purley System Support - RHEL7 Update 3 Support - SLES12 SP2 Support - Windows 2016 Support - VMware ESXi 6.5 Support Known Issues: - None =============================================================================== Firmware Version: 11.4.1186.4 Supported On: Lenovo System x Rack and Flex Problems Fixed: - [Lenovo 95350] Discovery Notify request not being sent after Discovered flag is cleared (202745) - Reported by VMW: [NPAR]No Traffic between partition nics sharing a same PF (205278) - NC-SI requiremennt V1.01 support (203733) Incremental Interoperability: - SLES 12 SP3 Support - RHEL 7.4 Support Known Issues: - None =============================================================================== Firmware Version: 11.2.1193.39 Supported On: Lenovo System x Rack and Flex Problems Fixed: * Virtual Addresses (MAC/WWWN) must be set to factory defaults when load defaults are applied * DMTF (upcoming DMTF standard for PLDM firmware download) * Xclarity can't select Gold (00AG570) add on Card to update FW on Apply/Activate page * PLDM_FW_DOWNLOAD: Lenovo IMM code parsing incorrect header format * PLDM_FW_DOWNLOAD: Lenovo upd header is not updated for all device ids supported for this feature * ReqID 7981: Lenovo IO CORE Monitoring Requirement involving 2 NC SI command and 2 pairs of statistics Incremental Interoperability: - Windows 2016 Support - VMware ESXi 6.5 Support Known Issues: - Beginning with software release 11.2, Emulex LightPulse adapters and OneConnect adapters have independent software kits. The "Broadcom Software Kit Migration Guide" document provides special instructions and considerations for using the driver kits for LightPulse and OneConnect adapters. https://docs.broadcom.com/docs/12378907 =============================================================================== Firmware Version: 10.6.228.36 Supported On: Lenovo System x Rack and Flex Problems Fixed: - FCoE initiator & target UCM settings are not shown in IMM web GUI (183090) - PXE VLAN config must be supported when multi-channel is disabled (182184) Incremental Interoperability: - This software/firmware is only supported with Lenovo Branded System BIOS Known Issues: - When running FCoE/NIC converged traffic, FCoE traffic will fail if the NIC traffic was running at line rate (around 9.5 GB/s). (161329) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) =============================================================================== Firmware Version: 10.6.228.13 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - This software/firmware is only supported with Lenovo Branded System BIOS Known Issues: - When running FCoE/NIC converged traffic, FCoE traffic will fail if the NIC traffic was running at line rate (around 9.5 GB/s). (161329) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) =============================================================================== Firmware Version: 10.4.303.12 Supported On: System x Rack Problems Fixed: - Improved handling of ASU/UCM deployed configurations - Lenovo xHMC support for inventory management - In FCoE configurations with BE3 in UFP mode on System x3690 X5 servers, a UEFI rescan is required to boot to the OS. This renders UFP not supportable in UFP/FCoE configurations(158669) Incremental Interoperability: - Lenovo branded versions of formerly IBM branded products Known Issues: - When running FCoE/NIC converged traffic, FCoE traffic will fail if the NIC traffic was running at line rate (around 9.5 GB/s). (161329) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. =============================================================================== Firmware Version: 10.3.138.0 Supported On: Flex Problems Fixed: - PSOD Heartbeat NMI in VMWare - Bandwidth and SVID are showing as N/A in PXE select utility when 2+2 adapter is in UFP mode Incremental Interoperability: - Lenovo branded versions of formerly IBM branded products Known Issues: - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. =============================================================================== Firmware Version: 10.3.83.7 Supported On: Flex Problems Fixed: - iSCSI SAN Boot failed after OS installation completed in UFP and vNIC1 mode with OCe12000 adapter Incremental Interoperability: - Grantley based ITE Known Issues: - Server hangs at "Scan for Fibre Devices", when port disable/enable performed on the Brocade Converged 10 Gb Ethernet Switch Module for IBM BladeCenter. Do not force the link down after POST has completed and then attempt a discovery of the SAN. If you do a reboot is needed.(159125) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - The server may hang at "Connecting Boot Devices and Adapters" if both FCoE ports have boot targets configured when using a 10Gb Ethernet Pass-Thru Module for IBM BladeCenter. The server will not hang if only a single port is configured with a boot target.(157137) - In FCoE configurations with BE3 in UFP mode on System x3690 X5 servers, a UEFI rescan is required to boot to the OS. This renders UFP not supportable in UFP/FCoE configurations(158669) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware Version: 10.2.377.18 & 10.2.377.23 Supported On: System x and BladeCenter Problems Fixed: - PXESelect showing incorrect MAC address, bandwidth & SVID values. This bug is a cosmetic issue and only impacts PXESelect. The values which are not displayed correctly in PXESelect can be viewed in UEFI, OCM, and on the switch.(151140) - "Configurations have changed" message seen in UEFI setup screen when no changes were made. This pop-up can be ignored and will be fixed in a future code release.(159086) - FCoE VLAN ID is being displayed incorrectly as hexadecimal in UEFI menu. This is a cosmetic issue only. The correct VLAN is properly assigned, so there is no functional impact. The actual VLAN ID can be observed elsewhere. (159561) - UE seen while booting the system after a CoreDump - When NIC traffic has run on all the 8 vNICs of an ASIC, the vnics of first port are not honoring the assigned BW - SRIOV performance & Fairness for 16VMs is low on Linux and VMware - ML2 VFA5 device name is show "N/A" under IMM web and CLI of server management>Adapters - BW changes not getting updated with vNIC1 mode - Enable 100M/10M speed in set speed for SGMII interface - UCM/vNIC2, Ports/PCI functions supporting FCoE need to be differentiated to make it easier for the user - UCM/vNIC2 ability to program VirtWWN/WWPN independent of BFS Enable - iSCSI boot not working on higher order PCI functions (PF8 11)on 4 port controllers on ARI enabled systems - Can't install Windows2012 onto the iSCSI LUN whose LUN ID is not 0 - Fixed several issues related to IBM Agentless inventory management Incremental Interoperability: - Grantley Rack Servers Known Issues: - Server hangs at "Scan for Fibre Devices", when port disable/enable performed on the Brocade Converged 10 Gb Ethernet Switch Module for IBM BladeCenter. Do not force the link down after POST has completed and then attempt a discovery of the SAN. If you do a reboot is needed.(159125) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - The server may hang at "Connecting Boot Devices and Adapters" if both FCoE ports have boot targets configured when using a 10Gb Ethernet Pass-Thru Module for IBM BladeCenter. The server will not hang if only a single port is configured with a boot target.(157137) - In FCoE configurations with BE3 in UFP mode on System x3690 X5 servers, a UEFI rescan is required to boot to the OS. This renders UFP not supportable in UFP/FCoE configurations(158669) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware Version: 10.2.261.41 Supported On: IBM System x Problems Fixed: - Hit BSOD during VMQ overnight traffic test with cpu affinity change - IBM VFA5 Adapter causing software panic on IBM G8246 Switch in UFP mode - LPVIDs are not getting saved in vNIC2 mode if we change mode from vNIC1 mode to vNIC2 - In MC VINC1, VNIC2, UFP modes, user should see only 4 functions per port for 2 port card Incremental Interoperability: - Emulex VFA5 2x10 GbE SFP+ Adapter and FCoE/iSCSI SW for IBM System x - Emulex VFA5 2x10 GbE SFP+ PCIe Adapter for IBM System x - Emulex VFA5 2x10 GbE SFP+ Integrated Adapter for IBM System x Known Issues: - Server hangs at "Scan for Fibre Devices", when port disable/enable performed on the Brocade Converged 10 Gb Ethernet Switch Module for IBM BladeCenter. Do not force the link down after POST has completed and then attempt a discovery of the SAN. If you do a reboot is needed.(159125) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - PXESelect showing incorrect MAC address, bandwidth & SVID values. This bug is a cosmetic issue and only impacts PXESelect. The values which are not displayed correctly in PXESelect can be viewed in UEFI, OCM, and on the switch.(151140) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - "Configurations have changed" message seen in UEFI setup screen when no changes were made. This pop-up can be ignored and will be fixed in a future code release.(159086) - The server may hang at "Connecting Boot Devices and Adapters" if both FCoE ports have boot targets configured when using a 10Gb Ethernet Pass-Thru Module for IBM BladeCenter. The server will not hang if only a single port is configured with a boot target.(157137) - FCoE VLAN ID is being displayed incorrectly as hexadecimal in UEFI menu. This is a cosmetic issue only. The correct VLAN is properly assigned, so there is no functional impact. The actual VLAN ID can be observed elsewhere. (159561) - In FCoE configurations with BE3 in UFP mode on System x3690 X5 servers, a UEFI rescan is required to boot to the OS. This renders UFP not supportable in UFP/FCoE configurations(158669) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware Version: 10.2.261.36 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - FRU VPD fields are now properly populated on CN4054 and CN4054R - Fixed UE resulting slow boot at UEFI splash screen and NIC devices removed - When using VMQ with Windows Server 2012 or 2012R2, the user may experience VM connectivity loss, packet drops, system hangs, inability to shutdown VMs and possible system crashes on shutdown. The complete solution also requires 10.2 based Windows NIC driver. Incremental Interoperability: - Inventory support for the Flex xHMC (Chassis Hardware Management Console) Known Issues: - Server hangs at "Scan for Fibre Devices", when port disable/enable performed on the Brocade Converged 10 Gb Ethernet Switch Module for IBM BladeCenter. Do not force the link down after POST has completed and then attempt a discovery of the SAN. If you do a reboot is needed.(159125) - To disable BFS after BOFM or IFM have been enabled, it must be done by setting BFS manually in the EFIBoot config utility or sending a BOFM default command to the port. This is working as expected and we are documenting it here.(159505) - Intermittently, 255 vPorts can not be created successfully in Windows with the error "ERROR: <127>: Not enough HBA or fabric resources. VPort not created!". There is no known workaround other than to reduce the number of configured vPorts.(155807) - Target login will fail if you dynamically reconfigure the SAN after POST has executed. If you do, then reboot the system.(159121) - PXESelect showing incorrect MAC address, bandwidth & SVID values. This bug is a cosmetic issue and only impacts PXESelect. The values which are not displayed correctly in PXESelect can be viewed in UEFI, OCM, and on the switch.(151140) - IFM Boot target is consumed on a particular port even though the target is not accessible to that port. IFM/BOFM boot target configuration is not supported. This limitation has existed for several code releases and we are now documenting it here.(159226) - "Configurations have changed" message seen in UEFI setup screen when no changes were made. This pop-up can be ignored and will be fixed in a future code release.(159086) - The server may hang at "Connecting Boot Devices and Adapters" if both FCoE ports have boot targets configured when using a 10Gb Ethernet Pass-Thru Module for IBM BladeCenter. The server will not hang if only a single port is configured with a boot target.(157137) - FCoE VLAN ID is being displayed incorrectly as hexadecimal in UEFI menu. This is a cosmetic issue only. The correct VLAN is properly assigned, so there is no functional impact. The actual VLAN ID can be observed elsewhere. (159561) - In FCoE configurations with BE3 in UFP mode on System x3690 X5 servers, a UEFI rescan is required to boot to the OS. This renders UFP not supportable in UFP/FCoE configurations(158669) - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware Version: 10.0.803.13 Supported On: IBM System x Problems Fixed: - IBM Feature on Demand for storage keys now works with Optimized Boot enabled - Link detection time improved from ~70 secs on DAC cable insertion/removal Incremental Interoperability: - Unified Fabric Protocol on IBM System x Servers Known Issues: - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware version: 4.6.281.26 (supported on IBM System x, BladeCenter and Flex) Problems Fixed: - CN4054 VPD is not valid during/after IMM flash - Links go down after 497 to 498 days in VMWare - Send 'link down' to switch if UE occurs on ASIC Incremental Interoperability: - Emulex Dual Port 10GbE SFP+ VFA IIIr for IBM System x - Emulex Dual Port 10GbE SFP+ Embedded VFA IIIr for IBM System x Known Issues: - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware version: 4.6.281.21 (supported on IBM System x, BladeCenter and Flex) Problems Fixed: - OCe10 VFA CNAs stuck in "Switch Independent Mode" - Port Identification fails on Port 1 under uEFI menu for PCIe rack VFA Incremental Interoperability: - IBM Flex System CN4054R 10Gb Virtual Fabric Adapter - IBM Ivy Bridge CPU Based Servers Known Issues: - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware version: 4.6.281.8 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - HW iSCSI offload is now supported in UFP Mode (UFP limited to IBM Flex) - Server may hang during boot with multichannel disabled with switch default settings. The IBM 10Gb Virtual Fabric Switch has trunking enabled and the Brocade Converged 10Gbe Switch Module has ports in "Access" mode. If the Brocade Converged 10Gbe Switch Module ports are switched to trunking mode, the hang no longer occurs. - When using Config Patterns with virtual MACs the bandwidth settings are are intermittently reset after some number of server reboots. Incremental Interoperability: - IBM Flex System Fabric SI4093 System Interconnect Module - IBM Flex System EN6131 40Gb Ethernet Switch - IBM Flex System x222 10Gb Virtual Fabric LOM - SLES 11.3 on IBM System x and BladeCenter - Complete NIC/iSCSI/FCoE support for ASU/Config Patterns for Flex and ASU for IBM System x & BladeCenter Known Issues: - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware version: 4.6.166.7 Supported On: IBM System x and BladeCenter Problems Fixed: - When connected to Cisco Nexus 4k and 5k switches, the virtual link may not recover on a switch port bounce (shut/no shut) in an ESXi environment. A switch port re-bounce will recover the link. - When connected to Cisco Nexus 4k and 5k switches, OneCommand Manager reports CRC's and there is poor network performance in an ESXi environment Incremental Interoperability: - ASU/Config Pattern configuration for iSCSI and FCoE (BE3) - UFP for iSCSI (BE3) - Support for using 10Gb ports with legacy 1Gb switches installed in a BladeCenter-S chassis or MSIM/MSIM-T expansion units in a BladeCenter-H or -HT chassis - Update Express System Packs now support Linux errata kernels - ESXi 5.1 Known Issues: - After making changes to any options in the sub-menus under 'controller configuration' in the Emulex UEFI NIC configuration utility and escaping to the prior menu, there is no prompt to save the changes. Be sure to select 'Save' before exiting the configuration menu. Applies to some legacy rack and blade servers. - Intermittently, when changing the multichannel mode, the system will not auto-reboot after exiting F1 UEFI setup and will continue through POST. Manually reboot the server to ensure all changes are applied. Applies to some legacy rack and blade servers. - Server may hang during boot with multichannel disabled with switch default settings. The IBM 10Gb Virtual Fabric Switch has trunking enabled and the Brocade Converged 10Gbe Switch Module has ports in "Access" mode. If the Brocade Converged 10Gbe Switch Module ports are switched to trunking mode, the hang no longer occurs. - When using a Cisco Nexus 4001i Switch Module, intermittently the 10Gb link may come up as 1Gb link speed. An internal switch port bounce will typically establish a 10Gb link. Setting the switch port speed to fixed 10000 will also allow a guaranteed 10Gb connection. More info available in the Cisco Nexus 4001i Switch Module firmware release notes. =============================================================================== Firmware Version: 4.4.180.3 Supported On: IBM Flex Problems Fixed: - 10Gb ports intermittenly linking at 1Gb speed with the IBM Flex System Fabric EN4093/EN4093R 10Gb Scalable Switch. - Switch Configuration after an erase configuration is not reflected after reboot - 'Logical Admin Link' not enabled after applying BOFM in Switch Agnostic mode - IBM pre-boot Diagnostics NIC physical loopback test fails in Command and GUI mode - Multiple cosmetic issues with iSCSI UEFI BIOS - 10Gb ports do not detect 'Link Down' with some switches - When adding an iSCSI hardware target in UEFI through discovery login, the boot target setting is ignored. Interoperability since previous release: - UEFI 2.3.1 Compliance - Increased VF count in SR-IOV (was 16 VFs per port, is now 32 VFs per port) - Support for next generation 10GbaseT PHY - Expose IBM Feature On Demand licensing information in UEFI configuration utility - IBM Universal Configuration Manager for Ethernet - IBM Unified Fabric Protocol - Ability to assign Switch Agnostic VLAN IDs without intermediate reboot - Changes to Personality or Multichannel in PXESelect allows further changes to be made w/o reboot - Support for IBM Flex System Fabric CN4093 10Gb Converged Scalable Switch Known Issues: - HW iSCSI offload is not currently supported in UFP mode =============================================================================== Firmware Version: 4.1.422.8 Supported On: IBM Flex Problems Fixed: - PHY Loopback fails on 10G ports - RH 6.2 and W2k8 servers crashed after shut/no shut the switch port with FCoE connection and 50 vlans configured - DSA: NIC physical loopback test fails in Command and GUI mode - Firmware update fails to update all VPD Code Level fields - DCBX does not converge with FW 4.1.422 on Cicso 5k. Interoperability Since Previous Release: - Support for IBM Flex System x440 10Gb Virtual Fabric LOM =============================================================================== Firmware Version: 4.1.422.0 Supported On: IBM Flex Problems Fixed: - SR IOV: Unable to ping IPV6 from DUT VF to Peer VF - Reading VPD from BE3 timesout making SR IOV fail in newer kernels - 1Gb devices are being suppressed by kernel during the PCI probe in SLES11 SP2 - MAC Loopback fails when Mgmt Adapter Creation fails - Failure of MAC loopback test after firmware update and warm reboot - Suppress efat dump mailbox commands in the efat dump log - Running ethtool with BE3 takes longer than expected after cable pull - Out of Order CQEs - TRACE_LOG function not printing the messages for some of the VF's - esx4/5 issue. svc target cannot be configured properly - RSCN is not generated by switch when a driver restarts - Port Link stays up even after switch removed Interoperability Since Previous Release: - Support for IBM Flex System EN4054 4-port 10Gb Ethernet Adapter - Support for IBM Flex System CN4054 10Gb Virtual Fabric Adapter - Support for IBM Flex System x240 10Gb Virtual Fabric LOM - Support for IBM Flex System Manager 10Gb Virtual Fabric LOM =============================================================================== Firmware Version: 4.2.433.3 Supported On: IBM System x and BladeCenter Problems Fixed: - Intermittent link status with IBM Virtual Fabric Switch when set to IBM Virtual Fabric Mode - The option to enable/disable SRIOV on HS23 LOM Port 0 is missing from OneCommand Manager. This selection is available in UEFI F1 setup. - Under high network stress, NMI errors can be seen with x3690 X5 and x3850 X5 - 10GbE ports running at 1GbE speed 10% below wirespeed at certain block sizes. - Unable to perform PXE boot with LPVID assigned to vNIC interface - 1Gb ports of HS23 do not handle tagging properly if multichannel mode enabled (pnic ok). Note: ESX driver patch also required. - Port flapping link after FW upgrade and "reboot" when connected to 10Gb PTM module (DAC) Interoperability Since Previous Release: - Support for Emulex Virtual Fabric Adapter Advanced (BE2) for IBM BladeCenter - Support for Emulex Virtual Fabric Adapter (BE2) for IBM BladeCenter - Support for Emulex Virtual Fabric Adapter (BE2) for IBM System X Known Issues: - All 10Gb KX4 ports must be used with 10Gb switches. They will not link at 1Gb. This applies to HS23 LOM and IBM BladeCenter VFA adapters only. - IPv6 IP address not visible via IMM GUI for IBM Option 90Y6456 =============================================================================== Firmware Version: 4.2.412.0 Supported On: IBM System x and BladeCenter Problems Fixed: - All VFS on Port 1 on optional Card unreachable from outside the host (x3750 M4) - 'Logical Admin Link' not enabled after applying BOFM in Switch Agnostic mode - IBM pre-boot Diagnostics NIC physical loopback test fails in Command and GUI mode - Multiple cosmetic issues with iSCSI UEFI BIOS - 10Gb ports do not detect 'Link Down' with some switches - 1Gb ports on HS23 LOM not able link up with IBM ICPM (FRU part 44W4486) in BCS chassis - Performance issues with specific DAC cables and SND top of rack switch Interoperability Since Previous Release: - UEFI 2.3.1 Compliance - Increased VF count in SR-IOV (was 16 VFs per port, is now 32 VFs per port) - Support for next generation 10GbaseT PHY - Expose IBM Feature On Demand licensing information in UEFI configuration utility Known Issues: - All 10Gb ports must be used with 10Gb switches. They will not link at 1Gb. - IPv6 IP address not visible via IMM GUI for IBM Option 90Y6456 - Intermittent link status with IBM Virtual Fabric Switch when set to IBM Virtual Fabric Mode - The option to enable/disable SRIOV on HS23 LOM Port 0 is missing from OneCommand Manager. This selection is available in UEFI F1 setup. =============================================================================== Firmware Version: 4.1.455.6 Supported On: IBM System x and BladeCenter Problems Fixed: - Fixed an issue where ASIC randomly hangs losing network connectivity when certain commands are sent from the host. Resulting behavior varies between OS's (Yellow bangs in Windows, PSOD in VMWare) - PHY Loopback fails on 10G ports - UE and hang on LOM or Robalo when powering off Janice in UEFI or PXESelect - Loopback on HS23 now functional in DSAp - DCBX does not converge with firmware on a FCoE fabric - Control C Broken on BE2 - 7 second delay in GET_CARD_ATTRIBUTES command - Emulex VF adapter FCoE link lost on VMware 5.0 - flow control set for 10G Base T phys for autonegotiation - Dynamic link update is not working with NIC UEFI driver - Remove unused definitions in UEFI NIC - Emulex driver expects that the Login Options field be zero - "map r" displays a iSCSI BLK device when no LUNs presented - "X/x Exit without save" does not work as expected - Display proper description as part of EFI Driver Model Related Information - Intermittent "Configuration changed. reset to apply it now?" - Garbled output being displayed at the LUN configuration section. - All Emulex Configuration settings get lost after a reset - Perform additional checks when no Luns are mapped to the target - CHAP name only accepts 6 to 10 characters - Hii ???iSCSI Initiator Configuration??? showing old version screen - iSCSI name is case sensitive - iSCSI HII allows user to select first 8 LUNs as bootable. - iSCSI Boot Enable/Disable option should be port specific - LUN Configuration shows LUN 0, even if no LUNs presented to initiator. Interoperability Since Previous Release: - Support for Emulex Virtual Fabric Adapter Advanced (CFFh, 2+2) for IBM BladeCenter =============================================================================== Firmware Version: 4.1.334.47 Supported On: IBM System x and BladeCenter Problems Fixed: - NCSI transmit workaround for IBM - Issues with data reporting when BOFM enabled - Improved network performance throughput - Issues with IBM Management network performance - Logical link status for IBM mezz card shown as 'N/A' - Serial Over LAN fails when more than one blade installed with a single switch - Removed UEFI menu items for unsupported functions on a port - Issues with diagnositics in virtual NIC/UMC mode - Issues with MAC and PHY loopback on OneConnect devices - Several issues with FCoE on some OEM fabric switches Interoperability Since Previous Release: - Support for HS23 Virtual Fabric LAN on Motherboard - Support for VFA III (95Y3762, 95Y3768), 90Y6456 - Support for IBM's Features on Demand (FOD) - Formal support for SR-IOV =============================================================================== Firmware Version: 4.0.1062.0 Supported On: IBM System x and BladeCenter Problems Fixed: - iLab Tucson - lpfc Device Resets and path failures continue after error inject. - Writes do not complete when running LIP test on BE3 port 2 - IPV6 implementations no longer experience an unrecognized routing type error. - VLAN tagging is now disabled on the OneConnect UCNA when multichannel is disabled. - UEFI/BIOS-assigned LPVIDs now function properly on the OneConnect 10102-based UCNA. - OneConnect UCNAs no longer send and receive only double-tagged packets and discard all other packets. - PHY, MAC, and External loopback tests no longer fail on all platforms. - Firmware now returns the proper OneConnect UCNA model number. - IPV6 implementations no longer experience an unrecognized routing type error. - Installing Windows 2008 or Windows 2008 R2 on remote iSCSI disks no longer causes the system to hang. - Issues no longer occur when resetting a server configured for multipathing on VMware ESX systems. - VLAN tagging is now disabled on the OneConnect UCNA when multichannel is disabled. - UEFI/BIOS-assigned LPVIDs now function properly on the OneConnect 10102-based UCNA. - OneConnect UCNAs no longer send and receive only double-tagged packets and discard all other packets. - An adapter hardware error no longer occurs when running I/O with LUNs on x86 platforms. - Changing from Convergence Enhanced Ethernet (CEE) to Chassis Internal Network (CIN) and then rebooting no longer fails to link-flap (CIN Fabric Login (FLOGI) now occurs). - PHY, MAC, and External loopback tests no longer fail on all platforms. - Firmware now returns the proper OneConnect UCNA model number. - DCB settings can now be configured. - (iSCSISelect) When enabling DHCP for an iSCSI initiator, the Emulex iSCSI Network Configuration utility now waits approximately 20 seconds to make a DHCP connection before returning control. - (FCoE BootBIOS) FCF CEE parameters for Boot from SAN configuration are now saved properly. Interoperability Since Previous Release: - Added support for switch-independent multichannel mode. - Added support for IBM Virtual Fabric 10Gb RackSwitch firmware 6.8.1.0. - Added support for IBM Virtual Fabric 10G Switch Module firmware 6.8.1.0. - Added support for Fibre Channel over Ethernet (FCoE) on rack systems. - Added support for iSCSI hardware offload on rack systems. - Added support for Cisco Nexus 5000-series switches. =============================================================================== Firmware Version: 2.103.397.3806 (IBM BladeCenter) Firmware Version: 2.103.269.31 (IBM System x) - Initial Release ===============================================================================
Lenovo Data Center Group Linux OS Support Home Page |
linux.lenovo.com |
© 2018-2019 Lenovo. All rights reserved |
=============================================================================== Emulex FCoE Device Driver for Linux =============================================================================== Device Driver Version: 12.0.1169.32 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Purley System Support - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - None =============================================================================== Firmware Version: 11.4.1186.5 Supported On: Lenovo System x Rack and Flex Problems Fixed: - RHEL7.4 Alpha Server with lpfc FCoE HBA does not boot w/NVMe lpfc patches (Regression for FCoE) (203422) Incremental Interoperability: - RHEL7 Update 4 Support - SLES12 SP3 Support Known Issues: - None =============================================================================== Firmware Version: 11.2.1112.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: * Support for increasing MAX_REMOTE_NODE for HLM mode. * EAdd support for Auto Transfer Ready in OCS_FC_SCST driver for SCST backend. Incremental Interoperability: - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - Beginning with software release 11.2, Emulex LightPulse adapters and OneConnect adapters have independent software kits. The "Broadcom Software Kit Migration Guide" document provides special instructions and considerations for using the driver kits for LightPulse and OneConnect adapters. https://docs.broadcom.com/docs/12378907 =============================================================================== Device Driver Version: 11.1.152.22 Supported On: Lenovo System x Rack and Flex Problems Fixed: - FCoE instability/lost link problems on NX548 switch failover tests (182149) Incremental Interoperability: - FC SRIOV with LPe31000 and LPe32000 Known Issues: - None =============================================================================== Device Driver Version: 11.0.240.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Support for Gen 6 Fibre Channel HBAs - Improved logging and debug capabilities Known Issues: - None =============================================================================== Device Driver Version: 10.6.228.31 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - SLES 12.1 - RHEL 7.2 Known Issues: - None =============================================================================== Device Driver Version: 10.6.205.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Incremental Interoperability: - Known Issues: - None =============================================================================== Device Driver Version: 10.4.299.0 Supported On: System x Rack Problems Fixed: - LPe16000 driver crash during performance testing Incremental Interoperability: - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.135.0 Supported On: Flex Problems Fixed: - Error recovery of a "hung" IO to a device is cancelling ALL IOs, not just the culprit one only - A NULL pointer dereference kernel panic occurs in lpfc_target_reset_handler() Incremental Interoperability: - SLES 12 - RHEL 6.6, 7.1 - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.71.0 Supported On: Flex Problems Fixed: - Error recovery of a "hung" IO to a device is cancelling ALL IOs, not just the culprit one only Incremental Interoperability: - Grantley based ITE Known Issues: - None =============================================================================== Device Driver Version: 10.2.375.0 Supported On: IBM System x and BladeCenter Problems Fixed: - ELS txcmplq list corruption cause crash during error injection - Cleaned up several issues around ExpressLane feature - Need a HAL workaround for upgrading 10.0 XE102/XE104 f/w to 10.2 - I/O hang on RHEL6.5 for small block (4K) random i/o on POWER8 Incremental Interoperability: - Grantley Rack Servers Known Issues: - None =============================================================================== Device Driver Version: 10.2.261.0 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - lpfc message number 0127 is not reporting the correct information - Replace CRC16 used in T10-PI related driver code Incremental Interoperability: - Crosslink & Express Lane on LPe16k based solutions Known Issues: - None =============================================================================== Device Driver Version: 10.0.800.0 Supported On: IBM System x Problems Fixed: - Low bi-di throughput seen on 1g ports of HS-23 LOM - Multiple performance improvements - Bonding driver takes approx 20 seconds to be ready Incremental Interoperability: - RHEL 5.10 - RHEL 6.5 - XE-102 and XE-104 ASIC (OCx14xxx) - Support for set and get channels ethtool operation Known Issues: - SR-IOV is supported with the following Linux distributions: * RHEL 6.4 and above 64-bit with KVM * SLES11 SP3 and above 64-bit with KVM NOTE: Guest operating system support is dependent on the Emulex driver version installed in the guest. =============================================================================== Device Driver Version: 8.3.7.27 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Execution of internal loopback causes a crash - Issue which can cause multipath failover to not work properly Incremental Interoperability: - IBM Flex System FC5052 2-port 16Gb FC Adapter - IBM Flex System FC5054 4-port 16Gb FC Adapter - IBM Tape unit interoperability - see SSIC for more information - SLES 11.3 on IBM System x and BladeCenter Known Issues: - None =============================================================================== Device Driver Version: 8.3.7.18 Supported On: IBM System x and BladeCenter Problems Fixed: - FLOGI sent with Class 2 in-order bit set when Class 2 is set as unsupported class of service - OXID Reused before RA_TOV without RRQ to recover - Intermittently storage paths will not recover after switch reboot Incremental Interoperability: - RHEL 5.9 - RHEL 6.4 Known Issues: - No known issues =============================================================================== Device Driver Version: 8.3.6.4 Supported On: IBM Flex Problems Fixed: - Fixed debug helper routine failed to dump CQ and EQ entries in non-MSI-X mode - FLOGI sent with Class 2 in-order bit set when Class 2 is set as unsupported class of service - Fixed boot from san failure when SLI4 FC device presented on the same PCI bus - Fix ABTS recovery to be FC spec compliant - Fixed not reporting logical link speed to SCSI midlayer when QoS not enabled - Fix performance problem in pt2pt mode - Fixed SCSI host create showing wrong link speed on SLI3 HBA ports - Update text for misconfigured port messages - Fix discovery bug when swapping cables Interoperability since previous release: - RHEL6.3 GA Kernel Known Issues: - No known issues =============================================================================== Device Driver Version: 8.3.5.86 Supported On: IBM System x and BladeCenter Problems Fixed: - Kernel panics with FCoE on HS23 in SLES 11.2 x86_64. Also requires a SLES11.2 kernel update - FLOGI sent with Class 2 in-order bit set when Class 2 is set as unsupported class of service Interoperability since previous release: - RHEL6.3 GA Kernel Known Issues: - No known issues =============================================================================== Device Driver Version: 8.3.5.74.1p Supported On: IBM System x and BladeCenter Problems Fixed: - Fixed kernel panic after scsi_eh escalation by checking the proper return status =============================================================================== Device Driver Version: 8.3.5.74 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Fixed system crash due to not providing SCSI error-handling host reset handler - Debug helper utility routines for dumping various SLI4 queues - Fix unsol abts xri lookup - Fixed issues with LPe16000 to LPe16000 discovery - Fix driver crash during back-to-back ramp events - Reregister VPI for SLI3 after cable moved to new Saturn port - Fix log message 2597 displayed when no error is detected - Fixed FCP LOG for easier Finisar trace correlation - Fix kernel panic when going into to sleep state - Fix error message displayed even when not an error - Fix Read Link status data - Fix unable to create vports on FCoE SLI4 adapter - Fix initiator sending flogi after acking flogi from target - Fix the case where an array controller does not apply FCP priority rules - Added support for handling SLI4-port XRI resource-provisioning profile change - Change default DA_ID support from disabled to enabled - Fix bug with driver unload leaving a scsi host for a vport around - Update copyright date for files modified in 2012 - Fix compilation errors on some kernels due to debugfs variable definitions - Used PCI configure space read to flush PCI function reset register write - Fixed system panic when extents enabled with large number of small blocks - Fixed system panic during EEH recovery - Fix resource leak when acc fails for received plogi - Fix SLI4 driver module load and unload test in a loop crashes the system - Fix missing CVL event causing round-robin FCF failover process to stop - Fix deadlock during adapter offline request - Fix same RPI registered multiple times after HBA reset =============================================================================== Device Driver Version: 8.3.5.66 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Fix SLI4 driver module load and unload test in a loop crashes the system - Fix missing CVL event causing round-robin FCF failover process to stop - Fix deadlock during adapter offline request - Fix same RPI registered multiple times after HBA reset - Fix driver handling of XRI Aborted CQE response - Fixed port and system failure in handling SLI4 FC port function reset - Fix bug with driver not sending a LOGO with vport delete - Make BA_ACC work on a fully qualified exchange - Fix for SLI4 Port delivery for BLS ABORT ACC - Fix KERNEL allocation while lock held - Incorrect usage of bghm for BlockGuard errors - Fix warning on i386 system - Fix ndlp nodelist not empty wait timeout during driver unloading - Fix mailbox and vpi memory leaks - Support character device to take a reference on the driver - Fix for FDISC failures after firmware reset or link bounce - Support debugfs for injecting T10 DIF errors - Fix SLI4 BlockGuard behavior when protection data is generated by HBA - Fix for driver using duplicate RPIs after 16Gb/s adapter port reset - Fix discovery problem when in pt2pt - Fixed failure in handling large CQ/EQ identifiers in an IOV environment - Fix Locking code raises IRQ twice - Fix driver not returning when bad ndlp found in abts error event handling - Support more driver logs in area of SLI4 port error attention and reset recovery - Fix bug with driver returning the wrong ndlp - Sync up with actual upstream lpfc driver code - Support SLI4 16Gb FC Adapter support for T10 DIF / BlockGuard - Merge from upstream: scsi: Fix up files implicitly depending on module.h inclusion - Fix driver behavior when receiving an ADISC - Fixed unbounded firmware revision string from port caused the system panic - Fix bug with driver processing dump command type 4 using 16Gb FC Adapter - Fix driver does not reset port when reset is needed during fw_dump - Fix ELS FDISC failing with local reject / invalid RPI - Changed SLI4 FC port internal loopback from outer serdes to inner internal - Fix REG_RPI fails on SLI4 HBA putting NPort into NPR state - Fix bug with driver processing an els command using 16Gb FC Adapter - Fix NMI seen due to CQE starvation - Fixed SLI4 FC port obtained link type and number dependent on link connection - Fixed SLI4 FC port internal loopback without SFP and external link/loopback plug - Fix driver incorrectly building fcpCdb during scsi command prep - Fixed FCP EQ memory check prevent initialize to single interrupt vector case - Fixed SLI4 16Gbit FC port internal and external diagnostic loopback tests - Fixed system panic during EEH recovery on SLI4 FC port - Fix FLOGI fails with invalid RPI (0x04) after connecting to a arbitrated loop - Fix ELS SCR failing with local reject / invalid RPI - Fix SLI4 port failures after firmware reset or dump - Fix port reset log messages indicate error when no error is seen - Fixed an issue where HBQ buffers are uninitialized in the base driver - Fixed missing shost lock protection on setting/clearing bit in vport->fc_flag - Fixed missing hbalock protection on setting/clearing bit in phba->link_flag - Fixed illegal state transition during driver unload - Fixed system crash on SLI4 FC port looopback diagnostic test with large buffers - Fixed a potential error path that might try to free DMA memory not allocated - Fixed SLI4 FC port Internal and External loopback diagnostic tests - Fix a crash while deleting 256 vports - Fix vport never finished discovery if switch runs out of resources - Fix kernel build warnings - Fix kernel panic when putting board in offline state twice - Fix memory leak when running hba resets to SLI-4 board - Fix driver message 3092 and 3116 to KERN_WARNING - Fix PCI read failure looks like a over temperature event - Fixed driver event reporting to mgmt on error attention due to forced dump - Fix not displaying firmware rev for BE UCNA boards - Fixed ABTS failure logic not rediscovering targets - Fix bug with driver not byte swapping extended mailbox data - Fix premature ndlp cleanup after second target LOGO - Fix Link pull causes I/O failures - Support loop for sli4 fc pass-through - No longer read fcoe parameters if board is not fcoe - Fix bug with driver passing a wrong count value - Support restore state and error log when sysfs board_mode attribute access failed - Fixed a typo in two of the log messages just introduced - Support for SLI4_CONFIG non-embedded COMN_GET_CNTL_ADDL_ATTR pass-through - Support for SLI4 FC port persistent linkdown through config region 23 - Fix for ABTS. Do not free original IOCB whenever ABTS fails - Fixed Linux driver inconsistency in reporting FC host port-state through OCM - Return an error if the mbox sysfs is called - Fix firmware upgrade attempt on unsupported adapter - Fix incomplete message number 2889 - Fix system crash when ASIC fails to initialize - Fix virtual fabrics bit set in FLOGI incorrectly - Fix bug with driver not returning extended mailbox data - Fix unsupported link speed accepted by SLI4 port - Incremented minor management version to reflect additional "protocol" sysfs attribute - Modify WQ handling to use entry_repost - Moved link type definitions to apply to all driver streams - Support a sysfs attribute "protocol" to report SLI4 port link protocol type - Support SLI4 INTF_TYPE and SLI_FAMILY as sub-field to the fwrev sysfs attribute - Support structure to support T10 DIF in SLI4 - Fix performance slowed due to too many register accesses - Fixed a warning when converting from simple_strtoul to strict_strtoul - Support Initial code to inject T10 DIF errors - Fixed a possible driver crash when cfg_fcp_eq_count is zero - Fixed non-embedded COMMON_GET_CNTL_ATTR mbox command failed with status 0x4 - Fixed null pointer dereference after clearing FCP policy rules - Changed the way of SLI4 device detecting physical port name - Fixed FCP policy exchange management rule parsing - Support check to lpfc_decode_firmware_rev to handle releases that have an 'X' - Fixed RPI leaks in ELS protocol handling - Fix kernel crash during boot with SLI4 card installed - Fixed NPIV FDISC failure on SLI4 if-type 2 ports - Fixed driver failed to follow IP reset procedure in SLI4 error attention handler - Fix sysfs lists fabric name for port that has been disconnected from fabric - Support FCP policy exchange management - Changed the timeout to 300 seconds for SLI_CONFIG (0x9B) mailbox command - Fixed fcp underrun reporting - Fixed a driver handling IP reset issue when PCI read return error - Fixed casting problem in calculating tx_words statistics - Support proper error code return for libdfc calling API to make proper decisions - Fix sli4 mailbox status code. Management software has dependencies - Fix compiler warning due to uninitialized local variable - Fix bus reset handler fails with bad failure code - Made proper error code return for libdfc calling API to make proper decisions - Made error log include PCI BAR bitmap from kernel when enable PCI device failed - Fixed sysfs API mismatch for building against SLES11 SP2 beta kernel - Fixed scsi midlayer queuecommand API mismatch for SLES11 SP2 beta kernel - Fixed ctlreg write bug not checking key string and limiting bytes to write - Fix default adapter name for the OCe15100 - Fixed incomplete scsi messages displayed - Fix cable pull failure on interface type 2 SLI-4 adapters - Fix enable_bg config parameter for SLI4 - Fixed not able to perform firmware reset through sysfs board_mode attribute - Fixed not recovering SLI port in handling error attention with RN bit set - Fixed two crash cases when unsolicted ELS ECHO_CMD is received - Fix direct connect does not come up for SLI4 FC ports - Fixed long wait when firmware reset to a SLI port without required privilege - Fix request firmware support for little endian systems - Fix driver firmware to match new firmware image format - Fixed SLI4 device firmware reset with SR-IOV virtual functions - Fix SLI4 CT handling for sequences > 4K - Fixed handling of unsolicited frames for vports - Fixed handling of CVL for vports - Fixed crash when aborting els IOs - Fix up CT and oxid/rxid for unsol rcv frames - Fixed mailbox double free panic - Support fcf priority record selection for fcf failover - Fixed compiler warning for putting large amount of memory on stack - Fixed driver build not building in debugfs even if kernel CONFIG_DEBUG_FS=y is set - Fix issue where the FC/FCoE Async Receive CQE did not scale for 16G FC adapters - FLOGI payload has Multiple N_port_ID set when lpfc_enable_npiv clear - Removed unused argument passed to the lpfc_bsg_diag_mode_enter routine - Fixed new ASIC device-mgmt BSG pass-through failed multi-buffer fw download - Support sysfs logging when writeable parameters are changed - Used PCI config space capability for sysfs get PF supported SR-IOV number of VFs - Support wait for SLI port status register for readyness following firmware reset - iDiag added SLI4 device PCI BAR memory mapped reigster access methods - Consolidated duplicating macro definitions - Fixed failure to enable PCIe AER prevent driver initialize and attach to device - Modified log messages for T10-PI processing - iDiag fixed mailbox capture and dump overlapping bitmap macro definition =============================================================================== Device Driver Version: 8.3.5.35 Supported On: IBM System x and BladeCenter - Initial Release ===============================================================================
Lenovo Data Center Group Linux OS Support Home Page |
linux.lenovo.com |
© 2018-2019 Lenovo. All rights reserved |
=============================================================================== Emulex iSCSI Device Driver for Linux =============================================================================== Device Driver Version: 12.0.1169.32 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Purley System Support - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - None =============================================================================== Device Driver Version: 11.4.1182.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Iscsiadm displays same iqn name for both the ports(even though different iqn configured in OCM) (204469) - Installer script(elx_iscsi_install.sh) failed while installing ISCSI driver (205145) Incremental Interoperability: - RHEL7 Update 4 Support - SLES12 SP3 Support Known Issues: - None =============================================================================== Firmware Version: 11.2.1193.12-2 Supported On: Lenovo System x Rack and Flex Problems Fixed: * SGL Chaining not working with iSCSI initiator mode Incremental Interoperability: - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - Beginning with software release 11.2, Emulex LightPulse adapters and OneConnect adapters have independent software kits. The "Broadcom Software Kit Migration Guide" document provides special instructions and considerations for using the driver kits for LightPulse and OneConnect adapters. https://docs.broadcom.com/docs/12378907 =============================================================================== Device Driver Version: 11.1.135.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Unable to configure iSCSI iface network parameters on RHEL 7.2 (iSCSI network settings ordering issue) (191012) Incremental Interoperability: - Update copyright information (191208) Known Issues: - None =============================================================================== Device Driver Version: 11.0.270.32 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Avago branded transceivers no longer throw errors when plugged to Emulex NICs and CNAs (186957) Known Issues: - None =============================================================================== Device Driver Version: 10.6.228.31 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Not able to collect OneCapture Logs on RHEL 7.1 OS (183840) Incremental Interoperability: - SLES 12.1 - RHEL 7.2 Known Issues: - None =============================================================================== Device Driver Version: 10.6.228.11 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Incremental Interoperability: - Known Issues: - None =============================================================================== Device Driver Version: 10.4.299.0 Supported On: System x Rack Problems Fixed: - ocs_iscsi_scst: Missing default CHAP name/secret - Fix DMA memory leak in driver Incremental Interoperability: - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.139.0 Supported On: Flex Problems Fixed: - None Incremental Interoperability: - SLES 12 - RHEL 6.6, 7.1 - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.74.0 Supported On: Flex Problems Fixed: - OCl14102 LOM UEFI RHEL7 MPIO not working Incremental Interoperability: - Grantley based ITE Known Issues: - None =============================================================================== Device Driver Version: 10.2.365.0 Supported On: System x and BladeCenter Problems Fixed: - be2iscsi rpm is failing with dependencies error - Unable to iSCSI SAN install RHEL/SLES/WIN2012R2 with UFP - Server is getting rebooted when we run IPv4+IPv6 bi-di NIC traffic simultaneously on UMC enabled interfaces Incremental Interoperability: - Grantley Rack Servers Known Issues: - None =============================================================================== Device Driver Version: 10.2.261.1 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Write error while running traffic - System crashed after running converged IO with driver load unload in a - Possible backlog of eq might happen if _process_cq keep processing for a long time: causes fw UE - Driver memory leak resulting in "DMA: Out of SW-IOMMU space" errors in dmesg file output Incremental Interoperability: - Adaptive interrupt coalescing - TCP window size / window scaling during offload Known Issues: - None. =============================================================================== Device Driver Version: 10.0.803.21 Supported On: IBM System x Problems Fixed: - Unable to OEMSEAddTarget or OEMSEDiscoverTaget successfully and get entries in mili2d.log - Emulex drivers are not working on RHEL5 and RHEL6 errata kernel updates - IO Read/Write almost halts with 'DMA: Out of SW IOMMU space' messages while running iSCSI Multipath I/O Incremental Interoperability: - RHEL 5.10 - RHEL 6.5 - XE-102 and XE-104 ASIC (OCx14xxx) Known Issues: - None =============================================================================== Device Driver Version: 4.6.267.4 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - RH5.9 kernel panic after rebooting the OS Incremental Interoperability: - None Known Issues: - None =============================================================================== Device Driver Version: 4.6.148.0 Supported On: IBM System x and BladeCenter Problems Fixed: - None Incremental Interoperability: - RHEL 5.9 - RHEL 6.4 - SLES 11.3 Known Issues: - None =============================================================================== Device Driver Version: 4.4.175.0 Supported On: IBM Flex Problems Fixed: - Cleaned up a cosmetic problem with warning messages during driver compile from source - System crashes when we run simultaneous IO to 6 vNIC�s with iSCSI IO - Linux kdump kernel crashes during shutdown after capturing dump, if be1iscsi driver running Incremental Interoperability: - RHEL6.3 GA Kernel Known Issues: - None =============================================================================== Device Driver Version: 4.2.398.0 Supported On: IBM System x and BladeCenter Problems Fixed: - Cleaned up a cosmetic problem with warning messages during driver compile from source Incremental Interoperability: - RHEL6.3 GA Kernel Known Issues: - None =============================================================================== Device Driver Version: 4.1.422.5 Supported On: IBM Flex - Initial release for IBM Flex =============================================================================== Device Driver Version: 4.1.402.0 Supported On: IBM System x and BladeCenter Problems Fixed: - LUN is not getting detected for OEM CNA in SLES 11 SP2 64 bit OS - ioctl semaphore deadlock when running load/unload loop with target - SLES 11 SP2 GMC: Enable support for Lnx iSCSI Driver - RHEL 5.8 GA: Enable support for Lnx iSCSI Driver =============================================================================== Device Driver Version 4.1.334.39 Supported On: IBM System x and BladeCenter Problems Fixed: - Emulex: Warning messages seen while building Emulex drivers on RHEL platform - Unable to load be1iscsi driver during iSCSI boot OS install - Modprobe be1iscsi hangs when system with max targets login - be1iscsi unload hangs the system - fw rebooted on 1st mbox cmd after driver load > unload > load - Failed to initialize iscsi driver - need to define DMA_32BIT_MASK / DMA_64BIT_MASK in Linux kernel 3.x.x =============================================================================== Device Driver Version: 2.103.397.0 Supported On: IBM System x and BladeCenter - Initial Release ===============================================================================
Lenovo Data Center Group Linux OS Support Home Page |
linux.lenovo.com |
© 2018-2019 Lenovo. All rights reserved |
=============================================================================== Emulex NIC Device Driver for Linux =============================================================================== Device Driver Version: 12.0.1169.32 Supported On: Lenovo System x Rack and Flex Problems Fixed: - None Incremental Interoperability: - Purley System Support - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - None =============================================================================== Device Driver Version: 11.4.1186.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Emulex be2net inbox driver delays the failover in bond after getting unrecoverable error in adapter (204427) - LPNIC driver installation failed through lpnic_install.sh installation script file (204826) Incremental Interoperability: - RHEL7 Update 4 Support - SLES12 SP3 Support Known Issues: - None =============================================================================== Firmware Version: 11.2.1193.31 Supported On: Lenovo System x Rack and Flex Problems Fixed: * Support asymmetric RX/TX queues * Enhance be2net to support PASSTHRU mode of SET_HSW_CONFIG cmd on BE3 Incremental Interoperability: - RHEL7 Update 3 Support - SLES12 SP2 Support Known Issues: - Beginning with software release 11.2, Emulex LightPulse adapters and OneConnect adapters have independent software kits. The "Broadcom Software Kit Migration Guide" document provides special instructions and considerations for using the driver kits for LightPulse and OneConnect adapters. https://docs.broadcom.com/docs/12378907 =============================================================================== Device Driver Version: 11.1.152.18 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Perf enhancement - Added support to switch off/on CPU power and P states to perf_tune_benet.sh script (184130) Incremental Interoperability: - None Known Issues: - None =============================================================================== Device Driver Version: 11.0.232.0 Supported On: Lenovo System x Rack and Flex Problems Fixed: - SRIOV 4port: Dmesg shows Max VF's count as zero for PF's except PF0 (187192) - Fixed several issues related to SRIOV Incremental Interoperability: - Support for Secure FW download feature Known Issues: - None =============================================================================== Device Driver Version: 10.6.228.32 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Not able to collect OneCapture Logs on RHEL 7.1 OS (183840) - Fixed race condition where repeated link toggle in promiscuous mode causes firmware to lock up (183575) - Fixed a memory leak in case of async firmware commands (184729) Incremental Interoperability: - SLES 12.1 - RHEL 6.7 & 7.2 - Support for Secure FW download feature Known Issues: - None =============================================================================== Device Driver Version: 10.6.228.9 Supported On: Lenovo System x Rack and Flex Problems Fixed: - Incremental Interoperability: - Known Issues: - None =============================================================================== Device Driver Version: 10.4.301.0 Supported On: System x Rack Problems Fixed: - Variety of performance improvements on small packet and UDP traffic Incremental Interoperability: - Add support in perf_tune_benet.sh to change the processor C-state dynamically - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.142.0 Supported On: Flex Problems Fixed: - None Incremental Interoperability: - SLES 12 - RHEL 6.6, 7.1 - Lenovo branded versions of formerly IBM branded products Known Issues: - None =============================================================================== Device Driver Version: 10.3.11.0 Supported On: Flex Problems Fixed: - None Incremental Interoperability: - Grantley based ITE Known Issues: - None =============================================================================== Device Driver Version: 10.2.363.0 Supported On: System x and BladeCenter Problems Fixed: - Multiple performance improvements in virtualized environments (SR-IOV,UMC) - Resetting Transparent Vlan to Default Vlan is not working - Cleanup ethtool set settings code - ARI on IBM x3650M4: All functions are not displayed in Linux Incremental Interoperability: - Grantley Rack Servers Known Issues: - None =============================================================================== Device Driver Version: 10.2.261.0 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Provide link state control for VFs - ethtool shows port as fibre though copper cable is connected. - Multi channel config should be determined via port resource descriptor - Showing WoL supported in ethtool for BE3 where it is not supported - be2net driver crash when umc enabled - ip link show command shows vlan for vfs although vlan creation is failed on the vfs - Emulex be2net driver does not support IOMMU pass through mode Incremental Interoperability: - PCI ARI support Known Issues: - None =============================================================================== Device Driver Version: 10.0.800.0 Supported On: IBM System x Problems Fixed: - Low bi-di throughput seen on 1g ports of HS-23 LOM - Multiple performance improvements - Bonding driver takes approx 20 seconds to be ready Incremental Interoperability: - RHEL 5.10 - RHEL 6.5 - XE-102 and XE-104 ASIC (OCx14xxx) - Support for set and get channels ethtool operation Known Issues: - SR-IOV is supported with the following Linux distributions: * RHEL 6.4 and above 64-bit with KVM * SLES11 SP3 and above 64-bit with KVM NOTE: Guest operating system support is dependent on the Emulex driver version installed in the guest. =============================================================================== Device Driver Version: 4.6.267.4 Supported On: IBM System x, BladeCenter, and Flex Problems Fixed: - Under certain conditions, performance issues may be seen with multi-channel in RHEL6.4 KVM configurations - Under certain conditions, performance issues may be seen with multi-channel in RHEL6.4 SR-IOV configurations - SR-IOV configuration section of the Linux NIC Driver Documentation has been modified to improve clarity Incremental Interoperability: - S3 Resume on IBM Flex - SLES 11.3 on IBM System x and BladeCenter Known Issues: - SR-IOV is supported with the following Linux distributions: * RHEL 6.3 and 6.4 64-bit with KVM * SLES11 SP2 and SP3 64-bit with KVM NOTE: Guest operating system support is dependent on the Emulex driver version installed in the guest. - Use of SR-IOV with the XEN hypervisor is not supported at this time. - Use of SR-IOV with the KVM hypervisor requires the use of macvlan network interfaces. =============================================================================== Device Driver Version: 4.6.148.0 Supported On: IBM System x and BladeCenter Problems Fixed: - Ethernet performance issues in a variety of workloads - VFs are not displayed after driver re-load when is VFs attached to VM - vLAN tagging in Switch Agnostic vNIC mode Incremental Interoperability: - RHEL 5.9 - RHEL 6.4 - Network performance varies with workload. An ethernet driver tuning script and instructions is now included with the be2net driver. - Switch Independent Mode within hypervisor OS's Known Issues: - Multi-channel is not supported in RHEL5.x XEN. This is a kernel limitation - Under certain conditions, performance issues may be seen with multi-channel in RHEL6.4 KVM configurations - Under certain conditions, performance issues may be seen with multi-channel in RHEL6.4 SR-IOV configurations - SR-IOV is not supported with RHEL5.x XEN =============================================================================== Device Driver Version: 4.4.180.0 Supported On: IBM Flex Device Driver Version: 4.2.389.0 Supported On IBM System x and BladeCenter Problems Fixed: - Unloading the driver hangs the system after running IPv6 traffic - Link status messages cleaned up when in Switch Agnostic mode - Miscellaneous network performance issues to improve throughput - Miscellaneous problems related to SRIOV to improve usability - Ethtool wol command displays wol enabled even when disabled - Ethtool speed reported as 10000Mb/s for disabled devices in Switch Agnostic mode Incremental Interoperability: - RHEL6.3 GA Kernel - Increase default B/W for SRIOV vf's to 10Gb from 100Mb =============================================================================== Device Driver Version: 4.1.440.0 Supported On: IBM System x and BladeCenter Device Driver Version: 4.1.422.5 Supported On: IBM Flex Problems Fixed: - Ethtool information missing with newest version of be2net =============================================================================== Device Driver Version: 4.1.413.0 Supported On: IBM System x and BladeCenter Problems Fixed: - WOL changes for Linux NIC driver as mentioned in Bug# 24543 - SR IOV UE when attempt to do multiple (3 times) driver load/unload - Speed is 10000Mb/s for Disabled devices in UMC Mode - Incorrect Duplex value is shown for unconnected port of NC552m in Rhel6.1 32bit - SLES 11 SP2 GMC: Enable support for Lnx NIC Driver - SRIOV: When VGT is configured for a VF, ip link show doesn't display it for the VF configured. - RHEL 5.8 GA: Enable support for Lnx NIC Driver - ip link command with vlan = 4095 has toggling effect - ip link command with vlan =0 should remove guest VGT privilege =============================================================================== Device Driver Version: 4.1.334.39 Supported On: IBM System x and BladeCenter Problems Fixed: - SR-IOV: Can't load be2net driver in a VM after the VM has been rebooted - SR-IOV: Unload PF driver with VF driver loaded causes system instability - SR-IOV: VF MAC always gets Emulex OUI - SR-IOV: Partition the limited per port VLAN entries across VFs and PFs - Non Member VLAN packets not received in interface promiscuous mode in BE3 - Avoid CQ full condition in HW - be2net: use RX_FILTER cmd to program multicast addresses - account for skb allocation failures in ethtool stats - Support 60 second delay to Linux NIC driver to allow FAT dump completion on recovery from EEH - Link status is up [ethtool and OCM] when the bandwidth is assign 0 for UMC configuration - Remove msix module parameter - Use V1 version of COMMON_NTWK_LINK_STATUS_QUERY for Lancer. - Emulex: Ethtool failing to update FW - Emulex: Panic in be2net driver in SLES11-SP2-B7 on adapter - TCP ACKs from different connections are getting received in single queue - Multicast filtering feature not working on PPC - Exit of mcast promiscous mode not working - VLAN promiscous mode cannot be set on Lancer - Can't communicate after changing MAC address via VLAN interface - Improper memory accounting in received packets - WOL changes for Linux NIC driver as mentioned in bug - Transparent VLAN tagging feature not working in SR-IOV =============================================================================== Device Driver Version: 2.103.397.0 Supported On: IBM System x and BladeCenter - Initial Release ===============================================================================
Lenovo Data Center Group Linux OS Support Home Page |
linux.lenovo.com |
© 2018-2019 Lenovo. All rights reserved |