Sponsored Content
Special Forums Hardware Fibre Channel HBA recommendations? Post 302653463 by hicksd8 on Saturday 9th of June 2012 05:31:45 AM
Old 06-09-2012
They are both excellent manufacturers with good support.

QLogic do supply their chips to motherboard manufacturers to mount on a variety of systems which means that their product is wide-spread and not just on their own HBA's. As a result my personal preference would QLogic but there really isn't much in it.
 

9 More Discussions You Might Find Interesting

1. AIX

Fibre channel drivers on RS/6000 aix 5L

Want to configure IBM raid strorage but the aix 5L cds do not have the drivers for the fibre channels. The machine is RS/6000. I have gone to IBM downlaodable sites but i can't find the drivers? help pliz:mad: (4 Replies)
Discussion started by: Zim-Aix-Guru
4 Replies

2. Shell Programming and Scripting

4 x fibre cards elimination

Hi, I have variables like mentioned below and want to separate as mentioned in output. var1="4 x fibre cards" var2="2 fibre cards" var3="6 - fibre cards" var4="4 x dual-port" I have variables like this.I want to separate numbers from this and fibre cards from this. output has to be... (4 Replies)
Discussion started by: ajilesh
4 Replies

3. AIX

Install Fibre Card AIX 5.3

Hello, I have two systems that are being prepared to be SAN attached .. can anyone tell me any specific checks I should perform prior to the cards being installed... I am aware of firmware / OS level and relevant drivers, is there anything else? thanks Chris. (8 Replies)
Discussion started by: chlawren
8 Replies

4. AIX

Power6 Virtual Fibre Channel Adapter

Hello, Searched in all IBM Redbooks and on the internet and couldn't find anything about the new feature of POWER 6 which Virtual Fibre ( Fiber ) channel adapter. It is similar to virtual scsi adapter. In my client partition I created the virtual Fibre Adapter mapped it with the VIO... (1 Reply)
Discussion started by: filosophizer
1 Replies

5. Solaris

USCSICMD ioctl calls for Fibre Channel(FC) devices on Solaris 10?

Hi , I have wrtitten a C program that issues USCSICMD ioctl call to the tape devices attached on solaris sparc 10. I was able to get the required information from all SCSI tape devices attached using the utility. But, whenever it is run on FC attached tape drives , the program returns an error... (0 Replies)
Discussion started by: naveen448
0 Replies

6. Solaris

How to enable hba port of a two port hba card

Hi, i have a two port hba card in solaris 10 system one of which is disabled. How to enable and configure the other port. ---------- Post updated at 08:17 AM ---------- Previous update was at 03:40 AM ---------- please help :( (7 Replies)
Discussion started by: rishiraaz
7 Replies

7. Solaris

Fibre channel link down on booting Solaris server

Hi I had power issue that affected a server, in which I had power ON the server SPARC T1-B3 running solaris 10. After power on the system stops at ok prompt, them I issued the following commands: {0} ok setenv auto-boot? false auto-boot? = false {0} ok reset-all SPARC T3-1B,... (10 Replies)
Discussion started by: fretagi
10 Replies

8. AIX

[ASK] - AIX Fibre Channel behavior

Hello all, Let me introduce about the context and my environment. We have an AIX 6.1 system, it has 4 FC channels / > lsdev -Cc adapter | grep fcs fcs0 Available 23-T1 Virtual Fibre Channel Client Adapter fcs1 Available 23-T1 Virtual Fibre Channel Client Adapter fcs2 Available 23-T1... (14 Replies)
Discussion started by: Phat
14 Replies

9. Solaris

Fibre Channel link not ready on Netra 240

Hi, One of my Netra 240 went into hung state and I had to reboot it. I powered it off and tried booting it again but unsuccessful. It is not connected to SAN and have local disks. Not able to boot in failsafe mode too. There are two disks of 72GB, both are mirrored in SVM. It complains about... (5 Replies)
Discussion started by: solaris_1977
5 Replies
BXE(4)							   BSD Kernel Interfaces Manual 						    BXE(4)

NAME
bxe -- QLogic NetXtreme II Ethernet 10Gb PCIe adapter driver SYNOPSIS
To compile this driver into the kernel, place the following lines in your kernel configuration file: device bxe Alternatively, to load the driver as a module at boot time, place the following line in loader.conf(5): if_bxe_load="YES" DESCRIPTION
The bxe driver provides support for PCIe 10Gb Ethernet adapters based on the QLogic NetXtreme II family of 10Gb chips. The driver supports Jumbo Frames, VLAN tagging, checksum offload (IPv4, TCP, UDP, IPv6-TCP, IPv6-UDP), MSI-X interrupts, TCP Segmentation Offload (TSO), Large Receive Offload (LRO), and Receive Side Scaling (RSS). HARDWARE
The bxe driver provides support for various NICs based on the QLogic NetXtreme II family of 10Gb Ethernet controller chips, including the following: o QLogic NetXtreme II BCM57710 10Gb o QLogic NetXtreme II BCM57711 10Gb o QLogic NetXtreme II BCM57711E 10Gb o QLogic NetXtreme II BCM57712 10Gb o QLogic NetXtreme II BCM57712-MF 10Gb o QLogic NetXtreme II BCM57800 10Gb o QLogic NetXtreme II BCM57800-MF 10Gb o QLogic NetXtreme II BCM57810 10Gb o QLogic NetXtreme II BCM57810-MF 10Gb o QLogic NetXtreme II BCM57840 10Gb / 20Gb o QLogic NetXtreme II BCM57840-MF 10Gb CONFIGURATION
There a number of configuration parameters that can be set to tweak the driver's behavior. These parameters can be set via the loader.conf(5) file to take affect during the next system boot. The following parameters affect ALL instances of the driver. hw.bxe.debug DEFAULT = 0 Sets the default logging level of the driver. See the Diagnostics and Debugging section below for more details. hw.bxe.interrupt_mode DEFAULT = 2 Sets the default interrupt mode: 0=IRQ, 1=MSI, 2=MSIX. If set to MSIX and allocation fails, the driver will roll back and attempt MSI allocation. If MSI allocation fails, the driver will roll back and attempt fixed level IRQ allocation. If IRQ allocation fails, then the driver load fails. With MSI/MSIX, the driver attempts to allocate a vector for each queue in addition to one more for default processing. hw.bxe.queue_count DEFAULT = 4 Sets the default number of fast path packet processing queues. Note that one MSI/MSIX interrupt vector is allocated per-queue. hw.bxe.max_rx_bufs DEFAULT = 0 Sets the maximum number of receive buffers to allocate per-queue. Zero(0) means to allocate a receive buffer for every buffer descriptor. By default this equates to 4080 buffers per-queue which is the maximum value for this config parameter. hw.bxe.hc_rx_ticks DEFAULT = 25 Sets the number of ticks for host interrupt coalescing in the receive path. hw.bxe.hc_tx_ticks DEFAULT = 50 Sets the number of ticks for host interrupt coalescing in the transmit path. hw.bxe.rx_budget DEFAULT = 0xffffffff Sets the maximum number of receive packets to process in an interrupt. If the budget is reached then the remaining/pending packets will be processed in a scheduled taskqueue. hw.bxe.max_aggregation_size DEFAULT = 32768 Sets the maximum LRO aggregration byte size. The higher the value the more packets the hardware will aggregate. Maximum is 65K. hw.bxe.mrrs DEFAULT = -1 Sets the PCI MRRS: -1=Auto, 0=128B, 1=256B, 2=512B, 3=1KB hw.bxe.autogreeen DEFAULT = 0 Set AutoGrEEEN: 0=HW_DEFAULT, 1=FORCE_ON, 2=FORCE_OFF hw.bxe.udp_rss DEFAULT = 0 Enable/Disable 4-tuple RSS for UDP: 0=DISABLED, 1=ENABLED Special care must be taken when modifying the number of queues and receive buffers. FreeBSD imposes a limit on the maximum number of mbuf(9) allocations. If buffer allocations fail, the interface initialization will fail and the interface will not be usable. The driver does not make a best effort for buffer allocations. It is an all or nothing effort. You can tweak the mbuf(9) allocation limit using sysctl(8) and view the current usage with netstat(1) as follows: # netstat -m # sysctl kern.ipc.nmbclusters # sysctl kern.ipc.nmbclusters=<#> There are additional configuration parameters that can be set on a per-instance basis to dynamically override the default configuration. The '#' below must be replaced with the driver instance / interface unit number: dev.bxe.#.debug DEFAULT = 0 Sets the default logging level of the driver instance. See hw.bxe.debug above and the Diagnostics and Debugging section below for more details. dev.bxe.#.rx_budget DEFAULT = 0xffffffff Sets the maximum number of receive packets to process in an interrupt for the driver instance. See hw.bxe.rx_budget above for more details. Additional items can be configured using ifconfig(8): MTU - Maximum Transmission Unit DEFAULT = 1500 RANGE = 46-9184 # ifconfig bxe# mtu <n> Promiscuous Mode DEFAULT = OFF # ifconfig bxe# [ promisc | -promisc ] Rx/Tx Checksum Offload DEFAULT = RX/TX CSUM ON Note that the Rx and Tx settings are not independent. # ifconfig bxe# [ rxcsum | -rxcsum | txcsum | -txcsum ] TSO - TCP Segmentation Offload DEFAULT = ON # ifconfig bxe# [ tso | -tso | tso6 | -tso6 ] LRO - TCP Large Receive Offload DEFAULT = ON # ifconfig bxe# [ lro | -lro ] DIAGNOSTICS AND DEBUGGING
There are many statistics exposed by bxe via sysctl(8). To dump the default driver configuration: # sysctl -a | grep hw.bxe To dump every instance's configuration and detailed statistics: # sysctl -a | grep dev.bxe To dump information for a single instance (replace the '#' with the driver instance / interface unit number): # sysctl -a | grep dev.bxe.# To dump information for all the queues of a single instance: # sysctl -a | grep dev.bxe.#.queue To dump information for a single queue of a single instance (replace the additional '#' with the queue number): # sysctl -a | grep dev.bxe.#.queue.# The bxe driver has the ability to dump a ton of debug messages to the system log. The default level of logging can be set with the hw.bxe.debug sysctl(8). Take care with this setting as it can result in too many logs being dumped. Since this parameter is the default one, it affects every instance and will dramatically change the timing in the driver. A better alternative to aid in debugging is to dynami- cally change the debug level of a specific instance with the dev.bxe.#.debug sysctl(8). This allows you to turn on/off logging of various debug groups on-the-fly. The different debug groups that can be toggled are: DBG_LOAD 0x00000001 /* load and unload */ DBG_INTR 0x00000002 /* interrupt handling */ DBG_SP 0x00000004 /* slowpath handling */ DBG_STATS 0x00000008 /* stats updates */ DBG_TX 0x00000010 /* packet transmit */ DBG_RX 0x00000020 /* packet receive */ DBG_PHY 0x00000040 /* phy/link handling */ DBG_IOCTL 0x00000080 /* ioctl handling */ DBG_MBUF 0x00000100 /* dumping mbuf info */ DBG_REGS 0x00000200 /* register access */ DBG_LRO 0x00000400 /* lro processing */ DBG_ASSERT 0x80000000 /* debug assert */ DBG_ALL 0xFFFFFFFF /* flying monkeys */ For example, to debug an issue in the receive path on bxe0: # sysctl dev.bxe.0.debug=0x22 When finished turn the logging back off: # sysctl dev.bxe.0.debug=0 SUPPORT
For support questions please contact your QLogic approved reseller or QLogic Technical Support at http://support.qlogic.com, or by E-mail at <support@qlogic.com>. SEE ALSO
netstat(1), altq(4), arp(4), netintro(4), ng_ether(4), vlan(4), ifconfig(8) HISTORY
The bxe device driver first appeared in FreeBSD 9.0. AUTHORS
The bxe driver was written by Eric Davis <edavis@broadcom.com>, David Christensen <davidch@broadcom.com>, and Gary Zambrano <zambrano@broadcom.com>. BSD
April 29, 2012 BSD
All times are GMT -4. The time now is 07:59 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy