Have you enabled the jumbo frame on both sides?

> The machines are connected by swich and JUMBO is enabled on switch
>  
> ProCurve Switch 6400cl-6XG# show vlan 1
>
>  Status and Counters - VLAN Information - Ports - VLAN 1
>
>   802.1Q VLAN ID : 1
>   Name : DEFAULT_VLAN
>   Status : Port-based
>   Voice : No
>   Jumbo : Yes
>
>   Port Information Mode     Unknown VLAN Status
>   ---------------- -------- ------------ ----------
>   1                Untagged Learn        Up
>   2                Untagged Learn        Up
>   3                Untagged Learn        Down
>   4                Untagged Learn        Down
>   5                Untagged Learn        Down
>   6                Untagged Learn        Down
>
>
>  
> Zhihui
>  
> 2008/6/25, Samuel Tu <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>:
>
>     Hi Zhihui,
>
>     Are the adapters connected back to back or through switch. It it's
>     switch, Please check the setting of the switch.
>
>     Thanks,
>     Samuel
>
>     zhihui Chen wrote:
>
>         I am not sure what the bug 6716683 means. But even if I send
>         out big-size (4096) packet and MTU is set to 8000, the packet
>         is still split to small packet, why?
>          intel5# ifconfig ixgbe0
>         ixgbe0:
>         flags=201000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4,CoS> mtu
>         8000 index 6
>                inet 10.0.7.5 <http://10.0.7.5/> <http://10.0.7.5
>         <http://10.0.7.5/>> netmask ffffff00 broadcast 10.0.7.255
>         <http://10.0.7.255/> <http://10.0.7.255 <http://10.0.7.255/>>
>                ether 0:30:48:94:62:e6
>         intel5#netperf -H intel3-7 -l 30 -t TCP_STREAM -- -m 4096
>         intel5#snoop -d ixgbe0 to host intel3-7
>         .......
>          intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820833814 Len=1460 Win=56940
>            intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820835274 Len=1460 Win=56940
>            intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820836734 Len=1176 Win=56940
>            intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820837910 Len=1460 Win=56940
>            intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820839370 Len=1460 Win=56940
>            intel5-7 -> intel3-7     TCP D=63923 S=43515 Ack=4133836232
>         Seq=1820840830 Len=1176 Win=56940
>         ........
>
>         zhihui
>          2008/6/25, Paul Guo - SUN ERI - Beijing China
>         <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>:
>
>            I guess you now hit the bug 6716683.
>            http://bugs.opensolaris.org/view_bug.do?bug_id=6716683
>            ixgbe tx could send package larger than max_frame_size
>
>            This bug sometimes causes traffic problems when a large
>            packet is passed to the driver (e.g jumbo frame or lso frame).
>            There is no workaround for it, but I will put-back the fix
>            along with the lso support soon.
>
>            Thanks,
>            Paul
>             > Thanks, I have used your method to set the MTU to 8000,
>         but big
>            issues
>             > rise up
>             > (1) Ping with msgsize>1476 will not work.
>             > (2) Ping with msgsize<=1476 works, but the latency
>         become longer.
>             > (3) Netperf testing can not be executed. Even if I kill
>         the netperf
>             > process, status of socket created by netperf still remains
>             > "ESTABLIBSHED", I think "FIN" packet have not been sent out
>             > successfully by each other.
>             > Thanks,
>             > Zhihui
>             >
>             > 2008/6/24, Paul Guo - SUN ERI - Beijing China
>         <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>             > <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>:
>             >
>             >     You can set default_mtu in /kernel/drv/ixgbe.conf to
>         enable jumbo
>             >     frame.
>             >     Please remember to rem_drv/add_drv or update_drv to
>         make it
>            to become
>             >     effective.
>             >     Currently we are working on adding/enabling/tuning more
>            features (such
>             >     as LSO/TSO, multiple
>             >     rings, etc) which will eventually increase the
>         performance
>             >     greatly. Stay
>             >     tuned.
>             >
>             >     -Paul
>             >
>             >     > Why the maximum MTU for ixgbe is only 1500, same
>         to e1000g?
>             >     > With 1500 mtu, the maximum throughput using
>         netperf for 10G
>             >     network is
>             >     > less than 3Gbit/s
>             >     >
>             >     >
>             >     > 2008/6/23, Samuel Tu <[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>             >     <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>             >     <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>>:
>
>             >     >
>             >     > Glad to know it works. We're go on the development
>         with more
>             >     features
>             >     > supporting. The LSO supporting is coming.
>             >     >
>             >     > Thanks,
>             >     > Samuel
>             >     >
>             >     > zhihui Chen wrote:
>             >     >
>             >     > Thanks a lot. it works now and I can ping from
>         this machine to
>             >     > another machine with another same-type adapter. We
>         will make a
>             >     > detail perfomance testing in next few weeks.
>             >     > Zhihui
>             >     >
>             >     >
>             >     > 2008/6/23, zhihui Chen <[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
>             >     > <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]> <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>
>             >     <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
>             >     > <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]> <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>>>:
>             >     >
>             >     > Thank Samuel. I have tried the "add_drv" command
>         and each
>            time it
>             >     > prints following information:
>             >     > ("ixgbe") already in use as a driver or alias.
>             >     > But I use "modinfo | grep ixgbe" can not find it.
>             >     > What is the problem?
>             >     > Zhihui
>             >     >
>             >     > 2008/6/23, Samuel Tu <[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
>             >     > <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]> <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>
>             >     <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
>             >     > <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]> <mailto:[EMAIL PROTECTED]
>         <mailto:[EMAIL PROTECTED]>>
>            <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>
>         <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>>>>:
>             >     >
>             >     >
>             >     > Hi Zhihui,
>             >     >
>             >     > Intel 82598EB (Oplin) is supported by ixgbe which
>         has been
>             >     > integrated
>             >     > into snv_90. The pci ids supported by the current
>         driver are:
>             >     > "pciex8086,10c6"
>             >     > "pciex8086,10c7"
>             >     >
>             >     > The CX4 adapter can also be supported by it but
>         we've not add
>             >     > it it
>             >     > since we've not verified it on such adapter.
>             >     >
>             >     > To make it work, Please manually config it:
>             >     > 1) add_drv -i '"pciex8086,10dd" ixgbe
>             >     > 2) ifconfig ixgbe_# plumb up
>             >     >
>             >     > Please let me know the result.
>             >     >
>             >     > Thanks,
>             >     > Samuel
>             >     >
>             >     > zhihui Chen wrote:
>             >     >
>             >     > I have one 10G network adapter (Intel 82598EB) in my
>             >     > machine. But after installation of SNV_91, I can
>         not find
>             >     > this adapter through "dladm show-dev". Following
>         is the
>             >     > output of "dladm show-dev" and "scanpci"
>             >     >
>             >     > intel5# dladm show-dev
>             >     > LINK STATE SPEED DUPLEX
>             >     > e1000g0 up 1000Mb full
>             >     > e1000g1 up 1000Mb full
>             >     > intel5#/usr/X11R6/bin/scanpci
>             >     > ......
>             >     > pci bus 0x0008 cardnum 0x00 function 0x00: vendor
>         0x8086
>             >     > device 0x10dd
>             >     > Intel Corporation 82598EB 10 Gigabit AT CX4
>         Network Connection
>             >     > pci bus 0x0008 cardnum 0x00 function 0x01: vendor
>         0x8086
>             >     > device 0x10dd
>             >     > Intel Corporation 82598EB 10 Gigabit AT CX4
>         Network Connection
>             >     > ......
>             >     >
>             >     > anyone can tell me how to config my system to find
>         the adapter?
>             >     >
>             >     > Thanks,
>             >     > Zhihui
>             >     >
>             >     >
>             >     >
>             >      
>          
> ------------------------------------------------------------------------
>             >     >
>             >     > _______________________________________________
>             >     > networking-discuss mailing list
>             >     > [email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>
>             >     <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>>
>             >     > <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>
>             >     <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>>>
>             >     > <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>
>             >     <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>>
>             >     > <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>
>             >     <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>>>>
>             >     >
>             >     >
>             >     >
>             >     >
>             >     > -- zhihui
>             >     > Intel OpenSolaris Team
>             >     >
>             >     >
>             >     >
>             >     > --
>             >     > zhihui
>             >     > Intel OpenSolaris Team
>             >     >
>             >     >
>             >     >
>             >     >
>             >     > --
>             >     > zhihui
>             >     > Intel OpenSolaris Team
>             >     >
>             >      
>          
> ------------------------------------------------------------------------
>             >     >
>             >     > _______________________________________________
>             >     > networking-discuss mailing list
>             >     > [email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>
>             >     <mailto:[email protected]
>         <mailto:[email protected]>
>            <mailto:[email protected]
>         <mailto:[email protected]>>>
>             >     >
>             >
>             >
>             >
>             >
>             > --
>             > zhihui
>             > Intel OpenSolaris Team
>
>
>
>
>         -- 
>         zhihui
>         Intel OpenSolaris Team
>
>
>
>
> -- 
> zhihui
> Intel OpenSolaris Team 

_______________________________________________
networking-discuss mailing list
[email protected]

Reply via email to