DPDK patches and discussions
 help / color / mirror / Atom feed
* [dpdk-dev] Issue while running testpmd with sr-iov setting
@ 2020-01-31  1:58 김영득
  0 siblings, 0 replies; 2+ messages in thread
From: 김영득 @ 2020-01-31  1:58 UTC (permalink / raw)
  To: dev

Hi.

 

I have a fail in SR-IOV

 

I use a HPE Eth 631SFP28 2port NIC

Setting up SR-IOV will cause problems on the second port (port 1).

The configuration of SR-IOV is assigned one VF on each port. 

I added this information to the last table.

 

l  Error Log


testpmd> port stop 0

Stopping ports...

Checking link statuses...

Done

 

testpmd> port start 0

bnxt_print_link_info(): Port 0 Link Up - speed 25000 Mbps - full-duplex

Port 0: 3A:73:4B:1D:BB:AE

Checking link statuses...

Done

 

testpmd> port stop 1

Stopping ports...

Done

 

testpmd> port start 1

bnxt_hwrm_ring_alloc(): hwrm_ring_alloc tx failed. rc:1

bnxt_init_chip(): HWRM ring alloc failure rc: 1

Fail to start port 1

Please stop the ports first

Done

 

l  Port Information


testpmd> show port info all

 

********************* Infos for port 0  *********************

MAC address: 3A:73:4B:1D:BB:AE

Device name: 0000:af:02.0

Driver name: net_bnxt

Connect to socket: 1

memory allocation on the socket: 1

Link status: up

Link speed: 25000 Mbps

Link duplex: full-duplex

MTU: 1500

Promiscuous mode: enabled

Allmulticast mode: disabled

Maximum number of MAC addresses: 4

Maximum number of MAC addresses of hash filtering: 0

VLAN offload:

  strip off

  filter off

  qinq(extend) off

Hash key size in bytes: 40

Redirection table size: 128

Supported RSS offload flow types:

  ipv4

  ipv4-tcp

  ipv4-udp

  ipv6

  ipv6-tcp

  ipv6-udp

Minimum size of RX buffer: 1

Maximum configurable length of RX packet: 9600

Maximum number of VMDq pools: 64

Current number of RX queues: 1

Max possible RX queues: 230

Max possible number of RXDs per queue: 8192

Min possible number of RXDs per queue: 16

RXDs number alignment: 1

Current number of TX queues: 1

Max possible TX queues: 230

Max possible number of TXDs per queue: 4096

Min possible number of TXDs per queue: 16

TXDs number alignment: 1

 

********************* Infos for port 1  *********************

MAC address: 6A:85:4C:58:B4:CF

Device name: 0000:af:03.0

Driver name: net_bnxt

Connect to socket: 1

memory allocation on the socket: 1

Link status: up

Link speed: 25000 Mbps

Link duplex: full-duplex

MTU: 1500

Promiscuous mode: enabled

Allmulticast mode: disabled

Maximum number of MAC addresses: 4

Maximum number of MAC addresses of hash filtering: 0

VLAN offload:

  strip off

  filter off

  qinq(extend) off

Hash key size in bytes: 40

Redirection table size: 128

Supported RSS offload flow types:

  ipv4

  ipv4-tcp

  ipv4-udp

  ipv6

  ipv6-tcp

  ipv6-udp

Minimum size of RX buffer: 1

Maximum configurable length of RX packet: 9600

Maximum number of VMDq pools: 64

Current number of RX queues: 1

Max possible RX queues: 230

Max possible number of RXDs per queue: 8192

Min possible number of RXDs per queue: 16

RXDs number alignment: 1

Current number of TX queues: 1

Max possible TX queues: 230

Max possible number of TXDs per queue: 4096

Min possible number of TXDs per queue: 16

TXDs number alignment: 1

 

l  VF configuration Info


$ ip link show dev ens4f0

4: ens4f0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
mode DEFAULT group default qlen 1000

    link/ether f4:03:43:ca:f8:d0 brd ff:ff:ff:ff:ff:ff

    vf 0 MAC 00:11:22:33:44:55, spoof checking off, link-state enable,
trust off

 

$ ip link show dev ens4f1d1

5: ens4f1d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
mode DEFAULT group default qlen 1000

    link/ether f4:03:43:ca:f8:d8 brd ff:ff:ff:ff:ff:ff

    vf 0 MAC 00:11:22:33:44:77, spoof checking off, link-state enable,
trust off

 

Thank for reading

Best regard.

 


^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: [dpdk-dev] Issue while running testpmd with sr-iov setting
@ 2020-01-31  2:23 김영득
  0 siblings, 0 replies; 2+ messages in thread
From: 김영득 @ 2020-01-31  2:23 UTC (permalink / raw)
  To: dev

 

Hi.

 

I have a fail in SR-IOV

 

I use a HPE Eth 631SFP28 2port NIC

Setting up SR-IOV will cause problems on the second port (port 1).

The configuration of SR-IOV is assigned one VF on each port. 

I added this information to the last table.

 

l  Error Log


testpmd> port stop 0

Stopping ports...

Checking link statuses...

Done

 

testpmd> port start 0

bnxt_print_link_info(): Port 0 Link Up - speed 25000 Mbps - full-duplex

Port 0: 3A:73:4B:1D:BB:AE

Checking link statuses...

Done

 

testpmd> port stop 1

Stopping ports...

Done

 

testpmd> port start 1

bnxt_hwrm_ring_alloc(): hwrm_ring_alloc tx failed. rc:1

bnxt_init_chip(): HWRM ring alloc failure rc: 1

Fail to start port 1

Please stop the ports first

Done

 

l  Port Information


testpmd> show port info all

 

********************* Infos for port 0  *********************

MAC address: 3A:73:4B:1D:BB:AE

Device name: 0000:af:02.0

Driver name: net_bnxt

Connect to socket: 1

memory allocation on the socket: 1

Link status: up

Link speed: 25000 Mbps

Link duplex: full-duplex

MTU: 1500

Promiscuous mode: enabled

Allmulticast mode: disabled

Maximum number of MAC addresses: 4

Maximum number of MAC addresses of hash filtering: 0

VLAN offload:

  strip off

  filter off

  qinq(extend) off

Hash key size in bytes: 40

Redirection table size: 128

Supported RSS offload flow types:

  ipv4

  ipv4-tcp

  ipv4-udp

  ipv6

  ipv6-tcp

  ipv6-udp

Minimum size of RX buffer: 1

Maximum configurable length of RX packet: 9600

Maximum number of VMDq pools: 64

Current number of RX queues: 1

Max possible RX queues: 230

Max possible number of RXDs per queue: 8192

Min possible number of RXDs per queue: 16

RXDs number alignment: 1

Current number of TX queues: 1

Max possible TX queues: 230

Max possible number of TXDs per queue: 4096

Min possible number of TXDs per queue: 16

TXDs number alignment: 1

 

********************* Infos for port 1  *********************

MAC address: 6A:85:4C:58:B4:CF

Device name: 0000:af:03.0

Driver name: net_bnxt

Connect to socket: 1

memory allocation on the socket: 1

Link status: up

Link speed: 25000 Mbps

Link duplex: full-duplex

MTU: 1500

Promiscuous mode: enabled

Allmulticast mode: disabled

Maximum number of MAC addresses: 4

Maximum number of MAC addresses of hash filtering: 0

VLAN offload:

  strip off

  filter off

  qinq(extend) off

Hash key size in bytes: 40

Redirection table size: 128

Supported RSS offload flow types:

  ipv4

  ipv4-tcp

  ipv4-udp

  ipv6

  ipv6-tcp

  ipv6-udp

Minimum size of RX buffer: 1

Maximum configurable length of RX packet: 9600

Maximum number of VMDq pools: 64

Current number of RX queues: 1

Max possible RX queues: 230

Max possible number of RXDs per queue: 8192

Min possible number of RXDs per queue: 16

RXDs number alignment: 1

Current number of TX queues: 1

Max possible TX queues: 230

Max possible number of TXDs per queue: 4096

Min possible number of TXDs per queue: 16

TXDs number alignment: 1

 

l  VF configuration Info


$ ip link show dev ens4f0

4: ens4f0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
mode DEFAULT group default qlen 1000

    link/ether f4:03:43:ca:f8:d0 brd ff:ff:ff:ff:ff:ff

    vf 0 MAC 00:11:22:33:44:55, spoof checking off, link-state enable,
trust off

 

$ ip link show dev ens4f1d1

5: ens4f1d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
mode DEFAULT group default qlen 1000

    link/ether f4:03:43:ca:f8:d8 brd ff:ff:ff:ff:ff:ff

    vf 0 MAC 00:11:22:33:44:77, spoof checking off, link-state enable,
trust off

 

Thank for reading

Best regard.

 


^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2020-01-31 16:17 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-01-31  1:58 [dpdk-dev] Issue while running testpmd with sr-iov setting 김영득
2020-01-31  2:23 김영득

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).