test suite reviews and discussions
 help / color / mirror / Atom feed
From: "Tu, Lijuan" <lijuan.tu@intel.com>
To: "Mo, YufengX" <yufengx.mo@intel.com>,
	"Zeng, XiaoxiaoX" <xiaoxiaox.zeng@intel.com>,
	"dts@dpdk.org" <dts@dpdk.org>
Cc: "Zeng, XiaoxiaoX" <xiaoxiaox.zeng@intel.com>
Subject: Re: [dts] [PATCH V1] tests/TestSuite_multiprocess: add	case	test_perf_multiprocess_performance
Date: Wed, 4 Sep 2019 05:34:43 +0000	[thread overview]
Message-ID: <8CE3E05A3F976642AAB0F4675D0AD20E0BB1C0CA@SHSMSX101.ccr.corp.intel.com> (raw)
In-Reply-To: <B8B15C44A3F2C044966E545C7B7371153E841DB4@shsmsx102.ccr.corp.intel.com>

Thanks yufeng's review

> -----Original Message-----
> From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of Mo, YufengX
> Sent: Monday, September 2, 2019 10:49 AM
> To: Zeng, XiaoxiaoX <xiaoxiaox.zeng@intel.com>; dts@dpdk.org
> Cc: Zeng, XiaoxiaoX <xiaoxiaox.zeng@intel.com>
> Subject: Re: [dts] [PATCH V1] tests/TestSuite_multiprocess: add case
> test_perf_multiprocess_performance
> 
> Hi, zeng,xiaoxiao
> 
> some questions about your test case `test_perf_multiprocess_performance`
> *. self.dut.kill_all() is used to kill dpdk revelant process, your test case hasn't
> run any dpdk process before it.
>   What's your target?
> *. You use self.dut.new_session() to create new session in a for loop, but you
> haven't close these sessions. And this suite
>   has created one `session_secondary` session in set_up_all method, why
> you want to create new one?
> *. When you run symmetric_mp, you haven't run server process, is it the test
> plan required?
> *. After you boot up first symmetric_mp, you haven't close it and run other
> one, is it the test plan required?
> *. Could you update this test case's test plan at the same time?
> 
> 
> > -----Original Message-----
> > From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of Xiaoxiao Zeng
> > Sent: Saturday, August 31, 2019 4:42 AM
> > To: dts@dpdk.org
> > Cc: Zeng, XiaoxiaoX <xiaoxiaox.zeng@intel.com>
> > Subject: [dts] [PATCH V1] tests/TestSuite_multiprocess: add case
> > test_perf_multiprocess_performance
> >
> > add testcase test_perf_multiprocess_performance
> > *.according to testplan, add test_perf_multiprocess_performance code
> >
> > Signed-off-by: Xiaoxiao Zeng <xiaoxiaox.zeng@intel.com>
> > ---
> >  tests/TestSuite_multiprocess.py | 127
> > ++++++++++++++++++++++++++------
> >  1 file changed, 103 insertions(+), 24 deletions(-)
> >
> > diff --git a/tests/TestSuite_multiprocess.py
> > b/tests/TestSuite_multiprocess.py index b690d81..4a4743e 100644
> > --- a/tests/TestSuite_multiprocess.py
> > +++ b/tests/TestSuite_multiprocess.py
> > @@ -37,6 +37,7 @@ Multi-process Test.
> >  import utils
> >  import time
> >  import os
> > +
> >  executions = []
> >  from test_case import TestCase
> >  from pktgen import PacketGeneratorHelper @@ -68,16 +69,17 @@ class
> > TestMultiprocess(TestCase):
> >          executions.append({'nprocs': 4, 'cores': '1S/2C/2T', 'pps': 0})
> >          executions.append({'nprocs': 4, 'cores': '1S/4C/1T', 'pps': 0})
> >          executions.append({'nprocs': 8, 'cores': '1S/4C/2T', 'pps': 0})
> > -        self.dut.alt_session.send_expect("cd dpdk","# ",5)
> > +        self.dut.alt_session.send_expect("cd dpdk", "# ", 5)
> >
> >          # start new session to run secondary
> >          self.session_secondary = self.dut.new_session()
> > +
> >          # get dts output path
> >          if self.logger.log_path.startswith(os.sep):
> >              self.output_path = self.logger.log_path
> >          else:
> >              cur_path = os.path.dirname(
> > -                                os.path.dirname(os.path.realpath(__file__)))
> > +                os.path.dirname(os.path.realpath(__file__)))
> >              self.output_path = os.sep.join([cur_path, self.logger.log_path])
> >          # create an instance to set stream field setting
> >          self.pktgen_helper = PacketGeneratorHelper() @@ -95,23 +97,28
> > @@ class TestMultiprocess(TestCase):
> >          # Send message from secondary to primary
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp -
> n 1 -c %s --proc-type=primary" % (self.target,
> > coremask), "Finished Process Init", 100)
> > +
> > +
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_m
> > + p -n 1 -c %s --proc-type=primary" % (self.target,
> > coremask),
> > +                             "Finished Process Init", 100)
> >          time.sleep(20)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> > -
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s --proc-type=secondary" %
> > (self.target, coremask), "Finished Process Init", 100)
> > +        self.session_secondary.send_expect(
> > +            "./examples/multi_process/simple_mp/%s/simple_mp -n 1 -c
> > + %s --proc-type=secondary" % (self.target, coremask), "Finished
> > Process Init",
> > +            100)
> >
> >          self.session_secondary.send_expect("send hello_primary", ">")
> >          out = self.dut.get_session_output()
> >          self.session_secondary.send_expect("quit", "# ")
> > -        self.dut.send_expect("quit","# ")
> > +        self.dut.send_expect("quit", "# ")
> >          self.verify("Received 'hello_primary'" in out, "Message not received
> on primary process")
> >          # Send message from primary to secondary
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s --proc-type=primary " %
> > (self.target, coremask), "Finished Process Init", 100)
> > +        self.session_secondary.send_expect(
> > +            "./examples/multi_process/simple_mp/%s/simple_mp -n 1 -c
> > + %s --proc-type=primary " % (self.target, coremask), "Finished
> > Process Init", 100)
> >          time.sleep(20)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> > -
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp -
> n 1 -c %s --proc-type=secondary" % (self.target,
> > coremask), "Finished Process Init", 100)
> > +
> > +
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_m
> > + p -n 1 -c %s --proc-type=secondary" % (self.target,
> > coremask),
> > +                             "Finished Process Init", 100)
> >          self.session_secondary.send_expect("send hello_secondary", ">")
> >          out = self.dut.get_session_output()
> >          self.session_secondary.send_expect("quit", "# ") @@ -127,10
> > +134,12 @@ class TestMultiprocess(TestCase):
> >
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s --proc-type=primary" %
> > (self.target, coremask), "Finished Process Init", 100)
> > +
> > + self.session_secondary.send_expect("./examples/multi_process/simple_
> > + mp/%s/simple_mp -n 1 -c %s --proc-type=primary" %
> > (self.target, coremask),
> > +                                           "Finished Process Init",
> > + 100)
> >          time.sleep(20)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> > -
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp -
> n 1 -c %s --proc-type=secondary" % (self.target,
> > coremask), "Finished Process Init", 100)
> > +
> > +
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_m
> > + p -n 1 -c %s --proc-type=secondary" % (self.target,
> > coremask),
> > +                             "Finished Process Init", 100)
> >          stringsSent = 0
> >          for line in open('/usr/share/dict/words', 'r').readlines():
> >              line = line.split('\n')[0] @@ -151,11 +160,13 @@ class
> > TestMultiprocess(TestCase):
> >          # Send message from secondary to primary (auto process type)
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -        out =
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp -
> n 1 -c %s --proc-type=auto " % (self.target,
> > coremask), "Finished Process Init", 100)
> > +        out =
> > +
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_m
> > + p -n 1 -c %s --proc-type=auto " % (self.target,
> > coremask),
> > +                                   "Finished Process Init", 100)
> >          self.verify("EAL: Auto-detected process type: PRIMARY" in out, "The
> type of process (PRIMARY) was not detected properly")
> >          time.sleep(20)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> > -        out =
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s --proc-type=auto" %
> > (self.target, coremask), "Finished Process Init", 100)
> > +        out = self.session_secondary.send_expect(
> > +            "./examples/multi_process/simple_mp/%s/simple_mp -n 1 -c
> > + %s --proc-type=auto" % (self.target, coremask), "Finished Process
> > Init", 100)
> >          self.verify("EAL: Auto-detected process type: SECONDARY" in out,
> >                      "The type of process (SECONDARY) was not detected
> > properly")
> >
> > @@ -168,13 +179,15 @@ class TestMultiprocess(TestCase):
> >          # Send message from primary to secondary (auto process type)
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -        out =
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s --proc-type=auto" %
> > (self.target, coremask), "Finished Process Init", 100)
> > +        out = self.session_secondary.send_expect(
> > +            "./examples/multi_process/simple_mp/%s/simple_mp -n 1 -c
> > + %s --proc-type=auto" % (self.target, coremask), "Finished Process
> > Init", 100)
> >          self.verify("EAL: Auto-detected process type: PRIMARY" in out, "The
> type of process (PRIMARY) was not detected properly")
> >          time.sleep(20)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> > -        out =
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp -
> n 1 -c %s --proc-type=auto" % (self.target,
> > coremask), "Finished Process Init", 100)
> > +        out =
> > +
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_m
> > + p -n 1 -c %s --proc-type=auto" % (self.target,
> > coremask),
> > +                                   "Finished Process Init", 100)
> >          self.verify("EAL: Auto-detected process type: SECONDARY" in out,
> "The type of process (SECONDARY) was not detected properly")
> > -        self.session_secondary.send_expect("send hello_secondary", ">",100)
> > +        self.session_secondary.send_expect("send hello_secondary",
> > + ">", 100)
> >          out = self.dut.get_session_output()
> >          self.session_secondary.send_expect("quit", "# ")
> >          self.dut.send_expect("quit", "# ") @@ -189,7 +202,8 @@ class
> > TestMultiprocess(TestCase):
> >
> >          cores = self.dut.get_core_list('1S/2C/1T')
> >          coremask = utils.create_mask(cores)
> > -
> self.session_secondary.send_expect("./examples/multi_process/simple_mp/
> %s/simple_mp -n 1 -c %s -m 64" % (self.target,
> > coremask), "Finished Process Init", 100)
> > +
> > + self.session_secondary.send_expect("./examples/multi_process/simple_
> > + mp/%s/simple_mp -n 1 -c %s -m 64" % (self.target,
> > coremask),
> > +                                           "Finished Process Init",
> > + 100)
> >          coremask = hex(int(coremask, 16) * 0x10).rstrip("L")
> >          out =
> >
> self.dut.send_expect("./examples/multi_process/simple_mp/%s/simple_mp
> > -n 1 -c %s" % (self.target, coremask), "# ", 100)
> >
> > @@ -198,6 +212,66 @@ class TestMultiprocess(TestCase):
> >
> >          self.session_secondary.send_expect("quit", "# ")
> >
> > +    def test_perf_multiprocess_performance(self):
> > +        """
> > +        Benchmark Multiprocess performance.
> > +        """
> > +        packet_count = 16
> > +        self.dut.kill_all()
> > +        self.dut.send_expect("fg", "# ")
> > +        dutPorts = self.dut.get_ports()
> > +        txPort = self.tester.get_local_port(dutPorts[0])
> > +        rxPort = self.tester.get_local_port(dutPorts[1])
> > +        mac = self.tester.get_mac(txPort)
> > +        dmac = self.dut.get_mac_address(dutPorts[0])
> > +        tgenInput = []
> > +
> > +        for i in range(packet_count):
> > +            package = r'flows = [Ether(src="%s",
> dst="%s")/IP(src="192.168.1.%d", dst="192.168.1.%d")/("X"*26)]' % (mac,
> dmac, i + 1, i + 2)
> > +            self.tester.scapy_append(package)
> > +            pcap = os.sep.join([self.output_path, "test_%d.pcap"%i])
> > +            self.tester.scapy_append('wrpcap("%s", flows)' % pcap)
> > +            tgenInput.append([txPort, rxPort, pcap])
> > +        self.tester.scapy_execute()
> > +
> > +        validExecutions = []
> > +        for execution in executions:
> > +            if len(self.dut.get_core_list(execution['cores'])) ==
> execution['nprocs']:
> > +                validExecutions.append(execution)
> > +
> > +        portMask = utils.create_mask([dutPorts[0], dutPorts[1]])
> > +
> > +        for n in range(len(validExecutions)):
> > +            execution = validExecutions[n]
> > +            coreList = self.dut.get_core_list(execution['cores'])
> > +            coreMask =
> utils.create_mask(self.dut.get_core_list(execution['cores']))
> > +            time.sleep(5)
> > +            self.dut_session = self.dut.new_session()
> > +            self.dut_session.send_expect(
> > +
> > + "./examples/multi_process/symmetric_mp/%s/symmetric_mp -c %s
> > + --proc-type=auto -- -p %s --num-procs=%d --proc-id=%d" %
> > (
> > +                    self.target, coreMask, portMask, execution['nprocs'], n),
> "Finished Process Init")
> > +            # clear streams before add new streams
> > +            self.tester.pktgen.clear_streams()
> > +            # run packet generator
> > +            streams =
> self.pktgen_helper.prepare_stream_from_tginput(tgenInput, 100,
> > +                                                                     None, self.tester.pktgen)
> > +            _, pps =
> > + self.tester.pktgen.measure_throughput(stream_ids=streams)
> > +
> > +            execution['pps'] = pps
> > +            #self.dut.kill_all()
> > +            time.sleep(5)
> > +
> > +        for n in range(len(executions)):
> > +            self.verify(executions[n]['pps'] is not 0, "No traffic
> > + detected")
> > +
> > +        self.result_table_create(['Num-procs',
> > + 'Sockets/Cores/Threads', 'Num Ports', 'Frame Size', '%-age Line
> > + Rate', 'Packet Rate(mpps)'])
> > +
> > +        for execution in validExecutions:
> > +            self.result_table_add(
> > +                [execution['nprocs'], execution['cores'], 2, 64,
> > + execution['pps'] / float(100000000 / (8 * 84)), execution['pps'] /
> > + float(1000000)])
> > +
> > +        self.result_table_print()
> > +
> >      def test_perf_multiprocess_client_serverperformance(self):
> >          """
> >          Benchmark Multiprocess client-server performance.
> > @@ -227,14 +301,16 @@ class TestMultiprocess(TestCase):
> >
> >              coreMask = utils.create_mask(self.dut.get_core_list('1S/1C/1T'))
> >              portMask = utils.create_mask([dutPorts[0], dutPorts[1]])
> > -
> self.dut.send_expect("./examples/multi_process/client_server_mp/mp_serve
> r/%s/mp_server -n %d -c %s -- -p %s -n %d" %
> > (self.target, self.dut.get_memory_channels(), "0xA0", portMask,
> > execution['nprocs']), "Finished Process Init", 20)
> > +
> self.dut.send_expect("./examples/multi_process/client_server_mp/mp_serve
> r/%s/mp_server -n %d -c %s -- -p %s -n %d" % (
> > +                self.target, self.dut.get_memory_channels(), "0xA0",
> > + portMask, execution['nprocs']), "Finished Process Init", 20)
> >              self.dut.send_expect("^Z", "\r\n")
> >              self.dut.send_expect("bg", "# ")
> >
> >              for n in range(execution['nprocs']):
> >                  time.sleep(5)
> >                  coreMask = utils.create_mask([coreList[n]])
> > -
> self.dut.send_expect("./examples/multi_process/client_server_mp/mp_clien
> t/%s/mp_client -n %d -c %s --proc-type=secondary
> > -- -n %d" % (self.target, self.dut.get_memory_channels(), coreMask,
> > n), "Finished Process Init")
> > +
> > + self.dut.send_expect("./examples/multi_process/client_server_mp/mp_c
> > + lient/%s/mp_client -n %d -c %s --proc-type=secondary
> > -- -n %d" % (
> > +                    self.target, self.dut.get_memory_channels(),
> > + coreMask, n), "Finished Process Init")
> >                  self.dut.send_expect("^Z", "\r\n")
> >                  self.dut.send_expect("bg", "# ")
> >
> > @@ -245,7 +321,7 @@ class TestMultiprocess(TestCase):
> >              self.tester.pktgen.clear_streams()
> >              # run packet generator
> >              streams =
> self.pktgen_helper.prepare_stream_from_tginput(tgenInput, 100,
> > -                                                None, self.tester.pktgen)
> > +
> > + None, self.tester.pktgen)
> >              _, pps =
> > self.tester.pktgen.measure_throughput(stream_ids=streams)
> >
> >              execution['pps'] = pps
> > @@ -255,20 +331,23 @@ class TestMultiprocess(TestCase):
> >          for n in range(len(executions)):
> >              self.verify(executions[n]['pps'] is not 0, "No traffic
> > detected")
> >
> > -        self.result_table_create(['Server threads', 'Server Cores/Threads',
> 'Num-procs', 'Sockets/Cores/Threads', 'Num Ports', 'Frame Size',
> > '%-age Line Rate', 'Packet Rate(mpps)'])
> > +        self.result_table_create(
> > +            ['Server threads', 'Server Cores/Threads', 'Num-procs',
> 'Sockets/Cores/Threads', 'Num Ports', 'Frame Size', '%-age Line Rate',
> > +             'Packet Rate(mpps)'])
> >
> >          for execution in validExecutions:
> > -            self.result_table_add([1, '1S/1C/1T', execution['nprocs'],
> execution['cores'], 2, 64, execution['pps'] / float(100000000 / (8 * 84)),
> > execution['pps'] / float(1000000)])
> > +            self.result_table_add([1, '1S/1C/1T', execution['nprocs'],
> execution['cores'], 2, 64, execution['pps'] / float(100000000 / (8 * 84)),
> > +                                   execution['pps'] /
> > + float(1000000)])
> >
> >          self.result_table_print()
> >
> >      def set_fields(self):
> >          ''' set ip protocol field behavior '''
> >          fields_config = {
> > -        'ip':  {
> > -            'src': {'range': 64, 'action': 'inc'},
> > -            'dst': {'range': 64, 'action': 'inc'},
> > -        },}
> > +            'ip': {
> > +                'src': {'range': 64, 'action': 'inc'},
> > +                'dst': {'range': 64, 'action': 'inc'},
> > +            }, }
> >
> >          return fields_config
> >
> > --
> > 2.17.0


      reply	other threads:[~2019-09-04  5:34 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-08-30 20:42 Xiaoxiao Zeng
2019-09-02  2:49 ` Mo, YufengX
2019-09-04  5:34   ` Tu, Lijuan [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=8CE3E05A3F976642AAB0F4675D0AD20E0BB1C0CA@SHSMSX101.ccr.corp.intel.com \
    --to=lijuan.tu@intel.com \
    --cc=dts@dpdk.org \
    --cc=xiaoxiaox.zeng@intel.com \
    --cc=yufengx.mo@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).