From patchwork Thu Mar 1 19:34:59 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Varghese, Vipin" X-Patchwork-Id: 35564 X-Patchwork-Delegate: jerinj@marvell.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id A5FB04CA1; Thu, 1 Mar 2018 14:53:23 +0100 (CET) Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by dpdk.org (Postfix) with ESMTP id 8DE384C7B for ; Thu, 1 Mar 2018 14:53:22 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga003.fm.intel.com ([10.253.24.29]) by fmsmga102.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 01 Mar 2018 05:53:21 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.47,408,1515484800"; d="scan'208";a="30696229" Received: from unknown (HELO saesrv02-S2600CWR.intel.com) ([10.224.122.203]) by FMSMGA003.fm.intel.com with ESMTP; 01 Mar 2018 05:53:20 -0800 From: Vipin Varghese To: dev@dpdk.org, harry.van.haaren@intel.com Cc: Vipin Varghese Date: Fri, 2 Mar 2018 01:04:59 +0530 Message-Id: <1519932900-10571-1-git-send-email-vipin.varghese@intel.com> X-Mailer: git-send-email 2.7.4 Subject: [dpdk-dev] [PATCH 1/2] event/sw: code refractor to reduce the fetch stall X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" With rearranging the code to prefetch the contents before loop check increases performance from single and multistage atomic pipeline. Signed-off-by: Vipin Varghese Acked-by: Harry van Haaren id; iq_dequeue_burst(sw, &qid->iq[iq_num], qes, count); - for (i = 0; i < count; i++) { - const struct rte_event *qe = &qes[i]; - const uint16_t flow_id = SW_HASH_FLOWID(qes[i].flow_id); - struct sw_fid_t *fid = &qid->fids[flow_id]; - int cq = fid->cq; + const struct rte_event *qe = &qes[0]; + const uint16_t flow_id = SW_HASH_FLOWID(qes[0].flow_id); + struct sw_fid_t *fid = &qid->fids[flow_id]; + int cq = fid->cq; + + for (i = 0; i < count; i++) { if (cq < 0) { uint32_t cq_idx = qid->cq_next_tx++; if (qid->cq_next_tx == qid->cq_num_mapped_cqs) @@ -101,6 +102,13 @@ sw_schedule_atomic_to_cq(struct sw_evdev *sw, struct sw_qid * const qid, &sw->cq_ring_space[cq]); p->cq_buf_count = 0; } + + if (likely(i+1 < count)) { + qe = (qes + i + 1); + flow_id = SW_HASH_FLOWID(qes[i + 1].flow_id); + fid = &qid->fids[flow_id]; + cq = fid->cq; + } } iq_put_back(sw, &qid->iq[iq_num], blocked_qes, nb_blocked);