From patchwork Thu Dec 1 06:46:51 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ganapati Kundapura X-Patchwork-Id: 120407 X-Patchwork-Delegate: jerinj@marvell.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 66ACCA00C5; Thu, 1 Dec 2022 07:47:14 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 1C6C242D27; Thu, 1 Dec 2022 07:47:01 +0100 (CET) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by mails.dpdk.org (Postfix) with ESMTP id D331142D0D for ; Thu, 1 Dec 2022 07:46:58 +0100 (CET) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1669877219; x=1701413219; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=Jk6rm1nQVoOnUGdSFMm3gJmVvl1oadw0zsv0ITSH3DA=; b=HR2AbtmnC8OnOh3WncrQL+WOLkVKafzFF2sx9PESLmXc3gdrQ69CRIWb fi2PWD5BvWaD8iHHIQlTx3A1XlEBYN+lMxhmYU0GxM09JStwZl2ClMM9L C2lg8v1ws4q3K69E/vX6gWl89ZmE1Je2IqmbQF+ZmLSZKxg9KqrSCI85a hy4b4gCfS/u1xnPRLtaMxwUKHodn7QuKnL2g53AsDKMZz60UqlHVPsOcG vZsX63//cQkKnfdNaRhyo8yInug/OBHvG1rz78UzUChxJBCQbIdgnV6Ig YLnZnF3g0DVLrbDmc4V+xx2W8/PRJpa7eYxLW6OxLmGw2FG1VZUIuGi3h A==; X-IronPort-AV: E=McAfee;i="6500,9779,10547"; a="314304305" X-IronPort-AV: E=Sophos;i="5.96,207,1665471600"; d="scan'208";a="314304305" Received: from orsmga003.jf.intel.com ([10.7.209.27]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 30 Nov 2022 22:46:57 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10547"; a="594950435" X-IronPort-AV: E=Sophos;i="5.96,207,1665471600"; d="scan'208";a="594950435" Received: from txandevlnx322.an.intel.com ([10.123.117.44]) by orsmga003.jf.intel.com with ESMTP; 30 Nov 2022 22:46:56 -0800 From: Ganapati Kundapura To: dev@dpdk.org, jerinj@marvell.com, s.v.naga.harish.k@intel.com, abhinandan.gujjar@intel.com Cc: jay.jayatheerthan@intel.com Subject: [PATCH v2 4/5] eventdev/crypto: fix overflow in circular buffer Date: Thu, 1 Dec 2022 00:46:51 -0600 Message-Id: <20221201064652.1885734-4-ganapati.kundapura@intel.com> X-Mailer: git-send-email 2.23.0 In-Reply-To: <20221201064652.1885734-1-ganapati.kundapura@intel.com> References: <20221130171014.1723899-1-ganapati.kundapura@intel.com> <20221201064652.1885734-1-ganapati.kundapura@intel.com> MIME-Version: 1.0 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Crypto adapter checks CPM backpressure once in enq_run() This leads to buffer overflow if some ops failed to flush to cryptodev. Checked CPM backpressure for every iteration in enq_run() Fixes: 7901eac3409a ("eventdev: add crypto adapter implementation") Signed-off-by: Ganapati Kundapura Acked-by: Volodymyr Fialko --- v2: * Updated subject line in commit message diff --git a/lib/eventdev/rte_event_crypto_adapter.c b/lib/eventdev/rte_event_crypto_adapter.c index 72deedd..1d39c5b 100644 --- a/lib/eventdev/rte_event_crypto_adapter.c +++ b/lib/eventdev/rte_event_crypto_adapter.c @@ -573,14 +573,15 @@ eca_crypto_adapter_enq_run(struct event_crypto_adapter *adapter, if (adapter->mode == RTE_EVENT_CRYPTO_ADAPTER_OP_NEW) return 0; - if (unlikely(adapter->stop_enq_to_cryptodev)) { - nb_enqueued += eca_crypto_enq_flush(adapter); + for (nb_enq = 0; nb_enq < max_enq; nb_enq += n) { - if (unlikely(adapter->stop_enq_to_cryptodev)) - goto skip_event_dequeue_burst; - } + if (unlikely(adapter->stop_enq_to_cryptodev)) { + nb_enqueued += eca_crypto_enq_flush(adapter); + + if (unlikely(adapter->stop_enq_to_cryptodev)) + break; + } - for (nb_enq = 0; nb_enq < max_enq; nb_enq += n) { stats->event_poll_count++; n = rte_event_dequeue_burst(event_dev_id, event_port_id, ev, BATCH_SIZE, 0); @@ -591,8 +592,6 @@ eca_crypto_adapter_enq_run(struct event_crypto_adapter *adapter, nb_enqueued += eca_enq_to_cryptodev(adapter, ev, n); } -skip_event_dequeue_burst: - if ((++adapter->transmit_loop_count & (CRYPTO_ENQ_FLUSH_THRESHOLD - 1)) == 0) { nb_enqueued += eca_crypto_enq_flush(adapter);