From patchwork Mon Dec 14 08:19:30 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Nithin Dabilpuram X-Patchwork-Id: 85136 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 5B0DBA09E9; Mon, 14 Dec 2020 09:19:50 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id EF51072D9; Mon, 14 Dec 2020 09:19:47 +0100 (CET) Received: from mx0b-0016f401.pphosted.com (mx0b-0016f401.pphosted.com [67.231.156.173]) by dpdk.org (Postfix) with ESMTP id 59FB129AC for ; Mon, 14 Dec 2020 09:19:45 +0100 (CET) Received: from pps.filterd (m0045851.ppops.net [127.0.0.1]) by mx0b-0016f401.pphosted.com (8.16.0.43/8.16.0.43) with SMTP id 0BE8GRmT012565; Mon, 14 Dec 2020 00:19:43 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h=from : to : cc : subject : date : message-id : in-reply-to : references : mime-version : content-type; s=pfpt0220; bh=ltVw5zX4Hs3AAcOJYXBOLzEScbrL6BcWMouBf0H75EY=; b=PelrIZq1mYzlku0yBLeIsovok65khvBbBpLvThW1ISOHcpOzBV5XCdSIqMnEh24LHYsE TGBF0xAiKFWD9MGLTrAUSS+dCMhxYUicPAcc50Atkri/qX8htjwPBxaTDY7El4/0nwlq OLg762UzaSxI52cVFXFL4F0FdaL92BtHxnBNj3ZqXtXuYS09jzAlPW6a4hHll9C+O5vH 1Hf77T9uDMKKtBNowc+ZJixIJG7z91Y9ml0AaI/1AV0ozUOy+/VNpmaA9isiTkUtqDgP 35W2W2kJX/gixVwksWdwPpy16h7XQYz7YZbOVanopMp4J4ohnshaKM2rDfv2J6x2T5T2 jg== Received: from sc-exch03.marvell.com ([199.233.58.183]) by mx0b-0016f401.pphosted.com with ESMTP id 35cx8t3d23-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-SHA384 bits=256 verify=NOT); Mon, 14 Dec 2020 00:19:43 -0800 Received: from DC5-EXCH01.marvell.com (10.69.176.38) by SC-EXCH03.marvell.com (10.93.176.83) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Mon, 14 Dec 2020 00:19:41 -0800 Received: from maili.marvell.com (10.69.176.80) by DC5-EXCH01.marvell.com (10.69.176.38) with Microsoft SMTP Server id 15.0.1497.2 via Frontend Transport; Mon, 14 Dec 2020 00:19:42 -0800 Received: from hyd1588t430.marvell.com (unknown [10.29.52.204]) by maili.marvell.com (Postfix) with ESMTP id 0E4883F703F; Mon, 14 Dec 2020 00:19:39 -0800 (PST) From: Nithin Dabilpuram To: , David Christensen , CC: , , Nithin Dabilpuram Date: Mon, 14 Dec 2020 13:49:30 +0530 Message-ID: <20201214081935.23577-1-ndabilpuram@marvell.com> X-Mailer: git-send-email 2.8.4 In-Reply-To: <20201012081106.10610-1-ndabilpuram@marvell.com> References: <20201012081106.10610-1-ndabilpuram@marvell.com> MIME-Version: 1.0 X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10434:6.0.343, 18.0.737 definitions=2020-12-14_03:2020-12-11, 2020-12-14 signatures=0 Subject: [dpdk-dev] [PATCH v5 0/4] fix issue with partial DMA unmap X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Partial DMA unmap is not supported by VFIO type1 IOMMU in Linux. Though the return value is zero, the returned DMA unmap size is not same as expected size. So add test case and fix to both heap triggered DMA mapping and user triggered DMA mapping/unmapping. Refer vfio_dma_do_unmap() in drivers/vfio/vfio_iommu_type1.c Snippet of comment is below. /* * vfio-iommu-type1 (v1) - User mappings were coalesced together to * avoid tracking individual mappings. This means that the granularity * of the original mapping was lost and the user was allowed to attempt * to unmap any range. Depending on the contiguousness of physical * memory and page sizes supported by the IOMMU, arbitrary unmaps may * or may not have worked. We only guaranteed unmap granularity * matching the original mapping; even though it was untracked here, * the original mappings are reflected in IOMMU mappings. This * resulted in a couple unusual behaviors. First, if a range is not * able to be unmapped, ex. a set of 4k pages that was mapped as a * 2M hugepage into the IOMMU, the unmap ioctl returns success but with * a zero sized unmap. Also, if an unmap request overlaps the first * address of a hugepage, the IOMMU will unmap the entire hugepage. * This also returns success and the returned unmap size reflects the * actual size unmapped. * We attempt to maintain compatibility with this "v1" interface, but * we take control out of the hands of the IOMMU. Therefore, an unmap * request offset from the beginning of the original mapping will * return success with zero sized unmap. And an unmap request covering * the first iova of mapping will unmap the entire range. This behavior can be verified by using first patch and add return check for dma_unmap.size != len in vfio_type1_dma_mem_map() v5: - Changed vfio test in test_vfio.c to use system pages allocated from heap instead of mmap() so that it comes in range of initially configured window for POWER9 System. - Added acked-by from David for 1/4, 2/4. v4: - Fixed issue with patch 4/4 on x86 builds. v3: - Fixed external memory test case(4/4) to use system page size instead of 4K. - Fixed check-git-log.sh issue and rebased. - Added acked-by from anatoly.burakov@intel.com to first 3 patches. v2: - Reverted earlier commit that enables mergin contiguous mapping for IOVA as PA. (see 1/3) - Updated documentation about kernel dma mapping limits and vfio module parameter. - Moved vfio test to test_vfio.c and handled comments from Anatoly. Nithin Dabilpuram (4): vfio: revert changes for map contiguous areas in one go vfio: fix DMA mapping granularity for type1 IOVA as VA test: add test case to validate VFIO DMA map/unmap test: change external memory test to use system page sz app/test/meson.build | 1 + app/test/test_external_mem.c | 3 +- app/test/test_vfio.c | 106 +++++++++++++++++++++++++++++++++ doc/guides/linux_gsg/linux_drivers.rst | 10 ++++ lib/librte_eal/linux/eal_vfio.c | 93 +++++++++++------------------ lib/librte_eal/linux/eal_vfio.h | 1 + 6 files changed, 156 insertions(+), 58 deletions(-) create mode 100644 app/test/test_vfio.c