summaryrefslogtreecommitdiff
path: root/kernel
diff options
context:
space:
mode:
authorBalbir Singh <balbirs@nvidia.com>2025-04-22 21:40:34 +1000
committerMarek Szyprowski <m.szyprowski@samsung.com>2025-04-22 15:06:33 +0200
commitcae5572ec9261f752af834cdaaf5a0ba0afcf256 (patch)
tree740864e5247d2d9cefe24b22f096bfbcf8e91b15 /kernel
parentc9b19ea63036fc537a69265acea1b18dabd1cbd3 (diff)
dma-mapping: Fix warning reported for missing prototype
lkp reported a warning about missing prototype for a recent patch. The kernel-doc style comments are out of sync, move them to the right function. Cc: Marek Szyprowski <m.szyprowski@samsung.com> Cc: Christoph Hellwig <hch@lst.de> Reported-by: kernel test robot <lkp@intel.com> Closes: https://lore.kernel.org/oe-kbuild-all/202504190615.g9fANxHw-lkp@intel.com/ Signed-off-by: Balbir Singh <balbirs@nvidia.com> [mszyprow: reformatted subject] Signed-off-by: Marek Szyprowski <m.szyprowski@samsung.com> Link: https://lore.kernel.org/r/20250422114034.3535515-1-balbirs@nvidia.com
Diffstat (limited to 'kernel')
-rw-r--r--kernel/dma/mapping.c16
1 files changed, 8 insertions, 8 deletions
diff --git a/kernel/dma/mapping.c b/kernel/dma/mapping.c
index 67da08fa6723..051a32988040 100644
--- a/kernel/dma/mapping.c
+++ b/kernel/dma/mapping.c
@@ -910,14 +910,6 @@ int dma_set_coherent_mask(struct device *dev, u64 mask)
}
EXPORT_SYMBOL(dma_set_coherent_mask);
-/**
- * dma_addressing_limited - return if the device is addressing limited
- * @dev: device to check
- *
- * Return %true if the devices DMA mask is too small to address all memory in
- * the system, else %false. Lack of addressing bits is the prime reason for
- * bounce buffering, but might not be the only one.
- */
static bool __dma_addressing_limited(struct device *dev)
{
const struct dma_map_ops *ops = get_dma_ops(dev);
@@ -931,6 +923,14 @@ static bool __dma_addressing_limited(struct device *dev)
return !dma_direct_all_ram_mapped(dev);
}
+/**
+ * dma_addressing_limited - return if the device is addressing limited
+ * @dev: device to check
+ *
+ * Return %true if the devices DMA mask is too small to address all memory in
+ * the system, else %false. Lack of addressing bits is the prime reason for
+ * bounce buffering, but might not be the only one.
+ */
bool dma_addressing_limited(struct device *dev)
{
if (!__dma_addressing_limited(dev))