1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | /* |
3 | * Copyright (c) 2014-2016, Intel Corporation. |
4 | */ |
5 | #include "test/nfit_test.h" |
6 | #include <linux/blkdev.h> |
7 | #include <linux/dax.h> |
8 | #include <pmem.h> |
9 | #include <nd.h> |
10 | |
11 | long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, |
12 | long nr_pages, enum dax_access_mode mode, void **kaddr, |
13 | pfn_t *pfn) |
14 | { |
15 | resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; |
16 | |
17 | if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, |
18 | PFN_PHYS(nr_pages)))) |
19 | return -EIO; |
20 | |
21 | /* |
22 | * Limit dax to a single page at a time given vmalloc()-backed |
23 | * in the nfit_test case. |
24 | */ |
25 | if (get_nfit_res(resource: pmem->phys_addr + offset)) { |
26 | struct page *page; |
27 | |
28 | if (kaddr) |
29 | *kaddr = pmem->virt_addr + offset; |
30 | page = vmalloc_to_page(addr: pmem->virt_addr + offset); |
31 | if (pfn) |
32 | *pfn = page_to_pfn_t(page); |
33 | pr_debug_ratelimited("%s: pmem: %p pgoff: %#lx pfn: %#lx\n" , |
34 | __func__, pmem, pgoff, page_to_pfn(page)); |
35 | |
36 | return 1; |
37 | } |
38 | |
39 | if (kaddr) |
40 | *kaddr = pmem->virt_addr + offset; |
41 | if (pfn) |
42 | *pfn = phys_to_pfn_t(pmem->phys_addr + offset, pmem->pfn_flags); |
43 | |
44 | /* |
45 | * If badblocks are present, limit known good range to the |
46 | * requested range. |
47 | */ |
48 | if (unlikely(pmem->bb.count)) |
49 | return nr_pages; |
50 | return PHYS_PFN(pmem->size - pmem->pfn_pad - offset); |
51 | } |
52 | |