aboutsummaryrefslogtreecommitdiff
path: root/sys/dev
diff options
context:
space:
mode:
authorJohn Baldwin <jhb@FreeBSD.org>2026-04-23 17:05:54 +0000
committerJohn Baldwin <jhb@FreeBSD.org>2026-04-23 17:05:54 +0000
commite27501388fc0dc2a29f90cb24ba8d36e9bb6631f (patch)
treebdabef7c7e09ab92e77860922eaf7bd2b7df7840 /sys/dev
parentec3a0b2a02ee2aa459aa72d751f6670b7f813f31 (diff)
Diffstat (limited to 'sys/dev')
-rw-r--r--sys/dev/drm2/ttm/ttm_bo_util.c4
-rw-r--r--sys/dev/gve/gve_qpl.c12
-rw-r--r--sys/dev/hwt/hwt_vm.c4
-rw-r--r--sys/dev/md/md.c11
-rw-r--r--sys/dev/pci/controller/pci_n1sdp.c8
-rw-r--r--sys/dev/spibus/spigen.c12
6 files changed, 25 insertions, 26 deletions
diff --git a/sys/dev/drm2/ttm/ttm_bo_util.c b/sys/dev/drm2/ttm/ttm_bo_util.c
index 9d3f4103f82e..1734a8103cde 100644
--- a/sys/dev/drm2/ttm/ttm_bo_util.c
+++ b/sys/dev/drm2/ttm/ttm_bo_util.c
@@ -517,7 +517,7 @@ static int ttm_bo_kmap_ttm(struct ttm_buffer_object *bo,
pmap_page_set_memattr(ttm->pages[start_page +
i], prot);
}
- pmap_qenter((vm_offset_t)map->virtual,
+ pmap_qenter(map->virtual,
&ttm->pages[start_page], num_pages);
}
}
@@ -571,7 +571,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map)
pmap_unmapdev(map->virtual, map->size);
break;
case ttm_bo_map_vmap:
- pmap_qremove((vm_offset_t)(map->virtual), map->num_pages);
+ pmap_qremove(map->virtual, map->num_pages);
kva_free((vm_offset_t)map->virtual,
map->num_pages * PAGE_SIZE);
break;
diff --git a/sys/dev/gve/gve_qpl.c b/sys/dev/gve/gve_qpl.c
index 0e7098dcd4a1..1c6a58372365 100644
--- a/sys/dev/gve/gve_qpl.c
+++ b/sys/dev/gve/gve_qpl.c
@@ -46,7 +46,7 @@ gve_free_qpl(struct gve_priv *priv, struct gve_queue_page_list *qpl)
}
if (qpl->kva) {
- pmap_qremove(qpl->kva, qpl->num_pages);
+ pmap_qremove((void *)qpl->kva, qpl->num_pages);
kva_free(qpl->kva, PAGE_SIZE * qpl->num_pages);
}
@@ -59,7 +59,7 @@ gve_free_qpl(struct gve_priv *priv, struct gve_queue_page_list *qpl)
*/
if (vm_page_unwire_noq(qpl->pages[i])) {
if (!qpl->kva) {
- pmap_qremove((vm_offset_t)qpl->dmas[i].cpu_addr, 1);
+ pmap_qremove(qpl->dmas[i].cpu_addr, 1);
kva_free((vm_offset_t)qpl->dmas[i].cpu_addr, PAGE_SIZE);
}
vm_page_free(qpl->pages[i]);
@@ -126,7 +126,7 @@ gve_alloc_qpl(struct gve_priv *priv, uint32_t id, int npages, bool single_kva)
err = ENOMEM;
goto abort;
}
- pmap_qenter((vm_offset_t)qpl->dmas[i].cpu_addr, &(qpl->pages[i]), 1);
+ pmap_qenter(qpl->dmas[i].cpu_addr, &(qpl->pages[i]), 1);
} else
qpl->dmas[i].cpu_addr = (void *)(qpl->kva + (PAGE_SIZE * i));
@@ -135,7 +135,7 @@ gve_alloc_qpl(struct gve_priv *priv, uint32_t id, int npages, bool single_kva)
}
if (single_kva)
- pmap_qenter(qpl->kva, qpl->pages, npages);
+ pmap_qenter((void *)qpl->kva, qpl->pages, npages);
for (i = 0; i < npages; i++) {
err = gve_dmamap_create(priv, /*size=*/PAGE_SIZE, /*align=*/PAGE_SIZE,
@@ -244,7 +244,7 @@ void
gve_mextadd_free(struct mbuf *mbuf)
{
vm_page_t page = (vm_page_t)mbuf->m_ext.ext_arg1;
- vm_offset_t va = (vm_offset_t)mbuf->m_ext.ext_arg2;
+ void *va = mbuf->m_ext.ext_arg2;
/*
* Free the page only if this is the last ref.
@@ -253,7 +253,7 @@ gve_mextadd_free(struct mbuf *mbuf)
*/
if (__predict_false(vm_page_unwire_noq(page))) {
pmap_qremove(va, 1);
- kva_free(va, PAGE_SIZE);
+ kva_free((vm_offset_t)va, PAGE_SIZE);
vm_page_free(page);
}
}
diff --git a/sys/dev/hwt/hwt_vm.c b/sys/dev/hwt/hwt_vm.c
index 6c55e218dcec..a3e906d71099 100644
--- a/sys/dev/hwt/hwt_vm.c
+++ b/sys/dev/hwt/hwt_vm.c
@@ -171,7 +171,7 @@ retry:
VM_OBJECT_WLOCK(vm->obj);
vm_page_insert(m, vm->obj, i);
if (kva_req)
- pmap_qenter(vm->kvaddr + i * PAGE_SIZE, &m, 1);
+ pmap_qenter((char *)vm->kvaddr + i * PAGE_SIZE, &m, 1);
VM_OBJECT_WUNLOCK(vm->obj);
}
@@ -442,7 +442,7 @@ hwt_vm_destroy_buffers(struct hwt_vm *vm)
int i;
if (vm->ctx->hwt_backend->kva_req && vm->kvaddr != 0) {
- pmap_qremove(vm->kvaddr, vm->npages);
+ pmap_qremove((void *)vm->kvaddr, vm->npages);
kva_free(vm->kvaddr, vm->npages * PAGE_SIZE);
}
VM_OBJECT_WLOCK(vm->obj);
diff --git a/sys/dev/md/md.c b/sys/dev/md/md.c
index 8d2908264aac..e2c0451b5843 100644
--- a/sys/dev/md/md.c
+++ b/sys/dev/md/md.c
@@ -286,7 +286,7 @@ struct md_s {
struct {
struct vnode *vnode;
char file[PATH_MAX];
- vm_offset_t kva;
+ char *kva;
} s_vnode;
/* MD_SWAP related fields */
@@ -983,8 +983,7 @@ unmapped_step:
("npages %d too large", npages));
pmap_qenter(sc->s_vnode.kva, &bp->bio_ma[atop(ma_offs)],
npages);
- aiov.iov_base = (void *)(sc->s_vnode.kva + (ma_offs &
- PAGE_MASK));
+ aiov.iov_base = sc->s_vnode.kva + (ma_offs & PAGE_MASK);
aiov.iov_len = iolen;
auio.uio_iov = &aiov;
auio.uio_iovcnt = 1;
@@ -1513,7 +1512,7 @@ mdcreate_vnode(struct md_s *sc, struct md_req *mdr, struct thread *td)
goto bad;
}
- sc->s_vnode.kva = kva_alloc(maxphys + PAGE_SIZE);
+ sc->s_vnode.kva = (char *)kva_alloc(maxphys + PAGE_SIZE);
return (0);
bad:
VOP_UNLOCK(nd.ni_vp);
@@ -1567,8 +1566,8 @@ mddestroy(struct md_s *sc, struct thread *td)
sc->flags & MD_READONLY ? FREAD : (FREAD|FWRITE),
sc->cred, td);
}
- if (sc->s_vnode.kva != 0)
- kva_free(sc->s_vnode.kva, maxphys + PAGE_SIZE);
+ if (sc->s_vnode.kva != NULL)
+ kva_free((vm_offset_t)sc->s_vnode.kva, maxphys + PAGE_SIZE);
break;
case MD_SWAP:
if (sc->s_swap.object != NULL)
diff --git a/sys/dev/pci/controller/pci_n1sdp.c b/sys/dev/pci/controller/pci_n1sdp.c
index 22f0ea27d45b..c1f8624e45aa 100644
--- a/sys/dev/pci/controller/pci_n1sdp.c
+++ b/sys/dev/pci/controller/pci_n1sdp.c
@@ -83,7 +83,7 @@ static int
n1sdp_init(struct generic_pcie_n1sdp_softc *sc)
{
struct pcie_discovery_data *shared_data;
- vm_offset_t vaddr;
+ void *vaddr;
vm_paddr_t paddr_rc;
vm_paddr_t paddr;
vm_page_t m[BDF_TABLE_SIZE / PAGE_SIZE];
@@ -100,8 +100,8 @@ n1sdp_init(struct generic_pcie_n1sdp_softc *sc)
MPASS(m[i] != NULL);
}
- vaddr = kva_alloc((vm_size_t)BDF_TABLE_SIZE);
- if (vaddr == 0) {
+ vaddr = (void *)kva_alloc((vm_size_t)BDF_TABLE_SIZE);
+ if (vaddr == NULL) {
printf("%s: Can't allocate KVA memory.", __func__);
error = ENXIO;
goto out;
@@ -130,7 +130,7 @@ n1sdp_init(struct generic_pcie_n1sdp_softc *sc)
out_pmap:
pmap_qremove(vaddr, nitems(m));
- kva_free(vaddr, (vm_size_t)BDF_TABLE_SIZE);
+ kva_free((vm_offset_t)vaddr, (vm_size_t)BDF_TABLE_SIZE);
out:
vm_phys_fictitious_unreg_range(paddr, paddr + BDF_TABLE_SIZE);
diff --git a/sys/dev/spibus/spigen.c b/sys/dev/spibus/spigen.c
index 94f276d239f3..400ae1e139ad 100644
--- a/sys/dev/spibus/spigen.c
+++ b/sys/dev/spibus/spigen.c
@@ -73,7 +73,7 @@ struct spigen_softc {
struct spigen_mmap {
vm_object_t bufobj;
- vm_offset_t kvaddr;
+ void *kvaddr;
size_t bufsize;
};
@@ -235,10 +235,10 @@ spigen_transfer_mmapped(struct cdev *cdev, struct spigen_transfer_mmapped *stm)
if (mmap->bufsize < stm->stm_command_length + stm->stm_data_length)
return (E2BIG);
- transfer.tx_cmd = transfer.rx_cmd = (void *)((uintptr_t)mmap->kvaddr);
+ transfer.tx_cmd = transfer.rx_cmd = mmap->kvaddr;
transfer.tx_cmd_sz = transfer.rx_cmd_sz = stm->stm_command_length;
transfer.tx_data = transfer.rx_data =
- (void *)((uintptr_t)mmap->kvaddr + stm->stm_command_length);
+ (void *)((char *)mmap->kvaddr + stm->stm_command_length);
transfer.tx_data_sz = transfer.rx_data_sz = stm->stm_data_length;
error = SPIBUS_TRANSFER(device_get_parent(dev), dev, &transfer);
@@ -283,9 +283,9 @@ spigen_mmap_cleanup(void *arg)
{
struct spigen_mmap *mmap = arg;
- if (mmap->kvaddr != 0) {
+ if (mmap->kvaddr != NULL) {
pmap_qremove(mmap->kvaddr, mmap->bufsize / PAGE_SIZE);
- kva_free(mmap->kvaddr, mmap->bufsize);
+ kva_free((vm_offset_t)mmap->kvaddr, mmap->bufsize);
}
if (mmap->bufobj != NULL)
vm_object_deallocate(mmap->bufobj);
@@ -312,7 +312,7 @@ spigen_mmap_single(struct cdev *cdev, vm_ooffset_t *offset,
return (EBUSY);
mmap = malloc(sizeof(*mmap), M_DEVBUF, M_ZERO | M_WAITOK);
- if ((mmap->kvaddr = kva_alloc(size)) == 0) {
+ if ((mmap->kvaddr = (void *)kva_alloc(size)) == 0) {
spigen_mmap_cleanup(mmap);
return (ENOMEM);
}