diff options
| author | John Baldwin <jhb@FreeBSD.org> | 2026-04-23 17:05:54 +0000 |
|---|---|---|
| committer | John Baldwin <jhb@FreeBSD.org> | 2026-04-23 17:05:54 +0000 |
| commit | e27501388fc0dc2a29f90cb24ba8d36e9bb6631f (patch) | |
| tree | bdabef7c7e09ab92e77860922eaf7bd2b7df7840 /sys/dev | |
| parent | ec3a0b2a02ee2aa459aa72d751f6670b7f813f31 (diff) | |
Diffstat (limited to 'sys/dev')
| -rw-r--r-- | sys/dev/drm2/ttm/ttm_bo_util.c | 4 | ||||
| -rw-r--r-- | sys/dev/gve/gve_qpl.c | 12 | ||||
| -rw-r--r-- | sys/dev/hwt/hwt_vm.c | 4 | ||||
| -rw-r--r-- | sys/dev/md/md.c | 11 | ||||
| -rw-r--r-- | sys/dev/pci/controller/pci_n1sdp.c | 8 | ||||
| -rw-r--r-- | sys/dev/spibus/spigen.c | 12 |
6 files changed, 25 insertions, 26 deletions
diff --git a/sys/dev/drm2/ttm/ttm_bo_util.c b/sys/dev/drm2/ttm/ttm_bo_util.c index 9d3f4103f82e..1734a8103cde 100644 --- a/sys/dev/drm2/ttm/ttm_bo_util.c +++ b/sys/dev/drm2/ttm/ttm_bo_util.c @@ -517,7 +517,7 @@ static int ttm_bo_kmap_ttm(struct ttm_buffer_object *bo, pmap_page_set_memattr(ttm->pages[start_page + i], prot); } - pmap_qenter((vm_offset_t)map->virtual, + pmap_qenter(map->virtual, &ttm->pages[start_page], num_pages); } } @@ -571,7 +571,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map) pmap_unmapdev(map->virtual, map->size); break; case ttm_bo_map_vmap: - pmap_qremove((vm_offset_t)(map->virtual), map->num_pages); + pmap_qremove(map->virtual, map->num_pages); kva_free((vm_offset_t)map->virtual, map->num_pages * PAGE_SIZE); break; diff --git a/sys/dev/gve/gve_qpl.c b/sys/dev/gve/gve_qpl.c index 0e7098dcd4a1..1c6a58372365 100644 --- a/sys/dev/gve/gve_qpl.c +++ b/sys/dev/gve/gve_qpl.c @@ -46,7 +46,7 @@ gve_free_qpl(struct gve_priv *priv, struct gve_queue_page_list *qpl) } if (qpl->kva) { - pmap_qremove(qpl->kva, qpl->num_pages); + pmap_qremove((void *)qpl->kva, qpl->num_pages); kva_free(qpl->kva, PAGE_SIZE * qpl->num_pages); } @@ -59,7 +59,7 @@ gve_free_qpl(struct gve_priv *priv, struct gve_queue_page_list *qpl) */ if (vm_page_unwire_noq(qpl->pages[i])) { if (!qpl->kva) { - pmap_qremove((vm_offset_t)qpl->dmas[i].cpu_addr, 1); + pmap_qremove(qpl->dmas[i].cpu_addr, 1); kva_free((vm_offset_t)qpl->dmas[i].cpu_addr, PAGE_SIZE); } vm_page_free(qpl->pages[i]); @@ -126,7 +126,7 @@ gve_alloc_qpl(struct gve_priv *priv, uint32_t id, int npages, bool single_kva) err = ENOMEM; goto abort; } - pmap_qenter((vm_offset_t)qpl->dmas[i].cpu_addr, &(qpl->pages[i]), 1); + pmap_qenter(qpl->dmas[i].cpu_addr, &(qpl->pages[i]), 1); } else qpl->dmas[i].cpu_addr = (void *)(qpl->kva + (PAGE_SIZE * i)); @@ -135,7 +135,7 @@ gve_alloc_qpl(struct gve_priv *priv, uint32_t id, int npages, bool single_kva) } if (single_kva) - pmap_qenter(qpl->kva, qpl->pages, npages); + pmap_qenter((void *)qpl->kva, qpl->pages, npages); for (i = 0; i < npages; i++) { err = gve_dmamap_create(priv, /*size=*/PAGE_SIZE, /*align=*/PAGE_SIZE, @@ -244,7 +244,7 @@ void gve_mextadd_free(struct mbuf *mbuf) { vm_page_t page = (vm_page_t)mbuf->m_ext.ext_arg1; - vm_offset_t va = (vm_offset_t)mbuf->m_ext.ext_arg2; + void *va = mbuf->m_ext.ext_arg2; /* * Free the page only if this is the last ref. @@ -253,7 +253,7 @@ gve_mextadd_free(struct mbuf *mbuf) */ if (__predict_false(vm_page_unwire_noq(page))) { pmap_qremove(va, 1); - kva_free(va, PAGE_SIZE); + kva_free((vm_offset_t)va, PAGE_SIZE); vm_page_free(page); } } diff --git a/sys/dev/hwt/hwt_vm.c b/sys/dev/hwt/hwt_vm.c index 6c55e218dcec..a3e906d71099 100644 --- a/sys/dev/hwt/hwt_vm.c +++ b/sys/dev/hwt/hwt_vm.c @@ -171,7 +171,7 @@ retry: VM_OBJECT_WLOCK(vm->obj); vm_page_insert(m, vm->obj, i); if (kva_req) - pmap_qenter(vm->kvaddr + i * PAGE_SIZE, &m, 1); + pmap_qenter((char *)vm->kvaddr + i * PAGE_SIZE, &m, 1); VM_OBJECT_WUNLOCK(vm->obj); } @@ -442,7 +442,7 @@ hwt_vm_destroy_buffers(struct hwt_vm *vm) int i; if (vm->ctx->hwt_backend->kva_req && vm->kvaddr != 0) { - pmap_qremove(vm->kvaddr, vm->npages); + pmap_qremove((void *)vm->kvaddr, vm->npages); kva_free(vm->kvaddr, vm->npages * PAGE_SIZE); } VM_OBJECT_WLOCK(vm->obj); diff --git a/sys/dev/md/md.c b/sys/dev/md/md.c index 8d2908264aac..e2c0451b5843 100644 --- a/sys/dev/md/md.c +++ b/sys/dev/md/md.c @@ -286,7 +286,7 @@ struct md_s { struct { struct vnode *vnode; char file[PATH_MAX]; - vm_offset_t kva; + char *kva; } s_vnode; /* MD_SWAP related fields */ @@ -983,8 +983,7 @@ unmapped_step: ("npages %d too large", npages)); pmap_qenter(sc->s_vnode.kva, &bp->bio_ma[atop(ma_offs)], npages); - aiov.iov_base = (void *)(sc->s_vnode.kva + (ma_offs & - PAGE_MASK)); + aiov.iov_base = sc->s_vnode.kva + (ma_offs & PAGE_MASK); aiov.iov_len = iolen; auio.uio_iov = &aiov; auio.uio_iovcnt = 1; @@ -1513,7 +1512,7 @@ mdcreate_vnode(struct md_s *sc, struct md_req *mdr, struct thread *td) goto bad; } - sc->s_vnode.kva = kva_alloc(maxphys + PAGE_SIZE); + sc->s_vnode.kva = (char *)kva_alloc(maxphys + PAGE_SIZE); return (0); bad: VOP_UNLOCK(nd.ni_vp); @@ -1567,8 +1566,8 @@ mddestroy(struct md_s *sc, struct thread *td) sc->flags & MD_READONLY ? FREAD : (FREAD|FWRITE), sc->cred, td); } - if (sc->s_vnode.kva != 0) - kva_free(sc->s_vnode.kva, maxphys + PAGE_SIZE); + if (sc->s_vnode.kva != NULL) + kva_free((vm_offset_t)sc->s_vnode.kva, maxphys + PAGE_SIZE); break; case MD_SWAP: if (sc->s_swap.object != NULL) diff --git a/sys/dev/pci/controller/pci_n1sdp.c b/sys/dev/pci/controller/pci_n1sdp.c index 22f0ea27d45b..c1f8624e45aa 100644 --- a/sys/dev/pci/controller/pci_n1sdp.c +++ b/sys/dev/pci/controller/pci_n1sdp.c @@ -83,7 +83,7 @@ static int n1sdp_init(struct generic_pcie_n1sdp_softc *sc) { struct pcie_discovery_data *shared_data; - vm_offset_t vaddr; + void *vaddr; vm_paddr_t paddr_rc; vm_paddr_t paddr; vm_page_t m[BDF_TABLE_SIZE / PAGE_SIZE]; @@ -100,8 +100,8 @@ n1sdp_init(struct generic_pcie_n1sdp_softc *sc) MPASS(m[i] != NULL); } - vaddr = kva_alloc((vm_size_t)BDF_TABLE_SIZE); - if (vaddr == 0) { + vaddr = (void *)kva_alloc((vm_size_t)BDF_TABLE_SIZE); + if (vaddr == NULL) { printf("%s: Can't allocate KVA memory.", __func__); error = ENXIO; goto out; @@ -130,7 +130,7 @@ n1sdp_init(struct generic_pcie_n1sdp_softc *sc) out_pmap: pmap_qremove(vaddr, nitems(m)); - kva_free(vaddr, (vm_size_t)BDF_TABLE_SIZE); + kva_free((vm_offset_t)vaddr, (vm_size_t)BDF_TABLE_SIZE); out: vm_phys_fictitious_unreg_range(paddr, paddr + BDF_TABLE_SIZE); diff --git a/sys/dev/spibus/spigen.c b/sys/dev/spibus/spigen.c index 94f276d239f3..400ae1e139ad 100644 --- a/sys/dev/spibus/spigen.c +++ b/sys/dev/spibus/spigen.c @@ -73,7 +73,7 @@ struct spigen_softc { struct spigen_mmap { vm_object_t bufobj; - vm_offset_t kvaddr; + void *kvaddr; size_t bufsize; }; @@ -235,10 +235,10 @@ spigen_transfer_mmapped(struct cdev *cdev, struct spigen_transfer_mmapped *stm) if (mmap->bufsize < stm->stm_command_length + stm->stm_data_length) return (E2BIG); - transfer.tx_cmd = transfer.rx_cmd = (void *)((uintptr_t)mmap->kvaddr); + transfer.tx_cmd = transfer.rx_cmd = mmap->kvaddr; transfer.tx_cmd_sz = transfer.rx_cmd_sz = stm->stm_command_length; transfer.tx_data = transfer.rx_data = - (void *)((uintptr_t)mmap->kvaddr + stm->stm_command_length); + (void *)((char *)mmap->kvaddr + stm->stm_command_length); transfer.tx_data_sz = transfer.rx_data_sz = stm->stm_data_length; error = SPIBUS_TRANSFER(device_get_parent(dev), dev, &transfer); @@ -283,9 +283,9 @@ spigen_mmap_cleanup(void *arg) { struct spigen_mmap *mmap = arg; - if (mmap->kvaddr != 0) { + if (mmap->kvaddr != NULL) { pmap_qremove(mmap->kvaddr, mmap->bufsize / PAGE_SIZE); - kva_free(mmap->kvaddr, mmap->bufsize); + kva_free((vm_offset_t)mmap->kvaddr, mmap->bufsize); } if (mmap->bufobj != NULL) vm_object_deallocate(mmap->bufobj); @@ -312,7 +312,7 @@ spigen_mmap_single(struct cdev *cdev, vm_ooffset_t *offset, return (EBUSY); mmap = malloc(sizeof(*mmap), M_DEVBUF, M_ZERO | M_WAITOK); - if ((mmap->kvaddr = kva_alloc(size)) == 0) { + if ((mmap->kvaddr = (void *)kva_alloc(size)) == 0) { spigen_mmap_cleanup(mmap); return (ENOMEM); } |
