aboutsummaryrefslogtreecommitdiff
path: root/sys
diff options
context:
space:
mode:
authorJohn Baldwin <jhb@FreeBSD.org>2022-09-22 22:09:19 +0000
committerJohn Baldwin <jhb@FreeBSD.org>2022-09-22 22:09:19 +0000
commitf49fd63a6a130ae464cdc7756e6f7d0d747c82c4 (patch)
tree218a911d9042b8afbfbb1c6643a2109a85dcf290 /sys
parent7ae99f80b6661760c5de3edd330b279f04b092a2 (diff)
downloadsrc-f49fd63a6a130ae464cdc7756e6f7d0d747c82c4.tar.gz
src-f49fd63a6a130ae464cdc7756e6f7d0d747c82c4.zip
Diffstat (limited to 'sys')
-rw-r--r--sys/amd64/amd64/mp_machdep.c12
-rw-r--r--sys/amd64/amd64/pmap.c2
-rw-r--r--sys/amd64/amd64/sys_machdep.c11
-rw-r--r--sys/amd64/amd64/vm_machdep.c2
-rw-r--r--sys/arm/arm/busdma_machdep.c6
-rw-r--r--sys/arm/arm/mp_machdep.c2
-rw-r--r--sys/arm/arm/pmap-v6.c8
-rw-r--r--sys/arm/freescale/imx/imx6_sdma.c8
-rw-r--r--sys/arm/nvidia/drm2/tegra_dc.c3
-rw-r--r--sys/arm/nvidia/drm2/tegra_drm.h2
-rw-r--r--sys/arm/nvidia/tegra_pcie.c2
-rw-r--r--sys/arm/nvidia/tegra_xhci.c10
-rw-r--r--sys/arm64/arm64/busdma_bounce.c6
-rw-r--r--sys/arm64/arm64/mp_machdep.c18
-rw-r--r--sys/arm64/arm64/pmap.c4
-rw-r--r--sys/compat/linuxkpi/common/include/linux/dma-mapping.h2
-rw-r--r--sys/compat/linuxkpi/common/src/linux_page.c6
-rw-r--r--sys/compat/linuxkpi/common/src/linux_pci.c4
-rw-r--r--sys/dev/agp/agp.c8
-rw-r--r--sys/dev/agp/agp_amd.c12
-rw-r--r--sys/dev/agp/agp_ati.c12
-rw-r--r--sys/dev/agp/agp_i810.c4
-rw-r--r--sys/dev/amd_ecc_inject/ecc_inject.c4
-rw-r--r--sys/dev/drm2/drmP.h2
-rw-r--r--sys/dev/drm2/drm_bufs.c4
-rw-r--r--sys/dev/drm2/drm_scatter.c16
-rw-r--r--sys/dev/hyperv/vmbus/hyperv.c5
-rw-r--r--sys/dev/iommu/busdma_iommu.c4
-rw-r--r--sys/dev/kvm_clock/kvm_clock.c2
-rw-r--r--sys/dev/liquidio/lio_network.h4
-rw-r--r--sys/dev/mlx5/mlx5_core/mlx5_fwdump.c21
-rw-r--r--sys/i386/i386/mp_machdep.c4
-rw-r--r--sys/i386/i386/pmap.c2
-rw-r--r--sys/kern/kern_malloc.c16
-rw-r--r--sys/kern/subr_busdma_bufalloc.c4
-rw-r--r--sys/powerpc/aim/mmu_radix.c2
-rw-r--r--sys/powerpc/powerpc/busdma_machdep.c4
-rw-r--r--sys/powerpc/powerpc/mp_machdep.c2
-rw-r--r--sys/riscv/riscv/busdma_bounce.c6
-rw-r--r--sys/riscv/riscv/mp_machdep.c8
-rw-r--r--sys/riscv/riscv/pmap.c2
-rw-r--r--sys/vm/uma_core.c4
-rw-r--r--sys/vm/vm_extern.h14
-rw-r--r--sys/vm/vm_init.c27
-rw-r--r--sys/vm/vm_kern.c48
-rw-r--r--sys/x86/iommu/intel_dmar.h2
-rw-r--r--sys/x86/iommu/intel_intrmap.c9
-rw-r--r--sys/x86/iommu/intel_qi.c4
-rw-r--r--sys/x86/x86/busdma_bounce.c6
-rw-r--r--sys/x86/x86/mp_x86.c3
50 files changed, 176 insertions, 197 deletions
diff --git a/sys/amd64/amd64/mp_machdep.c b/sys/amd64/amd64/mp_machdep.c
index 5cd9d93dd0f1..e5114b93328e 100644
--- a/sys/amd64/amd64/mp_machdep.c
+++ b/sys/amd64/amd64/mp_machdep.c
@@ -423,17 +423,17 @@ start_all_aps(void)
domain = acpi_pxm_get_cpu_locality(apic_id);
#endif
/* allocate and set up an idle stack data page */
- bootstacks[cpu] = (void *)kmem_malloc(kstack_pages * PAGE_SIZE,
+ bootstacks[cpu] = kmem_malloc(kstack_pages * PAGE_SIZE,
M_WAITOK | M_ZERO);
- doublefault_stack = (char *)kmem_malloc(DBLFAULT_STACK_SIZE,
+ doublefault_stack = kmem_malloc(DBLFAULT_STACK_SIZE,
M_WAITOK | M_ZERO);
- mce_stack = (char *)kmem_malloc(MCE_STACK_SIZE,
+ mce_stack = kmem_malloc(MCE_STACK_SIZE,
M_WAITOK | M_ZERO);
- nmi_stack = (char *)kmem_malloc_domainset(
+ nmi_stack = kmem_malloc_domainset(
DOMAINSET_PREF(domain), NMI_STACK_SIZE, M_WAITOK | M_ZERO);
- dbg_stack = (char *)kmem_malloc_domainset(
+ dbg_stack = kmem_malloc_domainset(
DOMAINSET_PREF(domain), DBG_STACK_SIZE, M_WAITOK | M_ZERO);
- dpcpu = (void *)kmem_malloc_domainset(DOMAINSET_PREF(domain),
+ dpcpu = kmem_malloc_domainset(DOMAINSET_PREF(domain),
DPCPU_SIZE, M_WAITOK | M_ZERO);
bootpcpu = &__pcpu[cpu];
diff --git a/sys/amd64/amd64/pmap.c b/sys/amd64/amd64/pmap.c
index c7fd0135880e..d7aeb8dcbd98 100644
--- a/sys/amd64/amd64/pmap.c
+++ b/sys/amd64/amd64/pmap.c
@@ -2390,7 +2390,7 @@ pmap_init_pv_table(void)
*/
s = (vm_size_t)pv_npg * sizeof(struct md_page);
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
TAILQ_INIT(&pv_dummy.pv_list);
diff --git a/sys/amd64/amd64/sys_machdep.c b/sys/amd64/amd64/sys_machdep.c
index c10b15896132..05c90f57ec6c 100644
--- a/sys/amd64/amd64/sys_machdep.c
+++ b/sys/amd64/amd64/sys_machdep.c
@@ -418,8 +418,7 @@ amd64_set_ioperm(td, uap)
*/
pcb = td->td_pcb;
if (pcb->pcb_tssp == NULL) {
- tssp = (struct amd64tss *)kmem_malloc(ctob(IOPAGES + 1),
- M_WAITOK);
+ tssp = kmem_malloc(ctob(IOPAGES + 1), M_WAITOK);
pmap_pti_add_kva((vm_offset_t)tssp, (vm_offset_t)tssp +
ctob(IOPAGES + 1), false);
iomap = (char *)&tssp[1];
@@ -523,8 +522,8 @@ user_ldt_alloc(struct proc *p, int force)
mtx_unlock(&dt_lock);
new_ldt = malloc(sizeof(struct proc_ldt), M_SUBPROC, M_WAITOK);
sz = max_ldt_segment * sizeof(struct user_segment_descriptor);
- sva = kmem_malloc(sz, M_WAITOK | M_ZERO);
- new_ldt->ldt_base = (caddr_t)sva;
+ new_ldt->ldt_base = kmem_malloc(sz, M_WAITOK | M_ZERO);
+ sva = (uintptr_t)new_ldt->ldt_base;
pmap_pti_add_kva(sva, sva + sz, false);
new_ldt->ldt_refcnt = 1;
sldt.ssd_base = sva;
@@ -539,7 +538,7 @@ user_ldt_alloc(struct proc *p, int force)
pldt = mdp->md_ldt;
if (pldt != NULL && !force) {
pmap_pti_remove_kva(sva, sva + sz);
- kmem_free(sva, sz);
+ kmem_free(new_ldt->ldt_base, sz);
free(new_ldt, M_SUBPROC);
return (pldt);
}
@@ -592,7 +591,7 @@ user_ldt_derefl(struct proc_ldt *pldt)
sva = (vm_offset_t)pldt->ldt_base;
sz = max_ldt_segment * sizeof(struct user_segment_descriptor);
pmap_pti_remove_kva(sva, sva + sz);
- kmem_free(sva, sz);
+ kmem_free(pldt->ldt_base, sz);
free(pldt, M_SUBPROC);
}
}
diff --git a/sys/amd64/amd64/vm_machdep.c b/sys/amd64/amd64/vm_machdep.c
index cc90e659ff11..5a49a291f313 100644
--- a/sys/amd64/amd64/vm_machdep.c
+++ b/sys/amd64/amd64/vm_machdep.c
@@ -373,7 +373,7 @@ cpu_thread_clean(struct thread *td)
if (pcb->pcb_tssp != NULL) {
pmap_pti_remove_kva((vm_offset_t)pcb->pcb_tssp,
(vm_offset_t)pcb->pcb_tssp + ctob(IOPAGES + 1));
- kmem_free((vm_offset_t)pcb->pcb_tssp, ctob(IOPAGES + 1));
+ kmem_free(pcb->pcb_tssp, ctob(IOPAGES + 1));
pcb->pcb_tssp = NULL;
}
}
diff --git a/sys/arm/arm/busdma_machdep.c b/sys/arm/arm/busdma_machdep.c
index 2540e15f75c5..a42f03d4be81 100644
--- a/sys/arm/arm/busdma_machdep.c
+++ b/sys/arm/arm/busdma_machdep.c
@@ -776,10 +776,10 @@ bus_dmamem_alloc(bus_dma_tag_t dmat, void **vaddr, int flags,
howmany(dmat->maxsize, MIN(dmat->maxsegsz, PAGE_SIZE)) &&
dmat->alignment <= PAGE_SIZE &&
(dmat->boundary % PAGE_SIZE) == 0) {
- *vaddr = (void *)kmem_alloc_attr(dmat->maxsize, mflags, 0,
+ *vaddr = kmem_alloc_attr(dmat->maxsize, mflags, 0,
dmat->lowaddr, memattr);
} else {
- *vaddr = (void *)kmem_alloc_contig(dmat->maxsize, mflags, 0,
+ *vaddr = kmem_alloc_contig(dmat->maxsize, mflags, 0,
dmat->lowaddr, dmat->alignment, dmat->boundary, memattr);
}
if (*vaddr == NULL) {
@@ -822,7 +822,7 @@ bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
!exclusion_bounce(dmat))
uma_zfree(bufzone->umazone, vaddr);
else
- kmem_free((vm_offset_t)vaddr, dmat->maxsize);
+ kmem_free(vaddr, dmat->maxsize);
dmat->map_count--;
if (map->flags & DMAMAP_COHERENT)
diff --git a/sys/arm/arm/mp_machdep.c b/sys/arm/arm/mp_machdep.c
index 6f772deee2d4..4b70869db52d 100644
--- a/sys/arm/arm/mp_machdep.c
+++ b/sys/arm/arm/mp_machdep.c
@@ -115,7 +115,7 @@ cpu_mp_start(void)
/* Reserve memory for application processors */
for(i = 0; i < (mp_ncpus - 1); i++)
- dpcpu[i] = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
+ dpcpu[i] = kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
dcache_wbinv_poc_all();
diff --git a/sys/arm/arm/pmap-v6.c b/sys/arm/arm/pmap-v6.c
index 107519be9dc0..2830bffdc23c 100644
--- a/sys/arm/arm/pmap-v6.c
+++ b/sys/arm/arm/pmap-v6.c
@@ -1780,7 +1780,7 @@ pmap_init(void)
*/
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
@@ -2213,7 +2213,7 @@ pmap_pinit(pmap_t pmap)
*/
if (pmap->pm_pt1 == NULL) {
- pmap->pm_pt1 = (pt1_entry_t *)kmem_alloc_contig(NB_IN_PT1,
+ pmap->pm_pt1 = kmem_alloc_contig(NB_IN_PT1,
M_NOWAIT | M_ZERO, 0, -1UL, NB_IN_PT1, 0, pt_memattr);
if (pmap->pm_pt1 == NULL)
return (0);
@@ -2229,7 +2229,7 @@ pmap_pinit(pmap_t pmap)
* be used no matter which process is current. Its mapping
* in PT2MAP can be used only for current process.
*/
- pmap->pm_pt2tab = (pt2_entry_t *)kmem_alloc_attr(NB_IN_PT2TAB,
+ pmap->pm_pt2tab = kmem_alloc_attr(NB_IN_PT2TAB,
M_NOWAIT | M_ZERO, 0, -1UL, pt_memattr);
if (pmap->pm_pt2tab == NULL) {
/*
@@ -2237,7 +2237,7 @@ pmap_pinit(pmap_t pmap)
* UMA_ZONE_NOFREE flag, it's important to leave
* no allocation in pmap if initialization failed.
*/
- kmem_free((vm_offset_t)pmap->pm_pt1, NB_IN_PT1);
+ kmem_free(pmap->pm_pt1, NB_IN_PT1);
pmap->pm_pt1 = NULL;
return (0);
}
diff --git a/sys/arm/freescale/imx/imx6_sdma.c b/sys/arm/freescale/imx/imx6_sdma.c
index ce33791ede0d..7183867bdb55 100644
--- a/sys/arm/freescale/imx/imx6_sdma.c
+++ b/sys/arm/freescale/imx/imx6_sdma.c
@@ -185,7 +185,7 @@ sdma_alloc(void)
chn = i;
/* Allocate area for buffer descriptors */
- channel->bd = (void *)kmem_alloc_contig(PAGE_SIZE, M_ZERO, 0, ~0,
+ channel->bd = kmem_alloc_contig(PAGE_SIZE, M_ZERO, 0, ~0,
PAGE_SIZE, 0, VM_MEMATTR_UNCACHEABLE);
return (chn);
@@ -202,7 +202,7 @@ sdma_free(int chn)
channel = &sc->channel[chn];
channel->in_use = 0;
- kmem_free((vm_offset_t)channel->bd, PAGE_SIZE);
+ kmem_free(channel->bd, PAGE_SIZE);
return (0);
}
@@ -396,7 +396,7 @@ boot_firmware(struct sdma_softc *sc)
sz = SDMA_N_CHANNELS * sizeof(struct sdma_channel_control) + \
sizeof(struct sdma_context_data);
- sc->ccb = (void *)kmem_alloc_contig(sz, M_ZERO, 0, ~0, PAGE_SIZE, 0,
+ sc->ccb = kmem_alloc_contig(sz, M_ZERO, 0, ~0, PAGE_SIZE, 0,
VM_MEMATTR_UNCACHEABLE);
sc->ccb_phys = vtophys(sc->ccb);
@@ -415,7 +415,7 @@ boot_firmware(struct sdma_softc *sc)
/* Channel 0 is used for booting firmware */
chn = 0;
- sc->bd0 = (void *)kmem_alloc_contig(PAGE_SIZE, M_ZERO, 0, ~0, PAGE_SIZE,
+ sc->bd0 = kmem_alloc_contig(PAGE_SIZE, M_ZERO, 0, ~0, PAGE_SIZE,
0, VM_MEMATTR_UNCACHEABLE);
bd0 = sc->bd0;
sc->ccb[chn].base_bd_ptr = vtophys(bd0);
diff --git a/sys/arm/nvidia/drm2/tegra_dc.c b/sys/arm/nvidia/drm2/tegra_dc.c
index 110f3611bf87..8133a40b5c38 100644
--- a/sys/arm/nvidia/drm2/tegra_dc.c
+++ b/sys/arm/nvidia/drm2/tegra_dc.c
@@ -1231,7 +1231,8 @@ dc_init_client(device_t dev, device_t host1x, struct tegra_drm *drm)
sc->tegra_crtc.cursor_vbase = kmem_alloc_contig(256 * 256 * 4,
M_WAITOK | M_ZERO, 0, -1UL, PAGE_SIZE, 0,
VM_MEMATTR_WRITE_COMBINING);
- sc->tegra_crtc.cursor_pbase = vtophys(sc->tegra_crtc.cursor_vbase);
+ sc->tegra_crtc.cursor_pbase =
+ vtophys((uintptr_t)sc->tegra_crtc.cursor_vbase);
return (0);
}
diff --git a/sys/arm/nvidia/drm2/tegra_drm.h b/sys/arm/nvidia/drm2/tegra_drm.h
index ada4f4434e65..ab5b2519fb69 100644
--- a/sys/arm/nvidia/drm2/tegra_drm.h
+++ b/sys/arm/nvidia/drm2/tegra_drm.h
@@ -64,7 +64,7 @@ struct tegra_crtc {
device_t dev;
int nvidia_head;
vm_paddr_t cursor_pbase; /* Cursor buffer */
- vm_offset_t cursor_vbase;
+ void *cursor_vbase;
};
struct tegra_drm_encoder {
diff --git a/sys/arm/nvidia/tegra_pcie.c b/sys/arm/nvidia/tegra_pcie.c
index ce3a32c80518..ecc58a93066f 100644
--- a/sys/arm/nvidia/tegra_pcie.c
+++ b/sys/arm/nvidia/tegra_pcie.c
@@ -1382,7 +1382,7 @@ tegra_pcib_attach_msi(device_t dev)
sc = device_get_softc(dev);
- sc->msi_page = kmem_alloc_contig(PAGE_SIZE, M_WAITOK, 0,
+ sc->msi_page = (uintptr_t)kmem_alloc_contig(PAGE_SIZE, M_WAITOK, 0,
BUS_SPACE_MAXADDR, PAGE_SIZE, 0, VM_MEMATTR_DEFAULT);
/* MSI BAR */
diff --git a/sys/arm/nvidia/tegra_xhci.c b/sys/arm/nvidia/tegra_xhci.c
index edf62f823d2c..4187271bbdb6 100644
--- a/sys/arm/nvidia/tegra_xhci.c
+++ b/sys/arm/nvidia/tegra_xhci.c
@@ -289,7 +289,7 @@ struct tegra_xhci_softc {
struct intr_config_hook irq_hook;
bool xhci_inited;
- vm_offset_t fw_vaddr;
+ void *fw_vaddr;
vm_size_t fw_size;
};
@@ -744,7 +744,7 @@ load_fw(struct tegra_xhci_softc *sc)
const struct firmware *fw;
const struct tegra_xusb_fw_hdr *fw_hdr;
vm_paddr_t fw_paddr, fw_base;
- vm_offset_t fw_vaddr;
+ void *fw_vaddr;
vm_size_t fw_size;
uint32_t code_tags, code_size;
struct clocktime fw_clock;
@@ -775,9 +775,9 @@ load_fw(struct tegra_xhci_softc *sc)
fw_vaddr = kmem_alloc_contig(fw_size, M_WAITOK, 0, -1UL, PAGE_SIZE, 0,
VM_MEMATTR_UNCACHEABLE);
- fw_paddr = vtophys(fw_vaddr);
+ fw_paddr = vtophys((uintptr_t)fw_vaddr);
fw_hdr = (const struct tegra_xusb_fw_hdr *)fw_vaddr;
- memcpy((void *)fw_vaddr, fw->data, fw_size);
+ memcpy(fw_vaddr, fw->data, fw_size);
firmware_put(fw, FIRMWARE_UNLOAD);
sc->fw_vaddr = fw_vaddr;
@@ -947,7 +947,7 @@ tegra_xhci_detach(device_t dev)
xhci_uninit(xsc);
if (sc->irq_hdl_mbox != NULL)
bus_teardown_intr(dev, sc->irq_res_mbox, sc->irq_hdl_mbox);
- if (sc->fw_vaddr != 0)
+ if (sc->fw_vaddr != NULL)
kmem_free(sc->fw_vaddr, sc->fw_size);
LOCK_DESTROY(sc);
return (0);
diff --git a/sys/arm64/arm64/busdma_bounce.c b/sys/arm64/arm64/busdma_bounce.c
index ce28e07d83a6..ced1656727e0 100644
--- a/sys/arm64/arm64/busdma_bounce.c
+++ b/sys/arm64/arm64/busdma_bounce.c
@@ -567,11 +567,11 @@ bounce_bus_dmamem_alloc(bus_dma_tag_t dmat, void** vaddr, int flags,
dmat->alloc_alignment <= PAGE_SIZE &&
(dmat->common.boundary % PAGE_SIZE) == 0) {
/* Page-based multi-segment allocations allowed */
- *vaddr = (void *)kmem_alloc_attr(dmat->alloc_size, mflags,
+ *vaddr = kmem_alloc_attr(dmat->alloc_size, mflags,
0ul, dmat->common.lowaddr, attr);
dmat->bounce_flags |= BF_KMEM_ALLOC;
} else {
- *vaddr = (void *)kmem_alloc_contig(dmat->alloc_size, mflags,
+ *vaddr = kmem_alloc_contig(dmat->alloc_size, mflags,
0ul, dmat->common.lowaddr, dmat->alloc_alignment != 0 ?
dmat->alloc_alignment : 1ul, dmat->common.boundary, attr);
dmat->bounce_flags |= BF_KMEM_ALLOC;
@@ -608,7 +608,7 @@ bounce_bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
if ((dmat->bounce_flags & BF_KMEM_ALLOC) == 0)
free(vaddr, M_DEVBUF);
else
- kmem_free((vm_offset_t)vaddr, dmat->alloc_size);
+ kmem_free(vaddr, dmat->alloc_size);
free(map, M_DEVBUF);
dmat->map_count--;
CTR3(KTR_BUSDMA, "%s: tag %p flags 0x%x", __func__, dmat,
diff --git a/sys/arm64/arm64/mp_machdep.c b/sys/arm64/arm64/mp_machdep.c
index ee512ce4e256..ef0ded31bcf7 100644
--- a/sys/arm64/arm64/mp_machdep.c
+++ b/sys/arm64/arm64/mp_machdep.c
@@ -319,8 +319,7 @@ smp_after_idle_runnable(void *arg __unused)
for (cpu = 1; cpu < mp_ncpus; cpu++) {
if (bootstacks[cpu] != NULL)
- kmem_free((vm_offset_t)bootstacks[cpu],
- MP_BOOTSTACK_SIZE);
+ kmem_free(bootstacks[cpu], MP_BOOTSTACK_SIZE);
}
}
SYSINIT(smp_after_idle_runnable, SI_SUB_SMP, SI_ORDER_ANY,
@@ -498,7 +497,6 @@ static bool
start_cpu(u_int cpuid, uint64_t target_cpu, int domain)
{
struct pcpu *pcpup;
- vm_offset_t pcpu_mem;
vm_size_t size;
vm_paddr_t pa;
int err, naps;
@@ -514,11 +512,9 @@ start_cpu(u_int cpuid, uint64_t target_cpu, int domain)
KASSERT(cpuid < MAXCPU, ("Too many CPUs"));
size = round_page(sizeof(*pcpup) + DPCPU_SIZE);
- pcpu_mem = kmem_malloc_domainset(DOMAINSET_PREF(domain), size,
+ pcpup = kmem_malloc_domainset(DOMAINSET_PREF(domain), size,
M_WAITOK | M_ZERO);
- pmap_disable_promotion(pcpu_mem, size);
-
- pcpup = (struct pcpu *)pcpu_mem;
+ pmap_disable_promotion((vm_offset_t)pcpup, size);
pcpu_init(pcpup, cpuid, sizeof(struct pcpu));
pcpup->pc_mpidr_low = target_cpu & CPU_AFF_MASK;
pcpup->pc_mpidr_high = (target_cpu & CPU_AFF_MASK) >> 32;
@@ -526,8 +522,8 @@ start_cpu(u_int cpuid, uint64_t target_cpu, int domain)
dpcpu[cpuid - 1] = (void *)(pcpup + 1);
dpcpu_init(dpcpu[cpuid - 1], cpuid);
- bootstacks[cpuid] = (void *)kmem_malloc_domainset(
- DOMAINSET_PREF(domain), MP_BOOTSTACK_SIZE, M_WAITOK | M_ZERO);
+ bootstacks[cpuid] = kmem_malloc_domainset(DOMAINSET_PREF(domain),
+ MP_BOOTSTACK_SIZE, M_WAITOK | M_ZERO);
naps = atomic_load_int(&aps_started);
bootstack = (char *)bootstacks[cpuid] + MP_BOOTSTACK_SIZE;
@@ -548,8 +544,8 @@ start_cpu(u_int cpuid, uint64_t target_cpu, int domain)
pcpu_destroy(pcpup);
dpcpu[cpuid - 1] = NULL;
- kmem_free((vm_offset_t)bootstacks[cpuid], MP_BOOTSTACK_SIZE);
- kmem_free(pcpu_mem, size);
+ kmem_free(bootstacks[cpuid], MP_BOOTSTACK_SIZE);
+ kmem_free(pcpup, size);
bootstacks[cpuid] = NULL;
mp_ncpus--;
return (false);
diff --git a/sys/arm64/arm64/pmap.c b/sys/arm64/arm64/pmap.c
index 2dc453e1aa33..697bf9665133 100644
--- a/sys/arm64/arm64/pmap.c
+++ b/sys/arm64/arm64/pmap.c
@@ -1247,7 +1247,7 @@ pmap_init_asids(struct asid_set *set, int bits)
* bit_alloc().
*/
set->asid_set_size = 1 << set->asid_bits;
- set->asid_set = (bitstr_t *)kmem_malloc(bitstr_size(set->asid_set_size),
+ set->asid_set = kmem_malloc(bitstr_size(set->asid_set_size),
M_WAITOK | M_ZERO);
for (i = 0; i < ASID_FIRST_AVAILABLE; i++)
bit_set(set->asid_set, i);
@@ -1326,7 +1326,7 @@ pmap_init(void)
*/
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
TAILQ_INIT(&pv_dummy.pv_list);
diff --git a/sys/compat/linuxkpi/common/include/linux/dma-mapping.h b/sys/compat/linuxkpi/common/include/linux/dma-mapping.h
index ef3570ef7639..8401006fbf5f 100644
--- a/sys/compat/linuxkpi/common/include/linux/dma-mapping.h
+++ b/sys/compat/linuxkpi/common/include/linux/dma-mapping.h
@@ -175,7 +175,7 @@ dma_free_coherent(struct device *dev, size_t size, void *cpu_addr,
{
linux_dma_unmap(dev, dma_addr, size);
- kmem_free((vm_offset_t)cpu_addr, size);
+ kmem_free(cpu_addr, size);
}
static inline dma_addr_t
diff --git a/sys/compat/linuxkpi/common/src/linux_page.c b/sys/compat/linuxkpi/common/src/linux_page.c
index df4a124cf3e2..26b0ed649372 100644
--- a/sys/compat/linuxkpi/common/src/linux_page.c
+++ b/sys/compat/linuxkpi/common/src/linux_page.c
@@ -170,7 +170,7 @@ vm_offset_t
linux_alloc_kmem(gfp_t flags, unsigned int order)
{
size_t size = ((size_t)PAGE_SIZE) << order;
- vm_offset_t addr;
+ void *addr;
if ((flags & GFP_DMA32) == 0) {
addr = kmem_malloc(size, flags & GFP_NATIVE_MASK);
@@ -178,7 +178,7 @@ linux_alloc_kmem(gfp_t flags, unsigned int order)
addr = kmem_alloc_contig(size, flags & GFP_NATIVE_MASK, 0,
BUS_SPACE_MAXADDR_32BIT, PAGE_SIZE, 0, VM_MEMATTR_DEFAULT);
}
- return (addr);
+ return ((vm_offset_t)addr);
}
void
@@ -186,7 +186,7 @@ linux_free_kmem(vm_offset_t addr, unsigned int order)
{
size_t size = ((size_t)PAGE_SIZE) << order;
- kmem_free(addr, size);
+ kmem_free((void *)addr, size);
}
static int
diff --git a/sys/compat/linuxkpi/common/src/linux_pci.c b/sys/compat/linuxkpi/common/src/linux_pci.c
index 732da9b36261..aaa6d646f04d 100644
--- a/sys/compat/linuxkpi/common/src/linux_pci.c
+++ b/sys/compat/linuxkpi/common/src/linux_pci.c
@@ -1124,13 +1124,13 @@ linux_dma_alloc_coherent(struct device *dev, size_t size,
align = PAGE_SIZE << get_order(size);
/* Always zero the allocation. */
flag |= M_ZERO;
- mem = (void *)kmem_alloc_contig(size, flag & GFP_NATIVE_MASK, 0, high,
+ mem = kmem_alloc_contig(size, flag & GFP_NATIVE_MASK, 0, high,
align, 0, VM_MEMATTR_DEFAULT);
if (mem != NULL) {
*dma_handle = linux_dma_map_phys_common(dev, vtophys(mem), size,
priv->dmat_coherent);
if (*dma_handle == 0) {
- kmem_free((vm_offset_t)mem, size);
+ kmem_free(mem, size);
mem = NULL;
}
} else {
diff --git a/sys/dev/agp/agp.c b/sys/dev/agp/agp.c
index b973b3d7ce3f..f3ffb91cef3e 100644
--- a/sys/dev/agp/agp.c
+++ b/sys/dev/agp/agp.c
@@ -153,9 +153,8 @@ agp_alloc_gatt(device_t dev)
return 0;
gatt->ag_entries = entries;
- gatt->ag_virtual = (void *)kmem_alloc_contig(entries *
- sizeof(u_int32_t), M_NOWAIT | M_ZERO, 0, ~0, PAGE_SIZE, 0,
- VM_MEMATTR_WRITE_COMBINING);
+ gatt->ag_virtual = kmem_alloc_contig(entries * sizeof(uint32_t),
+ M_NOWAIT | M_ZERO, 0, ~0, PAGE_SIZE, 0, VM_MEMATTR_WRITE_COMBINING);
if (!gatt->ag_virtual) {
if (bootverbose)
device_printf(dev, "contiguous allocation failed\n");
@@ -170,8 +169,7 @@ agp_alloc_gatt(device_t dev)
void
agp_free_gatt(struct agp_gatt *gatt)
{
- kmem_free((vm_offset_t)gatt->ag_virtual, gatt->ag_entries *
- sizeof(u_int32_t));
+ kmem_free(gatt->ag_virtual, gatt->ag_entries * sizeof(uint32_t));
free(gatt, M_AGP);
}
diff --git a/sys/dev/agp/agp_amd.c b/sys/dev/agp/agp_amd.c
index b104ec73f7f7..ff5f5475850c 100644
--- a/sys/dev/agp/agp_amd.c
+++ b/sys/dev/agp/agp_amd.c
@@ -101,7 +101,7 @@ agp_amd_alloc_gatt(device_t dev)
* directory.
*/
gatt->ag_entries = entries;
- gatt->ag_virtual = (void *)kmem_alloc_attr(entries * sizeof(u_int32_t),
+ gatt->ag_virtual = kmem_alloc_attr(entries * sizeof(uint32_t),
M_NOWAIT | M_ZERO, 0, ~0, VM_MEMATTR_WRITE_COMBINING);
if (!gatt->ag_virtual) {
if (bootverbose)
@@ -113,14 +113,13 @@ agp_amd_alloc_gatt(device_t dev)
/*
* Allocate the page directory.
*/
- gatt->ag_vdir = (void *)kmem_alloc_attr(AGP_PAGE_SIZE, M_NOWAIT |
+ gatt->ag_vdir = kmem_alloc_attr(AGP_PAGE_SIZE, M_NOWAIT |
M_ZERO, 0, ~0, VM_MEMATTR_WRITE_COMBINING);
if (!gatt->ag_vdir) {
if (bootverbose)
device_printf(dev,
"failed to allocate page directory\n");
- kmem_free((vm_offset_t)gatt->ag_virtual, entries *
- sizeof(u_int32_t));
+ kmem_free(gatt->ag_virtual, entries * sizeof(uint32_t));
free(gatt, M_AGP);
return 0;
}
@@ -168,9 +167,8 @@ agp_amd_alloc_gatt(device_t dev)
static void
agp_amd_free_gatt(struct agp_amd_gatt *gatt)
{
- kmem_free((vm_offset_t)gatt->ag_vdir, AGP_PAGE_SIZE);
- kmem_free((vm_offset_t)gatt->ag_virtual, gatt->ag_entries *
- sizeof(u_int32_t));
+ kmem_free(gatt->ag_vdir, AGP_PAGE_SIZE);
+ kmem_free(gatt->ag_virtual, gatt->ag_entries * sizeof(uint32_t));
free(gatt, M_AGP);
}
diff --git a/sys/dev/agp/agp_ati.c b/sys/dev/agp/agp_ati.c
index 8fc75a604c8c..2eaeb81cc75b 100644
--- a/sys/dev/agp/agp_ati.c
+++ b/sys/dev/agp/agp_ati.c
@@ -132,7 +132,7 @@ agp_ati_alloc_gatt(device_t dev)
/* Alloc the GATT -- pointers to pages of AGP memory */
sc->ag_entries = entries;
- sc->ag_virtual = (void *)kmem_alloc_attr(entries * sizeof(u_int32_t),
+ sc->ag_virtual = kmem_alloc_attr(entries * sizeof(uint32_t),
M_NOWAIT | M_ZERO, 0, ~0, VM_MEMATTR_WRITE_COMBINING);
if (sc->ag_virtual == NULL) {
if (bootverbose)
@@ -141,13 +141,12 @@ agp_ati_alloc_gatt(device_t dev)
}
/* Alloc the page directory -- pointers to each page of the GATT */
- sc->ag_vdir = (void *)kmem_alloc_attr(AGP_PAGE_SIZE, M_NOWAIT | M_ZERO,
+ sc->ag_vdir = kmem_alloc_attr(AGP_PAGE_SIZE, M_NOWAIT | M_ZERO,
0, ~0, VM_MEMATTR_WRITE_COMBINING);
if (sc->ag_vdir == NULL) {
if (bootverbose)
device_printf(dev, "pagedir allocation failed\n");
- kmem_free((vm_offset_t)sc->ag_virtual, entries *
- sizeof(u_int32_t));
+ kmem_free(sc->ag_virtual, entries * sizeof(uint32_t));
return ENOMEM;
}
sc->ag_pdir = vtophys((vm_offset_t)sc->ag_vdir);
@@ -263,9 +262,8 @@ agp_ati_detach(device_t dev)
temp = pci_read_config(dev, apsize_reg, 4);
pci_write_config(dev, apsize_reg, temp & ~1, 4);
- kmem_free((vm_offset_t)sc->ag_vdir, AGP_PAGE_SIZE);
- kmem_free((vm_offset_t)sc->ag_virtual, sc->ag_entries *
- sizeof(u_int32_t));
+ kmem_free(sc->ag_vdir, AGP_PAGE_SIZE);
+ kmem_free(sc->ag_virtual, sc->ag_entries * sizeof(uint32_t));
bus_release_resource(dev, SYS_RES_MEMORY, ATI_GART_MMADDR, sc->regs);
agp_free_res(dev);
diff --git a/sys/dev/agp/agp_i810.c b/sys/dev/agp/agp_i810.c
index 400e70a402c1..8d6b7da86eef 100644
--- a/sys/dev/agp/agp_i810.c
+++ b/sys/dev/agp/agp_i810.c
@@ -1189,7 +1189,7 @@ agp_i810_install_gatt(device_t dev)
sc->dcache_size = 0;
/* According to the specs the gatt on the i810 must be 64k. */
- sc->gatt->ag_virtual = (void *)kmem_alloc_contig(64 * 1024, M_NOWAIT |
+ sc->gatt->ag_virtual = kmem_alloc_contig(64 * 1024, M_NOWAIT |
M_ZERO, 0, ~0, PAGE_SIZE, 0, VM_MEMATTR_WRITE_COMBINING);
if (sc->gatt->ag_virtual == NULL) {
if (bootverbose)
@@ -1329,7 +1329,7 @@ agp_i810_deinstall_gatt(device_t dev)
sc = device_get_softc(dev);
bus_write_4(sc->sc_res[0], AGP_I810_PGTBL_CTL, 0);
- kmem_free((vm_offset_t)sc->gatt->ag_virtual, 64 * 1024);
+ kmem_free(sc->gatt->ag_virtual, 64 * 1024);
}
static void
diff --git a/sys/dev/amd_ecc_inject/ecc_inject.c b/sys/dev/amd_ecc_inject/ecc_inject.c
index 657e10407dbc..6e5fb2c3dc43 100644
--- a/sys/dev/amd_ecc_inject/ecc_inject.c
+++ b/sys/dev/amd_ecc_inject/ecc_inject.c
@@ -177,7 +177,7 @@ ecc_ei_inject_one(void *arg, size_t size)
static void
ecc_ei_inject(int count)
{
- vm_offset_t memory;
+ void *memory;
int injected;
KASSERT((quadrant & ~QUADRANT_MASK) == 0,
@@ -191,7 +191,7 @@ ecc_ei_inject(int count)
VM_MEMATTR_UNCACHEABLE);
for (injected = 0; injected < count; injected++) {
- ecc_ei_inject_one((void*)memory, PAGE_SIZE);
+ ecc_ei_inject_one(memory, PAGE_SIZE);
if (delay_ms != 0 && injected != count - 1)
pause_sbt("ecc_ei_inject", delay_ms * SBT_1MS, 0, 0);
}
diff --git a/sys/dev/drm2/drmP.h b/sys/dev/drm2/drmP.h
index ab11ad7f8cc0..f66f190958b5 100644
--- a/sys/dev/drm2/drmP.h
+++ b/sys/dev/drm2/drmP.h
@@ -497,7 +497,7 @@ struct drm_agp_head {
* Scatter-gather memory.
*/
struct drm_sg_mem {
- vm_offset_t vaddr;
+ void *vaddr;
vm_paddr_t *busaddr;
vm_pindex_t pages;
};
diff --git a/sys/dev/drm2/drm_bufs.c b/sys/dev/drm2/drm_bufs.c
index 9648acebea4f..c2eb7fe19258 100644
--- a/sys/dev/drm2/drm_bufs.c
+++ b/sys/dev/drm2/drm_bufs.c
@@ -392,8 +392,8 @@ static int drm_addmap_core(struct drm_device * dev, resource_size_t offset,
free(map, DRM_MEM_MAPS);
return -EINVAL;
}
- map->handle = (void *)(dev->sg->vaddr + offset);
- map->offset += dev->sg->vaddr;
+ map->handle = (char *)dev->sg->vaddr + offset;
+ map->offset += (uintptr_t)dev->sg->vaddr;
break;
case _DRM_CONSISTENT:
/* dma_addr_t is 64bit on i386 with CONFIG_HIGHMEM64G,
diff --git a/sys/dev/drm2/drm_scatter.c b/sys/dev/drm2/drm_scatter.c
index 1ccc88a5f69c..13040ee43b7b 100644
--- a/sys/dev/drm2/drm_scatter.c
+++ b/sys/dev/drm2/drm_scatter.c
@@ -35,7 +35,7 @@ __FBSDID("$FreeBSD$");
#define DEBUG_SCATTER 0
-static inline vm_offset_t drm_vmalloc_dma(vm_size_t size)
+static inline void *drm_vmalloc_dma(vm_size_t size)
{
return kmem_alloc_attr(size, M_NOWAIT | M_ZERO, 0,
BUS_SPACE_MAXADDR_32BIT, VM_MEMATTR_WRITE_COMBINING);
@@ -46,7 +46,7 @@ void drm_sg_cleanup(struct drm_sg_mem * entry)
if (entry == NULL)
return;
- if (entry->vaddr != 0)
+ if (entry->vaddr != NULL)
kmem_free(entry->vaddr, IDX_TO_OFF(entry->pages));
free(entry->busaddr, DRM_MEM_SGLISTS);
@@ -83,7 +83,7 @@ int drm_sg_alloc(struct drm_device *dev, struct drm_scatter_gather * request)
}
entry->vaddr = drm_vmalloc_dma(size);
- if (entry->vaddr == 0) {
+ if (entry->vaddr == NULL) {
free(entry->busaddr, DRM_MEM_DRIVER);
free(entry, DRM_MEM_DRIVER);
return -ENOMEM;
@@ -91,14 +91,14 @@ int drm_sg_alloc(struct drm_device *dev, struct drm_scatter_gather * request)
for (pindex = 0; pindex < entry->pages; pindex++) {
entry->busaddr[pindex] =
- vtophys(entry->vaddr + IDX_TO_OFF(pindex));
+ vtophys((uintptr_t)entry->vaddr + IDX_TO_OFF(pindex));
}
- request->handle = entry->vaddr;
+ request->handle = (uintptr_t)entry->vaddr;
dev->sg = entry;
- DRM_DEBUG("allocated %ju pages @ 0x%08zx, contents=%08lx\n",
+ DRM_DEBUG("allocated %ju pages @ %p, contents=%08lx\n",
entry->pages, entry->vaddr, *(unsigned long *)entry->vaddr);
return 0;
@@ -125,10 +125,10 @@ int drm_sg_free(struct drm_device *dev, void *data,
entry = dev->sg;
dev->sg = NULL;
- if (!entry || entry->vaddr != request->handle)
+ if (!entry || (uintptr_t)entry->vaddr != request->handle)
return -EINVAL;
- DRM_DEBUG("free 0x%zx\n", entry->vaddr);
+ DRM_DEBUG("free %p\n", entry->vaddr);
drm_sg_cleanup(entry);
diff --git a/sys/dev/hyperv/vmbus/hyperv.c b/sys/dev/hyperv/vmbus/hyperv.c
index 01e0ad9610d9..b2a74036f6c3 100644
--- a/sys/dev/hyperv/vmbus/hyperv.c
+++ b/sys/dev/hyperv/vmbus/hyperv.c
@@ -268,7 +268,7 @@ SYSINIT(hyperv_initialize, SI_SUB_HYPERVISOR, SI_ORDER_FIRST, hyperv_init,
static void
hypercall_memfree(void)
{
- kmem_free((vm_offset_t)hypercall_context.hc_addr, PAGE_SIZE);
+ kmem_free(hypercall_context.hc_addr, PAGE_SIZE);
hypercall_context.hc_addr = NULL;
}
@@ -286,8 +286,7 @@ hypercall_create(void *arg __unused)
* the NX bit.
* - Assume kmem_malloc() returns properly aligned memory.
*/
- hypercall_context.hc_addr = (void *)kmem_malloc(PAGE_SIZE, M_EXEC |
- M_WAITOK);
+ hypercall_context.hc_addr = kmem_malloc(PAGE_SIZE, M_EXEC | M_WAITOK);
hypercall_context.hc_paddr = vtophys(hypercall_context.hc_addr);
/* Get the 'reserved' bits, which requires preservation. */
diff --git a/sys/dev/iommu/busdma_iommu.c b/sys/dev/iommu/busdma_iommu.c
index 8f63d8b47f19..fac23b730162 100644
--- a/sys/dev/iommu/busdma_iommu.c
+++ b/sys/dev/iommu/busdma_iommu.c
@@ -519,7 +519,7 @@ iommu_bus_dmamem_alloc(bus_dma_tag_t dmat, void** vaddr, int flags,
DOMAINSET_PREF(tag->common.domain), mflags);
map->flags |= BUS_DMAMAP_IOMMU_MALLOC;
} else {
- *vaddr = (void *)kmem_alloc_attr_domainset(
+ *vaddr = kmem_alloc_attr_domainset(
DOMAINSET_PREF(tag->common.domain), tag->common.maxsize,
mflags, 0ul, BUS_SPACE_MAXADDR, attr);
map->flags |= BUS_DMAMAP_IOMMU_KMEM_ALLOC;
@@ -547,7 +547,7 @@ iommu_bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map1)
} else {
KASSERT((map->flags & BUS_DMAMAP_IOMMU_KMEM_ALLOC) != 0,
("iommu_bus_dmamem_free for non alloced map %p", map));
- kmem_free((vm_offset_t)vaddr, tag->common.maxsize);
+ kmem_free(vaddr, tag->common.maxsize);
map->flags &= ~BUS_DMAMAP_IOMMU_KMEM_ALLOC;
}
diff --git a/sys/dev/kvm_clock/kvm_clock.c b/sys/dev/kvm_clock/kvm_clock.c
index e6b991ba0f77..81c617755385 100644
--- a/sys/dev/kvm_clock/kvm_clock.c
+++ b/sys/dev/kvm_clock/kvm_clock.c
@@ -148,7 +148,7 @@ kvm_clock_attach(device_t dev)
(regs[0] & KVM_FEATURE_CLOCKSOURCE_STABLE_BIT) != 0;
/* Set up 'struct pvclock_vcpu_time_info' page(s): */
- sc->timeinfos = (struct pvclock_vcpu_time_info *)kmem_malloc(mp_ncpus *
+ sc->timeinfos = kmem_malloc(mp_ncpus *
sizeof(struct pvclock_vcpu_time_info), M_WAITOK | M_ZERO);
kvm_clock_system_time_enable(sc);
diff --git a/sys/dev/liquidio/lio_network.h b/sys/dev/liquidio/lio_network.h
index b1f4e1448fe9..5a843e4cf8ae 100644
--- a/sys/dev/liquidio/lio_network.h
+++ b/sys/dev/liquidio/lio_network.h
@@ -198,7 +198,7 @@ lio_dma_alloc(size_t size, vm_paddr_t *dma_handle)
void *mem;
align = PAGE_SIZE << lio_get_order(size);
- mem = (void *)kmem_alloc_contig(size, M_WAITOK, 0, ~0ul, align, 0,
+ mem = kmem_alloc_contig(size, M_WAITOK, 0, ~0ul, align, 0,
VM_MEMATTR_DEFAULT);
if (mem != NULL)
*dma_handle = vtophys(mem);
@@ -212,7 +212,7 @@ static inline void
lio_dma_free(size_t size, void *cpu_addr)
{
- kmem_free((vm_offset_t)cpu_addr, size);
+ kmem_free(cpu_addr, size);
}
static inline uint64_t
diff --git a/sys/dev/mlx5/mlx5_core/mlx5_fwdump.c b/sys/dev/mlx5/mlx5_core/mlx5_fwdump.c
index 315583601831..a272c759ea80 100644
--- a/sys/dev/mlx5/mlx5_core/mlx5_fwdump.c
+++ b/sys/dev/mlx5/mlx5_core/mlx5_fwdump.c
@@ -410,6 +410,7 @@ mlx5_ctl_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag,
struct mlx5_fw_update *fu;
struct firmware fake_fw;
struct mlx5_eeprom_get *eeprom_info;
+ void *fw_data;
int error;
error = 0;
@@ -461,21 +462,21 @@ mlx5_ctl_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag,
error = mlx5_dbsf_to_core(devaddr, &mdev);
if (error != 0)
break;
- bzero(&fake_fw, sizeof(fake_fw));
- fake_fw.name = "umlx_fw_up";
- fake_fw.datasize = fu->img_fw_data_len;
- fake_fw.version = 1;
- fake_fw.data = (void *)kmem_malloc(fu->img_fw_data_len,
- M_WAITOK);
+ fw_data = kmem_malloc(fu->img_fw_data_len, M_WAITOK);
if (fake_fw.data == NULL) {
error = ENOMEM;
break;
}
- error = copyin(fu->img_fw_data, __DECONST(void *, fake_fw.data),
- fu->img_fw_data_len);
- if (error == 0)
+ error = copyin(fu->img_fw_data, fw_data, fu->img_fw_data_len);
+ if (error == 0) {
+ bzero(&fake_fw, sizeof(fake_fw));
+ fake_fw.name = "umlx_fw_up";
+ fake_fw.datasize = fu->img_fw_data_len;
+ fake_fw.version = 1;
+ fake_fw.data = fw_data;
error = -mlx5_firmware_flash(mdev, &fake_fw);
- kmem_free((vm_offset_t)fake_fw.data, fu->img_fw_data_len);
+ }
+ kmem_free(fw_data, fu->img_fw_data_len);
break;
case MLX5_FW_RESET:
if ((fflag & FWRITE) == 0) {
diff --git a/sys/i386/i386/mp_machdep.c b/sys/i386/i386/mp_machdep.c
index bf1c7faf6182..98d716027060 100644
--- a/sys/i386/i386/mp_machdep.c
+++ b/sys/i386/i386/mp_machdep.c
@@ -393,9 +393,9 @@ start_all_aps(void)
apic_id = cpu_apic_ids[cpu];
/* allocate and set up a boot stack data page */
- bootstacks[cpu] = (char *)kmem_malloc(kstack_pages * PAGE_SIZE,
+ bootstacks[cpu] = kmem_malloc(kstack_pages * PAGE_SIZE,
M_WAITOK | M_ZERO);
- dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
+ dpcpu = kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
/* setup a vector to our boot code */
*((volatile u_short *) WARMBOOT_OFF) = WARMBOOT_TARGET;
*((volatile u_short *) WARMBOOT_SEG) = (boot_address >> 4);
diff --git a/sys/i386/i386/pmap.c b/sys/i386/i386/pmap.c
index 21aa5abee53b..372f4d9f4980 100644
--- a/sys/i386/i386/pmap.c
+++ b/sys/i386/i386/pmap.c
@@ -1050,7 +1050,7 @@ __CONCAT(PMTYPE, init)(void)
*/
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
diff --git a/sys/kern/kern_malloc.c b/sys/kern/kern_malloc.c
index 6bdcb48c5061..3ef16ec168a2 100644
--- a/sys/kern/kern_malloc.c
+++ b/sys/kern/kern_malloc.c
@@ -494,7 +494,7 @@ void
contigfree(void *addr, unsigned long size, struct malloc_type *type)
{
- kmem_free((vm_offset_t)addr, size);
+ kmem_free(addr, size);
malloc_type_freed(type, round_page(size));
}
@@ -588,17 +588,15 @@ static caddr_t __noinline
malloc_large(size_t size, struct malloc_type *mtp, struct domainset *policy,
int flags DEBUG_REDZONE_ARG_DEF)
{
- vm_offset_t kva;
- caddr_t va;
+ void *va;
size = roundup(size, PAGE_SIZE);
- kva = kmem_malloc_domainset(policy, size, flags);
- if (kva != 0) {
+ va = kmem_malloc_domainset(policy, size, flags);
+ if (va != NULL) {
/* The low bit is unused for slab pointers. */
- vsetzoneslab(kva, NULL, (void *)((size << 1) | 1));
+ vsetzoneslab((uintptr_t)va, NULL, (void *)((size << 1) | 1));
uma_total_inc(size);
}
- va = (caddr_t)kva;
malloc_type_allocated(mtp, va == NULL ? 0 : size);
if (__predict_false(va == NULL)) {
KASSERT((flags & M_WAITOK) == 0,
@@ -607,7 +605,7 @@ malloc_large(size_t size, struct malloc_type *mtp, struct domainset *policy,
#ifdef DEBUG_REDZONE
va = redzone_setup(va, osize);
#endif
- kasan_mark((void *)va, osize, size, KASAN_MALLOC_REDZONE);
+ kasan_mark(va, osize, size, KASAN_MALLOC_REDZONE);
}
return (va);
}
@@ -616,7 +614,7 @@ static void
free_large(void *addr, size_t size)
{
- kmem_free((vm_offset_t)addr, size);
+ kmem_free(addr, size);
uma_total_dec(size);
}
diff --git a/sys/kern/subr_busdma_bufalloc.c b/sys/kern/subr_busdma_bufalloc.c
index f201cde51519..ebf83d61ca18 100644
--- a/sys/kern/subr_busdma_bufalloc.c
+++ b/sys/kern/subr_busdma_bufalloc.c
@@ -158,7 +158,7 @@ busdma_bufalloc_alloc_uncacheable(uma_zone_t zone, vm_size_t size, int domain,
/* Inform UMA that this allocator uses kernel_arena/object. */
*pflag = UMA_SLAB_KERNEL;
- return ((void *)kmem_alloc_attr_domainset(DOMAINSET_FIXED(domain), size,
+ return (kmem_alloc_attr_domainset(DOMAINSET_FIXED(domain), size,
wait, 0, BUS_SPACE_MAXADDR, VM_MEMATTR_UNCACHEABLE));
#else
panic("VM_MEMATTR_UNCACHEABLE unavailable");
@@ -169,5 +169,5 @@ void
busdma_bufalloc_free_uncacheable(void *item, vm_size_t size, uint8_t pflag)
{
- kmem_free((vm_offset_t)item, size);
+ kmem_free(item, size);
}
diff --git a/sys/powerpc/aim/mmu_radix.c b/sys/powerpc/aim/mmu_radix.c
index c053de38034a..6e2203ccca45 100644
--- a/sys/powerpc/aim/mmu_radix.c
+++ b/sys/powerpc/aim/mmu_radix.c
@@ -3686,7 +3686,7 @@ mmu_radix_init(void)
*/
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
TAILQ_INIT(&pv_dummy.pv_list);
diff --git a/sys/powerpc/powerpc/busdma_machdep.c b/sys/powerpc/powerpc/busdma_machdep.c
index 3108b60599d4..1010a11d4d9d 100644
--- a/sys/powerpc/powerpc/busdma_machdep.c
+++ b/sys/powerpc/powerpc/busdma_machdep.c
@@ -483,7 +483,7 @@ bus_dmamem_alloc(bus_dma_tag_t dmat, void** vaddr, int flags,
* multi-seg allocations yet though.
* XXX Certain AGP hardware does.
*/
- *vaddr = (void *)kmem_alloc_contig(dmat->maxsize, mflags, 0ul,
+ *vaddr = kmem_alloc_contig(dmat->maxsize, mflags, 0ul,
dmat->lowaddr, dmat->alignment ? dmat->alignment : 1ul,
dmat->boundary, attr);
(*mapp)->contigalloc = 1;
@@ -511,7 +511,7 @@ bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
if (!map->contigalloc)
free(vaddr, M_DEVBUF);
else
- kmem_free((vm_offset_t)vaddr, dmat->maxsize);
+ kmem_free(vaddr, dmat->maxsize);
bus_dmamap_destroy(dmat, map);
CTR3(KTR_BUSDMA, "%s: tag %p flags 0x%x", __func__, dmat, dmat->flags);
}
diff --git a/sys/powerpc/powerpc/mp_machdep.c b/sys/powerpc/powerpc/mp_machdep.c
index 33ef870b8180..c83414bee104 100644
--- a/sys/powerpc/powerpc/mp_machdep.c
+++ b/sys/powerpc/powerpc/mp_machdep.c
@@ -176,7 +176,7 @@ cpu_mp_start(void)
void *dpcpu;
pc = &__pcpu[cpu.cr_cpuid];
- dpcpu = (void *)kmem_malloc_domainset(DOMAINSET_PREF(domain),
+ dpcpu = kmem_malloc_domainset(DOMAINSET_PREF(domain),
DPCPU_SIZE, M_WAITOK | M_ZERO);
pcpu_init(pc, cpu.cr_cpuid, sizeof(*pc));
dpcpu_init(dpcpu, cpu.cr_cpuid);
diff --git a/sys/riscv/riscv/busdma_bounce.c b/sys/riscv/riscv/busdma_bounce.c
index 5680e7702a2a..69f5b3438f77 100644
--- a/sys/riscv/riscv/busdma_bounce.c
+++ b/sys/riscv/riscv/busdma_bounce.c
@@ -454,11 +454,11 @@ bounce_bus_dmamem_alloc(bus_dma_tag_t dmat, void** vaddr, int flags,
dmat->common.alignment <= PAGE_SIZE &&
(dmat->common.boundary % PAGE_SIZE) == 0) {
/* Page-based multi-segment allocations allowed */
- *vaddr = (void *)kmem_alloc_attr(dmat->common.maxsize, mflags,
+ *vaddr = kmem_alloc_attr(dmat->common.maxsize, mflags,
0ul, dmat->common.lowaddr, attr);
dmat->bounce_flags |= BF_KMEM_ALLOC;
} else {
- *vaddr = (void *)kmem_alloc_contig(dmat->common.maxsize, mflags,
+ *vaddr = kmem_alloc_contig(dmat->common.maxsize, mflags,
0ul, dmat->common.lowaddr, dmat->common.alignment != 0 ?
dmat->common.alignment : 1ul, dmat->common.boundary, attr);
dmat->bounce_flags |= BF_KMEM_ALLOC;
@@ -495,7 +495,7 @@ bounce_bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
if ((dmat->bounce_flags & BF_KMEM_ALLOC) == 0)
free(vaddr, M_DEVBUF);
else
- kmem_free((vm_offset_t)vaddr, dmat->common.maxsize);
+ kmem_free(vaddr, dmat->common.maxsize);
free(map, M_DEVBUF);
dmat->map_count--;
CTR3(KTR_BUSDMA, "%s: tag %p flags 0x%x", __func__, dmat,
diff --git a/sys/riscv/riscv/mp_machdep.c b/sys/riscv/riscv/mp_machdep.c
index 2b441d49877b..5e8ae3db10e7 100644
--- a/sys/riscv/riscv/mp_machdep.c
+++ b/sys/riscv/riscv/mp_machdep.c
@@ -313,8 +313,7 @@ smp_after_idle_runnable(void *arg __unused)
for (cpu = 1; cpu <= mp_maxid; cpu++) {
if (bootstacks[cpu] != NULL)
- kmem_free((vm_offset_t)bootstacks[cpu],
- MP_BOOTSTACK_SIZE);
+ kmem_free(bootstacks[cpu], MP_BOOTSTACK_SIZE);
}
}
SYSINIT(smp_after_idle_runnable, SI_SUB_SMP, SI_ORDER_ANY,
@@ -475,11 +474,10 @@ cpu_init_fdt(u_int id, phandle_t node, u_int addr_size, pcell_t *reg)
pcpu_init(pcpup, cpuid, sizeof(struct pcpu));
pcpup->pc_hart = hart;
- dpcpu[cpuid - 1] = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
+ dpcpu[cpuid - 1] = kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
dpcpu_init(dpcpu[cpuid - 1], cpuid);
- bootstacks[cpuid] = (void *)kmem_malloc(MP_BOOTSTACK_SIZE,
- M_WAITOK | M_ZERO);
+ bootstacks[cpuid] = kmem_malloc(MP_BOOTSTACK_SIZE, M_WAITOK | M_ZERO);
naps = atomic_load_int(&aps_started);
bootstack = (char *)bootstacks[cpuid] + MP_BOOTSTACK_SIZE;
diff --git a/sys/riscv/riscv/pmap.c b/sys/riscv/riscv/pmap.c
index 56c571f73024..6aec89faf860 100644
--- a/sys/riscv/riscv/pmap.c
+++ b/sys/riscv/riscv/pmap.c
@@ -790,7 +790,7 @@ pmap_init(void)
*/
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
s = round_page(s);
- pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
+ pv_table = kmem_malloc(s, M_WAITOK | M_ZERO);
for (i = 0; i < pv_npg; i++)
TAILQ_INIT(&pv_table[i].pv_list);
TAILQ_INIT(&pv_dummy.pv_list);
diff --git a/sys/vm/uma_core.c b/sys/vm/uma_core.c
index 1351a54345d9..daa14cc3e3b3 100644
--- a/sys/vm/uma_core.c
+++ b/sys/vm/uma_core.c
@@ -1947,7 +1947,7 @@ page_alloc(uma_zone_t zone, vm_size_t bytes, int domain, uint8_t *pflag,
void *p; /* Returned page */
*pflag = UMA_SLAB_KERNEL;
- p = (void *)kmem_malloc_domainset(DOMAINSET_FIXED(domain), bytes, wait);
+ p = kmem_malloc_domainset(DOMAINSET_FIXED(domain), bytes, wait);
return (p);
}
@@ -2104,7 +2104,7 @@ page_free(void *mem, vm_size_t size, uint8_t flags)
KASSERT((flags & UMA_SLAB_KERNEL) != 0,
("UMA: page_free used with invalid flags %x", flags));
- kmem_free((vm_offset_t)mem, size);
+ kmem_free(mem, size);
}
/*
diff --git a/sys/vm/vm_extern.h b/sys/vm/vm_extern.h
index 680d13fbad09..107814c1178a 100644
--- a/sys/vm/vm_extern.h
+++ b/sys/vm/vm_extern.h
@@ -57,20 +57,20 @@ vm_offset_t kmap_alloc_wait(vm_map_t, vm_size_t);
void kmap_free_wakeup(vm_map_t, vm_offset_t, vm_size_t);
/* These operate on virtual addresses backed by memory. */
-vm_offset_t kmem_alloc_attr(vm_size_t size, int flags,
+void *kmem_alloc_attr(vm_size_t size, int flags,
vm_paddr_t low, vm_paddr_t high, vm_memattr_t memattr);
-vm_offset_t kmem_alloc_attr_domainset(struct domainset *ds, vm_size_t size,
+void *kmem_alloc_attr_domainset(struct domainset *ds, vm_size_t size,
int flags, vm_paddr_t low, vm_paddr_t high, vm_memattr_t memattr);
-vm_offset_t kmem_alloc_contig(vm_size_t size, int flags,
+void *kmem_alloc_contig(vm_size_t size, int flags,
vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
vm_memattr_t memattr);
-vm_offset_t kmem_alloc_contig_domainset(struct domainset *ds, vm_size_t size,
+void *kmem_alloc_contig_domainset(struct domainset *ds, vm_size_t size,
int flags, vm_paddr_t low, vm_paddr_t high, u_long alignment,
vm_paddr_t boundary, vm_memattr_t memattr);
-vm_offset_t kmem_malloc(vm_size_t size, int flags);
-vm_offset_t kmem_malloc_domainset(struct domainset *ds, vm_size_t size,
+void *kmem_malloc(vm_size_t size, int flags);
+void *kmem_malloc_domainset(struct domainset *ds, vm_size_t size,
int flags);
-void kmem_free(vm_offset_t addr, vm_size_t size);
+void kmem_free(void *addr, vm_size_t size);
/* This provides memory for previously allocated address space. */
int kmem_back(vm_object_t, vm_offset_t, vm_size_t, int);
diff --git a/sys/vm/vm_init.c b/sys/vm/vm_init.c
index 47b1eebbe139..c5a58c7a0ac7 100644
--- a/sys/vm/vm_init.c
+++ b/sys/vm/vm_init.c
@@ -150,8 +150,7 @@ vm_mem_init(void *dummy)
void
vm_ksubmap_init(struct kva_md_info *kmi)
{
- vm_offset_t firstaddr;
- caddr_t v;
+ caddr_t firstaddr, v;
vm_size_t size = 0;
long physmem_est;
vm_offset_t minaddr;
@@ -170,9 +169,9 @@ vm_ksubmap_init(struct kva_md_info *kmi)
* needed and allocates it. The second pass assigns virtual
* addresses to the various data structures.
*/
- firstaddr = 0;
+ firstaddr = NULL;
again:
- v = (caddr_t)firstaddr;
+ v = firstaddr;
/*
* Discount the physical memory larger than the size of kernel_map
@@ -186,7 +185,7 @@ again:
/*
* End of first pass, size has been calculated so allocate memory
*/
- if (firstaddr == 0) {
+ if (firstaddr == NULL) {
size = (vm_size_t)v;
#ifdef VM_FREELIST_DMA32
/*
@@ -195,10 +194,10 @@ again:
*/
firstaddr = kmem_alloc_attr(size, M_ZERO | M_NOWAIT,
(vm_paddr_t)1 << 32, ~(vm_paddr_t)0, VM_MEMATTR_DEFAULT);
- if (firstaddr == 0)
+ if (firstaddr == NULL)
#endif
firstaddr = kmem_malloc(size, M_ZERO | M_WAITOK);
- if (firstaddr == 0)
+ if (firstaddr == NULL)
panic("startup: no room for tables");
goto again;
}
@@ -206,15 +205,15 @@ again:
/*
* End of second pass, addresses have been assigned
*/
- if ((vm_size_t)((char *)v - firstaddr) != size)
+ if ((vm_size_t)(v - firstaddr) != size)
panic("startup: table size inconsistency");
/*
* Allocate the clean map to hold all of I/O virtual memory.
*/
size = (long)nbuf * BKVASIZE + (long)bio_transient_maxcnt * maxphys;
- kmi->clean_sva = firstaddr = kva_alloc(size);
- kmi->clean_eva = firstaddr + size;
+ kmi->clean_sva = kva_alloc(size);
+ kmi->clean_eva = kmi->clean_sva + size;
/*
* Allocate the buffer arena.
@@ -223,11 +222,10 @@ again:
* avoids lock contention at the expense of some fragmentation.
*/
size = (long)nbuf * BKVASIZE;
- kmi->buffer_sva = firstaddr;
+ kmi->buffer_sva = kmi->clean_sva;
kmi->buffer_eva = kmi->buffer_sva + size;
vmem_init(buffer_arena, "buffer arena", kmi->buffer_sva, size,
PAGE_SIZE, (mp_ncpus > 4) ? BKVASIZE * 8 : 0, M_WAITOK);
- firstaddr += size;
/*
* And optionally transient bio space.
@@ -235,11 +233,8 @@ again:
if (bio_transient_maxcnt != 0) {
size = (long)bio_transient_maxcnt * maxphys;
vmem_init(transient_arena, "transient arena",
- firstaddr, size, PAGE_SIZE, 0, M_WAITOK);
- firstaddr += size;
+ kmi->buffer_eva, size, PAGE_SIZE, 0, M_WAITOK);
}
- if (firstaddr != kmi->clean_eva)
- panic("Clean map calculation incorrect");
/*
* Allocate the pageable submaps. We may cache an exec map entry per
diff --git a/sys/vm/vm_kern.c b/sys/vm/vm_kern.c
index 7a61a29c81a4..10fe07cc86da 100644
--- a/sys/vm/vm_kern.c
+++ b/sys/vm/vm_kern.c
@@ -229,7 +229,7 @@ kmem_alloc_contig_pages(vm_object_t object, vm_pindex_t pindex, int domain,
* necessarily physically contiguous. If M_ZERO is specified through the
* given flags, then the pages are zeroed before they are mapped.
*/
-static vm_offset_t
+static void *
kmem_alloc_attr_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
vm_paddr_t high, vm_memattr_t memattr)
{
@@ -270,10 +270,10 @@ kmem_alloc_attr_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
}
VM_OBJECT_WUNLOCK(object);
kmem_alloc_san(addr, size, asize, flags);
- return (addr);
+ return ((void *)addr);
}
-vm_offset_t
+void *
kmem_alloc_attr(vm_size_t size, int flags, vm_paddr_t low, vm_paddr_t high,
vm_memattr_t memattr)
{
@@ -282,19 +282,19 @@ kmem_alloc_attr(vm_size_t size, int flags, vm_paddr_t low, vm_paddr_t high,
high, memattr));
}
-vm_offset_t
+void *
kmem_alloc_attr_domainset(struct domainset *ds, vm_size_t size, int flags,
vm_paddr_t low, vm_paddr_t high, vm_memattr_t memattr)
{
struct vm_domainset_iter di;
- vm_offset_t addr;
+ void *addr;
int domain;
vm_domainset_iter_policy_init(&di, ds, &domain, &flags);
do {
addr = kmem_alloc_attr_domain(domain, size, flags, low, high,
memattr);
- if (addr != 0)
+ if (addr != NULL)
break;
} while (vm_domainset_iter_policy(&di, &domain) == 0);
@@ -309,7 +309,7 @@ kmem_alloc_attr_domainset(struct domainset *ds, vm_size_t size, int flags,
* through the given flags, then the pages are zeroed before they are
* mapped.
*/
-static vm_offset_t
+static void *
kmem_alloc_contig_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
vm_memattr_t memattr)
@@ -326,7 +326,7 @@ kmem_alloc_contig_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
asize = round_page(size);
vmem = vm_dom[domain].vmd_kernel_arena;
if (vmem_alloc(vmem, asize, flags | M_BESTFIT, &addr))
- return (0);
+ return (NULL);
offset = addr - VM_MIN_KERNEL_ADDRESS;
pflags = malloc2vm_flags(flags) | VM_ALLOC_WIRED;
npages = atop(asize);
@@ -336,7 +336,7 @@ kmem_alloc_contig_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
if (m == NULL) {
VM_OBJECT_WUNLOCK(object);
vmem_free(vmem, addr, asize);
- return (0);
+ return (NULL);
}
KASSERT(vm_page_domain(m) == domain,
("kmem_alloc_contig_domain: Domain mismatch %d != %d",
@@ -353,10 +353,10 @@ kmem_alloc_contig_domain(int domain, vm_size_t size, int flags, vm_paddr_t low,
}
VM_OBJECT_WUNLOCK(object);
kmem_alloc_san(addr, size, asize, flags);
- return (addr);
+ return ((void *)addr);
}
-vm_offset_t
+void *
kmem_alloc_contig(vm_size_t size, int flags, vm_paddr_t low, vm_paddr_t high,
u_long alignment, vm_paddr_t boundary, vm_memattr_t memattr)
{
@@ -365,20 +365,20 @@ kmem_alloc_contig(vm_size_t size, int flags, vm_paddr_t low, vm_paddr_t high,
high, alignment, boundary, memattr));
}
-vm_offset_t
+void *
kmem_alloc_contig_domainset(struct domainset *ds, vm_size_t size, int flags,
vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
vm_memattr_t memattr)
{
struct vm_domainset_iter di;
- vm_offset_t addr;
+ void *addr;
int domain;
vm_domainset_iter_policy_init(&di, ds, &domain, &flags);
do {
addr = kmem_alloc_contig_domain(domain, size, flags, low, high,
alignment, boundary, memattr);
- if (addr != 0)
+ if (addr != NULL)
break;
} while (vm_domainset_iter_policy(&di, &domain) == 0);
@@ -423,7 +423,7 @@ kmem_subinit(vm_map_t map, vm_map_t parent, vm_offset_t *min, vm_offset_t *max,
*
* Allocate wired-down pages in the kernel's address space.
*/
-static vm_offset_t
+static void *
kmem_malloc_domain(int domain, vm_size_t size, int flags)
{
vmem_t *arena;
@@ -445,27 +445,27 @@ kmem_malloc_domain(int domain, vm_size_t size, int flags)
return (0);
}
kasan_mark((void *)addr, size, asize, KASAN_KMEM_REDZONE);
- return (addr);
+ return ((void *)addr);
}
-vm_offset_t
+void *
kmem_malloc(vm_size_t size, int flags)
{
return (kmem_malloc_domainset(DOMAINSET_RR(), size, flags));
}
-vm_offset_t
+void *
kmem_malloc_domainset(struct domainset *ds, vm_size_t size, int flags)
{
struct vm_domainset_iter di;
- vm_offset_t addr;
+ void *addr;
int domain;
vm_domainset_iter_policy_init(&di, ds, &domain, &flags);
do {
addr = kmem_malloc_domain(domain, size, flags);
- if (addr != 0)
+ if (addr != NULL)
break;
} while (vm_domainset_iter_policy(&di, &domain) == 0);
@@ -631,15 +631,15 @@ kmem_unback(vm_object_t object, vm_offset_t addr, vm_size_t size)
* original allocation.
*/
void
-kmem_free(vm_offset_t addr, vm_size_t size)
+kmem_free(void *addr, vm_size_t size)
{
struct vmem *arena;
size = round_page(size);
- kasan_mark((void *)addr, size, size, 0);
- arena = _kmem_unback(kernel_object, addr, size);
+ kasan_mark(addr, size, size, 0);
+ arena = _kmem_unback(kernel_object, (uintptr_t)addr, size);
if (arena != NULL)
- vmem_free(arena, addr, size);
+ vmem_free(arena, (uintptr_t)addr, size);
}
/*
diff --git a/sys/x86/iommu/intel_dmar.h b/sys/x86/iommu/intel_dmar.h
index 1234ee058ffd..21f56bcdefb4 100644
--- a/sys/x86/iommu/intel_dmar.h
+++ b/sys/x86/iommu/intel_dmar.h
@@ -158,7 +158,7 @@ struct dmar_unit {
/* QI */
int qi_enabled;
- vm_offset_t inv_queue;
+ char *inv_queue;
vm_size_t inv_queue_size;
uint32_t inv_queue_avail;
uint32_t inv_queue_tail;
diff --git a/sys/x86/iommu/intel_intrmap.c b/sys/x86/iommu/intel_intrmap.c
index c926f7a91ec7..acbf4b74b843 100644
--- a/sys/x86/iommu/intel_intrmap.c
+++ b/sys/x86/iommu/intel_intrmap.c
@@ -342,9 +342,9 @@ dmar_init_irt(struct dmar_unit *unit)
return (0);
}
unit->irte_cnt = clp2(num_io_irqs);
- unit->irt = (dmar_irte_t *)(uintptr_t)kmem_alloc_contig(
- unit->irte_cnt * sizeof(dmar_irte_t), M_ZERO | M_WAITOK, 0,
- dmar_high, PAGE_SIZE, 0, DMAR_IS_COHERENT(unit) ?
+ unit->irt = kmem_alloc_contig(unit->irte_cnt * sizeof(dmar_irte_t),
+ M_ZERO | M_WAITOK, 0, dmar_high, PAGE_SIZE, 0,
+ DMAR_IS_COHERENT(unit) ?
VM_MEMATTR_DEFAULT : VM_MEMATTR_UNCACHEABLE);
if (unit->irt == NULL)
return (ENOMEM);
@@ -378,7 +378,6 @@ dmar_fini_irt(struct dmar_unit *unit)
dmar_disable_ir(unit);
dmar_qi_invalidate_iec_glob(unit);
vmem_destroy(unit->irtids);
- kmem_free((vm_offset_t)unit->irt, unit->irte_cnt *
- sizeof(dmar_irte_t));
+ kmem_free(unit->irt, unit->irte_cnt * sizeof(dmar_irte_t));
}
}
diff --git a/sys/x86/iommu/intel_qi.c b/sys/x86/iommu/intel_qi.c
index 8a8e656083e3..89c9fb16485a 100644
--- a/sys/x86/iommu/intel_qi.c
+++ b/sys/x86/iommu/intel_qi.c
@@ -510,7 +510,7 @@ dmar_init_qi(struct dmar_unit *unit)
DMAR_LOCK(unit);
dmar_write8(unit, DMAR_IQT_REG, 0);
- iqa = pmap_kextract(unit->inv_queue);
+ iqa = pmap_kextract((uintptr_t)unit->inv_queue);
iqa |= qi_sz;
dmar_write8(unit, DMAR_IQA_REG, iqa);
dmar_enable_qi(unit);
@@ -552,7 +552,7 @@ dmar_fini_qi(struct dmar_unit *unit)
DMAR_UNLOCK(unit);
kmem_free(unit->inv_queue, unit->inv_queue_size);
- unit->inv_queue = 0;
+ unit->inv_queue = NULL;
unit->inv_queue_size = 0;
unit->qi_enabled = 0;
}
diff --git a/sys/x86/x86/busdma_bounce.c b/sys/x86/x86/busdma_bounce.c
index 925a1d633f7c..faed4b7353cc 100644
--- a/sys/x86/x86/busdma_bounce.c
+++ b/sys/x86/x86/busdma_bounce.c
@@ -449,12 +449,12 @@ bounce_bus_dmamem_alloc(bus_dma_tag_t dmat, void **vaddr, int flags,
dmat->common.alignment <= PAGE_SIZE &&
(dmat->common.boundary % PAGE_SIZE) == 0) {
/* Page-based multi-segment allocations allowed */
- *vaddr = (void *)kmem_alloc_attr_domainset(
+ *vaddr = kmem_alloc_attr_domainset(
DOMAINSET_PREF(dmat->common.domain), dmat->common.maxsize,
mflags, 0ul, dmat->common.lowaddr, attr);
dmat->bounce_flags |= BUS_DMA_KMEM_ALLOC;
} else {
- *vaddr = (void *)kmem_alloc_contig_domainset(
+ *vaddr = kmem_alloc_contig_domainset(
DOMAINSET_PREF(dmat->common.domain), dmat->common.maxsize,
mflags, 0ul, dmat->common.lowaddr,
dmat->common.alignment != 0 ? dmat->common.alignment : 1ul,
@@ -490,7 +490,7 @@ bounce_bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
if ((dmat->bounce_flags & BUS_DMA_KMEM_ALLOC) == 0)
free(vaddr, M_DEVBUF);
else
- kmem_free((vm_offset_t)vaddr, dmat->common.maxsize);
+ kmem_free(vaddr, dmat->common.maxsize);
CTR3(KTR_BUSDMA, "%s: tag %p flags 0x%x", __func__, dmat,
dmat->bounce_flags);
}
diff --git a/sys/x86/x86/mp_x86.c b/sys/x86/x86/mp_x86.c
index 6dab7987e208..61b08dd1ee97 100644
--- a/sys/x86/x86/mp_x86.c
+++ b/sys/x86/x86/mp_x86.c
@@ -1154,8 +1154,7 @@ smp_after_idle_runnable(void *arg __unused)
smp_no_rendezvous_barrier, NULL);
for (cpu = 1; cpu < mp_ncpus; cpu++) {
- kmem_free((vm_offset_t)bootstacks[cpu], kstack_pages *
- PAGE_SIZE);
+ kmem_free(bootstacks[cpu], kstack_pages * PAGE_SIZE);
}
}
SYSINIT(smp_after_idle_runnable, SI_SUB_SMP, SI_ORDER_ANY,