Lines Matching refs:len

304 dmareq_mapin(offset_t len, caddr_t addr, struct as *as, int flag)
314 buf.b_bcount = (size_t)len;
325 dmareq_pp_mapin(offset_t len, uint_t offset, page_t *pp, int flag)
334 buf.b_bcount = (size_t)len;
344 dmareq_pplist_mapin(uint_t len, caddr_t addr, page_t **pplist, struct as *as,
355 buf.b_bcount = len;
406 ddi_dmareq_mapout(caddr_t addr, offset_t len, int map_flags, page_t *pp,
418 buf.b_bcount = (size_t)len;
959 errdef.len = errdef_32.len;
1049 errdef_32.len = errdef.len;
1525 if (hp->len == INT_MAX - hp->offset)
1534 " len 0x%llx\n",
1536 hp->len);
1540 "handle no %d len 0x%llx"
1542 hp->len, (void *)hp->addr);
1635 hdlip->len = hp->len;
1645 hdlip->len = hp->len;
1786 ep->errdef.len) & ~LLSZMASK) >
1795 (errdefp->len == 0 ||
1796 hp->offset < errdefp->offset + errdefp->len) &&
1797 hp->offset + hp->len > errdefp->offset))) {
2246 log_acc_event(struct bofi_errent *ep, uint_t at, offset_t offset, off_t len,
2269 elem->size = len;
2298 off_t len;
2321 off < ep->errdef.offset + ep->errdef.len)) {
2342 * Make sure endaddr isn't greater than hp->addr + hp->len.
2343 * If endaddr becomes less than addr len becomes negative
2348 endaddr = (uint64_t *)((uintptr_t)(hp->addr + min(hp->len,
2349 ep->errdef.offset + ep->errdef.len)) & ~LLSZMASK);
2350 len = endaddr - addr;
2367 for (i = 0; i < len; i++)
2371 for (i = 0; i < len; i++)
2375 for (i = 0; i < len; i++)
2379 for (i = 0; i < len; i++)
2426 if (ep->errdef.len == 0)
2427 minlen = hp->len;
2429 minlen = min(hp->len, ep->errdef.len);
2542 if (ep->errdef.len == 0)
2543 minlen = hp->len;
2545 minlen = min(hp->len, ep->errdef.len);
2626 (caddr_t)addr - hp->addr >= hp->len)) { \
2629 (void *)addr, (void *)hp->addr, hp->len); \
2737 (caddr_t)addr - hp->addr >= hp->len)) { \
2740 (void *)addr, (void *)hp->addr, hp->len); \
2831 (caddr_t)(dev_addr + repcount) - hp->addr > hp->len)) { \
2834 (void *)dev_addr, (void *)hp->addr, hp->len); \
2836 (caddr_t)dev_addr - hp->addr >= hp->len) \
2838 repcount = (type *)(hp->addr + hp->len) - dev_addr; \
2952 (caddr_t)(dev_addr + repcount) - hp->addr > hp->len)) { \
2955 (void *)dev_addr, (void *)hp->addr, hp->len); \
2957 (caddr_t)dev_addr - hp->addr >= hp->len) \
2959 repcount = (type *)(hp->addr + hp->len) - dev_addr; \
3091 ddi_map_req_t *reqp, off_t offset, off_t len, caddr_t *vaddrp)
3106 retval = save_bus_ops.bus_map(dip, rdip, reqp, offset, len,
3137 if (len == 0)
3138 hp->len = INT_MAX - offset;
3140 hp->len = min(len, INT_MAX - offset);
3203 (ep->errdef.len == 0 ||
3204 offset < ep->errdef.offset + ep->errdef.len) &&
3205 offset + hp->len > ep->errdef.offset) {
3272 return (save_bus_ops.bus_map(dip, rdip, reqp, offset, len, vaddrp));
3298 ep->errdef.len) & ~LLSZMASK) >
3315 0, hp->len);
3327 xbcopy(void *from, void *to, u_longlong_t len)
3332 while (len--)
3554 hp->addr = ddi_dmareq_mapin(dmareqp, &hp->mapaddr, &hp->len);
3567 ((uintptr_t)hp->addr & pagemask) + hp->len,
3574 xbcopy(hp->origaddr, hp->addr, hp->len);
3576 dmareq.dmar_object.dmao_size = hp->len;
3618 ddi_dmareq_mapout(hp->mapaddr, hp->len, hp->map_flags,
3681 do_dma_corrupt(hp, ep, DDI_DMA_SYNC_FORCPU, 0, hp->len);
3697 xbcopy(hp->addr, hp->origaddr, hp->len);
3698 ddi_dmareq_mapout(hp->mapaddr, hp->len, hp->map_flags,
3714 ddi_dma_handle_t handle, off_t off, size_t len, uint_t flags)
3727 len, flags);
3750 len ? len : (hp->len - off));
3765 len ? len : (hp->len - off));
3777 len ? len : (hp->len - off));
3784 len, flags);
3895 for (i = 0; i < dummyhp->len; i++) {
3918 kmem_free(dummyhp->hparrayp, dummyhp->len *
3968 dummyhp->len = count;
4043 bofi_dvma_kaddr_load(ddi_dma_handle_t h, caddr_t a, uint_t len, uint_t index,
4091 hp->len = len;
4095 dummyhp->save.dvma_ops.dvma_kaddr_load(h, a, len, index, cp);
4111 ep->errdef.len) & ~LLSZMASK) >
4180 do_dma_corrupt(hp, ep, view, 0, hp->len);
4196 xbcopy(hp->addr, hp->origaddr, hp->len);
4248 xbcopy(hp->origaddr, hp->addr, hp->len);
4263 do_dma_corrupt(hp, ep, view, 0, hp->len);
4274 xbcopy(hp->addr, hp->origaddr, hp->len);