Lines Matching refs:len
37 * rl = zfs_range_lock(zp, off, len, lock_type);
39 * zfs_range_reduce(rl, off, len);
111 uint64_t len = new->r_len;
138 end_size = MAX(zp->z_size, new->r_off + len);
182 new->r_len = len;
254 zfs_range_new_proxy(avl_tree_t *tree, uint64_t off, uint64_t len)
258 ASSERT(len);
261 rl->r_len = len;
275 uint64_t len = new->r_len;
303 if (next == NULL || off + len <= next->r_off) {
322 if (off + len <= next->r_off)
330 if (off + len == next->r_off + next->r_len) {
336 if (off + len < next->r_off + next->r_len) {
338 next = zfs_range_split(tree, next, off + len);
342 ASSERT3U(off + len, >, next->r_off + next->r_len);
349 (off + len) - (prev->r_off + prev->r_len));
362 uint64_t len = new->r_len;
384 if (off + len < prev->r_off + prev->r_len)
397 if (off + len <= next->r_off)
407 if (off + len <= next->r_off + next->r_len)
426 zfs_range_lock(znode_t *zp, uint64_t off, uint64_t len, rl_type_t type)
435 if (len + off < off) /* overflow */
436 len = UINT64_MAX - off;
437 new->r_len = len;
467 uint64_t len;
499 for (len = remove->r_len; len != 0; rl = next) {
500 len -= rl->r_len;
501 if (len) {
568 zfs_range_reduce(rl_t *rl, uint64_t off, uint64_t len)
582 rl->r_len = len;