Lines Matching refs:cbuf

624 	md_raidcbuf_t		*cbuf;
649 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next) {
650 if (cbuf->cbuf_pwslot == -1)
652 col = &un->un_column_ic[cbuf->cbuf_column];
654 sb = &pw->pw_sb[cbuf->cbuf_pwslot];
673 md_raidcbuf_t *cbuf;
681 cbuf = cs->cs_pw_inval_list;
682 ASSERT(cbuf);
684 while (cbuf) {
685 pw = un->un_column_ic[cbuf->cbuf_column].un_pw_reserve;
687 ASSERT(sb[cbuf->cbuf_pwslot].sb_flags & SB_INVAL_PEND);
688 sb[cbuf->cbuf_pwslot].sb_flags = SB_UNUSED;
689 sb[cbuf->cbuf_pwslot].sb_cs = NULL;
692 cbuf_to_free = cbuf;
693 cbuf = cbuf->cbuf_next;
817 md_raidcbuf_t *cbuf;
829 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next) {
830 cbuf->cbuf_pwslot = raid_get_pwslot(cs, cbuf->cbuf_column);
1249 md_raidcbuf_t *cbuf, *cbuf1;
1264 cbuf = cs->cs_buflist;
1265 while (cbuf) {
1266 cbuf1 = cbuf->cbuf_next;
1267 kmem_cache_free(raid_cbuf_cache, cbuf);
1268 cbuf = cbuf1;
1679 md_raidcbuf_t *cbuf;
1687 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next)
1688 cbuf->cbuf_buffer =
1700 md_raidcbuf_t *cbuf;
1703 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next) {
1704 if (cbuf->cbuf_buffer == NULL)
1706 kmem_free(cbuf->cbuf_buffer, cbuf->cbuf_bcount + DEV_BSIZE);
1707 cbuf->cbuf_buffer = NULL;
1708 cbuf->cbuf_bcount = 0;
2174 md_raidcbuf_t *cbuf;
2193 cbuf = kmem_cache_alloc(raid_cbuf_cache,
2195 raid_cbuf_init(cbuf);
2196 cbuf->cbuf_un = cs->cs_un;
2197 cbuf->cbuf_ps = cs->cs_ps;
2198 cbuf->cbuf_column = j;
2199 cbuf->cbuf_bcount = dbtob(un->un_segsize);
2200 *cbufp = cbuf;
2201 cbufp = &cbuf->cbuf_next;
2446 md_raidcbuf_t *cbuf;
2502 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next) {
2506 dbuf = (uint_t *)(void *)(cbuf->cbuf_buffer + DEV_BSIZE);
2510 col = cbuf->cbuf_column;
2520 if (cbuf->cbuf_next == NULL) {
2550 if (cbuf->cbuf_next == NULL) {
2577 RAID_FILLIN_RPW(cbuf->cbuf_buffer, un, dsum, cs->cs_pcolumn,
2584 bp = &cbuf->cbuf_bp;
2585 bp->b_un.b_addr = cbuf->cbuf_buffer;
2586 bp->b_bcount = cbuf->cbuf_bcount + DEV_BSIZE;
2588 bp->b_lblkno = (cbuf->cbuf_pwslot * un->un_iosize) +
2904 md_raidcbuf_t *cbuf;
2918 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next)
2919 if (cbuf->cbuf_bp.b_flags & B_ERROR)
2920 (void) raid_error_state(un, &cbuf->cbuf_bp);
2943 md_raidcbuf_t *cbuf;
2970 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next)
2971 if ((cbuf->cbuf_bp.b_flags & B_ERROR) &&
2972 (COLUMN_STATE(un, cbuf->cbuf_column) != RCS_ERRED) &&
2973 (COLUMN_STATE(un, cbuf->cbuf_column) != RCS_LAST_ERRED))
2976 md_devname(setno, md_expldev(cbuf->cbuf_bp.b_edev),
3379 md_raidcbuf_t *cbuf;
3444 for (cbuf = cs->cs_buflist; cbuf; cbuf = cbuf->cbuf_next) {
3448 bp = &cbuf->cbuf_bp;
3451 bp->b_un.b_addr = cbuf->cbuf_buffer + DEV_BSIZE;
3452 bp->b_bcount = cbuf->cbuf_bcount;
3453 bp->b_bufsize = cbuf->cbuf_bcount;
3455 un->un_column[cbuf->cbuf_column].un_devstart +
3462 un->un_column[cbuf->cbuf_column].un_dev);