Lines Matching refs:tx

52 static boolean_t dbuf_undirty(dmu_buf_impl_t *db, dmu_tx_t *tx);
53 static void dbuf_write(dbuf_dirty_record_t *dr, arc_buf_t *data, dmu_tx_t *tx);
1255 dmu_tx_t *tx)
1259 uint64_t txg = tx->tx_txg;
1288 if (dbuf_undirty(db, tx)) {
1385 dbuf_new_size(dmu_buf_impl_t *db, int size, dmu_tx_t *tx)
1410 dmu_buf_will_dirty(&db->db, tx);
1428 ASSERT3U(db->db_last_dirty->dr_txg, ==, tx->tx_txg);
1433 dnode_willuse_space(dn, size-osize, tx);
1477 dbuf_dirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
1484 int txgoff = tx->tx_txg & TXG_MASK;
1486 ASSERT(tx->tx_txg != 0);
1488 DMU_TX_DIRTY_BUF(tx, db);
1502 ASSERT(!dmu_tx_is_syncing(tx) ||
1516 (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN));
1539 dn->dn_dirtyctx = (dmu_tx_is_syncing(tx) ?
1558 ASSERT(*drp == NULL || (*drp)->dr_txg <= tx->tx_txg ||
1560 while ((dr = *drp) != NULL && dr->dr_txg > tx->tx_txg)
1562 if (dr && dr->dr_txg == tx->tx_txg) {
1575 (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN));
1581 dn->dn_next_nlevels[(tx->tx_txg-1) & TXG_MASK] > db->db_level ||
1582 dn->dn_next_nlevels[(tx->tx_txg-2) & TXG_MASK] > db->db_level);
1595 ASSERT(!dmu_tx_is_syncing(tx) || DMU_OBJECT_IS_SPECIAL(dn->dn_object) ||
1612 dnode_willuse_space(dn, db->db.db_size, tx);
1627 dbuf_fix_old_data(db, tx->tx_txg);
1640 dbuf_fix_old_data(db, tx->tx_txg);
1655 dr->dr_txg = tx->tx_txg;
1678 dbuf_add_ref(db, (void *)(uintptr_t)tx->tx_txg);
1690 dnode_setdirty(dn, tx);
1719 dnode_willuse_space(dn, -willfree, tx);
1723 dnode_new_blkid(dn, db->db_blkid, tx, drop_struct_lock);
1743 di = dbuf_dirty(parent, tx);
1755 ASSERT3U(di->dr_txg, ==, tx->tx_txg);
1774 dnode_setdirty(dn, tx);
1784 dbuf_undirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
1787 uint64_t txg = tx->tx_txg;
1871 dmu_buf_will_dirty(dmu_buf_t *db_fake, dmu_tx_t *tx)
1876 ASSERT(tx->tx_txg != 0);
1888 dr != NULL && dr->dr_txg >= tx->tx_txg; dr = dr->dr_next) {
1894 if (dr->dr_txg == tx->tx_txg && db->db_state == DB_CACHED) {
1908 (void) dbuf_dirty(db, tx);
1912 dmu_buf_will_not_fill(dmu_buf_t *db_fake, dmu_tx_t *tx)
1918 dmu_buf_will_fill(db_fake, tx);
1922 dmu_buf_will_fill(dmu_buf_t *db_fake, dmu_tx_t *tx)
1927 ASSERT(tx->tx_txg != 0);
1932 dmu_tx_private_ok(tx));
1935 (void) dbuf_dirty(db, tx);
1941 dbuf_fill_done(dmu_buf_impl_t *db, dmu_tx_t *tx)
1964 dmu_tx_t *tx)
1982 dmu_buf_will_not_fill(dbuf, tx);
1984 ASSERT3U(db->db_last_dirty->dr_txg, ==, tx->tx_txg);
2002 dbuf_assign_arcbuf(dmu_buf_impl_t *db, arc_buf_t *buf, dmu_tx_t *tx)
2010 ASSERT(tx->tx_txg != 0);
2025 (void) dbuf_dirty(db, tx);
2037 if (dr != NULL && dr->dr_txg == tx->tx_txg) {
2056 (void) dbuf_dirty(db, tx);
2057 dmu_buf_fill_done(&db->db, tx);
2659 dbuf_spill_set_blksz(dmu_buf_t *db_fake, uint64_t blksz, dmu_tx_t *tx)
2674 dbuf_new_size(db, blksz, tx);
2682 dbuf_rm_spill(dnode_t *dn, dmu_tx_t *tx)
2684 dbuf_free_range(dn, DMU_SPILL_BLKID, DMU_SPILL_BLKID, tx);
2957 /* ASSERT(dmu_tx_is_syncing(tx) */
3000 dbuf_sync_indirect(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
3006 ASSERT(dmu_tx_is_syncing(tx));
3035 dbuf_write(dr, db->db_buf, tx);
3039 dbuf_sync_list(&dr->dt.di.dr_children, db->db_level - 1, tx);
3046 dbuf_sync_leaf(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
3052 uint64_t txg = tx->tx_txg;
3054 ASSERT(dmu_tx_is_syncing(tx));
3170 dbuf_write(dr, *datap, tx);
3190 dbuf_sync_list(list_t *list, int level, dmu_tx_t *tx)
3213 dbuf_sync_indirect(dr, tx);
3215 dbuf_sync_leaf(dr, tx);
3387 dmu_tx_t *tx = os->os_synctx;
3401 (void) dsl_dataset_block_kill(ds, bp_orig, tx, B_TRUE);
3402 dsl_dataset_block_born(ds, bp, tx);
3462 dbuf_rele_and_unlock(db, (void *)(uintptr_t)tx->tx_txg);
3506 dbuf_write(dbuf_dirty_record_t *dr, arc_buf_t *data, dmu_tx_t *tx)
3512 uint64_t txg = tx->tx_txg;
3518 ASSERT(dmu_tx_is_syncing(tx));