/illumos-gate/usr/src/uts/common/fs/zfs/sys/ |
H A D | dsl_pool.h | 144 void dsl_pool_dirty_space(dsl_pool_t *dp, int64_t space, dmu_tx_t *tx); 149 void dsl_pool_create_origin(dsl_pool_t *dp, dmu_tx_t *tx); 150 void dsl_pool_upgrade_clones(dsl_pool_t *dp, dmu_tx_t *tx); 151 void dsl_pool_upgrade_dir_clones(dsl_pool_t *dp, dmu_tx_t *tx); 164 const char *tag, uint64_t now, dmu_tx_t *tx); 166 const char *tag, dmu_tx_t *tx);
|
H A D | zfs_znode.h | 69 #define ZFS_ATTR_SET(zp, attr, value, pflags, tx) \ 76 &pflags, sizeof (pflags), tx)); \ 288 dmu_tx_t *tx); 309 extern void zfs_log_create(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 314 extern void zfs_log_remove(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 317 extern void zfs_log_link(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 319 extern void zfs_log_symlink(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 321 extern void zfs_log_rename(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 323 extern void zfs_log_write(zilog_t *zilog, dmu_tx_t *tx, int txtype, 325 extern void zfs_log_truncate(zilog_t *zilog, dmu_tx_t *tx, in [all...] |
H A D | dsl_deleg.h | 70 void dsl_deleg_set_create_perms(dsl_dir_t *dd, dmu_tx_t *tx, cred_t *cr); 73 int dsl_deleg_destroy(objset_t *os, uint64_t zapobj, dmu_tx_t *tx);
|
/illumos-gate/usr/src/uts/common/fs/zfs/ |
H A D | spa_errlog.c | 302 sync_error_list(spa_t *spa, avl_tree_t *t, uint64_t *obj, dmu_tx_t *tx) argument 313 0, tx); 322 *obj, buf, 1, strlen(name) + 1, name, tx); 345 dmu_tx_t *tx; local 368 tx = dmu_tx_create_assigned(spa->spa_dsl_pool, txg); 373 sync_error_list(spa, &last, &spa->spa_errlog_last, tx); 381 spa->spa_errlog_last, tx) == 0); 385 sync_error_list(spa, &scrub, &spa->spa_errlog_last, tx); 391 sync_error_list(spa, &scrub, &spa->spa_errlog_scrub, tx); 398 &spa->spa_errlog_last, tx); [all...] |
H A D | zvol.c | 142 * zvol maximum transfer in one DMU tx. 336 zvol_create_cb(objset_t *os, void *arg, cred_t *cr, dmu_tx_t *tx) argument 359 DMU_OT_NONE, 0, tx); 363 DMU_OT_NONE, 0, tx); 366 error = zap_update(os, ZVOL_ZAP_OBJ, "size", 8, 1, &volsize, tx); 398 dmu_tx_t *tx; local 416 tx = dmu_tx_create(os); 417 dmu_tx_hold_write(tx, ZVOL_OBJ, offset, length); 418 error = dmu_tx_assign(tx, TXG_WAIT); 420 dmu_tx_abort(tx); 687 dmu_tx_t *tx; local 725 dmu_tx_t *tx; local 1056 zvol_log_write(zvol_state_t *zv, dmu_tx_t *tx, offset_t off, ssize_t resid, boolean_t sync) argument 1296 dmu_tx_t *tx = dmu_tx_create(os); local 1459 dmu_tx_t *tx = dmu_tx_create(zv->zv_objset); local 1602 zvol_log_write_minor(void *minor_hdl, dmu_tx_t *tx, offset_t off, ssize_t resid, boolean_t sync) argument 1617 zvol_log_truncate(zvol_state_t *zv, dmu_tx_t *tx, uint64_t off, uint64_t len, boolean_t sync) argument 1783 dmu_tx_t *tx; local 1870 zfs_mvdev_dump_feature_check(void *arg, dmu_tx_t *tx) argument 1881 zfs_mvdev_dump_activate_feature_sync(void *arg, dmu_tx_t *tx) argument 1891 dmu_tx_t *tx; local 2052 dmu_tx_t *tx; local 2103 dmu_tx_t *tx; local [all...] |
H A D | zfs_vnops.c | 109 * cached atime changes. Third, zfs_zinactive() may require a new tx, 111 * If you must call VN_RELE() within a tx then use VN_RELE_ASYNC(). 126 * the tx assigns, and sometimes after (e.g. z_lock), then failing 130 * Thread B is in an already-assigned tx, and blocks for this lock. 132 * forever, because the previous txg can't quiesce until B's tx commits. 147 * (6) At the end of each vnode op, the DMU tx must always commit, 159 * tx = dmu_tx_create(...); // get DMU tx 161 * error = dmu_tx_assign(tx, waited ? TXG_WAITED : TXG_NOWAIT); 168 * dmu_tx_wait(tx); 659 dmu_tx_t *tx; local 1384 dmu_tx_t *tx; local 1641 dmu_tx_t *tx; local 1875 dmu_tx_t *tx; local 2057 dmu_tx_t *tx; local 2709 dmu_tx_t *tx; local 3448 dmu_tx_t *tx; local 3812 dmu_tx_t *tx; local 4009 dmu_tx_t *tx; local 4180 dmu_tx_t *tx; local 4419 dmu_tx_t *tx = dmu_tx_create(zfsvfs->z_os); local [all...] |
H A D | dmu.c | 229 dmu_set_bonus(dmu_buf_t *db_fake, int newsize, dmu_tx_t *tx) argument 243 dnode_setbonuslen(dn, newsize, tx); 252 dmu_set_bonustype(dmu_buf_t *db_fake, dmu_object_type_t type, dmu_tx_t *tx) argument 266 dnode_setbonus_type(dn, type, tx); 290 dmu_rm_spill(objset_t *os, uint64_t object, dmu_tx_t *tx) argument 296 dbuf_rm_spill(dn, tx); 298 dnode_rm_spill(dn, tx); 742 dmu_tx_t *tx = dmu_tx_create(os); local 743 dmu_tx_hold_free(tx, dn->dn_object, 750 dmu_tx_mark_netfree(tx); 792 dmu_tx_t *tx; local 815 dmu_free_range(objset_t *os, uint64_t object, uint64_t offset, uint64_t size, dmu_tx_t *tx) argument 889 dmu_write(objset_t *os, uint64_t object, uint64_t offset, uint64_t size, const void *buf, dmu_tx_t *tx) argument 931 dmu_prealloc(objset_t *os, uint64_t object, uint64_t offset, uint64_t size, dmu_tx_t *tx) argument 952 dmu_write_embedded(objset_t *os, uint64_t object, uint64_t offset, void *data, uint8_t etype, uint8_t comp, int uncompressed_size, int compressed_size, int byteorder, dmu_tx_t *tx) argument 1204 dmu_write_uio_dnode(dnode_t *dn, uio_t *uio, uint64_t size, dmu_tx_t *tx) argument 1265 dmu_write_uio_dbuf(dmu_buf_t *zdb, uio_t *uio, uint64_t size, dmu_tx_t *tx) argument 1289 dmu_write_uio(objset_t *os, uint64_t object, uio_t *uio, uint64_t size, dmu_tx_t *tx) argument 1310 dmu_write_pages(objset_t *os, uint64_t object, uint64_t offset, uint64_t size, page_t *pp, dmu_tx_t *tx) argument 1392 dmu_assign_arcbuf(dmu_buf_t *handle, uint64_t offset, arc_buf_t *buf, dmu_tx_t *tx) argument 1557 dmu_tx_t *tx; local 1737 dmu_object_set_blocksize(objset_t *os, uint64_t object, uint64_t size, int ibs, dmu_tx_t *tx) argument 1752 dmu_object_set_checksum(objset_t *os, uint64_t object, uint8_t checksum, dmu_tx_t *tx) argument 1772 dmu_object_set_compress(objset_t *os, uint64_t object, uint8_t compress, dmu_tx_t *tx) argument [all...] |
H A D | zfs_sa.c | 95 zfs_sa_symlink(znode_t *zp, char *link, int len, dmu_tx_t *tx) argument 101 len + ZFS_OLD_ZNODE_PHYS_SIZE, tx) == 0); 109 zfs_grow_blocksize(zp, len, tx); 113 dmu_buf_will_dirty(dbp, tx); 156 zfs_sa_set_scanstamp(znode_t *zp, xvattr_t *xvap, dmu_tx_t *tx) argument 166 sizeof (xoap->xoa_av_scanstamp), tx)); 176 VERIFY(dmu_set_bonus(db, len, tx) == 0); 182 &zp->z_pflags, sizeof (uint64_t), tx)); 195 zfs_sa_upgrade(sa_handle_t *hdl, dmu_tx_t *tx) argument 305 VERIFY(dmu_set_bonustype(db, DMU_OT_SA, tx) 319 zfs_sa_upgrade_txholds(dmu_tx_t *tx, znode_t *zp) argument [all...] |
H A D | zap_micro.c | 47 void *tag, dmu_tx_t *tx, zap_flags_t flags); 462 zap_lockdir_impl(dmu_buf_t *db, void *tag, dmu_tx_t *tx, argument 515 dmu_buf_will_dirty(db, tx); 521 if (zap->zap_ismicro && tx && adding && 528 int err = mzap_upgrade(zapp, tag, tx, 0); 533 VERIFY0(dmu_object_set_blocksize(os, obj, newsz, 0, tx)); 543 zap_lockdir_by_dnode(dnode_t *dn, dmu_tx_t *tx, argument 553 err = zap_lockdir_impl(db, tag, tx, lti, fatreader, adding, zapp); 561 zap_lockdir(objset_t *os, uint64_t obj, dmu_tx_t *tx, argument 570 err = zap_lockdir_impl(db, tag, tx, lt 584 mzap_upgrade(zap_t **zapp, void *tag, dmu_tx_t *tx, zap_flags_t flags) argument 635 mzap_create_impl(objset_t *os, uint64_t obj, int normflags, zap_flags_t flags, dmu_tx_t *tx) argument 669 zap_create_claim(objset_t *os, uint64_t obj, dmu_object_type_t ot, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 677 zap_create_claim_norm(objset_t *os, uint64_t obj, int normflags, dmu_object_type_t ot, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 691 zap_create(objset_t *os, dmu_object_type_t ot, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 698 zap_create_norm(objset_t *os, int normflags, dmu_object_type_t ot, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 708 zap_create_flags(objset_t *os, int normflags, zap_flags_t flags, dmu_object_type_t ot, int leaf_blockshift, int indirect_blockshift, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 727 zap_destroy(objset_t *os, uint64_t zapobj, dmu_tx_t *tx) argument 1070 zap_add(objset_t *os, uint64_t zapobj, const char *key, int integer_size, uint64_t num_integers, const void *val, dmu_tx_t *tx) argument 1115 zap_add_uint64(objset_t *os, uint64_t zapobj, const uint64_t *key, int key_numints, int integer_size, uint64_t num_integers, const void *val, dmu_tx_t *tx) argument 1140 zap_update(objset_t *os, uint64_t zapobj, const char *name, int integer_size, uint64_t num_integers, const void *val, dmu_tx_t *tx) argument 1198 zap_update_uint64(objset_t *os, uint64_t zapobj, const uint64_t *key, int key_numints, int integer_size, uint64_t num_integers, const void *val, dmu_tx_t *tx) argument 1223 zap_remove(objset_t *os, uint64_t zapobj, const char *name, dmu_tx_t *tx) argument 1229 zap_remove_norm(objset_t *os, uint64_t zapobj, const char *name, matchtype_t mt, dmu_tx_t *tx) argument 1264 zap_remove_uint64(objset_t *os, uint64_t zapobj, const uint64_t *key, int key_numints, dmu_tx_t *tx) argument [all...] |
H A D | dnode_sync.c | 40 dnode_increase_indirection(dnode_t *dn, dmu_tx_t *tx) argument 43 int txgoff = tx->tx_txg & TXG_MASK; 114 free_blocks(dnode_t *dn, blkptr_t *bp, int num, dmu_tx_t *tx) argument 125 bytesfreed += dsl_dataset_block_kill(ds, bp, tx, B_FALSE); 147 BP_SET_BIRTH(bp, dmu_tx_get_txg(tx), 0); 155 free_verify(dmu_buf_impl_t *db, uint64_t start, uint64_t end, dmu_tx_t *tx) argument 159 uint64_t txg = tx->tx_txg; 234 dmu_tx_t *tx) 274 FREE_VERIFY(db, start, end, tx); 275 free_blocks(dn, bp, end-start+1, tx); 233 free_children(dmu_buf_impl_t *db, uint64_t blkid, uint64_t nblks, dmu_tx_t *tx) argument 323 dnode_sync_free_range_impl(dnode_t *dn, uint64_t blkid, uint64_t nblks, dmu_tx_t *tx) argument 486 dnode_sync_free(dnode_t *dn, dmu_tx_t *tx) argument 547 dnode_sync(dnode_t *dn, dmu_tx_t *tx) argument [all...] |
H A D | dmu_objset.c | 765 dmu_objset_type_t type, dmu_tx_t *tx) 770 ASSERT(dmu_tx_is_syncing(tx)); 780 DN_MAX_INDBLKSHIFT, DMU_OT_NONE, 0, tx); 811 mdn->dn_next_nlevels[tx->tx_txg & TXG_MASK] = 824 dsl_dataset_dirty(ds, tx); 833 cred_t *cr, dmu_tx_t *tx); 841 dmu_objset_create_check(void *arg, dmu_tx_t *tx) argument 844 dsl_pool_t *dp = dmu_tx_pool(tx); 870 dmu_objset_create_sync(void *arg, dmu_tx_t *tx) argument 873 dsl_pool_t *dp = dmu_tx_pool(tx); 764 dmu_objset_create_impl(spa_t *spa, dsl_dataset_t *ds, blkptr_t *bp, dmu_objset_type_t type, dmu_tx_t *tx) argument 904 dmu_objset_create(const char *name, dmu_objset_type_t type, uint64_t flags, void (*func)(objset_t *os, void *arg, cred_t *cr, dmu_tx_t *tx), void *arg) argument 929 dmu_objset_clone_check(void *arg, dmu_tx_t *tx) argument 975 dmu_objset_clone_sync(void *arg, dmu_tx_t *tx) argument 1029 dmu_objset_sync_dnodes(list_t *list, list_t *newlist, dmu_tx_t *tx) argument 1095 dmu_tx_t *tx = os->os_synctx; local 1105 dmu_objset_sync(objset_t *os, zio_t *pio, dmu_tx_t *tx) argument 1151 dnode_sync(DMU_META_DNODE(os), tx); local 1158 dnode_sync(DMU_USERUSED_DNODE(os), tx); local 1160 dnode_sync(DMU_GROUPUSED_DNODE(os), tx); local 1248 do_userquota_cacheflush(objset_t *os, userquota_cache_t *cache, dmu_tx_t *tx) argument 1304 dmu_objset_do_userquota_updates(objset_t *os, dmu_tx_t *tx) argument 1377 dmu_objset_userquota_find_data(dmu_buf_impl_t *db, dmu_tx_t *tx) argument 1410 dmu_objset_userquota_get_ids(dnode_t *dn, boolean_t before, dmu_tx_t *tx) argument 1540 dmu_tx_t *tx; local [all...] |
H A D | bplist.c | 63 bplist_iterate(bplist_t *bpl, bplist_itor_t *func, void *arg, dmu_tx_t *tx) argument 72 func(arg, &bpe->bpe_blk, tx);
|
H A D | zfs_znode.c | 472 zfs_create_share_dir(zfsvfs_t *zfsvfs, dmu_tx_t *tx) argument 501 zfs_mknode(sharezp, &vattr, tx, kcred, IS_ROOT_NODE, &zp, &acl_ids); 506 ZFS_SHARES_DIR, 8, 1, &sharezp->z_id, tx); 750 * tx - dmu transaction id for zap operations 763 zfs_mknode(znode_t *dzp, vattr_t *vap, dmu_tx_t *tx, cred_t *cr, argument 790 gen = dmu_tx_get_txg(tx); 810 obj_type, bonuslen, tx)); 814 obj_type, bonuslen, tx); 820 obj_type, bonuslen, tx)); 824 obj_type, bonuslen, tx); 1011 zfs_xvattr_set(znode_t *zp, xvattr_t *xvap, dmu_tx_t *tx) argument 1262 zfs_znode_delete(znode_t *zp, dmu_tx_t *tx) argument 1401 zfs_grow_blocksize(znode_t *zp, uint64_t size, dmu_tx_t *tx) argument 1453 dmu_tx_t *tx; local 1568 dmu_tx_t *tx; local 1660 dmu_tx_t *tx; local 1725 zfs_create_fs(objset_t *os, cred_t *cr, nvlist_t *zplprops, dmu_tx_t *tx) argument [all...] |
H A D | zil.c | 377 zil_claim_log_block(zilog_t *zilog, blkptr_t *bp, void *tx, uint64_t first_txg) argument 381 * If tx == NULL, just verify that the block is claimable. 388 tx == NULL ? 0 : first_txg, bp, spa_claim_notify, NULL, 393 zil_claim_log_record(zilog_t *zilog, lr_t *lrc, void *tx, uint64_t first_txg) argument 412 return (zil_claim_log_block(zilog, &lr->lr_blkptr, tx, first_txg)); 417 zil_free_log_block(zilog_t *zilog, blkptr_t *bp, void *tx, uint64_t claim_txg) argument 419 zio_free_zil(zilog->zl_spa, dmu_tx_get_txg(tx), bp); 425 zil_free_log_record(zilog_t *zilog, lr_t *lrc, void *tx, uint64_t claim_txg) argument 436 zio_free(zilog->zl_spa, dmu_tx_get_txg(tx), bp); 529 dmu_tx_t *tx local 601 dmu_tx_t *tx; local 644 zil_destroy_sync(zilog_t *zilog, dmu_tx_t *tx) argument 654 dmu_tx_t *tx = txarg; local 682 dsl_dataset_dirty(dmu_objset_ds(os), tx); local 704 dsl_dataset_dirty(dmu_objset_ds(os), tx); local 719 zil_check_log_chain(dsl_pool_t *dp, dsl_dataset_t *ds, void *tx) argument 862 dmu_tx_t *tx = lwb->lwb_tx; local 953 dmu_tx_t *tx; local 1269 zil_itx_assign(zilog_t *zilog, itx_t *itx, dmu_tx_t *tx) argument 1636 zil_sync(zilog_t *zilog, dmu_tx_t *tx) argument 2157 zil_replaying(zilog_t *zilog, dmu_tx_t *tx) argument [all...] |
H A D | dnode.c | 351 dnode_setbonuslen(dnode_t *dn, int newsize, dmu_tx_t *tx) argument 355 dnode_setdirty(dn, tx); 361 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = DN_ZERO_BONUSLEN; 363 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = dn->dn_bonuslen; 368 dnode_setbonus_type(dnode_t *dn, dmu_object_type_t newtype, dmu_tx_t *tx) argument 371 dnode_setdirty(dn, tx); 374 dn->dn_next_bonustype[tx->tx_txg & TXG_MASK] = dn->dn_bonustype; 379 dnode_rm_spill(dnode_t *dn, dmu_tx_t *tx) argument 383 dnode_setdirty(dn, tx); 384 dn->dn_rm_spillblk[tx 533 dnode_allocate(dnode_t *dn, dmu_object_type_t ot, int blocksize, int ibs, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 616 dnode_reallocate(dnode_t *dn, dmu_object_type_t ot, int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument 1261 dnode_setdirty(dnode_t *dn, dmu_tx_t *tx) argument 1330 dnode_free(dnode_t *dn, dmu_tx_t *tx) argument 1367 dnode_set_blksz(dnode_t *dn, uint64_t size, int ibs, dmu_tx_t *tx) argument 1432 dnode_new_blkid(dnode_t *dn, uint64_t blkid, dmu_tx_t *tx, boolean_t have_read) argument 1514 dnode_dirty_l1(dnode_t *dn, uint64_t l1blkid, dmu_tx_t *tx) argument 1524 dnode_free_range(dnode_t *dn, uint64_t off, uint64_t len, dmu_tx_t *tx) argument 1812 dnode_willuse_space(dnode_t *dn, int64_t space, dmu_tx_t *tx) argument 1820 dsl_pool_dirty_space(dmu_tx_pool(tx), space, tx); local [all...] |
H A D | sa.c | 141 uint16_t buflen, dmu_tx_t *tx); 324 sa_data_op_t data_op, dmu_tx_t *tx) 345 if (tx && !(buftypes & SA_BONUS)) { 346 dmu_buf_will_dirty(hdl->sa_bonus, tx); 357 if (tx && !(buftypes & SA_SPILL) && 359 dmu_buf_will_dirty(hdl->sa_spill, tx); 390 bulk[i].sa_data, bulk[i].sa_length, tx); 394 bulk[i].sa_data, bulk[i].sa_length, tx); 406 uint64_t lot_num, uint64_t hash, boolean_t zapadd, dmu_tx_t *tx) 429 sa->sa_master_obj, SA_LAYOUTS, tx); 323 sa_attr_op(sa_handle_t *hdl, sa_bulk_attr_t *bulk, int count, sa_data_op_t data_op, dmu_tx_t *tx) argument 405 sa_add_layout_entry(objset_t *os, sa_attr_type_t *attrs, int attr_count, uint64_t lot_num, uint64_t hash, boolean_t zapadd, dmu_tx_t *tx) argument 462 sa_find_layout(objset_t *os, uint64_t hash, sa_attr_type_t *attrs, int count, dmu_tx_t *tx, sa_lot_t **lot) argument 492 sa_resize_spill(sa_handle_t *hdl, uint32_t size, dmu_tx_t *tx) argument 645 sa_build_layouts(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc, int attr_count, dmu_tx_t *tx) argument 1568 sa_attr_register_sync(sa_handle_t *hdl, dmu_tx_t *tx) argument 1610 sa_replace_all_by_template_locked(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc, int attr_count, dmu_tx_t *tx) argument 1621 sa_replace_all_by_template(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc, int attr_count, dmu_tx_t *tx) argument 1641 sa_modify_attrs(sa_handle_t *hdl, sa_attr_type_t newattr, sa_data_op_t action, sa_data_locator_t *locator, void *datastart, uint16_t buflen, dmu_tx_t *tx) argument 1788 sa_bulk_update_impl(sa_handle_t *hdl, sa_bulk_attr_t *bulk, int count, dmu_tx_t *tx) argument 1815 sa_update(sa_handle_t *hdl, sa_attr_type_t type, void *buf, uint32_t buflen, dmu_tx_t *tx) argument 1833 sa_update_from_cb(sa_handle_t *hdl, sa_attr_type_t attr, uint32_t buflen, sa_data_locator_t *locator, void *userdata, dmu_tx_t *tx) argument 1897 sa_bulk_update(sa_handle_t *hdl, sa_bulk_attr_t *attrs, int count, dmu_tx_t *tx) argument 1909 sa_remove(sa_handle_t *hdl, sa_attr_type_t attr, dmu_tx_t *tx) argument [all...] |
/illumos-gate/usr/src/cmd/svc/configd/ |
H A D | object.c | 658 backend_tx_t *tx; local 662 ret = backend_tx_begin(backend, &tx); 674 r = backend_tx_run(tx, q, tx_check_genid, data); 679 backend_tx_rollback(tx); 685 backend_tx_rollback(tx); 690 new_gen = backend_new_id(tx, BACKEND_ID_GENERATION); 692 backend_tx_rollback(tx); 698 data->txc_tx = tx; 700 r = backend_tx_run_update(tx, 706 backend_tx_rollback(tx); [all...] |
H A D | backend.c | 119 int bt_full; /* SQLITE_FULL during tx */ 2141 backend_tx_end(backend_tx_t *tx) argument 2145 be = tx->bt_be; 2147 if (tx->bt_full) { 2162 tx->bt_be = NULL; 2163 uu_free(tx); 2167 backend_tx_end_ro(backend_tx_t *tx) argument 2169 assert(tx->bt_readonly); 2170 backend_tx_end(tx); 2214 backend_tx_rollback(backend_tx_t *tx) argument 2241 backend_tx_commit(backend_tx_t *tx) argument 2305 backend_new_id(backend_tx_t *tx, enum id_space id) argument 2347 backend_tx_run(backend_tx_t *tx, backend_query_t *q, backend_run_callback_f *cb, void *data) argument 2379 backend_tx_run_single_int(backend_tx_t *tx, backend_query_t *q, uint32_t *buf) argument 2401 backend_tx_run_update(backend_tx_t *tx, const char *format, ...) argument 2430 backend_tx_run_update_changed(backend_tx_t *tx, const char *format, ...) argument [all...] |
/illumos-gate/usr/src/uts/common/io/comstar/lu/stmf_sbd/ |
H A D | sbd_zvol.c | 73 * dmu_tx_commit(tx) 74 * dmu_tx_abort(tx) 336 dmu_tx_t *tx; local 348 tx = dmu_tx_create(sl->sl_zvol_objset_hdl); 349 dmu_tx_hold_write(tx, ZVOL_OBJ, offset, (int)len); 350 error = dmu_tx_assign(tx, TXG_WAIT); 353 dmu_tx_abort(tx); 367 dmu_assign_arcbuf(sl->sl_zvol_bonus_hdl, toffset, abuf, tx); 374 zvol_log_write_minor(sl->sl_zvol_minor_hdl, tx, offset, 376 dmu_tx_commit(tx); 422 dmu_tx_t *tx; local [all...] |
/illumos-gate/usr/src/uts/common/sys/ |
H A D | fss.h | 152 #define FSSPROC(tx) ((fssproc_t *)(tx->t_cldata)) 155 #define FSSPROJ(tx) (FSSPROC(tx)->fss_proj)
|
/illumos-gate/usr/src/uts/common/disp/ |
H A D | ts.c | 1159 ts_parmsset(kthread_t *tx, void *parmsp, id_t reqpcid, cred_t *reqpcredp) argument 1165 tsproc_t *tspp = (tsproc_t *)tx->t_cldata; 1167 ASSERT(MUTEX_HELD(&(ttoproc(tx))->p_lock)); 1213 thread_lock(tx); 1221 thread_unlock(tx); 1226 ts_change_priority(tx, tspp); 1227 thread_unlock(tx); 1233 ia_parmsset(kthread_t *tx, void *parmsp, id_t reqpcid, cred_t *reqpcredp) argument 1235 tsproc_t *tspp = (tsproc_t *)tx->t_cldata; 1247 return (ts_parmsset(tx, parms 1916 kthread_t *tx; local 2194 kthread_t *tx; local [all...] |
/illumos-gate/usr/src/lib/libcurses/screen/ |
H A D | mvcur.c | 322 int nt, tx, x, stcost, iscont; local 333 tx = (nt > 0) ? (cx / TABSIZE + nt) * TABSIZE : cx; 338 for (x = tx; x < nx; ++x) { 370 for (x = tx; x < nx; ++x) { 396 int tx, nt, x, stcost; local 403 tx = cx; 413 tx = x; 416 stcost = nt * _COST(Back_tab) + (tx-nx) * _COST(Cursor_left); 431 for (; tx > nx; --tx) [all...] |
/illumos-gate/usr/src/cmd/ztest/ |
H A D | ztest.c | 1220 ztest_tx_assign(dmu_tx_t *tx, uint64_t txg_how, const char *tag) argument 1226 * Attempt to assign tx to some transaction group. 1228 error = dmu_tx_assign(tx, txg_how); 1232 dmu_tx_wait(tx); 1237 dmu_tx_abort(tx); 1240 txg = dmu_tx_get_txg(tx); 1319 ztest_log_create(ztest_ds_t *zd, dmu_tx_t *tx, lr_create_t *lr) argument 1325 if (zil_replaying(zd->zd_zilog, tx)) 1332 zil_itx_assign(zd->zd_zilog, itx, tx); 1336 ztest_log_remove(ztest_ds_t *zd, dmu_tx_t *tx, lr_remove_ argument 1354 ztest_log_write(ztest_ds_t *zd, dmu_tx_t *tx, lr_write_t *lr) argument 1387 ztest_log_truncate(ztest_ds_t *zd, dmu_tx_t *tx, lr_truncate_t *lr) argument 1403 ztest_log_setattr(ztest_ds_t *zd, dmu_tx_t *tx, lr_setattr_t *lr) argument 1428 dmu_tx_t *tx; local 1511 dmu_tx_t *tx; local 1566 dmu_tx_t *tx; local 1685 dmu_tx_t *tx; local 1724 dmu_tx_t *tx; local 2137 dmu_tx_t *tx; local 3175 ztest_objset_create_cb(objset_t *os, void *arg, cred_t *cr, dmu_tx_t *tx) argument 3551 dmu_tx_t *tx; local 3825 dmu_tx_t *tx; local 4121 dmu_tx_t *tx; local 4273 dmu_tx_t *tx; local 4298 dmu_tx_t *tx; local 4487 dmu_tx_t *tx; local 5053 dmu_tx_t *tx; local [all...] |
/illumos-gate/usr/src/cmd/zhack/ |
H A D | zhack.c | 288 zhack_feature_enable_sync(void *arg, dmu_tx_t *tx) argument 290 spa_t *spa = dmu_tx_pool(tx)->dp_spa; 293 feature_enable_sync(spa, feature, tx); 295 spa_history_log_internal(spa, "zhack enable feature", tx, 370 feature_incr_sync(void *arg, dmu_tx_t *tx) argument 372 spa_t *spa = dmu_tx_pool(tx)->dp_spa; 377 feature_sync(spa, feature, refcount + 1, tx); 378 spa_history_log_internal(spa, "zhack feature incr", tx, 383 feature_decr_sync(void *arg, dmu_tx_t *tx) argument 385 spa_t *spa = dmu_tx_pool(tx) [all...] |
/illumos-gate/usr/src/uts/sun4v/sys/ |
H A D | cnex.h | 84 cnex_intr_t tx; /* Transmit interrupt */ member in struct:cnex_ldc
|