Searched refs:tx (Results 1 - 25 of 198) sorted by relevance

12345678

/illumos-gate/usr/src/cmd/refer/
H A Dtick.c32 struct tbuffer tx; local
34 times(&tx);
36 user = tx.proc_user_time;
37 systm = tx.proc_system_time;
44 struct tbuffer tx; local
49 times(&tx);
52 use = (tx.proc_user_time - user)/60.;
53 sys = (tx.proc_system_time - systm)/60.;
H A Drefer6.c33 char *s, *tx; local
48 tx = s+3;
52 tx = s+2;
57 tx = s;
63 tx = caps(tx, buf1);
65 fprintf(stderr, " s %o tx %o %s\n", s, tx, tx);
70 fprintf(fo, "%s%c", tx, se
[all...]
/illumos-gate/usr/src/uts/common/fs/zfs/
H A Dtxg.c119 tx_state_t *tx = &dp->dp_tx; local
121 bzero(tx, sizeof (tx_state_t));
123 tx->tx_cpu = kmem_zalloc(max_ncpus * sizeof (tx_cpu_t), KM_SLEEP);
128 mutex_init(&tx->tx_cpu[c].tc_lock, NULL, MUTEX_DEFAULT, NULL);
129 mutex_init(&tx->tx_cpu[c].tc_open_lock, NULL, MUTEX_DEFAULT,
132 cv_init(&tx->tx_cpu[c].tc_cv[i], NULL, CV_DEFAULT,
134 list_create(&tx->tx_cpu[c].tc_callbacks[i],
140 mutex_init(&tx->tx_sync_lock, NULL, MUTEX_DEFAULT, NULL);
142 cv_init(&tx->tx_sync_more_cv, NULL, CV_DEFAULT, NULL);
143 cv_init(&tx
157 tx_state_t *tx = &dp->dp_tx; local
195 tx_state_t *tx = &dp->dp_tx; local
220 txg_thread_enter(tx_state_t *tx, callb_cpr_t *cpr) argument
227 txg_thread_exit(tx_state_t *tx, callb_cpr_t *cpr, kthread_t **tpp) argument
238 txg_thread_wait(tx_state_t *tx, callb_cpr_t *cpr, kcondvar_t *cv, clock_t time) argument
257 tx_state_t *tx = &dp->dp_tx; local
294 tx_state_t *tx = &dp->dp_tx; local
355 tx_state_t *tx = &dp->dp_tx; local
411 tx_state_t *tx = &dp->dp_tx; local
450 tx_state_t *tx = &dp->dp_tx; local
529 tx_state_t *tx = &dp->dp_tx; local
579 tx_state_t *tx = &dp->dp_tx; local
605 tx_state_t *tx = &dp->dp_tx; local
630 tx_state_t *tx = &dp->dp_tx; local
656 tx_state_t *tx = &dp->dp_tx; local
674 tx_state_t *tx = &dp->dp_tx; local
681 tx_state_t *tx = &dp->dp_tx; local
[all...]
H A Ddmu_tx.c43 typedef void (*dmu_tx_hold_func_t)(dmu_tx_t *tx, struct dnode *dn,
50 dmu_tx_t *tx = kmem_zalloc(sizeof (dmu_tx_t), KM_SLEEP); local
51 tx->tx_dir = dd;
53 tx->tx_pool = dd->dd_pool;
54 list_create(&tx->tx_holds, sizeof (dmu_tx_hold_t),
56 list_create(&tx->tx_callbacks, sizeof (dmu_tx_callback_t),
58 tx->tx_start = gethrtime();
60 refcount_create(&tx->tx_space_written);
61 refcount_create(&tx->tx_space_freed);
63 return (tx);
69 dmu_tx_t *tx = dmu_tx_create_dd(os->os_dsl_dataset->ds_dir); local
78 dmu_tx_t *tx = dmu_tx_create_dd(NULL); local
89 dmu_tx_is_syncing(dmu_tx_t *tx) argument
95 dmu_tx_private_ok(dmu_tx_t *tx) argument
101 dmu_tx_hold_object_impl(dmu_tx_t *tx, objset_t *os, uint64_t object, enum dmu_tx_hold_type type, uint64_t arg1, uint64_t arg2) argument
149 dmu_tx_add_new_object(dmu_tx_t *tx, objset_t *os, uint64_t object) argument
433 dmu_tx_hold_write(dmu_tx_t *tx, uint64_t object, uint64_t off, int len) argument
632 dmu_tx_mark_netfree(dmu_tx_t *tx) argument
653 dmu_tx_hold_free(dmu_tx_t *tx, uint64_t object, uint64_t off, uint64_t len) argument
744 dmu_tx_hold_zap(dmu_tx_t *tx, uint64_t object, int add, const char *name) argument
848 dmu_tx_hold_bonus(dmu_tx_t *tx, uint64_t object) argument
861 dmu_tx_hold_space(dmu_tx_t *tx, uint64_t space) argument
873 dmu_tx_holds(dmu_tx_t *tx, uint64_t object) argument
900 dmu_tx_dirty_buf(dmu_tx_t *tx, dmu_buf_impl_t *db) argument
1104 dmu_tx_delay(dmu_tx_t *tx, uint64_t dirty) argument
1156 dmu_tx_try_assign(dmu_tx_t *tx, txg_how_t txg_how) argument
1269 dmu_tx_unassign(dmu_tx_t *tx) argument
1321 dmu_tx_assign(dmu_tx_t *tx, txg_how_t txg_how) argument
1351 dmu_tx_wait(dmu_tx_t *tx) argument
1409 dmu_tx_willuse_space(dmu_tx_t *tx, int64_t delta) argument
1426 dmu_tx_destroy(dmu_tx_t *tx) argument
1463 dmu_tx_commit(dmu_tx_t *tx) argument
1506 dmu_tx_abort(dmu_tx_t *tx) argument
1520 dmu_tx_get_txg(dmu_tx_t *tx) argument
1527 dmu_tx_pool(dmu_tx_t *tx) argument
1535 dmu_tx_callback_register(dmu_tx_t *tx, dmu_tx_callback_func_t *func, void *data) argument
1577 dmu_tx_sa_registration_hold(sa_os_t *sa, dmu_tx_t *tx) argument
1598 dmu_tx_hold_spill(dmu_tx_t *tx, uint64_t object) argument
1635 dmu_tx_hold_sa_create(dmu_tx_t *tx, int attrsize) argument
1672 dmu_tx_hold_sa(dmu_tx_t *tx, sa_handle_t *hdl, boolean_t may_grow) argument
[all...]
H A Ddsl_destroy.c89 dsl_destroy_snapshot_check(void *arg, dmu_tx_t *tx) argument
92 dsl_pool_t *dp = dmu_tx_pool(tx);
96 if (!dmu_tx_is_syncing(tx))
144 process_old_cb(void *arg, const blkptr_t *bp, dmu_tx_t *tx) argument
152 dsl_deadlist_insert(&poa->ds->ds_deadlist, bp, tx);
163 dsl_free_sync(poa->pio, dp, tx->tx_txg, bp);
170 dsl_dataset_t *ds_next, boolean_t after_branch_point, dmu_tx_t *tx)
185 process_old_cb, &poa, tx));
191 -poa.used, -poa.comp, -poa.uncomp, tx);
207 dsl_dataset_remove_clones_key(dsl_dataset_t *ds, uint64_t mintxg, dmu_tx_t *tx) argument
169 process_old_deadlist(dsl_dataset_t *ds, dsl_dataset_t *ds_prev, dsl_dataset_t *ds_next, boolean_t after_branch_point, dmu_tx_t *tx) argument
239 dsl_destroy_snapshot_sync_impl(dsl_dataset_t *ds, boolean_t defer, dmu_tx_t *tx) argument
477 dsl_destroy_snapshot_sync(void *arg, dmu_tx_t *tx) argument
549 dmu_tx_t *tx; member in struct:killarg
558 dmu_tx_t *tx = ka->tx; local
581 old_synchronous_dataset_destroy(dsl_dataset_t *ds, dmu_tx_t *tx) argument
651 dsl_destroy_head_check(void *arg, dmu_tx_t *tx) argument
668 dsl_dir_destroy_sync(uint64_t ddobj, dmu_tx_t *tx) argument
715 dsl_destroy_head_sync_impl(dsl_dataset_t *ds, dmu_tx_t *tx) argument
796 zil_destroy_sync(dmu_objset_zil(os), tx); local
884 dsl_destroy_head_sync(void *arg, dmu_tx_t *tx) argument
896 dsl_destroy_head_begin_sync(void *arg, dmu_tx_t *tx) argument
[all...]
H A Ddmu_object.c36 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx)
95 dnode_allocate(dn, ot, blocksize, 0, bonustype, bonuslen, tx);
100 dmu_tx_add_new_object(tx, os, object);
106 int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx)
111 if (object == DMU_META_DNODE_OBJECT && !dmu_tx_private_ok(tx))
117 dnode_allocate(dn, ot, blocksize, 0, bonustype, bonuslen, tx);
120 dmu_tx_add_new_object(tx, os, object);
126 int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx)
139 dnode_reallocate(dn, ot, blocksize, bonustype, bonuslen, tx);
146 dmu_object_free(objset_t *os, uint64_t object, dmu_tx_t *tx) argument
35 dmu_object_alloc(objset_t *os, dmu_object_type_t ot, int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument
105 dmu_object_claim(objset_t *os, uint64_t object, dmu_object_type_t ot, int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument
125 dmu_object_reclaim(objset_t *os, uint64_t object, dmu_object_type_t ot, int blocksize, dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx) argument
192 dmu_object_zapify(objset_t *mos, uint64_t object, dmu_object_type_t old_type, dmu_tx_t *tx) argument
214 SPA_FEATURE_EXTENSIBLE_DATASET, tx); local
218 dmu_object_free_zapified(objset_t *mos, uint64_t object, dmu_tx_t *tx) argument
231 SPA_FEATURE_EXTENSIBLE_DATASET, tx); local
[all...]
H A Ddsl_synctask.c37 dsl_null_checkfunc(void *arg, dmu_tx_t *tx) argument
50 * dmu_tx_is_syncing(tx).
71 dmu_tx_t *tx; local
82 tx = dmu_tx_create_dd(dp->dp_mos_dir);
83 VERIFY0(dmu_tx_assign(tx, TXG_WAIT));
86 dst.dst_txg = dmu_tx_get_txg(tx);
96 err = dst.dst_checkfunc(arg, tx);
100 dmu_tx_commit(tx);
107 dmu_tx_commit(tx);
122 int blocks_modified, zfs_space_check_t space_check, dmu_tx_t *tx)
121 dsl_sync_task_nowait(dsl_pool_t *dp, dsl_syncfunc_t *syncfunc, void *arg, int blocks_modified, zfs_space_check_t space_check, dmu_tx_t *tx) argument
143 dsl_sync_task_sync(dsl_sync_task_t *dst, dmu_tx_t *tx) argument
[all...]
H A Dspa_history.c86 spa_history_create_obj(spa_t *spa, dmu_tx_t *tx) argument
95 sizeof (spa_history_phys_t), tx);
99 &spa->spa_history, tx) == 0);
105 dmu_buf_will_dirty(dbp, tx);
151 dmu_tx_t *tx)
170 dmu_write(mos, spa->spa_history, phys_eof, firstwrite, buf, tx);
176 len, (char *)buf + firstwrite, tx);
199 spa_history_log_sync(void *arg, dmu_tx_t *tx) argument
202 spa_t *spa = dmu_tx_pool(tx)->dp_spa;
217 spa_history_create_obj(spa, tx);
150 spa_history_write(spa_t *spa, void *buf, uint64_t len, spa_history_phys_t *shpp, dmu_tx_t *tx) argument
304 dmu_tx_t *tx; local
326 nvarg, 0, ZFS_SPACE_CHECK_NONE, tx); local
435 log_internal(nvlist_t *nvl, const char *operation, spa_t *spa, dmu_tx_t *tx, const char *fmt, va_list adx) argument
462 spa_history_log_sync, nvl, 0, ZFS_SPACE_CHECK_NONE, tx); local
468 spa_history_log_internal(spa_t *spa, const char *operation, dmu_tx_t *tx, const char *fmt, ...) argument
493 spa_history_log_internal_ds(dsl_dataset_t *ds, const char *operation, dmu_tx_t *tx, const char *fmt, ...) argument
512 spa_history_log_internal_dd(dsl_dir_t *dd, const char *operation, dmu_tx_t *tx, const char *fmt, ...) argument
[all...]
H A Dzfeature.c297 dmu_tx_t *tx)
304 sizeof (uint64_t), 1, &refcount, tx));
323 spa_activate_mos_feature(spa, feature->fi_guid, tx);
331 feature_enable_sync(spa_t *spa, zfeature_info_t *feature, dmu_tx_t *tx) argument
349 spa_feature_enable(spa, feature->fi_depends[i], tx);
353 feature->fi_desc, tx));
355 feature_sync(spa, feature, initial_refcount, tx);
358 uint64_t enabling_txg = dmu_tx_get_txg(tx);
364 DMU_POOL_FEATURE_ENABLED_TXG, tx);
366 spa_feature_incr(spa, SPA_FEATURE_ENABLED_TXG, tx);
296 feature_sync(spa_t *spa, zfeature_info_t *feature, uint64_t refcount, dmu_tx_t *tx) argument
375 feature_do_action(spa_t *spa, spa_feature_t fid, feature_action_t action, dmu_tx_t *tx) argument
410 spa_feature_create_zap_objects(spa_t *spa, dmu_tx_t *tx) argument
434 spa_feature_enable(spa_t *spa, spa_feature_t fid, dmu_tx_t *tx) argument
442 spa_feature_incr(spa_t *spa, spa_feature_t fid, dmu_tx_t *tx) argument
448 spa_feature_decr(spa_t *spa, spa_feature_t fid, dmu_tx_t *tx) argument
[all...]
H A Ddsl_deadlist.c147 dsl_deadlist_alloc(objset_t *os, dmu_tx_t *tx) argument
150 return (bpobj_alloc(os, SPA_OLD_MAXBLOCKSIZE, tx));
152 sizeof (dsl_deadlist_phys_t), tx));
156 dsl_deadlist_free(objset_t *os, uint64_t dlobj, dmu_tx_t *tx) argument
164 bpobj_free(os, dlobj, tx);
173 bpobj_decr_empty(os, tx);
175 bpobj_free(os, obj, tx);
178 VERIFY3U(0, ==, dmu_object_free(os, dlobj, tx));
183 const blkptr_t *bp, dmu_tx_t *tx)
187 uint64_t obj = bpobj_alloc(dl->dl_os, SPA_OLD_MAXBLOCKSIZE, tx);
182 dle_enqueue(dsl_deadlist_t *dl, dsl_deadlist_entry_t *dle, const blkptr_t *bp, dmu_tx_t *tx) argument
198 dle_enqueue_subobj(dsl_deadlist_t *dl, dsl_deadlist_entry_t *dle, uint64_t obj, dmu_tx_t *tx) argument
214 dsl_deadlist_insert(dsl_deadlist_t *dl, const blkptr_t *bp, dmu_tx_t *tx) argument
249 dsl_deadlist_add_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx) argument
273 dsl_deadlist_remove_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx) argument
300 dsl_deadlist_regenerate(objset_t *os, uint64_t dlobj, uint64_t mrs_obj, dmu_tx_t *tx) argument
324 dsl_deadlist_clone(dsl_deadlist_t *dl, uint64_t maxtxg, uint64_t mrs_obj, dmu_tx_t *tx) argument
418 dsl_deadlist_insert_bpobj(dsl_deadlist_t *dl, uint64_t obj, uint64_t birth, dmu_tx_t *tx) argument
448 dsl_deadlist_insert_cb(void *arg, const blkptr_t *bp, dmu_tx_t *tx) argument
460 dsl_deadlist_merge(dsl_deadlist_t *dl, uint64_t obj, dmu_tx_t *tx) argument
498 dsl_deadlist_move_bpobj(dsl_deadlist_t *dl, bpobj_t *bpo, uint64_t mintxg, dmu_tx_t *tx) argument
[all...]
H A Dbptree.c58 dmu_tx_t *ba_tx; /* caller supplied tx, NULL if not freeing */
62 bptree_alloc(objset_t *os, dmu_tx_t *tx) argument
70 sizeof (bptree_phys_t), tx);
77 dmu_buf_will_dirty(db, tx);
90 bptree_free(objset_t *os, uint64_t obj, dmu_tx_t *tx) argument
103 return (dmu_object_free(os, obj, tx));
122 uint64_t bytes, uint64_t comp, uint64_t uncomp, dmu_tx_t *tx)
133 ASSERT(dmu_tx_is_syncing(tx));
140 dmu_write(os, obj, bt->bt_end * sizeof (bte), sizeof (bte), &bte, tx);
142 dmu_buf_will_dirty(db, tx);
121 bptree_add(objset_t *os, uint64_t obj, blkptr_t *bp, uint64_t birth_txg, uint64_t bytes, uint64_t comp, uint64_t uncomp, dmu_tx_t *tx) argument
187 bptree_iterate(objset_t *os, uint64_t obj, boolean_t free, bptree_itor_t func, void *arg, dmu_tx_t *tx) argument
[all...]
H A Ddsl_pool.c344 dmu_tx_t *tx = dmu_tx_create_assigned(dp, txg); local
353 NULL, &dp->dp_meta_rootbp, DMU_OST_META, tx);
357 DMU_OT_OBJECT_DIRECTORY, DMU_OT_NONE, 0, tx);
364 dp->dp_root_dir_obj = dsl_dir_create_sync(dp, NULL, NULL, tx);
369 (void) dsl_dir_create_sync(dp, dp->dp_root_dir, MOS_DIR_NAME, tx);
376 FREE_DIR_NAME, tx);
381 obj = bpobj_alloc(dp->dp_meta_objset, SPA_OLD_MAXBLOCKSIZE, tx);
383 DMU_POOL_FREE_BPOBJ, sizeof (uint64_t), 1, &obj, tx) == 0);
389 dsl_pool_create_origin(dp, tx);
392 obj = dsl_dataset_create_sync_dd(dp->dp_root_dir, NULL, 0, tx);
428 dsl_pool_sync_mos(dsl_pool_t *dp, dmu_tx_t *tx) argument
459 dmu_tx_t *tx; local
645 dsl_pool_dirty_space(dsl_pool_t *dp, int64_t space, dmu_tx_t *tx) argument
677 dmu_tx_t *tx = arg; local
755 dsl_pool_upgrade_clones(dsl_pool_t *dp, dmu_tx_t *tx) argument
768 dmu_tx_t *tx = arg; local
794 dsl_pool_upgrade_dir_clones(dsl_pool_t *dp, dmu_tx_t *tx) argument
819 dsl_pool_create_origin(dsl_pool_t *dp, dmu_tx_t *tx) argument
890 dsl_pool_user_hold_create_obj(dsl_pool_t *dp, dmu_tx_t *tx) argument
902 dsl_pool_user_hold_rele_impl(dsl_pool_t *dp, uint64_t dsobj, const char *tag, uint64_t now, dmu_tx_t *tx, boolean_t holding) argument
940 dsl_pool_user_hold(dsl_pool_t *dp, uint64_t dsobj, const char *tag, uint64_t now, dmu_tx_t *tx) argument
950 dsl_pool_user_release(dsl_pool_t *dp, uint64_t dsobj, const char *tag, dmu_tx_t *tx) argument
[all...]
H A Ddsl_scan.c174 dsl_scan_setup_check(void *arg, dmu_tx_t *tx) argument
176 dsl_scan_t *scn = dmu_tx_pool(tx)->dp_scan;
185 dsl_scan_setup_sync(void *arg, dmu_tx_t *tx) argument
187 dsl_scan_t *scn = dmu_tx_pool(tx)->dp_scan;
199 scn->scn_phys.scn_max_txg = tx->tx_txg;
244 ot ? ot : DMU_OT_SCAN_QUEUE, DMU_OT_NONE, 0, tx);
246 dsl_scan_sync_state(scn, tx);
248 spa_history_log_internal(spa, "scan setup", tx,
255 dsl_scan_done(dsl_scan_t *scn, boolean_t complete, dmu_tx_t *tx) argument
276 DMU_POOL_DIRECTORY_OBJECT, old_names[i], tx);
342 dsl_scan_cancel_check(void *arg, dmu_tx_t *tx) argument
353 dsl_scan_cancel_sync(void *arg, dmu_tx_t *tx) argument
398 dsl_scan_sync_state(dsl_scan_t *scn, dmu_tx_t *tx) argument
620 dsl_scan_recurse(dsl_scan_t *scn, dsl_dataset_t *ds, dmu_objset_type_t ostype, dnode_phys_t *dnp, const blkptr_t *bp, const zbookmark_phys_t *zb, dmu_tx_t *tx) argument
719 dsl_scan_visitdnode(dsl_scan_t *scn, dsl_dataset_t *ds, dmu_objset_type_t ostype, dnode_phys_t *dnp, uint64_t object, dmu_tx_t *tx) argument
748 dsl_scan_visitbp(blkptr_t *bp, const zbookmark_phys_t *zb, dnode_phys_t *dnp, dsl_dataset_t *ds, dsl_scan_t *scn, dmu_objset_type_t ostype, dmu_tx_t *tx) argument
805 dsl_scan_visit_rootbp(dsl_scan_t *scn, dsl_dataset_t *ds, blkptr_t *bp, dmu_tx_t *tx) argument
819 dsl_scan_ds_destroyed(dsl_dataset_t *ds, dmu_tx_t *tx) argument
889 dsl_scan_ds_snapshotted(dsl_dataset_t *ds, dmu_tx_t *tx) argument
923 dsl_scan_ds_clone_swapped(dsl_dataset_t *ds1, dsl_dataset_t *ds2, dmu_tx_t *tx) argument
985 dmu_tx_t *tx; member in struct:enqueue_clones_arg
1023 dsl_scan_visitds(dsl_scan_t *scn, uint64_t dsobj, dmu_tx_t *tx) argument
1171 dmu_tx_t *tx = arg; local
1241 dsl_scan_ddt(dsl_scan_t *scn, dmu_tx_t *tx) argument
1281 dsl_scan_ddt_entry(dsl_scan_t *scn, enum zio_checksum checksum, ddt_entry_t *dde, dmu_tx_t *tx) argument
1304 dsl_scan_visit(dsl_scan_t *scn, dmu_tx_t *tx) argument
1409 dsl_scan_free_block_cb(void *arg, const blkptr_t *bp, dmu_tx_t *tx) argument
1450 dsl_scan_sync(dsl_pool_t *dp, dmu_tx_t *tx) argument
1682 dmu_tx_t *tx; local
1889 dsl_scan_restarting(dsl_scan_t *scn, dmu_tx_t *tx) argument
[all...]
/illumos-gate/usr/src/uts/common/io/nxge/
H A Dnxge_hv.c53 nxhv_dc_fp_t *tx; local
69 tx = &nhd->hio.tx;
71 tx->assign = &hv_niu_tx_dma_assign;
72 tx->unassign = &hv_niu_tx_dma_unassign;
73 tx->get_map = &hv_niu_vr_get_txmap;
76 tx->lp_conf = &hv_niu_tx_logical_page_conf;
77 tx->lp_info = &hv_niu_tx_logical_page_info;
79 tx->lp_cfgh_conf = &hv_niu_cfgh_tx_logical_page_conf;
80 tx
[all...]
/illumos-gate/usr/src/uts/common/fs/zfs/sys/
H A Ddmu_tx.h50 * No synchronization is needed because a tx can only be handled
128 int dmu_tx_assign(dmu_tx_t *tx, txg_how_t txg_how);
129 void dmu_tx_commit(dmu_tx_t *tx);
130 void dmu_tx_abort(dmu_tx_t *tx);
131 uint64_t dmu_tx_get_txg(dmu_tx_t *tx);
132 struct dsl_pool *dmu_tx_pool(dmu_tx_t *tx);
133 void dmu_tx_wait(dmu_tx_t *tx);
135 void dmu_tx_callback_register(dmu_tx_t *tx, dmu_tx_callback_func_t *dcb_func,
148 int dmu_tx_is_syncing(dmu_tx_t *tx);
149 int dmu_tx_private_ok(dmu_tx_t *tx);
[all...]
H A Ddsl_deadlist.h67 uint64_t dsl_deadlist_alloc(objset_t *os, dmu_tx_t *tx);
68 void dsl_deadlist_free(objset_t *os, uint64_t dlobj, dmu_tx_t *tx);
69 void dsl_deadlist_insert(dsl_deadlist_t *dl, const blkptr_t *bp, dmu_tx_t *tx);
70 void dsl_deadlist_add_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx);
71 void dsl_deadlist_remove_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx);
73 uint64_t mrs_obj, dmu_tx_t *tx);
79 void dsl_deadlist_merge(dsl_deadlist_t *dl, uint64_t obj, dmu_tx_t *tx);
81 dmu_tx_t *tx);
H A Dbpobj.h68 typedef int bpobj_itor_t(void *arg, const blkptr_t *bp, dmu_tx_t *tx);
70 uint64_t bpobj_alloc(objset_t *mos, int blocksize, dmu_tx_t *tx);
71 uint64_t bpobj_alloc_empty(objset_t *os, int blocksize, dmu_tx_t *tx);
72 void bpobj_free(objset_t *os, uint64_t obj, dmu_tx_t *tx);
73 void bpobj_decr_empty(objset_t *os, dmu_tx_t *tx);
78 int bpobj_iterate(bpobj_t *bpo, bpobj_itor_t func, void *arg, dmu_tx_t *tx);
81 void bpobj_enqueue_subobj(bpobj_t *bpo, uint64_t subobj, dmu_tx_t *tx);
82 void bpobj_enqueue(bpobj_t *bpo, const blkptr_t *bp, dmu_tx_t *tx);
H A Dbptree.h49 typedef int bptree_itor_t(void *arg, const blkptr_t *bp, dmu_tx_t *tx);
51 uint64_t bptree_alloc(objset_t *os, dmu_tx_t *tx);
52 int bptree_free(objset_t *os, uint64_t obj, dmu_tx_t *tx);
56 uint64_t bytes, uint64_t comp, uint64_t uncomp, dmu_tx_t *tx);
59 bptree_itor_t func, void *arg, dmu_tx_t *tx);
H A Dzap.h135 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx);
137 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx);
140 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx);
142 uint64_t parent_obj, const char *name, dmu_tx_t *tx);
148 zap_flags_t flags, dmu_tx_t *tx);
155 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx);
158 dmu_object_type_t bonustype, int bonuslen, dmu_tx_t *tx);
170 int zap_destroy(objset_t *ds, uint64_t zapobj, dmu_tx_t *tx);
237 const void *val, dmu_tx_t *tx);
240 const void *val, dmu_tx_t *tx);
[all...]
H A Dbplist.h45 typedef int bplist_itor_t(void *arg, const blkptr_t *bp, dmu_tx_t *tx);
51 void *arg, dmu_tx_t *tx);
H A Ddsl_userhold.h49 boolean_t temphold, struct dmu_tx *tx);
51 minor_t minor, uint64_t now, struct dmu_tx *tx);
H A Ddsl_dir.h109 /* gross estimate of space used by in-flight tx's */
133 const char *name, dmu_tx_t *tx);
137 void dsl_dir_dirty(dsl_dir_t *dd, dmu_tx_t *tx);
138 void dsl_dir_sync(dsl_dir_t *dd, dmu_tx_t *tx);
141 dmu_tx_t *tx);
142 void dsl_dir_tempreserve_clear(void *tr_cookie, dmu_tx_t *tx);
143 void dsl_dir_willuse_space(dsl_dir_t *dd, int64_t space, dmu_tx_t *tx);
145 int64_t used, int64_t compressed, int64_t uncompressed, dmu_tx_t *tx);
147 dd_used_t oldtype, dd_used_t newtype, dmu_tx_t *tx);
161 uint64_t reservation, cred_t *cr, dmu_tx_t *tx);
[all...]
/illumos-gate/usr/src/cmd/rcap/common/
H A Drcapd_conf.c169 scf_transaction_t *tx = NULL; local
179 if ((tx = scf_transaction_setup(simple_h)) == NULL) {
186 if (scf_set_count_property(tx, PRESSURE,
194 if (scf_set_count_property(tx, RECONFIG_INT,
201 if (scf_set_count_property(tx, RSS_SAMPLE_INT,
208 if (scf_set_count_property(tx, REPORT_INT,
215 if (scf_set_count_property(tx, WALK_INT,
222 if ((rval = scf_transaction_commit(tx)) == -1)
226 if (scf_transaction_restart(simple_h, tx)
238 if (tx !
[all...]
/illumos-gate/usr/src/uts/common/io/bge/
H A Dbge.conf175 bge-tx-rings = 1;
/illumos-gate/usr/src/lib/libmvec/common/
H A D__vcosbigf.c72 double tx, tt[3], ty[2], t, w, z, a; local
76 tx = *x;
77 hx = HI(&tx);
83 HI(&tx) = 0x41600000 | (hx & 0xfffff);
85 tt[0] = (double)((int) tx);
86 tx = (tx - tt[0]) * two24;
87 if (tx != zero)
90 tt[1] = (double)((int) tx);
91 tt[2] = (tx
[all...]

Completed in 148 milliseconds

12345678