Lines Matching defs:txg
849 vdev_metaslab_init(vdev_t *vd, uint64_t txg)
859 ASSERT(txg == 0 || spa_config_held(spa, SCL_ALLOC, RW_WRITER));
893 if (txg == 0) {
901 error = metaslab_init(vd->vdev_mg, m, object, txg,
907 if (txg == 0)
918 if (txg == 0)
1388 uint64_t txg = spa_last_synced_txg(spa) != 0 ?
1391 if ((label = vdev_label_read_config(vd, txg)) == NULL) {
1583 vdev_create(vdev_t *vd, uint64_t txg, boolean_t isreplacing)
1603 (error = vdev_label_init(vd, txg, isreplacing ?
1623 vdev_dirty(vdev_t *vd, int flags, void *arg, uint64_t txg)
1631 (void) txg_list_add(&vd->vdev_ms_list, arg, txg);
1634 (void) txg_list_add(&vd->vdev_dtl_list, arg, txg);
1636 (void) txg_list_add(&vd->vdev_spa->spa_vdev_txg_list, vd, txg);
1640 vdev_dirty_leaves(vdev_t *vd, int flags, uint64_t txg)
1643 vdev_dirty_leaves(vd->vdev_child[c], flags, txg);
1646 vdev_dirty(vd->vdev_top, flags, vd, txg);
1688 vdev_dtl_dirty(vdev_t *vd, vdev_dtl_type_t t, uint64_t txg, uint64_t size)
1697 if (!range_tree_contains(rt, txg, size))
1698 range_tree_add(rt, txg, size);
1703 vdev_dtl_contains(vdev_t *vd, vdev_dtl_type_t t, uint64_t txg, uint64_t size)
1713 dirty = range_tree_contains(rt, txg, size);
1733 * Returns the lowest txg in the DTL range.
1749 * Returns the highest txg in the DTL.
1787 * value to the highest txg value that exists in all DTLs. If this
1805 vdev_dtl_reassess(vdev_t *vd, uint64_t txg, uint64_t scrub_txg, int scrub_done)
1814 vdev_dtl_reassess(vd->vdev_child[c], txg,
1846 * that each txg in that range has refcnt -1 or 0.
1885 if (txg != 0)
1886 vdev_dirty(vd->vdev_top, VDD_DTL, vd, txg);
1999 vdev_dtl_sync(vdev_t *vd, uint64_t txg)
2012 tx = dmu_tx_create_assigned(spa->spa_dsl_pool, txg);
2071 zfs_dbgmsg("txg %llu, spa %s, DTL old object %llu, "
2072 "new object %llu", txg, spa_name(spa), object,
2119 * Determine if resilver is needed, and if so the txg range.
2227 vdev_remove(vdev_t *vd, uint64_t txg)
2233 tx = dmu_tx_create_assigned(spa_get_dsl(spa), txg);
2235 ASSERT3U(txg, ==, spa_syncing_txg(spa));
2286 vdev_sync_done(vdev_t *vd, uint64_t txg)
2289 boolean_t reassess = !txg_list_empty(&vd->vdev_ms_list, TXG_CLEAN(txg));
2293 while (msp = txg_list_remove(&vd->vdev_ms_list, TXG_CLEAN(txg)))
2294 metaslab_sync_done(msp, txg);
2301 vdev_sync(vdev_t *vd, uint64_t txg)
2312 tx = dmu_tx_create_assigned(spa->spa_dsl_pool, txg);
2324 vdev_remove(vd, txg);
2326 while ((msp = txg_list_remove(&vd->vdev_ms_list, txg)) != NULL) {
2327 metaslab_sync(msp, txg);
2328 (void) txg_list_add(&vd->vdev_ms_list, msp, TXG_CLEAN(txg));
2331 while ((lvd = txg_list_remove(&vd->vdev_dtl_list, txg)) != NULL)
2332 vdev_dtl_sync(lvd, txg);
2334 (void) txg_list_add(&spa->spa_vdev_txg_list, vd, TXG_CLEAN(txg));
2812 uint64_t txg = zio->io_txg;
2903 if (type == ZIO_TYPE_WRITE && txg != 0 &&
2910 * made by zil_claim() during spa_load() in the first txg.
2912 * txg as the block was born. In the scrub-induced repair
2923 uint64_t commit_txg = txg;
2927 vdev_dtl_dirty(vd, DTL_SCRUB, txg, 1);
2934 if (vdev_dtl_contains(vd, DTL_MISSING, txg, 1))
2937 vdev_dtl_dirty(pvd, DTL_PARTIAL, txg, 1);
2941 vdev_dtl_dirty(vd, DTL_MISSING, txg, 1);
3399 vdev_expand(vdev_t *vd, uint64_t txg)
3405 VERIFY(vdev_metaslab_init(vd, txg) == 0);