/illumos-gate/usr/src/uts/common/fs/zfs/ |
H A D | txg.c | 285 cv_wait(&tx->tx_exit_cv, &tx->tx_sync_lock); in txg_sync_stop() 498 tx->tx_synced_txg >= tx->tx_sync_txg_waiting && in txg_sync_thread() 513 if (tx->tx_quiesce_txg_waiting < tx->tx_open_txg+1) in txg_sync_thread() 514 tx->tx_quiesce_txg_waiting = tx->tx_open_txg+1; in txg_sync_thread() 520 txg_thread_exit(tx, &cpr, &tx->tx_sync_thread); in txg_sync_thread() 580 txg_thread_exit(tx, &cpr, &tx->tx_quiesce_thread); in txg_quiesce_thread() 667 cv_wait(&tx->tx_sync_done_cv, &tx->tx_sync_lock); in txg_wait_synced_impl() 730 tx->tx_quiesce_txg_waiting <= tx->tx_open_txg && in txg_kick() 731 tx->tx_sync_txg_waiting <= tx->tx_synced_txg && in txg_kick() 732 tx->tx_quiesced_txg <= tx->tx_synced_txg) { in txg_kick() [all …]
|
H A D | dmu_tx.c | 59 return (tx); in dmu_tx_create_dd() 67 return (tx); in dmu_tx_create() 296 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_write() 310 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_remap_l1indirect() 446 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_free() 521 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_zap() 547 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_bonus() 571 txh = dmu_tx_hold_object_impl(tx, tx->tx_objset, in dmu_tx_hold_space() 905 tx->tx_txg = txg_hold_open(tx->tx_pool, &tx->tx_txgh); in dmu_tx_try_assign() 943 asize, tx->tx_netfree, &tx->tx_tempreserve_cookie, tx); in dmu_tx_try_assign() [all …]
|
H A D | dsl_destroy.c | 221 dmu_tx_t *tx) in dsl_destroy_snapshot_handle_remaps() argument 309 tx)); in dsl_destroy_snapshot_sync_impl() 359 tx); in dsl_destroy_snapshot_sync_impl() 488 tx)); in dsl_destroy_snapshot_sync_impl() 491 tx)); in dsl_destroy_snapshot_sync_impl() 637 dmu_tx_t *tx; member 646 dmu_tx_t *tx = ka->tx; in kill_blkptr() local 657 dsl_free(ka->tx->tx_pool, ka->tx->tx_txg, bp); in kill_blkptr() 681 ka.tx = tx; in old_synchronous_dataset_destroy() 833 0, tx); in dsl_destroy_head_sync_impl() [all …]
|
H A D | dmu_object.c | 48 int dnodesize, dmu_tx_t *tx) in dmu_object_alloc_impl() argument 205 bonuslen, 0, tx)); in dmu_object_alloc() 211 dmu_tx_t *tx) in dmu_object_alloc_ibs() argument 222 bonuslen, dnodesize, tx)); in dmu_object_alloc_dnsize() 230 bonuslen, 0, tx)); in dmu_object_claim() 291 keep_spill, tx); in dmu_object_reclaim_dnsize() 310 dbuf_rm_spill(dn, tx); in dmu_object_rm_spill() 311 dnode_rm_spill(dn, tx); in dmu_object_rm_spill() 338 dnode_free(dn, tx); in dmu_object_free() 416 dmu_tx_t *tx) in dmu_object_zapify() argument [all …]
|
H A D | dsl_deadlist.c | 173 bpobj_free(os, dlobj, tx); in dsl_deadlist_free() 182 bpobj_decr_empty(os, tx); in dsl_deadlist_free() 184 bpobj_free(os, obj, tx); in dsl_deadlist_free() 202 dle->dle_mintxg, obj, tx)); in dle_enqueue() 209 uint64_t obj, dmu_tx_t *tx) in dle_enqueue_subobj() argument 251 dle_enqueue(dl, dle, bp, tx); in dsl_deadlist_insert() 279 mintxg, obj, tx)); in dsl_deadlist_add_key() 438 dmu_tx_t *tx) in dsl_deadlist_insert_bpobj() argument 509 dmu_buf_will_dirty(bonus, tx); in dsl_deadlist_merge() 520 dmu_tx_t *tx) in dsl_deadlist_move_bpobj() argument [all …]
|
H A D | dsl_dataset.c | 89 dmu_tx_t *tx); 222 dsl_free(tx->tx_pool, tx->tx_txg, bp); in dsl_dataset_block_kill() 236 dsl_free(tx->tx_pool, tx->tx_txg, bp); in dsl_dataset_block_kill() 1123 dmu_tx_t *tx) in dsl_dataset_remove_from_next_clones() argument 1131 obj, tx); in dsl_dataset_remove_from_next_clones() 1434 dmu_tx_t *tx) in dsl_dataset_snapshot_sync_impl() argument 1820 &ds->ds_resume_object[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 1823 &ds->ds_resume_offset[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 1826 &ds->ds_resume_bytes[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 3577 f, tx); in dsl_dataset_clone_swap_sync_impl() [all …]
|
H A D | dsl_pool.c | 482 FREE_DIR_NAME, tx); in dsl_pool_create() 526 dmu_tx_commit(tx); in dsl_pool_create() 601 dmu_tx_t *tx; in dsl_pool_sync() local 724 dsl_dir_sync(dd, tx); in dsl_pool_sync() 744 dsl_pool_sync_mos(dp, tx); in dsl_pool_sync() 767 dmu_tx_commit(tx); in dsl_pool_sync() 917 dmu_tx_t *tx = arg; in upgrade_clones_cb() local 1008 dmu_tx_t *tx = arg; in upgrade_dir_clones_cb() local 1021 0, tx); in upgrade_dir_clones_cb() 1026 ds->ds_object, tx)); in upgrade_dir_clones_cb() [all …]
|
H A D | zfeature.c | 297 dmu_tx_t *tx) in feature_sync() argument 304 sizeof (uint64_t), 1, &refcount, tx)); in feature_sync() 353 feature->fi_desc, tx)); in feature_enable_sync() 358 uint64_t enabling_txg = dmu_tx_get_txg(tx); in feature_enable_sync() 364 DMU_POOL_FEATURE_ENABLED_TXG, tx); in feature_enable_sync() 389 dmu_tx_t *tx) in feature_do_action() argument 400 ASSERT(dmu_tx_is_syncing(tx)); in feature_do_action() 419 feature_sync(spa, feature, refcount, tx); in feature_do_action() 434 DMU_POOL_FEATURES_FOR_READ, tx); in spa_feature_create_zap_objects() 437 DMU_POOL_FEATURES_FOR_WRITE, tx); in spa_feature_create_zap_objects() [all …]
|
H A D | dsl_synctask.c | 37 dsl_null_checkfunc(void *arg, dmu_tx_t *tx) in dsl_null_checkfunc() argument 48 dmu_tx_t *tx; in dsl_sync_task_common() local 59 tx = dmu_tx_create_dd(dp->dp_mos_dir); in dsl_sync_task_common() 60 VERIFY0(dmu_tx_assign(tx, TXG_WAIT)); in dsl_sync_task_common() 63 dst.dst_txg = dmu_tx_get_txg(tx); in dsl_sync_task_common() 73 err = dst.dst_checkfunc(arg, tx); in dsl_sync_task_common() 77 dmu_tx_commit(tx); in dsl_sync_task_common() 86 dmu_tx_commit(tx); in dsl_sync_task_common() 90 sigfunc(arg, tx); in dsl_sync_task_common() 179 dst->dst_txg = dmu_tx_get_txg(tx); in dsl_sync_task_nowait_common() [all …]
|
H A D | spa_history.c | 106 dmu_buf_will_dirty(dbp, tx); in spa_history_create_obj() 152 dmu_tx_t *tx) in spa_history_write() argument 293 dmu_buf_will_dirty(dbp, tx); in spa_history_log_sync() 386 dmu_tx_t *tx; in spa_history_log_nvl() local 400 dmu_tx_abort(tx); in spa_history_log_nvl() 414 dmu_tx_commit(tx); in spa_history_log_nvl() 558 dmu_tx_t *htx = tx; in spa_history_log_internal() 562 if (tx == NULL) { in spa_history_log_internal() 575 if (tx == NULL) in spa_history_log_internal() 587 ASSERT(tx != NULL); in spa_history_log_internal_ds() [all …]
|
H A D | zfs_dir.c | 594 dmu_tx_t *tx; in zfs_purgedir() local 623 dmu_tx_abort(tx); in zfs_purgedir() 635 dmu_tx_commit(tx); in zfs_purgedir() 651 dmu_tx_t *tx; in zfs_rmnode() local 728 dmu_tx_abort(tx); in zfs_rmnode() 751 dmu_tx_commit(tx); in zfs_rmnode() 833 8, 1, &value, tx); in zfs_link_create() 883 tx); in zfs_dropname() 1016 dmu_tx_t *tx; in zfs_make_xattrdir() local 1046 dmu_tx_abort(tx); in zfs_make_xattrdir() [all …]
|
H A D | dsl_dir.c | 533 dsl_dir_zapify(dd, tx); in dsl_dir_init_fs_ss_count() 759 dsl_dir_zapify(dd, tx); in dsl_dir_update_last_remap_txg_sync() 875 dmu_tx_t *tx) in dsl_fs_ss_count_adjust() argument 919 tx)); in dsl_fs_ss_count_adjust() 928 dmu_tx_t *tx) in dsl_dir_create_sync() argument 1265 uint64_t txg = tx->tx_txg; in dsl_dir_tempreserve_impl() 1445 ASSERT3U(tx->tx_txg, !=, 0); in dsl_dir_tempreserve_clear() 1488 dsl_dir_dirty(dd, tx); in dsl_dir_willuse_space() 1743 delta, 0, 0, tx); in dsl_dir_set_reservation_sync_impl() 1911 if (dmu_tx_is_syncing(tx)) { in dsl_dir_rename_check() [all …]
|
H A D | bpobj.c | 52 &dp->dp_empty_bpobj, tx) == 0); in bpobj_alloc_empty() 72 DMU_POOL_EMPTY_BPOBJ, tx)); in bpobj_decr_empty() 91 DMU_OT_BPOBJ_HDR, size, tx)); in bpobj_alloc() 132 bpobj_free(os, objarray[blkoff], tx); in bpobj_free() 248 err = func(arg, bp, tx); in bpobj_iterate_impl() 333 objarray[blkoff], tx); in bpobj_iterate_impl() 394 bpobj_decr_empty(bpo->bpo_os, tx); in bpobj_enqueue_subobj() 404 bpobj_free(bpo->bpo_os, subobj, tx); in bpobj_enqueue_subobj() 413 DMU_OT_NONE, 0, tx); in bpobj_enqueue_subobj() 422 sizeof (subobj), &subobj, tx); in bpobj_enqueue_subobj() [all …]
|
H A D | dsl_bookmark.c | 121 dmu_tx_t *tx) in dsl_bookmark_create_check_impl() argument 123 dsl_pool_t *dp = dmu_tx_pool(tx); in dsl_bookmark_create_check_impl() 156 dsl_pool_t *dp = dmu_tx_pool(tx); in dsl_bookmark_create_check() 172 nvpair_name(pair), tx); in dsl_bookmark_create_check() 189 dsl_pool_t *dp = dmu_tx_pool(tx); in dsl_bookmark_create_sync() 211 dsl_dataset_zapify(bmark_fs, tx); in dsl_bookmark_create_sync() 215 &bmark_fs->ds_bookmarks, tx)); in dsl_bookmark_create_sync() 388 SPA_FEATURE_BOOKMARK_V2, tx); in dsl_dataset_bookmark_remove() 398 dsl_pool_t *dp = dmu_tx_pool(tx); in dsl_bookmark_destroy_check() 433 if (dmu_tx_is_syncing(tx)) { in dsl_bookmark_destroy_check() [all …]
|
H A D | zap.c | 141 dmu_buf_will_dirty(db, tx); in fzap_upgrade() 171 dmu_tx_t *tx) in zap_table_grow() argument 246 dmu_tx_t *tx) in zap_table_store() argument 579 ASSERT(tx != NULL); in zap_set_idx_to_blk() 587 idx, blk, tx)); in zap_set_idx_to_blk() 991 dnodesize, tx)) > 0); in zap_create_link_dnsize() 993 tx)); in zap_create_link_dnsize() 1062 8, 1, &value, tx); in zap_join_key() 1073 dmu_tx_t *tx) in zap_join_increment() argument 1160 dmu_tx_t *tx) in zap_increment() argument [all …]
|
H A D | zvol.c | 406 dmu_tx_t *tx; in zvol_replay_write() local 428 dmu_tx_abort(tx); in zvol_replay_write() 692 dmu_tx_t *tx; in zvol_prealloc() local 730 dmu_tx_t *tx; in zvol_update_volsize() local 741 dmu_tx_abort(tx); in zvol_update_volsize() 747 &volsize, tx); in zvol_update_volsize() 1769 dmu_tx_t *tx; in zvol_ioctl() local 1913 dmu_tx_t *tx; in zvol_dump_init() local 2014 &vbs, tx); in zvol_dump_init() 2074 dmu_tx_t *tx; in zvol_dumpify() local [all …]
|
H A D | spa_checkpoint.c | 321 sdc.sdc_txg = tx->tx_txg; in spa_checkpoint_discard_thread_sync() 353 space_map_free(vd->vdev_checkpoint_sm, tx); in spa_checkpoint_discard_thread_sync() 451 spa_checkpoint_check(void *arg, dmu_tx_t *tx) in spa_checkpoint_check() argument 453 spa_t *spa = dmu_tx_pool(tx)->dp_spa; in spa_checkpoint_check() 475 spa_checkpoint_sync(void *arg, dmu_tx_t *tx) in spa_checkpoint_sync() argument 477 dsl_pool_t *dp = dmu_tx_pool(tx); in spa_checkpoint_sync() 514 &checkpoint, tx)); in spa_checkpoint_sync() 524 spa_history_log_internal(spa, "spa checkpoint", tx, in spa_checkpoint_sync() 574 spa_t *spa = dmu_tx_pool(tx)->dp_spa; in spa_checkpoint_discard_check() 592 spa_t *spa = dmu_tx_pool(tx)->dp_spa; in spa_checkpoint_discard_sync() [all …]
|
/illumos-gate/usr/src/uts/common/io/bnx/570x/driver/common/lmdev/ |
H A D | bnx_lm_send.c | 72 packet->u1.tx.dbg_start_bd = txq->prod_bd; in lm_send_packet() 214 packet->u1.tx.span_pages = 0; in lm_send_packet() 251 packet->u1.tx.span_pages++; in lm_send_packet() 285 packet->u1.tx.next_bd_idx = prod_idx; in lm_send_packet() 407 pkt_info = packet->u1.tx.tx_pkt_info; in lm_send_packet() 408 packet->u1.tx.size = 0; in lm_send_packet() 532 packet->u1.tx.span_pages = 0; in lm_send_packet() 555 packet->u1.tx.size += frag->size; in lm_send_packet() 603 packet->u1.tx.next_bd_idx = prod_idx; in lm_send_packet() 607 txq->prod_bseq += packet->u1.tx.size; in lm_send_packet() [all …]
|
/illumos-gate/usr/src/uts/common/fs/zfs/sys/ |
H A D | dmu_tx.h | 117 int dmu_tx_assign(dmu_tx_t *tx, uint64_t txg_how); 118 void dmu_tx_commit(dmu_tx_t *tx); 119 void dmu_tx_abort(dmu_tx_t *tx); 120 uint64_t dmu_tx_get_txg(dmu_tx_t *tx); 121 struct dsl_pool *dmu_tx_pool(dmu_tx_t *tx); 122 void dmu_tx_wait(dmu_tx_t *tx); 137 int dmu_tx_is_syncing(dmu_tx_t *tx); 138 int dmu_tx_private_ok(dmu_tx_t *tx); 141 void dmu_tx_hold_space(dmu_tx_t *tx, uint64_t space); 144 #define DMU_TX_DIRTY_BUF(tx, db) dmu_tx_dirty_buf(tx, db) argument [all …]
|
H A D | zap.h | 130 int dnodesize, dmu_tx_t *tx); 137 int dnodesize, dmu_tx_t *tx); 241 const void *val, dmu_tx_t *tx); 244 const void *val, dmu_tx_t *tx); 247 const void *val, dmu_tx_t *tx); 283 matchtype_t mt, dmu_tx_t *tx); 286 int key_numints, dmu_tx_t *tx); 312 uint64_t value, dmu_tx_t *tx); 316 dmu_tx_t *tx); 326 dmu_tx_t *tx); [all …]
|
H A D | dmu.h | 327 cred_t *cr, dmu_tx_t *tx); 465 dmu_tx_t *tx); 479 int ibs, dmu_tx_t *tx); 487 dmu_tx_t *tx); 494 dmu_tx_t *tx); 501 dmu_tx_t *tx); 798 void dmu_tx_wait(dmu_tx_t *tx); 850 dmu_tx_t *tx); 855 dmu_tx_t *tx); 857 dmu_tx_t *tx); [all …]
|
/illumos-gate/usr/src/cmd/refer/ |
H A D | refer6.c | 31 char *s, *tx; in putref() local 46 tx = s+3; in putref() 50 tx = s+2; in putref() 55 tx = s; in putref() 61 tx = caps(tx, buf1); in putref() 63 fprintf(stderr, " s %o tx %o %s\n", s, tx, tx); in putref() 71 lastype, tx, sep); in putref() 74 last(tx)) != 0); in putref() 86 tx = revauth(tx, buf2); in putref() 93 cch, tx, sep); in putref() [all …]
|
H A D | tick.c | 30 struct tbuffer tx; in tick() local 32 times(&tx); in tick() 34 user = tx.proc_user_time; in tick() 35 systm = tx.proc_system_time; in tick() 42 struct tbuffer tx; in tock() local 47 times(&tx); in tock() 50 use = (tx.proc_user_time - user)/60.; in tock() 51 sys = (tx.proc_system_time - systm)/60.; in tock()
|
/illumos-gate/usr/src/uts/common/io/nxge/ |
H A D | nxge_hv.c | 53 nxhv_dc_fp_t *tx; in nxge_hio_hv_init() local 69 tx = &nhd->hio.tx; in nxge_hio_hv_init() 71 tx->assign = &hv_niu_tx_dma_assign; in nxge_hio_hv_init() 72 tx->unassign = &hv_niu_tx_dma_unassign; in nxge_hio_hv_init() 73 tx->get_map = &hv_niu_vr_get_txmap; in nxge_hio_hv_init() 76 tx->lp_conf = &hv_niu_tx_logical_page_conf; in nxge_hio_hv_init() 77 tx->lp_info = &hv_niu_tx_logical_page_info; in nxge_hio_hv_init() 79 tx->lp_cfgh_conf = &hv_niu_cfgh_tx_logical_page_conf; in nxge_hio_hv_init() 80 tx->lp_cfgh_info = &hv_niu_cfgh_tx_logical_page_info; in nxge_hio_hv_init() 82 tx->getinfo = &hv_niu_vrtx_getinfo; in nxge_hio_hv_init()
|
/illumos-gate/usr/src/uts/common/io/igc/ |
H A D | igc_ring.c | 1294 tx->itx_buf_rem = 0; in igc_tx_ring_bind() 1296 list_insert_tail(&tx->itx_bufs, tx->itx_cur_buf); in igc_tx_ring_bind() 1321 if (tx->itx_cur_buf != NULL && (tx->itx_cur_buf->itb_bind || in igc_tx_ring_copy() 1332 list_insert_tail(&tx->itx_bufs, tx->itx_cur_buf); in igc_tx_ring_copy() 1333 tx->itx_buf_rem = tx->itx_cur_buf->itb_dma.idb_size; in igc_tx_ring_copy() 1339 tx->itx_ndescs++; in igc_tx_ring_copy() 1476 if (tx->itx_cksum != 0 || tx->itx_lso != 0) { in igc_tx_ring_write_descs() 1478 tx->itx_ndescs++; in igc_tx_ring_write_descs() 1503 tx->itx_ndescs--; in igc_tx_ring_write_descs() 1598 tx->itx_meoi.meoi_l2hlen - tx->itx_meoi.meoi_l3hlen - in igc_tx_ring_write_descs() [all …]
|