/linux/fs/gfs2/ |
H A D | inode.c | 136 &ip->i_gl); in gfs2_inode_lookup() 192 glock_set_object(ip->i_gl, ip); in gfs2_inode_lookup() 229 if (ip->i_gl) { in gfs2_inode_lookup() 230 gfs2_glock_put(ip->i_gl); in gfs2_inode_lookup() 231 ip->i_gl = NULL; in gfs2_inode_lookup() 779 glock_set_object(ip->i_gl, ip); in gfs2_create_inode() 824 glock_clear_object(ip->i_gl, ip); in gfs2_create_inode() 831 if (ip->i_gl) { in gfs2_create_inode() 832 gfs2_glock_put(ip->i_gl); in gfs2_create_inode() 833 ip->i_gl = NULL; in gfs2_create_inode() [all …]
|
H A D | super.c | 136 struct gfs2_glock *j_gl = ip->i_gl; in gfs2_make_fs_rw() 261 gfs2_trans_add_meta(m_ip->i_gl, m_bh); in update_statfs() 500 if (unlikely(!ip->i_gl)) { in gfs2_dirty_inode() 530 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_dirty_inode() 1179 struct gfs2_glock *gl = ip->i_gl; in gfs2_final_release_pages() 1411 if (!ret && ip->i_gl) in evict_unlinked_inode() 1438 gfs2_ail_flush(ip->i_gl, 0); in evict_linked_inode() 1524 if (ip->i_gl) { in gfs2_evict_inode() 1525 glock_clear_object(ip->i_gl, ip); in gfs2_evict_inode() 1527 gfs2_glock_add_to_lru(ip->i_gl); in gfs2_evict_inode() [all …]
|
H A D | xattr.c | 282 gfs2_trans_add_meta(ip->i_gl, bh); in ea_dealloc_unstuffed() 508 gfs2_trans_add_meta(ip->i_gl, bh[x]); in gfs2_iter_unstuffed() 650 *bhp = gfs2_meta_new(ip->i_gl, block); in ea_alloc_blk() 651 gfs2_trans_add_meta(ip->i_gl, *bhp); in ea_alloc_blk() 712 bh = gfs2_meta_new(ip->i_gl, block); in ea_write() 713 gfs2_trans_add_meta(ip->i_gl, bh); in ea_write() 885 gfs2_trans_add_meta(ip->i_gl, bh); in ea_set_simple_noalloc() 1005 gfs2_trans_add_meta(ip->i_gl, indbh); in ea_set_block() 1013 indbh = gfs2_meta_new(ip->i_gl, blk); in ea_set_block() 1336 gfs2_trans_add_meta(ip->i_gl, indbh); in ea_dealloc_indirect() [all …]
|
H A D | util.c | 61 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_NOEXP | in check_journal_clean() 123 struct gfs2_glock *i_gl; in signal_our_withdraw() local 134 i_gl = ip->i_gl; in signal_our_withdraw() 189 wait_on_bit(&i_gl->gl_flags, GLF_DEMOTE, TASK_UNINTERRUPTIBLE); in signal_our_withdraw() 209 if (i_gl->gl_ops->go_free) { in signal_our_withdraw() 210 set_bit(GLF_FREEING, &i_gl->gl_flags); in signal_our_withdraw() 211 wait_on_bit(&i_gl->gl_flags, GLF_FREEING, TASK_UNINTERRUPTIBLE); in signal_our_withdraw() 457 gfs2_dump_glock(NULL, ip->i_gl, 1); in gfs2_consist_inode_i()
|
H A D | acl.c | 70 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { in gfs2_get_acl() 71 int ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in gfs2_get_acl() 129 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { in gfs2_set_acl() 130 ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in gfs2_set_acl()
|
H A D | file.c | 169 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_fileattr_get() 245 gfs2_log_flush(sdp, ip->i_gl, in do_gfs2_set_flags() 264 gfs2_trans_add_meta(ip->i_gl, bh); in do_gfs2_set_flags() 432 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in gfs2_page_mkwrite() 457 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); in gfs2_page_mkwrite() 559 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_fault() 769 gfs2_ail_flush(ip->i_gl, 1); in gfs2_fsync() 839 gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, gh); in gfs2_file_direct_read() 902 gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, gh); in gfs2_file_direct_write() 982 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_file_read_iter() [all …]
|
H A D | dir.c | 94 bh = gfs2_meta_new(ip->i_gl, block); in gfs2_dir_get_new_buffer() 95 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_dir_get_new_buffer() 129 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_dir_write_stuffed() 210 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_dir_write_data() 232 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_dir_write_data() 681 gfs2_trans_add_meta(dip->i_gl, bh); in dirent_del() 720 gfs2_trans_add_meta(ip->i_gl, bh); in do_init_dirent() 877 bh = gfs2_meta_new(ip->i_gl, bn); in new_leaf() 882 gfs2_trans_add_meta(ip->i_gl, bh); in new_leaf() 1486 struct gfs2_glock *gl = ip->i_gl; in gfs2_dir_readahead() [all …]
|
H A D | dentry.c | 63 had_lock = (gfs2_glock_is_locked_by_me(dip->i_gl) != NULL); in gfs2_drevalidate() 65 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_drevalidate()
|
H A D | bmap.c | 81 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_unstuffer_folio() 127 gfs2_trans_add_meta(ip->i_gl, dibh); in __gfs2_unstuff_inode() 680 gfs2_trans_add_meta(ip->i_gl, dibh); in __gfs2_iomap_alloc() 752 gfs2_indirect_init(mp, ip->i_gl, i, in __gfs2_iomap_alloc() 1384 gfs2_trans_add_meta(ip->i_gl, dibh); in trunc_start() 1556 gfs2_trans_add_meta(ip->i_gl, bh); in sweep_bh_for_rgrps() 1590 gfs2_trans_add_meta(ip->i_gl, dibh); in sweep_bh_for_rgrps() 1957 gfs2_trans_add_meta(ip->i_gl, dibh); in punch_hole() 2002 gfs2_trans_add_meta(ip->i_gl, dibh); in trunc_end() 2101 gfs2_trans_add_meta(ip->i_gl, dibh); in do_grow() [all …]
|
H A D | aops.c | 57 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_trans_add_databufs() 157 if (gfs2_assert_withdraw(sdp, ip->i_gl->gl_state == LM_ST_EXCLUSIVE)) in gfs2_jdata_writepage() 397 gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL | in gfs2_jdata_writepages() 595 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_bmap()
|
H A D | meta_io.c | 437 if (!ip->i_gl) { in gfs2_journal_wipe() 446 bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); in gfs2_journal_wipe() 481 struct gfs2_glock *gl = ip->i_gl; in gfs2_meta_buffer()
|
H A D | lops.c | 784 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() 839 gfs2_inode_metasync(ip->i_gl); in buf_lo_after_scan() 845 gfs2_inode_metasync(ip->i_gl); in buf_lo_after_scan() 1008 struct gfs2_glock *gl = ip->i_gl; in databuf_lo_scan_elements() 1059 gfs2_inode_metasync(ip->i_gl); in databuf_lo_after_scan() 1066 gfs2_inode_metasync(ip->i_gl); in databuf_lo_after_scan()
|
H A D | recovery.c | 36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block() 351 gfs2_inode_metasync(ip->i_gl); in update_statfs_inode() 445 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in gfs2_recover_func()
|
H A D | quota.c | 426 error = gfs2_meta_read(ip->i_gl, iomap.addr >> inode->i_blkbits, in bh_get() 721 gfs2_trans_add_meta(ip->i_gl, qd->qd_bh); in do_qc() 800 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_write_buf_to_page() 952 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in do_sync() 1010 gfs2_log_flush(ip->i_gl->gl_name.ln_sbd, ip->i_gl, in do_sync() 1070 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); in do_glock() 1440 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); in gfs2_quota_init() 1728 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in gfs2_set_dqblk()
|
H A D | trace_gfs2.h | 455 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; 491 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; 523 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev;
|
H A D | export.c | 112 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_get_name()
|
H A D | ops_fstype.c | 561 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh); in gfs2_jindex_hold() 675 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID, in init_statfs() 770 sdp->sd_jinode_gl = ip->i_gl; in init_journal() 771 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in init_journal() 943 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID, in init_per_node()
|
H A D | incore.h | 391 struct gfs2_glock *i_gl; member
|
H A D | glops.c | 607 struct gfs2_glock *j_gl = ip->i_gl; in freeze_go_xmote_bh()
|
H A D | rgrp.c | 1037 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() 2458 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_alloc_blocks()
|
H A D | glock.c | 699 if (gl == m_ip->i_gl) in is_system_glock() 1026 gfs2_glock_poke(ip->i_gl); in gfs2_try_evict()
|