/linux/fs/gfs2/ |
H A D | glock.c | 257 GLOCK_BUG_ON(gl, gl->gl_lockref.count < 2); in gfs2_glock_queue_work() 622 gl->gl_target = gl->gl_state; in finish_xmote() 631 do_xmote(gl, gh, gl->gl_target); in finish_xmote() 701 GLOCK_BUG_ON(gl, gl->gl_state == target); in do_xmote() 702 GLOCK_BUG_ON(gl, gl->gl_state == gl->gl_target); in do_xmote() 862 gl->gl_target = gl->gl_demote_state; in run_queue() 873 do_xmote(gl, gh, gl->gl_target); in run_queue() 1088 finish_xmote(gl, gl->gl_reply); in glock_work_func() 1458 gl->gl_ops->go_callback(gl, remote); in request_demote() 2541 struct gfs2_glock *gl = gi->gl; in gfs2_glock_iter_next() local [all …]
|
H A D | glops.c | 44 gl->gl_name.ln_type, gl->gl_name.ln_number, in gfs2_ail_error() 45 gfs2_glock2aspace(gl)); in gfs2_ail_error() 77 gfs2_ail_error(gl, bh); in __gfs2_ail_flush() 82 GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count)); in __gfs2_ail_flush() 204 GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); in rgrp_go_sync() 256 ip = gl->gl_object; in gfs2_glock2inode() 268 rgd = gl->gl_object; in gfs2_glock2rgrp() 321 GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); in inode_go_sync() 323 gfs2_log_flush(gl->gl_name.ln_sbd, gl, GFS2_LOG_HEAD_FLUSH_NORMAL | in inode_go_sync() 627 gl->gl_lockref.count++; in iopen_go_callback() [all …]
|
H A D | glock.h | 127 void (*lm_put_lock) (struct gfs2_glock *gl); 130 void (*lm_cancel) (struct gfs2_glock *gl); 145 spin_lock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me() 155 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me() 162 if (gl->gl_ops->go_flags & GLOF_ASPACE) { in gfs2_glock2aspace() 174 void gfs2_glock_put(struct gfs2_glock *gl); 175 void gfs2_glock_put_async(struct gfs2_glock *gl); 206 gfs2_dump_glock(NULL, gl, true); \ 209 gfs2_dump_glock(NULL, gl, true); \ 213 gfs2_dump_glock(NULL, gl, true); \ [all …]
|
H A D | lock_dlm.c | 110 dstamp = gl->gl_dstamp; in gfs2_update_request_times() 132 if ((gl->gl_lksb.sb_flags & DLM_SBF_VALNOTVALID) && gl->gl_lksb.sb_lvbptr) in gdlm_ast() 138 gl->gl_ops->go_unlocked(gl); in gdlm_ast() 139 gfs2_glock_free(gl); in gdlm_ast() 156 ret = gl->gl_req; in gdlm_ast() 181 gl->gl_lksb.sb_lkid = 0; in gdlm_ast() 315 gfs2_glock_free(gl); in gdlm_put_lock() 326 gfs2_glock_free(gl); in gdlm_put_lock() 338 (!gl->gl_lksb.sb_lvbptr || gl->gl_state != LM_ST_EXCLUSIVE)) { in gdlm_put_lock() 345 NULL, gl); in gdlm_put_lock() [all …]
|
H A D | trace_gfs2.h | 94 TP_ARGS(gl, new_state), 109 __entry->glnum = gl->gl_name.ln_number; 110 __entry->gltype = gl->gl_name.ln_type; 115 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0); 131 TP_PROTO(const struct gfs2_glock *gl), 133 TP_ARGS(gl), 145 __entry->gltype = gl->gl_name.ln_type; 148 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0); 165 TP_ARGS(gl, remote), 183 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0); [all …]
|
H A D | meta_io.c | 116 struct address_space *mapping = gfs2_glock2aspace(gl); in gfs2_getbuf() 117 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_getbuf() 184 bh = gfs2_getbuf(gl, blkno, CREATE); in gfs2_meta_new() 251 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_meta_read() 261 *bhp = bh = gfs2_getbuf(gl, blkno, CREATE); in gfs2_meta_read() 274 bh = gfs2_getbuf(gl, blkno + 1, CREATE); in gfs2_meta_read() 481 struct gfs2_glock *gl = ip->i_gl; in gfs2_meta_buffer() local 489 ret = gfs2_meta_read(gl, num, DIO_WAIT, rahead, &bh); in gfs2_meta_buffer() 510 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_meta_ra() 522 first_bh = gfs2_getbuf(gl, dblock, CREATE); in gfs2_meta_ra() [all …]
|
H A D | main.c | 52 struct gfs2_glock *gl = foo; in gfs2_init_glock_once() local 54 spin_lock_init(&gl->gl_lockref.lock); in gfs2_init_glock_once() 55 INIT_LIST_HEAD(&gl->gl_holders); in gfs2_init_glock_once() 56 INIT_LIST_HEAD(&gl->gl_lru); in gfs2_init_glock_once() 57 INIT_LIST_HEAD(&gl->gl_ail_list); in gfs2_init_glock_once() 58 atomic_set(&gl->gl_ail_count, 0); in gfs2_init_glock_once() 59 atomic_set(&gl->gl_revokes, 0); in gfs2_init_glock_once()
|
H A D | incore.h | 38 typedef void (*gfs2_glop_bh_t) (struct gfs2_glock *gl, unsigned int ret); 218 int (*go_sync) (struct gfs2_glock *gl); 219 int (*go_xmote_bh)(struct gfs2_glock *gl); 220 void (*go_inval) (struct gfs2_glock *gl, int flags); 221 int (*go_instantiate) (struct gfs2_glock *gl); 223 void (*go_dump)(struct seq_file *seq, const struct gfs2_glock *gl, 225 void (*go_callback)(struct gfs2_glock *gl, bool remote); 226 void (*go_unlocked)(struct gfs2_glock *gl); 854 gl->gl_stats.stats[which]++; in gfs2_glstats_inc() 859 const struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_sbstats_inc() [all …]
|
H A D | trans.c | 163 static struct gfs2_bufdata *gfs2_alloc_bufdata(struct gfs2_glock *gl, in gfs2_alloc_bufdata() argument 170 bd->bd_gl = gl; in gfs2_alloc_bufdata() 192 void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh) in gfs2_trans_add_data() argument 195 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_trans_add_data() 209 bd = gfs2_alloc_bufdata(gl, bh); in gfs2_trans_add_data() 215 gfs2_assert(sdp, bd->bd_gl == gl); in gfs2_trans_add_data() 229 void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh) in gfs2_trans_add_meta() argument 232 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_trans_add_meta() 251 bd = gfs2_alloc_bufdata(gl, bh); in gfs2_trans_add_meta() 258 gfs2_assert(sdp, bd->bd_gl == gl); in gfs2_trans_add_meta()
|
H A D | meta_io.h | 53 struct buffer_head *gfs2_meta_new(struct gfs2_glock *gl, u64 blkno); 54 int gfs2_meta_read(struct gfs2_glock *gl, u64 blkno, int flags, 57 struct buffer_head *gfs2_getbuf(struct gfs2_glock *gl, u64 blkno, 75 struct buffer_head *gfs2_meta_ra(struct gfs2_glock *gl, u64 dblock, u32 extlen);
|
H A D | glops.h | 25 int gfs2_inode_metasync(struct gfs2_glock *gl); 26 void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync);
|
H A D | super.c | 1047 gfs2_glock_hold(gl); in gfs2_drop_inode() 1048 if (!gfs2_queue_try_to_evict(gl)) in gfs2_drop_inode() 1049 gfs2_glock_put_async(gl); in gfs2_drop_inode() 1179 struct gfs2_glock *gl = ip->i_gl; in gfs2_final_release_pages() local 1181 if (unlikely(!gl)) { in gfs2_final_release_pages() 1192 clear_bit(GLF_DIRTY, &gl->gl_flags); in gfs2_final_release_pages() 1255 gfs2_glock_put_async(gl); in gfs2_glock_put_eventually() 1257 gfs2_glock_put(gl); in gfs2_glock_put_eventually() 1518 glock_clear_object(gl, ip); in gfs2_evict_inode() 1519 gfs2_glock_hold(gl); in gfs2_evict_inode() [all …]
|
H A D | lops.c | 73 struct gfs2_glock *gl = bd->bd_gl; in maybe_release_space() local 74 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in maybe_release_space() 122 struct gfs2_glock *gl = bd->bd_gl; in gfs2_unpin() local 124 atomic_inc(&gl->gl_ail_count); in gfs2_unpin() 784 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() local 807 bh_ip = gfs2_meta_new(gl, blkno); in buf_lo_scan_elements() 894 struct gfs2_glock *gl; in gfs2_drain_revokes() local 899 gl = bd->bd_gl; in gfs2_drain_revokes() 900 gfs2_glock_remove_revoke(gl); in gfs2_drain_revokes() 1008 struct gfs2_glock *gl = ip->i_gl; in databuf_lo_scan_elements() local [all …]
|
H A D | trans.h | 44 void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh); 45 void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh);
|
H A D | rgrp.c | 720 struct gfs2_glock *gl; in gfs2_clear_rgrpd() local 724 gl = rgd->rd_gl; in gfs2_clear_rgrpd() 728 if (gl) { in gfs2_clear_rgrpd() 731 flush_delayed_work(&gl->gl_work); in gfs2_clear_rgrpd() 734 glock_clear_object(gl, rgd); in gfs2_clear_rgrpd() 735 gfs2_glock_put(gl); in gfs2_clear_rgrpd() 1037 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() local 1846 struct gfs2_glock *gl; in try_rgrp_unlink() local 1880 ip = gl->gl_object; in try_rgrp_unlink() 1883 gfs2_glock_put(gl); in try_rgrp_unlink() [all …]
|
H A D | log.c | 106 struct gfs2_glock *gl = NULL; in gfs2_ail1_start_one() local 135 if (gl == bd->bd_gl) in gfs2_ail1_start_one() 137 gl = bd->bd_gl; in gfs2_ail1_start_one() 772 struct gfs2_glock *gl = bd->bd_gl; in gfs2_add_revoke() local 775 if (atomic_inc_return(&gl->gl_revokes) == 1) in gfs2_add_revoke() 776 gfs2_glock_hold(gl); in gfs2_add_revoke() 781 set_bit(GLF_LFLUSH, &gl->gl_flags); in gfs2_add_revoke() 787 if (atomic_dec_return(&gl->gl_revokes) == 0) { in gfs2_glock_remove_revoke() 788 clear_bit(GLF_LFLUSH, &gl->gl_flags); in gfs2_glock_remove_revoke() 789 gfs2_glock_put_async(gl); in gfs2_glock_remove_revoke() [all …]
|
H A D | log.h | 86 void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl, 94 void gfs2_glock_remove_revoke(struct gfs2_glock *gl);
|
H A D | inode.c | 893 struct gfs2_glock *gl; in __gfs2_lookup() local 904 gl = GFS2_I(inode)->i_gl; in __gfs2_lookup() 905 error = gfs2_glock_nq_init(gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); in __gfs2_lookup() 1874 struct gfs2_glock *gl; in gfs2_permission() local 1879 gl = rcu_dereference_check(ip->i_gl, !may_not_block); in gfs2_permission() 1880 if (unlikely(!gl)) { in gfs2_permission() 1885 if (gfs2_glock_is_locked_by_me(gl) == NULL) { in gfs2_permission() 1888 error = gfs2_glock_nq_init(gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_permission() 2161 struct gfs2_glock *gl = ip->i_gl; in gfs2_update_time() local 2165 gh = gfs2_glock_is_locked_by_me(gl); in gfs2_update_time() [all …]
|
/linux/drivers/target/iscsi/cxgbit/ |
H A D | cxgbit_main.c | 231 gl->frags[i].page, in cxgbit_copy_frags() 232 gl->frags[i].offset, in cxgbit_copy_frags() 233 gl->frags[i].size); in cxgbit_copy_frags() 238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags() 260 if (unlikely(gl->nfrags > 1)) in cxgbit_lro_add_packet_gl() 308 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl() 413 if ((gl && (((skb_shinfo(skb)->nr_frags + gl->nfrags) > in cxgbit_lro_receive() 420 if (gl) in cxgbit_lro_receive() 462 rpl = gl ? (struct cpl_tx_data *)gl->va : in cxgbit_uld_lro_rx_handler() 474 if (!gl) { in cxgbit_uld_lro_rx_handler() [all …]
|
/linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 1486 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() 1521 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in t4vf_pktgl_to_skb() 1550 frag = gl->nfrags - 1; in t4vf_pktgl_free() 1575 t4vf_pktgl_free(gl); in do_gro() 1630 do_gro(rxq, gl, pkt); in t4vf_ethrx_handler() 1639 t4vf_pktgl_free(gl); in t4vf_ethrx_handler() 1777 struct pkt_gl gl; in process_responses() local 1799 gl.tot_len = len; in process_responses() 1817 gl.nfrags = frag+1; in process_responses() 1827 gl.va = (page_address(gl.frags[0].page) + in process_responses() [all …]
|
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_main.c | 339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument 349 skb = alloc_skb(gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 353 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 358 , gl->va + pktshift, in copy_gl_to_skb_pkt() 359 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 365 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument 371 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet() 439 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument 448 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler() 453 if (!gl) in chtls_uld_rx_handler() [all …]
|
/linux/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 1137 gl->va + pktshift, in copy_gl_to_skb_pkt() 1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 1167 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument 1174 if (gl == NULL) { in c4iw_uld_rx_handler() 1183 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler() 1189 } else if (unlikely(*(u8 *)rsp != *(u8 *)gl->va)) { in c4iw_uld_rx_handler() 1190 if (recv_rx_pkt(dev, gl, rsp)) in c4iw_uld_rx_handler() 1194 pci_name(ctx->lldi.pdev), gl->va, in c4iw_uld_rx_handler() 1196 be64_to_cpu(*(__force __be64 *)gl->va), in c4iw_uld_rx_handler() 1197 gl->tot_len); in c4iw_uld_rx_handler() [all …]
|
/linux/drivers/gpu/drm/ci/ |
H A D | image-tags.yml | 12 DEBIAN_X86_64_TEST_IMAGE_GL_PATH: "debian/x86_64_test-gl"
|
H A D | test.yml | 347 - .test-gl 358 - debian/x86_64_test-gl 368 - .test-gl 380 - debian/x86_64_test-gl
|
/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 3316 gl->frags[i].offset, in copy_frags() 3317 gl->frags[i].size); in copy_frags() 3320 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() 3346 __skb_put(skb, gl->tot_len); in cxgb4_pktgl_to_skb() 3347 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in cxgb4_pktgl_to_skb() 3355 copy_frags(skb, gl, pull_len); in cxgb4_pktgl_to_skb() 3356 skb->len = gl->tot_len; in cxgb4_pktgl_to_skb() 3376 for (p = gl->frags, n = gl->nfrags - 1; n--; p++) in t4_pktgl_free() 3391 t4_pktgl_free(gl); in handle_trace_pkt() 3440 t4_pktgl_free(gl); in do_gro() [all …]
|