i_gl 66 fs/gfs2/acl.c if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { i_gl 67 fs/gfs2/acl.c int ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, i_gl 123 fs/gfs2/acl.c if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { i_gl 124 fs/gfs2/acl.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 57 fs/gfs2/aops.c gfs2_trans_add_data(ip->i_gl, bh); i_gl 98 fs/gfs2/aops.c if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) i_gl 188 fs/gfs2/aops.c if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) i_gl 423 fs/gfs2/aops.c gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL | i_gl 524 fs/gfs2/aops.c gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 608 fs/gfs2/aops.c gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 695 fs/gfs2/aops.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); i_gl 94 fs/gfs2/bmap.c gfs2_trans_add_data(ip->i_gl, bh); i_gl 158 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 682 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 726 fs/gfs2/bmap.c gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++); i_gl 752 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[i-1]); i_gl 754 fs/gfs2/bmap.c gfs2_indirect_init(mp, ip->i_gl, i, i_gl 764 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[end_of_metadata]); i_gl 996 fs/gfs2/bmap.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &ip->i_gh); i_gl 1003 fs/gfs2/bmap.c error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, i_gl 1251 fs/gfs2/bmap.c set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); i_gl 1427 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 1597 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 1632 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 1840 fs/gfs2/bmap.c gfs2_metapath_ra(ip->i_gl, start, end); i_gl 1960 fs/gfs2/bmap.c gfs2_metapath_ra(ip->i_gl, start, end); i_gl 1995 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 2040 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 2147 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 2397 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 63 fs/gfs2/dentry.c had_lock = (gfs2_glock_is_locked_by_me(dip->i_gl) != NULL); i_gl 65 fs/gfs2/dentry.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); i_gl 97 fs/gfs2/dir.c bh = gfs2_meta_new(ip->i_gl, block); i_gl 98 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 111 fs/gfs2/dir.c error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh); i_gl 132 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 213 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 235 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 305 fs/gfs2/dir.c bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); i_gl 307 fs/gfs2/dir.c error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh); i_gl 682 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, bh); i_gl 721 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 764 fs/gfs2/dir.c error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, 0, bhp); i_gl 879 fs/gfs2/dir.c bh = gfs2_meta_new(ip->i_gl, bn); i_gl 884 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 973 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, dibh); i_gl 1033 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, obh); i_gl 1126 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, dibh); i_gl 1484 fs/gfs2/dir.c struct gfs2_glock *gl = ip->i_gl; i_gl 1750 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, obh); i_gl 1765 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 1949 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, bh); i_gl 2069 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, dibh); i_gl 112 fs/gfs2/export.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 64 fs/gfs2/file.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, i_gl 110 fs/gfs2/file.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); i_gl 164 fs/gfs2/file.c gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 231 fs/gfs2/file.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 265 fs/gfs2/file.c gfs2_log_flush(sdp, ip->i_gl, i_gl 284 fs/gfs2/file.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 467 fs/gfs2/file.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 475 fs/gfs2/file.c set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); i_gl 581 fs/gfs2/file.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, i_gl 651 fs/gfs2/file.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, i_gl 737 fs/gfs2/file.c gfs2_ail_flush(ip->i_gl, 1); i_gl 757 fs/gfs2/file.c gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, &gh); i_gl 788 fs/gfs2/file.c gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, &gh); i_gl 847 fs/gfs2/file.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 923 fs/gfs2/file.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 1118 fs/gfs2/file.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 1744 fs/gfs2/glock.c struct gfs2_glock *gl = ip->i_gl; i_gl 554 fs/gfs2/glops.c struct gfs2_glock *j_gl = ip->i_gl; i_gl 409 fs/gfs2/incore.h struct gfs2_glock *i_gl; /* Move into i_gh? */ i_gl 141 fs/gfs2/inode.c error = gfs2_glock_get(sdp, no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl); i_gl 144 fs/gfs2/inode.c flush_delayed_work(&ip->i_gl->gl_work); i_gl 156 fs/gfs2/inode.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, i_gl 169 fs/gfs2/inode.c glock_set_object(ip->i_gl, ip); i_gl 207 fs/gfs2/inode.c glock_clear_object(ip->i_gl, ip); i_gl 297 fs/gfs2/inode.c if (gfs2_glock_is_locked_by_me(dip->i_gl) == NULL) { i_gl 298 fs/gfs2/inode.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); i_gl 436 fs/gfs2/inode.c bh = gfs2_meta_new(ip->i_gl, ip->i_eattr); i_gl 437 fs/gfs2/inode.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 464 fs/gfs2/inode.c dibh = gfs2_meta_new(ip->i_gl, ip->i_no_addr); i_gl 465 fs/gfs2/inode.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 605 fs/gfs2/inode.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_EXCLUSIVE, 0, ghs); i_gl 703 fs/gfs2/inode.c error = gfs2_glock_get(sdp, ip->i_no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl); i_gl 706 fs/gfs2/inode.c flush_delayed_work(&ip->i_gl->gl_work); i_gl 707 fs/gfs2/inode.c glock_set_object(ip->i_gl, ip); i_gl 709 fs/gfs2/inode.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_SKIP, ghs + 1); i_gl 785 fs/gfs2/inode.c if (ip->i_gl) { i_gl 786 fs/gfs2/inode.c glock_clear_object(ip->i_gl, ip); i_gl 787 fs/gfs2/inode.c gfs2_glock_put(ip->i_gl); i_gl 852 fs/gfs2/inode.c gl = GFS2_I(inode)->i_gl; i_gl 912 fs/gfs2/inode.c gfs2_holder_init(dip->i_gl, LM_ST_EXCLUSIVE, 0, ghs); i_gl 913 fs/gfs2/inode.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, ghs + 1); i_gl 988 fs/gfs2/inode.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 1113 fs/gfs2/inode.c gfs2_holder_init(dip->i_gl, LM_ST_EXCLUSIVE, 0, ghs); i_gl 1114 fs/gfs2/inode.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, ghs + 1); i_gl 1391 fs/gfs2/inode.c gfs2_holder_init(odip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs); i_gl 1393 fs/gfs2/inode.c gfs2_holder_init(ndip->i_gl, LM_ST_EXCLUSIVE,GL_ASYNC, i_gl 1397 fs/gfs2/inode.c gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs + num_gh); i_gl 1401 fs/gfs2/inode.c gfs2_holder_init(nip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, i_gl 1628 fs/gfs2/inode.c gfs2_holder_init(odip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs); i_gl 1630 fs/gfs2/inode.c gfs2_holder_init(ndip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, i_gl 1634 fs/gfs2/inode.c gfs2_holder_init(oip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs + num_gh); i_gl 1637 fs/gfs2/inode.c gfs2_holder_init(nip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs + num_gh); i_gl 1761 fs/gfs2/inode.c gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); i_gl 1815 fs/gfs2/inode.c if (gfs2_glock_is_locked_by_me(ip->i_gl) == NULL) { i_gl 1818 fs/gfs2/inode.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); i_gl 1948 fs/gfs2/inode.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); i_gl 2003 fs/gfs2/inode.c if (gfs2_glock_is_locked_by_me(ip->i_gl) == NULL) { i_gl 2004 fs/gfs2/inode.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); i_gl 2038 fs/gfs2/inode.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 2059 fs/gfs2/inode.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 2078 fs/gfs2/inode.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh); i_gl 754 fs/gfs2/lops.c struct gfs2_glock *gl = ip->i_gl; i_gl 843 fs/gfs2/lops.c gfs2_meta_sync(ip->i_gl); i_gl 849 fs/gfs2/lops.c gfs2_meta_sync(ip->i_gl); i_gl 1007 fs/gfs2/lops.c struct gfs2_glock *gl = ip->i_gl; i_gl 1058 fs/gfs2/lops.c gfs2_meta_sync(ip->i_gl); i_gl 1065 fs/gfs2/lops.c gfs2_meta_sync(ip->i_gl); i_gl 377 fs/gfs2/meta_io.c bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); i_gl 406 fs/gfs2/meta_io.c struct gfs2_glock *gl = ip->i_gl; i_gl 553 fs/gfs2/ops_fstype.c error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh); i_gl 616 fs/gfs2/ops_fstype.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_NOEXP | i_gl 703 fs/gfs2/ops_fstype.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, i_gl 883 fs/gfs2/ops_fstype.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, i_gl 891 fs/gfs2/ops_fstype.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, i_gl 388 fs/gfs2/quota.c error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, 0, &bh); i_gl 653 fs/gfs2/quota.c gfs2_trans_add_meta(ip->i_gl, qd->qd_bh); i_gl 736 fs/gfs2/quota.c gfs2_trans_add_data(ip->i_gl, bh); i_gl 899 fs/gfs2/quota.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); i_gl 958 fs/gfs2/quota.c gfs2_log_flush(ip->i_gl->gl_name.ln_sbd, ip->i_gl, i_gl 1013 fs/gfs2/quota.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); i_gl 1374 fs/gfs2/quota.c bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); i_gl 1684 fs/gfs2/quota.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); i_gl 36 fs/gfs2/recovery.c struct gfs2_glock *gl = ip->i_gl; i_gl 331 fs/gfs2/recovery.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, i_gl 1037 fs/gfs2/rgrp.c struct gfs2_glock *gl = ip->i_gl; i_gl 2424 fs/gfs2/rgrp.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 159 fs/gfs2/super.c struct gfs2_glock *j_gl = ip->i_gl; i_gl 240 fs/gfs2/super.c error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, i_gl 291 fs/gfs2/super.c gfs2_trans_add_meta(l_ip->i_gl, l_bh); i_gl 319 fs/gfs2/super.c gfs2_trans_add_meta(l_ip->i_gl, l_bh); i_gl 320 fs/gfs2/super.c gfs2_trans_add_meta(m_ip->i_gl, m_bh); i_gl 345 fs/gfs2/super.c error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, i_gl 419 fs/gfs2/super.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh); i_gl 505 fs/gfs2/super.c struct address_space *metamapping = gfs2_glock2aspace(ip->i_gl); i_gl 511 fs/gfs2/super.c gfs2_log_flush(GFS2_SB(inode), ip->i_gl, i_gl 558 fs/gfs2/super.c if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { i_gl 559 fs/gfs2/super.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 565 fs/gfs2/super.c } else if (WARN_ON_ONCE(ip->i_gl->gl_state != LM_ST_EXCLUSIVE)) i_gl 579 fs/gfs2/super.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 1143 fs/gfs2/super.c struct gfs2_glock *gl = ip->i_gl; i_gl 1145 fs/gfs2/super.c truncate_inode_pages(gfs2_glock2aspace(ip->i_gl), 0); i_gl 1258 fs/gfs2/super.c BUG_ON(!gfs2_glock_is_locked_by_me(ip->i_gl)); i_gl 1268 fs/gfs2/super.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_SKIP, &gh); i_gl 1327 fs/gfs2/super.c glock_clear_object(ip->i_gl, ip); i_gl 1332 fs/gfs2/super.c gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL | i_gl 1334 fs/gfs2/super.c metamapping = gfs2_glock2aspace(ip->i_gl); i_gl 1335 fs/gfs2/super.c if (test_bit(GLF_DIRTY, &ip->i_gl->gl_flags)) { i_gl 1340 fs/gfs2/super.c gfs2_ail_flush(ip->i_gl, 0); i_gl 1363 fs/gfs2/super.c glock_clear_object(ip->i_gl, ip); i_gl 1374 fs/gfs2/super.c glock_clear_object(ip->i_gl, ip); i_gl 1376 fs/gfs2/super.c gfs2_glock_add_to_lru(ip->i_gl); i_gl 1377 fs/gfs2/super.c gfs2_glock_put_eventually(ip->i_gl); i_gl 1378 fs/gfs2/super.c ip->i_gl = NULL; i_gl 1398 fs/gfs2/super.c ip->i_gl = NULL; i_gl 457 fs/gfs2/trace_gfs2.h __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; i_gl 493 fs/gfs2/trace_gfs2.h __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; i_gl 525 fs/gfs2/trace_gfs2.h __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; i_gl 120 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &bh); i_gl 144 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, bn, DIO_WAIT, 0, &eabh); i_gl 271 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 411 fs/gfs2/xattr.c error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); i_gl 457 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, be64_to_cpu(*dataptrs), 0, 0, i_gl 490 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh[x]); i_gl 599 fs/gfs2/xattr.c if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { i_gl 600 fs/gfs2/xattr.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); i_gl 632 fs/gfs2/xattr.c *bhp = gfs2_meta_new(ip->i_gl, block); i_gl 633 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, *bhp); i_gl 694 fs/gfs2/xattr.c bh = gfs2_meta_new(ip->i_gl, block); i_gl 695 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 833 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, el->el_bh); i_gl 870 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); i_gl 894 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, es->es_bh); i_gl 968 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, i_gl 990 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, indbh); i_gl 998 fs/gfs2/xattr.c indbh = gfs2_meta_new(ip->i_gl, blk); i_gl 999 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, indbh); i_gl 1084 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, el->el_bh); i_gl 1231 fs/gfs2/xattr.c if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { i_gl 1232 fs/gfs2/xattr.c ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); i_gl 1236 fs/gfs2/xattr.c if (WARN_ON_ONCE(ip->i_gl->gl_state != LM_ST_EXCLUSIVE)) i_gl 1266 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &indbh); i_gl 1316 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, indbh); i_gl 1350 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, dibh); i_gl 1400 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, dibh);