Lines Matching refs:iclog

45 	struct xlog_in_core	**iclog,
61 struct xlog_in_core *iclog);
67 STATIC void xlog_state_done_syncing(xlog_in_core_t *iclog, int);
72 struct xlog_in_core *iclog);
77 struct xlog_in_core **iclog,
84 struct xlog_in_core *iclog);
88 struct xlog_in_core *iclog,
93 struct xlog_in_core *iclog);
119 struct xlog_in_core *iclog,
125 struct xlog_in_core *iclog,
515 struct xlog_in_core **iclog, in xfs_log_done() argument
527 (xlog_commit_record(log, ticket, iclog, &lsn)))) { in xfs_log_done()
568 struct xlog_in_core *iclog, in xfs_log_notify() argument
573 spin_lock(&iclog->ic_callback_lock); in xfs_log_notify()
574 abortflg = (iclog->ic_state & XLOG_STATE_IOERROR); in xfs_log_notify()
576 ASSERT_ALWAYS((iclog->ic_state == XLOG_STATE_ACTIVE) || in xfs_log_notify()
577 (iclog->ic_state == XLOG_STATE_WANT_SYNC)); in xfs_log_notify()
579 *(iclog->ic_callback_tail) = cb; in xfs_log_notify()
580 iclog->ic_callback_tail = &(cb->cb_next); in xfs_log_notify()
582 spin_unlock(&iclog->ic_callback_lock); in xfs_log_notify()
589 struct xlog_in_core *iclog) in xfs_log_release_iclog() argument
591 if (xlog_state_release_iclog(mp->m_log, iclog)) { in xfs_log_release_iclog()
782 xlog_in_core_t *iclog; in xfs_log_unmount_write() local
801 first_iclog = iclog = log->l_iclog; in xfs_log_unmount_write()
803 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xfs_log_unmount_write()
804 ASSERT(iclog->ic_state & XLOG_STATE_ACTIVE); in xfs_log_unmount_write()
805 ASSERT(iclog->ic_offset == 0); in xfs_log_unmount_write()
807 iclog = iclog->ic_next; in xfs_log_unmount_write()
808 } while (iclog != first_iclog); in xfs_log_unmount_write()
849 iclog = log->l_iclog; in xfs_log_unmount_write()
850 atomic_inc(&iclog->ic_refcnt); in xfs_log_unmount_write()
851 xlog_state_want_sync(log, iclog); in xfs_log_unmount_write()
853 error = xlog_state_release_iclog(log, iclog); in xfs_log_unmount_write()
856 if (!(iclog->ic_state == XLOG_STATE_ACTIVE || in xfs_log_unmount_write()
857 iclog->ic_state == XLOG_STATE_DIRTY)) { in xfs_log_unmount_write()
859 xlog_wait(&iclog->ic_force_wait, in xfs_log_unmount_write()
887 iclog = log->l_iclog; in xfs_log_unmount_write()
888 atomic_inc(&iclog->ic_refcnt); in xfs_log_unmount_write()
890 xlog_state_want_sync(log, iclog); in xfs_log_unmount_write()
892 error = xlog_state_release_iclog(log, iclog); in xfs_log_unmount_write()
896 if ( ! ( iclog->ic_state == XLOG_STATE_ACTIVE in xfs_log_unmount_write()
897 || iclog->ic_state == XLOG_STATE_DIRTY in xfs_log_unmount_write()
898 || iclog->ic_state == XLOG_STATE_IOERROR) ) { in xfs_log_unmount_write()
900 xlog_wait(&iclog->ic_force_wait, in xfs_log_unmount_write()
1169 struct xlog_in_core *iclog = bp->b_fspriv; in xlog_iodone() local
1170 struct xlog *l = iclog->ic_log; in xlog_iodone()
1187 } else if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_iodone()
1193 xlog_state_done_syncing(iclog, aborted); in xlog_iodone()
1332 xlog_in_core_t *iclog, *prev_iclog=NULL; in xlog_alloc_log() local
1432 iclog = *iclogp; in xlog_alloc_log()
1433 iclog->ic_prev = prev_iclog; in xlog_alloc_log()
1434 prev_iclog = iclog; in xlog_alloc_log()
1447 iclog->ic_bp = bp; in xlog_alloc_log()
1448 iclog->ic_data = bp->b_addr; in xlog_alloc_log()
1450 log->l_iclog_bak[i] = (xfs_caddr_t)&(iclog->ic_header); in xlog_alloc_log()
1452 head = &iclog->ic_header; in xlog_alloc_log()
1462 iclog->ic_size = BBTOB(bp->b_length) - log->l_iclog_hsize; in xlog_alloc_log()
1463 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_alloc_log()
1464 iclog->ic_log = log; in xlog_alloc_log()
1465 atomic_set(&iclog->ic_refcnt, 0); in xlog_alloc_log()
1466 spin_lock_init(&iclog->ic_callback_lock); in xlog_alloc_log()
1467 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_alloc_log()
1468 iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize; in xlog_alloc_log()
1470 init_waitqueue_head(&iclog->ic_force_wait); in xlog_alloc_log()
1471 init_waitqueue_head(&iclog->ic_write_wait); in xlog_alloc_log()
1473 iclogp = &iclog->ic_next; in xlog_alloc_log()
1484 for (iclog = log->l_iclog; iclog; iclog = prev_iclog) { in xlog_alloc_log()
1485 prev_iclog = iclog->ic_next; in xlog_alloc_log()
1486 if (iclog->ic_bp) in xlog_alloc_log()
1487 xfs_buf_free(iclog->ic_bp); in xlog_alloc_log()
1488 kmem_free(iclog); in xlog_alloc_log()
1507 struct xlog_in_core **iclog, in xlog_commit_record() argument
1522 ASSERT_ALWAYS(iclog); in xlog_commit_record()
1523 error = xlog_write(log, &vec, ticket, commitlsnp, iclog, in xlog_commit_record()
1599 struct xlog_in_core *iclog, in xlog_pack_data() argument
1603 int size = iclog->ic_offset + roundoff; in xlog_pack_data()
1607 cycle_lsn = CYCLE_LSN_DISK(iclog->ic_header.h_lsn); in xlog_pack_data()
1609 dp = iclog->ic_datap; in xlog_pack_data()
1613 iclog->ic_header.h_cycle_data[i] = *(__be32 *)dp; in xlog_pack_data()
1619 xlog_in_core_2_t *xhdr = iclog->ic_data; in xlog_pack_data()
1689 struct xlog_in_core *iclog = bp->b_fspriv; in xlog_bdstrat() local
1692 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_bdstrat()
1737 struct xlog_in_core *iclog) in xlog_sync() argument
1750 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_sync()
1753 count_init = log->l_iclog_hsize + iclog->ic_offset; in xlog_sync()
1775 xlog_pack_data(log, iclog, roundoff); in xlog_sync()
1778 size = iclog->ic_offset; in xlog_sync()
1781 iclog->ic_header.h_len = cpu_to_be32(size); in xlog_sync()
1783 bp = iclog->ic_bp; in xlog_sync()
1784 XFS_BUF_SET_ADDR(bp, BLOCK_LSN(be64_to_cpu(iclog->ic_header.h_lsn))); in xlog_sync()
1794 iclog->ic_bwritecnt = 2; in xlog_sync()
1803 dptr = (char *)&iclog->ic_header + count; in xlog_sync()
1813 iclog->ic_bwritecnt = 1; in xlog_sync()
1817 iclog->ic_header.h_crc = xlog_cksum(log, &iclog->ic_header, in xlog_sync()
1818 iclog->ic_datap, size); in xlog_sync()
1821 bp->b_fspriv = iclog; in xlog_sync()
1847 xlog_verify_iclog(log, iclog, count, true); in xlog_sync()
1863 bp = iclog->ic_log->l_xbuf; in xlog_sync()
1866 (char *)&iclog->ic_header + count, split); in xlog_sync()
1867 bp->b_fspriv = iclog; in xlog_sync()
1896 xlog_in_core_t *iclog, *next_iclog; in xlog_dealloc_log() local
1905 iclog = log->l_iclog; in xlog_dealloc_log()
1907 xfs_buf_lock(iclog->ic_bp); in xlog_dealloc_log()
1908 xfs_buf_unlock(iclog->ic_bp); in xlog_dealloc_log()
1909 iclog = iclog->ic_next; in xlog_dealloc_log()
1922 iclog = log->l_iclog; in xlog_dealloc_log()
1924 xfs_buf_free(iclog->ic_bp); in xlog_dealloc_log()
1925 next_iclog = iclog->ic_next; in xlog_dealloc_log()
1926 kmem_free(iclog); in xlog_dealloc_log()
1927 iclog = next_iclog; in xlog_dealloc_log()
1942 struct xlog_in_core *iclog, in xlog_state_finish_copy() argument
1948 be32_add_cpu(&iclog->ic_header.h_num_logops, record_cnt); in xlog_state_finish_copy()
1949 iclog->ic_offset += copy_bytes; in xlog_state_finish_copy()
2216 struct xlog_in_core *iclog, in xlog_write_copy_finish() argument
2230 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2233 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2239 if (iclog->ic_size - log_offset <= sizeof(xlog_op_header_t)) { in xlog_write_copy_finish()
2241 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2246 xlog_state_want_sync(log, iclog); in xlog_write_copy_finish()
2250 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2252 *commit_iclog = iclog; in xlog_write_copy_finish()
2307 struct xlog_in_core *iclog = NULL; in xlog_write() local
2348 error = xlog_state_get_iclog_space(log, len, &iclog, ticket, in xlog_write()
2353 ASSERT(log_offset <= iclog->ic_size - 1); in xlog_write()
2354 ptr = iclog->ic_datap + log_offset; in xlog_write()
2358 *start_lsn = be64_to_cpu(iclog->ic_header.h_lsn); in xlog_write()
2398 iclog->ic_size-log_offset, in xlog_write()
2414 error = xlog_write_copy_finish(log, iclog, flags, in xlog_write()
2455 xlog_state_finish_copy(log, iclog, record_cnt, data_cnt); in xlog_write()
2457 return xlog_state_release_iclog(log, iclog); in xlog_write()
2460 *commit_iclog = iclog; in xlog_write()
2484 xlog_in_core_t *iclog; in xlog_state_clean_log() local
2487 iclog = log->l_iclog; in xlog_state_clean_log()
2489 if (iclog->ic_state == XLOG_STATE_DIRTY) { in xlog_state_clean_log()
2490 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_state_clean_log()
2491 iclog->ic_offset = 0; in xlog_state_clean_log()
2492 ASSERT(iclog->ic_callback == NULL); in xlog_state_clean_log()
2502 (be32_to_cpu(iclog->ic_header.h_num_logops) == in xlog_state_clean_log()
2513 iclog->ic_header.h_num_logops = 0; in xlog_state_clean_log()
2514 memset(iclog->ic_header.h_cycle_data, 0, in xlog_state_clean_log()
2515 sizeof(iclog->ic_header.h_cycle_data)); in xlog_state_clean_log()
2516 iclog->ic_header.h_lsn = 0; in xlog_state_clean_log()
2517 } else if (iclog->ic_state == XLOG_STATE_ACTIVE) in xlog_state_clean_log()
2521 iclog = iclog->ic_next; in xlog_state_clean_log()
2522 } while (iclog != log->l_iclog); in xlog_state_clean_log()
2589 xlog_in_core_t *iclog; in xlog_state_do_callback() local
2603 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2618 iclog = log->l_iclog; in xlog_state_do_callback()
2625 if (iclog->ic_state & in xlog_state_do_callback()
2627 iclog = iclog->ic_next; in xlog_state_do_callback()
2638 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_do_callback()
2649 if (!(iclog->ic_state & in xlog_state_do_callback()
2675 be64_to_cpu(iclog->ic_header.h_lsn)) < 0) { in xlog_state_do_callback()
2676 iclog = iclog->ic_next; in xlog_state_do_callback()
2681 iclog->ic_state = XLOG_STATE_CALLBACK; in xlog_state_do_callback()
2702 be64_to_cpu(iclog->ic_header.h_lsn)) <= 0); in xlog_state_do_callback()
2703 if (iclog->ic_callback) in xlog_state_do_callback()
2705 be64_to_cpu(iclog->ic_header.h_lsn)); in xlog_state_do_callback()
2719 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2720 cb = iclog->ic_callback; in xlog_state_do_callback()
2722 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_state_do_callback()
2723 iclog->ic_callback = NULL; in xlog_state_do_callback()
2724 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2731 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2732 cb = iclog->ic_callback; in xlog_state_do_callback()
2739 ASSERT(iclog->ic_callback == NULL); in xlog_state_do_callback()
2740 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2741 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) in xlog_state_do_callback()
2742 iclog->ic_state = XLOG_STATE_DIRTY; in xlog_state_do_callback()
2751 wake_up_all(&iclog->ic_force_wait); in xlog_state_do_callback()
2753 iclog = iclog->ic_next; in xlog_state_do_callback()
2754 } while (first_iclog != iclog); in xlog_state_do_callback()
2771 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2773 ASSERT(iclog->ic_state != XLOG_STATE_DO_CALLBACK); in xlog_state_do_callback()
2783 if (iclog->ic_state == XLOG_STATE_WANT_SYNC || in xlog_state_do_callback()
2784 iclog->ic_state == XLOG_STATE_SYNCING || in xlog_state_do_callback()
2785 iclog->ic_state == XLOG_STATE_DONE_SYNC || in xlog_state_do_callback()
2786 iclog->ic_state == XLOG_STATE_IOERROR ) in xlog_state_do_callback()
2788 iclog = iclog->ic_next; in xlog_state_do_callback()
2789 } while (first_iclog != iclog); in xlog_state_do_callback()
2817 xlog_in_core_t *iclog, in xlog_state_done_syncing() argument
2820 struct xlog *log = iclog->ic_log; in xlog_state_done_syncing()
2824 ASSERT(iclog->ic_state == XLOG_STATE_SYNCING || in xlog_state_done_syncing()
2825 iclog->ic_state == XLOG_STATE_IOERROR); in xlog_state_done_syncing()
2826 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_state_done_syncing()
2827 ASSERT(iclog->ic_bwritecnt == 1 || iclog->ic_bwritecnt == 2); in xlog_state_done_syncing()
2836 if (iclog->ic_state != XLOG_STATE_IOERROR) { in xlog_state_done_syncing()
2837 if (--iclog->ic_bwritecnt == 1) { in xlog_state_done_syncing()
2841 iclog->ic_state = XLOG_STATE_DONE_SYNC; in xlog_state_done_syncing()
2849 wake_up_all(&iclog->ic_write_wait); in xlog_state_done_syncing()
2851 xlog_state_do_callback(log, aborted, iclog); /* also cleans log */ in xlog_state_done_syncing()
2884 xlog_in_core_t *iclog; in xlog_state_get_iclog_space() local
2894 iclog = log->l_iclog; in xlog_state_get_iclog_space()
2895 if (iclog->ic_state != XLOG_STATE_ACTIVE) { in xlog_state_get_iclog_space()
2903 head = &iclog->ic_header; in xlog_state_get_iclog_space()
2905 atomic_inc(&iclog->ic_refcnt); /* prevents sync */ in xlog_state_get_iclog_space()
2906 log_offset = iclog->ic_offset; in xlog_state_get_iclog_space()
2933 if (iclog->ic_size - iclog->ic_offset < 2*sizeof(xlog_op_header_t)) { in xlog_state_get_iclog_space()
2934 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
2943 if (!atomic_add_unless(&iclog->ic_refcnt, -1, 1)) { in xlog_state_get_iclog_space()
2946 error = xlog_state_release_iclog(log, iclog); in xlog_state_get_iclog_space()
2961 if (len <= iclog->ic_size - iclog->ic_offset) { in xlog_state_get_iclog_space()
2963 iclog->ic_offset += len; in xlog_state_get_iclog_space()
2966 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
2968 *iclogp = iclog; in xlog_state_get_iclog_space()
2970 ASSERT(iclog->ic_offset <= iclog->ic_size); in xlog_state_get_iclog_space()
3074 struct xlog_in_core *iclog) in xlog_state_release_iclog() argument
3078 if (iclog->ic_state & XLOG_STATE_IOERROR) in xlog_state_release_iclog()
3081 ASSERT(atomic_read(&iclog->ic_refcnt) > 0); in xlog_state_release_iclog()
3082 if (!atomic_dec_and_lock(&iclog->ic_refcnt, &log->l_icloglock)) in xlog_state_release_iclog()
3085 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_state_release_iclog()
3089 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE || in xlog_state_release_iclog()
3090 iclog->ic_state == XLOG_STATE_WANT_SYNC); in xlog_state_release_iclog()
3092 if (iclog->ic_state == XLOG_STATE_WANT_SYNC) { in xlog_state_release_iclog()
3096 iclog->ic_state = XLOG_STATE_SYNCING; in xlog_state_release_iclog()
3097 iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); in xlog_state_release_iclog()
3098 xlog_verify_tail_lsn(log, iclog, tail_lsn); in xlog_state_release_iclog()
3111 return xlog_sync(log, iclog); in xlog_state_release_iclog()
3126 struct xlog_in_core *iclog, in xlog_state_switch_iclogs() argument
3129 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); in xlog_state_switch_iclogs()
3131 eventual_size = iclog->ic_offset; in xlog_state_switch_iclogs()
3132 iclog->ic_state = XLOG_STATE_WANT_SYNC; in xlog_state_switch_iclogs()
3133 iclog->ic_header.h_prev_block = cpu_to_be32(log->l_prev_block); in xlog_state_switch_iclogs()
3154 ASSERT(iclog == log->l_iclog); in xlog_state_switch_iclogs()
3155 log->l_iclog = iclog->ic_next; in xlog_state_switch_iclogs()
3192 struct xlog_in_core *iclog; in _xfs_log_force() local
3201 iclog = log->l_iclog; in _xfs_log_force()
3202 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force()
3210 if (iclog->ic_state == XLOG_STATE_ACTIVE || in _xfs_log_force()
3211 iclog->ic_state == XLOG_STATE_DIRTY) { in _xfs_log_force()
3219 if (iclog->ic_state == XLOG_STATE_DIRTY || in _xfs_log_force()
3220 (atomic_read(&iclog->ic_refcnt) == 0 in _xfs_log_force()
3221 && iclog->ic_offset == 0)) { in _xfs_log_force()
3222 iclog = iclog->ic_prev; in _xfs_log_force()
3223 if (iclog->ic_state == XLOG_STATE_ACTIVE || in _xfs_log_force()
3224 iclog->ic_state == XLOG_STATE_DIRTY) in _xfs_log_force()
3229 if (atomic_read(&iclog->ic_refcnt) == 0) { in _xfs_log_force()
3236 atomic_inc(&iclog->ic_refcnt); in _xfs_log_force()
3237 lsn = be64_to_cpu(iclog->ic_header.h_lsn); in _xfs_log_force()
3238 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force()
3241 if (xlog_state_release_iclog(log, iclog)) in _xfs_log_force()
3247 if (be64_to_cpu(iclog->ic_header.h_lsn) == lsn && in _xfs_log_force()
3248 iclog->ic_state != XLOG_STATE_DIRTY) in _xfs_log_force()
3258 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force()
3276 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force()
3281 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in _xfs_log_force()
3287 if (iclog->ic_state & XLOG_STATE_IOERROR) in _xfs_log_force()
3340 struct xlog_in_core *iclog; in _xfs_log_force_lsn() local
3353 iclog = log->l_iclog; in _xfs_log_force_lsn()
3354 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force_lsn()
3360 if (be64_to_cpu(iclog->ic_header.h_lsn) != lsn) { in _xfs_log_force_lsn()
3361 iclog = iclog->ic_next; in _xfs_log_force_lsn()
3365 if (iclog->ic_state == XLOG_STATE_DIRTY) { in _xfs_log_force_lsn()
3370 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in _xfs_log_force_lsn()
3390 (iclog->ic_prev->ic_state & in _xfs_log_force_lsn()
3392 ASSERT(!(iclog->ic_state & XLOG_STATE_IOERROR)); in _xfs_log_force_lsn()
3396 xlog_wait(&iclog->ic_prev->ic_write_wait, in _xfs_log_force_lsn()
3403 atomic_inc(&iclog->ic_refcnt); in _xfs_log_force_lsn()
3404 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force_lsn()
3406 if (xlog_state_release_iclog(log, iclog)) in _xfs_log_force_lsn()
3414 !(iclog->ic_state & in _xfs_log_force_lsn()
3420 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force_lsn()
3425 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in _xfs_log_force_lsn()
3431 if (iclog->ic_state & XLOG_STATE_IOERROR) in _xfs_log_force_lsn()
3441 } while (iclog != log->l_iclog); in _xfs_log_force_lsn()
3473 struct xlog_in_core *iclog) in xlog_state_want_sync() argument
3477 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in xlog_state_want_sync()
3478 xlog_state_switch_iclogs(log, iclog, 0); in xlog_state_want_sync()
3480 ASSERT(iclog->ic_state & in xlog_state_want_sync()
3723 struct xlog_in_core *iclog, in xlog_verify_tail_lsn() argument
3731 if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize)) in xlog_verify_tail_lsn()
3740 if (blocks < BTOBB(iclog->ic_offset) + 1) in xlog_verify_tail_lsn()
3763 struct xlog_in_core *iclog, in xlog_verify_iclog() argument
3788 if (iclog->ic_header.h_magicno != cpu_to_be32(XLOG_HEADER_MAGIC_NUM)) in xlog_verify_iclog()
3791 ptr = (xfs_caddr_t) &iclog->ic_header; in xlog_verify_iclog()
3792 for (ptr += BBSIZE; ptr < ((xfs_caddr_t)&iclog->ic_header) + count; in xlog_verify_iclog()
3800 len = be32_to_cpu(iclog->ic_header.h_num_logops); in xlog_verify_iclog()
3801 ptr = iclog->ic_datap; in xlog_verify_iclog()
3804 xhdr = iclog->ic_data; in xlog_verify_iclog()
3814 idx = BTOBBT((xfs_caddr_t)&(ophead->oh_clientid) - iclog->ic_datap); in xlog_verify_iclog()
3822 iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3838 (__psint_t)iclog->ic_datap); in xlog_verify_iclog()
3844 op_len = be32_to_cpu(iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3859 xlog_in_core_t *iclog, *ic; in xlog_state_ioerror() local
3861 iclog = log->l_iclog; in xlog_state_ioerror()
3862 if (! (iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_ioerror()
3867 ic = iclog; in xlog_state_ioerror()
3871 } while (ic != iclog); in xlog_state_ioerror()
3975 xlog_in_core_t *iclog; in xfs_log_force_umount() local
3978 iclog = log->l_iclog; in xfs_log_force_umount()
3980 ASSERT(iclog->ic_callback == 0); in xfs_log_force_umount()
3981 iclog = iclog->ic_next; in xfs_log_force_umount()
3982 } while (iclog != log->l_iclog); in xfs_log_force_umount()
3994 xlog_in_core_t *iclog; in xlog_iclogs_empty() local
3996 iclog = log->l_iclog; in xlog_iclogs_empty()
4001 if (iclog->ic_header.h_num_logops) in xlog_iclogs_empty()
4003 iclog = iclog->ic_next; in xlog_iclogs_empty()
4004 } while (iclog != log->l_iclog); in xlog_iclogs_empty()