log              10213 arch/m68k/ifpsp060/src/fplsp.S # if enabled so the operating system can log the event.			#
log               993 arch/mips/kvm/mips.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
log              1002 arch/mips/kvm/mips.c 	r = kvm_get_dirty_log_protect(kvm, log, &flush);
log              1006 arch/mips/kvm/mips.c 		memslot = id_to_memslot(slots, log->slot);
log              1016 arch/mips/kvm/mips.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log)
log              1025 arch/mips/kvm/mips.c 	r = kvm_clear_dirty_log_protect(kvm, log, &flush);
log              1029 arch/mips/kvm/mips.c 		memslot = id_to_memslot(slots, log->slot);
log               279 arch/powerpc/include/asm/kvm_ppc.h 	int (*get_dirty_log)(struct kvm *kvm, struct kvm_dirty_log *log);
log               341 arch/powerpc/include/asm/rtas.h struct pseries_errorlog *get_pseries_errorlog(struct rtas_error_log *log,
log              1031 arch/powerpc/kernel/rtas.c struct pseries_errorlog *get_pseries_errorlog(struct rtas_error_log *log,
log              1035 arch/powerpc/kernel/rtas.c 		(struct rtas_ext_event_log_v6 *)log->buffer;
log              1038 arch/powerpc/kernel/rtas.c 	uint32_t ext_log_length = rtas_error_extended_log_length(log);
log              1048 arch/powerpc/kernel/rtas.c 	log_end = log->buffer + ext_log_length;
log               287 arch/powerpc/kernel/rtasd.c static void handle_rtas_event(const struct rtas_error_log *log)
log               289 arch/powerpc/kernel/rtasd.c 	if (rtas_error_type(log) != RTAS_TYPE_PRRN || !prrn_is_enabled())
log               295 arch/powerpc/kernel/rtasd.c 	handle_prrn_event(rtas_error_extended_log_length(log));
log               300 arch/powerpc/kernel/rtasd.c static void handle_rtas_event(const struct rtas_error_log *log)
log               832 arch/powerpc/kvm/book3s.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
log               834 arch/powerpc/kvm/book3s.c 	return kvm->arch.kvm_ops->get_dirty_log(kvm, log);
log              4393 arch/powerpc/kvm/book3s_hv.c 					 struct kvm_dirty_log *log)
log              4405 arch/powerpc/kvm/book3s_hv.c 	if (log->slot >= KVM_USER_MEM_SLOTS)
log              4409 arch/powerpc/kvm/book3s_hv.c 	memslot = id_to_memslot(slots, log->slot);
log              4449 arch/powerpc/kvm/book3s_hv.c 	if (copy_to_user(log->dirty_bitmap, buf, n))
log              1863 arch/powerpc/kvm/book3s_pr.c 					 struct kvm_dirty_log *log)
log              1875 arch/powerpc/kvm/book3s_pr.c 	r = kvm_get_dirty_log(kvm, log, &is_dirty);
log              1882 arch/powerpc/kvm/book3s_pr.c 		memslot = id_to_memslot(slots, log->slot);
log              1799 arch/powerpc/kvm/booke.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
log              2328 arch/powerpc/platforms/cell/spufs/file.c 	ctx->switch_log = kmalloc(struct_size(ctx->switch_log, log,
log              2365 arch/powerpc/platforms/cell/spufs/file.c 	p = ctx->switch_log->log + ctx->switch_log->tail % SWITCH_LOG_BUFSIZE;
log              2490 arch/powerpc/platforms/cell/spufs/file.c 		p = ctx->switch_log->log + ctx->switch_log->head;
log                64 arch/powerpc/platforms/cell/spufs/spufs.h 	} log[];
log               224 arch/powerpc/platforms/pseries/ras.c static void rtas_parse_epow_errlog(struct rtas_error_log *log)
log               231 arch/powerpc/platforms/pseries/ras.c 	pseries_log = get_pseries_errorlog(log, PSERIES_ELOG_SECT_ID_EPOW);
log               613 arch/s390/kvm/kvm-s390.c 			       struct kvm_dirty_log *log)
log               627 arch/s390/kvm/kvm-s390.c 	if (log->slot >= KVM_USER_MEM_SLOTS)
log               631 arch/s390/kvm/kvm-s390.c 	memslot = id_to_memslot(slots, log->slot);
log               637 arch/s390/kvm/kvm-s390.c 	r = kvm_get_dirty_log(kvm, log, &is_dirty);
log                88 arch/um/drivers/vhost_user.h 	u64 desc, used, avail, log;
log               604 arch/um/drivers/virtio_uml.c 				     u64 log)
log               613 arch/um/drivers/virtio_uml.c 		.payload.vring_addr.log = log,
log              4746 arch/x86/kvm/x86.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
log              4759 arch/x86/kvm/x86.c 	r = kvm_get_dirty_log_protect(kvm, log, &flush);
log              4773 arch/x86/kvm/x86.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log)
log              4786 arch/x86/kvm/x86.c 	r = kvm_clear_dirty_log_protect(kvm, log, &flush);
log               105 drivers/acpi/sbs.c static inline int battery_scale(int log)
log               108 drivers/acpi/sbs.c 	while (log--)
log               201 drivers/android/binder.c 	struct binder_transaction_log *log)
log               204 drivers/android/binder.c 	unsigned int cur = atomic_inc_return(&log->cur);
log               206 drivers/android/binder.c 	if (cur >= ARRAY_SIZE(log->entry))
log               207 drivers/android/binder.c 		log->full = true;
log               208 drivers/android/binder.c 	e = &log->entry[cur % ARRAY_SIZE(log->entry)];
log              6047 drivers/android/binder.c 	struct binder_transaction_log *log = m->private;
log              6048 drivers/android/binder.c 	unsigned int log_cur = atomic_read(&log->cur);
log              6054 drivers/android/binder.c 	cur = count < ARRAY_SIZE(log->entry) && !log->full ?
log              6055 drivers/android/binder.c 		0 : count % ARRAY_SIZE(log->entry);
log              6056 drivers/android/binder.c 	if (count > ARRAY_SIZE(log->entry) || log->full)
log              6057 drivers/android/binder.c 		count = ARRAY_SIZE(log->entry);
log              6059 drivers/android/binder.c 		unsigned int index = cur++ % ARRAY_SIZE(log->entry);
log              6061 drivers/android/binder.c 		print_binder_transaction_log_entry(m, &log->entry[index]);
log              2042 drivers/ata/libata-core.c unsigned int ata_read_log_page(struct ata_device *dev, u8 log,
log              2050 drivers/ata/libata-core.c 	DPRINTK("read log page - log 0x%x, page 0x%x\n", log, page);
log              2071 drivers/ata/libata-core.c 	tf.lbal = log;
log              2090 drivers/ata/libata-core.c static bool ata_log_supported(struct ata_device *dev, u8 log)
log              2096 drivers/ata/libata-core.c 	return get_unaligned_le16(&ap->sector_buf[log * 2]) ? true : false;
log                85 drivers/ata/libata.h extern unsigned int ata_read_log_page(struct ata_device *dev, u8 log,
log                48 drivers/base/devres.c module_param_named(log, log_devres, int, S_IRUGO | S_IWUSR);
log               225 drivers/block/paride/pg.c static int pg_identify(struct pg *dev, int log);
log               437 drivers/block/paride/pg.c static int pg_identify(struct pg *dev, int log)
log               452 drivers/block/paride/pg.c 	if (log) {
log               613 drivers/block/rsxx/cregs.c 		log_str = &card->log.buf[card->log.buf_len];
log               614 drivers/block/rsxx/cregs.c 		cnt = min(cmd->cnt8 - off, LOG_BUF_SIZE8 - card->log.buf_len);
log               618 drivers/block/rsxx/cregs.c 		card->log.buf_len += len;
log               625 drivers/block/rsxx/cregs.c 		    (card->log.buf_len == LOG_BUF_SIZE8)) {
log               626 drivers/block/rsxx/cregs.c 			if (card->log.buf_len != 1) /* Don't log blank lines. */
log               627 drivers/block/rsxx/cregs.c 				hw_log_msg(card, card->log.buf,
log               628 drivers/block/rsxx/cregs.c 					   card->log.buf_len);
log               629 drivers/block/rsxx/cregs.c 			card->log.buf_len = 0;
log               643 drivers/block/rsxx/cregs.c 			    sizeof(card->log.tmp), card->log.tmp,
log               144 drivers/block/rsxx/rsxx_priv.h 	} log;
log                51 drivers/char/tpm/eventlog/acpi.c 	struct tpm_bios_log *log;
log                56 drivers/char/tpm/eventlog/acpi.c 	log = &chip->log;
log                88 drivers/char/tpm/eventlog/acpi.c 	log->bios_event_log = kmalloc(len, GFP_KERNEL);
log                89 drivers/char/tpm/eventlog/acpi.c 	if (!log->bios_event_log)
log                92 drivers/char/tpm/eventlog/acpi.c 	log->bios_event_log_end = log->bios_event_log + len;
log                98 drivers/char/tpm/eventlog/acpi.c 	memcpy_fromio(log->bios_event_log, virt, len);
log               104 drivers/char/tpm/eventlog/acpi.c 	kfree(log->bios_event_log);
log               105 drivers/char/tpm/eventlog/acpi.c 	log->bios_event_log = NULL;
log                78 drivers/char/tpm/eventlog/common.c 	if (chip->log.bios_event_log != NULL) {
log                21 drivers/char/tpm/eventlog/efi.c 	struct tpm_bios_log *log;
log                33 drivers/char/tpm/eventlog/efi.c 	log = &chip->log;
log                52 drivers/char/tpm/eventlog/efi.c 	log->bios_event_log = kmemdup(log_tbl->log, log_size, GFP_KERNEL);
log                53 drivers/char/tpm/eventlog/efi.c 	if (!log->bios_event_log) {
log                58 drivers/char/tpm/eventlog/efi.c 	log->bios_event_log_end = log->bios_event_log + log_size;
log                73 drivers/char/tpm/eventlog/efi.c 		kfree(log->bios_event_log);
log                80 drivers/char/tpm/eventlog/efi.c 	tmp = krealloc(log->bios_event_log,
log                84 drivers/char/tpm/eventlog/efi.c 		kfree(log->bios_event_log);
log                89 drivers/char/tpm/eventlog/efi.c 	log->bios_event_log = tmp;
log                96 drivers/char/tpm/eventlog/efi.c 	memcpy((void *)log->bios_event_log + log_size,
log                99 drivers/char/tpm/eventlog/efi.c 	log->bios_event_log_end = log->bios_event_log +
log                25 drivers/char/tpm/eventlog/of.c 	struct tpm_bios_log *log;
log                29 drivers/char/tpm/eventlog/of.c 	log = &chip->log;
log                67 drivers/char/tpm/eventlog/of.c 	log->bios_event_log = kmemdup(__va(base), size, GFP_KERNEL);
log                68 drivers/char/tpm/eventlog/of.c 	if (!log->bios_event_log)
log                71 drivers/char/tpm/eventlog/of.c 	log->bios_event_log_end = log->bios_event_log + size;
log                74 drivers/char/tpm/eventlog/tpm1.c 	struct tpm_bios_log *log = &chip->log;
log                75 drivers/char/tpm/eventlog/tpm1.c 	void *addr = log->bios_event_log;
log                76 drivers/char/tpm/eventlog/tpm1.c 	void *limit = log->bios_event_log_end;
log               113 drivers/char/tpm/eventlog/tpm1.c 	struct tpm_bios_log *log = &chip->log;
log               114 drivers/char/tpm/eventlog/tpm1.c 	void *limit = log->bios_event_log_end;
log                45 drivers/char/tpm/eventlog/tpm2.c 	struct tpm_bios_log *log = &chip->log;
log                46 drivers/char/tpm/eventlog/tpm2.c 	void *addr = log->bios_event_log;
log                47 drivers/char/tpm/eventlog/tpm2.c 	void *limit = log->bios_event_log_end;
log                92 drivers/char/tpm/eventlog/tpm2.c 	struct tpm_bios_log *log = &chip->log;
log                93 drivers/char/tpm/eventlog/tpm2.c 	void *limit = log->bios_event_log_end;
log                98 drivers/char/tpm/eventlog/tpm2.c 	event_header = log->bios_event_log;
log               132 drivers/char/tpm/eventlog/tpm2.c 	struct tpm_bios_log *log = &chip->log;
log               133 drivers/char/tpm/eventlog/tpm2.c 	struct tcg_pcr_event *event_header = log->bios_event_log;
log               270 drivers/char/tpm/tpm-chip.c 	kfree(chip->log.bios_event_log);
log               129 drivers/edac/i3200_edac.c static unsigned long eccerrlog_syndrome(u64 log)
log               131 drivers/edac/i3200_edac.c 	return (log & I3200_ECCERRLOG_SYNDROME_BITS) >>
log               135 drivers/edac/i3200_edac.c static int eccerrlog_row(int channel, u64 log)
log               137 drivers/edac/i3200_edac.c 	u64 rank = ((log & I3200_ECCERRLOG_RANK_BITS) >>
log               223 drivers/edac/i3200_edac.c 	u64 log;
log               235 drivers/edac/i3200_edac.c 		log = info->eccerrlog[channel];
log               236 drivers/edac/i3200_edac.c 		if (log & I3200_ECCERRLOG_UE) {
log               239 drivers/edac/i3200_edac.c 					     eccerrlog_row(channel, log),
log               242 drivers/edac/i3200_edac.c 		} else if (log & I3200_ECCERRLOG_CE) {
log               244 drivers/edac/i3200_edac.c 					     0, 0, eccerrlog_syndrome(log),
log               245 drivers/edac/i3200_edac.c 					     eccerrlog_row(channel, log),
log               147 drivers/edac/ie31200_edac.c #define IE31200_ECCERRLOG_SYNDROME(log)		   \
log               148 drivers/edac/ie31200_edac.c 	((log & IE31200_ECCERRLOG_SYNDROME_BITS) >> \
log               241 drivers/edac/ie31200_edac.c static int eccerrlog_row(u64 log)
log               243 drivers/edac/ie31200_edac.c 	return ((log & IE31200_ECCERRLOG_RANK_BITS) >>
log               300 drivers/edac/ie31200_edac.c 	u64 log;
log               312 drivers/edac/ie31200_edac.c 		log = info->eccerrlog[channel];
log               313 drivers/edac/ie31200_edac.c 		if (log & IE31200_ECCERRLOG_UE) {
log               316 drivers/edac/ie31200_edac.c 					     eccerrlog_row(log),
log               319 drivers/edac/ie31200_edac.c 		} else if (log & IE31200_ECCERRLOG_CE) {
log               322 drivers/edac/ie31200_edac.c 					     IE31200_ECCERRLOG_SYNDROME(log),
log               323 drivers/edac/ie31200_edac.c 					     eccerrlog_row(log),
log               116 drivers/edac/x38_edac.c static unsigned long eccerrlog_syndrome(u64 log)
log               118 drivers/edac/x38_edac.c 	return (log & X38_ECCERRLOG_SYNDROME_BITS) >> 16;
log               121 drivers/edac/x38_edac.c static int eccerrlog_row(int channel, u64 log)
log               123 drivers/edac/x38_edac.c 	return ((log & X38_ECCERRLOG_RANK_BITS) >> 27) |
log               207 drivers/edac/x38_edac.c 	u64 log;
log               220 drivers/edac/x38_edac.c 		log = info->eccerrlog[channel];
log               221 drivers/edac/x38_edac.c 		if (log & X38_ECCERRLOG_UE) {
log               224 drivers/edac/x38_edac.c 					     eccerrlog_row(channel, log),
log               227 drivers/edac/x38_edac.c 		} else if (log & X38_ECCERRLOG_CE) {
log               229 drivers/edac/x38_edac.c 					     0, 0, eccerrlog_syndrome(log),
log               230 drivers/edac/x38_edac.c 					     eccerrlog_row(channel, log),
log               171 drivers/firmware/efi/libstub/tpm.c 	memcpy(log_tbl->log, (void *) first_entry_addr, log_size);
log                89 drivers/firmware/efi/tpm.c 						    log_tbl->log);
log                65 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	intel_guc_log_init_early(&guc->log);
log               113 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	u32 level = intel_guc_log_get_level(&guc->log);
log               154 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	u32 offset = intel_guc_ggtt_offset(guc, guc->log.vma) >> PAGE_SHIFT;
log               262 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	ret = intel_guc_log_create(&guc->log);
log               298 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	intel_guc_log_destroy(&guc->log);
log               324 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	intel_guc_log_destroy(&guc->log);
log               424 drivers/gpu/drm/i915/gt/uc/intel_guc.c 		intel_guc_log_handle_flush_event(&guc->log);
log                28 drivers/gpu/drm/i915/gt/uc/intel_guc.h 	struct intel_guc_log log;
log                13 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_capture_logs(struct intel_guc_log *log);
log                58 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static inline struct intel_guc *log_to_guc(struct intel_guc_log *log)
log                60 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	return container_of(log, struct intel_guc, log);
log                63 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_enable_flush_events(struct intel_guc_log *log)
log                65 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	intel_guc_enable_msg(log_to_guc(log),
log                70 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_disable_flush_events(struct intel_guc_log *log)
log                72 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	intel_guc_disable_msg(log_to_guc(log),
log               148 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_move_to_next_buf(struct intel_guc_log *log)
log               157 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	relay_reserve(log->relay.channel, log->vma->obj->base.size);
log               160 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	relay_flush(log->relay.channel);
log               163 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void *guc_get_write_buffer(struct intel_guc_log *log)
log               174 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	return relay_reserve(log->relay.channel, 0);
log               177 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static bool guc_check_log_buf_overflow(struct intel_guc_log *log,
log               181 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	unsigned int prev_full_cnt = log->stats[type].sampled_overflow;
log               187 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		log->stats[type].overflow = full_cnt;
log               188 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		log->stats[type].sampled_overflow += full_cnt - prev_full_cnt;
log               192 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 			log->stats[type].sampled_overflow += 16;
log               195 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		dev_notice_ratelimited(guc_to_gt(log_to_guc(log))->i915->drm.dev,
log               218 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_read_update_log_buffer(struct intel_guc_log *log)
log               227 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_lock(&log->relay.lock);
log               229 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	if (WARN_ON(!intel_guc_log_relay_enabled(log)))
log               233 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log_buf_state = src_data = log->relay.buf_addr;
log               236 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log_buf_snapshot_state = dst_data = guc_get_write_buffer(log);
log               244 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		log->relay.full_count++;
log               267 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		log->stats[type].flush += log_buf_state_local.flush_to_file;
log               268 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 		new_overflow = guc_check_log_buf_overflow(log, type, full_cnt);
log               315 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_move_to_next_buf(log);
log               318 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_unlock(&log->relay.lock);
log               323 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc_log *log =
log               326 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_capture_logs(log);
log               329 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static int guc_log_map(struct intel_guc_log *log)
log               333 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	lockdep_assert_held(&log->relay.lock);
log               335 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	if (!log->vma)
log               343 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	vaddr = i915_gem_object_pin_map(log->vma->obj, I915_MAP_WC);
log               347 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->relay.buf_addr = vaddr;
log               352 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_unmap(struct intel_guc_log *log)
log               354 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	lockdep_assert_held(&log->relay.lock);
log               356 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	i915_gem_object_unpin_map(log->vma->obj);
log               357 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->relay.buf_addr = NULL;
log               360 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_init_early(struct intel_guc_log *log)
log               362 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_init(&log->relay.lock);
log               363 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	INIT_WORK(&log->relay.flush_work, capture_logs_work);
log               366 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static int guc_log_relay_create(struct intel_guc_log *log)
log               368 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               374 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	lockdep_assert_held(&log->relay.lock);
log               375 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	GEM_BUG_ON(!log->vma);
log               378 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	subbuf_size = log->vma->size;
log               400 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->relay.channel = guc_log_relay_chan;
log               405 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_relay_destroy(struct intel_guc_log *log)
log               407 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	lockdep_assert_held(&log->relay.lock);
log               409 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	relay_close(log->relay.channel);
log               410 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->relay.channel = NULL;
log               413 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_capture_logs(struct intel_guc_log *log)
log               415 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               419 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_read_update_log_buffer(log);
log               429 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static u32 __get_default_log_level(struct intel_guc_log *log)
log               452 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_create(struct intel_guc_log *log)
log               454 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               459 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	GEM_BUG_ON(log->vma);
log               489 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->vma = vma;
log               491 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->level = __get_default_log_level(log);
log               493 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 			 log->level, enableddisabled(log->level),
log               494 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 			 yesno(GUC_LOG_LEVEL_IS_VERBOSE(log->level)),
log               495 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 			 GUC_LOG_LEVEL_TO_VERBOSITY(log->level));
log               504 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_destroy(struct intel_guc_log *log)
log               506 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	i915_vma_unpin_and_release(&log->vma, 0);
log               509 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
log               511 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               517 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	GEM_BUG_ON(!log->vma);
log               528 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	if (log->level == level)
log               541 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	log->level = level;
log               549 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c bool intel_guc_log_relay_enabled(const struct intel_guc_log *log)
log               551 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	return log->relay.buf_addr;
log               554 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_relay_open(struct intel_guc_log *log)
log               558 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	if (!log->vma)
log               561 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_lock(&log->relay.lock);
log               563 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	if (intel_guc_log_relay_enabled(log)) {
log               578 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	ret = guc_log_relay_create(log);
log               582 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	ret = guc_log_map(log);
log               586 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_unlock(&log->relay.lock);
log               588 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_enable_flush_events(log);
log               595 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	queue_work(system_highpri_wq, &log->relay.flush_work);
log               600 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_relay_destroy(log);
log               602 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_unlock(&log->relay.lock);
log               607 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_relay_flush(struct intel_guc_log *log)
log               609 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               617 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	flush_work(&log->relay.flush_work);
log               623 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_capture_logs(log);
log               626 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_relay_close(struct intel_guc_log *log)
log               628 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	struct intel_guc *guc = log_to_guc(log);
log               631 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_disable_flush_events(log);
log               634 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	flush_work(&log->relay.flush_work);
log               636 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_lock(&log->relay.lock);
log               637 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	GEM_BUG_ON(!intel_guc_log_relay_enabled(log));
log               638 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_unmap(log);
log               639 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	guc_log_relay_destroy(log);
log               640 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	mutex_unlock(&log->relay.lock);
log               643 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_handle_flush_event(struct intel_guc_log *log)
log               645 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c 	queue_work(system_highpri_wq, &log->relay.flush_work);
log                63 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_init_early(struct intel_guc_log *log);
log                64 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_create(struct intel_guc_log *log);
log                65 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_destroy(struct intel_guc_log *log);
log                67 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_set_level(struct intel_guc_log *log, u32 level);
log                68 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h bool intel_guc_log_relay_enabled(const struct intel_guc_log *log);
log                69 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_relay_open(struct intel_guc_log *log);
log                70 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_relay_flush(struct intel_guc_log *log);
log                71 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_relay_close(struct intel_guc_log *log);
log                73 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_handle_flush_event(struct intel_guc_log *log);
log                75 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h static inline u32 intel_guc_log_get_level(struct intel_guc_log *log)
log                77 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h 	return log->level;
log               114 drivers/gpu/drm/i915/gt/uc/intel_uc.c 	if (guc->log.vma && !uc->load_err_log)
log               115 drivers/gpu/drm/i915/gt/uc/intel_uc.c 		uc->load_err_log = i915_gem_object_get(guc->log.vma->obj);
log               120 drivers/gpu/drm/i915/gt/uc/intel_uc.c 	struct drm_i915_gem_object *log = fetch_and_zero(&uc->load_err_log);
log               122 drivers/gpu/drm/i915/gt/uc/intel_uc.c 	if (log)
log               123 drivers/gpu/drm/i915/gt/uc/intel_uc.c 		i915_gem_object_put(log);
log              1860 drivers/gpu/drm/i915/i915_debugfs.c 	struct intel_guc_log *log = &dev_priv->gt.uc.guc.log;
log              1863 drivers/gpu/drm/i915/i915_debugfs.c 	if (!intel_guc_log_relay_enabled(log)) {
log              1871 drivers/gpu/drm/i915/i915_debugfs.c 		   log->relay.full_count);
log              1876 drivers/gpu/drm/i915/i915_debugfs.c 			   log->stats[type].flush,
log              1877 drivers/gpu/drm/i915/i915_debugfs.c 			   log->stats[type].sampled_overflow);
log              1971 drivers/gpu/drm/i915/i915_debugfs.c 	u32 *log;
log              1979 drivers/gpu/drm/i915/i915_debugfs.c 	else if (dev_priv->gt.uc.guc.log.vma)
log              1980 drivers/gpu/drm/i915/i915_debugfs.c 		obj = dev_priv->gt.uc.guc.log.vma->obj;
log              1985 drivers/gpu/drm/i915/i915_debugfs.c 	log = i915_gem_object_pin_map(obj, I915_MAP_WC);
log              1986 drivers/gpu/drm/i915/i915_debugfs.c 	if (IS_ERR(log)) {
log              1989 drivers/gpu/drm/i915/i915_debugfs.c 		return PTR_ERR(log);
log              1994 drivers/gpu/drm/i915/i915_debugfs.c 			   *(log + i), *(log + i + 1),
log              1995 drivers/gpu/drm/i915/i915_debugfs.c 			   *(log + i + 2), *(log + i + 3));
log              2011 drivers/gpu/drm/i915/i915_debugfs.c 	*val = intel_guc_log_get_level(&dev_priv->gt.uc.guc.log);
log              2023 drivers/gpu/drm/i915/i915_debugfs.c 	return intel_guc_log_set_level(&dev_priv->gt.uc.guc.log, val);
log              2034 drivers/gpu/drm/i915/i915_debugfs.c 	struct intel_guc_log *log = &guc->log;
log              2039 drivers/gpu/drm/i915/i915_debugfs.c 	file->private_data = log;
log              2041 drivers/gpu/drm/i915/i915_debugfs.c 	return intel_guc_log_relay_open(log);
log              2050 drivers/gpu/drm/i915/i915_debugfs.c 	struct intel_guc_log *log = filp->private_data;
log              2052 drivers/gpu/drm/i915/i915_debugfs.c 	intel_guc_log_relay_flush(log);
log              2061 drivers/gpu/drm/i915/i915_debugfs.c 	intel_guc_log_relay_close(&guc->log);
log              1488 drivers/gpu/drm/i915/i915_gpu_error.c 						     uc->guc.log.vma,
log                36 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/gpio.h 	u8 log[2];
log               132 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c 	u32 log = ctr->logic_op;
log               142 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c 	nvkm_wr32(device, dom->addr + 0x044 + (ctr->slot * 0x08), log);
log                31 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c 	u32 log = ctr->logic_op;
log                40 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c 	nvkm_wr32(device, 0x00a420 + dom->addr + (ctr->slot * 0x40), log);
log                88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[0] = (info & 0x1800) >> 11,
log                89 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[1] = (info & 0x6000) >> 13,
log                98 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[0] = (info & 0x18000000) >> 27,
log                99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[1] = (info & 0x60000000) >> 29,
log               108 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[0] = (info1 & 0x30) >> 4,
log               109 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 				.log[1] = (info1 & 0xc0) >> 6,
log               140 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 					.log[0] = !!(conf & 0x02),
log               141 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c 					.log[1] =  !(conf & 0x02),
log              1598 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->fuc.r_funcMV[0] = (gpio.log[0] ^ 2) << 12;
log              1599 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->fuc.r_funcMV[1] = (gpio.log[1] ^ 2) << 12;
log              1605 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->fuc.r_func2E[0] = (gpio.log[0] ^ 2) << 12;
log              1606 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c 		ram->fuc.r_func2E[1] = (gpio.log[1] ^ 2) << 12;
log               483 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c 		if (!(func.log[1] & 1))
log               210 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c 		if (!(func.log[1] & 1))
log                70 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c 				.log[0] = 0,
log                71 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c 				.log[1] = 1,
log                88 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c 		int dir = !!(func.log[state] & 0x02);
log                89 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c 		int out = !!(func.log[state] & 0x01);
log               106 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c 			ret = (ret == (func.log[1] & 1));
log               236 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 		if (func.line != 16 && func.log[0] & DCB_GPIO_LOG_DIR_IN) {
log                50 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fanpwm.c 		if (card_type <= NV_40 || (fan->func.log[0] & 1))
log                75 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fanpwm.c 	if (card_type <= NV_40 || (fan->func.log[0] & 1))
log              1390 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h int vmw_host_log(const char *log);
log               479 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c int vmw_host_log(const char *log)
log               489 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c 	if (!log)
log               492 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c 	msg = kasprintf(GFP_KERNEL, "log %s", log);
log               100 drivers/ide/ide-cd.c 	int log = 0;
log               118 drivers/ide/ide-cd.c 		log = 1;
log               127 drivers/ide/ide-cd.c 		log = 1;
log               138 drivers/ide/ide-cd.c 		log = 1;
log               141 drivers/ide/ide-cd.c 	return log;
log               131 drivers/isdn/hardware/mISDN/avmfritz.c 	char			log[LOG_SIZE + 1];
log               429 drivers/isdn/hardware/mISDN/avmfritz.c 		snprintf(fc->log, LOG_SIZE, "B%1d-recv %s %d ",
log               431 drivers/isdn/hardware/mISDN/avmfritz.c 		print_hex_dump_bytes(fc->log, DUMP_PREFIX_OFFSET, p, count);
log               500 drivers/isdn/hardware/mISDN/avmfritz.c 		snprintf(fc->log, LOG_SIZE, "B%1d-send %s %d ",
log               502 drivers/isdn/hardware/mISDN/avmfritz.c 		print_hex_dump_bytes(fc->log, DUMP_PREFIX_OFFSET, p, count);
log                52 drivers/isdn/hardware/mISDN/ipac.h 	char			log[64];
log                53 drivers/isdn/hardware/mISDN/isar.h 	u8		log[256];
log               955 drivers/isdn/hardware/mISDN/mISDNipac.c 		snprintf(hscx->log, 64, "B%1d-recv %s %d ",
log               957 drivers/isdn/hardware/mISDN/mISDNipac.c 		print_hex_dump_bytes(hscx->log, DUMP_PREFIX_OFFSET, p, count);
log               972 drivers/isdn/hardware/mISDN/mISDNipac.c 		p = hscx->log;
log              1000 drivers/isdn/hardware/mISDN/mISDNipac.c 		snprintf(hscx->log, 64, "B%1d-send %s %d ",
log              1002 drivers/isdn/hardware/mISDN/mISDNipac.c 		print_hex_dump_bytes(hscx->log, DUMP_PREFIX_OFFSET, p, count);
log                74 drivers/isdn/hardware/mISDN/mISDNisar.c 						   isar->log, 256, 1);
log                76 drivers/isdn/hardware/mISDN/mISDNisar.c 					 __func__, l, isar->log);
log               103 drivers/isdn/hardware/mISDN/mISDNisar.c 						   1, isar->log, 256, 1);
log               105 drivers/isdn/hardware/mISDN/mISDNisar.c 					 __func__, l, isar->log);
log                84 drivers/isdn/hardware/mISDN/netjet.c 	char			log[LOG_SIZE];
log               413 drivers/isdn/hardware/mISDN/netjet.c 				snprintf(card->log, LOG_SIZE,
log               416 drivers/isdn/hardware/mISDN/netjet.c 				print_hex_dump_bytes(card->log,
log               517 drivers/isdn/hardware/mISDN/netjet.c 		snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ",
log               519 drivers/isdn/hardware/mISDN/netjet.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count);
log               591 drivers/isdn/hardware/mISDN/netjet.c 		snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ",
log               593 drivers/isdn/hardware/mISDN/netjet.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count);
log                68 drivers/isdn/hardware/mISDN/w6692.c 	char			log[64];
log               266 drivers/isdn/hardware/mISDN/w6692.c 		snprintf(card->log, 63, "D-recv %s %d ",
log               268 drivers/isdn/hardware/mISDN/w6692.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count);
log               301 drivers/isdn/hardware/mISDN/w6692.c 		snprintf(card->log, 63, "D-send %s %d ",
log               303 drivers/isdn/hardware/mISDN/w6692.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count);
log               477 drivers/isdn/hardware/mISDN/w6692.c 		snprintf(card->log, 63, "B%1d-recv %s %d ",
log               479 drivers/isdn/hardware/mISDN/w6692.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count);
log               521 drivers/isdn/hardware/mISDN/w6692.c 		snprintf(card->log, 63, "B%1d-send %s %d ",
log               523 drivers/isdn/hardware/mISDN/w6692.c 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count);
log               335 drivers/md/dm-crypt.c 	int log;
log               341 drivers/md/dm-crypt.c 	log = ilog2(bs);
log               346 drivers/md/dm-crypt.c 	if (1 << log != bs) {
log               351 drivers/md/dm-crypt.c 	if (log > 9) {
log               356 drivers/md/dm-crypt.c 	cc->iv_gen_private.benbi.shift = 9 - log;
log               190 drivers/md/dm-log-userspace-base.c static int userspace_ctr(struct dm_dirty_log *log, struct dm_target *ti,
log               320 drivers/md/dm-log-userspace-base.c 		log->context = lc;
log               326 drivers/md/dm-log-userspace-base.c static void userspace_dtr(struct dm_dirty_log *log)
log               328 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               352 drivers/md/dm-log-userspace-base.c static int userspace_presuspend(struct dm_dirty_log *log)
log               355 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               363 drivers/md/dm-log-userspace-base.c static int userspace_postsuspend(struct dm_dirty_log *log)
log               366 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               380 drivers/md/dm-log-userspace-base.c static int userspace_resume(struct dm_dirty_log *log)
log               383 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               392 drivers/md/dm-log-userspace-base.c static uint32_t userspace_get_region_size(struct dm_dirty_log *log)
log               394 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               407 drivers/md/dm-log-userspace-base.c static int userspace_is_clean(struct dm_dirty_log *log, region_t region)
log               413 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               434 drivers/md/dm-log-userspace-base.c static int userspace_in_sync(struct dm_dirty_log *log, region_t region,
log               441 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               562 drivers/md/dm-log-userspace-base.c static int userspace_flush(struct dm_dirty_log *log)
log               566 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               648 drivers/md/dm-log-userspace-base.c static void userspace_mark_region(struct dm_dirty_log *log, region_t region)
log               651 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               677 drivers/md/dm-log-userspace-base.c static void userspace_clear_region(struct dm_dirty_log *log, region_t region)
log               680 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               712 drivers/md/dm-log-userspace-base.c static int userspace_get_resync_work(struct dm_dirty_log *log, region_t *region)
log               716 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               739 drivers/md/dm-log-userspace-base.c static void userspace_set_region_sync(struct dm_dirty_log *log,
log               742 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               769 drivers/md/dm-log-userspace-base.c static region_t userspace_get_sync_count(struct dm_dirty_log *log)
log               774 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               794 drivers/md/dm-log-userspace-base.c static int userspace_status(struct dm_dirty_log *log, status_type_t status_type,
log               800 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               809 drivers/md/dm-log-userspace-base.c 			DMEMIT("%s 1 COM_FAILURE", log->type->name);
log               818 drivers/md/dm-log-userspace-base.c 		DMEMIT("%s %u %s ", log->type->name, lc->usr_argc, lc->uuid);
log               832 drivers/md/dm-log-userspace-base.c static int userspace_is_remote_recovering(struct dm_dirty_log *log,
log               837 drivers/md/dm-log-userspace-base.c 	struct log_c *lc = log->context;
log               153 drivers/md/dm-log.c 	struct dm_dirty_log *log;
log               155 drivers/md/dm-log.c 	log = kmalloc(sizeof(*log), GFP_KERNEL);
log               156 drivers/md/dm-log.c 	if (!log)
log               161 drivers/md/dm-log.c 		kfree(log);
log               165 drivers/md/dm-log.c 	log->flush_callback_fn = flush_callback_fn;
log               166 drivers/md/dm-log.c 	log->type = type;
log               167 drivers/md/dm-log.c 	if (type->ctr(log, ti, argc, argv)) {
log               168 drivers/md/dm-log.c 		kfree(log);
log               173 drivers/md/dm-log.c 	return log;
log               177 drivers/md/dm-log.c void dm_dirty_log_destroy(struct dm_dirty_log *log)
log               179 drivers/md/dm-log.c 	log->type->dtr(log);
log               180 drivers/md/dm-log.c 	put_type(log->type);
log               181 drivers/md/dm-log.c 	kfree(log);
log               316 drivers/md/dm-log.c static int read_header(struct log_c *log)
log               320 drivers/md/dm-log.c 	r = rw_header(log, REQ_OP_READ);
log               324 drivers/md/dm-log.c 	header_from_disk(&log->header, log->disk_header);
log               327 drivers/md/dm-log.c 	if (log->sync != DEFAULTSYNC || log->header.magic != MIRROR_MAGIC) {
log               328 drivers/md/dm-log.c 		log->header.magic = MIRROR_MAGIC;
log               329 drivers/md/dm-log.c 		log->header.version = MIRROR_DISK_VERSION;
log               330 drivers/md/dm-log.c 		log->header.nr_regions = 0;
log               334 drivers/md/dm-log.c 	if (log->header.version == 1)
log               335 drivers/md/dm-log.c 		log->header.version = 2;
log               338 drivers/md/dm-log.c 	if (log->header.version != MIRROR_DISK_VERSION) {
log               363 drivers/md/dm-log.c static int create_log_context(struct dm_dirty_log *log, struct dm_target *ti,
log               511 drivers/md/dm-log.c 	log->context = lc;
log               516 drivers/md/dm-log.c static int core_ctr(struct dm_dirty_log *log, struct dm_target *ti,
log               519 drivers/md/dm-log.c 	return create_log_context(log, ti, argc, argv, NULL);
log               529 drivers/md/dm-log.c static void core_dtr(struct dm_dirty_log *log)
log               531 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               542 drivers/md/dm-log.c static int disk_ctr(struct dm_dirty_log *log, struct dm_target *ti,
log               557 drivers/md/dm-log.c 	r = create_log_context(log, ti, argc - 1, argv + 1, dev);
log               566 drivers/md/dm-log.c static void disk_dtr(struct dm_dirty_log *log)
log               568 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               585 drivers/md/dm-log.c static int disk_resume(struct dm_dirty_log *log)
log               589 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               649 drivers/md/dm-log.c static uint32_t core_get_region_size(struct dm_dirty_log *log)
log               651 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               655 drivers/md/dm-log.c static int core_resume(struct dm_dirty_log *log)
log               657 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               662 drivers/md/dm-log.c static int core_is_clean(struct dm_dirty_log *log, region_t region)
log               664 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               668 drivers/md/dm-log.c static int core_in_sync(struct dm_dirty_log *log, region_t region, int block)
log               670 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               674 drivers/md/dm-log.c static int core_flush(struct dm_dirty_log *log)
log               680 drivers/md/dm-log.c static int disk_flush(struct dm_dirty_log *log)
log               683 drivers/md/dm-log.c 	struct log_c *lc = log->context;
log               689 drivers/md/dm-log.c 	if (lc->touched_cleaned && log->flush_callback_fn &&
log               690 drivers/md/dm-log.c 	    log->flush_callback_fn(lc->ti)) {
log               720 drivers/md/dm-log.c static void core_mark_region(struct dm_dirty_log *log, region_t region)
log               722 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               726 drivers/md/dm-log.c static void core_clear_region(struct dm_dirty_log *log, region_t region)
log               728 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               733 drivers/md/dm-log.c static int core_get_resync_work(struct dm_dirty_log *log, region_t *region)
log               735 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               755 drivers/md/dm-log.c static void core_set_region_sync(struct dm_dirty_log *log, region_t region,
log               758 drivers/md/dm-log.c 	struct log_c *lc = (struct log_c *) log->context;
log               770 drivers/md/dm-log.c static region_t core_get_sync_count(struct dm_dirty_log *log)
log               772 drivers/md/dm-log.c         struct log_c *lc = (struct log_c *) log->context;
log               781 drivers/md/dm-log.c static int core_status(struct dm_dirty_log *log, status_type_t status,
log               785 drivers/md/dm-log.c 	struct log_c *lc = log->context;
log               789 drivers/md/dm-log.c 		DMEMIT("1 %s", log->type->name);
log               793 drivers/md/dm-log.c 		DMEMIT("%s %u %u ", log->type->name,
log               801 drivers/md/dm-log.c static int disk_status(struct dm_dirty_log *log, status_type_t status,
log               805 drivers/md/dm-log.c 	struct log_c *lc = log->context;
log               809 drivers/md/dm-log.c 		DMEMIT("3 %s %s %c", log->type->name, lc->log_dev->name,
log               816 drivers/md/dm-log.c 		DMEMIT("%s %u %s %u ", log->type->name,
log               387 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log               404 drivers/md/dm-raid1.c 	    (log->type->get_sync_count(log) == ms->nr_regions)) {
log               439 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log               442 drivers/md/dm-raid1.c 	if (log->type->in_sync(log, region, 0))
log               685 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log               708 drivers/md/dm-raid1.c 		if (log->type->is_remote_recovering &&
log               709 drivers/md/dm-raid1.c 		    log->type->is_remote_recovering(log, region)) {
log              1189 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log              1202 drivers/md/dm-raid1.c 	r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0);
log              1303 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log              1333 drivers/md/dm-raid1.c 	if (log->type->presuspend && log->type->presuspend(log))
log              1349 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log              1351 drivers/md/dm-raid1.c 	if (log->type->postsuspend && log->type->postsuspend(log))
log              1359 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log              1362 drivers/md/dm-raid1.c 	if (log->type->resume && log->type->resume(log))
log              1399 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
log              1412 drivers/md/dm-raid1.c 		      (unsigned long long)log->type->get_sync_count(log),
log              1415 drivers/md/dm-raid1.c 		sz += log->type->status(log, type, result+sz, maxlen-sz);
log              1420 drivers/md/dm-raid1.c 		sz = log->type->status(log, type, result, maxlen);
log                62 drivers/md/dm-region-hash.c 	struct dm_dirty_log *log;
log               167 drivers/md/dm-region-hash.c 		struct dm_dirty_log *log, uint32_t region_size,
log               196 drivers/md/dm-region-hash.c 	rh->log = log;
log               251 drivers/md/dm-region-hash.c 	if (rh->log)
log               252 drivers/md/dm-region-hash.c 		dm_dirty_log_destroy(rh->log);
log               262 drivers/md/dm-region-hash.c 	return rh->log;
log               296 drivers/md/dm-region-hash.c 	nreg->state = rh->log->type->in_sync(rh->log, region, 1) ?
log               354 drivers/md/dm-region-hash.c 	r = rh->log->type->in_sync(rh->log, region, may_block);
log               368 drivers/md/dm-region-hash.c 	rh->log->type->set_region_sync(rh->log, reg->key, success);
log               398 drivers/md/dm-region-hash.c 	struct dm_dirty_log *log = rh->log;
log               412 drivers/md/dm-region-hash.c 	log->type->set_region_sync(log, region, 0);
log               484 drivers/md/dm-region-hash.c 		rh->log->type->clear_region(rh->log, reg->key);
log               495 drivers/md/dm-region-hash.c 		rh->log->type->clear_region(rh->log, reg->key);
log               499 drivers/md/dm-region-hash.c 	rh->log->type->flush(rh->log);
log               518 drivers/md/dm-region-hash.c 		rh->log->type->mark_region(rh->log, reg->key);
log               596 drivers/md/dm-region-hash.c 	r = rh->log->type->get_resync_work(rh->log, &region);
log               686 drivers/md/dm-region-hash.c 	return rh->log->type->flush(rh->log);
log               212 drivers/md/raid5-cache.c 	struct r5l_log *log;
log               252 drivers/md/raid5-cache.c bool r5c_is_writeback(struct r5l_log *log)
log               254 drivers/md/raid5-cache.c 	return (log != NULL &&
log               255 drivers/md/raid5-cache.c 		log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK);
log               258 drivers/md/raid5-cache.c static sector_t r5l_ring_add(struct r5l_log *log, sector_t start, sector_t inc)
log               261 drivers/md/raid5-cache.c 	if (start >= log->device_size)
log               262 drivers/md/raid5-cache.c 		start = start - log->device_size;
log               266 drivers/md/raid5-cache.c static sector_t r5l_ring_distance(struct r5l_log *log, sector_t start,
log               272 drivers/md/raid5-cache.c 		return end + log->device_size - start;
log               275 drivers/md/raid5-cache.c static bool r5l_has_free_space(struct r5l_log *log, sector_t size)
log               279 drivers/md/raid5-cache.c 	used_size = r5l_ring_distance(log, log->last_checkpoint,
log               280 drivers/md/raid5-cache.c 					log->log_start);
log               282 drivers/md/raid5-cache.c 	return log->device_size > used_size + size;
log               326 drivers/md/raid5-cache.c void r5l_wake_reclaim(struct r5l_log *log, sector_t space);
log               333 drivers/md/raid5-cache.c 	if (!r5c_is_writeback(conf->log))
log               349 drivers/md/raid5-cache.c 		r5l_wake_reclaim(conf->log, 0);
log               358 drivers/md/raid5-cache.c 	if (!r5c_is_writeback(conf->log))
log               368 drivers/md/raid5-cache.c 		r5l_wake_reclaim(conf->log, 0);
log               401 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log               403 drivers/md/raid5-cache.c 	if (!r5c_is_writeback(log))
log               407 drivers/md/raid5-cache.c 		((conf->max_degraded + 1) * atomic_read(&log->stripe_in_journal_count) +
log               418 drivers/md/raid5-cache.c static inline void r5c_update_log_state(struct r5l_log *log)
log               420 drivers/md/raid5-cache.c 	struct r5conf *conf = log->rdev->mddev->private;
log               425 drivers/md/raid5-cache.c 	if (!r5c_is_writeback(log))
log               428 drivers/md/raid5-cache.c 	free_space = r5l_ring_distance(log, log->log_start,
log               429 drivers/md/raid5-cache.c 				       log->last_checkpoint);
log               444 drivers/md/raid5-cache.c 		r5l_wake_reclaim(log, 0);
log               454 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log               456 drivers/md/raid5-cache.c 	BUG_ON(!r5c_is_writeback(log));
log               496 drivers/md/raid5-cache.c 	struct r5l_log *log = sh->raid_conf->log;
log               498 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) {
log               529 drivers/md/raid5-cache.c static void r5l_log_run_stripes(struct r5l_log *log)
log               533 drivers/md/raid5-cache.c 	lockdep_assert_held(&log->io_list_lock);
log               535 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
log               540 drivers/md/raid5-cache.c 		list_move_tail(&io->log_sibling, &log->finished_ios);
log               545 drivers/md/raid5-cache.c static void r5l_move_to_end_ios(struct r5l_log *log)
log               549 drivers/md/raid5-cache.c 	lockdep_assert_held(&log->io_list_lock);
log               551 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
log               555 drivers/md/raid5-cache.c 		list_move_tail(&io->log_sibling, &log->io_end_ios);
log               564 drivers/md/raid5-cache.c 	struct r5l_log *log = io->log;
log               570 drivers/md/raid5-cache.c 		md_error(log->rdev->mddev, log->rdev);
log               573 drivers/md/raid5-cache.c 	mempool_free(io->meta_page, &log->meta_pool);
log               575 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log               587 drivers/md/raid5-cache.c 	if (log->need_cache_flush && !list_empty(&io->stripe_list))
log               588 drivers/md/raid5-cache.c 		r5l_move_to_end_ios(log);
log               590 drivers/md/raid5-cache.c 		r5l_log_run_stripes(log);
log               591 drivers/md/raid5-cache.c 	if (!list_empty(&log->running_ios)) {
log               596 drivers/md/raid5-cache.c 		io_deferred = list_first_entry(&log->running_ios,
log               599 drivers/md/raid5-cache.c 			schedule_work(&log->deferred_io_work);
log               602 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log               604 drivers/md/raid5-cache.c 	if (log->need_cache_flush)
log               605 drivers/md/raid5-cache.c 		md_wakeup_thread(log->rdev->mddev->thread);
log               626 drivers/md/raid5-cache.c static void r5l_do_submit_io(struct r5l_log *log, struct r5l_io_unit *io)
log               630 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log               632 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log               663 drivers/md/raid5-cache.c 	struct r5l_log *log = container_of(work, struct r5l_log,
log               668 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log               669 drivers/md/raid5-cache.c 	if (!list_empty(&log->running_ios)) {
log               670 drivers/md/raid5-cache.c 		io = list_first_entry(&log->running_ios, struct r5l_io_unit,
log               677 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log               679 drivers/md/raid5-cache.c 		r5l_do_submit_io(log, io);
log               684 drivers/md/raid5-cache.c 	struct r5l_log *log = container_of(work, struct r5l_log,
log               686 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log               690 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH)
log               697 drivers/md/raid5-cache.c 		   conf->log == NULL ||
log               702 drivers/md/raid5-cache.c 		log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH;
log               708 drivers/md/raid5-cache.c static void r5l_submit_current_io(struct r5l_log *log)
log               710 drivers/md/raid5-cache.c 	struct r5l_io_unit *io = log->current_io;
log               721 drivers/md/raid5-cache.c 	crc = crc32c_le(log->uuid_checksum, block, PAGE_SIZE);
log               724 drivers/md/raid5-cache.c 	log->current_io = NULL;
log               725 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log               727 drivers/md/raid5-cache.c 		if (io != list_first_entry(&log->running_ios,
log               733 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log               735 drivers/md/raid5-cache.c 		r5l_do_submit_io(log, io);
log               738 drivers/md/raid5-cache.c static struct bio *r5l_bio_alloc(struct r5l_log *log)
log               740 drivers/md/raid5-cache.c 	struct bio *bio = bio_alloc_bioset(GFP_NOIO, BIO_MAX_PAGES, &log->bs);
log               743 drivers/md/raid5-cache.c 	bio_set_dev(bio, log->rdev->bdev);
log               744 drivers/md/raid5-cache.c 	bio->bi_iter.bi_sector = log->rdev->data_offset + log->log_start;
log               749 drivers/md/raid5-cache.c static void r5_reserve_log_entry(struct r5l_log *log, struct r5l_io_unit *io)
log               751 drivers/md/raid5-cache.c 	log->log_start = r5l_ring_add(log, log->log_start, BLOCK_SECTORS);
log               753 drivers/md/raid5-cache.c 	r5c_update_log_state(log);
log               761 drivers/md/raid5-cache.c 	if (log->log_start == 0)
log               764 drivers/md/raid5-cache.c 	io->log_end = log->log_start;
log               767 drivers/md/raid5-cache.c static struct r5l_io_unit *r5l_new_meta(struct r5l_log *log)
log               772 drivers/md/raid5-cache.c 	io = mempool_alloc(&log->io_pool, GFP_ATOMIC);
log               777 drivers/md/raid5-cache.c 	io->log = log;
log               783 drivers/md/raid5-cache.c 	io->meta_page = mempool_alloc(&log->meta_pool, GFP_NOIO);
log               788 drivers/md/raid5-cache.c 	block->seq = cpu_to_le64(log->seq);
log               789 drivers/md/raid5-cache.c 	block->position = cpu_to_le64(log->log_start);
log               791 drivers/md/raid5-cache.c 	io->log_start = log->log_start;
log               793 drivers/md/raid5-cache.c 	io->seq = log->seq++;
log               795 drivers/md/raid5-cache.c 	io->current_bio = r5l_bio_alloc(log);
log               800 drivers/md/raid5-cache.c 	r5_reserve_log_entry(log, io);
log               802 drivers/md/raid5-cache.c 	spin_lock_irq(&log->io_list_lock);
log               803 drivers/md/raid5-cache.c 	list_add_tail(&io->log_sibling, &log->running_ios);
log               804 drivers/md/raid5-cache.c 	spin_unlock_irq(&log->io_list_lock);
log               809 drivers/md/raid5-cache.c static int r5l_get_meta(struct r5l_log *log, unsigned int payload_size)
log               811 drivers/md/raid5-cache.c 	if (log->current_io &&
log               812 drivers/md/raid5-cache.c 	    log->current_io->meta_offset + payload_size > PAGE_SIZE)
log               813 drivers/md/raid5-cache.c 		r5l_submit_current_io(log);
log               815 drivers/md/raid5-cache.c 	if (!log->current_io) {
log               816 drivers/md/raid5-cache.c 		log->current_io = r5l_new_meta(log);
log               817 drivers/md/raid5-cache.c 		if (!log->current_io)
log               824 drivers/md/raid5-cache.c static void r5l_append_payload_meta(struct r5l_log *log, u16 type,
log               829 drivers/md/raid5-cache.c 	struct r5l_io_unit *io = log->current_io;
log               846 drivers/md/raid5-cache.c static void r5l_append_payload_page(struct r5l_log *log, struct page *page)
log               848 drivers/md/raid5-cache.c 	struct r5l_io_unit *io = log->current_io;
log               853 drivers/md/raid5-cache.c 		io->current_bio = r5l_bio_alloc(log);
log               861 drivers/md/raid5-cache.c 	r5_reserve_log_entry(log, io);
log               864 drivers/md/raid5-cache.c static void r5l_append_flush_payload(struct r5l_log *log, sector_t sect)
log               866 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log               880 drivers/md/raid5-cache.c 	mutex_lock(&log->io_mutex);
log               883 drivers/md/raid5-cache.c 	if (r5l_get_meta(log, meta_size)) {
log               884 drivers/md/raid5-cache.c 		mutex_unlock(&log->io_mutex);
log               889 drivers/md/raid5-cache.c 	io = log->current_io;
log               901 drivers/md/raid5-cache.c 	mutex_unlock(&log->io_mutex);
log               904 drivers/md/raid5-cache.c static int r5l_log_stripe(struct r5l_log *log, struct stripe_head *sh,
log               918 drivers/md/raid5-cache.c 	ret = r5l_get_meta(log, meta_size);
log               922 drivers/md/raid5-cache.c 	io = log->current_io;
log               934 drivers/md/raid5-cache.c 		    log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) {
log               942 drivers/md/raid5-cache.c 		r5l_append_payload_meta(log, R5LOG_PAYLOAD_DATA,
log               945 drivers/md/raid5-cache.c 		r5l_append_payload_page(log, sh->dev[i].page);
log               949 drivers/md/raid5-cache.c 		r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY,
log               952 drivers/md/raid5-cache.c 		r5l_append_payload_page(log, sh->dev[sh->pd_idx].page);
log               953 drivers/md/raid5-cache.c 		r5l_append_payload_page(log, sh->dev[sh->qd_idx].page);
log               955 drivers/md/raid5-cache.c 		r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY,
log               958 drivers/md/raid5-cache.c 		r5l_append_payload_page(log, sh->dev[sh->pd_idx].page);
log               966 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH)
log               972 drivers/md/raid5-cache.c 		spin_lock_irq(&log->stripe_in_journal_lock);
log               974 drivers/md/raid5-cache.c 			      &log->stripe_in_journal_list);
log               975 drivers/md/raid5-cache.c 		spin_unlock_irq(&log->stripe_in_journal_lock);
log               976 drivers/md/raid5-cache.c 		atomic_inc(&log->stripe_in_journal_count);
log               982 drivers/md/raid5-cache.c static inline void r5l_add_no_space_stripe(struct r5l_log *log,
log               985 drivers/md/raid5-cache.c 	spin_lock(&log->no_space_stripes_lock);
log               986 drivers/md/raid5-cache.c 	list_add_tail(&sh->log_list, &log->no_space_stripes);
log               987 drivers/md/raid5-cache.c 	spin_unlock(&log->no_space_stripes_lock);
log               994 drivers/md/raid5-cache.c int r5l_write_stripe(struct r5l_log *log, struct stripe_head *sh)
log              1004 drivers/md/raid5-cache.c 	if (!log)
log              1028 drivers/md/raid5-cache.c 		sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum,
log              1043 drivers/md/raid5-cache.c 	mutex_lock(&log->io_mutex);
log              1047 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) {
log              1048 drivers/md/raid5-cache.c 		if (!r5l_has_free_space(log, reserve)) {
log              1049 drivers/md/raid5-cache.c 			r5l_add_no_space_stripe(log, sh);
log              1052 drivers/md/raid5-cache.c 			ret = r5l_log_stripe(log, sh, data_pages, parity_pages);
log              1054 drivers/md/raid5-cache.c 				spin_lock_irq(&log->io_list_lock);
log              1056 drivers/md/raid5-cache.c 					      &log->no_mem_stripes);
log              1057 drivers/md/raid5-cache.c 				spin_unlock_irq(&log->io_list_lock);
log              1067 drivers/md/raid5-cache.c 			r5l_add_no_space_stripe(log, sh);
log              1070 drivers/md/raid5-cache.c 		} else if (!r5l_has_free_space(log, reserve)) {
log              1071 drivers/md/raid5-cache.c 			if (sh->log_start == log->last_checkpoint)
log              1074 drivers/md/raid5-cache.c 				r5l_add_no_space_stripe(log, sh);
log              1076 drivers/md/raid5-cache.c 			ret = r5l_log_stripe(log, sh, data_pages, parity_pages);
log              1078 drivers/md/raid5-cache.c 				spin_lock_irq(&log->io_list_lock);
log              1080 drivers/md/raid5-cache.c 					      &log->no_mem_stripes);
log              1081 drivers/md/raid5-cache.c 				spin_unlock_irq(&log->io_list_lock);
log              1086 drivers/md/raid5-cache.c 	mutex_unlock(&log->io_mutex);
log              1088 drivers/md/raid5-cache.c 		r5l_wake_reclaim(log, reserve);
log              1092 drivers/md/raid5-cache.c void r5l_write_stripe_run(struct r5l_log *log)
log              1094 drivers/md/raid5-cache.c 	if (!log)
log              1096 drivers/md/raid5-cache.c 	mutex_lock(&log->io_mutex);
log              1097 drivers/md/raid5-cache.c 	r5l_submit_current_io(log);
log              1098 drivers/md/raid5-cache.c 	mutex_unlock(&log->io_mutex);
log              1101 drivers/md/raid5-cache.c int r5l_handle_flush_request(struct r5l_log *log, struct bio *bio)
log              1103 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) {
log              1119 drivers/md/raid5-cache.c 			mutex_lock(&log->io_mutex);
log              1120 drivers/md/raid5-cache.c 			r5l_get_meta(log, 0);
log              1121 drivers/md/raid5-cache.c 			bio_list_add(&log->current_io->flush_barriers, bio);
log              1122 drivers/md/raid5-cache.c 			log->current_io->has_flush = 1;
log              1123 drivers/md/raid5-cache.c 			log->current_io->has_null_flush = 1;
log              1124 drivers/md/raid5-cache.c 			atomic_inc(&log->current_io->pending_stripe);
log              1125 drivers/md/raid5-cache.c 			r5l_submit_current_io(log);
log              1126 drivers/md/raid5-cache.c 			mutex_unlock(&log->io_mutex);
log              1134 drivers/md/raid5-cache.c static void r5l_run_no_space_stripes(struct r5l_log *log)
log              1138 drivers/md/raid5-cache.c 	spin_lock(&log->no_space_stripes_lock);
log              1139 drivers/md/raid5-cache.c 	while (!list_empty(&log->no_space_stripes)) {
log              1140 drivers/md/raid5-cache.c 		sh = list_first_entry(&log->no_space_stripes,
log              1146 drivers/md/raid5-cache.c 	spin_unlock(&log->no_space_stripes_lock);
log              1157 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              1161 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH)
log              1162 drivers/md/raid5-cache.c 		return log->next_checkpoint;
log              1164 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->stripe_in_journal_lock, flags);
log              1165 drivers/md/raid5-cache.c 	if (list_empty(&conf->log->stripe_in_journal_list)) {
log              1167 drivers/md/raid5-cache.c 		spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags);
log              1168 drivers/md/raid5-cache.c 		return log->next_checkpoint;
log              1170 drivers/md/raid5-cache.c 	sh = list_first_entry(&conf->log->stripe_in_journal_list,
log              1173 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags);
log              1177 drivers/md/raid5-cache.c static sector_t r5l_reclaimable_space(struct r5l_log *log)
log              1179 drivers/md/raid5-cache.c 	struct r5conf *conf = log->rdev->mddev->private;
log              1181 drivers/md/raid5-cache.c 	return r5l_ring_distance(log, log->last_checkpoint,
log              1185 drivers/md/raid5-cache.c static void r5l_run_no_mem_stripe(struct r5l_log *log)
log              1189 drivers/md/raid5-cache.c 	lockdep_assert_held(&log->io_list_lock);
log              1191 drivers/md/raid5-cache.c 	if (!list_empty(&log->no_mem_stripes)) {
log              1192 drivers/md/raid5-cache.c 		sh = list_first_entry(&log->no_mem_stripes,
log              1200 drivers/md/raid5-cache.c static bool r5l_complete_finished_ios(struct r5l_log *log)
log              1205 drivers/md/raid5-cache.c 	lockdep_assert_held(&log->io_list_lock);
log              1207 drivers/md/raid5-cache.c 	list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) {
log              1212 drivers/md/raid5-cache.c 		log->next_checkpoint = io->log_start;
log              1215 drivers/md/raid5-cache.c 		mempool_free(io, &log->io_pool);
log              1216 drivers/md/raid5-cache.c 		r5l_run_no_mem_stripe(log);
log              1226 drivers/md/raid5-cache.c 	struct r5l_log *log = io->log;
log              1227 drivers/md/raid5-cache.c 	struct r5conf *conf = log->rdev->mddev->private;
log              1230 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log              1233 drivers/md/raid5-cache.c 	if (!r5l_complete_finished_ios(log)) {
log              1234 drivers/md/raid5-cache.c 		spin_unlock_irqrestore(&log->io_list_lock, flags);
log              1238 drivers/md/raid5-cache.c 	if (r5l_reclaimable_space(log) > log->max_free_space ||
log              1240 drivers/md/raid5-cache.c 		r5l_wake_reclaim(log, 0);
log              1242 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log              1243 drivers/md/raid5-cache.c 	wake_up(&log->iounit_wait);
log              1259 drivers/md/raid5-cache.c 	struct r5l_log *log = container_of(bio, struct r5l_log,
log              1265 drivers/md/raid5-cache.c 		md_error(log->rdev->mddev, log->rdev);
log              1267 drivers/md/raid5-cache.c 	spin_lock_irqsave(&log->io_list_lock, flags);
log              1268 drivers/md/raid5-cache.c 	list_for_each_entry(io, &log->flushing_ios, log_sibling)
log              1270 drivers/md/raid5-cache.c 	list_splice_tail_init(&log->flushing_ios, &log->finished_ios);
log              1271 drivers/md/raid5-cache.c 	spin_unlock_irqrestore(&log->io_list_lock, flags);
log              1288 drivers/md/raid5-cache.c void r5l_flush_stripe_to_raid(struct r5l_log *log)
log              1292 drivers/md/raid5-cache.c 	if (!log || !log->need_cache_flush)
log              1295 drivers/md/raid5-cache.c 	spin_lock_irq(&log->io_list_lock);
log              1297 drivers/md/raid5-cache.c 	if (!list_empty(&log->flushing_ios)) {
log              1298 drivers/md/raid5-cache.c 		spin_unlock_irq(&log->io_list_lock);
log              1301 drivers/md/raid5-cache.c 	list_splice_tail_init(&log->io_end_ios, &log->flushing_ios);
log              1302 drivers/md/raid5-cache.c 	do_flush = !list_empty(&log->flushing_ios);
log              1303 drivers/md/raid5-cache.c 	spin_unlock_irq(&log->io_list_lock);
log              1307 drivers/md/raid5-cache.c 	bio_reset(&log->flush_bio);
log              1308 drivers/md/raid5-cache.c 	bio_set_dev(&log->flush_bio, log->rdev->bdev);
log              1309 drivers/md/raid5-cache.c 	log->flush_bio.bi_end_io = r5l_log_flush_endio;
log              1310 drivers/md/raid5-cache.c 	log->flush_bio.bi_opf = REQ_OP_WRITE | REQ_PREFLUSH;
log              1311 drivers/md/raid5-cache.c 	submit_bio(&log->flush_bio);
log              1314 drivers/md/raid5-cache.c static void r5l_write_super(struct r5l_log *log, sector_t cp);
log              1315 drivers/md/raid5-cache.c static void r5l_write_super_and_discard_space(struct r5l_log *log,
log              1318 drivers/md/raid5-cache.c 	struct block_device *bdev = log->rdev->bdev;
log              1321 drivers/md/raid5-cache.c 	r5l_write_super(log, end);
log              1326 drivers/md/raid5-cache.c 	mddev = log->rdev->mddev;
log              1346 drivers/md/raid5-cache.c 	if (log->last_checkpoint < end) {
log              1348 drivers/md/raid5-cache.c 				log->last_checkpoint + log->rdev->data_offset,
log              1349 drivers/md/raid5-cache.c 				end - log->last_checkpoint, GFP_NOIO, 0);
log              1352 drivers/md/raid5-cache.c 				log->last_checkpoint + log->rdev->data_offset,
log              1353 drivers/md/raid5-cache.c 				log->device_size - log->last_checkpoint,
log              1355 drivers/md/raid5-cache.c 		blkdev_issue_discard(bdev, log->rdev->data_offset, end,
log              1405 drivers/md/raid5-cache.c 	if (!conf->log)
log              1426 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              1434 drivers/md/raid5-cache.c 	if (!r5c_is_writeback(log))
log              1470 drivers/md/raid5-cache.c 		spin_lock_irqsave(&log->stripe_in_journal_lock, flags);
log              1472 drivers/md/raid5-cache.c 		list_for_each_entry(sh, &log->stripe_in_journal_list, r5c) {
log              1490 drivers/md/raid5-cache.c 		spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags);
log              1494 drivers/md/raid5-cache.c 		r5l_run_no_space_stripes(log);
log              1499 drivers/md/raid5-cache.c static void r5l_do_reclaim(struct r5l_log *log)
log              1501 drivers/md/raid5-cache.c 	struct r5conf *conf = log->rdev->mddev->private;
log              1502 drivers/md/raid5-cache.c 	sector_t reclaim_target = xchg(&log->reclaim_target, 0);
log              1507 drivers/md/raid5-cache.c 	spin_lock_irq(&log->io_list_lock);
log              1508 drivers/md/raid5-cache.c 	write_super = r5l_reclaimable_space(log) > log->max_free_space ||
log              1509 drivers/md/raid5-cache.c 		reclaim_target != 0 || !list_empty(&log->no_space_stripes);
log              1516 drivers/md/raid5-cache.c 		reclaimable = r5l_reclaimable_space(log);
log              1518 drivers/md/raid5-cache.c 		    (list_empty(&log->running_ios) &&
log              1519 drivers/md/raid5-cache.c 		     list_empty(&log->io_end_ios) &&
log              1520 drivers/md/raid5-cache.c 		     list_empty(&log->flushing_ios) &&
log              1521 drivers/md/raid5-cache.c 		     list_empty(&log->finished_ios)))
log              1524 drivers/md/raid5-cache.c 		md_wakeup_thread(log->rdev->mddev->thread);
log              1525 drivers/md/raid5-cache.c 		wait_event_lock_irq(log->iounit_wait,
log              1526 drivers/md/raid5-cache.c 				    r5l_reclaimable_space(log) > reclaimable,
log              1527 drivers/md/raid5-cache.c 				    log->io_list_lock);
log              1531 drivers/md/raid5-cache.c 	spin_unlock_irq(&log->io_list_lock);
log              1541 drivers/md/raid5-cache.c 	r5l_write_super_and_discard_space(log, next_checkpoint);
log              1543 drivers/md/raid5-cache.c 	mutex_lock(&log->io_mutex);
log              1544 drivers/md/raid5-cache.c 	log->last_checkpoint = next_checkpoint;
log              1545 drivers/md/raid5-cache.c 	r5c_update_log_state(log);
log              1546 drivers/md/raid5-cache.c 	mutex_unlock(&log->io_mutex);
log              1548 drivers/md/raid5-cache.c 	r5l_run_no_space_stripes(log);
log              1555 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              1557 drivers/md/raid5-cache.c 	if (!log)
log              1560 drivers/md/raid5-cache.c 	r5l_do_reclaim(log);
log              1563 drivers/md/raid5-cache.c void r5l_wake_reclaim(struct r5l_log *log, sector_t space)
log              1568 drivers/md/raid5-cache.c 	if (!log)
log              1571 drivers/md/raid5-cache.c 		target = log->reclaim_target;
log              1574 drivers/md/raid5-cache.c 	} while (cmpxchg(&log->reclaim_target, target, new) != target);
log              1575 drivers/md/raid5-cache.c 	md_wakeup_thread(log->reclaim_thread);
log              1578 drivers/md/raid5-cache.c void r5l_quiesce(struct r5l_log *log, int quiesce)
log              1584 drivers/md/raid5-cache.c 		mddev = log->rdev->mddev;
log              1586 drivers/md/raid5-cache.c 		kthread_park(log->reclaim_thread->tsk);
log              1587 drivers/md/raid5-cache.c 		r5l_wake_reclaim(log, MaxSector);
log              1588 drivers/md/raid5-cache.c 		r5l_do_reclaim(log);
log              1590 drivers/md/raid5-cache.c 		kthread_unpark(log->reclaim_thread->tsk);
log              1595 drivers/md/raid5-cache.c 	struct r5l_log *log;
log              1599 drivers/md/raid5-cache.c 	log = rcu_dereference(conf->log);
log              1601 drivers/md/raid5-cache.c 	if (!log)
log              1604 drivers/md/raid5-cache.c 		ret = test_bit(Faulty, &log->rdev->flags);
log              1634 drivers/md/raid5-cache.c static int r5l_recovery_allocate_ra_pool(struct r5l_log *log,
log              1639 drivers/md/raid5-cache.c 	ctx->ra_bio = bio_alloc_bioset(GFP_KERNEL, BIO_MAX_PAGES, &log->bs);
log              1663 drivers/md/raid5-cache.c static void r5l_recovery_free_ra_pool(struct r5l_log *log,
log              1679 drivers/md/raid5-cache.c static int r5l_recovery_fetch_ra_pool(struct r5l_log *log,
log              1684 drivers/md/raid5-cache.c 	bio_set_dev(ctx->ra_bio, log->rdev->bdev);
log              1686 drivers/md/raid5-cache.c 	ctx->ra_bio->bi_iter.bi_sector = log->rdev->data_offset + offset;
log              1696 drivers/md/raid5-cache.c 		offset = r5l_ring_add(log, offset, BLOCK_SECTORS);
log              1709 drivers/md/raid5-cache.c static int r5l_recovery_read_page(struct r5l_log *log,
log              1718 drivers/md/raid5-cache.c 		ret = r5l_recovery_fetch_ra_pool(log, ctx, offset);
log              1733 drivers/md/raid5-cache.c static int r5l_recovery_read_meta_block(struct r5l_log *log,
log              1741 drivers/md/raid5-cache.c 	ret = r5l_recovery_read_page(log, ctx, page, ctx->pos);
log              1755 drivers/md/raid5-cache.c 	crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE);
log              1768 drivers/md/raid5-cache.c r5l_recovery_create_empty_meta_block(struct r5l_log *log,
log              1783 drivers/md/raid5-cache.c static int r5l_log_write_empty_meta_block(struct r5l_log *log, sector_t pos,
log              1792 drivers/md/raid5-cache.c 	r5l_recovery_create_empty_meta_block(log, page, pos, seq);
log              1794 drivers/md/raid5-cache.c 	mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum,
log              1796 drivers/md/raid5-cache.c 	if (!sync_page_io(log->rdev, pos, PAGE_SIZE, page, REQ_OP_WRITE,
log              1812 drivers/md/raid5-cache.c static void r5l_recovery_load_data(struct r5l_log *log,
log              1818 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              1825 drivers/md/raid5-cache.c 	r5l_recovery_read_page(log, ctx, sh->dev[dd_idx].page, log_offset);
log              1834 drivers/md/raid5-cache.c static void r5l_recovery_load_parity(struct r5l_log *log,
log              1840 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              1844 drivers/md/raid5-cache.c 	r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset);
log              1851 drivers/md/raid5-cache.c 			log, ctx, sh->dev[sh->qd_idx].page,
log              1852 drivers/md/raid5-cache.c 			r5l_ring_add(log, log_offset, BLOCK_SECTORS));
log              1985 drivers/md/raid5-cache.c r5l_recovery_verify_data_checksum(struct r5l_log *log,
log              1993 drivers/md/raid5-cache.c 	r5l_recovery_read_page(log, ctx, page, log_offset);
log              1995 drivers/md/raid5-cache.c 	checksum = crc32c_le(log->uuid_checksum, addr, PAGE_SIZE);
log              2005 drivers/md/raid5-cache.c r5l_recovery_verify_data_checksum_for_mb(struct r5l_log *log,
log              2008 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2012 drivers/md/raid5-cache.c 	sector_t log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS);
log              2027 drivers/md/raid5-cache.c 				    log, ctx, page, log_offset,
log              2032 drivers/md/raid5-cache.c 				    log, ctx, page, log_offset,
log              2037 drivers/md/raid5-cache.c 				    log, ctx, page,
log              2038 drivers/md/raid5-cache.c 				    r5l_ring_add(log, log_offset,
log              2052 drivers/md/raid5-cache.c 			log_offset = r5l_ring_add(log, log_offset,
log              2078 drivers/md/raid5-cache.c r5c_recovery_analyze_meta_block(struct r5l_log *log,
log              2082 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2098 drivers/md/raid5-cache.c 	ret = r5l_recovery_verify_data_checksum_for_mb(log, ctx);
log              2106 drivers/md/raid5-cache.c 	log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS);
log              2189 drivers/md/raid5-cache.c 			r5l_recovery_load_data(log, sh, ctx, payload,
log              2192 drivers/md/raid5-cache.c 			r5l_recovery_load_parity(log, sh, ctx, payload,
log              2197 drivers/md/raid5-cache.c 		log_offset = r5l_ring_add(log, log_offset,
log              2212 drivers/md/raid5-cache.c static void r5c_recovery_load_one_stripe(struct r5l_log *log,
log              2243 drivers/md/raid5-cache.c static int r5c_recovery_flush_log(struct r5l_log *log,
log              2251 drivers/md/raid5-cache.c 		if (r5l_recovery_read_meta_block(log, ctx))
log              2254 drivers/md/raid5-cache.c 		ret = r5c_recovery_analyze_meta_block(log, ctx,
log              2263 drivers/md/raid5-cache.c 		ctx->pos = r5l_ring_add(log, ctx->pos, ctx->meta_total_blocks);
log              2277 drivers/md/raid5-cache.c 		r5c_recovery_load_one_stripe(log, sh);
log              2354 drivers/md/raid5-cache.c r5c_recovery_rewrite_data_only_stripes(struct r5l_log *log,
log              2358 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2378 drivers/md/raid5-cache.c 		r5l_recovery_create_empty_meta_block(log, page,
log              2382 drivers/md/raid5-cache.c 		write_pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS);
log              2398 drivers/md/raid5-cache.c 					crc32c_le(log->uuid_checksum, addr,
log              2401 drivers/md/raid5-cache.c 				sync_page_io(log->rdev, write_pos, PAGE_SIZE,
log              2403 drivers/md/raid5-cache.c 				write_pos = r5l_ring_add(log, write_pos,
log              2411 drivers/md/raid5-cache.c 		mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum,
log              2413 drivers/md/raid5-cache.c 		sync_page_io(log->rdev, ctx->pos, PAGE_SIZE, page,
log              2416 drivers/md/raid5-cache.c 		list_add_tail(&sh->r5c, &log->stripe_in_journal_list);
log              2417 drivers/md/raid5-cache.c 		atomic_inc(&log->stripe_in_journal_count);
log              2422 drivers/md/raid5-cache.c 	log->next_checkpoint = next_checkpoint;
log              2427 drivers/md/raid5-cache.c static void r5c_recovery_flush_data_only_stripes(struct r5l_log *log,
log              2430 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2437 drivers/md/raid5-cache.c 	log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_BACK;
log              2450 drivers/md/raid5-cache.c 	log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH;
log              2453 drivers/md/raid5-cache.c static int r5l_recovery_log(struct r5l_log *log)
log              2455 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2464 drivers/md/raid5-cache.c 	ctx->pos = log->last_checkpoint;
log              2465 drivers/md/raid5-cache.c 	ctx->seq = log->last_cp_seq;
log              2474 drivers/md/raid5-cache.c 	if (r5l_recovery_allocate_ra_pool(log, ctx) != 0) {
log              2479 drivers/md/raid5-cache.c 	ret = r5c_recovery_flush_log(log, ctx);
log              2496 drivers/md/raid5-cache.c 		log->next_checkpoint = ctx->pos;
log              2497 drivers/md/raid5-cache.c 		r5l_log_write_empty_meta_block(log, ctx->pos, ctx->seq++);
log              2498 drivers/md/raid5-cache.c 		ctx->pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS);
log              2499 drivers/md/raid5-cache.c 	} else if (r5c_recovery_rewrite_data_only_stripes(log, ctx)) {
log              2506 drivers/md/raid5-cache.c 	log->log_start = ctx->pos;
log              2507 drivers/md/raid5-cache.c 	log->seq = ctx->seq;
log              2508 drivers/md/raid5-cache.c 	log->last_checkpoint = pos;
log              2509 drivers/md/raid5-cache.c 	r5l_write_super(log, pos);
log              2511 drivers/md/raid5-cache.c 	r5c_recovery_flush_data_only_stripes(log, ctx);
log              2514 drivers/md/raid5-cache.c 	r5l_recovery_free_ra_pool(log, ctx);
log              2522 drivers/md/raid5-cache.c static void r5l_write_super(struct r5l_log *log, sector_t cp)
log              2524 drivers/md/raid5-cache.c 	struct mddev *mddev = log->rdev->mddev;
log              2526 drivers/md/raid5-cache.c 	log->rdev->journal_tail = cp;
log              2540 drivers/md/raid5-cache.c 	if (!conf || !conf->log) {
log              2545 drivers/md/raid5-cache.c 	switch (conf->log->r5c_journal_mode) {
log              2580 drivers/md/raid5-cache.c 	if (!conf || !conf->log)
log              2588 drivers/md/raid5-cache.c 	conf->log->r5c_journal_mode = mode;
log              2639 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              2648 drivers/md/raid5-cache.c 	BUG_ON(!r5c_is_writeback(log));
log              2698 drivers/md/raid5-cache.c 		spin_lock(&log->tree_lock);
log              2699 drivers/md/raid5-cache.c 		pslot = radix_tree_lookup_slot(&log->big_stripe_tree,
log              2703 drivers/md/raid5-cache.c 				pslot, &log->tree_lock) >>
log              2706 drivers/md/raid5-cache.c 				&log->big_stripe_tree, pslot,
log              2714 drivers/md/raid5-cache.c 				&log->big_stripe_tree, tree_index,
log              2717 drivers/md/raid5-cache.c 				spin_unlock(&log->tree_lock);
log              2722 drivers/md/raid5-cache.c 		spin_unlock(&log->tree_lock);
log              2806 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              2813 drivers/md/raid5-cache.c 	if (!log || !test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags))
log              2819 drivers/md/raid5-cache.c 	if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH)
log              2841 drivers/md/raid5-cache.c 	spin_lock_irq(&log->stripe_in_journal_lock);
log              2843 drivers/md/raid5-cache.c 	spin_unlock_irq(&log->stripe_in_journal_lock);
log              2846 drivers/md/raid5-cache.c 	atomic_dec(&log->stripe_in_journal_count);
log              2847 drivers/md/raid5-cache.c 	r5c_update_log_state(log);
log              2853 drivers/md/raid5-cache.c 		spin_lock(&log->tree_lock);
log              2854 drivers/md/raid5-cache.c 		pslot = radix_tree_lookup_slot(&log->big_stripe_tree,
log              2858 drivers/md/raid5-cache.c 			pslot, &log->tree_lock) >>
log              2861 drivers/md/raid5-cache.c 			radix_tree_delete(&log->big_stripe_tree, tree_index);
log              2864 drivers/md/raid5-cache.c 				&log->big_stripe_tree, pslot,
log              2866 drivers/md/raid5-cache.c 		spin_unlock(&log->tree_lock);
log              2881 drivers/md/raid5-cache.c 	r5l_append_flush_payload(log, sh->sector);
log              2887 drivers/md/raid5-cache.c int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh)
log              2895 drivers/md/raid5-cache.c 	BUG_ON(!log);
log              2903 drivers/md/raid5-cache.c 		sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum,
log              2917 drivers/md/raid5-cache.c 	mutex_lock(&log->io_mutex);
log              2923 drivers/md/raid5-cache.c 		r5l_add_no_space_stripe(log, sh);
log              2924 drivers/md/raid5-cache.c 	else if (!r5l_has_free_space(log, reserve)) {
log              2925 drivers/md/raid5-cache.c 		if (sh->log_start == log->last_checkpoint)
log              2928 drivers/md/raid5-cache.c 			r5l_add_no_space_stripe(log, sh);
log              2930 drivers/md/raid5-cache.c 		ret = r5l_log_stripe(log, sh, pages, 0);
log              2932 drivers/md/raid5-cache.c 			spin_lock_irq(&log->io_list_lock);
log              2933 drivers/md/raid5-cache.c 			list_add_tail(&sh->log_list, &log->no_mem_stripes);
log              2934 drivers/md/raid5-cache.c 			spin_unlock_irq(&log->io_list_lock);
log              2938 drivers/md/raid5-cache.c 	mutex_unlock(&log->io_mutex);
log              2945 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              2949 drivers/md/raid5-cache.c 	if (!log)
log              2954 drivers/md/raid5-cache.c 	slot = radix_tree_lookup(&log->big_stripe_tree, tree_index);
log              2958 drivers/md/raid5-cache.c static int r5l_load_log(struct r5l_log *log)
log              2960 drivers/md/raid5-cache.c 	struct md_rdev *rdev = log->rdev;
log              2963 drivers/md/raid5-cache.c 	sector_t cp = log->rdev->journal_tail;
log              2988 drivers/md/raid5-cache.c 	expected_crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE);
log              2999 drivers/md/raid5-cache.c 		log->last_cp_seq = prandom_u32();
log              3001 drivers/md/raid5-cache.c 		r5l_log_write_empty_meta_block(log, cp, log->last_cp_seq);
log              3007 drivers/md/raid5-cache.c 		r5l_write_super(log, cp);
log              3009 drivers/md/raid5-cache.c 		log->last_cp_seq = le64_to_cpu(mb->seq);
log              3011 drivers/md/raid5-cache.c 	log->device_size = round_down(rdev->sectors, BLOCK_SECTORS);
log              3012 drivers/md/raid5-cache.c 	log->max_free_space = log->device_size >> RECLAIM_MAX_FREE_SPACE_SHIFT;
log              3013 drivers/md/raid5-cache.c 	if (log->max_free_space > RECLAIM_MAX_FREE_SPACE)
log              3014 drivers/md/raid5-cache.c 		log->max_free_space = RECLAIM_MAX_FREE_SPACE;
log              3015 drivers/md/raid5-cache.c 	log->last_checkpoint = cp;
log              3020 drivers/md/raid5-cache.c 		log->log_start = r5l_ring_add(log, cp, BLOCK_SECTORS);
log              3021 drivers/md/raid5-cache.c 		log->seq = log->last_cp_seq + 1;
log              3022 drivers/md/raid5-cache.c 		log->next_checkpoint = cp;
log              3024 drivers/md/raid5-cache.c 		ret = r5l_recovery_log(log);
log              3026 drivers/md/raid5-cache.c 	r5c_update_log_state(log);
log              3033 drivers/md/raid5-cache.c int r5l_start(struct r5l_log *log)
log              3037 drivers/md/raid5-cache.c 	if (!log)
log              3040 drivers/md/raid5-cache.c 	ret = r5l_load_log(log);
log              3042 drivers/md/raid5-cache.c 		struct mddev *mddev = log->rdev->mddev;
log              3053 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              3055 drivers/md/raid5-cache.c 	if (!log)
log              3060 drivers/md/raid5-cache.c 	    conf->log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK)
log              3061 drivers/md/raid5-cache.c 		schedule_work(&log->disable_writeback_work);
log              3067 drivers/md/raid5-cache.c 	struct r5l_log *log;
log              3092 drivers/md/raid5-cache.c 	log = kzalloc(sizeof(*log), GFP_KERNEL);
log              3093 drivers/md/raid5-cache.c 	if (!log)
log              3095 drivers/md/raid5-cache.c 	log->rdev = rdev;
log              3097 drivers/md/raid5-cache.c 	log->need_cache_flush = test_bit(QUEUE_FLAG_WC, &q->queue_flags) != 0;
log              3099 drivers/md/raid5-cache.c 	log->uuid_checksum = crc32c_le(~0, rdev->mddev->uuid,
log              3102 drivers/md/raid5-cache.c 	mutex_init(&log->io_mutex);
log              3104 drivers/md/raid5-cache.c 	spin_lock_init(&log->io_list_lock);
log              3105 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->running_ios);
log              3106 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->io_end_ios);
log              3107 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->flushing_ios);
log              3108 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->finished_ios);
log              3109 drivers/md/raid5-cache.c 	bio_init(&log->flush_bio, NULL, 0);
log              3111 drivers/md/raid5-cache.c 	log->io_kc = KMEM_CACHE(r5l_io_unit, 0);
log              3112 drivers/md/raid5-cache.c 	if (!log->io_kc)
log              3115 drivers/md/raid5-cache.c 	ret = mempool_init_slab_pool(&log->io_pool, R5L_POOL_SIZE, log->io_kc);
log              3119 drivers/md/raid5-cache.c 	ret = bioset_init(&log->bs, R5L_POOL_SIZE, 0, BIOSET_NEED_BVECS);
log              3123 drivers/md/raid5-cache.c 	ret = mempool_init_page_pool(&log->meta_pool, R5L_POOL_SIZE, 0);
log              3127 drivers/md/raid5-cache.c 	spin_lock_init(&log->tree_lock);
log              3128 drivers/md/raid5-cache.c 	INIT_RADIX_TREE(&log->big_stripe_tree, GFP_NOWAIT | __GFP_NOWARN);
log              3130 drivers/md/raid5-cache.c 	log->reclaim_thread = md_register_thread(r5l_reclaim_thread,
log              3131 drivers/md/raid5-cache.c 						 log->rdev->mddev, "reclaim");
log              3132 drivers/md/raid5-cache.c 	if (!log->reclaim_thread)
log              3134 drivers/md/raid5-cache.c 	log->reclaim_thread->timeout = R5C_RECLAIM_WAKEUP_INTERVAL;
log              3136 drivers/md/raid5-cache.c 	init_waitqueue_head(&log->iounit_wait);
log              3138 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->no_mem_stripes);
log              3140 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->no_space_stripes);
log              3141 drivers/md/raid5-cache.c 	spin_lock_init(&log->no_space_stripes_lock);
log              3143 drivers/md/raid5-cache.c 	INIT_WORK(&log->deferred_io_work, r5l_submit_io_async);
log              3144 drivers/md/raid5-cache.c 	INIT_WORK(&log->disable_writeback_work, r5c_disable_writeback_async);
log              3146 drivers/md/raid5-cache.c 	log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH;
log              3147 drivers/md/raid5-cache.c 	INIT_LIST_HEAD(&log->stripe_in_journal_list);
log              3148 drivers/md/raid5-cache.c 	spin_lock_init(&log->stripe_in_journal_lock);
log              3149 drivers/md/raid5-cache.c 	atomic_set(&log->stripe_in_journal_count, 0);
log              3151 drivers/md/raid5-cache.c 	rcu_assign_pointer(conf->log, log);
log              3157 drivers/md/raid5-cache.c 	mempool_exit(&log->meta_pool);
log              3159 drivers/md/raid5-cache.c 	bioset_exit(&log->bs);
log              3161 drivers/md/raid5-cache.c 	mempool_exit(&log->io_pool);
log              3163 drivers/md/raid5-cache.c 	kmem_cache_destroy(log->io_kc);
log              3165 drivers/md/raid5-cache.c 	kfree(log);
log              3171 drivers/md/raid5-cache.c 	struct r5l_log *log = conf->log;
log              3173 drivers/md/raid5-cache.c 	conf->log = NULL;
log              3178 drivers/md/raid5-cache.c 	flush_work(&log->disable_writeback_work);
log              3179 drivers/md/raid5-cache.c 	md_unregister_thread(&log->reclaim_thread);
log              3180 drivers/md/raid5-cache.c 	mempool_exit(&log->meta_pool);
log              3181 drivers/md/raid5-cache.c 	bioset_exit(&log->bs);
log              3182 drivers/md/raid5-cache.c 	mempool_exit(&log->io_pool);
log              3183 drivers/md/raid5-cache.c 	kmem_cache_destroy(log->io_kc);
log              3184 drivers/md/raid5-cache.c 	kfree(log);
log                 7 drivers/md/raid5-log.h extern int r5l_write_stripe(struct r5l_log *log, struct stripe_head *head_sh);
log                 8 drivers/md/raid5-log.h extern void r5l_write_stripe_run(struct r5l_log *log);
log                 9 drivers/md/raid5-log.h extern void r5l_flush_stripe_to_raid(struct r5l_log *log);
log                11 drivers/md/raid5-log.h extern int r5l_handle_flush_request(struct r5l_log *log, struct bio *bio);
log                12 drivers/md/raid5-log.h extern void r5l_quiesce(struct r5l_log *log, int quiesce);
log                14 drivers/md/raid5-log.h extern bool r5c_is_writeback(struct r5l_log *log);
log                23 drivers/md/raid5-log.h extern void r5l_wake_reclaim(struct r5l_log *log, sector_t space);
log                26 drivers/md/raid5-log.h extern int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh);
log                35 drivers/md/raid5-log.h extern int r5l_start(struct r5l_log *log);
log                47 drivers/md/raid5-log.h extern int ppl_handle_flush_request(struct r5l_log *log, struct bio *bio);
log                64 drivers/md/raid5-log.h 	if (conf->log) {
log                69 drivers/md/raid5-log.h 			return r5l_write_stripe(conf->log, sh);
log                72 drivers/md/raid5-log.h 			return r5c_cache_data(conf->log, sh);
log                85 drivers/md/raid5-log.h 	if (conf->log)
log                93 drivers/md/raid5-log.h 	if (conf->log)
log                94 drivers/md/raid5-log.h 		r5l_write_stripe_run(conf->log);
log               101 drivers/md/raid5-log.h 	if (conf->log)
log               102 drivers/md/raid5-log.h 		r5l_flush_stripe_to_raid(conf->log);
log               111 drivers/md/raid5-log.h 	if (conf->log)
log               112 drivers/md/raid5-log.h 		ret = r5l_handle_flush_request(conf->log, bio);
log               114 drivers/md/raid5-log.h 		ret = ppl_handle_flush_request(conf->log, bio);
log               121 drivers/md/raid5-log.h 	if (conf->log)
log               122 drivers/md/raid5-log.h 		r5l_quiesce(conf->log, quiesce);
log               129 drivers/md/raid5-log.h 	if (conf->log)
log               136 drivers/md/raid5-ppl.c 	struct ppl_log *log;
log               232 drivers/md/raid5-ppl.c static struct ppl_io_unit *ppl_new_iounit(struct ppl_log *log,
log               235 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               248 drivers/md/raid5-ppl.c 	io->log = log;
log               266 drivers/md/raid5-ppl.c static int ppl_log_stripe(struct ppl_log *log, struct stripe_head *sh)
log               268 drivers/md/raid5-ppl.c 	struct ppl_io_unit *io = log->current_io;
log               279 drivers/md/raid5-ppl.c 	if (io && (io->pp_size == log->entry_space ||
log               288 drivers/md/raid5-ppl.c 		io = ppl_new_iounit(log, sh);
log               291 drivers/md/raid5-ppl.c 		spin_lock_irq(&log->io_list_lock);
log               292 drivers/md/raid5-ppl.c 		list_add_tail(&io->log_sibling, &log->io_list);
log               293 drivers/md/raid5-ppl.c 		spin_unlock_irq(&log->io_list_lock);
log               295 drivers/md/raid5-ppl.c 		log->current_io = io;
log               364 drivers/md/raid5-ppl.c 	struct ppl_log *log;
log               373 drivers/md/raid5-ppl.c 	log = &ppl_conf->child_logs[sh->pd_idx];
log               375 drivers/md/raid5-ppl.c 	mutex_lock(&log->io_mutex);
log               377 drivers/md/raid5-ppl.c 	if (!log->rdev || test_bit(Faulty, &log->rdev->flags)) {
log               378 drivers/md/raid5-ppl.c 		mutex_unlock(&log->io_mutex);
log               386 drivers/md/raid5-ppl.c 	if (ppl_log_stripe(log, sh)) {
log               392 drivers/md/raid5-ppl.c 	mutex_unlock(&log->io_mutex);
log               400 drivers/md/raid5-ppl.c 	struct ppl_log *log = io->log;
log               401 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               407 drivers/md/raid5-ppl.c 		md_error(ppl_conf->mddev, log->rdev);
log               431 drivers/md/raid5-ppl.c 	struct ppl_log *log = io->log;
log               432 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               440 drivers/md/raid5-ppl.c 	if (!log->rdev || test_bit(Faulty, &log->rdev->flags)) {
log               461 drivers/md/raid5-ppl.c 	if (log->use_multippl &&
log               462 drivers/md/raid5-ppl.c 	    log->rdev->ppl.sector + log->rdev->ppl.size - log->next_io_sector <
log               464 drivers/md/raid5-ppl.c 		log->next_io_sector = log->rdev->ppl.sector;
log               469 drivers/md/raid5-ppl.c 	bio_set_dev(bio, log->rdev->bdev);
log               470 drivers/md/raid5-ppl.c 	bio->bi_iter.bi_sector = log->next_io_sector;
log               475 drivers/md/raid5-ppl.c 	    (unsigned long long)log->next_io_sector);
log               477 drivers/md/raid5-ppl.c 	if (log->use_multippl)
log               478 drivers/md/raid5-ppl.c 		log->next_io_sector += (PPL_HEADER_SIZE + io->pp_size) >> 9;
log               480 drivers/md/raid5-ppl.c 	WARN_ON(log->disk_flush_bitmap != 0);
log               488 drivers/md/raid5-ppl.c 				set_bit(i, &log->disk_flush_bitmap);
log               515 drivers/md/raid5-ppl.c static void ppl_submit_current_io(struct ppl_log *log)
log               519 drivers/md/raid5-ppl.c 	spin_lock_irq(&log->io_list_lock);
log               521 drivers/md/raid5-ppl.c 	io = list_first_entry_or_null(&log->io_list, struct ppl_io_unit,
log               526 drivers/md/raid5-ppl.c 	spin_unlock_irq(&log->io_list_lock);
log               531 drivers/md/raid5-ppl.c 		if (io == log->current_io)
log               532 drivers/md/raid5-ppl.c 			log->current_io = NULL;
log               541 drivers/md/raid5-ppl.c 	struct ppl_log *log;
log               545 drivers/md/raid5-ppl.c 		log = &ppl_conf->child_logs[i];
log               547 drivers/md/raid5-ppl.c 		mutex_lock(&log->io_mutex);
log               548 drivers/md/raid5-ppl.c 		ppl_submit_current_io(log);
log               549 drivers/md/raid5-ppl.c 		mutex_unlock(&log->io_mutex);
log               555 drivers/md/raid5-ppl.c 	struct ppl_log *log = io->log;
log               556 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               564 drivers/md/raid5-ppl.c 	spin_lock(&log->io_list_lock);
log               566 drivers/md/raid5-ppl.c 	spin_unlock(&log->io_list_lock);
log               590 drivers/md/raid5-ppl.c 	struct ppl_log *log = io->log;
log               591 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               617 drivers/md/raid5-ppl.c 	struct ppl_log *log = io->log;
log               618 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               626 drivers/md/raid5-ppl.c 	for_each_set_bit(i, &log->disk_flush_bitmap, raid_disks) {
log               654 drivers/md/raid5-ppl.c 	log->disk_flush_bitmap = 0;
log               663 drivers/md/raid5-ppl.c 					    struct ppl_log *log)
log               667 drivers/md/raid5-ppl.c 	io = list_first_entry_or_null(&log->io_list, struct ppl_io_unit,
log               680 drivers/md/raid5-ppl.c 			struct ppl_log *log = &ppl_conf->child_logs[i];
log               682 drivers/md/raid5-ppl.c 			spin_lock_irq(&log->io_list_lock);
log               684 drivers/md/raid5-ppl.c 					    ppl_no_io_unit_submitted(conf, log),
log               685 drivers/md/raid5-ppl.c 					    log->io_list_lock);
log               686 drivers/md/raid5-ppl.c 			spin_unlock_irq(&log->io_list_lock);
log               691 drivers/md/raid5-ppl.c int ppl_handle_flush_request(struct r5l_log *log, struct bio *bio)
log               709 drivers/md/raid5-ppl.c 		if (io->log->disk_flush_bitmap)
log               795 drivers/md/raid5-ppl.c static int ppl_recover_entry(struct ppl_log *log, struct ppl_header_entry *e,
log               798 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               928 drivers/md/raid5-ppl.c 			if (!sync_page_io(log->rdev,
log               929 drivers/md/raid5-ppl.c 					ppl_sector - log->rdev->data_offset + i,
log               934 drivers/md/raid5-ppl.c 				md_error(mddev, log->rdev);
log               948 drivers/md/raid5-ppl.c 		BUG_ON(parity_rdev->bdev->bd_dev != log->rdev->bdev->bd_dev);
log               970 drivers/md/raid5-ppl.c static int ppl_recover(struct ppl_log *log, struct ppl_header *pplhdr,
log               973 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log               974 drivers/md/raid5-ppl.c 	struct md_rdev *rdev = log->rdev;
log              1030 drivers/md/raid5-ppl.c 			ret = ppl_recover_entry(log, e, ppl_sector);
log              1046 drivers/md/raid5-ppl.c static int ppl_write_empty_header(struct ppl_log *log)
log              1050 drivers/md/raid5-ppl.c 	struct md_rdev *rdev = log->rdev;
log              1063 drivers/md/raid5-ppl.c 			    log->rdev->ppl.size, GFP_NOIO, 0);
log              1065 drivers/md/raid5-ppl.c 	pplhdr->signature = cpu_to_le32(log->ppl_conf->signature);
log              1079 drivers/md/raid5-ppl.c static int ppl_load_distributed(struct ppl_log *log)
log              1081 drivers/md/raid5-ppl.c 	struct ppl_conf *ppl_conf = log->ppl_conf;
log              1082 drivers/md/raid5-ppl.c 	struct md_rdev *rdev = log->rdev;
log              1180 drivers/md/raid5-ppl.c 		ret = ppl_recover(log, pplhdr, pplhdr_offset);
log              1184 drivers/md/raid5-ppl.c 		ret = ppl_write_empty_header(log);
log              1203 drivers/md/raid5-ppl.c 		struct ppl_log *log = &ppl_conf->child_logs[i];
log              1206 drivers/md/raid5-ppl.c 		if (!log->rdev)
log              1209 drivers/md/raid5-ppl.c 		ret = ppl_load_distributed(log);
log              1309 drivers/md/raid5-ppl.c static void ppl_init_child_log(struct ppl_log *log, struct md_rdev *rdev)
log              1315 drivers/md/raid5-ppl.c 		log->use_multippl = true;
log              1317 drivers/md/raid5-ppl.c 			&log->ppl_conf->mddev->flags);
log              1318 drivers/md/raid5-ppl.c 		log->entry_space = PPL_SPACE_SIZE;
log              1320 drivers/md/raid5-ppl.c 		log->use_multippl = false;
log              1321 drivers/md/raid5-ppl.c 		log->entry_space = (log->rdev->ppl.size << 9) -
log              1324 drivers/md/raid5-ppl.c 	log->next_io_sector = rdev->ppl.sector;
log              1328 drivers/md/raid5-ppl.c 		log->wb_cache_on = true;
log              1417 drivers/md/raid5-ppl.c 		struct ppl_log *log = &ppl_conf->child_logs[i];
log              1420 drivers/md/raid5-ppl.c 		mutex_init(&log->io_mutex);
log              1421 drivers/md/raid5-ppl.c 		spin_lock_init(&log->io_list_lock);
log              1422 drivers/md/raid5-ppl.c 		INIT_LIST_HEAD(&log->io_list);
log              1424 drivers/md/raid5-ppl.c 		log->ppl_conf = ppl_conf;
log              1425 drivers/md/raid5-ppl.c 		log->rdev = rdev;
log              1432 drivers/md/raid5-ppl.c 			ppl_init_child_log(log, rdev);
log              1468 drivers/md/raid5-ppl.c 	struct ppl_log *log;
log              1485 drivers/md/raid5-ppl.c 	log = &ppl_conf->child_logs[rdev->raid_disk];
log              1487 drivers/md/raid5-ppl.c 	mutex_lock(&log->io_mutex);
log              1491 drivers/md/raid5-ppl.c 			log->rdev = rdev;
log              1492 drivers/md/raid5-ppl.c 			ret = ppl_write_empty_header(log);
log              1493 drivers/md/raid5-ppl.c 			ppl_init_child_log(log, rdev);
log              1496 drivers/md/raid5-ppl.c 		log->rdev = NULL;
log              1498 drivers/md/raid5-ppl.c 	mutex_unlock(&log->io_mutex);
log               221 drivers/md/raid5.c 	if (r5c_is_writeback(conf->log))
log               233 drivers/md/raid5.c 	    (conf->quiesce && r5c_is_writeback(conf->log) &&
log               271 drivers/md/raid5.c 			if (!r5c_is_writeback(conf->log))
log               657 drivers/md/raid5.c 				r5l_wake_reclaim(conf->log, 0);
log              1659 drivers/md/raid5.c 	if (r5c_is_writeback(sh->raid_conf->log))
log              1765 drivers/md/raid5.c 							     r5c_is_writeback(conf->log));
log              1767 drivers/md/raid5.c 					    !r5c_is_writeback(conf->log)) {
log              4861 drivers/md/raid5.c 		if (!r5c_is_writeback(conf->log)) {
log              5330 drivers/md/raid5.c 	bool second_try = !r5c_is_writeback(conf->log) &&
log              6209 drivers/md/raid5.c 	r5l_flush_stripe_to_raid(conf->log);
log              6265 drivers/md/raid5.c 	r5l_flush_stripe_to_raid(conf->log);
log              6352 drivers/md/raid5.c 	r5l_flush_stripe_to_raid(conf->log);
log              7604 drivers/md/raid5.c 	if (test_bit(Journal, &rdev->flags) && conf->log) {
log              7690 drivers/md/raid5.c 		if (conf->log)
log              7702 drivers/md/raid5.c 		ret = r5l_start(conf->log);
log              8406 drivers/md/raid5.c 	return r5l_start(conf->log);
log               681 drivers/md/raid5.h 	struct r5l_log		*log;
log               125 drivers/media/dvb-core/dvb_math.c 	u64 log;
log               132 drivers/media/dvb-core/dvb_math.c 	log = intlog2(value);
log               139 drivers/media/dvb-core/dvb_math.c 	return (log * 646456993) >> 31;
log              1831 drivers/media/v4l2-core/v4l2-ctrls.c 	.log = std_log,
log              2833 drivers/media/v4l2-core/v4l2-ctrls.c 	ctrl->type_ops->log(ctrl);
log                87 drivers/mtd/nand/raw/nandsim.c static uint log            = CONFIG_NANDSIM_LOG;
log               120 drivers/mtd/nand/raw/nandsim.c module_param(log,            uint, 0400);
log               145 drivers/mtd/nand/raw/nandsim.c MODULE_PARM_DESC(log,            "Perform logging if not zero");
log               173 drivers/mtd/nand/raw/nandsim.c 	do { if (log) pr_debug(" log: " args); } while(0)
log                90 drivers/mtd/tests/nandbiterrs.c static int write_page(int log)
log                92 drivers/mtd/tests/nandbiterrs.c 	if (log)
log                99 drivers/mtd/tests/nandbiterrs.c static int rewrite_page(int log)
log               104 drivers/mtd/tests/nandbiterrs.c 	if (log)
log               128 drivers/mtd/tests/nandbiterrs.c static int read_page(int log)
log               134 drivers/mtd/tests/nandbiterrs.c 	if (log)
log               154 drivers/mtd/tests/nandbiterrs.c static int verify_page(int log)
log               158 drivers/mtd/tests/nandbiterrs.c 	if (log)
log              2834 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	struct mbox_cmd_log *log = NULL;
log              2842 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	log = padap->mbox_log;
log              2851 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		entry_idx = log->cursor + k;
log              2852 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		if (entry_idx >= log->size)
log              2853 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 			entry_idx -= log->size;
log              2855 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		entry = mbox_cmd_log_entry(log, entry_idx);
log               451 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h static inline struct mbox_cmd *mbox_cmd_log_entry(struct mbox_cmd_log *log,
log               454 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	return &((struct mbox_cmd *)&(log)[1])[entry_idx];
log               987 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct fw_devlog_e log[0];	/* Firmware Device Log */
log              1009 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		e = &dinfo->log[index];
log              1108 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   dparams->start, dparams->size, (__be32 *)dinfo->log,
log              1120 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		struct fw_devlog_e *e = &dinfo->log[index];
log              1156 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log              1168 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	entry_idx = log->cursor + ((uintptr_t)v - 2);
log              1169 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	if (entry_idx >= log->size)
log              1170 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		entry_idx -= log->size;
log              1171 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	entry = mbox_cmd_log_entry(log, entry_idx);
log              1194 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log              1196 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	return ((pos <= log->size) ? (void *)(uintptr_t)(pos + 1) : NULL);
log               241 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log               245 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry = mbox_cmd_log_entry(log, log->cursor++);
log               246 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	if (log->cursor == log->size)
log               247 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		log->cursor = 0;
log               254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->seqno = log->seqno++;
log              1963 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log              1975 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	entry_idx = log->cursor + ((uintptr_t)v - 2);
log              1976 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	if (entry_idx >= log->size)
log              1977 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		entry_idx -= log->size;
log              1978 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	entry = mbox_cmd_log_entry(log, entry_idx);
log              2001 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log              2003 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	return ((pos <= log->size) ? (void *)(uintptr_t)(pos + 1) : NULL);
log               294 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h static inline struct mbox_cmd *mbox_cmd_log_entry(struct mbox_cmd_log *log,
log               297 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h 	return &((struct mbox_cmd *)&(log)[1])[entry_idx];
log                90 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	struct mbox_cmd_log *log = adapter->mbox_log;
log                94 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry = mbox_cmd_log_entry(log, log->cursor++);
log                95 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	if (log->cursor == log->size)
log                96 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		log->cursor = 0;
log               103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->seqno = log->seqno++;
log              2468 drivers/net/ethernet/pensando/ionic/ionic_if.h 	struct ionic_log_event log;
log               492 drivers/net/wimax/i2400m/rx.c 	struct i2400m_roq_log *log;
log               598 drivers/net/wimax/i2400m/rx.c 	if (roq->log->in - roq->log->out == I2400M_ROQ_LOG_LENGTH)
log               599 drivers/net/wimax/i2400m/rx.c 		roq->log->out++;
log               600 drivers/net/wimax/i2400m/rx.c 	cnt_idx = roq->log->in++ % I2400M_ROQ_LOG_LENGTH;
log               601 drivers/net/wimax/i2400m/rx.c 	e = &roq->log->entry[cnt_idx];
log               623 drivers/net/wimax/i2400m/rx.c 	BUG_ON(roq->log->out > roq->log->in);
log               624 drivers/net/wimax/i2400m/rx.c 	for (cnt = roq->log->out; cnt < roq->log->in; cnt++) {
log               626 drivers/net/wimax/i2400m/rx.c 		e = &roq->log->entry[cnt_idx];
log               630 drivers/net/wimax/i2400m/rx.c 	roq->log->in = roq->log->out = 0;
log               934 drivers/net/wimax/i2400m/rx.c 	kfree(i2400m->rx_roq[0].log);
log              1361 drivers/net/wimax/i2400m/rx.c 		rd = kcalloc(I2400M_RO_CIN + 1, sizeof(*i2400m->rx_roq[0].log),
log              1370 drivers/net/wimax/i2400m/rx.c 			i2400m->rx_roq[itr].log = &rd[itr];
log               352 drivers/net/wireless/broadcom/b43/debugfs.c 	struct b43_txstatus_log *log = &dev->dfsentry->txstatlog;
log               357 drivers/net/wireless/broadcom/b43/debugfs.c 	if (log->end < 0) {
log               365 drivers/net/wireless/broadcom/b43/debugfs.c 	i = log->end + 1;
log               370 drivers/net/wireless/broadcom/b43/debugfs.c 		stat = &(log->log[i]);
log               385 drivers/net/wireless/broadcom/b43/debugfs.c 		if (i == log->end)
log               682 drivers/net/wireless/broadcom/b43/debugfs.c 	struct b43_txstatus_log *log;
log               692 drivers/net/wireless/broadcom/b43/debugfs.c 	log = &e->txstatlog;
log               693 drivers/net/wireless/broadcom/b43/debugfs.c 	log->log = kcalloc(B43_NR_LOGGED_TXSTATUS,
log               695 drivers/net/wireless/broadcom/b43/debugfs.c 	if (!log->log) {
log               700 drivers/net/wireless/broadcom/b43/debugfs.c 	log->end = -1;
log               764 drivers/net/wireless/broadcom/b43/debugfs.c 	kfree(e->txstatlog.log);
log               772 drivers/net/wireless/broadcom/b43/debugfs.c 	struct b43_txstatus_log *log;
log               778 drivers/net/wireless/broadcom/b43/debugfs.c 	log = &e->txstatlog;
log               779 drivers/net/wireless/broadcom/b43/debugfs.c 	i = log->end + 1;
log               782 drivers/net/wireless/broadcom/b43/debugfs.c 	log->end = i;
log               783 drivers/net/wireless/broadcom/b43/debugfs.c 	cur = &(log->log[i]);
log                30 drivers/net/wireless/broadcom/b43/debugfs.h 	struct b43_txstatus *log;
log               127 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	struct b43legacy_txstatus_log *log = &dev->dfsentry->txstatlog;
log               133 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	spin_lock_irqsave(&log->lock, flags);
log               134 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	if (log->end < 0) {
log               142 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	i = log->end + 1;
log               147 drivers/net/wireless/broadcom/b43legacy/debugfs.c 		stat = &(log->log[i]);
log               162 drivers/net/wireless/broadcom/b43legacy/debugfs.c 		if (i == log->end)
log               167 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	spin_unlock_irqrestore(&log->lock, flags);
log               371 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	struct b43legacy_txstatus_log *log;
log               381 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	log = &e->txstatlog;
log               382 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	log->log = kcalloc(B43legacy_NR_LOGGED_TXSTATUS,
log               384 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	if (!log->log) {
log               389 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	log->end = -1;
log               390 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	spin_lock_init(&log->lock);
log               436 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	kfree(e->txstatlog.log);
log               444 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	struct b43legacy_txstatus_log *log;
log               450 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	log = &e->txstatlog;
log               452 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	spin_lock(&log->lock);
log               453 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	i = log->end + 1;
log               456 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	log->end = i;
log               457 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	cur = &(log->log[i]);
log               459 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	spin_unlock(&log->lock);
log                25 drivers/net/wireless/broadcom/b43legacy/debugfs.h 	struct b43legacy_txstatus *log;
log               667 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			  error->log[i].time,
log               668 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			  error->log[i].data, error->log[i].event);
log              1219 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				  u32 log_len, struct ipw_event *log)
log              1226 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				  (u8 *) log, sizeof(*log) * log_len);
log              1239 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			sizeof(*error->log) * log_len, GFP_ATOMIC);
log              1251 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	error->log = (struct ipw_event *)(error->elem + elem_len);
log              1253 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_capture_event_log(priv, log_len, error->log);
log              1268 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct ipw_event *log;
log              1272 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	log_size = PAGE_SIZE / sizeof(*log) > log_len ?
log              1273 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			sizeof(*log) * log_len : PAGE_SIZE;
log              1274 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	log = kzalloc(log_size, GFP_KERNEL);
log              1275 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	if (!log) {
log              1279 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	log_len = log_size / sizeof(*log);
log              1280 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_capture_event_log(priv, log_len, log);
log              1286 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				log[i].time, log[i].event, log[i].data);
log              1288 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	kfree(log);
log              1322 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				priv->error->log[i].time,
log              1323 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				priv->error->log[i].event,
log              1324 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				priv->error->log[i].data);
log              1110 drivers/net/wireless/intel/ipw2x00/ipw2200.h 	struct ipw_event *log;
log              1520 drivers/net/wireless/marvell/mwifiex/main.h 			   struct mwifiex_ds_get_stats *log);
log              1200 drivers/net/wireless/marvell/mwifiex/sta_ioctl.c 		       struct mwifiex_ds_get_stats *log)
log              1203 drivers/net/wireless/marvell/mwifiex/sta_ioctl.c 				HostCmd_ACT_GEN_GET, 0, log, true);
log               207 drivers/nvdimm/btt.c 			struct log_group *log)
log               210 drivers/nvdimm/btt.c 			arena->logoff + (lane * LOG_GRP_SIZE), log,
log               271 drivers/nvdimm/btt.c static u32 log_seq(struct log_group *log, int log_idx)
log               273 drivers/nvdimm/btt.c 	return le32_to_cpu(log->ent[log_idx].seq);
log               285 drivers/nvdimm/btt.c static int btt_log_get_old(struct arena_info *a, struct log_group *log)
log               296 drivers/nvdimm/btt.c 	if (log_seq(log, idx0) == 0) {
log               297 drivers/nvdimm/btt.c 		log->ent[idx0].seq = cpu_to_le32(1);
log               301 drivers/nvdimm/btt.c 	if (log_seq(log, idx0) == log_seq(log, idx1))
log               303 drivers/nvdimm/btt.c 	if (log_seq(log, idx0) + log_seq(log, idx1) > 5)
log               306 drivers/nvdimm/btt.c 	if (log_seq(log, idx0) < log_seq(log, idx1)) {
log               307 drivers/nvdimm/btt.c 		if ((log_seq(log, idx1) - log_seq(log, idx0)) == 1)
log               312 drivers/nvdimm/btt.c 		if ((log_seq(log, idx0) - log_seq(log, idx1)) == 1)
log               332 drivers/nvdimm/btt.c 	struct log_group log;
log               334 drivers/nvdimm/btt.c 	ret = btt_log_group_read(arena, lane, &log);
log               338 drivers/nvdimm/btt.c 	old_ent = btt_log_get_old(arena, &log);
log               342 drivers/nvdimm/btt.c 				old_ent, lane, log.ent[arena->log_index[0]].seq,
log               343 drivers/nvdimm/btt.c 				log.ent[arena->log_index[1]].seq);
log               351 drivers/nvdimm/btt.c 		memcpy(ent, &log.ent[arena->log_index[ret_ent]], LOG_ENT_SIZE);
log               628 drivers/nvdimm/btt.c 	struct log_group log;
log               632 drivers/nvdimm/btt.c 		ret = btt_log_group_read(arena, i, &log);
log               638 drivers/nvdimm/btt.c 				if (ent_is_padding(&log.ent[j])) {
log               666 drivers/nvdimm/btt.c 					if (ent_is_padding(&log.ent[j]))
log               678 drivers/nvdimm/btt.c 					if (!ent_is_padding(&log.ent[j]))
log              1317 drivers/nvdimm/btt.c 	struct log_entry log;
log              1388 drivers/nvdimm/btt.c 		log.lba = cpu_to_le32(premap);
log              1389 drivers/nvdimm/btt.c 		log.old_map = cpu_to_le32(old_postmap);
log              1390 drivers/nvdimm/btt.c 		log.new_map = cpu_to_le32(new_postmap);
log              1391 drivers/nvdimm/btt.c 		log.seq = cpu_to_le32(arena->freelist[lane].seq);
log              1393 drivers/nvdimm/btt.c 		ret = btt_flog_write(arena, lane, sub, &log);
log              2700 drivers/nvme/host/core.c 		void *log, size_t size, u64 offset)
log              2714 drivers/nvme/host/core.c 	return nvme_submit_sync_cmd(ctrl->admin_q, &c, log, size);
log              3700 drivers/nvme/host/core.c 	__le32 *log;
log              3703 drivers/nvme/host/core.c 	log = kzalloc(log_size, GFP_KERNEL);
log              3704 drivers/nvme/host/core.c 	if (!log)
log              3713 drivers/nvme/host/core.c 	error = nvme_get_log(ctrl, NVME_NSID_ALL, NVME_LOG_CHANGED_NS, 0, log,
log              3719 drivers/nvme/host/core.c 	kfree(log);
log              3863 drivers/nvme/host/core.c 	struct nvme_fw_slot_info_log *log;
log              3865 drivers/nvme/host/core.c 	log = kmalloc(sizeof(*log), GFP_KERNEL);
log              3866 drivers/nvme/host/core.c 	if (!log)
log              3869 drivers/nvme/host/core.c 	if (nvme_get_log(ctrl, NVME_NSID_ALL, NVME_LOG_FW_SLOT, 0, log,
log              3870 drivers/nvme/host/core.c 			sizeof(*log), 0))
log              3872 drivers/nvme/host/core.c 	kfree(log);
log               508 drivers/nvme/host/nvme.h 		void *log, size_t size, u64 offset);
log               133 drivers/nvme/target/admin-cmd.c 	struct nvme_smart_log *log;
log               137 drivers/nvme/target/admin-cmd.c 	if (req->data_len != sizeof(*log))
log               140 drivers/nvme/target/admin-cmd.c 	log = kzalloc(sizeof(*log), GFP_KERNEL);
log               141 drivers/nvme/target/admin-cmd.c 	if (!log)
log               145 drivers/nvme/target/admin-cmd.c 		status = nvmet_get_smart_log_all(req, log);
log               147 drivers/nvme/target/admin-cmd.c 		status = nvmet_get_smart_log_nsid(req, log);
log               153 drivers/nvme/target/admin-cmd.c 			&log->num_err_log_entries);
log               156 drivers/nvme/target/admin-cmd.c 	status = nvmet_copy_to_sgl(req, 0, log, sizeof(*log));
log               158 drivers/nvme/target/admin-cmd.c 	kfree(log);
log               166 drivers/nvme/target/admin-cmd.c 	struct nvme_effects_log *log;
log               168 drivers/nvme/target/admin-cmd.c 	log = kzalloc(sizeof(*log), GFP_KERNEL);
log               169 drivers/nvme/target/admin-cmd.c 	if (!log)
log               172 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_get_log_page]	= cpu_to_le32(1 << 0);
log               173 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_identify]		= cpu_to_le32(1 << 0);
log               174 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_abort_cmd]		= cpu_to_le32(1 << 0);
log               175 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_set_features]	= cpu_to_le32(1 << 0);
log               176 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_get_features]	= cpu_to_le32(1 << 0);
log               177 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_async_event]	= cpu_to_le32(1 << 0);
log               178 drivers/nvme/target/admin-cmd.c 	log->acs[nvme_admin_keep_alive]		= cpu_to_le32(1 << 0);
log               180 drivers/nvme/target/admin-cmd.c 	log->iocs[nvme_cmd_read]		= cpu_to_le32(1 << 0);
log               181 drivers/nvme/target/admin-cmd.c 	log->iocs[nvme_cmd_write]		= cpu_to_le32(1 << 0);
log               182 drivers/nvme/target/admin-cmd.c 	log->iocs[nvme_cmd_flush]		= cpu_to_le32(1 << 0);
log               183 drivers/nvme/target/admin-cmd.c 	log->iocs[nvme_cmd_dsm]			= cpu_to_le32(1 << 0);
log               184 drivers/nvme/target/admin-cmd.c 	log->iocs[nvme_cmd_write_zeroes]	= cpu_to_le32(1 << 0);
log               186 drivers/nvme/target/admin-cmd.c 	status = nvmet_copy_to_sgl(req, 0, log, sizeof(*log));
log               188 drivers/nvme/target/admin-cmd.c 	kfree(log);
log               154 drivers/pci/pcie/dpc.c 	u32 status, mask, sev, syserr, exc, dw0, dw1, dw2, dw3, log, prefix;
log               193 drivers/pci/pcie/dpc.c 	pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_IMPSPEC_LOG, &log);
log               194 drivers/pci/pcie/dpc.c 	pci_err(pdev, "RP PIO ImpSpec Log %#010x\n", log);
log               917 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 	struct cygnus_mux_log *log;
log               928 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 			log = &pinctrl->mux_log[i * CYGNUS_NUM_MUX_PER_REG
log               930 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 			log->mux.offset = i * 4;
log               931 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 			log->mux.shift = j * 4;
log               932 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 			log->mux.alt = 0;
log               933 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c 			log->is_configured = false;
log               984 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	struct ns2_mux_log *log;
log               996 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	log = &pinctrl->mux_log[0];
log               997 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	log->mux.base = NS2_PIN_MUX_BASE0;
log               998 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	log->mux.offset = 0;
log               999 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	log->mux.shift = 31;
log              1000 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 	log->mux.alt = 0;
log              1008 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log = &pinctrl->mux_log[i];
log              1009 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.base = NS2_PIN_MUX_BASE0;
log              1010 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.offset = NS2_MUX_PAD_FUNC1_OFFSET;
log              1011 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.shift = 32 - (i * 2);
log              1012 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.alt = 0;
log              1020 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log = &pinctrl->mux_log[(NS2_NUM_IOMUX - NS2_NUM_PWM_MUX) + i];
log              1021 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.base = NS2_PIN_MUX_BASE1;
log              1022 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.offset = 0;
log              1023 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.shift = i;
log              1024 drivers/pinctrl/bcm/pinctrl-ns2-mux.c 		log->mux.alt =  0;
log               538 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 	struct nsp_mux_log *log;
log               549 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 		log = &pinctrl->mux_log[i];
log               550 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 		log->mux.base = nsp_pin_groups[i].mux.base;
log               551 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 		log->mux.shift = nsp_pin_groups[i].mux.shift;
log               552 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 		log->mux.alt = 0;
log               553 drivers/pinctrl/bcm/pinctrl-nsp-mux.c 		log->is_configured = false;
log               353 drivers/regulator/max77802-regulator.c #define regulator_77802_desc_p_ldo(num, supply, log)	{		\
log               359 drivers/regulator/max77802-regulator.c 	.ops		= &max77802_ldo_ops_logic##log,			\
log               374 drivers/regulator/max77802-regulator.c #define regulator_77802_desc_n_ldo(num, supply, log)   {		\
log               380 drivers/regulator/max77802-regulator.c 	.ops		= &max77802_ldo_ops_logic##log,			\
log               132 drivers/s390/scsi/zfcp_dbf.c 		goto log;
log               148 drivers/s390/scsi/zfcp_dbf.c log:
log               424 drivers/s390/scsi/zfcp_fsf.h 	u8 log[FSF_QTCB_LOG_SIZE];
log               931 drivers/scsi/esas2r/atvda.h 	struct atto_vda_ae_log log;
log                86 drivers/scsi/ibmvscsi/ibmvfc.c 	int log;
log              1474 drivers/scsi/ibmvscsi/ibmvfc.c 		logerr = cmd_status[index].log;
log               140 drivers/scsi/mesh.c 	struct dbglog log[N_DBG_LOG];
log               179 drivers/scsi/mesh.c 	struct dbglog log[N_DBG_SLOG];
log               218 drivers/scsi/mesh.c 	tlp = &tp->log[tp->log_ix];
log               219 drivers/scsi/mesh.c 	slp = &ms->log[ms->log_ix];
log               251 drivers/scsi/mesh.c 		lp = &tp->log[i];
log               276 drivers/scsi/mesh.c 		lp = &ms->log[i];
log               184 drivers/staging/comedi/drivers/usbduxfast.c 				u8 len, u8 op, u8 out, u8 log)
log               192 drivers/staging/comedi/drivers/usbduxfast.c 	devpriv->duxbuf[1 + 0x18 + index] = log;
log                70 drivers/staging/greybus/log.c 	struct gb_log *log;
log                80 drivers/staging/greybus/log.c 	log = kzalloc(sizeof(*log), GFP_KERNEL);
log                81 drivers/staging/greybus/log.c 	if (!log)
log                91 drivers/staging/greybus/log.c 	log->connection = connection;
log                92 drivers/staging/greybus/log.c 	greybus_set_drvdata(bundle, log);
log               103 drivers/staging/greybus/log.c 	kfree(log);
log               109 drivers/staging/greybus/log.c 	struct gb_log *log = greybus_get_drvdata(bundle);
log               110 drivers/staging/greybus/log.c 	struct gb_connection *connection = log->connection;
log               115 drivers/staging/greybus/log.c 	kfree(log);
log                43 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct proc_dir_entry *log;	/* log entry */
log               330 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->log = proc_create_data(pd->log_name,
log               352 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if (pd->log)
log               479 drivers/usb/atm/cxacru.c 			char log[CMD_MAX_CONFIG * 12 + 1]; /* %02x=%08x */
log               491 drivers/usb/atm/cxacru.c 				snprintf(log + tmp*12, 13, " %02x=%08x",
log               494 drivers/usb/atm/cxacru.c 			atm_info(instance->usbatm, "config%s\n", log);
log               504 drivers/usb/host/isp116x.h 	isp116x_show_regs(isp116x, log, NULL);
log              1248 drivers/usb/storage/ene_ub6250.c 	u16 log;
log              1254 drivers/usb/storage/ene_ub6250.c 	log = info->MS_Lib.Phy2LogMap[phyblk];
log              1256 drivers/usb/storage/ene_ub6250.c 	if (log < info->MS_Lib.NumberOfLogBlock)
log              1257 drivers/usb/storage/ene_ub6250.c 		info->MS_Lib.Log2PhyMap[log] = MS_LB_NOT_USED;
log              1314 drivers/usb/storage/ene_ub6250.c 	u16 log;
log              1320 drivers/usb/storage/ene_ub6250.c 	log = info->MS_Lib.Phy2LogMap[phyblk];
log              1322 drivers/usb/storage/ene_ub6250.c 	if (log < info->MS_Lib.NumberOfLogBlock)
log              1323 drivers/usb/storage/ene_ub6250.c 		info->MS_Lib.Log2PhyMap[log] = MS_LB_NOT_USED;
log              1030 drivers/vhost/net.c 		       struct vhost_log *log,
log              1051 drivers/vhost/net.c 				      &in, log, log_num);
log              1066 drivers/vhost/net.c 		if (unlikely(log)) {
log              1068 drivers/vhost/net.c 			log += *log_num;
log              1079 drivers/vhost/net.c 	if (unlikely(log))
log              1099 drivers/vhost/net.c 	unsigned uninitialized_var(in), log;
log              1138 drivers/vhost/net.c 		vq->log : NULL;
log              1149 drivers/vhost/net.c 					vhost_len, &in, vq_log, &log,
log              1227 drivers/vhost/net.c 			vhost_log_write(vq, vq_log, log, vhost_len,
log               374 drivers/vhost/vhost.c 	kfree(vq->log);
log               375 drivers/vhost/vhost.c 	vq->log = NULL;
log               391 drivers/vhost/vhost.c 		vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log),
log               395 drivers/vhost/vhost.c 		if (!vq->indirect || !vq->log || !vq->heads)
log               482 drivers/vhost/vhost.c 		vq->log = NULL;
log               759 drivers/vhost/vhost.c 		bool log;
log               762 drivers/vhost/vhost.c 		log = log_all || vhost_has_feature(d->vqs[i], VHOST_F_LOG_ALL);
log               766 drivers/vhost/vhost.c 						 umem, log);
log              1826 drivers/vhost/vhost.c 	unsigned long log = (unsigned long)addr;
log              1829 drivers/vhost/vhost.c 	int bit = nr + (log % PAGE_SIZE) * 8;
log              1832 drivers/vhost/vhost.c 	r = get_user_pages_fast(log, 1, FOLL_WRITE, &page);
log              1855 drivers/vhost/vhost.c 		u64 log = base + write_page / 8;
log              1857 drivers/vhost/vhost.c 		if ((u64)(unsigned long)log != log)
log              1859 drivers/vhost/vhost.c 		r = set_bit_to_user(bit, (void __user *)(unsigned long)log);
log              1933 drivers/vhost/vhost.c int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log,
log              1952 drivers/vhost/vhost.c 		u64 l = min(log[i].len, len);
log              1953 drivers/vhost/vhost.c 		r = log_write(vq->log_base, log[i].addr, l);
log              2105 drivers/vhost/vhost.c 			struct vhost_log *log, unsigned int *log_num,
log              2181 drivers/vhost/vhost.c 			if (unlikely(log && ret)) {
log              2182 drivers/vhost/vhost.c 				log[*log_num].addr = vhost64_to_cpu(vq, desc.addr);
log              2183 drivers/vhost/vhost.c 				log[*log_num].len = vhost32_to_cpu(vq, desc.len);
log              2211 drivers/vhost/vhost.c 		      struct vhost_log *log, unsigned int *log_num)
log              2269 drivers/vhost/vhost.c 	if (unlikely(log))
log              2295 drivers/vhost/vhost.c 					   log, log_num, &desc);
log              2322 drivers/vhost/vhost.c 			if (unlikely(log && ret)) {
log              2323 drivers/vhost/vhost.c 				log[*log_num].addr = vhost64_to_cpu(vq, desc.addr);
log              2324 drivers/vhost/vhost.c 				log[*log_num].len = vhost32_to_cpu(vq, desc.len);
log               138 drivers/vhost/vhost.h 	struct vhost_log *log;
log               196 drivers/vhost/vhost.h 		      struct vhost_log *log, unsigned int *log_num);
log               212 drivers/vhost/vhost.h int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log,
log              1404 drivers/virt/vboxguest/vboxguest_core.c static int vbg_ioctl_log(struct vbg_ioctl_log *log)
log              1406 drivers/virt/vboxguest/vboxguest_core.c 	if (log->hdr.size_out != sizeof(log->hdr))
log              1409 drivers/virt/vboxguest/vboxguest_core.c 	vbg_info("%.*s", (int)(log->hdr.size_in - sizeof(log->hdr)),
log              1410 drivers/virt/vboxguest/vboxguest_core.c 		 log->u.in.msg);
log               105 fs/btrfs/tree-log.c 				       struct btrfs_root *log,
log               289 fs/btrfs/tree-log.c 	int (*process_func)(struct btrfs_root *log, struct extent_buffer *eb,
log               296 fs/btrfs/tree-log.c static int process_one_buffer(struct btrfs_root *log,
log               300 fs/btrfs/tree-log.c 	struct btrfs_fs_info *fs_info = log->fs_info;
log               943 fs/btrfs/tree-log.c static noinline int backref_in_log(struct btrfs_root *log,
log               962 fs/btrfs/tree-log.c 	ret = btrfs_search_slot(NULL, log, key, path, 0, 0);
log              1418 fs/btrfs/tree-log.c 				  struct btrfs_root *log,
log              1505 fs/btrfs/tree-log.c 				ret = __add_inode_ref(trans, root, path, log,
log              2217 fs/btrfs/tree-log.c 				      struct btrfs_root *log,
log              2253 fs/btrfs/tree-log.c 		if (log && dir_key->type == BTRFS_DIR_ITEM_KEY) {
log              2254 fs/btrfs/tree-log.c 			log_di = btrfs_lookup_dir_item(trans, log, log_path,
log              2257 fs/btrfs/tree-log.c 		} else if (log && dir_key->type == BTRFS_DIR_INDEX_KEY) {
log              2258 fs/btrfs/tree-log.c 			log_di = btrfs_lookup_dir_index_item(trans, log,
log              2320 fs/btrfs/tree-log.c 			      struct btrfs_root *log,
log              2373 fs/btrfs/tree-log.c 			log_di = btrfs_lookup_xattr(NULL, log, log_path, ino,
log              2428 fs/btrfs/tree-log.c 				       struct btrfs_root *log,
log              2463 fs/btrfs/tree-log.c 			ret = find_dir_range(log, path, dirid, key_type,
log              2494 fs/btrfs/tree-log.c 			ret = check_item_in_log(trans, root, log, path,
log              2535 fs/btrfs/tree-log.c static int replay_one_buffer(struct btrfs_root *log, struct extent_buffer *eb,
log              2584 fs/btrfs/tree-log.c 			ret = replay_xattr_deletes(wc->trans, root, log,
log              2591 fs/btrfs/tree-log.c 					 root, log, path, key.objectid, 0);
log              2659 fs/btrfs/tree-log.c 			ret = add_inode_ref(wc->trans, root, log, path,
log              2857 fs/btrfs/tree-log.c 			 struct btrfs_root *log, struct walk_control *wc)
log              2859 fs/btrfs/tree-log.c 	struct btrfs_fs_info *fs_info = log->fs_info;
log              2870 fs/btrfs/tree-log.c 	level = btrfs_header_level(log->node);
log              2872 fs/btrfs/tree-log.c 	path->nodes[level] = log->node;
log              2873 fs/btrfs/tree-log.c 	extent_buffer_get(log->node);
log              2877 fs/btrfs/tree-log.c 		wret = walk_down_log_tree(trans, log, path, &level, wc);
log              2885 fs/btrfs/tree-log.c 		wret = walk_up_log_tree(trans, log, path, &level, wc);
log              2896 fs/btrfs/tree-log.c 		ret = wc->process_func(log, path->nodes[orig_level], wc,
log              2917 fs/btrfs/tree-log.c 			WARN_ON(log->root_key.objectid !=
log              2936 fs/btrfs/tree-log.c 			   struct btrfs_root *log,
log              2939 fs/btrfs/tree-log.c 	struct btrfs_fs_info *fs_info = log->fs_info;
log              2942 fs/btrfs/tree-log.c 	if (log->log_transid == 1) {
log              2945 fs/btrfs/tree-log.c 				&log->root_key, root_item);
log              2948 fs/btrfs/tree-log.c 				&log->root_key, root_item);
log              3044 fs/btrfs/tree-log.c 	struct btrfs_root *log = root->log_root;
log              3101 fs/btrfs/tree-log.c 	ret = btrfs_write_marked_extents(fs_info, &log->dirty_log_pages, mark);
log              3123 fs/btrfs/tree-log.c 	btrfs_set_root_node(&log->root_item, log->node);
log              3124 fs/btrfs/tree-log.c 	memcpy(&new_root_item, &log->root_item, sizeof(new_root_item));
log              3127 fs/btrfs/tree-log.c 	log->log_transid = root->log_transid;
log              3155 fs/btrfs/tree-log.c 	ret = update_log_root(trans, log, &new_root_item);
log              3174 fs/btrfs/tree-log.c 		btrfs_wait_tree_log_extents(log, mark);
log              3191 fs/btrfs/tree-log.c 		ret = btrfs_wait_tree_log_extents(log, mark);
log              3215 fs/btrfs/tree-log.c 		btrfs_wait_tree_log_extents(log, mark);
log              3231 fs/btrfs/tree-log.c 	ret = btrfs_wait_tree_log_extents(log, mark);
log              3299 fs/btrfs/tree-log.c 			  struct btrfs_root *log)
log              3307 fs/btrfs/tree-log.c 	ret = walk_log_tree(trans, log, &wc);
log              3312 fs/btrfs/tree-log.c 			btrfs_handle_fs_error(log->fs_info, ret, NULL);
log              3315 fs/btrfs/tree-log.c 	clear_extent_bits(&log->dirty_log_pages, 0, (u64)-1,
log              3317 fs/btrfs/tree-log.c 	free_extent_buffer(log->node);
log              3318 fs/btrfs/tree-log.c 	kfree(log);
log              3394 fs/btrfs/tree-log.c 	struct btrfs_root *log;
log              3411 fs/btrfs/tree-log.c 	log = root->log_root;
log              3418 fs/btrfs/tree-log.c 	di = btrfs_lookup_dir_item(trans, log, path, dir_ino,
log              3425 fs/btrfs/tree-log.c 		ret = btrfs_delete_one_dir_name(trans, log, path, di);
log              3433 fs/btrfs/tree-log.c 	di = btrfs_lookup_dir_index_item(trans, log, path, dir_ino,
log              3440 fs/btrfs/tree-log.c 		ret = btrfs_delete_one_dir_name(trans, log, path, di);
log              3459 fs/btrfs/tree-log.c 		ret = btrfs_search_slot(trans, log, &key, path, 0, 1);
log              3502 fs/btrfs/tree-log.c 	struct btrfs_root *log;
log              3512 fs/btrfs/tree-log.c 	log = root->log_root;
log              3515 fs/btrfs/tree-log.c 	ret = btrfs_del_inode_ref(trans, log, name, name_len, btrfs_ino(inode),
log              3534 fs/btrfs/tree-log.c 				       struct btrfs_root *log,
log              3549 fs/btrfs/tree-log.c 	ret = btrfs_insert_empty_item(trans, log, path, &key, sizeof(*item));
log              3574 fs/btrfs/tree-log.c 	struct btrfs_root *log = root->log_root;
log              3584 fs/btrfs/tree-log.c 	log = root->log_root;
log              3630 fs/btrfs/tree-log.c 			ret = overwrite_item(trans, log, dst_path,
log              3668 fs/btrfs/tree-log.c 			ret = overwrite_item(trans, log, dst_path, src, i,
log              3726 fs/btrfs/tree-log.c 			ret = overwrite_item(trans, log, dst_path,
log              3746 fs/btrfs/tree-log.c 		ret = insert_dir_log_key(trans, log, path, key_type,
log              3804 fs/btrfs/tree-log.c 				  struct btrfs_root *log,
log              3818 fs/btrfs/tree-log.c 		ret = btrfs_search_slot(trans, log, &key, path, -1, 1);
log              3840 fs/btrfs/tree-log.c 		ret = btrfs_del_items(trans, log, path, start_slot,
log              3913 fs/btrfs/tree-log.c 			  struct btrfs_root *log, struct btrfs_path *path,
log              3919 fs/btrfs/tree-log.c 	ret = btrfs_insert_empty_item(trans, log, path,
log              3963 fs/btrfs/tree-log.c 	struct btrfs_root *log = inode->root->log_root;
log              3989 fs/btrfs/tree-log.c 	ret = btrfs_insert_empty_items(trans, log, dst_path,
log              4075 fs/btrfs/tree-log.c 			ret = log_csums(trans, log, sums);
log              4148 fs/btrfs/tree-log.c 	struct btrfs_root *log = root->log_root;
log              4158 fs/btrfs/tree-log.c 	ret = log_extent_csums(trans, inode, log, em);
log              4162 fs/btrfs/tree-log.c 	ret = __btrfs_drop_extents(trans, log, &inode->vfs_inode, path, em->start,
log              4173 fs/btrfs/tree-log.c 		ret = btrfs_insert_empty_item(trans, log, path, &key,
log              4467 fs/btrfs/tree-log.c static int logged_inode_size(struct btrfs_root *log, struct btrfs_inode *inode,
log              4477 fs/btrfs/tree-log.c 	ret = btrfs_search_slot(NULL, log, &key, path, 0, 0);
log              5036 fs/btrfs/tree-log.c 	struct btrfs_root *log = root->log_root;
log              5114 fs/btrfs/tree-log.c 		ret = drop_objectid_items(trans, log, path, ino, max_key_type);
log              5130 fs/btrfs/tree-log.c 			err = logged_inode_size(log, inode, path, &logged_isize);
log              5138 fs/btrfs/tree-log.c 				ret = drop_objectid_items(trans, log, path, ino,
log              5147 fs/btrfs/tree-log.c 						log, &inode->vfs_inode, 0, 0);
log              5158 fs/btrfs/tree-log.c 			ret = drop_objectid_items(trans, log, path, ino,
log              5323 fs/btrfs/tree-log.c 		err = log_inode_item(trans, log, dst_path, inode);
log              5550 fs/btrfs/tree-log.c 	struct btrfs_root *log = root->log_root;
log              5584 fs/btrfs/tree-log.c 		ret = btrfs_search_forward(log, &min_key, path, trans->transid);
log              5653 fs/btrfs/tree-log.c 			ret = btrfs_next_leaf(log, path);
log              6165 fs/btrfs/tree-log.c 	struct btrfs_root *log;
log              6218 fs/btrfs/tree-log.c 		log = btrfs_read_fs_root(log_root_tree, &found_key);
log              6219 fs/btrfs/tree-log.c 		if (IS_ERR(log)) {
log              6220 fs/btrfs/tree-log.c 			ret = PTR_ERR(log);
log              6247 fs/btrfs/tree-log.c 							log->node->start,
log              6248 fs/btrfs/tree-log.c 							log->node->len);
log              6249 fs/btrfs/tree-log.c 			free_extent_buffer(log->node);
log              6250 fs/btrfs/tree-log.c 			free_extent_buffer(log->commit_root);
log              6251 fs/btrfs/tree-log.c 			kfree(log);
log              6260 fs/btrfs/tree-log.c 		wc.replay_dest->log_root = log;
log              6262 fs/btrfs/tree-log.c 		ret = walk_log_tree(trans, log, &wc);
log              6287 fs/btrfs/tree-log.c 		free_extent_buffer(log->node);
log              6288 fs/btrfs/tree-log.c 		free_extent_buffer(log->commit_root);
log              6289 fs/btrfs/tree-log.c 		kfree(log);
log               364 fs/fs_context.c 	if (fc->log)
log               365 fs/fs_context.c 		refcount_inc(&fc->log->usage);
log               391 fs/fs_context.c 	struct fc_log *log = fc ? fc->log : NULL;
log               418 fs/fs_context.c 	if (log && within_module_core((unsigned long)p, log->owner))
log               429 fs/fs_context.c 	if (!log) {
log               444 fs/fs_context.c 		unsigned int logsize = ARRAY_SIZE(log->buffer);
log               447 fs/fs_context.c 		index = log->head & (logsize - 1);
log               448 fs/fs_context.c 		BUILD_BUG_ON(sizeof(log->head) != sizeof(u8) ||
log               449 fs/fs_context.c 			     sizeof(log->tail) != sizeof(u8));
log               450 fs/fs_context.c 		if ((u8)(log->head - log->tail) == logsize) {
log               452 fs/fs_context.c 			if (log->need_free & (1 << index))
log               453 fs/fs_context.c 				kfree(log->buffer[index]);
log               454 fs/fs_context.c 			log->tail++;
log               457 fs/fs_context.c 		log->buffer[index] = q;
log               458 fs/fs_context.c 		log->need_free &= ~(1 << index);
log               459 fs/fs_context.c 		log->need_free |= freeable << index;
log               460 fs/fs_context.c 		log->head++;
log               471 fs/fs_context.c 	struct fc_log *log = fc->log;
log               474 fs/fs_context.c 	if (log) {
log               475 fs/fs_context.c 		if (refcount_dec_and_test(&log->usage)) {
log               476 fs/fs_context.c 			fc->log = NULL;
log               478 fs/fs_context.c 				if (log->need_free & (1 << i))
log               479 fs/fs_context.c 					kfree(log->buffer[i]);
log               480 fs/fs_context.c 			kfree(log);
log                28 fs/fsopen.c    	struct fc_log *log = fc->log;
log                29 fs/fsopen.c    	unsigned int logsize = ARRAY_SIZE(log->buffer);
log                39 fs/fsopen.c    	if (log->head == log->tail) {
log                44 fs/fsopen.c    	index = log->tail & (logsize - 1);
log                45 fs/fsopen.c    	p = log->buffer[index];
log                46 fs/fsopen.c    	need_free = log->need_free & (1 << index);
log                47 fs/fsopen.c    	log->buffer[index] = NULL;
log                48 fs/fsopen.c    	log->need_free &= ~(1 << index);
log                49 fs/fsopen.c    	log->tail++;
log               100 fs/fsopen.c    	fc->log = kzalloc(sizeof(*fc->log), GFP_KERNEL);
log               101 fs/fsopen.c    	if (!fc->log)
log               103 fs/fsopen.c    	refcount_set(&fc->log->usage, 1);
log               104 fs/fsopen.c    	fc->log->owner = fc->fs_type->owner;
log                32 fs/jfs/file.c  		jfs_flush_journal(JFS_SBI(inode->i_sb)->log, 1);
log               128 fs/jfs/inode.c 		jfs_flush_journal(JFS_SBI(inode->i_sb)->log, wait);
log               421 fs/jfs/jfs_dmap.c 	struct jfs_log *log;
log               436 fs/jfs/jfs_dmap.c 	log = (struct jfs_log *) JFS_SBI(tblk->sb)->log;
log               437 fs/jfs/jfs_dmap.c 	logdiff(difft, lsn, log);
log               528 fs/jfs/jfs_dmap.c 		LOGSYNC_LOCK(log, flags);
log               531 fs/jfs/jfs_dmap.c 			logdiff(diffp, mp->lsn, log);
log               540 fs/jfs/jfs_dmap.c 			logdiff(difft, tblk->clsn, log);
log               541 fs/jfs/jfs_dmap.c 			logdiff(diffp, mp->clsn, log);
log               545 fs/jfs/jfs_dmap.c 			mp->log = log;
log               549 fs/jfs/jfs_dmap.c 			log->count++;
log               554 fs/jfs/jfs_dmap.c 		LOGSYNC_UNLOCK(log, flags);
log              2726 fs/jfs/jfs_imap.c 	struct jfs_log *log;
log              2801 fs/jfs/jfs_imap.c 	log = JFS_SBI(tblk->sb)->log;
log              2802 fs/jfs/jfs_imap.c 	LOGSYNC_LOCK(log, flags);
log              2805 fs/jfs/jfs_imap.c 		logdiff(difft, lsn, log);
log              2806 fs/jfs/jfs_imap.c 		logdiff(diffp, mp->lsn, log);
log              2814 fs/jfs/jfs_imap.c 		logdiff(difft, tblk->clsn, log);
log              2815 fs/jfs/jfs_imap.c 		logdiff(diffp, mp->clsn, log);
log              2819 fs/jfs/jfs_imap.c 		mp->log = log;
log              2822 fs/jfs/jfs_imap.c 		log->count++;
log              2826 fs/jfs/jfs_imap.c 	LOGSYNC_UNLOCK(log, flags);
log               158 fs/jfs/jfs_incore.h 	struct jfs_log	*log;		/* log			*/
log               214 fs/jfs/jfs_incore.h 	if (JFS_SBI(inode->i_sb)->log)
log                80 fs/jfs/jfs_logmgr.c #define LOG_LOCK_INIT(log)	mutex_init(&(log)->loglock)
log                81 fs/jfs/jfs_logmgr.c #define LOG_LOCK(log)		mutex_lock(&((log)->loglock))
log                82 fs/jfs/jfs_logmgr.c #define LOG_UNLOCK(log)		mutex_unlock(&((log)->loglock))
log                89 fs/jfs/jfs_logmgr.c #define LOGGC_LOCK_INIT(log)	spin_lock_init(&(log)->gclock)
log                90 fs/jfs/jfs_logmgr.c #define LOGGC_LOCK(log)		spin_lock_irq(&(log)->gclock)
log                91 fs/jfs/jfs_logmgr.c #define LOGGC_UNLOCK(log)	spin_unlock_irq(&(log)->gclock)
log               163 fs/jfs/jfs_logmgr.c static int lmWriteRecord(struct jfs_log * log, struct tblock * tblk,
log               166 fs/jfs/jfs_logmgr.c static int lmNextPage(struct jfs_log * log);
log               167 fs/jfs/jfs_logmgr.c static int lmLogFileSystem(struct jfs_log * log, struct jfs_sb_info *sbi,
log               172 fs/jfs/jfs_logmgr.c static int lbmLogInit(struct jfs_log * log);
log               173 fs/jfs/jfs_logmgr.c static void lbmLogShutdown(struct jfs_log * log);
log               174 fs/jfs/jfs_logmgr.c static struct lbuf *lbmAllocate(struct jfs_log * log, int);
log               177 fs/jfs/jfs_logmgr.c static int lbmRead(struct jfs_log * log, int pn, struct lbuf ** bpp);
log               178 fs/jfs/jfs_logmgr.c static void lbmWrite(struct jfs_log * log, struct lbuf * bp, int flag, int cant_block);
log               179 fs/jfs/jfs_logmgr.c static void lbmDirectWrite(struct jfs_log * log, struct lbuf * bp, int flag);
log               183 fs/jfs/jfs_logmgr.c static void lmGCwrite(struct jfs_log * log, int cant_block);
log               184 fs/jfs/jfs_logmgr.c static int lmLogSync(struct jfs_log * log, int hard_sync);
log               201 fs/jfs/jfs_logmgr.c static void write_special_inodes(struct jfs_log *log,
log               206 fs/jfs/jfs_logmgr.c 	list_for_each_entry(sbi, &log->sb_list, log_list) {
log               225 fs/jfs/jfs_logmgr.c int lmLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               234 fs/jfs/jfs_logmgr.c 		 log, tblk, lrd, tlck);
log               236 fs/jfs/jfs_logmgr.c 	LOG_LOCK(log);
log               250 fs/jfs/jfs_logmgr.c 	lsn = log->lsn;
log               252 fs/jfs/jfs_logmgr.c 	LOGSYNC_LOCK(log, flags);
log               258 fs/jfs/jfs_logmgr.c 		mp->log = log;
log               260 fs/jfs/jfs_logmgr.c 		log->count++;
log               263 fs/jfs/jfs_logmgr.c 		list_add_tail(&mp->synclist, &log->synclist);
log               287 fs/jfs/jfs_logmgr.c 		log->count++;
log               297 fs/jfs/jfs_logmgr.c 		logdiff(diffp, mp->lsn, log);
log               298 fs/jfs/jfs_logmgr.c 		logdiff(difft, tblk->lsn, log);
log               308 fs/jfs/jfs_logmgr.c 	LOGSYNC_UNLOCK(log, flags);
log               314 fs/jfs/jfs_logmgr.c 	lsn = lmWriteRecord(log, tblk, lrd, tlck);
log               319 fs/jfs/jfs_logmgr.c 	logdiff(diffp, lsn, log);
log               320 fs/jfs/jfs_logmgr.c 	if (diffp >= log->nextsync)
log               321 fs/jfs/jfs_logmgr.c 		lsn = lmLogSync(log, 0);
log               324 fs/jfs/jfs_logmgr.c 	log->lsn = lsn;
log               326 fs/jfs/jfs_logmgr.c 	LOG_UNLOCK(log);
log               344 fs/jfs/jfs_logmgr.c lmWriteRecord(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               367 fs/jfs/jfs_logmgr.c 	bp = (struct lbuf *) log->bp;
log               369 fs/jfs/jfs_logmgr.c 	dstoffset = log->eor;
log               416 fs/jfs/jfs_logmgr.c 			lmNextPage(log);
log               418 fs/jfs/jfs_logmgr.c 			bp = log->bp;
log               441 fs/jfs/jfs_logmgr.c 			lmNextPage(log);
log               443 fs/jfs/jfs_logmgr.c 			bp = (struct lbuf *) log->bp;
log               495 fs/jfs/jfs_logmgr.c 		log->eor = dstoffset;
log               497 fs/jfs/jfs_logmgr.c 		lsn = (log->page << L2LOGPSIZE) + dstoffset;
log               514 fs/jfs/jfs_logmgr.c 			LOGGC_LOCK(log);
log               518 fs/jfs/jfs_logmgr.c 			tblk->bp = log->bp;
log               519 fs/jfs/jfs_logmgr.c 			tblk->pn = log->page;
log               520 fs/jfs/jfs_logmgr.c 			tblk->eor = log->eor;
log               523 fs/jfs/jfs_logmgr.c 			list_add_tail(&tblk->cqueue, &log->cqueue);
log               525 fs/jfs/jfs_logmgr.c 			LOGGC_UNLOCK(log);
log               529 fs/jfs/jfs_logmgr.c 			le16_to_cpu(lrd->type), log->bp, log->page, dstoffset);
log               537 fs/jfs/jfs_logmgr.c 		lmNextPage(log);
log               539 fs/jfs/jfs_logmgr.c 		bp = (struct lbuf *) log->bp;
log               560 fs/jfs/jfs_logmgr.c static int lmNextPage(struct jfs_log * log)
log               570 fs/jfs/jfs_logmgr.c 	pn = log->page;
log               571 fs/jfs/jfs_logmgr.c 	bp = log->bp;
log               575 fs/jfs/jfs_logmgr.c 	LOGGC_LOCK(log);
log               581 fs/jfs/jfs_logmgr.c 	if (list_empty(&log->cqueue))
log               584 fs/jfs/jfs_logmgr.c 		tblk = list_entry(log->cqueue.prev, struct tblock, cqueue);
log               600 fs/jfs/jfs_logmgr.c 		if (log->cflag & logGC_PAGEOUT) {
log               608 fs/jfs/jfs_logmgr.c 				lbmWrite(log, bp, 0, 0);
log               613 fs/jfs/jfs_logmgr.c 			log->cflag |= logGC_PAGEOUT;
log               614 fs/jfs/jfs_logmgr.c 			lmGCwrite(log, 0);
log               624 fs/jfs/jfs_logmgr.c 		lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmFREE, 0);
log               626 fs/jfs/jfs_logmgr.c 	LOGGC_UNLOCK(log);
log               634 fs/jfs/jfs_logmgr.c 	log->page = (pn == log->size - 1) ? 2 : pn + 1;
log               635 fs/jfs/jfs_logmgr.c 	log->eor = LOGPHDRSIZE;	/* ? valid page empty/full at logRedo() */
log               638 fs/jfs/jfs_logmgr.c 	nextbp = lbmAllocate(log, log->page);
log               639 fs/jfs/jfs_logmgr.c 	nextbp->l_eor = log->eor;
log               640 fs/jfs/jfs_logmgr.c 	log->bp = nextbp;
log               666 fs/jfs/jfs_logmgr.c int lmGroupCommit(struct jfs_log * log, struct tblock * tblk)
log               670 fs/jfs/jfs_logmgr.c 	LOGGC_LOCK(log);
log               677 fs/jfs/jfs_logmgr.c 		LOGGC_UNLOCK(log);
log               680 fs/jfs/jfs_logmgr.c 	jfs_info("lmGroup Commit: tblk = 0x%p, gcrtc = %d", tblk, log->gcrtc);
log               685 fs/jfs/jfs_logmgr.c 	if ((!(log->cflag & logGC_PAGEOUT)) && (!list_empty(&log->cqueue)) &&
log               686 fs/jfs/jfs_logmgr.c 	    (!(tblk->xflag & COMMIT_LAZY) || test_bit(log_FLUSH, &log->flag)
log               693 fs/jfs/jfs_logmgr.c 		log->cflag |= logGC_PAGEOUT;
log               695 fs/jfs/jfs_logmgr.c 		lmGCwrite(log, 0);
log               702 fs/jfs/jfs_logmgr.c 		LOGGC_UNLOCK(log);
log               712 fs/jfs/jfs_logmgr.c 		LOGGC_UNLOCK(log);
log               718 fs/jfs/jfs_logmgr.c 	log->gcrtc++;
log               722 fs/jfs/jfs_logmgr.c 		     LOGGC_LOCK(log), LOGGC_UNLOCK(log));
log               728 fs/jfs/jfs_logmgr.c 	LOGGC_UNLOCK(log);
log               745 fs/jfs/jfs_logmgr.c static void lmGCwrite(struct jfs_log * log, int cant_write)
log               760 fs/jfs/jfs_logmgr.c 	gcpn = list_entry(log->cqueue.next, struct tblock, cqueue)->pn;
log               762 fs/jfs/jfs_logmgr.c 	list_for_each_entry(tblk, &log->cqueue, cqueue) {
log               785 fs/jfs/jfs_logmgr.c 		lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmGC,
log               793 fs/jfs/jfs_logmgr.c 		lbmWrite(log, bp, lbmWRITE | lbmGC, cant_write);
log               813 fs/jfs/jfs_logmgr.c 	struct jfs_log *log = bp->l_log;
log               818 fs/jfs/jfs_logmgr.c 	spin_lock_irqsave(&log->gclock, flags);
log               825 fs/jfs/jfs_logmgr.c 	list_for_each_entry_safe(tblk, temp, &log->cqueue, cqueue) {
log               840 fs/jfs/jfs_logmgr.c 		if (tblk == log->flush_tblk) {
log               842 fs/jfs/jfs_logmgr.c 			clear_bit(log_FLUSH, &log->flag);
log               843 fs/jfs/jfs_logmgr.c 			log->flush_tblk = NULL;
log               859 fs/jfs/jfs_logmgr.c 				log->gcrtc--;
log               878 fs/jfs/jfs_logmgr.c 			lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmFREE,
log               891 fs/jfs/jfs_logmgr.c 	if ((!list_empty(&log->cqueue)) &&
log               892 fs/jfs/jfs_logmgr.c 	    ((log->gcrtc > 0) || (tblk->bp->l_wqnext != NULL) ||
log               893 fs/jfs/jfs_logmgr.c 	     test_bit(log_FLUSH, &log->flag) || jfs_tlocks_low))
log               897 fs/jfs/jfs_logmgr.c 		lmGCwrite(log, 1);
log               905 fs/jfs/jfs_logmgr.c 		log->cflag &= ~logGC_PAGEOUT;
log               908 fs/jfs/jfs_logmgr.c 	spin_unlock_irqrestore(&log->gclock, flags);
log               929 fs/jfs/jfs_logmgr.c static int lmLogSync(struct jfs_log * log, int hard_sync)
log               943 fs/jfs/jfs_logmgr.c 		write_special_inodes(log, filemap_fdatawrite);
log               945 fs/jfs/jfs_logmgr.c 		write_special_inodes(log, filemap_flush);
log               954 fs/jfs/jfs_logmgr.c 	if (log->sync == log->syncpt) {
log               955 fs/jfs/jfs_logmgr.c 		LOGSYNC_LOCK(log, flags);
log               956 fs/jfs/jfs_logmgr.c 		if (list_empty(&log->synclist))
log               957 fs/jfs/jfs_logmgr.c 			log->sync = log->lsn;
log               959 fs/jfs/jfs_logmgr.c 			lp = list_entry(log->synclist.next,
log               961 fs/jfs/jfs_logmgr.c 			log->sync = lp->lsn;
log               963 fs/jfs/jfs_logmgr.c 		LOGSYNC_UNLOCK(log, flags);
log               971 fs/jfs/jfs_logmgr.c 	if (log->sync != log->syncpt) {
log               976 fs/jfs/jfs_logmgr.c 		lrd.log.syncpt.sync = cpu_to_le32(log->sync);
log               977 fs/jfs/jfs_logmgr.c 		lsn = lmWriteRecord(log, NULL, &lrd, NULL);
log               979 fs/jfs/jfs_logmgr.c 		log->syncpt = log->sync;
log               981 fs/jfs/jfs_logmgr.c 		lsn = log->lsn;
log               986 fs/jfs/jfs_logmgr.c 	logsize = log->logsize;
log               988 fs/jfs/jfs_logmgr.c 	logdiff(written, lsn, log);
log              1013 fs/jfs/jfs_logmgr.c 		log->syncpt = log->sync = lsn;
log              1014 fs/jfs/jfs_logmgr.c 		log->nextsync = delta;
log              1017 fs/jfs/jfs_logmgr.c 		log->nextsync = written + more;
log              1024 fs/jfs/jfs_logmgr.c 	if (!test_bit(log_SYNCBARRIER, &log->flag) &&
log              1025 fs/jfs/jfs_logmgr.c 	    (written > LOGSYNC_BARRIER(logsize)) && log->active) {
log              1026 fs/jfs/jfs_logmgr.c 		set_bit(log_SYNCBARRIER, &log->flag);
log              1028 fs/jfs/jfs_logmgr.c 			 log->syncpt);
log              1032 fs/jfs/jfs_logmgr.c 		jfs_flush_journal(log, 0);
log              1046 fs/jfs/jfs_logmgr.c void jfs_syncpt(struct jfs_log *log, int hard_sync)
log              1047 fs/jfs/jfs_logmgr.c {	LOG_LOCK(log);
log              1048 fs/jfs/jfs_logmgr.c 	if (!test_bit(log_QUIESCE, &log->flag))
log              1049 fs/jfs/jfs_logmgr.c 		lmLogSync(log, hard_sync);
log              1050 fs/jfs/jfs_logmgr.c 	LOG_UNLOCK(log);
log              1070 fs/jfs/jfs_logmgr.c 	struct jfs_log *log;
log              1080 fs/jfs/jfs_logmgr.c 	list_for_each_entry(log, &jfs_external_logs, journal_list) {
log              1081 fs/jfs/jfs_logmgr.c 		if (log->bdev->bd_dev == sbi->logdev) {
log              1082 fs/jfs/jfs_logmgr.c 			if (!uuid_equal(&log->uuid, &sbi->loguuid)) {
log              1090 fs/jfs/jfs_logmgr.c 			if ((rc = lmLogFileSystem(log, sbi, 1))) {
log              1098 fs/jfs/jfs_logmgr.c 	if (!(log = kzalloc(sizeof(struct jfs_log), GFP_KERNEL))) {
log              1102 fs/jfs/jfs_logmgr.c 	INIT_LIST_HEAD(&log->sb_list);
log              1103 fs/jfs/jfs_logmgr.c 	init_waitqueue_head(&log->syncwait);
log              1112 fs/jfs/jfs_logmgr.c 				 log);
log              1118 fs/jfs/jfs_logmgr.c 	log->bdev = bdev;
log              1119 fs/jfs/jfs_logmgr.c 	uuid_copy(&log->uuid, &sbi->loguuid);
log              1124 fs/jfs/jfs_logmgr.c 	if ((rc = lmLogInit(log)))
log              1127 fs/jfs/jfs_logmgr.c 	list_add(&log->journal_list, &jfs_external_logs);
log              1132 fs/jfs/jfs_logmgr.c 	if ((rc = lmLogFileSystem(log, sbi, 1)))
log              1136 fs/jfs/jfs_logmgr.c 	LOG_LOCK(log);
log              1137 fs/jfs/jfs_logmgr.c 	list_add(&sbi->log_list, &log->sb_list);
log              1138 fs/jfs/jfs_logmgr.c 	sbi->log = log;
log              1139 fs/jfs/jfs_logmgr.c 	LOG_UNLOCK(log);
log              1148 fs/jfs/jfs_logmgr.c 	list_del(&log->journal_list);
log              1149 fs/jfs/jfs_logmgr.c 	lbmLogShutdown(log);
log              1156 fs/jfs/jfs_logmgr.c 	kfree(log);
log              1164 fs/jfs/jfs_logmgr.c 	struct jfs_log *log;
log              1167 fs/jfs/jfs_logmgr.c 	if (!(log = kzalloc(sizeof(struct jfs_log), GFP_KERNEL)))
log              1169 fs/jfs/jfs_logmgr.c 	INIT_LIST_HEAD(&log->sb_list);
log              1170 fs/jfs/jfs_logmgr.c 	init_waitqueue_head(&log->syncwait);
log              1172 fs/jfs/jfs_logmgr.c 	set_bit(log_INLINELOG, &log->flag);
log              1173 fs/jfs/jfs_logmgr.c 	log->bdev = sb->s_bdev;
log              1174 fs/jfs/jfs_logmgr.c 	log->base = addressPXD(&JFS_SBI(sb)->logpxd);
log              1175 fs/jfs/jfs_logmgr.c 	log->size = lengthPXD(&JFS_SBI(sb)->logpxd) >>
log              1177 fs/jfs/jfs_logmgr.c 	log->l2bsize = sb->s_blocksize_bits;
log              1183 fs/jfs/jfs_logmgr.c 	if ((rc = lmLogInit(log))) {
log              1184 fs/jfs/jfs_logmgr.c 		kfree(log);
log              1189 fs/jfs/jfs_logmgr.c 	list_add(&JFS_SBI(sb)->log_list, &log->sb_list);
log              1190 fs/jfs/jfs_logmgr.c 	JFS_SBI(sb)->log = log;
log              1223 fs/jfs/jfs_logmgr.c 	JFS_SBI(sb)->log = dummy_log;
log              1248 fs/jfs/jfs_logmgr.c int lmLogInit(struct jfs_log * log)
log              1258 fs/jfs/jfs_logmgr.c 	jfs_info("lmLogInit: log:0x%p", log);
log              1261 fs/jfs/jfs_logmgr.c 	LOGGC_LOCK_INIT(log);
log              1264 fs/jfs/jfs_logmgr.c 	LOG_LOCK_INIT(log);
log              1266 fs/jfs/jfs_logmgr.c 	LOGSYNC_LOCK_INIT(log);
log              1268 fs/jfs/jfs_logmgr.c 	INIT_LIST_HEAD(&log->synclist);
log              1270 fs/jfs/jfs_logmgr.c 	INIT_LIST_HEAD(&log->cqueue);
log              1271 fs/jfs/jfs_logmgr.c 	log->flush_tblk = NULL;
log              1273 fs/jfs/jfs_logmgr.c 	log->count = 0;
log              1278 fs/jfs/jfs_logmgr.c 	if ((rc = lbmLogInit(log)))
log              1281 fs/jfs/jfs_logmgr.c 	if (!test_bit(log_INLINELOG, &log->flag))
log              1282 fs/jfs/jfs_logmgr.c 		log->l2bsize = L2LOGPSIZE;
log              1285 fs/jfs/jfs_logmgr.c 	if (log->no_integrity) {
log              1291 fs/jfs/jfs_logmgr.c 		bp = lbmAllocate(log , 0);
log              1292 fs/jfs/jfs_logmgr.c 		log->bp = bp;
log              1298 fs/jfs/jfs_logmgr.c 		if ((rc = lbmRead(log, 1, &bpsuper)))
log              1317 fs/jfs/jfs_logmgr.c 		if (test_bit(log_INLINELOG,&log->flag)) {
log              1318 fs/jfs/jfs_logmgr.c 			if (log->size != le32_to_cpu(logsuper->size)) {
log              1323 fs/jfs/jfs_logmgr.c 				 log, (unsigned long long)log->base, log->size);
log              1325 fs/jfs/jfs_logmgr.c 			if (!uuid_equal(&logsuper->uuid, &log->uuid)) {
log              1329 fs/jfs/jfs_logmgr.c 			log->size = le32_to_cpu(logsuper->size);
log              1330 fs/jfs/jfs_logmgr.c 			log->l2bsize = le32_to_cpu(logsuper->l2bsize);
log              1332 fs/jfs/jfs_logmgr.c 				 log, (unsigned long long)log->base, log->size);
log              1335 fs/jfs/jfs_logmgr.c 		log->page = le32_to_cpu(logsuper->end) / LOGPSIZE;
log              1336 fs/jfs/jfs_logmgr.c 		log->eor = le32_to_cpu(logsuper->end) - (LOGPSIZE * log->page);
log              1342 fs/jfs/jfs_logmgr.c 		if ((rc = lbmRead(log, log->page, &bp)))
log              1348 fs/jfs/jfs_logmgr.c 			 le32_to_cpu(logsuper->end), log->page, log->eor,
log              1351 fs/jfs/jfs_logmgr.c 		log->bp = bp;
log              1352 fs/jfs/jfs_logmgr.c 		bp->l_pn = log->page;
log              1353 fs/jfs/jfs_logmgr.c 		bp->l_eor = log->eor;
log              1356 fs/jfs/jfs_logmgr.c 		if (log->eor >= LOGPSIZE - LOGPTLRSIZE)
log              1357 fs/jfs/jfs_logmgr.c 			lmNextPage(log);
log              1373 fs/jfs/jfs_logmgr.c 		lrd.log.syncpt.sync = 0;
log              1374 fs/jfs/jfs_logmgr.c 		lsn = lmWriteRecord(log, NULL, &lrd, NULL);
log              1375 fs/jfs/jfs_logmgr.c 		bp = log->bp;
log              1379 fs/jfs/jfs_logmgr.c 		lbmWrite(log, bp, lbmWRITE | lbmSYNC, 0);
log              1387 fs/jfs/jfs_logmgr.c 		log->serial = le32_to_cpu(logsuper->serial) + 1;
log              1388 fs/jfs/jfs_logmgr.c 		logsuper->serial = cpu_to_le32(log->serial);
log              1389 fs/jfs/jfs_logmgr.c 		lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC);
log              1395 fs/jfs/jfs_logmgr.c 	log->logsize = (log->size - 2) << L2LOGPSIZE;
log              1396 fs/jfs/jfs_logmgr.c 	log->lsn = lsn;
log              1397 fs/jfs/jfs_logmgr.c 	log->syncpt = lsn;
log              1398 fs/jfs/jfs_logmgr.c 	log->sync = log->syncpt;
log              1399 fs/jfs/jfs_logmgr.c 	log->nextsync = LOGSYNC_DELTA(log->logsize);
log              1402 fs/jfs/jfs_logmgr.c 		 log->lsn, log->syncpt, log->sync);
log              1407 fs/jfs/jfs_logmgr.c 	log->clsn = lsn;
log              1415 fs/jfs/jfs_logmgr.c 	log->wqueue = NULL;
log              1423 fs/jfs/jfs_logmgr.c 	lbmLogShutdown(log);
log              1445 fs/jfs/jfs_logmgr.c 	struct jfs_log *log = sbi->log;
log              1449 fs/jfs/jfs_logmgr.c 	jfs_info("lmLogClose: log:0x%p", log);
log              1452 fs/jfs/jfs_logmgr.c 	LOG_LOCK(log);
log              1454 fs/jfs/jfs_logmgr.c 	LOG_UNLOCK(log);
log              1455 fs/jfs/jfs_logmgr.c 	sbi->log = NULL;
log              1463 fs/jfs/jfs_logmgr.c 	if (test_bit(log_INLINELOG, &log->flag)) {
log              1467 fs/jfs/jfs_logmgr.c 		rc = lmLogShutdown(log);
log              1468 fs/jfs/jfs_logmgr.c 		kfree(log);
log              1472 fs/jfs/jfs_logmgr.c 	if (!log->no_integrity)
log              1473 fs/jfs/jfs_logmgr.c 		lmLogFileSystem(log, sbi, 0);
log              1475 fs/jfs/jfs_logmgr.c 	if (!list_empty(&log->sb_list))
log              1485 fs/jfs/jfs_logmgr.c 	if (log->no_integrity)
log              1491 fs/jfs/jfs_logmgr.c 	list_del(&log->journal_list);
log              1492 fs/jfs/jfs_logmgr.c 	bdev = log->bdev;
log              1493 fs/jfs/jfs_logmgr.c 	rc = lmLogShutdown(log);
log              1497 fs/jfs/jfs_logmgr.c 	kfree(log);
log              1516 fs/jfs/jfs_logmgr.c void jfs_flush_journal(struct jfs_log *log, int wait)
log              1522 fs/jfs/jfs_logmgr.c 	if (!log)
log              1525 fs/jfs/jfs_logmgr.c 	jfs_info("jfs_flush_journal: log:0x%p wait=%d", log, wait);
log              1527 fs/jfs/jfs_logmgr.c 	LOGGC_LOCK(log);
log              1529 fs/jfs/jfs_logmgr.c 	if (!list_empty(&log->cqueue)) {
log              1534 fs/jfs/jfs_logmgr.c 		target = list_entry(log->cqueue.prev, struct tblock, cqueue);
log              1536 fs/jfs/jfs_logmgr.c 		if (test_bit(log_FLUSH, &log->flag)) {
log              1543 fs/jfs/jfs_logmgr.c 			if (log->flush_tblk)
log              1544 fs/jfs/jfs_logmgr.c 				log->flush_tblk = target;
log              1547 fs/jfs/jfs_logmgr.c 			log->flush_tblk = target;
log              1548 fs/jfs/jfs_logmgr.c 			set_bit(log_FLUSH, &log->flag);
log              1553 fs/jfs/jfs_logmgr.c 			if (!(log->cflag & logGC_PAGEOUT)) {
log              1554 fs/jfs/jfs_logmgr.c 				log->cflag |= logGC_PAGEOUT;
log              1555 fs/jfs/jfs_logmgr.c 				lmGCwrite(log, 0);
log              1559 fs/jfs/jfs_logmgr.c 	if ((wait > 1) || test_bit(log_SYNCBARRIER, &log->flag)) {
log              1561 fs/jfs/jfs_logmgr.c 		set_bit(log_FLUSH, &log->flag);
log              1562 fs/jfs/jfs_logmgr.c 		log->flush_tblk = NULL;
log              1570 fs/jfs/jfs_logmgr.c 		LOGGC_UNLOCK(log);
log              1572 fs/jfs/jfs_logmgr.c 		LOGGC_LOCK(log);
log              1575 fs/jfs/jfs_logmgr.c 	LOGGC_UNLOCK(log);
log              1580 fs/jfs/jfs_logmgr.c 	write_special_inodes(log, filemap_fdatawrite);
log              1586 fs/jfs/jfs_logmgr.c 	if ((!list_empty(&log->cqueue)) || !list_empty(&log->synclist)) {
log              1589 fs/jfs/jfs_logmgr.c 			write_special_inodes(log, filemap_fdatawrite);
log              1590 fs/jfs/jfs_logmgr.c 			if (list_empty(&log->cqueue) &&
log              1591 fs/jfs/jfs_logmgr.c 			    list_empty(&log->synclist))
log              1595 fs/jfs/jfs_logmgr.c 	assert(list_empty(&log->cqueue));
log              1598 fs/jfs/jfs_logmgr.c 	if (!list_empty(&log->synclist)) {
log              1602 fs/jfs/jfs_logmgr.c 		list_for_each_entry(lp, &log->synclist, synclist) {
log              1619 fs/jfs/jfs_logmgr.c 	WARN_ON(!list_empty(&log->synclist));
log              1621 fs/jfs/jfs_logmgr.c 	clear_bit(log_FLUSH, &log->flag);
log              1638 fs/jfs/jfs_logmgr.c int lmLogShutdown(struct jfs_log * log)
log              1648 fs/jfs/jfs_logmgr.c 	jfs_info("lmLogShutdown: log:0x%p", log);
log              1650 fs/jfs/jfs_logmgr.c 	jfs_flush_journal(log, 2);
log              1660 fs/jfs/jfs_logmgr.c 	lrd.log.syncpt.sync = 0;
log              1662 fs/jfs/jfs_logmgr.c 	lsn = lmWriteRecord(log, NULL, &lrd, NULL);
log              1663 fs/jfs/jfs_logmgr.c 	bp = log->bp;
log              1666 fs/jfs/jfs_logmgr.c 	lbmWrite(log, log->bp, lbmWRITE | lbmRELEASE | lbmSYNC, 0);
log              1667 fs/jfs/jfs_logmgr.c 	lbmIOWait(log->bp, lbmFREE);
log              1668 fs/jfs/jfs_logmgr.c 	log->bp = NULL;
log              1675 fs/jfs/jfs_logmgr.c 	if ((rc = lbmRead(log, 1, &bpsuper)))
log              1681 fs/jfs/jfs_logmgr.c 	lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC);
log              1685 fs/jfs/jfs_logmgr.c 		 lsn, log->page, log->eor);
log              1691 fs/jfs/jfs_logmgr.c 	lbmLogShutdown(log);
log              1714 fs/jfs/jfs_logmgr.c static int lmLogFileSystem(struct jfs_log * log, struct jfs_sb_info *sbi,
log              1726 fs/jfs/jfs_logmgr.c 	if ((rc = lbmRead(log, 1, &bpsuper)))
log              1768 fs/jfs/jfs_logmgr.c 	lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC);
log              1794 fs/jfs/jfs_logmgr.c static int lbmLogInit(struct jfs_log * log)
log              1799 fs/jfs/jfs_logmgr.c 	jfs_info("lbmLogInit: log:0x%p", log);
log              1802 fs/jfs/jfs_logmgr.c 	log->bp = NULL;
log              1805 fs/jfs/jfs_logmgr.c 	log->wqueue = NULL;
log              1816 fs/jfs/jfs_logmgr.c 	init_waitqueue_head(&log->free_wait);
log              1818 fs/jfs/jfs_logmgr.c 	log->lbuf_free = NULL;
log              1840 fs/jfs/jfs_logmgr.c 			lbuf->l_log = log;
log              1843 fs/jfs/jfs_logmgr.c 			lbuf->l_freelist = log->lbuf_free;
log              1844 fs/jfs/jfs_logmgr.c 			log->lbuf_free = lbuf;
log              1852 fs/jfs/jfs_logmgr.c 	lbmLogShutdown(log);
log              1862 fs/jfs/jfs_logmgr.c static void lbmLogShutdown(struct jfs_log * log)
log              1866 fs/jfs/jfs_logmgr.c 	jfs_info("lbmLogShutdown: log:0x%p", log);
log              1868 fs/jfs/jfs_logmgr.c 	lbuf = log->lbuf_free;
log              1883 fs/jfs/jfs_logmgr.c static struct lbuf *lbmAllocate(struct jfs_log * log, int pn)
log              1892 fs/jfs/jfs_logmgr.c 	LCACHE_SLEEP_COND(log->free_wait, (bp = log->lbuf_free), flags);
log              1893 fs/jfs/jfs_logmgr.c 	log->lbuf_free = bp->l_freelist;
log              1902 fs/jfs/jfs_logmgr.c 	bp->l_blkno = log->base + (pn << (L2LOGPSIZE - log->l2bsize));
log              1927 fs/jfs/jfs_logmgr.c 	struct jfs_log *log = bp->l_log;
log              1934 fs/jfs/jfs_logmgr.c 	bp->l_freelist = log->lbuf_free;
log              1935 fs/jfs/jfs_logmgr.c 	log->lbuf_free = bp;
log              1937 fs/jfs/jfs_logmgr.c 	wake_up(&log->free_wait);
log              1969 fs/jfs/jfs_logmgr.c static int lbmRead(struct jfs_log * log, int pn, struct lbuf ** bpp)
log              1977 fs/jfs/jfs_logmgr.c 	*bpp = bp = lbmAllocate(log, pn);
log              1984 fs/jfs/jfs_logmgr.c 	bio->bi_iter.bi_sector = bp->l_blkno << (log->l2bsize - 9);
log              1985 fs/jfs/jfs_logmgr.c 	bio_set_dev(bio, log->bdev);
log              1994 fs/jfs/jfs_logmgr.c 	if (log->no_integrity) {
log              2022 fs/jfs/jfs_logmgr.c static void lbmWrite(struct jfs_log * log, struct lbuf * bp, int flag,
log              2032 fs/jfs/jfs_logmgr.c 	    log->base + (bp->l_pn << (L2LOGPSIZE - log->l2bsize));
log              2047 fs/jfs/jfs_logmgr.c 	tail = log->wqueue;
log              2053 fs/jfs/jfs_logmgr.c 			log->wqueue = bp;
log              2056 fs/jfs/jfs_logmgr.c 			log->wqueue = bp;
log              2077 fs/jfs/jfs_logmgr.c 		LOGGC_UNLOCK(log);
log              2079 fs/jfs/jfs_logmgr.c 		LOGGC_LOCK(log);
log              2090 fs/jfs/jfs_logmgr.c static void lbmDirectWrite(struct jfs_log * log, struct lbuf * bp, int flag)
log              2102 fs/jfs/jfs_logmgr.c 	    log->base + (bp->l_pn << (L2LOGPSIZE - log->l2bsize));
log              2123 fs/jfs/jfs_logmgr.c 	struct jfs_log *log = bp->l_log;
log              2128 fs/jfs/jfs_logmgr.c 	bio->bi_iter.bi_sector = bp->l_blkno << (log->l2bsize - 9);
log              2129 fs/jfs/jfs_logmgr.c 	bio_set_dev(bio, log->bdev);
log              2139 fs/jfs/jfs_logmgr.c 	if (log->no_integrity) {
log              2183 fs/jfs/jfs_logmgr.c 	struct jfs_log *log;
log              2233 fs/jfs/jfs_logmgr.c 	log = bp->l_log;
log              2234 fs/jfs/jfs_logmgr.c 	log->clsn = (bp->l_pn << L2LOGPSIZE) + bp->l_ceor;
log              2242 fs/jfs/jfs_logmgr.c 	tail = log->wqueue;
log              2250 fs/jfs/jfs_logmgr.c 			log->wqueue = NULL;
log              2364 fs/jfs/jfs_logmgr.c int lmLogFormat(struct jfs_log *log, s64 logAddress, int logSize)
log              2378 fs/jfs/jfs_logmgr.c 	sbi = list_entry(log->sb_list.next, struct jfs_sb_info, log_list);
log              2381 fs/jfs/jfs_logmgr.c 	bp = lbmAllocate(log, 1);
log              2449 fs/jfs/jfs_logmgr.c 	lrd_ptr->log.syncpt.sync = 0;
log               336 fs/jfs/jfs_logmgr.h 	} log;
log               477 fs/jfs/jfs_logmgr.h #define LOGSYNC_LOCK_INIT(log) spin_lock_init(&(log)->synclock)
log               478 fs/jfs/jfs_logmgr.h #define LOGSYNC_LOCK(log, flags) spin_lock_irqsave(&(log)->synclock, flags)
log               479 fs/jfs/jfs_logmgr.h #define LOGSYNC_UNLOCK(log, flags) \
log               480 fs/jfs/jfs_logmgr.h 	spin_unlock_irqrestore(&(log)->synclock, flags)
log               483 fs/jfs/jfs_logmgr.h #define logdiff(diff, lsn, log)\
log               485 fs/jfs/jfs_logmgr.h 	diff = (lsn) - (log)->syncpt;\
log               487 fs/jfs/jfs_logmgr.h 		diff += (log)->logsize;\
log               492 fs/jfs/jfs_logmgr.h extern int lmLogShutdown(struct jfs_log * log);
log               493 fs/jfs/jfs_logmgr.h extern int lmLogInit(struct jfs_log * log);
log               494 fs/jfs/jfs_logmgr.h extern int lmLogFormat(struct jfs_log *log, s64 logAddress, int logSize);
log               497 fs/jfs/jfs_logmgr.h extern void jfs_flush_journal(struct jfs_log * log, int wait);
log               498 fs/jfs/jfs_logmgr.h extern void jfs_syncpt(struct jfs_log *log, int hard_sync);
log               182 fs/jfs/jfs_metapage.c 		mp->log = NULL;
log               281 fs/jfs/jfs_metapage.c 	struct jfs_log *log = mp->log;
log               287 fs/jfs/jfs_metapage.c 	if (!log)
log               290 fs/jfs/jfs_metapage.c 	LOGSYNC_LOCK(log, flags);
log               292 fs/jfs/jfs_metapage.c 		mp->log = NULL;
log               295 fs/jfs/jfs_metapage.c 		log->count--;
log               298 fs/jfs/jfs_metapage.c 	LOGSYNC_UNLOCK(log, flags);
log               373 fs/jfs/jfs_metapage.c 			if (mp->log && !(mp->log->cflag & logGC_PAGEOUT))
log               374 fs/jfs/jfs_metapage.c 				jfs_flush_journal(mp->log, 0);
log                34 fs/jfs/jfs_metapage.h 	struct jfs_log *log;
log               427 fs/jfs/jfs_mount.c 		j_sb->s_logdev = cpu_to_le32(new_encode_dev(sbi->log->bdev->bd_dev));
log               428 fs/jfs/jfs_mount.c 		j_sb->s_logserial = cpu_to_le32(sbi->log->serial);
log               483 fs/jfs/jfs_mount.c 	struct jfs_log *log = JFS_SBI(sb)->log;
log               491 fs/jfs/jfs_mount.c 	lmLog(log, NULL, &lrd, NULL);
log               151 fs/jfs/jfs_txnmgr.c static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               153 fs/jfs/jfs_txnmgr.c static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               155 fs/jfs/jfs_txnmgr.c static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               157 fs/jfs/jfs_txnmgr.c static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               162 fs/jfs/jfs_txnmgr.c static int txLog(struct jfs_log * log, struct tblock * tblk,
log               166 fs/jfs/jfs_txnmgr.c static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log               352 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log               355 fs/jfs/jfs_txnmgr.c 	log = JFS_SBI(sb)->log;
log               366 fs/jfs/jfs_txnmgr.c 		if (test_bit(log_SYNCBARRIER, &log->flag) ||
log               367 fs/jfs/jfs_txnmgr.c 		    test_bit(log_QUIESCE, &log->flag)) {
log               369 fs/jfs/jfs_txnmgr.c 			TXN_SLEEP(&log->syncwait);
log               421 fs/jfs/jfs_txnmgr.c 	++log->logtid;
log               422 fs/jfs/jfs_txnmgr.c 	tblk->logtid = log->logtid;
log               424 fs/jfs/jfs_txnmgr.c 	++log->active;
log               449 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log               451 fs/jfs/jfs_txnmgr.c 	log = JFS_SBI(sb)->log;
log               460 fs/jfs/jfs_txnmgr.c 	if (test_bit(log_SYNCBARRIER, &log->flag) ||
log               461 fs/jfs/jfs_txnmgr.c 	    test_bit(log_QUIESCE, &log->flag)) {
log               463 fs/jfs/jfs_txnmgr.c 		TXN_SLEEP(&log->syncwait);
log               490 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log               501 fs/jfs/jfs_txnmgr.c 	log = JFS_SBI(tblk->sb)->log;
log               515 fs/jfs/jfs_txnmgr.c 		spin_lock_irq(&log->gclock);	// LOGGC_LOCK
log               517 fs/jfs/jfs_txnmgr.c 		spin_unlock_irq(&log->gclock);	// LOGGC_UNLOCK
log               534 fs/jfs/jfs_txnmgr.c 	if (--log->active == 0) {
log               535 fs/jfs/jfs_txnmgr.c 		clear_bit(log_FLUSH, &log->flag);
log               540 fs/jfs/jfs_txnmgr.c 		if (test_bit(log_SYNCBARRIER, &log->flag)) {
log               544 fs/jfs/jfs_txnmgr.c 			jfs_syncpt(log, 1);
log               546 fs/jfs/jfs_txnmgr.c 			jfs_info("log barrier off: 0x%x", log->lsn);
log               549 fs/jfs/jfs_txnmgr.c 			clear_bit(log_SYNCBARRIER, &log->flag);
log               552 fs/jfs/jfs_txnmgr.c 			TXN_WAKEUP(&log->syncwait);
log               901 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log               906 fs/jfs/jfs_txnmgr.c 	log = JFS_SBI(tblk->sb)->log;
log               930 fs/jfs/jfs_txnmgr.c 			LOGSYNC_LOCK(log, flags);
log               932 fs/jfs/jfs_txnmgr.c 				logdiff(difft, tblk->clsn, log);
log               933 fs/jfs/jfs_txnmgr.c 				logdiff(diffp, mp->clsn, log);
log               938 fs/jfs/jfs_txnmgr.c 			LOGSYNC_UNLOCK(log, flags);
log               969 fs/jfs/jfs_txnmgr.c 		LOGSYNC_LOCK(log, flags);
log               970 fs/jfs/jfs_txnmgr.c 		log->count--;
log               972 fs/jfs/jfs_txnmgr.c 		LOGSYNC_UNLOCK(log, flags);
log              1129 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log              1155 fs/jfs/jfs_txnmgr.c 	log = JFS_SBI(sb)->log;
log              1156 fs/jfs/jfs_txnmgr.c 	cd.log = log;
log              1259 fs/jfs/jfs_txnmgr.c 	if ((rc = txLog(log, tblk, &cd)))
log              1298 fs/jfs/jfs_txnmgr.c 	lmLog(log, tblk, lrd, NULL);
log              1300 fs/jfs/jfs_txnmgr.c 	lmGroupCommit(log, tblk);
log              1368 fs/jfs/jfs_txnmgr.c static int txLog(struct jfs_log * log, struct tblock * tblk, struct commit * cd)
log              1387 fs/jfs/jfs_txnmgr.c 		lrd->log.redopage.fileset = cpu_to_le32(JFS_IP(ip)->fileset);
log              1388 fs/jfs/jfs_txnmgr.c 		lrd->log.redopage.inode = cpu_to_le32(ip->i_ino);
log              1393 fs/jfs/jfs_txnmgr.c 			xtLog(log, tblk, lrd, tlck);
log              1397 fs/jfs/jfs_txnmgr.c 			dtLog(log, tblk, lrd, tlck);
log              1401 fs/jfs/jfs_txnmgr.c 			diLog(log, tblk, lrd, tlck, cd);
log              1405 fs/jfs/jfs_txnmgr.c 			mapLog(log, tblk, lrd, tlck);
log              1409 fs/jfs/jfs_txnmgr.c 			dataLog(log, tblk, lrd, tlck);
log              1425 fs/jfs/jfs_txnmgr.c static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log              1436 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.type = cpu_to_le16(LOG_INODE);
log              1437 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.l2linesize = cpu_to_le16(L2INODESLOTSIZE);
log              1439 fs/jfs/jfs_txnmgr.c 	pxd = &lrd->log.redopage.pxd;
log              1450 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1480 fs/jfs/jfs_txnmgr.c 		lrd->log.noredoinoext.iagnum =
log              1482 fs/jfs/jfs_txnmgr.c 		lrd->log.noredoinoext.inoext_idx =
log              1487 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              1514 fs/jfs/jfs_txnmgr.c 				lrd->log.updatemap.type =
log              1517 fs/jfs/jfs_txnmgr.c 				lrd->log.updatemap.type =
log              1519 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.nxd = cpu_to_le16(1);
log              1520 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.pxd = pxdlock->pxd;
log              1522 fs/jfs/jfs_txnmgr.c 			    cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              1538 fs/jfs/jfs_txnmgr.c static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log              1547 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.type = cpu_to_le16(LOG_DATA);
log              1548 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.l2linesize = cpu_to_le16(L2DATASLOTSIZE);
log              1550 fs/jfs/jfs_txnmgr.c 	pxd = &lrd->log.redopage.pxd;
log              1571 fs/jfs/jfs_txnmgr.c 	lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1584 fs/jfs/jfs_txnmgr.c static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log              1594 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.type = cpu_to_le16(LOG_DTREE);
log              1595 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.l2linesize = cpu_to_le16(L2DTSLOTSIZE);
log              1597 fs/jfs/jfs_txnmgr.c 	pxd = &lrd->log.redopage.pxd;
log              1600 fs/jfs/jfs_txnmgr.c 		lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT);
log              1615 fs/jfs/jfs_txnmgr.c 			lrd->log.redopage.type |= cpu_to_le16(LOG_EXTEND);
log              1617 fs/jfs/jfs_txnmgr.c 			lrd->log.redopage.type |= cpu_to_le16(LOG_NEW);
log              1621 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1650 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1672 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              1687 fs/jfs/jfs_txnmgr.c static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log              1704 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.type = cpu_to_le16(LOG_XTREE);
log              1705 fs/jfs/jfs_txnmgr.c 	lrd->log.redopage.l2linesize = cpu_to_le16(L2XTSLOTSIZE);
log              1707 fs/jfs/jfs_txnmgr.c 	page_pxd = &lrd->log.redopage.pxd;
log              1710 fs/jfs/jfs_txnmgr.c 		lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT);
log              1713 fs/jfs/jfs_txnmgr.c 			lrd->log.redopage.type |=
log              1740 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1831 fs/jfs/jfs_txnmgr.c 			    cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              1837 fs/jfs/jfs_txnmgr.c 				    cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1846 fs/jfs/jfs_txnmgr.c 		lrd->log.updatemap.type = cpu_to_le16(LOG_FREEXADLIST);
log              1849 fs/jfs/jfs_txnmgr.c 		lrd->log.updatemap.nxd =
log              1855 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1950 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              1964 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD);
log              1965 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.nxd = cpu_to_le16(1);
log              1966 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.pxd = pxdlock->pxd;
log              1969 fs/jfs/jfs_txnmgr.c 			    cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              1981 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.type =
log              1985 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.nxd =
log              1992 fs/jfs/jfs_txnmgr.c 			    cpu_to_le32(lmLog(log, tblk, lrd, tlck));
log              2070 fs/jfs/jfs_txnmgr.c static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
log              2090 fs/jfs/jfs_txnmgr.c 		pxd = &lrd->log.redopage.pxd;
log              2092 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              2105 fs/jfs/jfs_txnmgr.c 		lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD);
log              2106 fs/jfs/jfs_txnmgr.c 		lrd->log.updatemap.nxd = cpu_to_le16(1);
log              2107 fs/jfs/jfs_txnmgr.c 		lrd->log.updatemap.pxd = pxdlock->pxd;
log              2108 fs/jfs/jfs_txnmgr.c 		lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              2132 fs/jfs/jfs_txnmgr.c 				lrd->log.updatemap.type =
log              2135 fs/jfs/jfs_txnmgr.c 				lrd->log.updatemap.type =
log              2137 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.nxd = cpu_to_le16(1);
log              2138 fs/jfs/jfs_txnmgr.c 			lrd->log.updatemap.pxd = pxdlock->pxd;
log              2140 fs/jfs/jfs_txnmgr.c 			    cpu_to_le32(lmLog(log, tblk, lrd, NULL));
log              2684 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log;
log              2698 fs/jfs/jfs_txnmgr.c 	log = (struct jfs_log *) JFS_SBI(tblk->sb)->log;
log              2700 fs/jfs/jfs_txnmgr.c 	spin_lock_irq(&log->gclock);	// LOGGC_LOCK
log              2705 fs/jfs/jfs_txnmgr.c 		log->gcrtc--;
log              2713 fs/jfs/jfs_txnmgr.c 		spin_unlock_irq(&log->gclock);	// LOGGC_UNLOCK
log              2718 fs/jfs/jfs_txnmgr.c 		spin_unlock_irq(&log->gclock);	// LOGGC_UNLOCK
log              2825 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log = mp->log;
log              2828 fs/jfs/jfs_txnmgr.c 	assert(log);
log              2846 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log = JFS_SBI(sb)->log;
log              2849 fs/jfs/jfs_txnmgr.c 	set_bit(log_QUIESCE, &log->flag);
log              2890 fs/jfs/jfs_txnmgr.c 	jfs_flush_journal(log, 0);
log              2900 fs/jfs/jfs_txnmgr.c 	struct jfs_log *log = JFS_SBI(sb)->log;
log              2902 fs/jfs/jfs_txnmgr.c 	clear_bit(log_QUIESCE, &log->flag);
log              2903 fs/jfs/jfs_txnmgr.c 	TXN_WAKEUP(&log->syncwait);
log               263 fs/jfs/jfs_txnmgr.h 	struct jfs_log *log;	/* log */
log                44 fs/jfs/jfs_umount.c 	struct jfs_log *log;
log                54 fs/jfs/jfs_umount.c 	if ((log = sbi->log))
log                58 fs/jfs/jfs_umount.c 		jfs_flush_journal(log, 2);
log               108 fs/jfs/jfs_umount.c 	if (log) {		/* log = NULL if read-only mount */
log               126 fs/jfs/jfs_umount.c 	struct jfs_log *log = sbi->log;
log               128 fs/jfs/jfs_umount.c 	if (!log)
log               136 fs/jfs/jfs_umount.c 	jfs_flush_journal(log, 2);
log              3168 fs/jfs/jfs_xtree.c 	int freed, log;
log              3183 fs/jfs/jfs_xtree.c 		log = 1;
log              3185 fs/jfs/jfs_xtree.c 		log = 0;
log              3245 fs/jfs/jfs_xtree.c 		if (log)
log              3274 fs/jfs/jfs_xtree.c 	if (log) {
log              3352 fs/jfs/jfs_xtree.c 			if (log) {	/* COMMIT_PWMAP */
log              3393 fs/jfs/jfs_xtree.c 			if (!log) {	/* COMMIT_WAMP */
log              3415 fs/jfs/jfs_xtree.c 	if (log) {		/* COMMIT_PWMAP */
log              3439 fs/jfs/jfs_xtree.c 		if (log) {	/* COMMIT_PWMAP */
log              3483 fs/jfs/jfs_xtree.c 			if (log) {	/* COMMIT_PWMAP */
log              3532 fs/jfs/jfs_xtree.c 	if (log && mp->lid && (tblk->last != mp->lid) &&
log              3558 fs/jfs/jfs_xtree.c 		if (log) {	/* COMMIT_PWMAP */
log              3597 fs/jfs/jfs_xtree.c 			if (log) {	/* COMMIT_PWMAP */
log                56 fs/jfs/resize.c 	struct jfs_log *log = sbi->log;
log               184 fs/jfs/resize.c 		if ((rc = lmLogFormat(log, newLogAddress, newLogSize)))
log               207 fs/jfs/resize.c 		lmLogShutdown(log);
log               248 fs/jfs/resize.c 			if ((rc = lmLogFormat(log, newLogAddress, newLogSize)))
log               254 fs/jfs/resize.c 		log->base = newLogAddress;
log               255 fs/jfs/resize.c 		log->size = newLogSize >> (L2LOGPSIZE - sb->s_blocksize_bits);
log               256 fs/jfs/resize.c 		if ((rc = lmLogInit(log)))
log               499 fs/jfs/resize.c 	j_sb->s_logserial = cpu_to_le32(log->serial);
log               569 fs/jfs/super.c 		sbi->log = NULL;
log               625 fs/jfs/super.c 	struct jfs_log *log = sbi->log;
log               630 fs/jfs/super.c 		rc = lmLogShutdown(log);
log               655 fs/jfs/super.c 	struct jfs_log *log = sbi->log;
log               664 fs/jfs/super.c 		rc = lmLogInit(log);
log               681 fs/jfs/super.c 	struct jfs_log *log = JFS_SBI(sb)->log;
log               684 fs/jfs/super.c 	if (log) {
log               690 fs/jfs/super.c 		jfs_flush_journal(log, wait);
log               691 fs/jfs/super.c 		jfs_syncpt(log, 0);
log                37 fs/xfs/libxfs/xfs_log_format.h #define XLOG_BTOLSUNIT(log, b)  (((b)+(log)->l_mp->m_sb.sb_logsunit-1) / \
log                38 fs/xfs/libxfs/xfs_log_format.h                                  (log)->l_mp->m_sb.sb_logsunit)
log                39 fs/xfs/libxfs/xfs_log_format.h #define XLOG_LSUNITTOB(log, su) ((su) * (log)->l_mp->m_sb.sb_logsunit)
log                46 fs/xfs/libxfs/xfs_log_format.h #define XLOG_REC_SHIFT(log) \
log                47 fs/xfs/libxfs/xfs_log_format.h 	BTOBB(1 << (xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? \
log                49 fs/xfs/libxfs/xfs_log_format.h #define XLOG_TOTAL_REC_SHIFT(log) \
log                50 fs/xfs/libxfs/xfs_log_format.h 	BTOBB(XLOG_MAX_ICLOGS << (xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? \
log               447 fs/xfs/libxfs/xfs_rtbitmap.c 	int		log,		/* log2 of extent size */
log               463 fs/xfs/libxfs/xfs_rtbitmap.c 	so = XFS_SUMOFFS(mp, log, bbno);
log               501 fs/xfs/libxfs/xfs_rtbitmap.c 			if (*sp == 0 && log == mp->m_rsum_cache[bbno])
log               503 fs/xfs/libxfs/xfs_rtbitmap.c 			if (*sp != 0 && log < mp->m_rsum_cache[bbno])
log               504 fs/xfs/libxfs/xfs_rtbitmap.c 				mp->m_rsum_cache[bbno] = log;
log               517 fs/xfs/libxfs/xfs_rtbitmap.c 	int		log,		/* log2 of extent size */
log               523 fs/xfs/libxfs/xfs_rtbitmap.c 	return xfs_rtmodify_summary_int(mp, tp, log, bbno,
log                29 fs/xfs/xfs_log.c 	struct xlog		*log,
log                42 fs/xfs/xfs_log.c 	struct xlog		*log,
log                46 fs/xfs/xfs_log.c 	struct xlog		*log);
log                54 fs/xfs/xfs_log.c 	struct xlog		*log,
log                62 fs/xfs/xfs_log.c 	struct xlog		*log,
log                66 fs/xfs/xfs_log.c 	struct xlog		*log,
log                71 fs/xfs/xfs_log.c 	struct xlog		*log,
log                76 fs/xfs/xfs_log.c 	struct xlog		*log,
log                80 fs/xfs/xfs_log.c 	struct xlog		*log,
log                84 fs/xfs/xfs_log.c 	struct xlog		*log,
log                90 fs/xfs/xfs_log.c 	struct xlog		*log,
log                94 fs/xfs/xfs_log.c 	struct xlog *log);
log                97 fs/xfs/xfs_log.c 	struct xlog		*log,
log               102 fs/xfs/xfs_log.c 	struct xlog		*log,
log               114 fs/xfs/xfs_log.c 	struct xlog		*log);
log               118 fs/xfs/xfs_log.c 	struct xlog		*log,
log               132 fs/xfs/xfs_log.c 			space += log->l_logsize;
log               144 fs/xfs/xfs_log.c 	struct xlog		*log,
log               157 fs/xfs/xfs_log.c 		tmp = log->l_logsize - space;
log               194 fs/xfs/xfs_log.c 	struct xlog		*log,
log               198 fs/xfs/xfs_log.c 	if (head == &log->l_write_head) {
log               211 fs/xfs/xfs_log.c 	struct xlog		*log,
log               242 fs/xfs/xfs_log.c 		need_bytes = xlog_ticket_reservation(log, head, tic);
log               245 fs/xfs/xfs_log.c 				xlog_grant_push_ail(log, need_bytes);
log               250 fs/xfs/xfs_log.c 		trace_xfs_log_grant_wake_up(log, tic);
log               260 fs/xfs/xfs_log.c 	struct xlog		*log,
log               269 fs/xfs/xfs_log.c 		if (XLOG_FORCED_SHUTDOWN(log))
log               271 fs/xfs/xfs_log.c 		xlog_grant_push_ail(log, need_bytes);
log               276 fs/xfs/xfs_log.c 		XFS_STATS_INC(log->l_mp, xs_sleep_logspace);
log               278 fs/xfs/xfs_log.c 		trace_xfs_log_grant_sleep(log, tic);
log               280 fs/xfs/xfs_log.c 		trace_xfs_log_grant_wake(log, tic);
log               283 fs/xfs/xfs_log.c 		if (XLOG_FORCED_SHUTDOWN(log))
log               285 fs/xfs/xfs_log.c 	} while (xlog_space_left(log, &head->grant) < need_bytes);
log               313 fs/xfs/xfs_log.c 	struct xlog		*log,
log               321 fs/xfs/xfs_log.c 	ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY));
log               329 fs/xfs/xfs_log.c 	*need_bytes = xlog_ticket_reservation(log, head, tic);
log               330 fs/xfs/xfs_log.c 	free_bytes = xlog_space_left(log, &head->grant);
log               333 fs/xfs/xfs_log.c 		if (!xlog_grant_head_wake(log, head, &free_bytes) ||
log               335 fs/xfs/xfs_log.c 			error = xlog_grant_head_wait(log, head, tic,
log               341 fs/xfs/xfs_log.c 		error = xlog_grant_head_wait(log, head, tic, *need_bytes);
log               380 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log               384 fs/xfs/xfs_log.c 	if (XLOG_FORCED_SHUTDOWN(log))
log               397 fs/xfs/xfs_log.c 	xlog_grant_push_ail(log, tic->t_unit_res);
log               405 fs/xfs/xfs_log.c 	trace_xfs_log_regrant(log, tic);
log               407 fs/xfs/xfs_log.c 	error = xlog_grant_head_check(log, &log->l_write_head, tic,
log               412 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_write_head.grant, need_bytes);
log               413 fs/xfs/xfs_log.c 	trace_xfs_log_regrant_exit(log, tic);
log               414 fs/xfs/xfs_log.c 	xlog_verify_grant_tail(log);
log               445 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log               452 fs/xfs/xfs_log.c 	if (XLOG_FORCED_SHUTDOWN(log))
log               458 fs/xfs/xfs_log.c 	tic = xlog_ticket_alloc(log, unit_bytes, cnt, client, permanent, 0);
log               461 fs/xfs/xfs_log.c 	xlog_grant_push_ail(log, tic->t_cnt ? tic->t_unit_res * tic->t_cnt
log               464 fs/xfs/xfs_log.c 	trace_xfs_log_reserve(log, tic);
log               466 fs/xfs/xfs_log.c 	error = xlog_grant_head_check(log, &log->l_reserve_head, tic,
log               471 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_reserve_head.grant, need_bytes);
log               472 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_write_head.grant, need_bytes);
log               473 fs/xfs/xfs_log.c 	trace_xfs_log_reserve_exit(log, tic);
log               474 fs/xfs/xfs_log.c 	xlog_verify_grant_tail(log);
log               517 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log               520 fs/xfs/xfs_log.c 	if (XLOG_FORCED_SHUTDOWN(log) ||
log               526 fs/xfs/xfs_log.c 	     (xlog_commit_record(log, ticket, iclog, &lsn)))) {
log               533 fs/xfs/xfs_log.c 		trace_xfs_log_done_nonperm(log, ticket);
log               539 fs/xfs/xfs_log.c 		xlog_ungrant_log_space(log, ticket);
log               541 fs/xfs/xfs_log.c 		trace_xfs_log_done_perm(log, ticket);
log               543 fs/xfs/xfs_log.c 		xlog_regrant_reserve_log_space(log, ticket);
log               829 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log               856 fs/xfs/xfs_log.c 	error = xlog_write(log, &vec, tic, &lsn, NULL, flags);
log               865 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log               866 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log               868 fs/xfs/xfs_log.c 	xlog_state_want_sync(log, iclog);
log               869 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log               870 fs/xfs/xfs_log.c 	error = xlog_state_release_iclog(log, iclog);
log               872 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log               875 fs/xfs/xfs_log.c 		if (!XLOG_FORCED_SHUTDOWN(log)) {
log               876 fs/xfs/xfs_log.c 			xlog_wait(&iclog->ic_force_wait, &log->l_icloglock);
log               882 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log               887 fs/xfs/xfs_log.c 		trace_xfs_log_umount_write(log, tic);
log               888 fs/xfs/xfs_log.c 		xlog_ungrant_log_space(log, tic);
log               904 fs/xfs/xfs_log.c 	struct xlog	 *log = mp->m_log;
log               916 fs/xfs/xfs_log.c 	    xfs_readonly_buftarg(log->l_targ)) {
log               922 fs/xfs/xfs_log.c 	ASSERT(error || !(XLOG_FORCED_SHUTDOWN(log)));
log               925 fs/xfs/xfs_log.c 	first_iclog = iclog = log->l_iclog;
log               934 fs/xfs/xfs_log.c 	if (! (XLOG_FORCED_SHUTDOWN(log))) {
log               950 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log               951 fs/xfs/xfs_log.c 		iclog = log->l_iclog;
log               954 fs/xfs/xfs_log.c 		xlog_state_want_sync(log, iclog);
log               955 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log               956 fs/xfs/xfs_log.c 		error =  xlog_state_release_iclog(log, iclog);
log               958 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log               965 fs/xfs/xfs_log.c 							&log->l_icloglock);
log               967 fs/xfs/xfs_log.c 			spin_unlock(&log->l_icloglock);
log              1051 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              1054 fs/xfs/xfs_log.c 	if (XLOG_FORCED_SHUTDOWN(log))
log              1057 fs/xfs/xfs_log.c 	if (!list_empty_careful(&log->l_write_head.waiters)) {
log              1058 fs/xfs/xfs_log.c 		ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY));
log              1060 fs/xfs/xfs_log.c 		spin_lock(&log->l_write_head.lock);
log              1061 fs/xfs/xfs_log.c 		free_bytes = xlog_space_left(log, &log->l_write_head.grant);
log              1062 fs/xfs/xfs_log.c 		xlog_grant_head_wake(log, &log->l_write_head, &free_bytes);
log              1063 fs/xfs/xfs_log.c 		spin_unlock(&log->l_write_head.lock);
log              1066 fs/xfs/xfs_log.c 	if (!list_empty_careful(&log->l_reserve_head.waiters)) {
log              1067 fs/xfs/xfs_log.c 		ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY));
log              1069 fs/xfs/xfs_log.c 		spin_lock(&log->l_reserve_head.lock);
log              1070 fs/xfs/xfs_log.c 		free_bytes = xlog_space_left(log, &log->l_reserve_head.grant);
log              1071 fs/xfs/xfs_log.c 		xlog_grant_head_wake(log, &log->l_reserve_head, &free_bytes);
log              1072 fs/xfs/xfs_log.c 		spin_unlock(&log->l_reserve_head.lock);
log              1096 fs/xfs/xfs_log.c 	struct xlog	*log = mp->m_log;
log              1102 fs/xfs/xfs_log.c 	if (!xlog_cil_empty(log))
log              1105 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              1106 fs/xfs/xfs_log.c 	switch (log->l_covered_state) {
log              1113 fs/xfs/xfs_log.c 		if (xfs_ail_min_lsn(log->l_ailp))
log              1115 fs/xfs/xfs_log.c 		if (!xlog_iclogs_empty(log))
log              1119 fs/xfs/xfs_log.c 		if (log->l_covered_state == XLOG_STATE_COVER_NEED)
log              1120 fs/xfs/xfs_log.c 			log->l_covered_state = XLOG_STATE_COVER_DONE;
log              1122 fs/xfs/xfs_log.c 			log->l_covered_state = XLOG_STATE_COVER_DONE2;
log              1128 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              1139 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              1154 fs/xfs/xfs_log.c 		tail_lsn = atomic64_read(&log->l_last_sync_lsn);
log              1155 fs/xfs/xfs_log.c 	trace_xfs_log_assign_tail_lsn(log, tail_lsn);
log              1156 fs/xfs/xfs_log.c 	atomic64_set(&log->l_tail_lsn, tail_lsn);
log              1189 fs/xfs/xfs_log.c 	struct xlog	*log,
log              1199 fs/xfs/xfs_log.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_bytes);
log              1202 fs/xfs/xfs_log.c 		free_bytes = log->l_logsize - (head_bytes - tail_bytes);
log              1214 fs/xfs/xfs_log.c 		xfs_alert(log->l_mp, "xlog_space_left: head behind tail");
log              1215 fs/xfs/xfs_log.c 		xfs_alert(log->l_mp,
log              1218 fs/xfs/xfs_log.c 		xfs_alert(log->l_mp,
log              1222 fs/xfs/xfs_log.c 		free_bytes = log->l_logsize;
log              1234 fs/xfs/xfs_log.c 	struct xlog		*log = iclog->ic_log;
log              1248 fs/xfs/xfs_log.c 	if (XFS_TEST_ERROR(error, log->l_mp, XFS_ERRTAG_IODONE_IOERR)) {
log              1249 fs/xfs/xfs_log.c 		xfs_alert(log->l_mp, "log I/O error %d", error);
log              1250 fs/xfs/xfs_log.c 		xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR);
log              1284 fs/xfs/xfs_log.c 	struct xlog		*log)
log              1291 fs/xfs/xfs_log.c 	log->l_iclog_bufs = mp->m_logbufs;
log              1292 fs/xfs/xfs_log.c 	log->l_iclog_size = mp->m_logbsize;
log              1297 fs/xfs/xfs_log.c 	log->l_iclog_heads =
log              1299 fs/xfs/xfs_log.c 	log->l_iclog_hsize = log->l_iclog_heads << BBSHIFT;
log              1319 fs/xfs/xfs_log.c 	struct xlog		*log = container_of(to_delayed_work(work),
log              1321 fs/xfs/xfs_log.c 	struct xfs_mount	*mp = log->l_mp;
log              1359 fs/xfs/xfs_log.c 	struct xlog		*log;
log              1367 fs/xfs/xfs_log.c 	log = kmem_zalloc(sizeof(struct xlog), KM_MAYFAIL);
log              1368 fs/xfs/xfs_log.c 	if (!log) {
log              1373 fs/xfs/xfs_log.c 	log->l_mp	   = mp;
log              1374 fs/xfs/xfs_log.c 	log->l_targ	   = log_target;
log              1375 fs/xfs/xfs_log.c 	log->l_logsize     = BBTOB(num_bblks);
log              1376 fs/xfs/xfs_log.c 	log->l_logBBstart  = blk_offset;
log              1377 fs/xfs/xfs_log.c 	log->l_logBBsize   = num_bblks;
log              1378 fs/xfs/xfs_log.c 	log->l_covered_state = XLOG_STATE_COVER_IDLE;
log              1379 fs/xfs/xfs_log.c 	log->l_flags	   |= XLOG_ACTIVE_RECOVERY;
log              1380 fs/xfs/xfs_log.c 	INIT_DELAYED_WORK(&log->l_work, xfs_log_worker);
log              1382 fs/xfs/xfs_log.c 	log->l_prev_block  = -1;
log              1384 fs/xfs/xfs_log.c 	xlog_assign_atomic_lsn(&log->l_tail_lsn, 1, 0);
log              1385 fs/xfs/xfs_log.c 	xlog_assign_atomic_lsn(&log->l_last_sync_lsn, 1, 0);
log              1386 fs/xfs/xfs_log.c 	log->l_curr_cycle  = 1;	    /* 0 is bad since this is initial value */
log              1388 fs/xfs/xfs_log.c 	xlog_grant_head_init(&log->l_reserve_head);
log              1389 fs/xfs/xfs_log.c 	xlog_grant_head_init(&log->l_write_head);
log              1408 fs/xfs/xfs_log.c 		if (log2_size && log->l_logBBstart > 0 &&
log              1416 fs/xfs/xfs_log.c 	log->l_sectBBsize = 1 << log2_size;
log              1418 fs/xfs/xfs_log.c 	xlog_get_iclog_buffer_size(mp, log);
log              1420 fs/xfs/xfs_log.c 	spin_lock_init(&log->l_icloglock);
log              1421 fs/xfs/xfs_log.c 	init_waitqueue_head(&log->l_flush_wait);
log              1423 fs/xfs/xfs_log.c 	iclogp = &log->l_iclog;
log              1431 fs/xfs/xfs_log.c 	ASSERT(log->l_iclog_size >= 4096);
log              1432 fs/xfs/xfs_log.c 	for (i = 0; i < log->l_iclog_bufs; i++) {
log              1434 fs/xfs/xfs_log.c 		size_t bvec_size = howmany(log->l_iclog_size, PAGE_SIZE) *
log              1445 fs/xfs/xfs_log.c 		iclog->ic_data = kmem_alloc_io(log->l_iclog_size, align_mask,
log              1450 fs/xfs/xfs_log.c 		log->l_iclog_bak[i] = &iclog->ic_header;
log              1456 fs/xfs/xfs_log.c 			xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? 2 : 1);
log              1457 fs/xfs/xfs_log.c 		head->h_size = cpu_to_be32(log->l_iclog_size);
log              1462 fs/xfs/xfs_log.c 		iclog->ic_size = log->l_iclog_size - log->l_iclog_hsize;
log              1464 fs/xfs/xfs_log.c 		iclog->ic_log = log;
log              1468 fs/xfs/xfs_log.c 		iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize;
log              1477 fs/xfs/xfs_log.c 	*iclogp = log->l_iclog;			/* complete ring */
log              1478 fs/xfs/xfs_log.c 	log->l_iclog->ic_prev = prev_iclog;	/* re-write 1st prev ptr */
log              1480 fs/xfs/xfs_log.c 	log->l_ioend_workqueue = alloc_workqueue("xfs-log/%s",
log              1483 fs/xfs/xfs_log.c 	if (!log->l_ioend_workqueue)
log              1486 fs/xfs/xfs_log.c 	error = xlog_cil_init(log);
log              1489 fs/xfs/xfs_log.c 	return log;
log              1492 fs/xfs/xfs_log.c 	destroy_workqueue(log->l_ioend_workqueue);
log              1494 fs/xfs/xfs_log.c 	for (iclog = log->l_iclog; iclog; iclog = prev_iclog) {
log              1498 fs/xfs/xfs_log.c 		if (prev_iclog == log->l_iclog)
log              1502 fs/xfs/xfs_log.c 	kmem_free(log);
log              1514 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1519 fs/xfs/xfs_log.c 	struct xfs_mount *mp = log->l_mp;
log              1532 fs/xfs/xfs_log.c 	error = xlog_write(log, &vec, ticket, commitlsnp, iclog,
log              1548 fs/xfs/xfs_log.c 	struct xlog	*log,
log              1559 fs/xfs/xfs_log.c 	ASSERT(BTOBB(need_bytes) < log->l_logBBsize);
log              1561 fs/xfs/xfs_log.c 	free_bytes = xlog_space_left(log, &log->l_reserve_head.grant);
log              1570 fs/xfs/xfs_log.c 	free_threshold = max(free_threshold, (log->l_logBBsize >> 2));
log              1575 fs/xfs/xfs_log.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &threshold_cycle,
log              1578 fs/xfs/xfs_log.c 	if (threshold_block >= log->l_logBBsize) {
log              1579 fs/xfs/xfs_log.c 		threshold_block -= log->l_logBBsize;
log              1589 fs/xfs/xfs_log.c 	last_sync_lsn = atomic64_read(&log->l_last_sync_lsn);
log              1598 fs/xfs/xfs_log.c 	if (!XLOG_FORCED_SHUTDOWN(log))
log              1599 fs/xfs/xfs_log.c 		xfs_ail_push(log->l_ailp, threshold_lsn);
log              1607 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1627 fs/xfs/xfs_log.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log              1638 fs/xfs/xfs_log.c 		for (i = 1; i < log->l_iclog_heads; i++)
log              1651 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1664 fs/xfs/xfs_log.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log              1715 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1721 fs/xfs/xfs_log.c 	ASSERT(bno < log->l_logBBsize);
log              1748 fs/xfs/xfs_log.c 	bio_set_dev(&iclog->ic_bio, log->l_targ->bt_bdev);
log              1749 fs/xfs/xfs_log.c 	iclog->ic_bio.bi_iter.bi_sector = log->l_logBBstart + bno;
log              1764 fs/xfs/xfs_log.c 	if (bno + BTOBB(count) > log->l_logBBsize) {
log              1767 fs/xfs/xfs_log.c 		split = bio_split(&iclog->ic_bio, log->l_logBBsize - bno,
log              1773 fs/xfs/xfs_log.c 		iclog->ic_bio.bi_iter.bi_sector = log->l_logBBstart;
log              1786 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1791 fs/xfs/xfs_log.c 	unsigned int		split_offset = BBTOB(log->l_logBBsize - bno);
log              1805 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1812 fs/xfs/xfs_log.c 	use_lsunit = xfs_sb_version_haslogv2(&log->l_mp->m_sb) &&
log              1813 fs/xfs/xfs_log.c 			log->l_mp->m_sb.sb_logsunit > 1;
log              1816 fs/xfs/xfs_log.c 	count_init = log->l_iclog_hsize + iclog->ic_offset;
log              1821 fs/xfs/xfs_log.c 		count = XLOG_LSUNITTOB(log, XLOG_BTOLSUNIT(log, count_init));
log              1830 fs/xfs/xfs_log.c 		ASSERT(*roundoff < log->l_mp->m_sb.sb_logsunit);
log              1861 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1872 fs/xfs/xfs_log.c 	count = xlog_calc_iclog_size(log, iclog, &roundoff);
log              1875 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_reserve_head.grant, roundoff);
log              1876 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_write_head.grant, roundoff);
log              1879 fs/xfs/xfs_log.c 	xlog_pack_data(log, iclog, roundoff); 
log              1883 fs/xfs/xfs_log.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb))
log              1887 fs/xfs/xfs_log.c 	XFS_STATS_INC(log->l_mp, xs_log_writes);
log              1888 fs/xfs/xfs_log.c 	XFS_STATS_ADD(log->l_mp, xs_log_blocks, BTOBB(count));
log              1893 fs/xfs/xfs_log.c 	if (bno + BTOBB(count) > log->l_logBBsize) {
log              1894 fs/xfs/xfs_log.c 		xlog_split_iclog(log, &iclog->ic_header, bno, count);
log              1899 fs/xfs/xfs_log.c 	iclog->ic_header.h_crc = xlog_cksum(log, &iclog->ic_header,
log              1909 fs/xfs/xfs_log.c 	if (XFS_TEST_ERROR(false, log->l_mp, XFS_ERRTAG_LOG_BAD_CRC)) {
log              1912 fs/xfs/xfs_log.c 		xfs_warn(log->l_mp,
log              1926 fs/xfs/xfs_log.c 	if (log->l_targ != log->l_mp->m_ddev_targp || split) {
log              1927 fs/xfs/xfs_log.c 		xfs_blkdev_issue_flush(log->l_mp->m_ddev_targp);
log              1931 fs/xfs/xfs_log.c 	xlog_verify_iclog(log, iclog, count);
log              1932 fs/xfs/xfs_log.c 	xlog_write_iclog(log, iclog, bno, count, need_flush);
log              1940 fs/xfs/xfs_log.c 	struct xlog	*log)
log              1945 fs/xfs/xfs_log.c 	xlog_cil_destroy(log);
log              1951 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              1952 fs/xfs/xfs_log.c 	for (i = 0; i < log->l_iclog_bufs; i++) {
log              1958 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              1959 fs/xfs/xfs_log.c 	for (i = 0; i < log->l_iclog_bufs; i++) {
log              1966 fs/xfs/xfs_log.c 	log->l_mp->m_log = NULL;
log              1967 fs/xfs/xfs_log.c 	destroy_workqueue(log->l_ioend_workqueue);
log              1968 fs/xfs/xfs_log.c 	kmem_free(log);
log              1977 fs/xfs/xfs_log.c 	struct xlog		*log,
log              1982 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              1987 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              2175 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2198 fs/xfs/xfs_log.c 		xfs_warn(log->l_mp,
log              2258 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2273 fs/xfs/xfs_log.c 		xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt);
log              2276 fs/xfs/xfs_log.c 		return xlog_state_release_iclog(log, iclog);
log              2284 fs/xfs/xfs_log.c 		xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt);
log              2288 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log              2289 fs/xfs/xfs_log.c 		xlog_state_want_sync(log, iclog);
log              2290 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              2293 fs/xfs/xfs_log.c 			return xlog_state_release_iclog(log, iclog);
log              2343 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2382 fs/xfs/xfs_log.c 		xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES,
log              2384 fs/xfs/xfs_log.c 		xlog_print_tic_res(log->l_mp, ticket);
log              2385 fs/xfs/xfs_log.c 		xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR);
log              2395 fs/xfs/xfs_log.c 		error = xlog_state_get_iclog_space(log, len, &iclog, ticket,
log              2437 fs/xfs/xfs_log.c 			ophdr = xlog_write_setup_ophdr(log, ptr, ticket, flags);
log              2450 fs/xfs/xfs_log.c 			xlog_verify_dest_ptr(log, ptr);
log              2470 fs/xfs/xfs_log.c 			error = xlog_write_copy_finish(log, iclog, flags,
log              2511 fs/xfs/xfs_log.c 	xlog_state_finish_copy(log, iclog, record_cnt, data_cnt);
log              2513 fs/xfs/xfs_log.c 		return xlog_state_release_iclog(log, iclog);
log              2546 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2557 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              2592 fs/xfs/xfs_log.c 	} while (iclog != log->l_iclog);
log              2609 fs/xfs/xfs_log.c 		switch (log->l_covered_state) {
log              2613 fs/xfs/xfs_log.c 			log->l_covered_state = XLOG_STATE_COVER_NEED;
log              2618 fs/xfs/xfs_log.c 				log->l_covered_state = XLOG_STATE_COVER_NEED2;
log              2620 fs/xfs/xfs_log.c 				log->l_covered_state = XLOG_STATE_COVER_NEED;
log              2625 fs/xfs/xfs_log.c 				log->l_covered_state = XLOG_STATE_COVER_IDLE;
log              2627 fs/xfs/xfs_log.c 				log->l_covered_state = XLOG_STATE_COVER_NEED;
log              2638 fs/xfs/xfs_log.c 	struct xlog		*log)
log              2640 fs/xfs/xfs_log.c 	struct xlog_in_core	*iclog = log->l_iclog;
log              2650 fs/xfs/xfs_log.c 	} while ((iclog = iclog->ic_next) != log->l_iclog);
log              2679 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2685 fs/xfs/xfs_log.c 	ASSERT(XFS_LSN_CMP(atomic64_read(&log->l_last_sync_lsn),
log              2691 fs/xfs/xfs_log.c 	atomic64_set(&log->l_last_sync_lsn, header_lsn);
log              2692 fs/xfs/xfs_log.c 	xlog_grant_push_ail(log, 0);
log              2702 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2752 fs/xfs/xfs_log.c 	lowest_lsn = xlog_get_lowest_lsn(log);
log              2756 fs/xfs/xfs_log.c 	xlog_state_set_callback(log, iclog, header_lsn);
log              2772 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2776 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              2793 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              2811 fs/xfs/xfs_log.c 	struct xlog		*log)
log              2813 fs/xfs/xfs_log.c 	struct xlog_in_core	*first_iclog = log->l_iclog;
log              2841 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2853 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              2863 fs/xfs/xfs_log.c 		first_iclog = log->l_iclog;
log              2864 fs/xfs/xfs_log.c 		iclog = log->l_iclog;
log              2870 fs/xfs/xfs_log.c 			if (xlog_state_iodone_process_iclog(log, iclog,
log              2885 fs/xfs/xfs_log.c 			xlog_state_do_iclog_callbacks(log, iclog, aborted);
log              2887 fs/xfs/xfs_log.c 			xlog_state_clean_iclog(log, iclog);
log              2896 fs/xfs/xfs_log.c 			xfs_warn(log->l_mp,
log              2903 fs/xfs/xfs_log.c 		xlog_state_callback_check_state(log);
log              2905 fs/xfs/xfs_log.c 	if (log->l_iclog->ic_state & (XLOG_STATE_ACTIVE|XLOG_STATE_IOERROR))
log              2906 fs/xfs/xfs_log.c 		wake_up_all(&log->l_flush_wait);
log              2908 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              2930 fs/xfs/xfs_log.c 	struct xlog		*log = iclog->ic_log;
log              2932 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              2953 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              2954 fs/xfs/xfs_log.c 	xlog_state_do_callback(log, aborted, iclog);	/* also cleans log */
log              2978 fs/xfs/xfs_log.c 	struct xlog		*log,
log              2991 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              2992 fs/xfs/xfs_log.c 	if (XLOG_FORCED_SHUTDOWN(log)) {
log              2993 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              2997 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              2999 fs/xfs/xfs_log.c 		XFS_STATS_INC(log->l_mp, xs_log_noiclogs);
log              3002 fs/xfs/xfs_log.c 		xlog_wait(&log->l_flush_wait, &log->l_icloglock);
log              3017 fs/xfs/xfs_log.c 		ticket->t_curr_res -= log->l_iclog_hsize;
log              3019 fs/xfs/xfs_log.c 				    log->l_iclog_hsize,
log              3021 fs/xfs/xfs_log.c 		head->h_cycle = cpu_to_be32(log->l_curr_cycle);
log              3023 fs/xfs/xfs_log.c 			xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block));
log              3024 fs/xfs/xfs_log.c 		ASSERT(log->l_curr_block >= 0);
log              3037 fs/xfs/xfs_log.c 		xlog_state_switch_iclogs(log, iclog, iclog->ic_size);
log              3048 fs/xfs/xfs_log.c 			spin_unlock(&log->l_icloglock);
log              3049 fs/xfs/xfs_log.c 			error = xlog_state_release_iclog(log, iclog);
log              3053 fs/xfs/xfs_log.c 			spin_unlock(&log->l_icloglock);
log              3069 fs/xfs/xfs_log.c 		xlog_state_switch_iclogs(log, iclog, iclog->ic_size);
log              3074 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3089 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3092 fs/xfs/xfs_log.c 	trace_xfs_log_regrant_reserve_enter(log, ticket);
log              3097 fs/xfs/xfs_log.c 	xlog_grant_sub_space(log, &log->l_reserve_head.grant,
log              3099 fs/xfs/xfs_log.c 	xlog_grant_sub_space(log, &log->l_write_head.grant,
log              3104 fs/xfs/xfs_log.c 	trace_xfs_log_regrant_reserve_sub(log, ticket);
log              3110 fs/xfs/xfs_log.c 	xlog_grant_add_space(log, &log->l_reserve_head.grant,
log              3113 fs/xfs/xfs_log.c 	trace_xfs_log_regrant_reserve_exit(log, ticket);
log              3136 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3144 fs/xfs/xfs_log.c 	trace_xfs_log_ungrant_enter(log, ticket);
log              3145 fs/xfs/xfs_log.c 	trace_xfs_log_ungrant_sub(log, ticket);
log              3157 fs/xfs/xfs_log.c 	xlog_grant_sub_space(log, &log->l_reserve_head.grant, bytes);
log              3158 fs/xfs/xfs_log.c 	xlog_grant_sub_space(log, &log->l_write_head.grant, bytes);
log              3160 fs/xfs/xfs_log.c 	trace_xfs_log_ungrant_exit(log, ticket);
log              3162 fs/xfs/xfs_log.c 	xfs_log_space_wake(log->l_mp);
log              3176 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3185 fs/xfs/xfs_log.c 	if (!atomic_dec_and_lock(&iclog->ic_refcnt, &log->l_icloglock))
log              3189 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              3197 fs/xfs/xfs_log.c 		xfs_lsn_t tail_lsn = xlog_assign_tail_lsn(log->l_mp);
log              3201 fs/xfs/xfs_log.c 		xlog_verify_tail_lsn(log, iclog, tail_lsn);
log              3204 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3214 fs/xfs/xfs_log.c 		xlog_sync(log, iclog);
log              3228 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3236 fs/xfs/xfs_log.c 	iclog->ic_header.h_prev_block = cpu_to_be32(log->l_prev_block);
log              3237 fs/xfs/xfs_log.c 	log->l_prev_block = log->l_curr_block;
log              3238 fs/xfs/xfs_log.c 	log->l_prev_cycle = log->l_curr_cycle;
log              3241 fs/xfs/xfs_log.c 	log->l_curr_block += BTOBB(eventual_size)+BTOBB(log->l_iclog_hsize);
log              3244 fs/xfs/xfs_log.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb) &&
log              3245 fs/xfs/xfs_log.c 	    log->l_mp->m_sb.sb_logsunit > 1) {
log              3246 fs/xfs/xfs_log.c 		uint32_t sunit_bb = BTOBB(log->l_mp->m_sb.sb_logsunit);
log              3247 fs/xfs/xfs_log.c 		log->l_curr_block = roundup(log->l_curr_block, sunit_bb);
log              3250 fs/xfs/xfs_log.c 	if (log->l_curr_block >= log->l_logBBsize) {
log              3258 fs/xfs/xfs_log.c 		log->l_curr_block -= log->l_logBBsize;
log              3259 fs/xfs/xfs_log.c 		ASSERT(log->l_curr_block >= 0);
log              3261 fs/xfs/xfs_log.c 		log->l_curr_cycle++;
log              3262 fs/xfs/xfs_log.c 		if (log->l_curr_cycle == XLOG_HEADER_MAGIC_NUM)
log              3263 fs/xfs/xfs_log.c 			log->l_curr_cycle++;
log              3265 fs/xfs/xfs_log.c 	ASSERT(iclog == log->l_iclog);
log              3266 fs/xfs/xfs_log.c 	log->l_iclog = iclog->ic_next;
log              3301 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              3308 fs/xfs/xfs_log.c 	xlog_cil_force(log);
log              3310 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              3311 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              3341 fs/xfs/xfs_log.c 			xlog_state_switch_iclogs(log, iclog, 0);
log              3342 fs/xfs/xfs_log.c 			spin_unlock(&log->l_icloglock);
log              3344 fs/xfs/xfs_log.c 			if (xlog_state_release_iclog(log, iclog))
log              3347 fs/xfs/xfs_log.c 			spin_lock(&log->l_icloglock);
log              3359 fs/xfs/xfs_log.c 			xlog_state_switch_iclogs(log, iclog, 0);
log              3375 fs/xfs/xfs_log.c 	xlog_wait(&iclog->ic_force_wait, &log->l_icloglock);
log              3381 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3384 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3396 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              3399 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              3400 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              3406 fs/xfs/xfs_log.c 		if (iclog == log->l_iclog)
log              3437 fs/xfs/xfs_log.c 					&log->l_icloglock);
log              3441 fs/xfs/xfs_log.c 		xlog_state_switch_iclogs(log, iclog, 0);
log              3442 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              3443 fs/xfs/xfs_log.c 		if (xlog_state_release_iclog(log, iclog))
log              3447 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log              3458 fs/xfs/xfs_log.c 	xlog_wait(&iclog->ic_force_wait, &log->l_icloglock);
log              3464 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3467 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3514 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3517 fs/xfs/xfs_log.c 	assert_spin_locked(&log->l_icloglock);
log              3520 fs/xfs/xfs_log.c 		xlog_state_switch_iclogs(log, iclog, 0);
log              3565 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              3624 fs/xfs/xfs_log.c 	iclog_space = log->l_iclog_size - log->l_iclog_hsize;
log              3636 fs/xfs/xfs_log.c 	unit_bytes += log->l_iclog_hsize * num_headers;
log              3639 fs/xfs/xfs_log.c 	unit_bytes += log->l_iclog_hsize;
log              3658 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3672 fs/xfs/xfs_log.c 	unit_res = xfs_log_calc_unit_res(log->l_mp, unit_bytes);
log              3707 fs/xfs/xfs_log.c 	struct xlog	*log,
log              3713 fs/xfs/xfs_log.c 	for (i = 0; i < log->l_iclog_bufs; i++) {
log              3714 fs/xfs/xfs_log.c 		if (ptr >= log->l_iclog_bak[i] &&
log              3715 fs/xfs/xfs_log.c 		    ptr <= log->l_iclog_bak[i] + log->l_iclog_size)
log              3720 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: invalid ptr", __func__);
log              3736 fs/xfs/xfs_log.c 	struct xlog	*log)
log              3741 fs/xfs/xfs_log.c 	xlog_crack_grant_head(&log->l_write_head.grant, &cycle, &space);
log              3742 fs/xfs/xfs_log.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_blocks);
log              3745 fs/xfs/xfs_log.c 		    !(log->l_flags & XLOG_TAIL_WARN)) {
log              3746 fs/xfs/xfs_log.c 			xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES,
log              3748 fs/xfs/xfs_log.c 			log->l_flags |= XLOG_TAIL_WARN;
log              3752 fs/xfs/xfs_log.c 		    !(log->l_flags & XLOG_TAIL_WARN)) {
log              3753 fs/xfs/xfs_log.c 			xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES,
log              3755 fs/xfs/xfs_log.c 			log->l_flags |= XLOG_TAIL_WARN;
log              3763 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3769 fs/xfs/xfs_log.c     if (CYCLE_LSN(tail_lsn) == log->l_prev_cycle) {
log              3771 fs/xfs/xfs_log.c 	    log->l_logBBsize - (log->l_prev_block - BLOCK_LSN(tail_lsn));
log              3772 fs/xfs/xfs_log.c 	if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize))
log              3773 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: ran out of log space", __func__);
log              3775 fs/xfs/xfs_log.c 	ASSERT(CYCLE_LSN(tail_lsn)+1 == log->l_prev_cycle);
log              3777 fs/xfs/xfs_log.c 	if (BLOCK_LSN(tail_lsn) == log->l_prev_block)
log              3778 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: tail wrapped", __func__);
log              3780 fs/xfs/xfs_log.c 	blocks = BLOCK_LSN(tail_lsn) - log->l_prev_block;
log              3782 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: ran out of log space", __func__);
log              3803 fs/xfs/xfs_log.c 	struct xlog		*log,
log              3817 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              3818 fs/xfs/xfs_log.c 	icptr = log->l_iclog;
log              3819 fs/xfs/xfs_log.c 	for (i = 0; i < log->l_iclog_bufs; i++, icptr = icptr->ic_next)
log              3822 fs/xfs/xfs_log.c 	if (icptr != log->l_iclog)
log              3823 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: corrupt iclog ring", __func__);
log              3824 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3828 fs/xfs/xfs_log.c 		xfs_emerg(log->l_mp, "%s: invalid magic num", __func__);
log              3834 fs/xfs/xfs_log.c 			xfs_emerg(log->l_mp, "%s: unexpected magic num",
log              3864 fs/xfs/xfs_log.c 			xfs_warn(log->l_mp,
log              3895 fs/xfs/xfs_log.c 	struct xlog	*log)
log              3899 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              3939 fs/xfs/xfs_log.c 	struct xlog	*log;
log              3942 fs/xfs/xfs_log.c 	log = mp->m_log;
log              3948 fs/xfs/xfs_log.c 	if (!log ||
log              3949 fs/xfs/xfs_log.c 	    log->l_flags & XLOG_ACTIVE_RECOVERY) {
log              3960 fs/xfs/xfs_log.c 	if (logerror && log->l_iclog->ic_state & XLOG_STATE_IOERROR) {
log              3961 fs/xfs/xfs_log.c 		ASSERT(XLOG_FORCED_SHUTDOWN(log));
log              3979 fs/xfs/xfs_log.c 	spin_lock(&log->l_icloglock);
log              3988 fs/xfs/xfs_log.c 	log->l_flags |= XLOG_IO_ERROR;
log              3989 fs/xfs/xfs_log.c 	retval = xlog_state_ioerror(log);
log              3990 fs/xfs/xfs_log.c 	spin_unlock(&log->l_icloglock);
log              3999 fs/xfs/xfs_log.c 	xlog_grant_head_wake_all(&log->l_reserve_head);
log              4000 fs/xfs/xfs_log.c 	xlog_grant_head_wake_all(&log->l_write_head);
log              4008 fs/xfs/xfs_log.c 	spin_lock(&log->l_cilp->xc_push_lock);
log              4009 fs/xfs/xfs_log.c 	wake_up_all(&log->l_cilp->xc_commit_wait);
log              4010 fs/xfs/xfs_log.c 	spin_unlock(&log->l_cilp->xc_push_lock);
log              4011 fs/xfs/xfs_log.c 	xlog_state_do_callback(log, true, NULL);
log              4017 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log              4018 fs/xfs/xfs_log.c 		iclog = log->l_iclog;
log              4022 fs/xfs/xfs_log.c 		} while (iclog != log->l_iclog);
log              4023 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              4032 fs/xfs/xfs_log.c 	struct xlog	*log)
log              4036 fs/xfs/xfs_log.c 	iclog = log->l_iclog;
log              4044 fs/xfs/xfs_log.c 	} while (iclog != log->l_iclog);
log              4057 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              4079 fs/xfs/xfs_log.c 		spin_lock(&log->l_icloglock);
log              4084 fs/xfs/xfs_log.c 			 log->l_curr_cycle, log->l_curr_block);
log              4085 fs/xfs/xfs_log.c 		spin_unlock(&log->l_icloglock);
log              4095 fs/xfs/xfs_log.c 	struct xlog		*log = mp->m_log;
log              4097 fs/xfs/xfs_log.c 	return log->l_flags & XLOG_ACTIVE_RECOVERY;
log                36 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log                40 fs/xfs/xfs_log_cil.c 	tic = xlog_ticket_alloc(log, 0, 1, XFS_TRANSACTION, 0,
log                63 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log                65 fs/xfs/xfs_log_cil.c 	log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log);
log                66 fs/xfs/xfs_log_cil.c 	log->l_cilp->xc_ctx->sequence = 1;
log               126 fs/xfs/xfs_log_cil.c 	struct xlog		*log,
log               226 fs/xfs/xfs_log_cil.c 	struct xlog		*log,
log               267 fs/xfs/xfs_log_cil.c 		lv->lv_item->li_seq = log->l_cilp->xc_ctx->sequence;
log               300 fs/xfs/xfs_log_cil.c 	struct xlog		*log,
log               375 fs/xfs/xfs_log_cil.c 		xfs_cil_prepare_item(log, lv, old_lv, diff_len, diff_iovecs);
log               388 fs/xfs/xfs_log_cil.c 	struct xlog		*log,
log               391 fs/xfs/xfs_log_cil.c 	struct xfs_cil		*cil = log->l_cilp;
log               405 fs/xfs/xfs_log_cil.c 	xlog_cil_insert_format_items(log, tp, &len, &diff_iovecs);
log               432 fs/xfs/xfs_log_cil.c 	iclog_space = log->l_iclog_size - log->l_iclog_hsize;
log               437 fs/xfs/xfs_log_cil.c 		split_res *= log->l_iclog_hsize + sizeof(struct xlog_op_header);
log               451 fs/xfs/xfs_log_cil.c 		xfs_warn(log->l_mp, "Transaction log reservation overrun:");
log               452 fs/xfs/xfs_log_cil.c 		xfs_warn(log->l_mp,
log               455 fs/xfs/xfs_log_cil.c 		xfs_warn(log->l_mp, "  split region headers: %d bytes",
log               457 fs/xfs/xfs_log_cil.c 		xfs_warn(log->l_mp, "  ctx ticket: %d bytes", ctx_res);
log               484 fs/xfs/xfs_log_cil.c 		xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR);
log               644 fs/xfs/xfs_log_cil.c 	struct xlog		*log)
log               646 fs/xfs/xfs_log_cil.c 	struct xfs_cil		*cil = log->l_cilp;
log               664 fs/xfs/xfs_log_cil.c 	new_ctx->ticket = xlog_cil_ticket_alloc(log);
log               806 fs/xfs/xfs_log_cil.c 	error = xlog_write(log, &lvhdr, tic, &ctx->start_lsn, NULL, 0);
log               822 fs/xfs/xfs_log_cil.c 		if (XLOG_FORCED_SHUTDOWN(log)) {
log               845 fs/xfs/xfs_log_cil.c 	commit_lsn = xfs_log_done(log->l_mp, tic, &commit_iclog, false);
log               870 fs/xfs/xfs_log_cil.c 	return xfs_log_release_iclog(log->l_mp, commit_iclog);
log               903 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log               905 fs/xfs/xfs_log_cil.c 	struct xfs_cil	*cil = log->l_cilp;
log               917 fs/xfs/xfs_log_cil.c 	if (cil->xc_ctx->space_used < XLOG_CIL_SPACE_LIMIT(log))
log               923 fs/xfs/xfs_log_cil.c 		queue_work(log->l_mp->m_cil_workqueue, &cil->xc_push_work);
log               937 fs/xfs/xfs_log_cil.c 	struct xlog	*log,
log               940 fs/xfs/xfs_log_cil.c 	struct xfs_cil	*cil = log->l_cilp;
log               961 fs/xfs/xfs_log_cil.c 	queue_work(log->l_mp->m_cil_workqueue, &cil->xc_push_work);
log               967 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log               969 fs/xfs/xfs_log_cil.c 	struct xfs_cil	*cil = log->l_cilp;
log               999 fs/xfs/xfs_log_cil.c 	struct xlog		*log = mp->m_log;
log              1000 fs/xfs/xfs_log_cil.c 	struct xfs_cil		*cil = log->l_cilp;
log              1009 fs/xfs/xfs_log_cil.c 	xlog_cil_alloc_shadow_bufs(log, tp);
log              1014 fs/xfs/xfs_log_cil.c 	xlog_cil_insert_items(log, tp);
log              1041 fs/xfs/xfs_log_cil.c 	xlog_cil_push_background(log);
log              1058 fs/xfs/xfs_log_cil.c 	struct xlog	*log,
log              1061 fs/xfs/xfs_log_cil.c 	struct xfs_cil		*cil = log->l_cilp;
log              1073 fs/xfs/xfs_log_cil.c 	xlog_cil_push_now(log, sequence);
log              1088 fs/xfs/xfs_log_cil.c 		if (XLOG_FORCED_SHUTDOWN(log))
log              1177 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log              1207 fs/xfs/xfs_log_cil.c 	cil->xc_log = log;
log              1208 fs/xfs/xfs_log_cil.c 	log->l_cilp = cil;
log              1214 fs/xfs/xfs_log_cil.c 	struct xlog	*log)
log              1216 fs/xfs/xfs_log_cil.c 	if (log->l_cilp->xc_ctx) {
log              1217 fs/xfs/xfs_log_cil.c 		if (log->l_cilp->xc_ctx->ticket)
log              1218 fs/xfs/xfs_log_cil.c 			xfs_log_ticket_put(log->l_cilp->xc_ctx->ticket);
log              1219 fs/xfs/xfs_log_cil.c 		kmem_free(log->l_cilp->xc_ctx);
log              1222 fs/xfs/xfs_log_cil.c 	ASSERT(list_empty(&log->l_cilp->xc_cil));
log              1223 fs/xfs/xfs_log_cil.c 	kmem_free(log->l_cilp);
log               332 fs/xfs/xfs_log_priv.h #define XLOG_CIL_SPACE_LIMIT(log)	(log->l_logsize >> 3)
log               409 fs/xfs/xfs_log_priv.h #define XLOG_BUF_CANCEL_BUCKET(log, blkno) \
log               410 fs/xfs/xfs_log_priv.h 	((log)->l_buf_cancel_table + ((uint64_t)blkno % XLOG_BC_TABLE_SIZE))
log               412 fs/xfs/xfs_log_priv.h #define XLOG_FORCED_SHUTDOWN(log)	((log)->l_flags & XLOG_IO_ERROR)
log               417 fs/xfs/xfs_log_priv.h 	struct xlog		*log);
log               420 fs/xfs/xfs_log_priv.h 	struct xlog		*log);
log               424 fs/xfs/xfs_log_priv.h extern __le32	 xlog_cksum(struct xlog *log, struct xlog_rec_header *rhead,
log               430 fs/xfs/xfs_log_priv.h 	struct xlog	*log,
log               450 fs/xfs/xfs_log_priv.h 	struct xlog		*log,
log               515 fs/xfs/xfs_log_priv.h int	xlog_cil_init(struct xlog *log);
log               516 fs/xfs/xfs_log_priv.h void	xlog_cil_init_post_recovery(struct xlog *log);
log               517 fs/xfs/xfs_log_priv.h void	xlog_cil_destroy(struct xlog *log);
log               518 fs/xfs/xfs_log_priv.h bool	xlog_cil_empty(struct xlog *log);
log               525 fs/xfs/xfs_log_priv.h 	struct xlog *log,
log               529 fs/xfs/xfs_log_priv.h xlog_cil_force(struct xlog *log)
log               531 fs/xfs/xfs_log_priv.h 	xlog_cil_force_lsn(log, log->l_cilp->xc_current_sequence);
log               564 fs/xfs/xfs_log_priv.h 	struct xlog	*log,
log               583 fs/xfs/xfs_log_priv.h 	cur_cycle = READ_ONCE(log->l_curr_cycle);
log               585 fs/xfs/xfs_log_priv.h 	cur_block = READ_ONCE(log->l_curr_block);
log               594 fs/xfs/xfs_log_priv.h 		spin_lock(&log->l_icloglock);
log               595 fs/xfs/xfs_log_priv.h 		cur_cycle = log->l_curr_cycle;
log               596 fs/xfs/xfs_log_priv.h 		cur_block = log->l_curr_block;
log               597 fs/xfs/xfs_log_priv.h 		spin_unlock(&log->l_icloglock);
log                52 fs/xfs/xfs_log_recover.c #define	xlog_recover_check_summary(log)
log                80 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log                84 fs/xfs/xfs_log_recover.c 	if (blk_no < 0 || blk_no >= log->l_logBBsize)
log                86 fs/xfs/xfs_log_recover.c 	if (bbcount <= 0 || (blk_no + bbcount) > log->l_logBBsize)
log                97 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               100 fs/xfs/xfs_log_recover.c 	int align_mask = xfs_buftarg_dma_alignment(log->l_targ);
log               106 fs/xfs/xfs_log_recover.c 	if (!xlog_verify_bno(log, 0, nbblks)) {
log               107 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "Invalid block length (0x%x) for buffer",
log               109 fs/xfs/xfs_log_recover.c 		XFS_ERROR_REPORT(__func__, XFS_ERRLEVEL_HIGH, log->l_mp);
log               127 fs/xfs/xfs_log_recover.c 	if (nbblks > 1 && log->l_sectBBsize > 1)
log               128 fs/xfs/xfs_log_recover.c 		nbblks += log->l_sectBBsize;
log               129 fs/xfs/xfs_log_recover.c 	nbblks = round_up(nbblks, log->l_sectBBsize);
log               139 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               142 fs/xfs/xfs_log_recover.c 	return BBTOB(blk_no & ((xfs_daddr_t)log->l_sectBBsize - 1));
log               147 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log               155 fs/xfs/xfs_log_recover.c 	if (!xlog_verify_bno(log, blk_no, nbblks)) {
log               156 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp,
log               159 fs/xfs/xfs_log_recover.c 		XFS_ERROR_REPORT(__func__, XFS_ERRLEVEL_HIGH, log->l_mp);
log               163 fs/xfs/xfs_log_recover.c 	blk_no = round_down(blk_no, log->l_sectBBsize);
log               164 fs/xfs/xfs_log_recover.c 	nbblks = round_up(nbblks, log->l_sectBBsize);
log               167 fs/xfs/xfs_log_recover.c 	error = xfs_rw_bdev(log->l_targ->bt_bdev, log->l_logBBstart + blk_no,
log               169 fs/xfs/xfs_log_recover.c 	if (error && !XFS_FORCED_SHUTDOWN(log->l_mp)) {
log               170 fs/xfs/xfs_log_recover.c 		xfs_alert(log->l_mp,
log               180 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               185 fs/xfs/xfs_log_recover.c 	return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ);
log               190 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               198 fs/xfs/xfs_log_recover.c 	error = xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ);
log               200 fs/xfs/xfs_log_recover.c 		*offset = data + xlog_align(log, blk_no);
log               206 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               211 fs/xfs/xfs_log_recover.c 	return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_WRITE);
log               327 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               342 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, mid_blk, 1, buffer, &offset);
log               370 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               390 fs/xfs/xfs_log_recover.c 	while (bufblks > log->l_logBBsize)
log               392 fs/xfs/xfs_log_recover.c 	while (!(buffer = xlog_alloc_buffer(log, bufblks))) {
log               394 fs/xfs/xfs_log_recover.c 		if (bufblks < log->l_sectBBsize)
log               403 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, i, bcount, buffer, &buf);
log               439 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log               455 fs/xfs/xfs_log_recover.c 	buffer = xlog_alloc_buffer(log, num_blks);
log               457 fs/xfs/xfs_log_recover.c 		buffer = xlog_alloc_buffer(log, 1);
log               462 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, start_blk, num_blks, buffer, &offset);
log               471 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log               479 fs/xfs/xfs_log_recover.c 			error = xlog_bread(log, i, 1, buffer, &offset);
log               507 fs/xfs/xfs_log_recover.c 	if ((error = xlog_header_check_mount(log->l_mp, head)))
log               517 fs/xfs/xfs_log_recover.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log               551 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               560 fs/xfs/xfs_log_recover.c 	int		error, log_bbnum = log->l_logBBsize;
log               563 fs/xfs/xfs_log_recover.c 	error = xlog_find_zeroed(log, &first_blk);
log               565 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "empty log check failed");
log               577 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp, "totally zeroed log");
log               584 fs/xfs/xfs_log_recover.c 	buffer = xlog_alloc_buffer(log, 1);
log               588 fs/xfs/xfs_log_recover.c 	error = xlog_bread(log, 0, 1, buffer, &offset);
log               595 fs/xfs/xfs_log_recover.c 	error = xlog_bread(log, last_blk, 1, buffer, &offset);
log               665 fs/xfs/xfs_log_recover.c 		error = xlog_find_cycle_start(log, buffer, first_blk, &head_blk,
log               678 fs/xfs/xfs_log_recover.c 	num_scan_bblks = min_t(int, log_bbnum, XLOG_TOTAL_REC_SHIFT(log));
log               685 fs/xfs/xfs_log_recover.c 		if ((error = xlog_find_verify_cycle(log,
log               722 fs/xfs/xfs_log_recover.c 		if ((error = xlog_find_verify_cycle(log, start_blk,
log               738 fs/xfs/xfs_log_recover.c 		if ((error = xlog_find_verify_cycle(log,
log               751 fs/xfs/xfs_log_recover.c 	num_scan_bblks = XLOG_REC_SHIFT(log);
log               756 fs/xfs/xfs_log_recover.c 		error = xlog_find_verify_log_record(log, start_blk, &head_blk, 0);
log               764 fs/xfs/xfs_log_recover.c 		error = xlog_find_verify_log_record(log, start_blk, &head_blk, 0);
log               774 fs/xfs/xfs_log_recover.c 			error = xlog_find_verify_log_record(log, start_blk,
log               802 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "failed to find log head");
log               816 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log               839 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, i, 1, buffer, &offset);
log               857 fs/xfs/xfs_log_recover.c 		for (i = log->l_logBBsize - 1; i >= (int) tail_blk; i--) {
log               858 fs/xfs/xfs_log_recover.c 			error = xlog_bread(log, i, 1, buffer, &offset);
log               890 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log               911 fs/xfs/xfs_log_recover.c 	end_blk = head_blk > tail_blk ? head_blk : log->l_logBBsize - 1;
log               913 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, i, 1, buffer, &offset);
log               931 fs/xfs/xfs_log_recover.c 			error = xlog_bread(log, i, 1, buffer, &offset);
log               957 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log               964 fs/xfs/xfs_log_recover.c 	return tail_blk + (log->l_logBBsize - head_blk);
log               988 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1001 fs/xfs/xfs_log_recover.c 	buffer = xlog_alloc_buffer(log, 1);
log              1009 fs/xfs/xfs_log_recover.c 	error = xlog_seek_logrec_hdr(log, head_blk, *tail_blk, 1, buffer,
log              1024 fs/xfs/xfs_log_recover.c 	error = xlog_do_recovery_pass(log, head_blk, *tail_blk,
log              1033 fs/xfs/xfs_log_recover.c 		tail_distance = xlog_tail_distance(log, head_blk, first_bad);
log              1038 fs/xfs/xfs_log_recover.c 		error = xlog_seek_logrec_hdr(log, head_blk, first_bad, 2,
log              1045 fs/xfs/xfs_log_recover.c 		error = xlog_do_recovery_pass(log, head_blk, *tail_blk,
log              1050 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp,
log              1073 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1095 fs/xfs/xfs_log_recover.c 	tmp_buffer = xlog_alloc_buffer(log, 1);
log              1098 fs/xfs/xfs_log_recover.c 	error = xlog_rseek_logrec_hdr(log, *head_blk, *tail_blk,
log              1110 fs/xfs/xfs_log_recover.c 	error = xlog_do_recovery_pass(log, *head_blk, tmp_rhead_blk,
log              1118 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp,
log              1130 fs/xfs/xfs_log_recover.c 		found = xlog_rseek_logrec_hdr(log, first_bad, *tail_blk, 1,
log              1156 fs/xfs/xfs_log_recover.c 	return xlog_verify_tail(log, *head_blk, tail_blk,
log              1170 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1175 fs/xfs/xfs_log_recover.c 	div_s64_rem(bno, log->l_logBBsize, &mod);
log              1186 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1212 fs/xfs/xfs_log_recover.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log              1228 fs/xfs/xfs_log_recover.c 	after_umount_blk = xlog_wrap_logbno(log,
log              1233 fs/xfs/xfs_log_recover.c 		umount_data_blk = xlog_wrap_logbno(log, rhead_blk + hblks);
log              1234 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, umount_data_blk, 1, buffer, &offset);
log              1245 fs/xfs/xfs_log_recover.c 			xlog_assign_atomic_lsn(&log->l_tail_lsn,
log              1246 fs/xfs/xfs_log_recover.c 					log->l_curr_cycle, after_umount_blk);
log              1247 fs/xfs/xfs_log_recover.c 			xlog_assign_atomic_lsn(&log->l_last_sync_lsn,
log              1248 fs/xfs/xfs_log_recover.c 					log->l_curr_cycle, after_umount_blk);
log              1260 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1276 fs/xfs/xfs_log_recover.c 	log->l_prev_block = rhead_blk;
log              1277 fs/xfs/xfs_log_recover.c 	log->l_curr_block = (int)head_blk;
log              1278 fs/xfs/xfs_log_recover.c 	log->l_curr_cycle = be32_to_cpu(rhead->h_cycle);
log              1280 fs/xfs/xfs_log_recover.c 		log->l_curr_cycle++;
log              1281 fs/xfs/xfs_log_recover.c 	atomic64_set(&log->l_tail_lsn, be64_to_cpu(rhead->h_tail_lsn));
log              1282 fs/xfs/xfs_log_recover.c 	atomic64_set(&log->l_last_sync_lsn, be64_to_cpu(rhead->h_lsn));
log              1283 fs/xfs/xfs_log_recover.c 	xlog_assign_grant_head(&log->l_reserve_head.grant, log->l_curr_cycle,
log              1284 fs/xfs/xfs_log_recover.c 					BBTOB(log->l_curr_block));
log              1285 fs/xfs/xfs_log_recover.c 	xlog_assign_grant_head(&log->l_write_head.grant, log->l_curr_cycle,
log              1286 fs/xfs/xfs_log_recover.c 					BBTOB(log->l_curr_block));
log              1307 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1323 fs/xfs/xfs_log_recover.c 	if ((error = xlog_find_head(log, head_blk)))
log              1327 fs/xfs/xfs_log_recover.c 	buffer = xlog_alloc_buffer(log, 1);
log              1331 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, 0, 1, buffer, &offset);
log              1347 fs/xfs/xfs_log_recover.c 	error = xlog_rseek_logrec_hdr(log, *head_blk, *head_blk, 1, buffer,
log              1352 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: couldn't find sync record", __func__);
log              1360 fs/xfs/xfs_log_recover.c 	xlog_set_state(log, *head_blk, rhead, rhead_blk, wrapped);
log              1361 fs/xfs/xfs_log_recover.c 	tail_lsn = atomic64_read(&log->l_tail_lsn);
log              1367 fs/xfs/xfs_log_recover.c 	error = xlog_check_unmount_rec(log, head_blk, tail_blk, rhead,
log              1385 fs/xfs/xfs_log_recover.c 		error = xlog_verify_head(log, head_blk, tail_blk, buffer,
log              1392 fs/xfs/xfs_log_recover.c 			xlog_set_state(log, *head_blk, rhead, rhead_blk,
log              1394 fs/xfs/xfs_log_recover.c 			tail_lsn = atomic64_read(&log->l_tail_lsn);
log              1395 fs/xfs/xfs_log_recover.c 			error = xlog_check_unmount_rec(log, head_blk, tail_blk,
log              1409 fs/xfs/xfs_log_recover.c 		log->l_mp->m_flags |= XFS_MOUNT_WAS_CLEAN;
log              1430 fs/xfs/xfs_log_recover.c 	if (!xfs_readonly_buftarg(log->l_targ))
log              1431 fs/xfs/xfs_log_recover.c 		error = xlog_clear_stale_blocks(log, tail_lsn);
log              1437 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "failed to locate log tail");
log              1459 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log              1467 fs/xfs/xfs_log_recover.c 	int	        error, log_bbnum = log->l_logBBsize;
log              1472 fs/xfs/xfs_log_recover.c 	buffer = xlog_alloc_buffer(log, 1);
log              1475 fs/xfs/xfs_log_recover.c 	error = xlog_bread(log, 0, 1, buffer, &offset);
log              1487 fs/xfs/xfs_log_recover.c 	error = xlog_bread(log, log_bbnum-1, 1, buffer, &offset);
log              1499 fs/xfs/xfs_log_recover.c 	error = xlog_find_cycle_start(log, buffer, 0, &last_blk, 0);
log              1509 fs/xfs/xfs_log_recover.c 	num_scan_bblks = XLOG_TOTAL_REC_SHIFT(log);
log              1522 fs/xfs/xfs_log_recover.c 	if ((error = xlog_find_verify_cycle(log, start_blk,
log              1532 fs/xfs/xfs_log_recover.c 	error = xlog_find_verify_log_record(log, start_blk, &last_blk, 0);
log              1553 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1566 fs/xfs/xfs_log_recover.c 			xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? 2 : 1);
log              1570 fs/xfs/xfs_log_recover.c 	memcpy(&recp->h_fs_uuid, &log->l_mp->m_sb.sb_uuid, sizeof(uuid_t));
log              1575 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log              1585 fs/xfs/xfs_log_recover.c 	int		sectbb = log->l_sectBBsize;
log              1598 fs/xfs/xfs_log_recover.c 	while (bufblks > log->l_logBBsize)
log              1600 fs/xfs/xfs_log_recover.c 	while (!(buffer = xlog_alloc_buffer(log, bufblks))) {
log              1612 fs/xfs/xfs_log_recover.c 		error = xlog_bread_noalign(log, start_block, 1, buffer);
log              1631 fs/xfs/xfs_log_recover.c 			error = xlog_bread_noalign(log, ealign, sectbb,
log              1638 fs/xfs/xfs_log_recover.c 		offset = buffer + xlog_align(log, start_block);
log              1640 fs/xfs/xfs_log_recover.c 			xlog_add_record(log, offset, cycle, i+j,
log              1644 fs/xfs/xfs_log_recover.c 		error = xlog_bwrite(log, start_block, endcount, buffer);
log              1674 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log              1685 fs/xfs/xfs_log_recover.c 	head_cycle = log->l_curr_cycle;
log              1686 fs/xfs/xfs_log_recover.c 	head_block = log->l_curr_block;
log              1702 fs/xfs/xfs_log_recover.c 		if (unlikely(head_block < tail_block || head_block >= log->l_logBBsize)) {
log              1704 fs/xfs/xfs_log_recover.c 					 XFS_ERRLEVEL_LOW, log->l_mp);
log              1707 fs/xfs/xfs_log_recover.c 		tail_distance = tail_block + (log->l_logBBsize - head_block);
log              1716 fs/xfs/xfs_log_recover.c 					 XFS_ERRLEVEL_LOW, log->l_mp);
log              1731 fs/xfs/xfs_log_recover.c 	max_distance = XLOG_TOTAL_REC_SHIFT(log);
log              1741 fs/xfs/xfs_log_recover.c 	if ((head_block + max_distance) <= log->l_logBBsize) {
log              1749 fs/xfs/xfs_log_recover.c 		error = xlog_write_log_records(log, (head_cycle - 1),
log              1762 fs/xfs/xfs_log_recover.c 		distance = log->l_logBBsize - head_block;
log              1763 fs/xfs/xfs_log_recover.c 		error = xlog_write_log_records(log, (head_cycle - 1),
log              1778 fs/xfs/xfs_log_recover.c 		distance = max_distance - (log->l_logBBsize - head_block);
log              1779 fs/xfs/xfs_log_recover.c 		error = xlog_write_log_records(log, head_cycle, 0, distance,
log              1846 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1868 fs/xfs/xfs_log_recover.c 				trace_xfs_log_recover_item_reorder_head(log,
log              1890 fs/xfs/xfs_log_recover.c 			trace_xfs_log_recover_item_reorder_tail(log,
log              1895 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log              1936 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              1947 fs/xfs/xfs_log_recover.c 		trace_xfs_log_recover_buf_not_cancel(log, buf_f);
log              1955 fs/xfs/xfs_log_recover.c 	bucket = XLOG_BUF_CANCEL_BUCKET(log, buf_f->blf_blkno);
log              1960 fs/xfs/xfs_log_recover.c 			trace_xfs_log_recover_buf_cancel_ref_inc(log, buf_f);
log              1971 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_buf_cancel_add(log, buf_f);
log              1982 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              1990 fs/xfs/xfs_log_recover.c 	if (!log->l_buf_cancel_table) {
log              1996 fs/xfs/xfs_log_recover.c 	bucket = XLOG_BUF_CANCEL_BUCKET(log, blkno);
log              2022 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              2029 fs/xfs/xfs_log_recover.c 	bcp = xlog_peek_buffer_cancelled(log, blkno, len, flags);
log              2664 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              2671 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_buf_dquot_buf(log, buf_f);
log              2689 fs/xfs/xfs_log_recover.c 	if (log->l_quotaoffs_flag & type)
log              2721 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              2727 fs/xfs/xfs_log_recover.c 	xfs_mount_t		*mp = log->l_mp;
log              2737 fs/xfs/xfs_log_recover.c 	if (xlog_check_buffer_cancelled(log, buf_f->blf_blkno,
log              2739 fs/xfs/xfs_log_recover.c 		trace_xfs_log_recover_buf_cancel(log, buf_f);
log              2743 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_buf_recover(log, buf_f);
log              2780 fs/xfs/xfs_log_recover.c 		trace_xfs_log_recover_buf_skip(log, buf_f);
log              2793 fs/xfs/xfs_log_recover.c 		dirty = xlog_recover_do_dquot_buffer(mp, log, item, bp, buf_f);
log              2817 fs/xfs/xfs_log_recover.c 	    (BBTOB(bp->b_length) != M_IGEO(log->l_mp)->inode_cluster_size)) {
log              2913 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              2919 fs/xfs/xfs_log_recover.c 	xfs_mount_t		*mp = log->l_mp;
log              2946 fs/xfs/xfs_log_recover.c 	if (xlog_check_buffer_cancelled(log, in_f->ilf_blkno,
log              2949 fs/xfs/xfs_log_recover.c 		trace_xfs_log_recover_inode_cancel(log, in_f);
log              2952 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_inode_recover(log, in_f);
log              3003 fs/xfs/xfs_log_recover.c 			trace_xfs_log_recover_inode_skip(log, in_f);
log              3027 fs/xfs/xfs_log_recover.c 			trace_xfs_log_recover_inode_skip(log, in_f);
log              3167 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp, "%s: Invalid flag", __func__);
log              3181 fs/xfs/xfs_log_recover.c 	xfs_dinode_calc_crc(log->l_mp, dip);
log              3202 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3213 fs/xfs/xfs_log_recover.c 		log->l_quotaoffs_flag |= XFS_DQ_USER;
log              3215 fs/xfs/xfs_log_recover.c 		log->l_quotaoffs_flag |= XFS_DQ_PROJ;
log              3217 fs/xfs/xfs_log_recover.c 		log->l_quotaoffs_flag |= XFS_DQ_GROUP;
log              3227 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3232 fs/xfs/xfs_log_recover.c 	xfs_mount_t		*mp = log->l_mp;
log              3249 fs/xfs/xfs_log_recover.c 		xfs_alert(log->l_mp, "NULL dquot in %s.", __func__);
log              3253 fs/xfs/xfs_log_recover.c 		xfs_alert(log->l_mp, "dquot too small (%d) in %s.",
log              3263 fs/xfs/xfs_log_recover.c 	if (log->l_quotaoffs_flag & type)
log              3340 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3345 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3359 fs/xfs/xfs_log_recover.c 	spin_lock(&log->l_ailp->ail_lock);
log              3366 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_update(log->l_ailp, &efip->efi_item, lsn);
log              3381 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3389 fs/xfs/xfs_log_recover.c 	struct xfs_ail		*ailp = log->l_ailp;
log              3436 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3441 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3455 fs/xfs/xfs_log_recover.c 	spin_lock(&log->l_ailp->ail_lock);
log              3462 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_update(log->l_ailp, &ruip->rui_item, lsn);
log              3477 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3485 fs/xfs/xfs_log_recover.c 	struct xfs_ail			*ailp = log->l_ailp;
log              3552 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3557 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3571 fs/xfs/xfs_log_recover.c 	spin_lock(&log->l_ailp->ail_lock);
log              3578 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_update(log->l_ailp, &cuip->cui_item, lsn);
log              3593 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3601 fs/xfs/xfs_log_recover.c 	struct xfs_ail			*ailp = log->l_ailp;
log              3669 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3674 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3690 fs/xfs/xfs_log_recover.c 	spin_lock(&log->l_ailp->ail_lock);
log              3697 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_update(log->l_ailp, &buip->bui_item, lsn);
log              3712 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3720 fs/xfs/xfs_log_recover.c 	struct xfs_ail			*ailp = log->l_ailp;
log              3766 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              3770 fs/xfs/xfs_log_recover.c 	struct xfs_mount	*mp = log->l_mp;
log              3785 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad type");
log              3790 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad icl size");
log              3796 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agno");
log              3801 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agbno");
log              3806 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad isize");
log              3811 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad count");
log              3816 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad length");
log              3826 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp,
log              3833 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp,
log              3852 fs/xfs/xfs_log_recover.c 		if (xlog_check_buffer_cancelled(log, daddr, bb_per_cluster, 0))
log              3871 fs/xfs/xfs_log_recover.c 		trace_xfs_log_recover_icreate_cancel(log, icl);
log              3875 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_icreate_recover(log, icl);
log              3882 fs/xfs/xfs_log_recover.c 	struct xlog                     *log,
log              3886 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3888 fs/xfs/xfs_log_recover.c 	if (xlog_peek_buffer_cancelled(log, buf_f->blf_blkno,
log              3899 fs/xfs/xfs_log_recover.c 	struct xlog                     *log,
log              3904 fs/xfs/xfs_log_recover.c 	struct xfs_mount		*mp = log->l_mp;
log              3917 fs/xfs/xfs_log_recover.c 	if (xlog_peek_buffer_cancelled(log, ilfp->ilf_blkno, ilfp->ilf_len, 0))
log              3926 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3929 fs/xfs/xfs_log_recover.c 	struct xfs_mount	*mp = log->l_mp;
log              3947 fs/xfs/xfs_log_recover.c 	if (log->l_quotaoffs_flag & type)
log              3955 fs/xfs/xfs_log_recover.c 	if (xlog_peek_buffer_cancelled(log, dq_f->qlf_blkno, len, 0))
log              3964 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3969 fs/xfs/xfs_log_recover.c 		xlog_recover_buffer_ra_pass2(log, item);
log              3972 fs/xfs/xfs_log_recover.c 		xlog_recover_inode_ra_pass2(log, item);
log              3975 fs/xfs/xfs_log_recover.c 		xlog_recover_dquot_ra_pass2(log, item);
log              3993 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              3997 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_item_recover(log, trans, item, XLOG_RECOVER_PASS1);
log              4001 fs/xfs/xfs_log_recover.c 		return xlog_recover_buffer_pass1(log, item);
log              4003 fs/xfs/xfs_log_recover.c 		return xlog_recover_quotaoff_pass1(log, item);
log              4018 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: invalid item type (%d)",
log              4027 fs/xfs/xfs_log_recover.c 	struct xlog			*log,
log              4032 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_item_recover(log, trans, item, XLOG_RECOVER_PASS2);
log              4036 fs/xfs/xfs_log_recover.c 		return xlog_recover_buffer_pass2(log, buffer_list, item,
log              4039 fs/xfs/xfs_log_recover.c 		return xlog_recover_inode_pass2(log, buffer_list, item,
log              4042 fs/xfs/xfs_log_recover.c 		return xlog_recover_efi_pass2(log, item, trans->r_lsn);
log              4044 fs/xfs/xfs_log_recover.c 		return xlog_recover_efd_pass2(log, item);
log              4046 fs/xfs/xfs_log_recover.c 		return xlog_recover_rui_pass2(log, item, trans->r_lsn);
log              4048 fs/xfs/xfs_log_recover.c 		return xlog_recover_rud_pass2(log, item);
log              4050 fs/xfs/xfs_log_recover.c 		return xlog_recover_cui_pass2(log, item, trans->r_lsn);
log              4052 fs/xfs/xfs_log_recover.c 		return xlog_recover_cud_pass2(log, item);
log              4054 fs/xfs/xfs_log_recover.c 		return xlog_recover_bui_pass2(log, item, trans->r_lsn);
log              4056 fs/xfs/xfs_log_recover.c 		return xlog_recover_bud_pass2(log, item);
log              4058 fs/xfs/xfs_log_recover.c 		return xlog_recover_dquot_pass2(log, buffer_list, item,
log              4061 fs/xfs/xfs_log_recover.c 		return xlog_recover_do_icreate_pass2(log, buffer_list, item);
log              4066 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: invalid item type (%d)",
log              4075 fs/xfs/xfs_log_recover.c 	struct xlog                     *log,
log              4084 fs/xfs/xfs_log_recover.c 		error = xlog_recover_commit_pass2(log, trans,
log              4101 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4117 fs/xfs/xfs_log_recover.c 	error = xlog_recover_reorder_trans(log, trans, pass);
log              4124 fs/xfs/xfs_log_recover.c 			error = xlog_recover_commit_pass1(log, trans, item);
log              4127 fs/xfs/xfs_log_recover.c 			xlog_recover_ra_pass2(log, item);
log              4131 fs/xfs/xfs_log_recover.c 				error = xlog_recover_items_pass2(log, trans,
log              4149 fs/xfs/xfs_log_recover.c 			error = xlog_recover_items_pass2(log, trans,
log              4173 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4189 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp, "%s: bad header length", __func__);
log              4210 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_item_add_cont(log, trans, item, 0);
log              4229 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4243 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp, "%s: bad header magic number",
log              4250 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp, "%s: bad header length", __func__);
log              4283 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log              4301 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_item_add(log, trans, item, 0);
log              4337 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4361 fs/xfs/xfs_log_recover.c 		error = xlog_recover_add_to_trans(log, trans, dp, len);
log              4364 fs/xfs/xfs_log_recover.c 		error = xlog_recover_add_to_cont_trans(log, trans, dp, len);
log              4367 fs/xfs/xfs_log_recover.c 		error = xlog_recover_commit_trans(log, trans, pass,
log              4376 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: Unmount LR", __func__);
log              4381 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: bad flag 0x%x", __func__, flags);
log              4444 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4460 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: bad clientid 0x%x",
log              4471 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: bad length 0x%x", __func__, len);
log              4505 fs/xfs/xfs_log_recover.c 	if (log->l_recovery_lsn != trans->r_lsn &&
log              4510 fs/xfs/xfs_log_recover.c 		log->l_recovery_lsn = trans->r_lsn;
log              4513 fs/xfs/xfs_log_recover.c 	return xlog_recovery_process_trans(log, trans, dp, len,
log              4528 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              4544 fs/xfs/xfs_log_recover.c 	if (xlog_header_check_recover(log->l_mp, rhead))
log              4547 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover_record(log, rhead, pass);
log              4555 fs/xfs/xfs_log_recover.c 		error = xlog_recover_process_ophdr(log, rhash, rhead, ohead,
log              4792 fs/xfs/xfs_log_recover.c 	struct xlog		*log)
log              4812 fs/xfs/xfs_log_recover.c 	error = xfs_trans_alloc_empty(log->l_mp, &parent_tp);
log              4816 fs/xfs/xfs_log_recover.c 	ailp = log->l_ailp;
log              4820 fs/xfs/xfs_log_recover.c 	last_lsn = xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block);
log              4850 fs/xfs/xfs_log_recover.c 			error = xlog_recover_process_efi(log->l_mp, ailp, lip);
log              4853 fs/xfs/xfs_log_recover.c 			error = xlog_recover_process_rui(log->l_mp, ailp, lip);
log              4882 fs/xfs/xfs_log_recover.c 	struct xlog		*log)
log              4888 fs/xfs/xfs_log_recover.c 	ailp = log->l_ailp;
log              4906 fs/xfs/xfs_log_recover.c 			xlog_recover_cancel_efi(log->l_mp, ailp, lip);
log              4909 fs/xfs/xfs_log_recover.c 			xlog_recover_cancel_rui(log->l_mp, ailp, lip);
log              4912 fs/xfs/xfs_log_recover.c 			xlog_recover_cancel_cui(log->l_mp, ailp, lip);
log              4915 fs/xfs/xfs_log_recover.c 			xlog_recover_cancel_bui(log->l_mp, ailp, lip);
log              5051 fs/xfs/xfs_log_recover.c 	struct xlog	*log)
log              5061 fs/xfs/xfs_log_recover.c 	mp = log->l_mp;
log              5105 fs/xfs/xfs_log_recover.c 	struct xlog		*log)
log              5115 fs/xfs/xfs_log_recover.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log              5131 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              5141 fs/xfs/xfs_log_recover.c 	crc = xlog_cksum(log, rhead, dp, be32_to_cpu(rhead->h_len));
log              5163 fs/xfs/xfs_log_recover.c 		if (old_crc || xfs_sb_version_hascrc(&log->l_mp->m_sb)) {
log              5164 fs/xfs/xfs_log_recover.c 			xfs_alert(log->l_mp,
log              5175 fs/xfs/xfs_log_recover.c 		if (xfs_sb_version_hascrc(&log->l_mp->m_sb))
log              5179 fs/xfs/xfs_log_recover.c 	xlog_unpack_data(rhead, dp, log);
log              5181 fs/xfs/xfs_log_recover.c 	return xlog_recover_process_data(log, rhash, rhead, dp, pass,
log              5187 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              5195 fs/xfs/xfs_log_recover.c 				XFS_ERRLEVEL_LOW, log->l_mp);
log              5201 fs/xfs/xfs_log_recover.c 		xfs_warn(log->l_mp, "%s: unrecognised log version (%d).",
log              5210 fs/xfs/xfs_log_recover.c 				XFS_ERRLEVEL_LOW, log->l_mp);
log              5213 fs/xfs/xfs_log_recover.c 	if (unlikely( blkno > log->l_logBBsize || blkno > INT_MAX )) {
log              5215 fs/xfs/xfs_log_recover.c 				XFS_ERRLEVEL_LOW, log->l_mp);
log              5231 fs/xfs/xfs_log_recover.c 	struct xlog		*log,
log              5260 fs/xfs/xfs_log_recover.c 	if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) {
log              5266 fs/xfs/xfs_log_recover.c 		hbp = xlog_alloc_buffer(log, 1);
log              5270 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, tail_blk, 1, hbp, &offset);
log              5275 fs/xfs/xfs_log_recover.c 		error = xlog_valid_rec_header(log, rhead, tail_blk);
log              5293 fs/xfs/xfs_log_recover.c 			if (h_len <= log->l_mp->m_logbsize &&
log              5295 fs/xfs/xfs_log_recover.c 				xfs_warn(log->l_mp,
log              5297 fs/xfs/xfs_log_recover.c 					 h_size, log->l_mp->m_logbsize);
log              5298 fs/xfs/xfs_log_recover.c 				h_size = log->l_mp->m_logbsize;
log              5309 fs/xfs/xfs_log_recover.c 			hbp = xlog_alloc_buffer(log, hblks);
log              5314 fs/xfs/xfs_log_recover.c 		ASSERT(log->l_sectBBsize == 1);
log              5316 fs/xfs/xfs_log_recover.c 		hbp = xlog_alloc_buffer(log, 1);
log              5322 fs/xfs/xfs_log_recover.c 	dbp = xlog_alloc_buffer(log, BTOBB(h_size));
log              5335 fs/xfs/xfs_log_recover.c 		while (blk_no < log->l_logBBsize) {
log              5342 fs/xfs/xfs_log_recover.c 			if (blk_no + hblks <= log->l_logBBsize) {
log              5344 fs/xfs/xfs_log_recover.c 				error = xlog_bread(log, blk_no, hblks, hbp,
log              5350 fs/xfs/xfs_log_recover.c 				if (blk_no != log->l_logBBsize) {
log              5353 fs/xfs/xfs_log_recover.c 					split_hblks = log->l_logBBsize - (int)blk_no;
log              5355 fs/xfs/xfs_log_recover.c 					error = xlog_bread(log, blk_no,
log              5375 fs/xfs/xfs_log_recover.c 				error = xlog_bread_noalign(log, 0,
log              5382 fs/xfs/xfs_log_recover.c 			error = xlog_valid_rec_header(log, rhead,
log              5397 fs/xfs/xfs_log_recover.c 			if (blk_no + bblks <= log->l_logBBsize ||
log              5398 fs/xfs/xfs_log_recover.c 			    blk_no >= log->l_logBBsize) {
log              5399 fs/xfs/xfs_log_recover.c 				rblk_no = xlog_wrap_logbno(log, blk_no);
log              5400 fs/xfs/xfs_log_recover.c 				error = xlog_bread(log, rblk_no, bblks, dbp,
log              5409 fs/xfs/xfs_log_recover.c 				if (blk_no != log->l_logBBsize) {
log              5415 fs/xfs/xfs_log_recover.c 						log->l_logBBsize - (int)blk_no;
log              5417 fs/xfs/xfs_log_recover.c 					error = xlog_bread(log, blk_no,
log              5436 fs/xfs/xfs_log_recover.c 				error = xlog_bread_noalign(log, 0,
log              5443 fs/xfs/xfs_log_recover.c 			error = xlog_recover_process(log, rhash, rhead, offset,
log              5452 fs/xfs/xfs_log_recover.c 		ASSERT(blk_no >= log->l_logBBsize);
log              5453 fs/xfs/xfs_log_recover.c 		blk_no -= log->l_logBBsize;
log              5459 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, blk_no, hblks, hbp, &offset);
log              5464 fs/xfs/xfs_log_recover.c 		error = xlog_valid_rec_header(log, rhead, blk_no);
log              5470 fs/xfs/xfs_log_recover.c 		error = xlog_bread(log, blk_no+hblks, bblks, dbp,
log              5475 fs/xfs/xfs_log_recover.c 		error = xlog_recover_process(log, rhash, rhead, offset, pass,
log              5530 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log              5542 fs/xfs/xfs_log_recover.c 	log->l_buf_cancel_table = kmem_zalloc(XLOG_BC_TABLE_SIZE *
log              5546 fs/xfs/xfs_log_recover.c 		INIT_LIST_HEAD(&log->l_buf_cancel_table[i]);
log              5548 fs/xfs/xfs_log_recover.c 	error = xlog_do_recovery_pass(log, head_blk, tail_blk,
log              5551 fs/xfs/xfs_log_recover.c 		kmem_free(log->l_buf_cancel_table);
log              5552 fs/xfs/xfs_log_recover.c 		log->l_buf_cancel_table = NULL;
log              5559 fs/xfs/xfs_log_recover.c 	error = xlog_do_recovery_pass(log, head_blk, tail_blk,
log              5566 fs/xfs/xfs_log_recover.c 			ASSERT(list_empty(&log->l_buf_cancel_table[i]));
log              5570 fs/xfs/xfs_log_recover.c 	kmem_free(log->l_buf_cancel_table);
log              5571 fs/xfs/xfs_log_recover.c 	log->l_buf_cancel_table = NULL;
log              5581 fs/xfs/xfs_log_recover.c 	struct xlog	*log,
log              5585 fs/xfs/xfs_log_recover.c 	struct xfs_mount *mp = log->l_mp;
log              5590 fs/xfs/xfs_log_recover.c 	trace_xfs_log_recover(log, head_blk, tail_blk);
log              5595 fs/xfs/xfs_log_recover.c 	error = xlog_do_log_recovery(log, head_blk, tail_blk);
log              5651 fs/xfs/xfs_log_recover.c 	xlog_recover_check_summary(log);
log              5654 fs/xfs/xfs_log_recover.c 	log->l_flags &= ~XLOG_ACTIVE_RECOVERY;
log              5665 fs/xfs/xfs_log_recover.c 	struct xlog	*log)
log              5671 fs/xfs/xfs_log_recover.c 	error = xlog_find_tail(log, &head_blk, &tail_blk);
log              5680 fs/xfs/xfs_log_recover.c 	if (xfs_sb_version_hascrc(&log->l_mp->m_sb) &&
log              5681 fs/xfs/xfs_log_recover.c 	    !xfs_log_check_lsn(log->l_mp, log->l_mp->m_sb.sb_lsn))
log              5696 fs/xfs/xfs_log_recover.c 		if ((error = xfs_dev_is_read_only(log->l_mp, "recovery"))) {
log              5707 fs/xfs/xfs_log_recover.c 		if (XFS_SB_VERSION_NUM(&log->l_mp->m_sb) == XFS_SB_VERSION_5 &&
log              5708 fs/xfs/xfs_log_recover.c 		    xfs_sb_has_incompat_log_feature(&log->l_mp->m_sb,
log              5710 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log              5712 fs/xfs/xfs_log_recover.c 				(log->l_mp->m_sb.sb_features_log_incompat &
log              5714 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log              5716 fs/xfs/xfs_log_recover.c 			xfs_warn(log->l_mp,
log              5727 fs/xfs/xfs_log_recover.c 			xfs_notice(log->l_mp,
log              5733 fs/xfs/xfs_log_recover.c 		xfs_notice(log->l_mp, "Starting recovery (logdev: %s)",
log              5734 fs/xfs/xfs_log_recover.c 				log->l_mp->m_logname ? log->l_mp->m_logname
log              5737 fs/xfs/xfs_log_recover.c 		error = xlog_do_recover(log, head_blk, tail_blk);
log              5738 fs/xfs/xfs_log_recover.c 		log->l_flags |= XLOG_RECOVERY_NEEDED;
log              5754 fs/xfs/xfs_log_recover.c 	struct xlog	*log)
log              5764 fs/xfs/xfs_log_recover.c 	if (log->l_flags & XLOG_RECOVERY_NEEDED) {
log              5766 fs/xfs/xfs_log_recover.c 		error = xlog_recover_process_intents(log);
log              5768 fs/xfs/xfs_log_recover.c 			xfs_alert(log->l_mp, "Failed to recover intents");
log              5778 fs/xfs/xfs_log_recover.c 		xfs_log_force(log->l_mp, XFS_LOG_SYNC);
log              5780 fs/xfs/xfs_log_recover.c 		xlog_recover_process_iunlinks(log);
log              5782 fs/xfs/xfs_log_recover.c 		xlog_recover_check_summary(log);
log              5784 fs/xfs/xfs_log_recover.c 		xfs_notice(log->l_mp, "Ending recovery (logdev: %s)",
log              5785 fs/xfs/xfs_log_recover.c 				log->l_mp->m_logname ? log->l_mp->m_logname
log              5787 fs/xfs/xfs_log_recover.c 		log->l_flags &= ~XLOG_RECOVERY_NEEDED;
log              5789 fs/xfs/xfs_log_recover.c 		xfs_info(log->l_mp, "Ending clean mount");
log              5796 fs/xfs/xfs_log_recover.c 	struct xlog	*log)
log              5798 fs/xfs/xfs_log_recover.c 	if (log->l_flags & XLOG_RECOVERY_NEEDED)
log              5799 fs/xfs/xfs_log_recover.c 		xlog_recover_cancel_intents(log);
log              5809 fs/xfs/xfs_log_recover.c 	struct xlog	*log)
log              5821 fs/xfs/xfs_log_recover.c 	mp = log->l_mp;
log                33 fs/xfs/xfs_rtalloc.c 	int		log,		/* log2 of extent size */
log                39 fs/xfs/xfs_rtalloc.c 	return xfs_rtmodify_summary_int(mp, tp, log, bbno, 0, rbpp, rsb, sum);
log                58 fs/xfs/xfs_rtalloc.c 	int		log;		/* loop counter, log2 of ext. size */
log                68 fs/xfs/xfs_rtalloc.c 	for (log = low; log <= high; log++) {
log                72 fs/xfs/xfs_rtalloc.c 		error = xfs_rtget_summary(mp, tp, log, bbno, rbpp, rsb, &sum);
log                90 fs/xfs/xfs_rtalloc.c 	if (mp->m_rsum_cache && log > mp->m_rsum_cache[bbno])
log                91 fs/xfs/xfs_rtalloc.c 		mp->m_rsum_cache[bbno] = log;
log               109 fs/xfs/xfs_rtalloc.c 	int		log;		/* summary level number (log length) */
log               114 fs/xfs/xfs_rtalloc.c 	for (log = omp->m_rsumlevels - 1; log >= 0; log--) {
log               118 fs/xfs/xfs_rtalloc.c 			error = xfs_rtget_summary(omp, tp, log, bbno, &bp,
log               124 fs/xfs/xfs_rtalloc.c 			error = xfs_rtmodify_summary(omp, tp, log, bbno, -sum,
log               128 fs/xfs/xfs_rtalloc.c 			error = xfs_rtmodify_summary(nmp, tp, log, bbno, sum,
log               117 fs/xfs/xfs_rtalloc.h 			     int log, xfs_rtblock_t bbno, int delta,
log               120 fs/xfs/xfs_rtalloc.h int xfs_rtmodify_summary(struct xfs_mount *mp, struct xfs_trans *tp, int log,
log               322 fs/xfs/xfs_sysfs.c 	struct xlog *log = to_xlog(kobject);
log               324 fs/xfs/xfs_sysfs.c 	spin_lock(&log->l_icloglock);
log               325 fs/xfs/xfs_sysfs.c 	cycle = log->l_curr_cycle;
log               326 fs/xfs/xfs_sysfs.c 	block = log->l_curr_block;
log               327 fs/xfs/xfs_sysfs.c 	spin_unlock(&log->l_icloglock);
log               340 fs/xfs/xfs_sysfs.c 	struct xlog *log = to_xlog(kobject);
log               342 fs/xfs/xfs_sysfs.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block);
log               355 fs/xfs/xfs_sysfs.c 	struct xlog *log = to_xlog(kobject);
log               357 fs/xfs/xfs_sysfs.c 	xlog_crack_grant_head(&log->l_reserve_head.grant, &cycle, &bytes);
log               369 fs/xfs/xfs_sysfs.c 	struct xlog *log = to_xlog(kobject);
log               371 fs/xfs/xfs_sysfs.c 	xlog_crack_grant_head(&log->l_write_head.grant, &cycle, &bytes);
log               931 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xlog_ticket *tic),
log               932 fs/xfs/xfs_trace.h 	TP_ARGS(log, tic),
log               951 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log               957 fs/xfs/xfs_trace.h 		__entry->reserveq = list_empty(&log->l_reserve_head.waiters);
log               958 fs/xfs/xfs_trace.h 		__entry->writeq = list_empty(&log->l_write_head.waiters);
log               959 fs/xfs/xfs_trace.h 		xlog_crack_grant_head(&log->l_reserve_head.grant,
log               962 fs/xfs/xfs_trace.h 		xlog_crack_grant_head(&log->l_write_head.grant,
log               965 fs/xfs/xfs_trace.h 		__entry->curr_cycle = log->l_curr_cycle;
log               966 fs/xfs/xfs_trace.h 		__entry->curr_block = log->l_curr_block;
log               967 fs/xfs/xfs_trace.h 		__entry->tail_lsn = atomic64_read(&log->l_tail_lsn);
log               996 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xlog_ticket *tic), \
log               997 fs/xfs/xfs_trace.h 	TP_ARGS(log, tic))
log              1104 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, xfs_lsn_t new_lsn),
log              1105 fs/xfs/xfs_trace.h 	TP_ARGS(log, new_lsn),
log              1113 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              1115 fs/xfs/xfs_trace.h 		__entry->old_lsn = atomic64_read(&log->l_tail_lsn);
log              1116 fs/xfs/xfs_trace.h 		__entry->last_sync_lsn = atomic64_read(&log->l_last_sync_lsn);
log              1950 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, xfs_daddr_t headblk, xfs_daddr_t tailblk),
log              1951 fs/xfs/xfs_trace.h 	TP_ARGS(log, headblk, tailblk),
log              1958 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              1968 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xlog_rec_header *rhead, int pass),
log              1969 fs/xfs/xfs_trace.h 	TP_ARGS(log, rhead, pass),
log              1978 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              1991 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xlog_recover *trans,
log              1993 fs/xfs/xfs_trace.h 	TP_ARGS(log, trans, item, pass),
log              2005 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              2028 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xlog_recover *trans, \
log              2030 fs/xfs/xfs_trace.h 	TP_ARGS(log, trans, item, pass))
log              2039 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_buf_log_format *buf_f),
log              2040 fs/xfs/xfs_trace.h 	TP_ARGS(log, buf_f),
log              2050 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              2069 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_buf_log_format *buf_f), \
log              2070 fs/xfs/xfs_trace.h 	TP_ARGS(log, buf_f))
log              2083 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_inode_log_format *in_f),
log              2084 fs/xfs/xfs_trace.h 	TP_ARGS(log, in_f),
log              2097 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              2121 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_inode_log_format *in_f), \
log              2122 fs/xfs/xfs_trace.h 	TP_ARGS(log, in_f))
log              2129 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_icreate_log *in_f),
log              2130 fs/xfs/xfs_trace.h 	TP_ARGS(log, in_f),
log              2141 fs/xfs/xfs_trace.h 		__entry->dev = log->l_mp->m_super->s_dev;
log              2156 fs/xfs/xfs_trace.h 	TP_PROTO(struct xlog *log, struct xfs_icreate_log *in_f), \
log              2157 fs/xfs/xfs_trace.h 	TP_ARGS(log, in_f))
log               323 include/linux/bpf_verifier.h static inline bool bpf_verifier_log_full(const struct bpf_verifier_log *log)
log               325 include/linux/bpf_verifier.h 	return log->len_used >= log->len_total - 1;
log               334 include/linux/bpf_verifier.h static inline bool bpf_verifier_log_needed(const struct bpf_verifier_log *log)
log               336 include/linux/bpf_verifier.h 	return log->level && log->ubuf && !bpf_verifier_log_full(log);
log               369 include/linux/bpf_verifier.h 	struct bpf_verifier_log log;
log               396 include/linux/bpf_verifier.h __printf(2, 0) void bpf_verifier_vlog(struct bpf_verifier_log *log,
log                35 include/linux/dm-dirty-log.h 	int (*ctr)(struct dm_dirty_log *log, struct dm_target *ti,
log                37 include/linux/dm-dirty-log.h 	void (*dtr)(struct dm_dirty_log *log);
log                43 include/linux/dm-dirty-log.h 	int (*presuspend)(struct dm_dirty_log *log);
log                44 include/linux/dm-dirty-log.h 	int (*postsuspend)(struct dm_dirty_log *log);
log                45 include/linux/dm-dirty-log.h 	int (*resume)(struct dm_dirty_log *log);
log                51 include/linux/dm-dirty-log.h 	uint32_t (*get_region_size)(struct dm_dirty_log *log);
log                57 include/linux/dm-dirty-log.h 	int (*is_clean)(struct dm_dirty_log *log, region_t region);
log                70 include/linux/dm-dirty-log.h 	int (*in_sync)(struct dm_dirty_log *log, region_t region,
log                77 include/linux/dm-dirty-log.h 	int (*flush)(struct dm_dirty_log *log);
log                85 include/linux/dm-dirty-log.h 	void (*mark_region)(struct dm_dirty_log *log, region_t region);
log                86 include/linux/dm-dirty-log.h 	void (*clear_region)(struct dm_dirty_log *log, region_t region);
log               100 include/linux/dm-dirty-log.h 	int (*get_resync_work)(struct dm_dirty_log *log, region_t *region);
log               107 include/linux/dm-dirty-log.h 	void (*set_region_sync)(struct dm_dirty_log *log,
log               113 include/linux/dm-dirty-log.h 	region_t (*get_sync_count)(struct dm_dirty_log *log);
log               118 include/linux/dm-dirty-log.h 	int (*status)(struct dm_dirty_log *log, status_type_t status_type,
log               129 include/linux/dm-dirty-log.h 	int (*is_remote_recovering)(struct dm_dirty_log *log, region_t region);
log               143 include/linux/dm-dirty-log.h void dm_dirty_log_destroy(struct dm_dirty_log *log);
log                41 include/linux/dm-region-hash.h 		struct dm_dirty_log *log, uint32_t region_size,
log              1721 include/linux/efi.h 	u8	log[];
log                96 include/linux/fs_context.h 	struct fc_log		*log;		/* Logging buffer */
log               809 include/linux/kvm_host.h 			struct kvm_dirty_log *log, int *is_dirty);
log               812 include/linux/kvm_host.h 			      struct kvm_dirty_log *log, bool *flush);
log               814 include/linux/kvm_host.h 				struct kvm_clear_dirty_log *log, bool *flush);
log               822 include/linux/kvm_host.h 				struct kvm_dirty_log *log);
log               824 include/linux/kvm_host.h 				  struct kvm_clear_dirty_log *log);
log               121 include/linux/tpm.h 	struct tpm_bios_log log;
log               110 include/media/v4l2-ctrls.h 	void (*log)(const struct v4l2_ctrl *ctrl);
log                35 include/net/netfilter/nf_log.h 		} log;
log                75 include/net/udp.h 	unsigned int		log;
log               253 kernel/bpf/btf.c 	struct bpf_verifier_log log;
log               664 kernel/bpf/btf.c __printf(2, 3) static void __btf_verifier_log(struct bpf_verifier_log *log,
log               670 kernel/bpf/btf.c 	bpf_verifier_vlog(log, fmt, args);
log               677 kernel/bpf/btf.c 	struct bpf_verifier_log *log = &env->log;
log               680 kernel/bpf/btf.c 	if (!bpf_verifier_log_needed(log))
log               684 kernel/bpf/btf.c 	bpf_verifier_vlog(log, fmt, args);
log               693 kernel/bpf/btf.c 	struct bpf_verifier_log *log = &env->log;
log               698 kernel/bpf/btf.c 	if (!bpf_verifier_log_needed(log))
log               701 kernel/bpf/btf.c 	__btf_verifier_log(log, "[%u] %s %s%s",
log               711 kernel/bpf/btf.c 		__btf_verifier_log(log, " ");
log               713 kernel/bpf/btf.c 		bpf_verifier_vlog(log, fmt, args);
log               717 kernel/bpf/btf.c 	__btf_verifier_log(log, "\n");
log               731 kernel/bpf/btf.c 	struct bpf_verifier_log *log = &env->log;
log               735 kernel/bpf/btf.c 	if (!bpf_verifier_log_needed(log))
log               748 kernel/bpf/btf.c 		__btf_verifier_log(log,
log               755 kernel/bpf/btf.c 		__btf_verifier_log(log, "\t%s type_id=%u bits_offset=%u",
log               760 kernel/bpf/btf.c 		__btf_verifier_log(log, " ");
log               762 kernel/bpf/btf.c 		bpf_verifier_vlog(log, fmt, args);
log               766 kernel/bpf/btf.c 	__btf_verifier_log(log, "\n");
log               775 kernel/bpf/btf.c 	struct bpf_verifier_log *log = &env->log;
log               778 kernel/bpf/btf.c 	if (!bpf_verifier_log_needed(log))
log               783 kernel/bpf/btf.c 	__btf_verifier_log(log, "\t type_id=%u offset=%u size=%u",
log               786 kernel/bpf/btf.c 		__btf_verifier_log(log, " ");
log               788 kernel/bpf/btf.c 		bpf_verifier_vlog(log, fmt, args);
log               792 kernel/bpf/btf.c 	__btf_verifier_log(log, "\n");
log               798 kernel/bpf/btf.c 	struct bpf_verifier_log *log = &env->log;
log               802 kernel/bpf/btf.c 	if (!bpf_verifier_log_needed(log))
log               806 kernel/bpf/btf.c 	__btf_verifier_log(log, "magic: 0x%x\n", hdr->magic);
log               807 kernel/bpf/btf.c 	__btf_verifier_log(log, "version: %u\n", hdr->version);
log               808 kernel/bpf/btf.c 	__btf_verifier_log(log, "flags: 0x%x\n", hdr->flags);
log               809 kernel/bpf/btf.c 	__btf_verifier_log(log, "hdr_len: %u\n", hdr->hdr_len);
log               810 kernel/bpf/btf.c 	__btf_verifier_log(log, "type_off: %u\n", hdr->type_off);
log               811 kernel/bpf/btf.c 	__btf_verifier_log(log, "type_len: %u\n", hdr->type_len);
log               812 kernel/bpf/btf.c 	__btf_verifier_log(log, "str_off: %u\n", hdr->str_off);
log               813 kernel/bpf/btf.c 	__btf_verifier_log(log, "str_len: %u\n", hdr->str_len);
log               814 kernel/bpf/btf.c 	__btf_verifier_log(log, "btf_total_size: %u\n", btf_data_size);
log              3290 kernel/bpf/btf.c 	struct bpf_verifier_log *log;
log              3302 kernel/bpf/btf.c 	log = &env->log;
log              3307 kernel/bpf/btf.c 		log->level = log_level;
log              3308 kernel/bpf/btf.c 		log->ubuf = log_ubuf;
log              3309 kernel/bpf/btf.c 		log->len_total = log_size;
log              3312 kernel/bpf/btf.c 		if (log->len_total < 128 || log->len_total > UINT_MAX >> 8 ||
log              3313 kernel/bpf/btf.c 		    !log->level || !log->ubuf) {
log              3354 kernel/bpf/btf.c 	if (log->level && bpf_verifier_log_full(log)) {
log               232 kernel/bpf/verifier.c void bpf_verifier_vlog(struct bpf_verifier_log *log, const char *fmt,
log               237 kernel/bpf/verifier.c 	n = vscnprintf(log->kbuf, BPF_VERIFIER_TMP_LOG_SIZE, fmt, args);
log               242 kernel/bpf/verifier.c 	n = min(log->len_total - log->len_used - 1, n);
log               243 kernel/bpf/verifier.c 	log->kbuf[n] = '\0';
log               245 kernel/bpf/verifier.c 	if (!copy_to_user(log->ubuf + log->len_used, log->kbuf, n + 1))
log               246 kernel/bpf/verifier.c 		log->len_used += n;
log               248 kernel/bpf/verifier.c 		log->ubuf = NULL;
log               260 kernel/bpf/verifier.c 	if (!bpf_verifier_log_needed(&env->log))
log               264 kernel/bpf/verifier.c 	bpf_verifier_vlog(&env->log, fmt, args);
log               274 kernel/bpf/verifier.c 	if (!bpf_verifier_log_needed(&env->log))
log               278 kernel/bpf/verifier.c 	bpf_verifier_vlog(&env->log, fmt, args);
log               296 kernel/bpf/verifier.c 	if (!bpf_verifier_log_needed(&env->log))
log               307 kernel/bpf/verifier.c 		bpf_verifier_vlog(&env->log, prefix_fmt, args);
log              1153 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL2)
log              1459 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL) {
log              1718 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL)
log              1805 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL) {
log              2198 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL)
log              3825 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL) {
log              3865 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL) {
log              6016 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_LEVEL)
log              7264 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL2)
log              7278 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL2)
log              7604 kernel/bpf/verifier.c 			if (env->log.level & BPF_LOG_LEVEL) {
log              7622 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL2 ||
log              7623 kernel/bpf/verifier.c 		    (env->log.level & BPF_LOG_LEVEL && do_print_state)) {
log              7624 kernel/bpf/verifier.c 			if (env->log.level & BPF_LOG_LEVEL2)
log              7635 kernel/bpf/verifier.c 		if (env->log.level & BPF_LOG_LEVEL) {
log              9273 kernel/bpf/verifier.c 	if (env->log.level & BPF_LOG_STATS) {
log              9298 kernel/bpf/verifier.c 	struct bpf_verifier_log *log;
log              9312 kernel/bpf/verifier.c 	log = &env->log;
log              9334 kernel/bpf/verifier.c 		log->level = attr->log_level;
log              9335 kernel/bpf/verifier.c 		log->ubuf = (char __user *) (unsigned long) attr->log_buf;
log              9336 kernel/bpf/verifier.c 		log->len_total = attr->log_size;
log              9340 kernel/bpf/verifier.c 		if (log->len_total < 128 || log->len_total > UINT_MAX >> 2 ||
log              9341 kernel/bpf/verifier.c 		    !log->level || !log->ubuf || log->level & ~BPF_LOG_MASK)
log              9436 kernel/bpf/verifier.c 	if (log->level && bpf_verifier_log_full(log))
log              9438 kernel/bpf/verifier.c 	if (log->level && !log->ubuf) {
log               130 kernel/seccomp.c 	bool log;
log               536 kernel/seccomp.c 		filter->log = true;
log               637 kernel/seccomp.c 	bool log = false;
log               643 kernel/seccomp.c 		log = requested && seccomp_actions_logged & SECCOMP_LOG_TRAP;
log               646 kernel/seccomp.c 		log = requested && seccomp_actions_logged & SECCOMP_LOG_ERRNO;
log               649 kernel/seccomp.c 		log = requested && seccomp_actions_logged & SECCOMP_LOG_TRACE;
log               652 kernel/seccomp.c 		log = requested && seccomp_actions_logged & SECCOMP_LOG_USER_NOTIF;
log               655 kernel/seccomp.c 		log = seccomp_actions_logged & SECCOMP_LOG_LOG;
log               658 kernel/seccomp.c 		log = seccomp_actions_logged & SECCOMP_LOG_KILL_THREAD;
log               662 kernel/seccomp.c 		log = seccomp_actions_logged & SECCOMP_LOG_KILL_PROCESS;
log               671 kernel/seccomp.c 	if (!log)
log               909 kernel/seccomp.c 	seccomp_log(this_syscall, 0, action, match ? match->log : false);
log              1562 kernel/seccomp.c 	if (filter->log)
log              1597 kernel/seccomp.c 	u32		log;
log              1623 kernel/seccomp.c 		if (!(actions_logged & cur->log))
log              1654 kernel/seccomp.c 			*action_logged = cur->log;
log              1215 mm/vmalloc.c   	unsigned int log;
log              1217 mm/vmalloc.c   	log = fls(num_online_cpus());
log              1219 mm/vmalloc.c   	return log * (32UL * 1024 * 1024 / PAGE_SIZE);
log                87 net/bridge/netfilter/ebt_log.c 	       '0' + loginfo->u.log.level, prefix,
log                93 net/bridge/netfilter/ebt_log.c 		bitmask = loginfo->u.log.logflags;
log               186 net/bridge/netfilter/ebt_log.c 	li.u.log.level = info->loglevel;
log               187 net/bridge/netfilter/ebt_log.c 	li.u.log.logflags = info->bitmask;
log               149 net/ipv4/netfilter/ip_tables.c 		.log = {
log                28 net/ipv4/netfilter/nf_log_arp.c 		.log = {
log                27 net/ipv4/netfilter/nf_log_ipv4.c 		.log = {
log                44 net/ipv4/netfilter/nf_log_ipv4.c 		logflags = info->u.log.logflags;
log               280 net/ipv4/netfilter/nf_log_ipv4.c 		logflags = info->u.log.logflags;
log               131 net/ipv4/udp.c 			       struct sock *sk, unsigned int log)
log               152 net/ipv4/udp.c 				__set_bit(udp_sk(sk2)->udp_port_hash >> log,
log               255 net/ipv4/udp.c 					    udptable->log);
log               265 net/ipv4/udp.c 				    !test_bit(snum >> udptable->log, bitmap) &&
log              3008 net/ipv4/udp.c 					      &table->log,
log               174 net/ipv6/netfilter/ip6_tables.c 		.log = {
log                28 net/ipv6/netfilter/nf_log_ipv6.c 		.log = {
log                50 net/ipv6/netfilter/nf_log_ipv6.c 		logflags = info->u.log.logflags;
log               293 net/ipv6/netfilter/nf_log_ipv6.c 		logflags = info->u.log.logflags;
log                30 net/netfilter/nf_log.c 	struct nf_logger *log;
log                37 net/netfilter/nf_log.c 		log = nft_log_dereference(loggers[pf][i]);
log                38 net/netfilter/nf_log.c 		if (!strncasecmp(str_logger, log->name, strlen(log->name)))
log                39 net/netfilter/nf_log.c 			return log;
log                47 net/netfilter/nf_log.c 	const struct nf_logger *log;
log                53 net/netfilter/nf_log.c 	log = nft_log_dereference(net->nf.nf_loggers[pf]);
log                54 net/netfilter/nf_log.c 	if (log == NULL)
log                66 net/netfilter/nf_log.c 	const struct nf_logger *log;
log                70 net/netfilter/nf_log.c 		log = nft_log_dereference(net->nf.nf_loggers[i]);
log                71 net/netfilter/nf_log.c 		if (log == logger)
log               114 net/netfilter/nf_log.c 	const struct nf_logger *log;
log               119 net/netfilter/nf_log.c 		log = nft_log_dereference(loggers[i][logger->type]);
log               120 net/netfilter/nf_log.c 		if (log == logger)
log               160 net/netfilter/nf_log_common.c 	       '0' + loginfo->u.log.level, prefix,
log               101 net/netfilter/nfnetlink_log.c __instance_lookup(struct nfnl_log_net *log, u_int16_t group_num)
log               106 net/netfilter/nfnetlink_log.c 	head = &log->instance_table[instance_hashfn(group_num)];
log               121 net/netfilter/nfnetlink_log.c instance_lookup_get(struct nfnl_log_net *log, u_int16_t group_num)
log               126 net/netfilter/nfnetlink_log.c 	inst = __instance_lookup(log, group_num);
log               158 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(net);
log               161 net/netfilter/nfnetlink_log.c 	spin_lock_bh(&log->instances_lock);
log               162 net/netfilter/nfnetlink_log.c 	if (__instance_lookup(log, group_num)) {
log               198 net/netfilter/nfnetlink_log.c 		       &log->instance_table[instance_hashfn(group_num)]);
log               201 net/netfilter/nfnetlink_log.c 	spin_unlock_bh(&log->instances_lock);
log               206 net/netfilter/nfnetlink_log.c 	spin_unlock_bh(&log->instances_lock);
log               235 net/netfilter/nfnetlink_log.c instance_destroy(struct nfnl_log_net *log,
log               238 net/netfilter/nfnetlink_log.c 	spin_lock_bh(&log->instances_lock);
log               240 net/netfilter/nfnetlink_log.c 	spin_unlock_bh(&log->instances_lock);
log               442 net/netfilter/nfnetlink_log.c __build_packet_message(struct nfnl_log_net *log,
log               627 net/netfilter/nfnetlink_log.c 			 htonl(atomic_inc_return(&log->global_seq))))
log               689 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(net);
log               699 net/netfilter/nfnetlink_log.c 	inst = instance_lookup_get(log, li->u.ulog.group);
log               793 net/netfilter/nfnetlink_log.c 	__build_packet_message(log, inst, skb, data_len, pf,
log               822 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(n->net);
log               828 net/netfilter/nfnetlink_log.c 		spin_lock_bh(&log->instances_lock);
log               832 net/netfilter/nfnetlink_log.c 			struct hlist_head *head = &log->instance_table[i];
log               839 net/netfilter/nfnetlink_log.c 		spin_unlock_bh(&log->instances_lock);
log               881 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(net);
log               899 net/netfilter/nfnetlink_log.c 	inst = instance_lookup_get(log, group_num);
log               949 net/netfilter/nfnetlink_log.c 			instance_destroy(log, inst);
log              1018 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log;
log              1022 net/netfilter/nfnetlink_log.c 	log = nfnl_log_pernet(net);
log              1025 net/netfilter/nfnetlink_log.c 		struct hlist_head *head = &log->instance_table[st->bucket];
log              1038 net/netfilter/nfnetlink_log.c 		struct nfnl_log_net *log;
log              1044 net/netfilter/nfnetlink_log.c 		log = nfnl_log_pernet(net);
log              1045 net/netfilter/nfnetlink_log.c 		head = &log->instance_table[st->bucket];
log              1106 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(net);
log              1114 net/netfilter/nfnetlink_log.c 		INIT_HLIST_HEAD(&log->instance_table[i]);
log              1115 net/netfilter/nfnetlink_log.c 	spin_lock_init(&log->instances_lock);
log              1133 net/netfilter/nfnetlink_log.c 	struct nfnl_log_net *log = nfnl_log_pernet(net);
log              1141 net/netfilter/nfnetlink_log.c 		WARN_ON_ONCE(!hlist_empty(&log->instance_table[i]));
log               111 net/netfilter/nft_log.c 	    priv->loginfo.u.log.level == NFT_LOGLEVEL_AUDIT) {
log               163 net/netfilter/nft_log.c 			li->u.log.level =
log               166 net/netfilter/nft_log.c 			li->u.log.level = NFT_LOGLEVEL_WARNING;
log               168 net/netfilter/nft_log.c 		if (li->u.log.level > NFT_LOGLEVEL_AUDIT) {
log               174 net/netfilter/nft_log.c 			li->u.log.logflags =
log               176 net/netfilter/nft_log.c 			if (li->u.log.logflags & ~NF_LOG_MASK) {
log               196 net/netfilter/nft_log.c 	if (li->u.log.level == NFT_LOGLEVEL_AUDIT)
log               220 net/netfilter/nft_log.c 	if (li->u.log.level == NFT_LOGLEVEL_AUDIT)
log               236 net/netfilter/nft_log.c 		if (nla_put_be32(skb, NFTA_LOG_LEVEL, htonl(li->u.log.level)))
log               239 net/netfilter/nft_log.c 		if (li->u.log.logflags) {
log               241 net/netfilter/nft_log.c 					 htonl(li->u.log.logflags)))
log                36 net/netfilter/xt_LOG.c 	li.u.log.level = loginfo->level;
log                37 net/netfilter/xt_LOG.c 	li.u.log.logflags = loginfo->logflags;
log              1323 net/openvswitch/conntrack.c 			     const struct sw_flow_key *key, bool log)
log              1332 net/openvswitch/conntrack.c 		OVS_NLERR(log, "Unknown helper \"%s\"", name);
log              1348 net/openvswitch/conntrack.c 			OVS_NLERR(log, "Failed to load \"%s\" NAT helper, error: %d",
log              1361 net/openvswitch/conntrack.c 		     struct ovs_conntrack_info *info, bool log)
log              1386 net/openvswitch/conntrack.c 			OVS_NLERR(log, "Unknown NAT attribute (type=%d, max=%d)",
log              1392 net/openvswitch/conntrack.c 			OVS_NLERR(log, "NAT attribute type %d has unexpected length (%d != %d)",
log              1402 net/openvswitch/conntrack.c 				OVS_NLERR(log, "Only one type of NAT may be specified");
log              1447 net/openvswitch/conntrack.c 			OVS_NLERR(log, "Unknown nat attribute (%d)", type);
log              1453 net/openvswitch/conntrack.c 		OVS_NLERR(log, "NAT attribute has %d unknown bytes", rem);
log              1459 net/openvswitch/conntrack.c 			OVS_NLERR(log,
log              1466 net/openvswitch/conntrack.c 		OVS_NLERR(log,
log              1507 net/openvswitch/conntrack.c 		    const char **helper, bool log)
log              1518 net/openvswitch/conntrack.c 			OVS_NLERR(log,
log              1527 net/openvswitch/conntrack.c 			OVS_NLERR(log,
log              1550 net/openvswitch/conntrack.c 				OVS_NLERR(log, "ct_mark mask cannot be 0");
log              1562 net/openvswitch/conntrack.c 				OVS_NLERR(log, "ct_labels mask cannot be 0");
log              1572 net/openvswitch/conntrack.c 				OVS_NLERR(log, "Invalid conntrack helper");
log              1578 net/openvswitch/conntrack.c 			int err = parse_nat(a, info, log);
log              1593 net/openvswitch/conntrack.c 				OVS_NLERR(log, "Invalid conntrack timeout");
log              1600 net/openvswitch/conntrack.c 			OVS_NLERR(log, "Unknown conntrack attr (%d)",
log              1608 net/openvswitch/conntrack.c 		OVS_NLERR(log,
log              1615 net/openvswitch/conntrack.c 		OVS_NLERR(log,
log              1621 net/openvswitch/conntrack.c 		OVS_NLERR(log, "Conntrack attr has %d unknown bytes", rem);
log              1650 net/openvswitch/conntrack.c 		       struct sw_flow_actions **sfa,  bool log)
log              1659 net/openvswitch/conntrack.c 		OVS_NLERR(log, "ct family unspecified");
log              1669 net/openvswitch/conntrack.c 	err = parse_ct(attr, &ct_info, &helper, log);
log              1676 net/openvswitch/conntrack.c 		OVS_NLERR(log, "Failed to allocate conntrack template");
log              1692 net/openvswitch/conntrack.c 		err = ovs_ct_add_helper(&ct_info, helper, key, log);
log              1698 net/openvswitch/conntrack.c 				 sizeof(ct_info), log);
log                21 net/openvswitch/conntrack.h 		       bool log);
log                51 net/openvswitch/conntrack.h 				     struct sw_flow_actions **acts, bool log)
log               547 net/openvswitch/datapath.c 	bool log = !a[OVS_PACKET_ATTR_PROBE];
log               577 net/openvswitch/datapath.c 					     packet, &flow->key, log);
log               582 net/openvswitch/datapath.c 				   &flow->key, &acts, log);
log               909 net/openvswitch/datapath.c 	bool log = !a[OVS_FLOW_ATTR_PROBE];
log               914 net/openvswitch/datapath.c 		OVS_NLERR(log, "Flow key attr not present in new flow.");
log               918 net/openvswitch/datapath.c 		OVS_NLERR(log, "Flow actions attr not present in new flow.");
log               934 net/openvswitch/datapath.c 				  a[OVS_FLOW_ATTR_MASK], log);
log               940 net/openvswitch/datapath.c 				       &new_flow->key, log);
log               952 net/openvswitch/datapath.c 				     &new_flow->key, &acts, log);
log               954 net/openvswitch/datapath.c 		OVS_NLERR(log, "Flow actions may not be safe on all matching packets.");
log              1064 net/openvswitch/datapath.c 						bool log)
log              1071 net/openvswitch/datapath.c 	error = ovs_nla_copy_actions(net, a, &masked_key, &acts, log);
log              1073 net/openvswitch/datapath.c 		OVS_NLERR(log,
log              1100 net/openvswitch/datapath.c 			      bool log)
log              1108 net/openvswitch/datapath.c 					  a[OVS_FLOW_ATTR_MASK], log);
log              1115 net/openvswitch/datapath.c 			OVS_NLERR(log,
log              1122 net/openvswitch/datapath.c 					 &mask, log);
log              1149 net/openvswitch/datapath.c 	bool log = !a[OVS_FLOW_ATTR_PROBE];
log              1152 net/openvswitch/datapath.c 	ufid_present = ovs_nla_get_ufid(&sfid, a[OVS_FLOW_ATTR_UFID], log);
log              1154 net/openvswitch/datapath.c 		OVS_NLERR(log,
log              1160 net/openvswitch/datapath.c 					      &acts, log);
log              1250 net/openvswitch/datapath.c 	bool log = !a[OVS_FLOW_ATTR_PROBE];
log              1253 net/openvswitch/datapath.c 	ufid_present = ovs_nla_get_ufid(&ufid, a[OVS_FLOW_ATTR_UFID], log);
log              1257 net/openvswitch/datapath.c 					log);
log              1259 net/openvswitch/datapath.c 		OVS_NLERR(log,
log              1309 net/openvswitch/datapath.c 	bool log = !a[OVS_FLOW_ATTR_PROBE];
log              1312 net/openvswitch/datapath.c 	ufid_present = ovs_nla_get_ufid(&ufid, a[OVS_FLOW_ATTR_UFID], log);
log              1316 net/openvswitch/datapath.c 					NULL, log);
log               900 net/openvswitch/flow.c 				   struct sw_flow_key *key, bool log)
log               906 net/openvswitch/flow.c 	err = parse_flow_nlattrs(attr, a, &attrs, log);
log               911 net/openvswitch/flow.c 	err = ovs_nla_get_flow_metadata(net, a, attrs, key, log);
log               280 net/openvswitch/flow.h 				   struct sw_flow_key *key, bool log);
log               151 net/openvswitch/flow_netlink.c 			   u64 key_attrs, u64 mask_attrs, bool log)
log               290 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Missing key (keys=%llx, expected=%llx)",
log               298 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Unexpected mask (mask=%llx, allowed=%llx)",
log               463 net/openvswitch/flow_netlink.c 				u64 *attrsp, bool log, bool nz)
log               475 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Key type %d is out of range max %d",
log               481 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Duplicate key (type %d).", type);
log               487 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Key %d has unexpected len %d expected %d",
log               498 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Message has %d unknown bytes.", rem);
log               508 net/openvswitch/flow_netlink.c 				   bool log)
log               510 net/openvswitch/flow_netlink.c 	return __parse_flow_nlattrs(attr, a, attrsp, log, true);
log               514 net/openvswitch/flow_netlink.c 		       u64 *attrsp, bool log)
log               516 net/openvswitch/flow_netlink.c 	return __parse_flow_nlattrs(attr, a, attrsp, log, false);
log               521 net/openvswitch/flow_netlink.c 				     bool log)
log               526 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Geneve option length err (len %d, max %zu).",
log               532 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Geneve opt len %d is not a multiple of 4.",
log               556 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Geneve option len %d != mask len %d",
log               572 net/openvswitch/flow_netlink.c 				     bool log)
log               586 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "VXLAN extension %d out of range max %d",
log               593 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "VXLAN extension %d has unexpected len %d expected %d",
log               604 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Unknown VXLAN extension attribute %d",
log               610 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "VXLAN extension message has %d unknown bytes.",
log               628 net/openvswitch/flow_netlink.c 				      bool log)
log               636 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "ERSPAN option length err (len %d, max %zu).",
log               655 net/openvswitch/flow_netlink.c 			      bool log)
log               669 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Tunnel attr %d out of range max %d",
log               676 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Tunnel attr %d has unexpected len %d expected %d",
log               735 net/openvswitch/flow_netlink.c 				OVS_NLERR(log, "Multiple metadata blocks provided");
log               739 net/openvswitch/flow_netlink.c 			err = genev_tun_opt_from_nlattr(a, match, is_mask, log);
log               748 net/openvswitch/flow_netlink.c 				OVS_NLERR(log, "Multiple metadata blocks provided");
log               752 net/openvswitch/flow_netlink.c 			err = vxlan_tun_opt_from_nlattr(a, match, is_mask, log);
log               763 net/openvswitch/flow_netlink.c 				OVS_NLERR(log, "Multiple metadata blocks provided");
log               768 net/openvswitch/flow_netlink.c 							 log);
log               780 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Unknown IP tunnel attribute %d",
log               794 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "IP tunnel attribute has %d unknown bytes.",
log               800 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Mixed IPv4 and IPv6 tunnel attributes");
log               806 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IP tunnel dst address not specified");
log               818 net/openvswitch/flow_netlink.c 					OVS_NLERR(log, "IPv4 tun info is not correct");
log               822 net/openvswitch/flow_netlink.c 				OVS_NLERR(log, "IPv4 tunnel dst address is zero");
log               827 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IPv6 tunnel dst address is zero");
log               832 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IP tunnel TTL not specified.");
log               985 net/openvswitch/flow_netlink.c 				      const struct nlattr **a, bool log)
log               998 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Invalid %s frame", (inner) ? "C-VLAN" : "VLAN");
log              1007 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "%s TCI does not have VLAN_CFI_MASK bit set.",
log              1012 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Truncated %s header has non-zero encap attribute.",
log              1023 net/openvswitch/flow_netlink.c 					   const struct nlattr **a, bool log)
log              1038 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Encap mask attribute is set for non-%s frame.",
log              1050 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Must have an exact match on %s TPID (mask=%x).",
log              1055 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "%s TCI mask does not have exact match for VLAN_CFI_MASK bit.",
log              1066 net/openvswitch/flow_netlink.c 				     bool log)
log              1073 net/openvswitch/flow_netlink.c 						 a, log);
log              1076 net/openvswitch/flow_netlink.c 						      a, log);
log              1091 net/openvswitch/flow_netlink.c 		err = parse_flow_nlattrs(encap, a, key_attrs, log);
log              1093 net/openvswitch/flow_netlink.c 		err = parse_flow_mask_nlattrs(encap, a, key_attrs, log);
log              1100 net/openvswitch/flow_netlink.c 				   bool is_mask, bool log)
log              1106 net/openvswitch/flow_netlink.c 					is_mask, log);
log              1113 net/openvswitch/flow_netlink.c 						is_mask, log);
log              1123 net/openvswitch/flow_netlink.c 				       bool is_mask, bool log)
log              1132 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "EtherType %x is less than min %x",
log              1144 net/openvswitch/flow_netlink.c 				 bool is_mask, bool log)
log              1174 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Port %d exceeds max allowable %d",
log              1193 net/openvswitch/flow_netlink.c 				       is_mask, log) < 0)
log              1203 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "ct_state flags %08x unsupported",
log              1275 net/openvswitch/flow_netlink.c 						   log);
log              1376 net/openvswitch/flow_netlink.c 				   bool is_push_nsh, bool log)
log              1394 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "nsh attr %d is out of range max %d",
log              1402 net/openvswitch/flow_netlink.c 			    log,
log              1446 net/openvswitch/flow_netlink.c 				    log,
log              1455 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Unknown nsh attribute %d",
log              1462 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "nsh attribute has %d unknown bytes.", rem);
log              1497 net/openvswitch/flow_netlink.c 				bool is_mask, bool log)
log              1501 net/openvswitch/flow_netlink.c 	err = metadata_from_nlattrs(net, match, &attrs, a, is_mask, log);
log              1519 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "VLAN attribute unexpected.");
log              1525 net/openvswitch/flow_netlink.c 							  log);
log              1532 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Either Ethernet header or EtherType is required.");
log              1541 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IPv4 frag type %d is out of range max %d",
log              1565 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IPv6 frag type %d is out of range max %d",
log              1571 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "IPv6 flow label %x is out of range (max=%x)",
log              1603 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Unknown ARP opcode (opcode=%d).",
log              1624 net/openvswitch/flow_netlink.c 					    is_mask, false, log) < 0)
log              1711 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "Unknown key attributes %llx",
log              1761 net/openvswitch/flow_netlink.c 		      bool log)
log              1769 net/openvswitch/flow_netlink.c 	err = parse_flow_nlattrs(nla_key, a, &key_attrs, log);
log              1773 net/openvswitch/flow_netlink.c 	err = parse_vlan_from_nlattrs(match, &key_attrs, a, false, log);
log              1777 net/openvswitch/flow_netlink.c 	err = ovs_key_from_nlattrs(net, match, key_attrs, a, false, log);
log              1811 net/openvswitch/flow_netlink.c 		err = parse_flow_mask_nlattrs(nla_mask, a, &mask_attrs, log);
log              1819 net/openvswitch/flow_netlink.c 		err = parse_vlan_from_nlattrs(match, &mask_attrs, a, true, log);
log              1824 net/openvswitch/flow_netlink.c 					   log);
log              1829 net/openvswitch/flow_netlink.c 	if (!match_validate(match, key_attrs, mask_attrs, log))
log              1837 net/openvswitch/flow_netlink.c static size_t get_ufid_len(const struct nlattr *attr, bool log)
log              1846 net/openvswitch/flow_netlink.c 		OVS_NLERR(log, "ufid size %u bytes exceeds the range (1, %d)",
log              1858 net/openvswitch/flow_netlink.c 		      bool log)
log              1860 net/openvswitch/flow_netlink.c 	sfid->ufid_len = get_ufid_len(attr, log);
log              1868 net/openvswitch/flow_netlink.c 			   const struct sw_flow_key *key, bool log)
log              1872 net/openvswitch/flow_netlink.c 	if (ovs_nla_get_ufid(sfid, ufid, log))
log              1912 net/openvswitch/flow_netlink.c 			      u64 attrs, struct sw_flow_key *key, bool log)
log              1928 net/openvswitch/flow_netlink.c 	return metadata_from_nlattrs(net, &match, &attrs, a, false, log);
log              2317 net/openvswitch/flow_netlink.c 				       int attr_len, bool log)
log              2333 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Flow action size exceeds max %u",
log              2356 net/openvswitch/flow_netlink.c 				   int attrtype, void *data, int len, bool log)
log              2360 net/openvswitch/flow_netlink.c 	a = reserve_sfa_size(sfa, nla_attr_size(len), log);
log              2375 net/openvswitch/flow_netlink.c 		       int len, bool log)
log              2379 net/openvswitch/flow_netlink.c 	a = __add_action(sfa, attrtype, data, len, log);
log              2385 net/openvswitch/flow_netlink.c 					  int attrtype, bool log)
log              2390 net/openvswitch/flow_netlink.c 	err = ovs_nla_add_action(sfa, attrtype, NULL, 0, log);
log              2409 net/openvswitch/flow_netlink.c 				  __be16 eth_type, __be16 vlan_tci, bool log);
log              2415 net/openvswitch/flow_netlink.c 				    bool log, bool last)
log              2442 net/openvswitch/flow_netlink.c 	start = add_nested_action_start(sfa, OVS_ACTION_ATTR_SAMPLE, log);
log              2461 net/openvswitch/flow_netlink.c 				 log);
log              2466 net/openvswitch/flow_netlink.c 				     eth_type, vlan_tci, log);
log              2481 net/openvswitch/flow_netlink.c 				   bool log, bool last)
log              2489 net/openvswitch/flow_netlink.c 	start = add_nested_action_start(sfa, OVS_ACTION_ATTR_CLONE, log);
log              2496 net/openvswitch/flow_netlink.c 				 sizeof(exec), log);
log              2501 net/openvswitch/flow_netlink.c 				     eth_type, vlan_tci, log);
log              2557 net/openvswitch/flow_netlink.c 				     struct sw_flow_actions **sfa, bool log)
log              2570 net/openvswitch/flow_netlink.c 	opts_type = ip_tun_from_nlattr(nla_data(attr), &match, false, log);
log              2591 net/openvswitch/flow_netlink.c 	start = add_nested_action_start(sfa, OVS_ACTION_ATTR_SET, log);
log              2608 net/openvswitch/flow_netlink.c 			 sizeof(*ovs_tun), log);
log              2638 net/openvswitch/flow_netlink.c 			 bool is_push_nsh, bool log)
log              2646 net/openvswitch/flow_netlink.c 				      is_push_nsh, log);
log              2667 net/openvswitch/flow_netlink.c 			u8 mac_proto, __be16 eth_type, bool masked, bool log)
log              2709 net/openvswitch/flow_netlink.c 		err = validate_and_copy_set_tun(a, sfa, log);
log              2795 net/openvswitch/flow_netlink.c 		if (!validate_nsh(nla_data(a), masked, false, log))
log              2812 net/openvswitch/flow_netlink.c 						log);
log              2816 net/openvswitch/flow_netlink.c 		at = __add_action(sfa, key_type, NULL, len, log);
log              2867 net/openvswitch/flow_netlink.c 					   bool log, bool last)
log              2894 net/openvswitch/flow_netlink.c 					log);
log              2905 net/openvswitch/flow_netlink.c 				 sizeof(arg), log);
log              2910 net/openvswitch/flow_netlink.c 		OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_LESS_EQUAL, log);
log              2915 net/openvswitch/flow_netlink.c 				     eth_type, vlan_tci, log);
log              2923 net/openvswitch/flow_netlink.c 		OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_GREATER, log);
log              2928 net/openvswitch/flow_netlink.c 				     eth_type, vlan_tci, log);
log              2939 net/openvswitch/flow_netlink.c 		       struct sw_flow_actions **sfa, bool log)
log              2944 net/openvswitch/flow_netlink.c 	to = reserve_sfa_size(sfa, from->nla_len, log);
log              2955 net/openvswitch/flow_netlink.c 				  __be16 eth_type, __be16 vlan_tci, bool log)
log              3091 net/openvswitch/flow_netlink.c 					   false, log);
log              3099 net/openvswitch/flow_netlink.c 					   true, log);
log              3109 net/openvswitch/flow_netlink.c 						       log, last);
log              3117 net/openvswitch/flow_netlink.c 			err = ovs_ct_copy_action(net, a, key, sfa, log);
log              3179 net/openvswitch/flow_netlink.c 						      log, last);
log              3191 net/openvswitch/flow_netlink.c 							      vlan_tci, log,
log              3200 net/openvswitch/flow_netlink.c 			OVS_NLERR(log, "Unknown Action type %d", type);
log              3204 net/openvswitch/flow_netlink.c 			err = copy_action(a, sfa, log);
log              3219 net/openvswitch/flow_netlink.c 			 struct sw_flow_actions **sfa, bool log)
log              3229 net/openvswitch/flow_netlink.c 				     key->eth.vlan.tci, log);
log                36 net/openvswitch/flow_netlink.h 		       u64 *attrsp, bool log);
log                39 net/openvswitch/flow_netlink.h 			      u64 attrs, struct sw_flow_key *key, bool log);
log                47 net/openvswitch/flow_netlink.h 		      bool log);
log                52 net/openvswitch/flow_netlink.h bool ovs_nla_get_ufid(struct sw_flow_id *, const struct nlattr *, bool log);
log                54 net/openvswitch/flow_netlink.h 			   const struct sw_flow_key *key, bool log);
log                59 net/openvswitch/flow_netlink.h 			 struct sw_flow_actions **sfa, bool log);
log                61 net/openvswitch/flow_netlink.h 		       void *data, int len, bool log);
log               703 net/sched/sch_fq.c static int fq_resize(struct Qdisc *sch, u32 log)
log               710 net/sched/sch_fq.c 	if (q->fq_root && log == q->fq_trees_log)
log               714 net/sched/sch_fq.c 	array = kvmalloc_node(sizeof(struct rb_root) << log, GFP_KERNEL | __GFP_RETRY_MAYFAIL,
log               719 net/sched/sch_fq.c 	for (idx = 0; idx < (1U << log); idx++)
log               726 net/sched/sch_fq.c 		fq_rehash(q, old_fq_root, q->fq_trees_log, array, log);
log               729 net/sched/sch_fq.c 	q->fq_trees_log = log;
log               304 security/tomoyo/audit.c 	char *log;
log               383 security/tomoyo/audit.c 	entry->log = buf;
log               456 security/tomoyo/audit.c 		head->read_buf = ptr->log;
log              1085 tools/perf/util/auxtrace.c 			synth_opts->log = true;
log                98 tools/perf/util/auxtrace.h 	bool			log;
log              3244 tools/perf/util/intel-pt.c 	if (pt->synth_opts.log)
log              1126 tools/perf/util/s390-cpumsf.c 	sf->use_logfile = session->itrace_synth_opts->log;
log               322 tools/power/acpi/tools/acpidbg/acpidbg.c 					ACPI_AML_BATCH_DO(fd, read, log, ret);
log               324 tools/power/acpi/tools/acpidbg/acpidbg.c 					ACPI_AML_DO(fd, read, log, ret);
log               328 tools/power/acpi/tools/acpidbg/acpidbg.c 					ACPI_AML_BATCH_DO(STDOUT_FILENO, write, log, ret);
log               330 tools/power/acpi/tools/acpidbg/acpidbg.c 					ACPI_AML_DO(STDOUT_FILENO, write, log, ret);
log               864 tools/testing/selftests/bpf/test_verifier.c static bool cmp_str_seq(const char *log, const char *exp)
log               882 tools/testing/selftests/bpf/test_verifier.c 		q = strstr(log, needle);
log               888 tools/testing/selftests/bpf/test_verifier.c 		log = q + len;
log                48 tools/testing/selftests/bpf/test_verifier_log.c static int load(char *log, size_t log_len, int log_level)
log                57 tools/testing/selftests/bpf/test_verifier_log.c 	attr.log_buf = ptr_to_u64(log);
log                88 tools/testing/selftests/bpf/test_verifier_log.c static void test_log_good(char *log, size_t buf_len, size_t log_len,
log                94 tools/testing/selftests/bpf/test_verifier_log.c 	memset(log, 1, buf_len);
log                96 tools/testing/selftests/bpf/test_verifier_log.c 	ret = load(log, log_len, 1);
log                99 tools/testing/selftests/bpf/test_verifier_log.c 	len = strnlen(log, buf_len);
log               110 tools/testing/selftests/bpf/test_verifier_log.c 	if (strchr(log, 1)) {
log               115 tools/testing/selftests/bpf/test_verifier_log.c 	check_ones(log + len + 1, buf_len - len - 1,
log               118 tools/testing/selftests/bpf/test_verifier_log.c 	if (memcmp(full_log, log, LOG_SIZE)) {
log               124 tools/testing/selftests/bpf/test_verifier_log.c static void test_log_bad(char *log, size_t log_len, int log_level)
log               128 tools/testing/selftests/bpf/test_verifier_log.c 	ret = load(log, log_len, log_level);
log               130 tools/testing/selftests/bpf/test_verifier_log.c 	if (log)
log               131 tools/testing/selftests/bpf/test_verifier_log.c 		check_ones(log, LOG_SIZE,
log               138 tools/testing/selftests/bpf/test_verifier_log.c 	char log[LOG_SIZE];
log               142 tools/testing/selftests/bpf/test_verifier_log.c 	memset(log, 1, LOG_SIZE);
log               146 tools/testing/selftests/bpf/test_verifier_log.c 	test_log_bad(log, LOG_SIZE, 0);
log               149 tools/testing/selftests/bpf/test_verifier_log.c 	test_log_bad(log, 15, 1);
log               161 tools/testing/selftests/bpf/test_verifier_log.c 	test_log_good(log, LOG_SIZE, want_len + 2, want_len, EACCES, full_log);
log               168 tools/testing/selftests/bpf/test_verifier_log.c 		test_log_good(log, LOG_SIZE, want_len + 1 - i, want_len - i,
log                76 tools/testing/selftests/kvm/include/kvm_util.h void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log);
log                77 tools/testing/selftests/kvm/include/kvm_util.h void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log,
log               277 tools/testing/selftests/kvm/lib/kvm_util.c void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log)
log               279 tools/testing/selftests/kvm/lib/kvm_util.c 	struct kvm_dirty_log args = { .dirty_bitmap = log, .slot = slot };
log               287 tools/testing/selftests/kvm/lib/kvm_util.c void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log,
log               290 tools/testing/selftests/kvm/lib/kvm_util.c 	struct kvm_clear_dirty_log args = { .dirty_bitmap = log, .slot = slot,
log              1018 tools/testing/selftests/seccomp/seccomp_bpf.c 	struct sock_fprog log;
log              1073 tools/testing/selftests/seccomp/seccomp_bpf.c 	FILTER_ALLOC(log);
log              1084 tools/testing/selftests/seccomp/seccomp_bpf.c 	FILTER_FREE(log);
log              1102 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1128 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1161 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1184 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1211 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1234 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1254 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1278 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1301 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1321 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1339 tools/testing/selftests/seccomp/seccomp_bpf.c 	ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log);
log              1245 virt/kvm/arm/arm.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
log              1252 virt/kvm/arm/arm.c 	r = kvm_get_dirty_log_protect(kvm, log, &flush);
log              1261 virt/kvm/arm/arm.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log)
log              1268 virt/kvm/arm/arm.c 	r = kvm_clear_dirty_log_protect(kvm, log, &flush);
log              1176 virt/kvm/kvm_main.c 			struct kvm_dirty_log *log, int *is_dirty)
log              1184 virt/kvm/kvm_main.c 	as_id = log->slot >> 16;
log              1185 virt/kvm/kvm_main.c 	id = (u16)log->slot;
log              1199 virt/kvm/kvm_main.c 	if (copy_to_user(log->dirty_bitmap, memslot->dirty_bitmap, n))
log              1232 virt/kvm/kvm_main.c 			struct kvm_dirty_log *log, bool *flush)
log              1241 virt/kvm/kvm_main.c 	as_id = log->slot >> 16;
log              1242 virt/kvm/kvm_main.c 	id = (u16)log->slot;
log              1288 virt/kvm/kvm_main.c 	if (copy_to_user(log->dirty_bitmap, dirty_bitmap_buffer, n))
log              1302 virt/kvm/kvm_main.c 				struct kvm_clear_dirty_log *log, bool *flush)
log              1312 virt/kvm/kvm_main.c 	as_id = log->slot >> 16;
log              1313 virt/kvm/kvm_main.c 	id = (u16)log->slot;
log              1317 virt/kvm/kvm_main.c 	if (log->first_page & 63)
log              1327 virt/kvm/kvm_main.c 	n = ALIGN(log->num_pages, BITS_PER_LONG) / 8;
log              1329 virt/kvm/kvm_main.c 	if (log->first_page > memslot->npages ||
log              1330 virt/kvm/kvm_main.c 	    log->num_pages > memslot->npages - log->first_page ||
log              1331 virt/kvm/kvm_main.c 	    (log->num_pages < memslot->npages - log->first_page && (log->num_pages & 63)))
log              1336 virt/kvm/kvm_main.c 	if (copy_from_user(dirty_bitmap_buffer, log->dirty_bitmap, n))
log              1340 virt/kvm/kvm_main.c 	for (offset = log->first_page, i = offset / BITS_PER_LONG,
log              1341 virt/kvm/kvm_main.c 		 n = DIV_ROUND_UP(log->num_pages, BITS_PER_LONG); n--;
log              3382 virt/kvm/kvm_main.c 		struct kvm_dirty_log log;
log              3385 virt/kvm/kvm_main.c 		if (copy_from_user(&log, argp, sizeof(log)))
log              3387 virt/kvm/kvm_main.c 		r = kvm_vm_ioctl_get_dirty_log(kvm, &log);
log              3392 virt/kvm/kvm_main.c 		struct kvm_clear_dirty_log log;
log              3395 virt/kvm/kvm_main.c 		if (copy_from_user(&log, argp, sizeof(log)))
log              3397 virt/kvm/kvm_main.c 		r = kvm_vm_ioctl_clear_dirty_log(kvm, &log);
log              3558 virt/kvm/kvm_main.c 		struct kvm_dirty_log log;
log              3563 virt/kvm/kvm_main.c 		log.slot	 = compat_log.slot;
log              3564 virt/kvm/kvm_main.c 		log.padding1	 = compat_log.padding1;
log              3565 virt/kvm/kvm_main.c 		log.padding2	 = compat_log.padding2;
log              3566 virt/kvm/kvm_main.c 		log.dirty_bitmap = compat_ptr(compat_log.dirty_bitmap);
log              3568 virt/kvm/kvm_main.c 		r = kvm_vm_ioctl_get_dirty_log(kvm, &log);