managed           136 arch/x86/kernel/apic/vector.c 	bool managed = irqd_affinity_is_managed(irqd);
managed           166 arch/x86/kernel/apic/vector.c 				managed);
managed           332 arch/x86/kernel/apic/vector.c 	bool managed = irqd_affinity_is_managed(irqd);
managed           344 arch/x86/kernel/apic/vector.c 	irq_matrix_free(vector_matrix, apicd->cpu, vector, managed);
managed           353 arch/x86/kernel/apic/vector.c 	irq_matrix_free(vector_matrix, apicd->prev_cpu, vector, managed);
managed           838 arch/x86/kernel/apic/vector.c 	bool managed = apicd->is_managed;
managed           846 arch/x86/kernel/apic/vector.c 	WARN_ON_ONCE(managed);
managed           848 arch/x86/kernel/apic/vector.c 	trace_vector_free_moved(apicd->irq, cpu, vector, managed);
managed           849 arch/x86/kernel/apic/vector.c 	irq_matrix_free(vector_matrix, cpu, vector, managed);
managed            34 drivers/clk/renesas/r9a06g032-clocks.c 	uint32_t managed: 1;
managed            71 drivers/clk/renesas/r9a06g032-clocks.c 		.managed = 1, .gate = I_GATE(__VA_ARGS__) }
managed           396 drivers/clk/renesas/r9a06g032-clocks.c 		    r9a06g032_clocks[index].managed) {
managed             6 drivers/gpu/drm/nouveau/include/nvif/if000c.h 	__u8  managed;
managed            33 drivers/gpu/drm/nouveau/include/nvif/vmm.h int nvif_vmm_init(struct nvif_mmu *, s32 oclass, bool managed, u64 addr,
managed           115 drivers/gpu/drm/nouveau/nvif/vmm.c nvif_vmm_init(struct nvif_mmu *mmu, s32 oclass, bool managed, u64 addr,
managed           128 drivers/gpu/drm/nouveau/nvif/vmm.c 	args->managed = managed;
managed            31 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h 		int (*ctor)(struct nvkm_mmu *, bool managed, u64 addr, u64 size,
managed           384 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c 	bool managed;
managed           387 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c 		managed = args->v0.managed != 0;
managed           399 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c 		ret = mmu->func->vmm.ctor(mmu, managed, addr, size, argv, argc,
managed          1035 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	      u32 pd_header, bool managed, u64 addr, u64 size,
managed          1093 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (managed) {
managed          1141 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	      u32 hdr, bool managed, u64 addr, u64 size,
managed          1147 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	return nvkm_vmm_ctor(func, mmu, hdr, managed, addr, size, key, name, *pvmm);
managed           163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 		  u32 pd_header, bool managed, u64 addr, u64 size,
managed           167 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 		  u32 pd_header, bool managed, u64 addr, u64 size,
managed           403 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	       struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           408 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	case 16: return nv04_vmm_new_(func_16, mmu, 0, managed, addr, size,
managed           410 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	case 17: return nv04_vmm_new_(func_17, mmu, 0, managed, addr, size,
managed           419 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c gf100_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           423 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	return gf100_vmm_new_(&gf100_vmm_16, &gf100_vmm_17, mmu, managed, addr,
managed            98 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c gk104_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           102 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c 	return gf100_vmm_new_(&gk104_vmm_16, &gk104_vmm_17, mmu, managed, addr,
managed            67 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c gk20a_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            71 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c 	return gf100_vmm_new_(&gk20a_vmm_16, &gk20a_vmm_17, mmu, managed, addr,
managed           144 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	       struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           168 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	return nvkm_vmm_new_(func, mmu, 0, managed, addr, size, key, name, pvmm);
managed           172 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c gm200_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           176 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	return gm200_vmm_new_(&gm200_vmm_16, &gm200_vmm_17, mmu, managed, addr,
managed           181 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c gm200_vmm_new_fixed(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           185 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	return gf100_vmm_new_(&gm200_vmm_16, &gm200_vmm_17, mmu, managed, addr,
managed            57 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c gm20b_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            61 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c 	return gm200_vmm_new_(&gm20b_vmm_16, &gm20b_vmm_17, mmu, managed, addr,
managed            66 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c gm20b_vmm_new_fixed(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c 	return gf100_vmm_new_(&gm20b_vmm_16, &gm20b_vmm_17, mmu, managed, addr,
managed           509 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	       struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           528 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	ret = nvkm_vmm_new_(func, mmu, 0, managed, addr, size, key, name, pvmm);
managed           537 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c gp100_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           541 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	return gp100_vmm_new_(&gp100_vmm, mmu, managed, addr, size,
managed            45 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c gp10b_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            49 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c 	return gp100_vmm_new_(&gp10b_vmm, mmu, managed, addr, size,
managed            83 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c gv100_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            87 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c 	return gp100_vmm_new_(&gv100_vmm, mmu, managed, addr, size,
managed            39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c mcp77_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            43 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c 	return nv04_vmm_new_(&mcp77_vmm, mmu, 0, managed, addr, size,
managed           103 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	      u32 pd_header, bool managed, u64 addr, u64 size,
managed           112 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	ret = nvkm_vmm_new_(func, mmu, pd_header, managed, addr, size,
managed           121 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c nv04_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           129 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	ret = nv04_vmm_new_(&nv04_vmm, mmu, 8, managed, addr, size,
managed           107 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c nv41_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           111 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c 	return nv04_vmm_new_(&nv41_vmm, mmu, 0, managed, addr, size,
managed           208 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c nv44_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           216 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c 	ret = nv04_vmm_new_(&nv44_vmm, mmu, 0, managed, addr, size,
managed           379 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c nv50_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed           383 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	return nv04_vmm_new_(&nv50_vmm, mmu, 0, managed, addr, size,
managed            72 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c tu102_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
managed            76 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c 	return gp100_vmm_new_(&tu102_vmm, mmu, managed, addr, size,
managed           257 drivers/net/phy/phylink.c 	const char *managed;
managed           264 drivers/net/phy/phylink.c 	if (fwnode_property_read_string(fwnode, "managed", &managed) == 0 &&
managed           265 drivers/net/phy/phylink.c 	    strcmp(managed, "in-band-status") == 0) {
managed           431 drivers/of/of_mdio.c 	const char *managed;
managed           440 drivers/of/of_mdio.c 	err = of_property_read_string(np, "managed", &managed);
managed           441 drivers/of/of_mdio.c 	if (err == 0 && strcmp(managed, "auto") != 0)
managed           458 drivers/of/of_mdio.c 	const char *managed;
managed           460 drivers/of/of_mdio.c 	if (of_property_read_string(np, "managed", &managed) == 0 &&
managed           461 drivers/of/of_mdio.c 	    strcmp(managed, "in-band-status") == 0) {
managed          1216 fs/namei.c     	unsigned managed;
managed          1223 fs/namei.c     	while (managed = READ_ONCE(path->dentry->d_flags),
managed          1224 fs/namei.c     	       managed &= DCACHE_MANAGED_DENTRY,
managed          1225 fs/namei.c     	       unlikely(managed != 0)) {
managed          1228 fs/namei.c     		if (managed & DCACHE_MANAGE_TRANSIT) {
managed          1237 fs/namei.c     		if (managed & DCACHE_MOUNTED) {
managed          1256 fs/namei.c     		if (managed & DCACHE_NEED_AUTOMOUNT) {
managed          1405 fs/namei.c     	unsigned managed;
managed          1408 fs/namei.c     	while (managed = READ_ONCE(path->dentry->d_flags),
managed          1409 fs/namei.c     	       unlikely(managed & DCACHE_MANAGED_DENTRY)) {
managed          1420 fs/namei.c     		if (managed & DCACHE_MANAGE_TRANSIT) {
managed          1429 fs/namei.c     		if (managed & DCACHE_MOUNTED) {
managed          1194 include/linux/irq.h 		     unsigned int bit, bool managed);
managed            78 include/trace/events/irq_matrix.h 		__field(	unsigned int,	managed			)
managed            91 include/trace/events/irq_matrix.h 		__entry->managed		= cmap->managed;
managed           101 include/trace/events/irq_matrix.h 		  __entry->managed, __entry->online_maps,
managed            49 include/uapi/linux/icmpv6.h 					managed:1;
managed            52 include/uapi/linux/icmpv6.h 			__u8		managed:1,
managed            76 include/uapi/linux/icmpv6.h #define icmp6_addrconf_managed	icmp6_dataun.u_nd_ra.managed
managed           269 include/uapi/sound/compress_params.h 	__u32 managed;
managed            16 kernel/irq/matrix.c 	unsigned int		managed;
managed            86 kernel/irq/matrix.c 		cm->available -= cm->managed + m->systembits_inalloc;
managed           111 kernel/irq/matrix.c 				      unsigned int num, bool managed)
managed           121 kernel/irq/matrix.c 	if (managed)
managed           221 kernel/irq/matrix.c 		cm->managed++;
managed           259 kernel/irq/matrix.c 		if (WARN_ON_ONCE(!cm->managed))
managed           271 kernel/irq/matrix.c 		cm->managed--;
managed           412 kernel/irq/matrix.c 		     unsigned int bit, bool managed)
managed           421 kernel/irq/matrix.c 	if(managed)
managed           427 kernel/irq/matrix.c 	if (!managed) {
managed           498 kernel/irq/matrix.c 			   cpu, cm->available, cm->managed,