kmem_cache_alloc   27 arch/arm64/mm/pgd.c 		return kmem_cache_alloc(pgd_cache, gfp);
kmem_cache_alloc   10 arch/powerpc/include/asm/book3s/32/pgalloc.h 	return kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE),
kmem_cache_alloc   55 arch/powerpc/include/asm/book3s/64/pgalloc.h 	pgd = kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE),
kmem_cache_alloc   97 arch/powerpc/include/asm/book3s/64/pgalloc.h 	pud = kmem_cache_alloc(PGT_CACHE(PUD_CACHE_INDEX),
kmem_cache_alloc   22 arch/powerpc/include/asm/nohash/64/pgalloc.h 	return kmem_cache_alloc(PGT_CACHE(PUD_INDEX_SIZE),
kmem_cache_alloc   50 arch/powerpc/include/asm/nohash/64/pgalloc.h 	return kmem_cache_alloc(PGT_CACHE(PMD_CACHE_INDEX),
kmem_cache_alloc   22 arch/powerpc/include/asm/nohash/pgalloc.h 	return kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE),
kmem_cache_alloc  356 arch/powerpc/kvm/book3s_64_mmu_radix.c 	return kmem_cache_alloc(kvm_pte_cache, GFP_KERNEL);
kmem_cache_alloc  366 arch/powerpc/kvm/book3s_64_mmu_radix.c 	return kmem_cache_alloc(kvm_pmd_cache, GFP_KERNEL);
kmem_cache_alloc   73 arch/powerpc/mm/hugetlbpage.c 		new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL));
kmem_cache_alloc  696 arch/powerpc/perf/hv-24x7.c 		kmem_cache_alloc(hv_page_cache, GFP_KERNEL);
kmem_cache_alloc  933 arch/powerpc/perf/hv-24x7.c 	void *page = kmem_cache_alloc(hv_page_cache, GFP_USER);
kmem_cache_alloc 1013 arch/powerpc/perf/hv-24x7.c 	void *page = kmem_cache_alloc(hv_page_cache, GFP_USER);	\
kmem_cache_alloc   51 arch/powerpc/platforms/cell/spufs/inode.c 	ei = kmem_cache_alloc(spufs_inode_cache, GFP_KERNEL);
kmem_cache_alloc   91 arch/powerpc/platforms/pseries/lpar.c 		dtl = kmem_cache_alloc(dtl_cache, GFP_KERNEL);
kmem_cache_alloc  644 arch/powerpc/sysdev/xive/native.c 		p = kmem_cache_alloc(xive_provision_cache, GFP_KERNEL);
kmem_cache_alloc   89 arch/s390/kernel/nmi.c 	origin = (unsigned long) kmem_cache_alloc(mcesa_cache, GFP_KERNEL);
kmem_cache_alloc  109 arch/s390/kernel/nmi.c 	origin = (unsigned long) kmem_cache_alloc(mcesa_cache, GFP_KERNEL);
kmem_cache_alloc  340 arch/s390/mm/pgalloc.c 	table = kmem_cache_alloc(base_pgt_cache, GFP_KERNEL);
kmem_cache_alloc   32 arch/s390/pci/pci_dma.c 	table = kmem_cache_alloc(dma_region_table_cache, GFP_ATOMIC);
kmem_cache_alloc   50 arch/s390/pci/pci_dma.c 	table = kmem_cache_alloc(dma_page_table_cache, GFP_ATOMIC);
kmem_cache_alloc   23 arch/sh/kernel/cpu/fpu.c 		tsk->thread.xstate = kmem_cache_alloc(task_xstate_cachep,
kmem_cache_alloc  161 arch/sh/kernel/cpu/sh4/sq.c 	map = kmem_cache_alloc(sq_cache, GFP_KERNEL);
kmem_cache_alloc   32 arch/sh/kernel/process.c 		dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep,
kmem_cache_alloc   35 arch/sh/mm/pgtable.c 	return kmem_cache_alloc(pgd_cachep, PGALLOC_GFP);
kmem_cache_alloc   51 arch/sh/mm/pgtable.c 	return kmem_cache_alloc(pmd_cachep, PGALLOC_GFP);
kmem_cache_alloc   28 arch/sparc/include/asm/pgalloc_64.h 	return kmem_cache_alloc(pgtable_cache, GFP_KERNEL);
kmem_cache_alloc   45 arch/sparc/include/asm/pgalloc_64.h 	return kmem_cache_alloc(pgtable_cache, GFP_KERNEL);
kmem_cache_alloc   55 arch/sparc/include/asm/pgalloc_64.h 	return kmem_cache_alloc(pgtable_cache, GFP_KERNEL);
kmem_cache_alloc  393 arch/x86/mm/pgtable.c 	return kmem_cache_alloc(pgd_cache, GFP_PGTABLE_USER);
kmem_cache_alloc  222 block/bio.c    		bvl = kmem_cache_alloc(bvs->slab, __gfp_mask);
kmem_cache_alloc  168 drivers/block/rsxx/cregs.c 	cmd = kmem_cache_alloc(creg_cmd_pool, GFP_KERNEL);
kmem_cache_alloc  161 drivers/block/rsxx/dev.c 	bio_meta = kmem_cache_alloc(bio_meta_pool, GFP_KERNEL);
kmem_cache_alloc  608 drivers/block/rsxx/dma.c 	dma = kmem_cache_alloc(rsxx_dma_pool, GFP_KERNEL);
kmem_cache_alloc  642 drivers/block/rsxx/dma.c 	dma = kmem_cache_alloc(rsxx_dma_pool, GFP_KERNEL);
kmem_cache_alloc 2609 drivers/block/skd_main.c 	buf = kmem_cache_alloc(s, gfp);
kmem_cache_alloc 1050 drivers/crypto/axis/artpec6_crypto.c 	common->dma = kmem_cache_alloc(ac->dma_cache, flags);
kmem_cache_alloc  463 drivers/crypto/caam/qi.c 	return kmem_cache_alloc(qi_cache, flags);
kmem_cache_alloc  325 drivers/crypto/ccp/ccp-dmaengine.c 	cmd = kmem_cache_alloc(chan->ccp->dma_cmd_cache, GFP_NOWAIT);
kmem_cache_alloc  434 drivers/dax/super.c 	dax_dev = kmem_cache_alloc(dax_cache, GFP_KERNEL);
kmem_cache_alloc   87 drivers/dma-buf/st-dma-fence.c 	f = kmem_cache_alloc(slab_fences, GFP_KERNEL);
kmem_cache_alloc   85 drivers/dma/ioat/prep.c 	sed = kmem_cache_alloc(ioat_sed_cache, flags);
kmem_cache_alloc 1258 drivers/firewire/net.c 	ptask = kmem_cache_alloc(fwnet_packet_task_cache, GFP_ATOMIC);
kmem_cache_alloc  145 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	fence = kmem_cache_alloc(amdgpu_fence_slab, GFP_KERNEL);
kmem_cache_alloc  171 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	e = kmem_cache_alloc(amdgpu_sync_slab, GFP_KERNEL);
kmem_cache_alloc 1306 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	runqueue_node = kmem_cache_alloc(g2d->runqueue_slab, GFP_KERNEL);
kmem_cache_alloc   89 drivers/gpu/drm/i915/gem/i915_gem_context.c 	return kmem_cache_alloc(global.slab_luts, GFP_KERNEL);
kmem_cache_alloc  200 drivers/gpu/drm/i915/i915_active.c 	prealloc = kmem_cache_alloc(global.slab_cache, GFP_KERNEL);
kmem_cache_alloc  607 drivers/gpu/drm/i915/i915_active.c 			node = kmem_cache_alloc(global.slab_cache, GFP_KERNEL);
kmem_cache_alloc  350 drivers/gpu/drm/i915/i915_request.c 	cb = kmem_cache_alloc(global.slab_execute_cbs, gfp);
kmem_cache_alloc  618 drivers/gpu/drm/i915/i915_request.c 	rq = kmem_cache_alloc(global.slab_requests,
kmem_cache_alloc  631 drivers/gpu/drm/i915/i915_request.c 	return kmem_cache_alloc(global.slab_requests, gfp);
kmem_cache_alloc  676 drivers/gpu/drm/i915/i915_request.c 	rq = kmem_cache_alloc(global.slab_requests,
kmem_cache_alloc  109 drivers/gpu/drm/i915/i915_scheduler.c 		p = kmem_cache_alloc(global.slab_priorities, GFP_ATOMIC);
kmem_cache_alloc  401 drivers/gpu/drm/i915/i915_scheduler.c 	return kmem_cache_alloc(global.slab_dependencies, GFP_KERNEL);
kmem_cache_alloc  820 drivers/infiniband/hw/hfi1/user_sdma.c 		tx = kmem_cache_alloc(pq->txreq_cache, GFP_KERNEL);
kmem_cache_alloc  103 drivers/infiniband/hw/hfi1/verbs_txreq.c 		tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
kmem_cache_alloc   83 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
kmem_cache_alloc  182 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx = kmem_cache_alloc(dd->vnic.txreq_cache, GFP_ATOMIC);
kmem_cache_alloc  959 drivers/infiniband/hw/qib/qib_user_sdma.c 			pkt = kmem_cache_alloc(pq->pkt_slab, GFP_KERNEL);
kmem_cache_alloc  665 drivers/infiniband/ulp/srpt/ib_srpt.c 	ioctx->buf = kmem_cache_alloc(buf_cache, GFP_KERNEL);
kmem_cache_alloc 3718 drivers/iommu/amd_iommu.c 	table->table = kmem_cache_alloc(amd_iommu_irq_cache, GFP_KERNEL);
kmem_cache_alloc  532 drivers/iommu/intel-iommu.c 	return kmem_cache_alloc(iommu_domain_cache, GFP_ATOMIC);
kmem_cache_alloc  542 drivers/iommu/intel-iommu.c 	return kmem_cache_alloc(iommu_devinfo_cache, GFP_ATOMIC);
kmem_cache_alloc  380 drivers/md/dm-bufio.c 		return kmem_cache_alloc(c->slab_cache, gfp_mask);
kmem_cache_alloc  444 drivers/md/dm-bufio.c 	struct dm_buffer *b = kmem_cache_alloc(c->slab_buffer, gfp_mask);
kmem_cache_alloc  174 drivers/md/dm-cache-background-tracker.c 	return kmem_cache_alloc(b->work_cache, GFP_NOWAIT);
kmem_cache_alloc  724 drivers/md/dm-snap.c 	e = kmem_cache_alloc(exception_cache, gfp);
kmem_cache_alloc  726 drivers/md/dm-snap.c 		e = kmem_cache_alloc(exception_cache, GFP_ATOMIC);
kmem_cache_alloc  210 drivers/md/raid5-ppl.c 	io = kmem_cache_alloc(kc, gfp_mask);
kmem_cache_alloc  158 drivers/misc/lkdtm/heap.c 	val = kmem_cache_alloc(double_free_cache, GFP_KERNEL);
kmem_cache_alloc  175 drivers/misc/lkdtm/heap.c 	val = kmem_cache_alloc(a_cache, GFP_KERNEL);
kmem_cache_alloc  204 drivers/misc/lkdtm/usercopy.c 	buf = kmem_cache_alloc(whitelist_cache, GFP_KERNEL);
kmem_cache_alloc 1579 drivers/misc/mic/scif/scif_dma.c 		temp = kmem_cache_alloc(unaligned_cache, GFP_KERNEL);
kmem_cache_alloc 1535 drivers/mtd/nand/raw/nandsim.c 		mypage->byte = kmem_cache_alloc(ns->nand_pages_slab, GFP_NOFS);
kmem_cache_alloc 1053 drivers/mtd/ubi/fastmap.c 		e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
kmem_cache_alloc 1698 drivers/mtd/ubi/wl.c 	e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
kmem_cache_alloc 1775 drivers/mtd/ubi/wl.c 		e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
kmem_cache_alloc 1797 drivers/mtd/ubi/wl.c 			e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
kmem_cache_alloc 3944 drivers/s390/block/dasd_eckd.c 			char *copy = kmem_cache_alloc(dasd_page_cache,
kmem_cache_alloc  514 drivers/s390/block/dasd_fba.c 			char *copy = kmem_cache_alloc(dasd_page_cache,
kmem_cache_alloc 3700 drivers/s390/net/qeth_core_main.c 	*hdr = kmem_cache_alloc(qeth_core_header_cache, GFP_ATOMIC);
kmem_cache_alloc  566 drivers/s390/net/qeth_l2_main.c 		hdr = kmem_cache_alloc(qeth_core_header_cache, GFP_ATOMIC);
kmem_cache_alloc  697 drivers/s390/scsi/zfcp_fsf.c 		qtcb = kmem_cache_alloc(zfcp_fsf_qtcb_cache, GFP_ATOMIC);
kmem_cache_alloc  233 drivers/scsi/aic94xx/aic94xx_hwi.h 	return kmem_cache_alloc(asd_dma_token_cache, flags);
kmem_cache_alloc 4620 drivers/scsi/pmcraid.c 			kmem_cache_alloc(pinstance->cmd_cachep, GFP_KERNEL);
kmem_cache_alloc  328 drivers/soc/fsl/qbman/qman_test_stash.c 		handler = kmem_cache_alloc(hp_handler_slab, GFP_KERNEL);
kmem_cache_alloc 3395 drivers/staging/exfat/exfat_super.c 	ei = kmem_cache_alloc(exfat_inode_cachep, GFP_NOFS);
kmem_cache_alloc  285 drivers/staging/fieldbus/anybuss/host.c 	t = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  235 drivers/staging/fwserial/fwserial.c 	txn = kmem_cache_alloc(fwtty_txn_cache, GFP_ATOMIC);
kmem_cache_alloc  727 drivers/staging/fwserial/fwserial.c 		txn = kmem_cache_alloc(fwtty_txn_cache, GFP_ATOMIC);
kmem_cache_alloc 2452 drivers/usb/dwc2/hcd.c 		qh->dw_align_buf = kmem_cache_alloc(hsotg->unaligned_cache,
kmem_cache_alloc  205 drivers/usb/mon/mon_text.c 	    (ep = kmem_cache_alloc(rp->e_slab, GFP_ATOMIC)) == NULL) {
kmem_cache_alloc  277 drivers/usb/mon/mon_text.c 	    (ep = kmem_cache_alloc(rp->e_slab, GFP_ATOMIC)) == NULL) {
kmem_cache_alloc  226 fs/9p/vfs_inode.c 	v9inode = (struct v9fs_inode *)kmem_cache_alloc(v9fs_inode_cache,
kmem_cache_alloc  274 fs/adfs/super.c 	ei = kmem_cache_alloc(adfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  103 fs/affs/super.c 	i = kmem_cache_alloc(affs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  670 fs/afs/super.c 	vnode = kmem_cache_alloc(afs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 1031 fs/aio.c       	req = kmem_cache_alloc(kiocb_cachep, GFP_KERNEL);
kmem_cache_alloc  279 fs/befs/linuxvfs.c 	bi = kmem_cache_alloc(befs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  243 fs/bfs/inode.c 	bi = kmem_cache_alloc(bfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  773 fs/block_dev.c 	struct bdev_inode *ei = kmem_cache_alloc(bdev_cachep, GFP_KERNEL);
kmem_cache_alloc  345 fs/btrfs/backref.c 	ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask);
kmem_cache_alloc  683 fs/btrfs/backref.c 			new_ref = kmem_cache_alloc(btrfs_prelim_ref_cache,
kmem_cache_alloc  935 fs/btrfs/delayed-ref.c 	ref = kmem_cache_alloc(btrfs_delayed_tree_ref_cachep, GFP_NOFS);
kmem_cache_alloc  939 fs/btrfs/delayed-ref.c 	head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS);
kmem_cache_alloc 1030 fs/btrfs/delayed-ref.c 	ref = kmem_cache_alloc(btrfs_delayed_data_ref_cachep, GFP_NOFS);
kmem_cache_alloc 1046 fs/btrfs/delayed-ref.c 	head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS);
kmem_cache_alloc 1108 fs/btrfs/delayed-ref.c 	head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS);
kmem_cache_alloc  299 fs/btrfs/delayed-ref.h 	return kmem_cache_alloc(btrfs_delayed_extent_op_cachep, GFP_NOFS);
kmem_cache_alloc  742 fs/btrfs/disk-io.c 	end_io_wq = kmem_cache_alloc(btrfs_end_io_wq_cache, GFP_NOFS);
kmem_cache_alloc  301 fs/btrfs/extent_io.c 	state = kmem_cache_alloc(extent_state_cache, mask);
kmem_cache_alloc 9385 fs/btrfs/inode.c 	ei = kmem_cache_alloc(btrfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  102 fs/cachefiles/bind.c 	fsdef = kmem_cache_alloc(cachefiles_object_jar, GFP_KERNEL);
kmem_cache_alloc   43 fs/cachefiles/interface.c 	object = kmem_cache_alloc(cachefiles_object_jar, cachefiles_gfp);
kmem_cache_alloc  229 fs/ceph/caps.c 		cap = kmem_cache_alloc(ceph_cap_cachep, GFP_NOFS);
kmem_cache_alloc  333 fs/ceph/caps.c 		cap = kmem_cache_alloc(ceph_cap_cachep, GFP_NOFS);
kmem_cache_alloc 1666 fs/ceph/caps.c 	return kmem_cache_alloc(ceph_cap_flush_cachep, GFP_KERNEL);
kmem_cache_alloc  426 fs/ceph/inode.c 	ci = kmem_cache_alloc(ceph_inode_cachep, GFP_NOFS);
kmem_cache_alloc  326 fs/cifs/cifsfs.c 	cifs_inode = kmem_cache_alloc(cifs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   46 fs/coda/inode.c 	ei = kmem_cache_alloc(coda_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 1688 fs/dcache.c    	dentry = kmem_cache_alloc(dentry_cache, GFP_KERNEL);
kmem_cache_alloc   96 fs/dcookies.c  	struct dcookie_struct *dcs = kmem_cache_alloc(dcookie_cache,
kmem_cache_alloc 1199 fs/direct-io.c 	dio = kmem_cache_alloc(dio_cache, GFP_KERNEL);
kmem_cache_alloc 1414 fs/ecryptfs/crypto.c 	page_virt = kmem_cache_alloc(ecryptfs_header_cache, GFP_USER);
kmem_cache_alloc 1646 fs/ecryptfs/crypto.c 	tmp_tfm = kmem_cache_alloc(ecryptfs_key_tfm_cache, GFP_KERNEL);
kmem_cache_alloc  327 fs/ecryptfs/inode.c 	dentry_info = kmem_cache_alloc(ecryptfs_dentry_info_cache, GFP_KERNEL);
kmem_cache_alloc 2415 fs/ecryptfs/keystore.c 	key_rec = kmem_cache_alloc(ecryptfs_key_record_cache, GFP_KERNEL);
kmem_cache_alloc 2496 fs/ecryptfs/keystore.c 	new_key_sig = kmem_cache_alloc(ecryptfs_key_sig_cache, GFP_KERNEL);
kmem_cache_alloc  418 fs/ecryptfs/mmap.c 	xattr_virt = kmem_cache_alloc(ecryptfs_xattr_cache, GFP_KERNEL);
kmem_cache_alloc   41 fs/ecryptfs/super.c 	inode_info = kmem_cache_alloc(ecryptfs_inode_info_cache, GFP_KERNEL);
kmem_cache_alloc   71 fs/efs/super.c 	ei = kmem_cache_alloc(efs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   59 fs/erofs/super.c 		kmem_cache_alloc(erofs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  409 fs/erofs/zdata.c 	pcl = kmem_cache_alloc(pcluster_cachep, GFP_NOFS);
kmem_cache_alloc 1322 fs/eventpoll.c 	if (epi->nwait >= 0 && (pwq = kmem_cache_alloc(pwq_cache, GFP_KERNEL))) {
kmem_cache_alloc 1510 fs/eventpoll.c 	if (!(epi = kmem_cache_alloc(epi_cache, GFP_KERNEL)))
kmem_cache_alloc 1200 fs/exec.c      		newsighand = kmem_cache_alloc(sighand_cachep, GFP_KERNEL);
kmem_cache_alloc  183 fs/ext2/super.c 	ei = kmem_cache_alloc(ext2_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   95 fs/ext4/block_validity.c 		new_entry = kmem_cache_alloc(ext4_system_zone_cachep,
kmem_cache_alloc  449 fs/ext4/extents_status.c 	es = kmem_cache_alloc(ext4_es_cachep, GFP_ATOMIC);
kmem_cache_alloc 1876 fs/ext4/extents_status.c 	pr = kmem_cache_alloc(ext4_pending_cachep, GFP_ATOMIC);
kmem_cache_alloc 3659 fs/ext4/mballoc.c 	pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS);
kmem_cache_alloc 3753 fs/ext4/mballoc.c 	pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS);
kmem_cache_alloc 4898 fs/ext4/mballoc.c 		new_entry = kmem_cache_alloc(ext4_free_data_cachep,
kmem_cache_alloc 1084 fs/ext4/super.c 	ei = kmem_cache_alloc(ext4_inode_cachep, GFP_NOFS);
kmem_cache_alloc  209 fs/f2fs/extent_cache.c 	en = kmem_cache_alloc(extent_node_slab, GFP_ATOMIC);
kmem_cache_alloc 2209 fs/f2fs/f2fs.h 	entry = kmem_cache_alloc(cachep, flags);
kmem_cache_alloc 2211 fs/f2fs/f2fs.h 		entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL);
kmem_cache_alloc  151 fs/f2fs/node.c 		new = kmem_cache_alloc(nat_entry_slab, GFP_F2FS_ZERO);
kmem_cache_alloc  887 fs/f2fs/super.c 	fi = kmem_cache_alloc(f2fs_inode_cachep, GFP_F2FS_ZERO);
kmem_cache_alloc   64 fs/fat/cache.c 	return kmem_cache_alloc(fat_cache_cachep, GFP_NOFS);
kmem_cache_alloc  747 fs/fat/inode.c 	ei = kmem_cache_alloc(fat_inode_cachep, GFP_NOFS);
kmem_cache_alloc  895 fs/fcntl.c     	return kmem_cache_alloc(fasync_cache, GFP_KERNEL);
kmem_cache_alloc  280 fs/file.c      	newf = kmem_cache_alloc(files_cachep, GFP_KERNEL);
kmem_cache_alloc  127 fs/freevxfs/vxfs_super.c 	vi = kmem_cache_alloc(vxfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  114 fs/fs_struct.c 	struct fs_struct *fs = kmem_cache_alloc(fs_cachep, GFP_KERNEL);
kmem_cache_alloc   76 fs/fuse/inode.c 	fi = kmem_cache_alloc(fuse_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  816 fs/gfs2/glock.c 	gl = kmem_cache_alloc(cachep, GFP_NOFS);
kmem_cache_alloc 1394 fs/gfs2/super.c 	ip = kmem_cache_alloc(gfs2_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  166 fs/hfs/super.c 	i = kmem_cache_alloc(hfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   85 fs/hfsplus/attributes.c 	return kmem_cache_alloc(hfsplus_attr_tree_cachep, GFP_KERNEL);
kmem_cache_alloc  628 fs/hfsplus/super.c 	i = kmem_cache_alloc(hfsplus_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  236 fs/hpfs/super.c 	ei = kmem_cache_alloc(hpfs_inode_cachep, GFP_NOFS);
kmem_cache_alloc 1038 fs/hugetlbfs/inode.c 	p = kmem_cache_alloc(hugetlbfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  233 fs/inode.c     		inode = kmem_cache_alloc(inode_cachep, GFP_KERNEL);
kmem_cache_alloc  624 fs/io_uring.c  		req = kmem_cache_alloc(req_cachep, gfp);
kmem_cache_alloc  639 fs/io_uring.c  			state->reqs[0] = kmem_cache_alloc(req_cachep, gfp);
kmem_cache_alloc   73 fs/isofs/inode.c 	ei = kmem_cache_alloc(isofs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 2345 fs/jbd2/journal.c 		ptr = kmem_cache_alloc(get_slab(size), flags);
kmem_cache_alloc  145 fs/jbd2/revoke.c 	record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask);
kmem_cache_alloc  224 fs/jbd2/revoke.c 	table = kmem_cache_alloc(jbd2_revoke_table_cache, GFP_KERNEL);
kmem_cache_alloc  130 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(full_dnode_slab, GFP_KERNEL);
kmem_cache_alloc  144 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(raw_dirent_slab, GFP_KERNEL);
kmem_cache_alloc  158 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(raw_inode_slab, GFP_KERNEL);
kmem_cache_alloc  172 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(tmp_dnode_info_slab, GFP_KERNEL);
kmem_cache_alloc  188 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(raw_node_ref_slab, GFP_KERNEL);
kmem_cache_alloc  251 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(node_frag_slab, GFP_KERNEL);
kmem_cache_alloc  265 fs/jffs2/malloc.c 	ret = kmem_cache_alloc(inode_cache_slab, GFP_KERNEL);
kmem_cache_alloc   42 fs/jffs2/super.c 	f = kmem_cache_alloc(jffs2_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  105 fs/jfs/super.c 	jfs_inode = kmem_cache_alloc(jfs_inode_cachep, GFP_NOFS);
kmem_cache_alloc  258 fs/locks.c     	ctx = kmem_cache_alloc(flctx_cache, GFP_KERNEL);
kmem_cache_alloc   88 fs/mbcache.c   	entry = kmem_cache_alloc(mb_entry_cache, mask);
kmem_cache_alloc   66 fs/minix/inode.c 	ei = kmem_cache_alloc(minix_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 2068 fs/nfs/inode.c 	nfsi = kmem_cache_alloc(nfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   87 fs/nfs/write.c 			p = kmem_cache_alloc(nfs_cdata_cachep, GFP_NOIO |
kmem_cache_alloc  143 fs/nfsd/filecache.c 		new = kmem_cache_alloc(nfsd_file_mark_slab, GFP_KERNEL);
kmem_cache_alloc  176 fs/nfsd/filecache.c 	nf = kmem_cache_alloc(nfsd_file_slab, GFP_KERNEL);
kmem_cache_alloc  421 fs/nfsd/nfs4layouts.c 	new = kmem_cache_alloc(nfs4_layout_cache, GFP_KERNEL);
kmem_cache_alloc 3937 fs/nfsd/nfs4state.c 	return kmem_cache_alloc(file_slab, GFP_KERNEL);
kmem_cache_alloc 4058 fs/nfsd/nfs4state.c 	sop = kmem_cache_alloc(slab, GFP_KERNEL);
kmem_cache_alloc   98 fs/nfsd/nfscache.c 	rp = kmem_cache_alloc(nn->drc_slab, GFP_KERNEL);
kmem_cache_alloc   28 fs/nilfs2/btree.c 	path = kmem_cache_alloc(nilfs_btree_path_cache, GFP_NOFS);
kmem_cache_alloc   38 fs/nilfs2/segbuf.c 	segbuf = kmem_cache_alloc(nilfs_segbuf_cachep, GFP_NOFS);
kmem_cache_alloc  165 fs/nilfs2/segment.c 		ti = kmem_cache_alloc(nilfs_transaction_cachep, GFP_NOFS);
kmem_cache_alloc  148 fs/nilfs2/super.c 	ii = kmem_cache_alloc(nilfs_inode_cachep, GFP_NOFS);
kmem_cache_alloc  295 fs/notify/dnotify/dnotify.c 	dn = kmem_cache_alloc(dnotify_struct_cache, GFP_KERNEL);
kmem_cache_alloc  302 fs/notify/dnotify/dnotify.c 	new_dn_mark = kmem_cache_alloc(dnotify_mark_cache, GFP_KERNEL);
kmem_cache_alloc  311 fs/notify/fanotify/fanotify.c 		pevent = kmem_cache_alloc(fanotify_perm_event_cachep, gfp);
kmem_cache_alloc  319 fs/notify/fanotify/fanotify.c 	event = kmem_cache_alloc(fanotify_event_cachep, gfp);
kmem_cache_alloc  688 fs/notify/fanotify/fanotify_user.c 	mark = kmem_cache_alloc(fanotify_mark_cache, GFP_KERNEL);
kmem_cache_alloc  570 fs/notify/inotify/inotify_user.c 	tmp_i_mark = kmem_cache_alloc(inotify_inode_mark_cachep, GFP_KERNEL);
kmem_cache_alloc  479 fs/notify/mark.c 	conn = kmem_cache_alloc(fsnotify_mark_connector_cachep, GFP_KERNEL);
kmem_cache_alloc 1259 fs/ntfs/attrib.c 	ctx = kmem_cache_alloc(ntfs_attr_ctx_cache, GFP_NOFS);
kmem_cache_alloc   29 fs/ntfs/index.c 	ictx = kmem_cache_alloc(ntfs_index_ctx_cache, GFP_NOFS);
kmem_cache_alloc  312 fs/ntfs/inode.c 	ni = kmem_cache_alloc(ntfs_big_inode_cache, GFP_NOFS);
kmem_cache_alloc  331 fs/ntfs/inode.c 	ni = kmem_cache_alloc(ntfs_inode_cache, GFP_NOFS);
kmem_cache_alloc  255 fs/ntfs/unistr.c 		ucs = kmem_cache_alloc(ntfs_name_cache, GFP_NOFS);
kmem_cache_alloc  788 fs/ocfs2/dlm/dlmmaster.c 		alloc_mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS);
kmem_cache_alloc 1569 fs/ocfs2/dlm/dlmmaster.c 			mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS);
kmem_cache_alloc 2573 fs/ocfs2/dlm/dlmmaster.c 	mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS);
kmem_cache_alloc 3123 fs/ocfs2/dlm/dlmmaster.c 	mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS);
kmem_cache_alloc  328 fs/ocfs2/dlmfs/dlmfs.c 	ip = kmem_cache_alloc(dlmfs_inode_cache, GFP_NOFS);
kmem_cache_alloc  258 fs/ocfs2/quota_local.c 		newchunk = kmem_cache_alloc(ocfs2_qf_chunk_cachep, GFP_NOFS);
kmem_cache_alloc  990 fs/ocfs2/quota_local.c 	chunk = kmem_cache_alloc(ocfs2_qf_chunk_cachep, GFP_NOFS);
kmem_cache_alloc  552 fs/ocfs2/super.c 	oi = kmem_cache_alloc(ocfs2_inode_cachep, GFP_NOFS);
kmem_cache_alloc  399 fs/ocfs2/uptodate.c 	new = kmem_cache_alloc(ocfs2_uptodate_cachep, GFP_NOFS);
kmem_cache_alloc  410 fs/ocfs2/uptodate.c 			tree[i] = kmem_cache_alloc(ocfs2_uptodate_cachep,
kmem_cache_alloc  334 fs/openpromfs/inode.c 	oi = kmem_cache_alloc(op_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  109 fs/orangefs/super.c 	orangefs_inode = kmem_cache_alloc(orangefs_inode_cache, GFP_KERNEL);
kmem_cache_alloc  173 fs/overlayfs/super.c 	struct ovl_inode *oi = kmem_cache_alloc(ovl_inode_cachep, GFP_KERNEL);
kmem_cache_alloc   62 fs/proc/inode.c 	ei = kmem_cache_alloc(proc_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  356 fs/proc/inode.c 		pdeo = kmem_cache_alloc(pde_opener_cache, GFP_KERNEL);
kmem_cache_alloc  342 fs/qnx4/inode.c 	ei = kmem_cache_alloc(qnx4_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  602 fs/qnx6/inode.c 	ei = kmem_cache_alloc(qnx6_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  642 fs/reiserfs/super.c 	ei = kmem_cache_alloc(reiserfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  377 fs/romfs/super.c 	inode = kmem_cache_alloc(romfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  470 fs/squashfs/super.c 		kmem_cache_alloc(squashfs_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  310 fs/sysv/inode.c 	si = kmem_cache_alloc(sysv_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  253 fs/ubifs/super.c 	ui = kmem_cache_alloc(ubifs_inode_slab, GFP_NOFS);
kmem_cache_alloc  144 fs/udf/super.c 	ei = kmem_cache_alloc(udf_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 1448 fs/ufs/super.c 	ei = kmem_cache_alloc(ufs_inode_cachep, GFP_NOFS);
kmem_cache_alloc  694 fs/userfaultfd.c 		ctx = kmem_cache_alloc(userfaultfd_ctx_cachep, GFP_KERNEL);
kmem_cache_alloc 1962 fs/userfaultfd.c 	ctx = kmem_cache_alloc(userfaultfd_ctx_cachep, GFP_KERNEL);
kmem_cache_alloc  128 fs/xfs/kmem.c  		ptr = kmem_cache_alloc(zone, lflags);
kmem_cache_alloc 2566 include/linux/fs.h #define __getname()		kmem_cache_alloc(names_cachep, GFP_KERNEL)
kmem_cache_alloc 1454 include/linux/jbd2.h 	return kmem_cache_alloc(jbd2_inode_cache, gfp_flags);
kmem_cache_alloc  394 include/linux/slab.h void *kmem_cache_alloc(struct kmem_cache *, gfp_t flags) __assume_slab_alignment __malloc;
kmem_cache_alloc  427 include/linux/slab.h 	return kmem_cache_alloc(s, flags);
kmem_cache_alloc  452 include/linux/slab.h 	void *ret = kmem_cache_alloc(s, flags);
kmem_cache_alloc  680 include/linux/slab.h 	return kmem_cache_alloc(k, flags | __GFP_ZERO);
kmem_cache_alloc   84 include/net/request_sock.h 	req = kmem_cache_alloc(ops->slab, GFP_ATOMIC | __GFP_NOWARN);
kmem_cache_alloc   54 include/trace/events/kmem.h DEFINE_EVENT(kmem_alloc, kmem_cache_alloc,
kmem_cache_alloc  427 ipc/mqueue.c   	ei = kmem_cache_alloc(mqueue_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 1681 kernel/audit.c 	ab = kmem_cache_alloc(audit_buffer_cache, gfp_mask);
kmem_cache_alloc  258 kernel/cred.c  	new = kmem_cache_alloc(cred_jar, GFP_KERNEL);
kmem_cache_alloc  686 kernel/cred.c  	new = kmem_cache_alloc(cred_jar, GFP_KERNEL);
kmem_cache_alloc  348 kernel/fork.c  	vma = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL);
kmem_cache_alloc  356 kernel/fork.c  	struct vm_area_struct *new = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL);
kmem_cache_alloc  677 kernel/fork.c  #define allocate_mm()	(kmem_cache_alloc(mm_cachep, GFP_KERNEL))
kmem_cache_alloc 1507 kernel/fork.c  	sig = kmem_cache_alloc(sighand_cachep, GFP_KERNEL);
kmem_cache_alloc   49 kernel/nsproxy.c 	nsproxy = kmem_cache_alloc(nsproxy_cachep, GFP_KERNEL);
kmem_cache_alloc  169 kernel/pid.c   	pid = kmem_cache_alloc(ns->pid_cachep, GFP_KERNEL);
kmem_cache_alloc 6941 kernel/sched/core.c 	tg = kmem_cache_alloc(task_group_cache, GFP_KERNEL | __GFP_ZERO);
kmem_cache_alloc  434 kernel/signal.c 		q = kmem_cache_alloc(sigqueue_cachep, flags);
kmem_cache_alloc  111 kernel/trace/trace_events.c 	field = kmem_cache_alloc(field_cachep, GFP_TRACE);
kmem_cache_alloc 2249 kernel/trace/trace_events.c 	file = kmem_cache_alloc(file_cachep, GFP_TRACE);
kmem_cache_alloc   34 kernel/utsname.c 	uts_ns = kmem_cache_alloc(uts_ns_cache, GFP_KERNEL);
kmem_cache_alloc   83 lib/btree.c    	return kmem_cache_alloc(btree_cachep, gfp_mask);
kmem_cache_alloc  134 lib/lru_cache.c 		void *p = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  262 lib/radix-tree.c 		ret = kmem_cache_alloc(radix_tree_node_cachep,
kmem_cache_alloc  285 lib/radix-tree.c 	ret = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
kmem_cache_alloc  347 lib/radix-tree.c 		node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
kmem_cache_alloc  392 lib/test_kasan.c 	p = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  423 lib/test_kasan.c 		p = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  552 lib/test_kasan.c 	p = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  577 lib/test_kasan.c 	p = kmem_cache_alloc(cache, GFP_KERNEL);
kmem_cache_alloc  206 lib/test_meminit.c 		buf = kmem_cache_alloc(c, alloc_mask);
kmem_cache_alloc  263 lib/test_meminit.c 	buf = kmem_cache_alloc(c, GFP_KERNEL);
kmem_cache_alloc  281 lib/test_meminit.c 		buf = kmem_cache_alloc(c, GFP_KERNEL);
kmem_cache_alloc  304 lib/xarray.c   	xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp);
kmem_cache_alloc  335 lib/xarray.c   		xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp);
kmem_cache_alloc  338 lib/xarray.c   		xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp);
kmem_cache_alloc  371 lib/xarray.c   		node = kmem_cache_alloc(radix_tree_node_cachep, gfp);
kmem_cache_alloc   53 mm/kmemleak-test.c 		kmem_cache_alloc(files_cachep, GFP_KERNEL));
kmem_cache_alloc   55 mm/kmemleak-test.c 		kmem_cache_alloc(files_cachep, GFP_KERNEL));
kmem_cache_alloc  423 mm/kmemleak.c  		object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp));
kmem_cache_alloc  799 mm/kmemleak.c  		area = kmem_cache_alloc(scan_area_cache, gfp_kmemleak_mask(gfp));
kmem_cache_alloc  402 mm/ksm.c       	return kmem_cache_alloc(stable_node_cache, GFP_KERNEL | __GFP_HIGH);
kmem_cache_alloc 4661 mm/memory.c    	ptl = kmem_cache_alloc(page_ptl_cachep, GFP_KERNEL);
kmem_cache_alloc  276 mm/mempolicy.c 	policy = kmem_cache_alloc(policy_cache, GFP_KERNEL);
kmem_cache_alloc 2224 mm/mempolicy.c 	struct mempolicy *new = kmem_cache_alloc(policy_cache, GFP_KERNEL);
kmem_cache_alloc 2489 mm/mempolicy.c 	n = kmem_cache_alloc(sn_cache, GFP_KERNEL);
kmem_cache_alloc 2561 mm/mempolicy.c 	n_new = kmem_cache_alloc(sn_cache, GFP_KERNEL);
kmem_cache_alloc 2564 mm/mempolicy.c 	mpol_new = kmem_cache_alloc(policy_cache, GFP_KERNEL);
kmem_cache_alloc  513 mm/mempool.c   	return kmem_cache_alloc(mem, gfp_mask);
kmem_cache_alloc 1398 mm/nommu.c     	region = kmem_cache_alloc(vm_region_jar, GFP_KERNEL);
kmem_cache_alloc   83 mm/rmap.c      	anon_vma = kmem_cache_alloc(anon_vma_cachep, GFP_KERNEL);
kmem_cache_alloc  130 mm/rmap.c      	return kmem_cache_alloc(anon_vma_chain_cachep, gfp);
kmem_cache_alloc 3741 mm/shmem.c     	info = kmem_cache_alloc(shmem_inode_cachep, GFP_KERNEL);
kmem_cache_alloc 3490 mm/slab.c      EXPORT_SYMBOL(kmem_cache_alloc);
kmem_cache_alloc  122 mm/slab_common.c 		void *x = p[i] = kmem_cache_alloc(s, flags);
kmem_cache_alloc 1796 mm/slab_common.c EXPORT_TRACEPOINT_SYMBOL(kmem_cache_alloc);
kmem_cache_alloc  626 mm/slob.c      EXPORT_SYMBOL(kmem_cache_alloc);
kmem_cache_alloc 2786 mm/slub.c      EXPORT_SYMBOL(kmem_cache_alloc);
kmem_cache_alloc 4156 mm/slub.c      		n = kmem_cache_alloc(kmem_cache_node, GFP_KERNEL);
kmem_cache_alloc  973 mm/vmalloc.c   			lva = kmem_cache_alloc(vmap_area_cachep, GFP_NOWAIT);
kmem_cache_alloc  201 mm/z3fold.c    	slots = kmem_cache_alloc(pool->c_handle,
kmem_cache_alloc  353 mm/zsmalloc.c  	return (unsigned long)kmem_cache_alloc(pool->handle_cachep,
kmem_cache_alloc  364 mm/zsmalloc.c  	return kmem_cache_alloc(pool->zspage_cachep,
kmem_cache_alloc  251 mm/zswap.c     	entry = kmem_cache_alloc(zswap_entry_cache, gfp);
kmem_cache_alloc  229 net/9p/client.c 		fc->sdata = kmem_cache_alloc(c->fcall_cache, GFP_NOFS);
kmem_cache_alloc  270 net/9p/client.c 	struct p9_req_t *req = kmem_cache_alloc(p9_req_cache, GFP_NOFS);
kmem_cache_alloc  475 net/batman-adv/translation-table.c 	tt_change_node = kmem_cache_alloc(batadv_tt_change_cache, GFP_ATOMIC);
kmem_cache_alloc  711 net/batman-adv/translation-table.c 	tt_local = kmem_cache_alloc(batadv_tl_cache, GFP_ATOMIC);
kmem_cache_alloc 2896 net/batman-adv/translation-table.c 	tt_req_node = kmem_cache_alloc(batadv_tt_req_cache, GFP_ATOMIC);
kmem_cache_alloc 3725 net/batman-adv/translation-table.c 		tt_roam_node = kmem_cache_alloc(batadv_tt_roam_cache,
kmem_cache_alloc  492 net/bridge/br_fdb.c 	fdb = kmem_cache_alloc(br_fdb_cache, GFP_ATOMIC);
kmem_cache_alloc  459 net/can/af_can.c 	rcv = kmem_cache_alloc(rcv_cache, GFP_KERNEL);
kmem_cache_alloc 1041 net/can/gw.c   	gwj = kmem_cache_alloc(cgw_cache, GFP_KERNEL);
kmem_cache_alloc  587 net/ceph/osd_client.c 		req = kmem_cache_alloc(ceph_osd_request_cache, gfp_flags);
kmem_cache_alloc   93 net/core/dst.c 	dst = kmem_cache_alloc(ops->kmem_cachep, GFP_ATOMIC);
kmem_cache_alloc  310 net/core/skbuff.c 	skb = kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC);
kmem_cache_alloc 1448 net/core/skbuff.c 		n = kmem_cache_alloc(skbuff_head_cache, gfp_mask);
kmem_cache_alloc 5981 net/core/skbuff.c 	struct skb_ext *new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC);
kmem_cache_alloc 5999 net/core/skbuff.c 	new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC);
kmem_cache_alloc 1599 net/core/sock.c 		sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
kmem_cache_alloc   55 net/dccp/ackvec.c 	avr = kmem_cache_alloc(dccp_ackvec_record_slab, GFP_ATOMIC);
kmem_cache_alloc  148 net/dccp/ccid.c 	ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab :
kmem_cache_alloc   40 net/dccp/ccids/lib/loss_interval.c 		lh->ring[LIH_INDEX(lh->counter)] = kmem_cache_alloc(tfrc_lh_slab,
kmem_cache_alloc   51 net/dccp/ccids/lib/packet_history.c 	struct tfrc_tx_hist_entry *entry = kmem_cache_alloc(tfrc_tx_hist_slab, gfp_any());
kmem_cache_alloc  339 net/dccp/ccids/lib/packet_history.c 		h->ring[i] = kmem_cache_alloc(tfrc_rx_hist_slab, GFP_ATOMIC);
kmem_cache_alloc  360 net/ipv4/fib_trie.c 	kv = kmem_cache_alloc(trie_leaf_kmem, GFP_KERNEL);
kmem_cache_alloc 1204 net/ipv4/fib_trie.c 			new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL);
kmem_cache_alloc 1258 net/ipv4/fib_trie.c 	new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL);
kmem_cache_alloc 1761 net/ipv4/fib_trie.c 			new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL);
kmem_cache_alloc   67 net/ipv4/inet_hashtables.c 	struct inet_bind_bucket *tb = kmem_cache_alloc(cachep, GFP_ATOMIC);
kmem_cache_alloc  164 net/ipv4/inet_timewait_sock.c 	tw = kmem_cache_alloc(sk->sk_prot_creator->twsk_prot->twsk_slab,
kmem_cache_alloc  216 net/ipv4/inetpeer.c 		p = kmem_cache_alloc(peer_cachep, GFP_ATOMIC);
kmem_cache_alloc  147 net/ipv6/xfrm6_tunnel.c 	x6spi = kmem_cache_alloc(xfrm6_tunnel_spi_kmem, GFP_ATOMIC);
kmem_cache_alloc  237 net/mac80211/mesh.c 	p = kmem_cache_alloc(rm_cache, GFP_ATOMIC);
kmem_cache_alloc  910 net/netfilter/ipvs/ip_vs_conn.c 	cp = kmem_cache_alloc(ip_vs_conn_cachep, GFP_ATOMIC);
kmem_cache_alloc  183 net/netfilter/nf_conncount.c 	conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC);
kmem_cache_alloc  358 net/netfilter/nf_conncount.c 	rbconn = kmem_cache_alloc(conncount_rb_cachep, GFP_ATOMIC);
kmem_cache_alloc  362 net/netfilter/nf_conncount.c 	conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC);
kmem_cache_alloc 1371 net/netfilter/nf_conntrack_core.c 	ct = kmem_cache_alloc(nf_conntrack_cachep, gfp);
kmem_cache_alloc  286 net/netfilter/nf_conntrack_expect.c 	new = kmem_cache_alloc(nf_ct_expect_cachep, GFP_ATOMIC);
kmem_cache_alloc  246 net/netfilter/xt_hashlimit.c 		ent = kmem_cache_alloc(hashlimit_cachep, GFP_ATOMIC);
kmem_cache_alloc  262 net/rds/ib_recv.c 		ibinc = kmem_cache_alloc(rds_ib_incoming_slab, slab_mask);
kmem_cache_alloc  288 net/rds/ib_recv.c 		frag = kmem_cache_alloc(rds_ib_frag_slab, slab_mask);
kmem_cache_alloc  383 net/rds/tcp.c  		tc = kmem_cache_alloc(rds_tcp_conn_slab, gfp);
kmem_cache_alloc  173 net/rds/tcp_recv.c 			tinc = kmem_cache_alloc(rds_tcp_incoming_slab,
kmem_cache_alloc 8538 net/sctp/socket.c 	pp = kmem_cache_alloc(sctp_bucket_cachep, GFP_ATOMIC);
kmem_cache_alloc  239 net/socket.c   	ei = kmem_cache_alloc(sock_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  200 net/sunrpc/rpc_pipe.c 	rpci = kmem_cache_alloc(rpc_inode_cachep, GFP_KERNEL);
kmem_cache_alloc  615 net/xfrm/xfrm_state.c 	x = kmem_cache_alloc(xfrm_state_cache, GFP_ATOMIC | __GFP_ZERO);
kmem_cache_alloc  105 security/integrity/iint.c 	iint = kmem_cache_alloc(iint_cache, GFP_NOFS);
kmem_cache_alloc   20 tools/testing/radix-tree/linux/slab.h void *kmem_cache_alloc(struct kmem_cache *cachep, int flags);