eh_entries         87 fs/ext4/ext4_extents.h 	__le16	eh_entries;	/* number of valid entries */
eh_entries        167 fs/ext4/ext4_extents.h 	(le16_to_cpu((__path__)->p_hdr->eh_entries) \
eh_entries        170 fs/ext4/ext4_extents.h 	(EXT_FIRST_EXTENT((__hdr__)) + le16_to_cpu((__hdr__)->eh_entries) - 1)
eh_entries        172 fs/ext4/ext4_extents.h 	(EXT_FIRST_INDEX((__hdr__)) + le16_to_cpu((__hdr__)->eh_entries) - 1)
eh_entries        396 fs/ext4/extents.c 	if (eh->eh_entries == 0)
eh_entries        399 fs/ext4/extents.c 	entries = le16_to_cpu(eh->eh_entries);
eh_entries        461 fs/ext4/extents.c 	if (unlikely(le16_to_cpu(eh->eh_entries) > le16_to_cpu(eh->eh_max))) {
eh_entries        488 fs/ext4/extents.c 			 le16_to_cpu(eh->eh_entries), le16_to_cpu(eh->eh_max),
eh_entries        508 fs/ext4/extents.c 	for (i = le16_to_cpu(eh->eh_entries); i > 0; i--, ex++) {
eh_entries        674 fs/ext4/extents.c 	for (i = 0; i < le16_to_cpu(eh->eh_entries); i++, ex++) {
eh_entries        772 fs/ext4/extents.c 		for (k = 0; k < le16_to_cpu(eh->eh_entries); k++, ix++) {
eh_entries        806 fs/ext4/extents.c 	if (eh->eh_entries == 0) {
eh_entries        843 fs/ext4/extents.c 		for (k = 0; k < le16_to_cpu(eh->eh_entries); k++, ex++) {
eh_entries        862 fs/ext4/extents.c 	eh->eh_entries = 0;
eh_entries        912 fs/ext4/extents.c 			  ppos, le16_to_cpu(eh->eh_entries), le16_to_cpu(eh->eh_max));
eh_entries        977 fs/ext4/extents.c 	if (unlikely(le16_to_cpu(curp->p_hdr->eh_entries)
eh_entries        981 fs/ext4/extents.c 				 le16_to_cpu(curp->p_hdr->eh_entries),
eh_entries       1012 fs/ext4/extents.c 	le16_add_cpu(&curp->p_hdr->eh_entries, 1);
eh_entries       1117 fs/ext4/extents.c 	neh->eh_entries = 0;
eh_entries       1123 fs/ext4/extents.c 	if (unlikely(path[depth].p_hdr->eh_entries !=
eh_entries       1126 fs/ext4/extents.c 				 path[depth].p_hdr->eh_entries,
eh_entries       1138 fs/ext4/extents.c 		le16_add_cpu(&neh->eh_entries, m);
eh_entries       1143 fs/ext4/extents.c 		sizeof(struct ext4_extent) * le16_to_cpu(neh->eh_entries);
eh_entries       1160 fs/ext4/extents.c 		le16_add_cpu(&path[depth].p_hdr->eh_entries, -m);
eh_entries       1194 fs/ext4/extents.c 		neh->eh_entries = cpu_to_le16(1);
eh_entries       1222 fs/ext4/extents.c 			le16_add_cpu(&neh->eh_entries, m);
eh_entries       1226 fs/ext4/extents.c 		   (sizeof(struct ext4_extent) * le16_to_cpu(neh->eh_entries));
eh_entries       1244 fs/ext4/extents.c 			le16_add_cpu(&path[i].p_hdr->eh_entries, -m);
eh_entries       1345 fs/ext4/extents.c 	neh->eh_entries = cpu_to_le16(1);
eh_entries       1354 fs/ext4/extents.c 		  le16_to_cpu(neh->eh_entries), le16_to_cpu(neh->eh_max),
eh_entries       1426 fs/ext4/extents.c 		if (path[depth].p_hdr->eh_entries == path[depth].p_hdr->eh_max) {
eh_entries       1818 fs/ext4/extents.c 		le16_add_cpu(&eh->eh_entries, -1);
eh_entries       1820 fs/ext4/extents.c 		WARN_ON(eh->eh_entries == 0);
eh_entries       1821 fs/ext4/extents.c 		if (!eh->eh_entries)
eh_entries       1841 fs/ext4/extents.c 	    (le16_to_cpu(path[0].p_hdr->eh_entries) != 1) ||
eh_entries       1842 fs/ext4/extents.c 	    (le16_to_cpu(path[1].p_hdr->eh_entries) > max_root))
eh_entries       1857 fs/ext4/extents.c 	s = le16_to_cpu(path[1].p_hdr->eh_entries) *
eh_entries       2060 fs/ext4/extents.c 	if (le16_to_cpu(eh->eh_entries) < le16_to_cpu(eh->eh_max))
eh_entries       2076 fs/ext4/extents.c 		if (le16_to_cpu(eh->eh_entries) < le16_to_cpu(eh->eh_max)) {
eh_entries       2078 fs/ext4/extents.c 				  le16_to_cpu(eh->eh_entries));
eh_entries       2083 fs/ext4/extents.c 			  le16_to_cpu(eh->eh_entries), le16_to_cpu(eh->eh_max));
eh_entries       2151 fs/ext4/extents.c 	le16_add_cpu(&eh->eh_entries, 1);
eh_entries       2451 fs/ext4/extents.c 	if (unlikely(path->p_hdr->eh_entries == 0)) {
eh_entries       2465 fs/ext4/extents.c 	le16_add_cpu(&path->p_hdr->eh_entries, -1);
eh_entries       2505 fs/ext4/extents.c 		if (le16_to_cpu(path[depth].p_hdr->eh_entries)
eh_entries       2871 fs/ext4/extents.c 			le16_add_cpu(&eh->eh_entries, -1);
eh_entries       2885 fs/ext4/extents.c 	if (correct_index && eh->eh_entries)
eh_entries       2913 fs/ext4/extents.c 	if (err == 0 && eh->eh_entries == 0 && path[depth].p_bh != NULL)
eh_entries       2936 fs/ext4/extents.c 	if (le16_to_cpu(path->p_hdr->eh_entries) == path->p_block)
eh_entries       3061 fs/ext4/extents.c 				le16_to_cpu(path[k].p_hdr->eh_entries)+1;
eh_entries       3101 fs/ext4/extents.c 			path[i].p_block = le16_to_cpu(path[i].p_hdr->eh_entries)+1;
eh_entries       3104 fs/ext4/extents.c 				  le16_to_cpu(path[i].p_hdr->eh_entries));
eh_entries       3138 fs/ext4/extents.c 			path[i].p_block = le16_to_cpu(path[i].p_hdr->eh_entries);
eh_entries       3142 fs/ext4/extents.c 			if (path[i].p_hdr->eh_entries == 0 && i > 0) {
eh_entries       3157 fs/ext4/extents.c 					 path->p_hdr->eh_entries);
eh_entries       3177 fs/ext4/extents.c 	if (path->p_hdr->eh_entries == 0) {
eh_entries       3927 fs/ext4/extents.c 	if (unlikely(!eh->eh_entries))
eh_entries        392 fs/ext4/migrate.c 		for (i = 0; i < le16_to_cpu(eh->eh_entries); i++, ix++) {
eh_entries        420 fs/ext4/migrate.c 	for (i = 0; i < le16_to_cpu(eh->eh_entries); i++, ix++) {
eh_entries        656 fs/ext4/migrate.c 	    eh->eh_depth != 0 || le16_to_cpu(eh->eh_entries) > 1) {
eh_entries        660 fs/ext4/migrate.c 	if (eh->eh_entries == 0)
eh_entries       2204 include/trace/events/ext4.h 		 int depth, struct partial_cluster *pc, __le16 eh_entries),
eh_entries       2206 include/trace/events/ext4.h 	TP_ARGS(inode, start, end, depth, pc, eh_entries),
eh_entries       2217 include/trace/events/ext4.h 		__field(	unsigned short,	eh_entries	)
eh_entries       2229 include/trace/events/ext4.h 		__entry->eh_entries	= le16_to_cpu(eh_entries);
eh_entries       2243 include/trace/events/ext4.h 		  (unsigned short) __entry->eh_entries)