1/*
2 * QLogic qlcnic NIC Driver
3 * Copyright (c) 2009-2013 QLogic Corporation
4 *
5 * See LICENSE.qlcnic for copyright and licensing details.
6 */
7
8#include "qlcnic.h"
9#include "qlcnic_hdr.h"
10#include "qlcnic_83xx_hw.h"
11#include "qlcnic_hw.h"
12
13#include <net/ip.h>
14
15#define QLC_83XX_MINIDUMP_FLASH		0x520000
16#define QLC_83XX_OCM_INDEX			3
17#define QLC_83XX_PCI_INDEX			0
18#define QLC_83XX_DMA_ENGINE_INDEX		8
19
20static const u32 qlcnic_ms_read_data[] = {
21	0x410000A8, 0x410000AC, 0x410000B8, 0x410000BC
22};
23
24#define QLCNIC_DUMP_WCRB	BIT_0
25#define QLCNIC_DUMP_RWCRB	BIT_1
26#define QLCNIC_DUMP_ANDCRB	BIT_2
27#define QLCNIC_DUMP_ORCRB	BIT_3
28#define QLCNIC_DUMP_POLLCRB	BIT_4
29#define QLCNIC_DUMP_RD_SAVE	BIT_5
30#define QLCNIC_DUMP_WRT_SAVED	BIT_6
31#define QLCNIC_DUMP_MOD_SAVE_ST	BIT_7
32#define QLCNIC_DUMP_SKIP	BIT_7
33
34#define QLCNIC_DUMP_MASK_MAX	0xff
35
36struct qlcnic_pex_dma_descriptor {
37	u32	read_data_size;
38	u32	dma_desc_cmd;
39	u32	src_addr_low;
40	u32	src_addr_high;
41	u32	dma_bus_addr_low;
42	u32	dma_bus_addr_high;
43	u32	rsvd[6];
44} __packed;
45
46struct qlcnic_common_entry_hdr {
47	u32     type;
48	u32     offset;
49	u32     cap_size;
50#if defined(__LITTLE_ENDIAN)
51	u8      mask;
52	u8      rsvd[2];
53	u8      flags;
54#else
55	u8      flags;
56	u8      rsvd[2];
57	u8      mask;
58#endif
59} __packed;
60
61struct __crb {
62	u32	addr;
63#if defined(__LITTLE_ENDIAN)
64	u8	stride;
65	u8	rsvd1[3];
66#else
67	u8	rsvd1[3];
68	u8	stride;
69#endif
70	u32	data_size;
71	u32	no_ops;
72	u32	rsvd2[4];
73} __packed;
74
75struct __ctrl {
76	u32	addr;
77#if defined(__LITTLE_ENDIAN)
78	u8	stride;
79	u8	index_a;
80	u16	timeout;
81#else
82	u16	timeout;
83	u8	index_a;
84	u8	stride;
85#endif
86	u32	data_size;
87	u32	no_ops;
88#if defined(__LITTLE_ENDIAN)
89	u8	opcode;
90	u8	index_v;
91	u8	shl_val;
92	u8	shr_val;
93#else
94	u8	shr_val;
95	u8	shl_val;
96	u8	index_v;
97	u8	opcode;
98#endif
99	u32	val1;
100	u32	val2;
101	u32	val3;
102} __packed;
103
104struct __cache {
105	u32	addr;
106#if defined(__LITTLE_ENDIAN)
107	u16	stride;
108	u16	init_tag_val;
109#else
110	u16	init_tag_val;
111	u16	stride;
112#endif
113	u32	size;
114	u32	no_ops;
115	u32	ctrl_addr;
116	u32	ctrl_val;
117	u32	read_addr;
118#if defined(__LITTLE_ENDIAN)
119	u8	read_addr_stride;
120	u8	read_addr_num;
121	u8	rsvd1[2];
122#else
123	u8	rsvd1[2];
124	u8	read_addr_num;
125	u8	read_addr_stride;
126#endif
127} __packed;
128
129struct __ocm {
130	u8	rsvd[8];
131	u32	size;
132	u32	no_ops;
133	u8	rsvd1[8];
134	u32	read_addr;
135	u32	read_addr_stride;
136} __packed;
137
138struct __mem {
139	u32	desc_card_addr;
140	u32	dma_desc_cmd;
141	u32	start_dma_cmd;
142	u32	rsvd[3];
143	u32	addr;
144	u32	size;
145} __packed;
146
147struct __mux {
148	u32	addr;
149	u8	rsvd[4];
150	u32	size;
151	u32	no_ops;
152	u32	val;
153	u32	val_stride;
154	u32	read_addr;
155	u8	rsvd2[4];
156} __packed;
157
158struct __queue {
159	u32	sel_addr;
160#if defined(__LITTLE_ENDIAN)
161	u16	stride;
162	u8	rsvd[2];
163#else
164	u8	rsvd[2];
165	u16	stride;
166#endif
167	u32	size;
168	u32	no_ops;
169	u8	rsvd2[8];
170	u32	read_addr;
171#if defined(__LITTLE_ENDIAN)
172	u8	read_addr_stride;
173	u8	read_addr_cnt;
174	u8	rsvd3[2];
175#else
176	u8	rsvd3[2];
177	u8	read_addr_cnt;
178	u8	read_addr_stride;
179#endif
180} __packed;
181
182struct __pollrd {
183	u32	sel_addr;
184	u32	read_addr;
185	u32	sel_val;
186#if defined(__LITTLE_ENDIAN)
187	u16	sel_val_stride;
188	u16	no_ops;
189#else
190	u16	no_ops;
191	u16	sel_val_stride;
192#endif
193	u32	poll_wait;
194	u32	poll_mask;
195	u32	data_size;
196	u8	rsvd[4];
197} __packed;
198
199struct __mux2 {
200	u32	sel_addr1;
201	u32	sel_addr2;
202	u32	sel_val1;
203	u32	sel_val2;
204	u32	no_ops;
205	u32	sel_val_mask;
206	u32	read_addr;
207#if defined(__LITTLE_ENDIAN)
208	u8	sel_val_stride;
209	u8	data_size;
210	u8	rsvd[2];
211#else
212	u8	rsvd[2];
213	u8	data_size;
214	u8	sel_val_stride;
215#endif
216} __packed;
217
218struct __pollrdmwr {
219	u32	addr1;
220	u32	addr2;
221	u32	val1;
222	u32	val2;
223	u32	poll_wait;
224	u32	poll_mask;
225	u32	mod_mask;
226	u32	data_size;
227} __packed;
228
229struct qlcnic_dump_entry {
230	struct qlcnic_common_entry_hdr hdr;
231	union {
232		struct __crb		crb;
233		struct __cache		cache;
234		struct __ocm		ocm;
235		struct __mem		mem;
236		struct __mux		mux;
237		struct __queue		que;
238		struct __ctrl		ctrl;
239		struct __pollrdmwr	pollrdmwr;
240		struct __mux2		mux2;
241		struct __pollrd		pollrd;
242	} region;
243} __packed;
244
245enum qlcnic_minidump_opcode {
246	QLCNIC_DUMP_NOP		= 0,
247	QLCNIC_DUMP_READ_CRB	= 1,
248	QLCNIC_DUMP_READ_MUX	= 2,
249	QLCNIC_DUMP_QUEUE	= 3,
250	QLCNIC_DUMP_BRD_CONFIG	= 4,
251	QLCNIC_DUMP_READ_OCM	= 6,
252	QLCNIC_DUMP_PEG_REG	= 7,
253	QLCNIC_DUMP_L1_DTAG	= 8,
254	QLCNIC_DUMP_L1_ITAG	= 9,
255	QLCNIC_DUMP_L1_DATA	= 11,
256	QLCNIC_DUMP_L1_INST	= 12,
257	QLCNIC_DUMP_L2_DTAG	= 21,
258	QLCNIC_DUMP_L2_ITAG	= 22,
259	QLCNIC_DUMP_L2_DATA	= 23,
260	QLCNIC_DUMP_L2_INST	= 24,
261	QLCNIC_DUMP_POLL_RD	= 35,
262	QLCNIC_READ_MUX2	= 36,
263	QLCNIC_READ_POLLRDMWR	= 37,
264	QLCNIC_DUMP_READ_ROM	= 71,
265	QLCNIC_DUMP_READ_MEM	= 72,
266	QLCNIC_DUMP_READ_CTRL	= 98,
267	QLCNIC_DUMP_TLHDR	= 99,
268	QLCNIC_DUMP_RDEND	= 255
269};
270
271inline u32 qlcnic_82xx_get_saved_state(void *t_hdr, u32 index)
272{
273	struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr;
274
275	return hdr->saved_state[index];
276}
277
278inline void qlcnic_82xx_set_saved_state(void *t_hdr, u32 index,
279					u32 value)
280{
281	struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr;
282
283	hdr->saved_state[index] = value;
284}
285
286void qlcnic_82xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump)
287{
288	struct qlcnic_82xx_dump_template_hdr *hdr;
289
290	hdr = fw_dump->tmpl_hdr;
291	fw_dump->tmpl_hdr_size = hdr->size;
292	fw_dump->version = hdr->version;
293	fw_dump->num_entries = hdr->num_entries;
294	fw_dump->offset = hdr->offset;
295
296	hdr->drv_cap_mask = hdr->cap_mask;
297	fw_dump->cap_mask = hdr->cap_mask;
298
299	fw_dump->use_pex_dma = (hdr->capabilities & BIT_0) ? true : false;
300}
301
302inline u32 qlcnic_82xx_get_cap_size(void *t_hdr, int index)
303{
304	struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr;
305
306	return hdr->cap_sizes[index];
307}
308
309void qlcnic_82xx_set_sys_info(void *t_hdr, int idx, u32 value)
310{
311	struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr;
312
313	hdr->sys_info[idx] = value;
314}
315
316void qlcnic_82xx_store_cap_mask(void *tmpl_hdr, u32 mask)
317{
318	struct qlcnic_82xx_dump_template_hdr *hdr = tmpl_hdr;
319
320	hdr->drv_cap_mask = mask;
321}
322
323inline u32 qlcnic_83xx_get_saved_state(void *t_hdr, u32 index)
324{
325	struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr;
326
327	return hdr->saved_state[index];
328}
329
330inline void qlcnic_83xx_set_saved_state(void *t_hdr, u32 index,
331					u32 value)
332{
333	struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr;
334
335	hdr->saved_state[index] = value;
336}
337
338#define QLCNIC_TEMPLATE_VERSION (0x20001)
339
340void qlcnic_83xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump)
341{
342	struct qlcnic_83xx_dump_template_hdr *hdr;
343
344	hdr = fw_dump->tmpl_hdr;
345	fw_dump->tmpl_hdr_size = hdr->size;
346	fw_dump->version = hdr->version;
347	fw_dump->num_entries = hdr->num_entries;
348	fw_dump->offset = hdr->offset;
349
350	hdr->drv_cap_mask = hdr->cap_mask;
351	fw_dump->cap_mask = hdr->cap_mask;
352
353	fw_dump->use_pex_dma = (fw_dump->version & 0xfffff) >=
354			       QLCNIC_TEMPLATE_VERSION;
355}
356
357inline u32 qlcnic_83xx_get_cap_size(void *t_hdr, int index)
358{
359	struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr;
360
361	return hdr->cap_sizes[index];
362}
363
364void qlcnic_83xx_set_sys_info(void *t_hdr, int idx, u32 value)
365{
366	struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr;
367
368	hdr->sys_info[idx] = value;
369}
370
371void qlcnic_83xx_store_cap_mask(void *tmpl_hdr, u32 mask)
372{
373	struct qlcnic_83xx_dump_template_hdr *hdr;
374
375	hdr = tmpl_hdr;
376	hdr->drv_cap_mask = mask;
377}
378
379struct qlcnic_dump_operations {
380	enum qlcnic_minidump_opcode opcode;
381	u32 (*handler)(struct qlcnic_adapter *, struct qlcnic_dump_entry *,
382		       __le32 *);
383};
384
385static u32 qlcnic_dump_crb(struct qlcnic_adapter *adapter,
386			   struct qlcnic_dump_entry *entry, __le32 *buffer)
387{
388	int i;
389	u32 addr, data;
390	struct __crb *crb = &entry->region.crb;
391
392	addr = crb->addr;
393
394	for (i = 0; i < crb->no_ops; i++) {
395		data = qlcnic_ind_rd(adapter, addr);
396		*buffer++ = cpu_to_le32(addr);
397		*buffer++ = cpu_to_le32(data);
398		addr += crb->stride;
399	}
400	return crb->no_ops * 2 * sizeof(u32);
401}
402
403static u32 qlcnic_dump_ctrl(struct qlcnic_adapter *adapter,
404			    struct qlcnic_dump_entry *entry, __le32 *buffer)
405{
406	void *hdr = adapter->ahw->fw_dump.tmpl_hdr;
407	struct __ctrl *ctr = &entry->region.ctrl;
408	int i, k, timeout = 0;
409	u32 addr, data, temp;
410	u8 no_ops;
411
412	addr = ctr->addr;
413	no_ops = ctr->no_ops;
414
415	for (i = 0; i < no_ops; i++) {
416		k = 0;
417		for (k = 0; k < 8; k++) {
418			if (!(ctr->opcode & (1 << k)))
419				continue;
420			switch (1 << k) {
421			case QLCNIC_DUMP_WCRB:
422				qlcnic_ind_wr(adapter, addr, ctr->val1);
423				break;
424			case QLCNIC_DUMP_RWCRB:
425				data = qlcnic_ind_rd(adapter, addr);
426				qlcnic_ind_wr(adapter, addr, data);
427				break;
428			case QLCNIC_DUMP_ANDCRB:
429				data = qlcnic_ind_rd(adapter, addr);
430				qlcnic_ind_wr(adapter, addr,
431					      (data & ctr->val2));
432				break;
433			case QLCNIC_DUMP_ORCRB:
434				data = qlcnic_ind_rd(adapter, addr);
435				qlcnic_ind_wr(adapter, addr,
436					      (data | ctr->val3));
437				break;
438			case QLCNIC_DUMP_POLLCRB:
439				while (timeout <= ctr->timeout) {
440					data = qlcnic_ind_rd(adapter, addr);
441					if ((data & ctr->val2) == ctr->val1)
442						break;
443					usleep_range(1000, 2000);
444					timeout++;
445				}
446				if (timeout > ctr->timeout) {
447					dev_info(&adapter->pdev->dev,
448					"Timed out, aborting poll CRB\n");
449					return -EINVAL;
450				}
451				break;
452			case QLCNIC_DUMP_RD_SAVE:
453				temp = ctr->index_a;
454				if (temp)
455					addr = qlcnic_get_saved_state(adapter,
456								      hdr,
457								      temp);
458				data = qlcnic_ind_rd(adapter, addr);
459				qlcnic_set_saved_state(adapter, hdr,
460						       ctr->index_v, data);
461				break;
462			case QLCNIC_DUMP_WRT_SAVED:
463				temp = ctr->index_v;
464				if (temp)
465					data = qlcnic_get_saved_state(adapter,
466								      hdr,
467								      temp);
468				else
469					data = ctr->val1;
470
471				temp = ctr->index_a;
472				if (temp)
473					addr = qlcnic_get_saved_state(adapter,
474								      hdr,
475								      temp);
476				qlcnic_ind_wr(adapter, addr, data);
477				break;
478			case QLCNIC_DUMP_MOD_SAVE_ST:
479				data = qlcnic_get_saved_state(adapter, hdr,
480							      ctr->index_v);
481				data <<= ctr->shl_val;
482				data >>= ctr->shr_val;
483				if (ctr->val2)
484					data &= ctr->val2;
485				data |= ctr->val3;
486				data += ctr->val1;
487				qlcnic_set_saved_state(adapter, hdr,
488						       ctr->index_v, data);
489				break;
490			default:
491				dev_info(&adapter->pdev->dev,
492					 "Unknown opcode\n");
493				break;
494			}
495		}
496		addr += ctr->stride;
497	}
498	return 0;
499}
500
501static u32 qlcnic_dump_mux(struct qlcnic_adapter *adapter,
502			   struct qlcnic_dump_entry *entry, __le32 *buffer)
503{
504	int loop;
505	u32 val, data = 0;
506	struct __mux *mux = &entry->region.mux;
507
508	val = mux->val;
509	for (loop = 0; loop < mux->no_ops; loop++) {
510		qlcnic_ind_wr(adapter, mux->addr, val);
511		data = qlcnic_ind_rd(adapter, mux->read_addr);
512		*buffer++ = cpu_to_le32(val);
513		*buffer++ = cpu_to_le32(data);
514		val += mux->val_stride;
515	}
516	return 2 * mux->no_ops * sizeof(u32);
517}
518
519static u32 qlcnic_dump_que(struct qlcnic_adapter *adapter,
520			   struct qlcnic_dump_entry *entry, __le32 *buffer)
521{
522	int i, loop;
523	u32 cnt, addr, data, que_id = 0;
524	struct __queue *que = &entry->region.que;
525
526	addr = que->read_addr;
527	cnt = que->read_addr_cnt;
528
529	for (loop = 0; loop < que->no_ops; loop++) {
530		qlcnic_ind_wr(adapter, que->sel_addr, que_id);
531		addr = que->read_addr;
532		for (i = 0; i < cnt; i++) {
533			data = qlcnic_ind_rd(adapter, addr);
534			*buffer++ = cpu_to_le32(data);
535			addr += que->read_addr_stride;
536		}
537		que_id += que->stride;
538	}
539	return que->no_ops * cnt * sizeof(u32);
540}
541
542static u32 qlcnic_dump_ocm(struct qlcnic_adapter *adapter,
543			   struct qlcnic_dump_entry *entry, __le32 *buffer)
544{
545	int i;
546	u32 data;
547	void __iomem *addr;
548	struct __ocm *ocm = &entry->region.ocm;
549
550	addr = adapter->ahw->pci_base0 + ocm->read_addr;
551	for (i = 0; i < ocm->no_ops; i++) {
552		data = readl(addr);
553		*buffer++ = cpu_to_le32(data);
554		addr += ocm->read_addr_stride;
555	}
556	return ocm->no_ops * sizeof(u32);
557}
558
559static u32 qlcnic_read_rom(struct qlcnic_adapter *adapter,
560			   struct qlcnic_dump_entry *entry, __le32 *buffer)
561{
562	int i, count = 0;
563	u32 fl_addr, size, val, lck_val, addr;
564	struct __mem *rom = &entry->region.mem;
565
566	fl_addr = rom->addr;
567	size = rom->size / 4;
568lock_try:
569	lck_val = QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_LOCK);
570	if (!lck_val && count < MAX_CTL_CHECK) {
571		usleep_range(10000, 11000);
572		count++;
573		goto lock_try;
574	}
575	QLC_SHARED_REG_WR32(adapter, QLCNIC_FLASH_LOCK_OWNER,
576			    adapter->ahw->pci_func);
577	for (i = 0; i < size; i++) {
578		addr = fl_addr & 0xFFFF0000;
579		qlcnic_ind_wr(adapter, FLASH_ROM_WINDOW, addr);
580		addr = LSW(fl_addr) + FLASH_ROM_DATA;
581		val = qlcnic_ind_rd(adapter, addr);
582		fl_addr += 4;
583		*buffer++ = cpu_to_le32(val);
584	}
585	QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_UNLOCK);
586	return rom->size;
587}
588
589static u32 qlcnic_dump_l1_cache(struct qlcnic_adapter *adapter,
590				struct qlcnic_dump_entry *entry, __le32 *buffer)
591{
592	int i;
593	u32 cnt, val, data, addr;
594	struct __cache *l1 = &entry->region.cache;
595
596	val = l1->init_tag_val;
597
598	for (i = 0; i < l1->no_ops; i++) {
599		qlcnic_ind_wr(adapter, l1->addr, val);
600		qlcnic_ind_wr(adapter, l1->ctrl_addr, LSW(l1->ctrl_val));
601		addr = l1->read_addr;
602		cnt = l1->read_addr_num;
603		while (cnt) {
604			data = qlcnic_ind_rd(adapter, addr);
605			*buffer++ = cpu_to_le32(data);
606			addr += l1->read_addr_stride;
607			cnt--;
608		}
609		val += l1->stride;
610	}
611	return l1->no_ops * l1->read_addr_num * sizeof(u32);
612}
613
614static u32 qlcnic_dump_l2_cache(struct qlcnic_adapter *adapter,
615				struct qlcnic_dump_entry *entry, __le32 *buffer)
616{
617	int i;
618	u32 cnt, val, data, addr;
619	u8 poll_mask, poll_to, time_out = 0;
620	struct __cache *l2 = &entry->region.cache;
621
622	val = l2->init_tag_val;
623	poll_mask = LSB(MSW(l2->ctrl_val));
624	poll_to = MSB(MSW(l2->ctrl_val));
625
626	for (i = 0; i < l2->no_ops; i++) {
627		qlcnic_ind_wr(adapter, l2->addr, val);
628		if (LSW(l2->ctrl_val))
629			qlcnic_ind_wr(adapter, l2->ctrl_addr,
630				      LSW(l2->ctrl_val));
631		if (!poll_mask)
632			goto skip_poll;
633		do {
634			data = qlcnic_ind_rd(adapter, l2->ctrl_addr);
635			if (!(data & poll_mask))
636				break;
637			usleep_range(1000, 2000);
638			time_out++;
639		} while (time_out <= poll_to);
640
641		if (time_out > poll_to) {
642			dev_err(&adapter->pdev->dev,
643				"Timeout exceeded in %s, aborting dump\n",
644				__func__);
645			return -EINVAL;
646		}
647skip_poll:
648		addr = l2->read_addr;
649		cnt = l2->read_addr_num;
650		while (cnt) {
651			data = qlcnic_ind_rd(adapter, addr);
652			*buffer++ = cpu_to_le32(data);
653			addr += l2->read_addr_stride;
654			cnt--;
655		}
656		val += l2->stride;
657	}
658	return l2->no_ops * l2->read_addr_num * sizeof(u32);
659}
660
661static u32 qlcnic_read_memory_test_agent(struct qlcnic_adapter *adapter,
662					 struct __mem *mem, __le32 *buffer,
663					 int *ret)
664{
665	u32 addr, data, test;
666	int i, reg_read;
667
668	reg_read = mem->size;
669	addr = mem->addr;
670	/* check for data size of multiple of 16 and 16 byte alignment */
671	if ((addr & 0xf) || (reg_read%16)) {
672		dev_info(&adapter->pdev->dev,
673			 "Unaligned memory addr:0x%x size:0x%x\n",
674			 addr, reg_read);
675		*ret = -EINVAL;
676		return 0;
677	}
678
679	mutex_lock(&adapter->ahw->mem_lock);
680
681	while (reg_read != 0) {
682		qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_LO, addr);
683		qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_HI, 0);
684		qlcnic_ind_wr(adapter, QLCNIC_MS_CTRL, QLCNIC_TA_START_ENABLE);
685
686		for (i = 0; i < MAX_CTL_CHECK; i++) {
687			test = qlcnic_ind_rd(adapter, QLCNIC_MS_CTRL);
688			if (!(test & TA_CTL_BUSY))
689				break;
690		}
691		if (i == MAX_CTL_CHECK) {
692			if (printk_ratelimit()) {
693				dev_err(&adapter->pdev->dev,
694					"failed to read through agent\n");
695				*ret = -EIO;
696				goto out;
697			}
698		}
699		for (i = 0; i < 4; i++) {
700			data = qlcnic_ind_rd(adapter, qlcnic_ms_read_data[i]);
701			*buffer++ = cpu_to_le32(data);
702		}
703		addr += 16;
704		reg_read -= 16;
705		ret += 16;
706	}
707out:
708	mutex_unlock(&adapter->ahw->mem_lock);
709	return mem->size;
710}
711
712/* DMA register base address */
713#define QLC_DMA_REG_BASE_ADDR(dma_no)	(0x77320000 + (dma_no * 0x10000))
714
715/* DMA register offsets w.r.t base address */
716#define QLC_DMA_CMD_BUFF_ADDR_LOW	0
717#define QLC_DMA_CMD_BUFF_ADDR_HI	4
718#define QLC_DMA_CMD_STATUS_CTRL		8
719
720static int qlcnic_start_pex_dma(struct qlcnic_adapter *adapter,
721				struct __mem *mem)
722{
723	struct device *dev = &adapter->pdev->dev;
724	u32 dma_no, dma_base_addr, temp_addr;
725	int i, ret, dma_sts;
726	void *tmpl_hdr;
727
728	tmpl_hdr = adapter->ahw->fw_dump.tmpl_hdr;
729	dma_no = qlcnic_get_saved_state(adapter, tmpl_hdr,
730					QLC_83XX_DMA_ENGINE_INDEX);
731	dma_base_addr = QLC_DMA_REG_BASE_ADDR(dma_no);
732
733	temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_LOW;
734	ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr);
735	if (ret)
736		return ret;
737
738	temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_HI;
739	ret = qlcnic_ind_wr(adapter, temp_addr, 0);
740	if (ret)
741		return ret;
742
743	temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL;
744	ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd);
745	if (ret)
746		return ret;
747
748	/* Wait for DMA to complete */
749	temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL;
750	for (i = 0; i < 400; i++) {
751		dma_sts = qlcnic_ind_rd(adapter, temp_addr);
752
753		if (dma_sts & BIT_1)
754			usleep_range(250, 500);
755		else
756			break;
757	}
758
759	if (i >= 400) {
760		dev_info(dev, "PEX DMA operation timed out");
761		ret = -EIO;
762	}
763
764	return ret;
765}
766
767static u32 qlcnic_read_memory_pexdma(struct qlcnic_adapter *adapter,
768				     struct __mem *mem,
769				     __le32 *buffer, int *ret)
770{
771	struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump;
772	u32 temp, dma_base_addr, size = 0, read_size = 0;
773	struct qlcnic_pex_dma_descriptor *dma_descr;
774	struct device *dev = &adapter->pdev->dev;
775	dma_addr_t dma_phys_addr;
776	void *dma_buffer;
777	void *tmpl_hdr;
778
779	tmpl_hdr = fw_dump->tmpl_hdr;
780
781	/* Check if DMA engine is available */
782	temp = qlcnic_get_saved_state(adapter, tmpl_hdr,
783				      QLC_83XX_DMA_ENGINE_INDEX);
784	dma_base_addr = QLC_DMA_REG_BASE_ADDR(temp);
785	temp = qlcnic_ind_rd(adapter,
786			     dma_base_addr + QLC_DMA_CMD_STATUS_CTRL);
787
788	if (!(temp & BIT_31)) {
789		dev_info(dev, "%s: DMA engine is not available\n", __func__);
790		*ret = -EIO;
791		return 0;
792	}
793
794	/* Create DMA descriptor */
795	dma_descr = kzalloc(sizeof(struct qlcnic_pex_dma_descriptor),
796			    GFP_KERNEL);
797	if (!dma_descr) {
798		*ret = -ENOMEM;
799		return 0;
800	}
801
802	/* dma_desc_cmd  0:15  = 0
803	 * dma_desc_cmd 16:19  = mem->dma_desc_cmd 0:3
804	 * dma_desc_cmd 20:23  = pci function number
805	 * dma_desc_cmd 24:31  = mem->dma_desc_cmd 8:15
806	 */
807	dma_phys_addr = fw_dump->phys_addr;
808	dma_buffer = fw_dump->dma_buffer;
809	temp = 0;
810	temp = mem->dma_desc_cmd & 0xff0f;
811	temp |= (adapter->ahw->pci_func & 0xf) << 4;
812	dma_descr->dma_desc_cmd = (temp << 16) & 0xffff0000;
813	dma_descr->dma_bus_addr_low = LSD(dma_phys_addr);
814	dma_descr->dma_bus_addr_high = MSD(dma_phys_addr);
815	dma_descr->src_addr_high = 0;
816
817	/* Collect memory dump using multiple DMA operations if required */
818	while (read_size < mem->size) {
819		if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE)
820			size = QLC_PEX_DMA_READ_SIZE;
821		else
822			size = mem->size - read_size;
823
824		dma_descr->src_addr_low = mem->addr + read_size;
825		dma_descr->read_data_size = size;
826
827		/* Write DMA descriptor to MS memory*/
828		temp = sizeof(struct qlcnic_pex_dma_descriptor) / 16;
829		*ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr,
830					      (u32 *)dma_descr, temp);
831		if (*ret) {
832			dev_info(dev, "Failed to write DMA descriptor to MS memory at address 0x%x\n",
833				 mem->desc_card_addr);
834			goto free_dma_descr;
835		}
836
837		*ret = qlcnic_start_pex_dma(adapter, mem);
838		if (*ret) {
839			dev_info(dev, "Failed to start PEX DMA operation\n");
840			goto free_dma_descr;
841		}
842
843		memcpy(buffer, dma_buffer, size);
844		buffer += size / 4;
845		read_size += size;
846	}
847
848free_dma_descr:
849	kfree(dma_descr);
850
851	return read_size;
852}
853
854static u32 qlcnic_read_memory(struct qlcnic_adapter *adapter,
855			      struct qlcnic_dump_entry *entry, __le32 *buffer)
856{
857	struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump;
858	struct device *dev = &adapter->pdev->dev;
859	struct __mem *mem = &entry->region.mem;
860	u32 data_size;
861	int ret = 0;
862
863	if (fw_dump->use_pex_dma) {
864		data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer,
865						      &ret);
866		if (ret)
867			dev_info(dev,
868				 "Failed to read memory dump using PEX DMA: mask[0x%x]\n",
869				 entry->hdr.mask);
870		else
871			return data_size;
872	}
873
874	data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret);
875	if (ret) {
876		dev_info(dev,
877			 "Failed to read memory dump using test agent method: mask[0x%x]\n",
878			 entry->hdr.mask);
879		return 0;
880	} else {
881		return data_size;
882	}
883}
884
885static u32 qlcnic_dump_nop(struct qlcnic_adapter *adapter,
886			   struct qlcnic_dump_entry *entry, __le32 *buffer)
887{
888	entry->hdr.flags |= QLCNIC_DUMP_SKIP;
889	return 0;
890}
891
892static int qlcnic_valid_dump_entry(struct device *dev,
893				   struct qlcnic_dump_entry *entry, u32 size)
894{
895	int ret = 1;
896	if (size != entry->hdr.cap_size) {
897		dev_err(dev,
898			"Invalid entry, Type:%d\tMask:%d\tSize:%dCap_size:%d\n",
899			entry->hdr.type, entry->hdr.mask, size,
900			entry->hdr.cap_size);
901		ret = 0;
902	}
903	return ret;
904}
905
906static u32 qlcnic_read_pollrdmwr(struct qlcnic_adapter *adapter,
907				 struct qlcnic_dump_entry *entry,
908				 __le32 *buffer)
909{
910	struct __pollrdmwr *poll = &entry->region.pollrdmwr;
911	u32 data, wait_count, poll_wait, temp;
912
913	poll_wait = poll->poll_wait;
914
915	qlcnic_ind_wr(adapter, poll->addr1, poll->val1);
916	wait_count = 0;
917
918	while (wait_count < poll_wait) {
919		data = qlcnic_ind_rd(adapter, poll->addr1);
920		if ((data & poll->poll_mask) != 0)
921			break;
922		wait_count++;
923	}
924
925	if (wait_count == poll_wait) {
926		dev_err(&adapter->pdev->dev,
927			"Timeout exceeded in %s, aborting dump\n",
928			__func__);
929		return 0;
930	}
931
932	data = qlcnic_ind_rd(adapter, poll->addr2) & poll->mod_mask;
933	qlcnic_ind_wr(adapter, poll->addr2, data);
934	qlcnic_ind_wr(adapter, poll->addr1, poll->val2);
935	wait_count = 0;
936
937	while (wait_count < poll_wait) {
938		temp = qlcnic_ind_rd(adapter, poll->addr1);
939		if ((temp & poll->poll_mask) != 0)
940			break;
941		wait_count++;
942	}
943
944	*buffer++ = cpu_to_le32(poll->addr2);
945	*buffer++ = cpu_to_le32(data);
946
947	return 2 * sizeof(u32);
948
949}
950
951static u32 qlcnic_read_pollrd(struct qlcnic_adapter *adapter,
952			      struct qlcnic_dump_entry *entry, __le32 *buffer)
953{
954	struct __pollrd *pollrd = &entry->region.pollrd;
955	u32 data, wait_count, poll_wait, sel_val;
956	int i;
957
958	poll_wait = pollrd->poll_wait;
959	sel_val = pollrd->sel_val;
960
961	for (i = 0; i < pollrd->no_ops; i++) {
962		qlcnic_ind_wr(adapter, pollrd->sel_addr, sel_val);
963		wait_count = 0;
964		while (wait_count < poll_wait) {
965			data = qlcnic_ind_rd(adapter, pollrd->sel_addr);
966			if ((data & pollrd->poll_mask) != 0)
967				break;
968			wait_count++;
969		}
970
971		if (wait_count == poll_wait) {
972			dev_err(&adapter->pdev->dev,
973				"Timeout exceeded in %s, aborting dump\n",
974				__func__);
975			return 0;
976		}
977
978		data = qlcnic_ind_rd(adapter, pollrd->read_addr);
979		*buffer++ = cpu_to_le32(sel_val);
980		*buffer++ = cpu_to_le32(data);
981		sel_val += pollrd->sel_val_stride;
982	}
983	return pollrd->no_ops * (2 * sizeof(u32));
984}
985
986static u32 qlcnic_read_mux2(struct qlcnic_adapter *adapter,
987			    struct qlcnic_dump_entry *entry, __le32 *buffer)
988{
989	struct __mux2 *mux2 = &entry->region.mux2;
990	u32 data;
991	u32 t_sel_val, sel_val1, sel_val2;
992	int i;
993
994	sel_val1 = mux2->sel_val1;
995	sel_val2 = mux2->sel_val2;
996
997	for (i = 0; i < mux2->no_ops; i++) {
998		qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val1);
999		t_sel_val = sel_val1 & mux2->sel_val_mask;
1000		qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val);
1001		data = qlcnic_ind_rd(adapter, mux2->read_addr);
1002		*buffer++ = cpu_to_le32(t_sel_val);
1003		*buffer++ = cpu_to_le32(data);
1004		qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val2);
1005		t_sel_val = sel_val2 & mux2->sel_val_mask;
1006		qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val);
1007		data = qlcnic_ind_rd(adapter, mux2->read_addr);
1008		*buffer++ = cpu_to_le32(t_sel_val);
1009		*buffer++ = cpu_to_le32(data);
1010		sel_val1 += mux2->sel_val_stride;
1011		sel_val2 += mux2->sel_val_stride;
1012	}
1013
1014	return mux2->no_ops * (4 * sizeof(u32));
1015}
1016
1017static u32 qlcnic_83xx_dump_rom(struct qlcnic_adapter *adapter,
1018				struct qlcnic_dump_entry *entry, __le32 *buffer)
1019{
1020	u32 fl_addr, size;
1021	struct __mem *rom = &entry->region.mem;
1022
1023	fl_addr = rom->addr;
1024	size = rom->size / 4;
1025
1026	if (!qlcnic_83xx_lockless_flash_read32(adapter, fl_addr,
1027					       (u8 *)buffer, size))
1028		return rom->size;
1029
1030	return 0;
1031}
1032
1033static const struct qlcnic_dump_operations qlcnic_fw_dump_ops[] = {
1034	{QLCNIC_DUMP_NOP, qlcnic_dump_nop},
1035	{QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb},
1036	{QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux},
1037	{QLCNIC_DUMP_QUEUE, qlcnic_dump_que},
1038	{QLCNIC_DUMP_BRD_CONFIG, qlcnic_read_rom},
1039	{QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm},
1040	{QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl},
1041	{QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache},
1042	{QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache},
1043	{QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache},
1044	{QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache},
1045	{QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache},
1046	{QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache},
1047	{QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache},
1048	{QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache},
1049	{QLCNIC_DUMP_READ_ROM, qlcnic_read_rom},
1050	{QLCNIC_DUMP_READ_MEM, qlcnic_read_memory},
1051	{QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl},
1052	{QLCNIC_DUMP_TLHDR, qlcnic_dump_nop},
1053	{QLCNIC_DUMP_RDEND, qlcnic_dump_nop},
1054};
1055
1056static const struct qlcnic_dump_operations qlcnic_83xx_fw_dump_ops[] = {
1057	{QLCNIC_DUMP_NOP, qlcnic_dump_nop},
1058	{QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb},
1059	{QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux},
1060	{QLCNIC_DUMP_QUEUE, qlcnic_dump_que},
1061	{QLCNIC_DUMP_BRD_CONFIG, qlcnic_83xx_dump_rom},
1062	{QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm},
1063	{QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl},
1064	{QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache},
1065	{QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache},
1066	{QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache},
1067	{QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache},
1068	{QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache},
1069	{QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache},
1070	{QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache},
1071	{QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache},
1072	{QLCNIC_DUMP_POLL_RD, qlcnic_read_pollrd},
1073	{QLCNIC_READ_MUX2, qlcnic_read_mux2},
1074	{QLCNIC_READ_POLLRDMWR, qlcnic_read_pollrdmwr},
1075	{QLCNIC_DUMP_READ_ROM, qlcnic_83xx_dump_rom},
1076	{QLCNIC_DUMP_READ_MEM, qlcnic_read_memory},
1077	{QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl},
1078	{QLCNIC_DUMP_TLHDR, qlcnic_dump_nop},
1079	{QLCNIC_DUMP_RDEND, qlcnic_dump_nop},
1080};
1081
1082static uint32_t qlcnic_temp_checksum(uint32_t *temp_buffer, u32 temp_size)
1083{
1084	uint64_t sum = 0;
1085	int count = temp_size / sizeof(uint32_t);
1086	while (count-- > 0)
1087		sum += *temp_buffer++;
1088	while (sum >> 32)
1089		sum = (sum & 0xFFFFFFFF) + (sum >> 32);
1090	return ~sum;
1091}
1092
1093static int qlcnic_fw_flash_get_minidump_temp(struct qlcnic_adapter *adapter,
1094					     u8 *buffer, u32 size)
1095{
1096	int ret = 0;
1097
1098	if (qlcnic_82xx_check(adapter))
1099		return -EIO;
1100
1101	if (qlcnic_83xx_lock_flash(adapter))
1102		return -EIO;
1103
1104	ret = qlcnic_83xx_lockless_flash_read32(adapter,
1105						QLC_83XX_MINIDUMP_FLASH,
1106						buffer, size / sizeof(u32));
1107
1108	qlcnic_83xx_unlock_flash(adapter);
1109
1110	return ret;
1111}
1112
1113static int
1114qlcnic_fw_flash_get_minidump_temp_size(struct qlcnic_adapter *adapter,
1115				       struct qlcnic_cmd_args *cmd)
1116{
1117	struct qlcnic_83xx_dump_template_hdr tmp_hdr;
1118	u32 size = sizeof(tmp_hdr) / sizeof(u32);
1119	int ret = 0;
1120
1121	if (qlcnic_82xx_check(adapter))
1122		return -EIO;
1123
1124	if (qlcnic_83xx_lock_flash(adapter))
1125		return -EIO;
1126
1127	ret = qlcnic_83xx_lockless_flash_read32(adapter,
1128						QLC_83XX_MINIDUMP_FLASH,
1129						(u8 *)&tmp_hdr, size);
1130
1131	qlcnic_83xx_unlock_flash(adapter);
1132
1133	cmd->rsp.arg[2] = tmp_hdr.size;
1134	cmd->rsp.arg[3] = tmp_hdr.version;
1135
1136	return ret;
1137}
1138
1139static int qlcnic_fw_get_minidump_temp_size(struct qlcnic_adapter *adapter,
1140					    u32 *version, u32 *temp_size,
1141					    u8 *use_flash_temp)
1142{
1143	int err = 0;
1144	struct qlcnic_cmd_args cmd;
1145
1146	if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_TEMP_SIZE))
1147		return -ENOMEM;
1148
1149	err = qlcnic_issue_cmd(adapter, &cmd);
1150	if (err != QLCNIC_RCODE_SUCCESS) {
1151		if (qlcnic_fw_flash_get_minidump_temp_size(adapter, &cmd)) {
1152			qlcnic_free_mbx_args(&cmd);
1153			return -EIO;
1154		}
1155		*use_flash_temp = 1;
1156	}
1157
1158	*temp_size = cmd.rsp.arg[2];
1159	*version = cmd.rsp.arg[3];
1160	qlcnic_free_mbx_args(&cmd);
1161
1162	if (!(*temp_size))
1163		return -EIO;
1164
1165	return 0;
1166}
1167
1168static int __qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter,
1169					     u32 *buffer, u32 temp_size)
1170{
1171	int err = 0, i;
1172	void *tmp_addr;
1173	__le32 *tmp_buf;
1174	struct qlcnic_cmd_args cmd;
1175	dma_addr_t tmp_addr_t = 0;
1176
1177	tmp_addr = dma_alloc_coherent(&adapter->pdev->dev, temp_size,
1178				      &tmp_addr_t, GFP_KERNEL);
1179	if (!tmp_addr)
1180		return -ENOMEM;
1181
1182	if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_GET_TEMP_HDR)) {
1183		err = -ENOMEM;
1184		goto free_mem;
1185	}
1186
1187	cmd.req.arg[1] = LSD(tmp_addr_t);
1188	cmd.req.arg[2] = MSD(tmp_addr_t);
1189	cmd.req.arg[3] = temp_size;
1190	err = qlcnic_issue_cmd(adapter, &cmd);
1191
1192	tmp_buf = tmp_addr;
1193	if (err == QLCNIC_RCODE_SUCCESS) {
1194		for (i = 0; i < temp_size / sizeof(u32); i++)
1195			*buffer++ = __le32_to_cpu(*tmp_buf++);
1196	}
1197
1198	qlcnic_free_mbx_args(&cmd);
1199
1200free_mem:
1201	dma_free_coherent(&adapter->pdev->dev, temp_size, tmp_addr, tmp_addr_t);
1202
1203	return err;
1204}
1205
1206int qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter)
1207{
1208	struct qlcnic_hardware_context *ahw;
1209	struct qlcnic_fw_dump *fw_dump;
1210	u32 version, csum, *tmp_buf;
1211	u8 use_flash_temp = 0;
1212	u32 temp_size = 0;
1213	void *temp_buffer;
1214	int err;
1215
1216	ahw = adapter->ahw;
1217	fw_dump = &ahw->fw_dump;
1218	err = qlcnic_fw_get_minidump_temp_size(adapter, &version, &temp_size,
1219					       &use_flash_temp);
1220	if (err) {
1221		dev_err(&adapter->pdev->dev,
1222			"Can't get template size %d\n", err);
1223		return -EIO;
1224	}
1225
1226	fw_dump->tmpl_hdr = vzalloc(temp_size);
1227	if (!fw_dump->tmpl_hdr)
1228		return -ENOMEM;
1229
1230	tmp_buf = (u32 *)fw_dump->tmpl_hdr;
1231	if (use_flash_temp)
1232		goto flash_temp;
1233
1234	err = __qlcnic_fw_cmd_get_minidump_temp(adapter, tmp_buf, temp_size);
1235
1236	if (err) {
1237flash_temp:
1238		err = qlcnic_fw_flash_get_minidump_temp(adapter, (u8 *)tmp_buf,
1239							temp_size);
1240
1241		if (err) {
1242			dev_err(&adapter->pdev->dev,
1243				"Failed to get minidump template header %d\n",
1244				err);
1245			vfree(fw_dump->tmpl_hdr);
1246			fw_dump->tmpl_hdr = NULL;
1247			return -EIO;
1248		}
1249	}
1250
1251	csum = qlcnic_temp_checksum((uint32_t *)tmp_buf, temp_size);
1252
1253	if (csum) {
1254		dev_err(&adapter->pdev->dev,
1255			"Template header checksum validation failed\n");
1256		vfree(fw_dump->tmpl_hdr);
1257		fw_dump->tmpl_hdr = NULL;
1258		return -EIO;
1259	}
1260
1261	qlcnic_cache_tmpl_hdr_values(adapter, fw_dump);
1262
1263	if (fw_dump->use_pex_dma) {
1264		fw_dump->dma_buffer = NULL;
1265		temp_buffer = dma_alloc_coherent(&adapter->pdev->dev,
1266						 QLC_PEX_DMA_READ_SIZE,
1267						 &fw_dump->phys_addr,
1268						 GFP_KERNEL);
1269		if (!temp_buffer)
1270			fw_dump->use_pex_dma = false;
1271		else
1272			fw_dump->dma_buffer = temp_buffer;
1273	}
1274
1275
1276	dev_info(&adapter->pdev->dev,
1277		 "Default minidump capture mask 0x%x\n",
1278		 fw_dump->cap_mask);
1279
1280	qlcnic_enable_fw_dump_state(adapter);
1281
1282	return 0;
1283}
1284
1285int qlcnic_dump_fw(struct qlcnic_adapter *adapter)
1286{
1287	struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump;
1288	static const struct qlcnic_dump_operations *fw_dump_ops;
1289	struct qlcnic_83xx_dump_template_hdr *hdr_83xx;
1290	u32 entry_offset, dump, no_entries, buf_offset = 0;
1291	int i, k, ops_cnt, ops_index, dump_size = 0;
1292	struct device *dev = &adapter->pdev->dev;
1293	struct qlcnic_hardware_context *ahw;
1294	struct qlcnic_dump_entry *entry;
1295	void *tmpl_hdr;
1296	u32 ocm_window;
1297	__le32 *buffer;
1298	char mesg[64];
1299	char *msg[] = {mesg, NULL};
1300
1301	ahw = adapter->ahw;
1302	tmpl_hdr = fw_dump->tmpl_hdr;
1303
1304	/* Return if we don't have firmware dump template header */
1305	if (!tmpl_hdr)
1306		return -EIO;
1307
1308	if (!qlcnic_check_fw_dump_state(adapter)) {
1309		dev_info(&adapter->pdev->dev, "Dump not enabled\n");
1310		return -EIO;
1311	}
1312
1313	if (fw_dump->clr) {
1314		dev_info(&adapter->pdev->dev,
1315			 "Previous dump not cleared, not capturing dump\n");
1316		return -EIO;
1317	}
1318
1319	netif_info(adapter->ahw, drv, adapter->netdev, "Take FW dump\n");
1320	/* Calculate the size for dump data area only */
1321	for (i = 2, k = 1; (i & QLCNIC_DUMP_MASK_MAX); i <<= 1, k++)
1322		if (i & fw_dump->cap_mask)
1323			dump_size += qlcnic_get_cap_size(adapter, tmpl_hdr, k);
1324
1325	if (!dump_size)
1326		return -EIO;
1327
1328	fw_dump->data = vzalloc(dump_size);
1329	if (!fw_dump->data)
1330		return -ENOMEM;
1331
1332	buffer = fw_dump->data;
1333	fw_dump->size = dump_size;
1334	no_entries = fw_dump->num_entries;
1335	entry_offset = fw_dump->offset;
1336	qlcnic_set_sys_info(adapter, tmpl_hdr, 0, QLCNIC_DRIVER_VERSION);
1337	qlcnic_set_sys_info(adapter, tmpl_hdr, 1, adapter->fw_version);
1338
1339	if (qlcnic_82xx_check(adapter)) {
1340		ops_cnt = ARRAY_SIZE(qlcnic_fw_dump_ops);
1341		fw_dump_ops = qlcnic_fw_dump_ops;
1342	} else {
1343		hdr_83xx = tmpl_hdr;
1344		ops_cnt = ARRAY_SIZE(qlcnic_83xx_fw_dump_ops);
1345		fw_dump_ops = qlcnic_83xx_fw_dump_ops;
1346		ocm_window = hdr_83xx->ocm_wnd_reg[ahw->pci_func];
1347		hdr_83xx->saved_state[QLC_83XX_OCM_INDEX] = ocm_window;
1348		hdr_83xx->saved_state[QLC_83XX_PCI_INDEX] = ahw->pci_func;
1349	}
1350
1351	for (i = 0; i < no_entries; i++) {
1352		entry = tmpl_hdr + entry_offset;
1353		if (!(entry->hdr.mask & fw_dump->cap_mask)) {
1354			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
1355			entry_offset += entry->hdr.offset;
1356			continue;
1357		}
1358
1359		/* Find the handler for this entry */
1360		ops_index = 0;
1361		while (ops_index < ops_cnt) {
1362			if (entry->hdr.type == fw_dump_ops[ops_index].opcode)
1363				break;
1364			ops_index++;
1365		}
1366
1367		if (ops_index == ops_cnt) {
1368			dev_info(dev, "Skipping unknown entry opcode %d\n",
1369				 entry->hdr.type);
1370			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
1371			entry_offset += entry->hdr.offset;
1372			continue;
1373		}
1374
1375		/* Collect dump for this entry */
1376		dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer);
1377		if (!qlcnic_valid_dump_entry(dev, entry, dump)) {
1378			entry->hdr.flags |= QLCNIC_DUMP_SKIP;
1379			entry_offset += entry->hdr.offset;
1380			continue;
1381		}
1382
1383		buf_offset += entry->hdr.cap_size;
1384		entry_offset += entry->hdr.offset;
1385		buffer = fw_dump->data + buf_offset;
1386	}
1387
1388	fw_dump->clr = 1;
1389	snprintf(mesg, sizeof(mesg), "FW_DUMP=%s", adapter->netdev->name);
1390	netdev_info(adapter->netdev,
1391		    "Dump data %d bytes captured, template header size %d bytes\n",
1392		    fw_dump->size, fw_dump->tmpl_hdr_size);
1393	/* Send a udev event to notify availability of FW dump */
1394	kobject_uevent_env(&dev->kobj, KOBJ_CHANGE, msg);
1395
1396	return 0;
1397}
1398
1399void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *adapter)
1400{
1401	u32 prev_version, current_version;
1402	struct qlcnic_hardware_context *ahw = adapter->ahw;
1403	struct qlcnic_fw_dump *fw_dump = &ahw->fw_dump;
1404	struct pci_dev *pdev = adapter->pdev;
1405
1406	prev_version = adapter->fw_version;
1407	current_version = qlcnic_83xx_get_fw_version(adapter);
1408
1409	if (fw_dump->tmpl_hdr == NULL || current_version > prev_version) {
1410		vfree(fw_dump->tmpl_hdr);
1411		if (!qlcnic_fw_cmd_get_minidump_temp(adapter))
1412			dev_info(&pdev->dev, "Supports FW dump capability\n");
1413	}
1414}
1415