1
2
3
4
5#include "main.h"
6#include "fw.h"
7#include "tx.h"
8#include "reg.h"
9#include "debug.h"
10
11static void rtw_fw_c2h_cmd_handle_ext(struct rtw_dev *rtwdev,
12 struct sk_buff *skb)
13{
14 struct rtw_c2h_cmd *c2h;
15 u8 sub_cmd_id;
16
17 c2h = get_c2h_from_skb(skb);
18 sub_cmd_id = c2h->payload[0];
19
20 switch (sub_cmd_id) {
21 case C2H_CCX_RPT:
22 rtw_tx_report_handle(rtwdev, skb);
23 break;
24 default:
25 break;
26 }
27}
28
29void rtw_fw_c2h_cmd_handle(struct rtw_dev *rtwdev, struct sk_buff *skb)
30{
31 struct rtw_c2h_cmd *c2h;
32 u32 pkt_offset;
33 u8 len;
34
35 pkt_offset = *((u32 *)skb->cb);
36 c2h = (struct rtw_c2h_cmd *)(skb->data + pkt_offset);
37 len = skb->len - pkt_offset - 2;
38
39 rtw_dbg(rtwdev, RTW_DBG_FW, "recv C2H, id=0x%02x, seq=0x%02x, len=%d\n",
40 c2h->id, c2h->seq, len);
41
42 switch (c2h->id) {
43 case C2H_HALMAC:
44 rtw_fw_c2h_cmd_handle_ext(rtwdev, skb);
45 break;
46 default:
47 break;
48 }
49}
50
51static void rtw_fw_send_h2c_command(struct rtw_dev *rtwdev,
52 u8 *h2c)
53{
54 u8 box;
55 u8 box_state;
56 u32 box_reg, box_ex_reg;
57 u32 h2c_wait;
58 int idx;
59
60 rtw_dbg(rtwdev, RTW_DBG_FW,
61 "send H2C content %02x%02x%02x%02x %02x%02x%02x%02x\n",
62 h2c[3], h2c[2], h2c[1], h2c[0],
63 h2c[7], h2c[6], h2c[5], h2c[4]);
64
65 spin_lock(&rtwdev->h2c.lock);
66
67 box = rtwdev->h2c.last_box_num;
68 switch (box) {
69 case 0:
70 box_reg = REG_HMEBOX0;
71 box_ex_reg = REG_HMEBOX0_EX;
72 break;
73 case 1:
74 box_reg = REG_HMEBOX1;
75 box_ex_reg = REG_HMEBOX1_EX;
76 break;
77 case 2:
78 box_reg = REG_HMEBOX2;
79 box_ex_reg = REG_HMEBOX2_EX;
80 break;
81 case 3:
82 box_reg = REG_HMEBOX3;
83 box_ex_reg = REG_HMEBOX3_EX;
84 break;
85 default:
86 WARN(1, "invalid h2c mail box number\n");
87 goto out;
88 }
89
90 h2c_wait = 20;
91 do {
92 box_state = rtw_read8(rtwdev, REG_HMETFR);
93 } while ((box_state >> box) & 0x1 && --h2c_wait > 0);
94
95 if (!h2c_wait) {
96 rtw_err(rtwdev, "failed to send h2c command\n");
97 goto out;
98 }
99
100 for (idx = 0; idx < 4; idx++)
101 rtw_write8(rtwdev, box_reg + idx, h2c[idx]);
102 for (idx = 0; idx < 4; idx++)
103 rtw_write8(rtwdev, box_ex_reg + idx, h2c[idx + 4]);
104
105 if (++rtwdev->h2c.last_box_num >= 4)
106 rtwdev->h2c.last_box_num = 0;
107
108out:
109 spin_unlock(&rtwdev->h2c.lock);
110}
111
112static void rtw_fw_send_h2c_packet(struct rtw_dev *rtwdev, u8 *h2c_pkt)
113{
114 int ret;
115
116 spin_lock(&rtwdev->h2c.lock);
117
118 FW_OFFLOAD_H2C_SET_SEQ_NUM(h2c_pkt, rtwdev->h2c.seq);
119 ret = rtw_hci_write_data_h2c(rtwdev, h2c_pkt, H2C_PKT_SIZE);
120 if (ret)
121 rtw_err(rtwdev, "failed to send h2c packet\n");
122 rtwdev->h2c.seq++;
123
124 spin_unlock(&rtwdev->h2c.lock);
125}
126
127void
128rtw_fw_send_general_info(struct rtw_dev *rtwdev)
129{
130 struct rtw_fifo_conf *fifo = &rtwdev->fifo;
131 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
132 u16 total_size = H2C_PKT_HDR_SIZE + 4;
133
134 rtw_h2c_pkt_set_header(h2c_pkt, H2C_PKT_GENERAL_INFO);
135
136 SET_PKT_H2C_TOTAL_LEN(h2c_pkt, total_size);
137
138 GENERAL_INFO_SET_FW_TX_BOUNDARY(h2c_pkt,
139 fifo->rsvd_fw_txbuf_addr -
140 fifo->rsvd_boundary);
141
142 rtw_fw_send_h2c_packet(rtwdev, h2c_pkt);
143}
144
145void
146rtw_fw_send_phydm_info(struct rtw_dev *rtwdev)
147{
148 struct rtw_hal *hal = &rtwdev->hal;
149 struct rtw_efuse *efuse = &rtwdev->efuse;
150 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
151 u16 total_size = H2C_PKT_HDR_SIZE + 8;
152 u8 fw_rf_type = 0;
153
154 if (hal->rf_type == RF_1T1R)
155 fw_rf_type = FW_RF_1T1R;
156 else if (hal->rf_type == RF_2T2R)
157 fw_rf_type = FW_RF_2T2R;
158
159 rtw_h2c_pkt_set_header(h2c_pkt, H2C_PKT_PHYDM_INFO);
160
161 SET_PKT_H2C_TOTAL_LEN(h2c_pkt, total_size);
162 PHYDM_INFO_SET_REF_TYPE(h2c_pkt, efuse->rfe_option);
163 PHYDM_INFO_SET_RF_TYPE(h2c_pkt, fw_rf_type);
164 PHYDM_INFO_SET_CUT_VER(h2c_pkt, hal->cut_version);
165 PHYDM_INFO_SET_RX_ANT_STATUS(h2c_pkt, hal->antenna_tx);
166 PHYDM_INFO_SET_TX_ANT_STATUS(h2c_pkt, hal->antenna_rx);
167
168 rtw_fw_send_h2c_packet(rtwdev, h2c_pkt);
169}
170
171void rtw_fw_do_iqk(struct rtw_dev *rtwdev, struct rtw_iqk_para *para)
172{
173 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
174 u16 total_size = H2C_PKT_HDR_SIZE + 1;
175
176 rtw_h2c_pkt_set_header(h2c_pkt, H2C_PKT_IQK);
177 SET_PKT_H2C_TOTAL_LEN(h2c_pkt, total_size);
178 IQK_SET_CLEAR(h2c_pkt, para->clear);
179 IQK_SET_SEGMENT_IQK(h2c_pkt, para->segment_iqk);
180
181 rtw_fw_send_h2c_packet(rtwdev, h2c_pkt);
182}
183
184void rtw_fw_send_rssi_info(struct rtw_dev *rtwdev, struct rtw_sta_info *si)
185{
186 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
187 u8 rssi = ewma_rssi_read(&si->avg_rssi);
188 bool stbc_en = si->stbc_en ? true : false;
189
190 SET_H2C_CMD_ID_CLASS(h2c_pkt, H2C_CMD_RSSI_MONITOR);
191
192 SET_RSSI_INFO_MACID(h2c_pkt, si->mac_id);
193 SET_RSSI_INFO_RSSI(h2c_pkt, rssi);
194 SET_RSSI_INFO_STBC(h2c_pkt, stbc_en);
195
196 rtw_fw_send_h2c_command(rtwdev, h2c_pkt);
197}
198
199void rtw_fw_send_ra_info(struct rtw_dev *rtwdev, struct rtw_sta_info *si)
200{
201 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
202 bool no_update = si->updated;
203 bool disable_pt = true;
204
205 SET_H2C_CMD_ID_CLASS(h2c_pkt, H2C_CMD_RA_INFO);
206
207 SET_RA_INFO_MACID(h2c_pkt, si->mac_id);
208 SET_RA_INFO_RATE_ID(h2c_pkt, si->rate_id);
209 SET_RA_INFO_INIT_RA_LVL(h2c_pkt, si->init_ra_lv);
210 SET_RA_INFO_SGI_EN(h2c_pkt, si->sgi_enable);
211 SET_RA_INFO_BW_MODE(h2c_pkt, si->bw_mode);
212 SET_RA_INFO_LDPC(h2c_pkt, si->ldpc_en);
213 SET_RA_INFO_NO_UPDATE(h2c_pkt, no_update);
214 SET_RA_INFO_VHT_EN(h2c_pkt, si->vht_enable);
215 SET_RA_INFO_DIS_PT(h2c_pkt, disable_pt);
216 SET_RA_INFO_RA_MASK0(h2c_pkt, (si->ra_mask & 0xff));
217 SET_RA_INFO_RA_MASK1(h2c_pkt, (si->ra_mask & 0xff00) >> 8);
218 SET_RA_INFO_RA_MASK2(h2c_pkt, (si->ra_mask & 0xff0000) >> 16);
219 SET_RA_INFO_RA_MASK3(h2c_pkt, (si->ra_mask & 0xff000000) >> 24);
220
221 si->init_ra_lv = 0;
222 si->updated = true;
223
224 rtw_fw_send_h2c_command(rtwdev, h2c_pkt);
225}
226
227void rtw_fw_media_status_report(struct rtw_dev *rtwdev, u8 mac_id, bool connect)
228{
229 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
230
231 SET_H2C_CMD_ID_CLASS(h2c_pkt, H2C_CMD_MEDIA_STATUS_RPT);
232 MEDIA_STATUS_RPT_SET_OP_MODE(h2c_pkt, connect);
233 MEDIA_STATUS_RPT_SET_MACID(h2c_pkt, mac_id);
234
235 rtw_fw_send_h2c_command(rtwdev, h2c_pkt);
236}
237
238void rtw_fw_set_pwr_mode(struct rtw_dev *rtwdev)
239{
240 struct rtw_lps_conf *conf = &rtwdev->lps_conf;
241 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
242
243 SET_H2C_CMD_ID_CLASS(h2c_pkt, H2C_CMD_SET_PWR_MODE);
244
245 SET_PWR_MODE_SET_MODE(h2c_pkt, conf->mode);
246 SET_PWR_MODE_SET_RLBM(h2c_pkt, conf->rlbm);
247 SET_PWR_MODE_SET_SMART_PS(h2c_pkt, conf->smart_ps);
248 SET_PWR_MODE_SET_AWAKE_INTERVAL(h2c_pkt, conf->awake_interval);
249 SET_PWR_MODE_SET_PORT_ID(h2c_pkt, conf->port_id);
250 SET_PWR_MODE_SET_PWR_STATE(h2c_pkt, conf->state);
251
252 rtw_fw_send_h2c_command(rtwdev, h2c_pkt);
253}
254
255static u8 rtw_get_rsvd_page_location(struct rtw_dev *rtwdev,
256 enum rtw_rsvd_packet_type type)
257{
258 struct rtw_rsvd_page *rsvd_pkt;
259 u8 location = 0;
260
261 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list) {
262 if (type == rsvd_pkt->type)
263 location = rsvd_pkt->page;
264 }
265
266 return location;
267}
268
269void rtw_send_rsvd_page_h2c(struct rtw_dev *rtwdev)
270{
271 u8 h2c_pkt[H2C_PKT_SIZE] = {0};
272 u8 location = 0;
273
274 SET_H2C_CMD_ID_CLASS(h2c_pkt, H2C_CMD_RSVD_PAGE);
275
276 location = rtw_get_rsvd_page_location(rtwdev, RSVD_PROBE_RESP);
277 *(h2c_pkt + 1) = location;
278 rtw_dbg(rtwdev, RTW_DBG_FW, "RSVD_PROBE_RESP loc: %d\n", location);
279
280 location = rtw_get_rsvd_page_location(rtwdev, RSVD_PS_POLL);
281 *(h2c_pkt + 2) = location;
282 rtw_dbg(rtwdev, RTW_DBG_FW, "RSVD_PS_POLL loc: %d\n", location);
283
284 location = rtw_get_rsvd_page_location(rtwdev, RSVD_NULL);
285 *(h2c_pkt + 3) = location;
286 rtw_dbg(rtwdev, RTW_DBG_FW, "RSVD_NULL loc: %d\n", location);
287
288 location = rtw_get_rsvd_page_location(rtwdev, RSVD_QOS_NULL);
289 *(h2c_pkt + 4) = location;
290 rtw_dbg(rtwdev, RTW_DBG_FW, "RSVD_QOS_NULL loc: %d\n", location);
291
292 rtw_fw_send_h2c_command(rtwdev, h2c_pkt);
293}
294
295static struct sk_buff *
296rtw_beacon_get(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
297{
298 struct sk_buff *skb_new;
299
300 if (vif->type != NL80211_IFTYPE_AP &&
301 vif->type != NL80211_IFTYPE_ADHOC &&
302 !ieee80211_vif_is_mesh(vif)) {
303 skb_new = alloc_skb(1, GFP_KERNEL);
304 if (!skb_new)
305 return NULL;
306 skb_put(skb_new, 1);
307 } else {
308 skb_new = ieee80211_beacon_get(hw, vif);
309 }
310
311 return skb_new;
312}
313
314static struct sk_buff *rtw_get_rsvd_page_skb(struct ieee80211_hw *hw,
315 struct ieee80211_vif *vif,
316 enum rtw_rsvd_packet_type type)
317{
318 struct sk_buff *skb_new;
319
320 switch (type) {
321 case RSVD_BEACON:
322 skb_new = rtw_beacon_get(hw, vif);
323 break;
324 case RSVD_PS_POLL:
325 skb_new = ieee80211_pspoll_get(hw, vif);
326 break;
327 case RSVD_PROBE_RESP:
328 skb_new = ieee80211_proberesp_get(hw, vif);
329 break;
330 case RSVD_NULL:
331 skb_new = ieee80211_nullfunc_get(hw, vif, false);
332 break;
333 case RSVD_QOS_NULL:
334 skb_new = ieee80211_nullfunc_get(hw, vif, true);
335 break;
336 default:
337 return NULL;
338 }
339
340 if (!skb_new)
341 return NULL;
342
343 return skb_new;
344}
345
346static void rtw_fill_rsvd_page_desc(struct rtw_dev *rtwdev, struct sk_buff *skb)
347{
348 struct rtw_tx_pkt_info pkt_info;
349 struct rtw_chip_info *chip = rtwdev->chip;
350 u8 *pkt_desc;
351
352 memset(&pkt_info, 0, sizeof(pkt_info));
353 rtw_rsvd_page_pkt_info_update(rtwdev, &pkt_info, skb);
354 pkt_desc = skb_push(skb, chip->tx_pkt_desc_sz);
355 memset(pkt_desc, 0, chip->tx_pkt_desc_sz);
356 rtw_tx_fill_tx_desc(&pkt_info, skb);
357}
358
359static inline u8 rtw_len_to_page(unsigned int len, u8 page_size)
360{
361 return DIV_ROUND_UP(len, page_size);
362}
363
364static void rtw_rsvd_page_list_to_buf(struct rtw_dev *rtwdev, u8 page_size,
365 u8 page_margin, u32 page, u8 *buf,
366 struct rtw_rsvd_page *rsvd_pkt)
367{
368 struct sk_buff *skb = rsvd_pkt->skb;
369
370 if (rsvd_pkt->add_txdesc)
371 rtw_fill_rsvd_page_desc(rtwdev, skb);
372
373 if (page >= 1)
374 memcpy(buf + page_margin + page_size * (page - 1),
375 skb->data, skb->len);
376 else
377 memcpy(buf, skb->data, skb->len);
378}
379
380void rtw_add_rsvd_page(struct rtw_dev *rtwdev, enum rtw_rsvd_packet_type type,
381 bool txdesc)
382{
383 struct rtw_rsvd_page *rsvd_pkt;
384
385 lockdep_assert_held(&rtwdev->mutex);
386
387 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list) {
388 if (rsvd_pkt->type == type)
389 return;
390 }
391
392 rsvd_pkt = kmalloc(sizeof(*rsvd_pkt), GFP_KERNEL);
393 if (!rsvd_pkt)
394 return;
395
396 rsvd_pkt->type = type;
397 rsvd_pkt->add_txdesc = txdesc;
398 list_add_tail(&rsvd_pkt->list, &rtwdev->rsvd_page_list);
399}
400
401void rtw_reset_rsvd_page(struct rtw_dev *rtwdev)
402{
403 struct rtw_rsvd_page *rsvd_pkt, *tmp;
404
405 lockdep_assert_held(&rtwdev->mutex);
406
407 list_for_each_entry_safe(rsvd_pkt, tmp, &rtwdev->rsvd_page_list, list) {
408 if (rsvd_pkt->type == RSVD_BEACON)
409 continue;
410 list_del(&rsvd_pkt->list);
411 kfree(rsvd_pkt);
412 }
413}
414
415int rtw_fw_write_data_rsvd_page(struct rtw_dev *rtwdev, u16 pg_addr,
416 u8 *buf, u32 size)
417{
418 u8 bckp[2];
419 u8 val;
420 u16 rsvd_pg_head;
421 int ret;
422
423 lockdep_assert_held(&rtwdev->mutex);
424
425 if (!size)
426 return -EINVAL;
427
428 pg_addr &= BIT_MASK_BCN_HEAD_1_V1;
429 rtw_write16(rtwdev, REG_FIFOPAGE_CTRL_2, pg_addr | BIT_BCN_VALID_V1);
430
431 val = rtw_read8(rtwdev, REG_CR + 1);
432 bckp[0] = val;
433 val |= BIT_ENSWBCN >> 8;
434 rtw_write8(rtwdev, REG_CR + 1, val);
435
436 val = rtw_read8(rtwdev, REG_FWHW_TXQ_CTRL + 2);
437 bckp[1] = val;
438 val &= ~(BIT_EN_BCNQ_DL >> 16);
439 rtw_write8(rtwdev, REG_FWHW_TXQ_CTRL + 2, val);
440
441 ret = rtw_hci_write_data_rsvd_page(rtwdev, buf, size);
442 if (ret) {
443 rtw_err(rtwdev, "failed to write data to rsvd page\n");
444 goto restore;
445 }
446
447 if (!check_hw_ready(rtwdev, REG_FIFOPAGE_CTRL_2, BIT_BCN_VALID_V1, 1)) {
448 rtw_err(rtwdev, "error beacon valid\n");
449 ret = -EBUSY;
450 }
451
452restore:
453 rsvd_pg_head = rtwdev->fifo.rsvd_boundary;
454 rtw_write16(rtwdev, REG_FIFOPAGE_CTRL_2,
455 rsvd_pg_head | BIT_BCN_VALID_V1);
456 rtw_write8(rtwdev, REG_FWHW_TXQ_CTRL + 2, bckp[1]);
457 rtw_write8(rtwdev, REG_CR + 1, bckp[0]);
458
459 return ret;
460}
461
462static int rtw_download_drv_rsvd_page(struct rtw_dev *rtwdev, u8 *buf, u32 size)
463{
464 u32 pg_size;
465 u32 pg_num = 0;
466 u16 pg_addr = 0;
467
468 pg_size = rtwdev->chip->page_size;
469 pg_num = size / pg_size + ((size & (pg_size - 1)) ? 1 : 0);
470 if (pg_num > rtwdev->fifo.rsvd_drv_pg_num)
471 return -ENOMEM;
472
473 pg_addr = rtwdev->fifo.rsvd_drv_addr;
474
475 return rtw_fw_write_data_rsvd_page(rtwdev, pg_addr, buf, size);
476}
477
478static u8 *rtw_build_rsvd_page(struct rtw_dev *rtwdev,
479 struct ieee80211_vif *vif, u32 *size)
480{
481 struct ieee80211_hw *hw = rtwdev->hw;
482 struct rtw_chip_info *chip = rtwdev->chip;
483 struct sk_buff *iter;
484 struct rtw_rsvd_page *rsvd_pkt;
485 u32 page = 0;
486 u8 total_page = 0;
487 u8 page_size, page_margin, tx_desc_sz;
488 u8 *buf;
489
490 page_size = chip->page_size;
491 tx_desc_sz = chip->tx_pkt_desc_sz;
492 page_margin = page_size - tx_desc_sz;
493
494 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list) {
495 iter = rtw_get_rsvd_page_skb(hw, vif, rsvd_pkt->type);
496 if (!iter) {
497 rtw_err(rtwdev, "fail to build rsvd packet\n");
498 goto release_skb;
499 }
500 rsvd_pkt->skb = iter;
501 rsvd_pkt->page = total_page;
502 if (rsvd_pkt->add_txdesc)
503 total_page += rtw_len_to_page(iter->len + tx_desc_sz,
504 page_size);
505 else
506 total_page += rtw_len_to_page(iter->len, page_size);
507 }
508
509 if (total_page > rtwdev->fifo.rsvd_drv_pg_num) {
510 rtw_err(rtwdev, "rsvd page over size: %d\n", total_page);
511 goto release_skb;
512 }
513
514 *size = (total_page - 1) * page_size + page_margin;
515 buf = kzalloc(*size, GFP_KERNEL);
516 if (!buf)
517 goto release_skb;
518
519 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list) {
520 rtw_rsvd_page_list_to_buf(rtwdev, page_size, page_margin,
521 page, buf, rsvd_pkt);
522 page += rtw_len_to_page(rsvd_pkt->skb->len, page_size);
523 }
524 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list)
525 kfree_skb(rsvd_pkt->skb);
526
527 return buf;
528
529release_skb:
530 list_for_each_entry(rsvd_pkt, &rtwdev->rsvd_page_list, list)
531 kfree_skb(rsvd_pkt->skb);
532
533 return NULL;
534}
535
536static int
537rtw_download_beacon(struct rtw_dev *rtwdev, struct ieee80211_vif *vif)
538{
539 struct ieee80211_hw *hw = rtwdev->hw;
540 struct sk_buff *skb;
541 int ret = 0;
542
543 skb = rtw_beacon_get(hw, vif);
544 if (!skb) {
545 rtw_err(rtwdev, "failed to get beacon skb\n");
546 ret = -ENOMEM;
547 goto out;
548 }
549
550 ret = rtw_download_drv_rsvd_page(rtwdev, skb->data, skb->len);
551 if (ret)
552 rtw_err(rtwdev, "failed to download drv rsvd page\n");
553
554 dev_kfree_skb(skb);
555
556out:
557 return ret;
558}
559
560int rtw_fw_download_rsvd_page(struct rtw_dev *rtwdev, struct ieee80211_vif *vif)
561{
562 u8 *buf;
563 u32 size;
564 int ret;
565
566 buf = rtw_build_rsvd_page(rtwdev, vif, &size);
567 if (!buf) {
568 rtw_err(rtwdev, "failed to build rsvd page pkt\n");
569 return -ENOMEM;
570 }
571
572 ret = rtw_download_drv_rsvd_page(rtwdev, buf, size);
573 if (ret) {
574 rtw_err(rtwdev, "failed to download drv rsvd page\n");
575 goto free;
576 }
577
578 ret = rtw_download_beacon(rtwdev, vif);
579 if (ret) {
580 rtw_err(rtwdev, "failed to download beacon\n");
581 goto free;
582 }
583
584free:
585 kfree(buf);
586
587 return ret;
588}
589
590int rtw_dump_drv_rsvd_page(struct rtw_dev *rtwdev,
591 u32 offset, u32 size, u32 *buf)
592{
593 struct rtw_fifo_conf *fifo = &rtwdev->fifo;
594 u32 residue, i;
595 u16 start_pg;
596 u16 idx = 0;
597 u16 ctl;
598 u8 rcr;
599
600 if (size & 0x3) {
601 rtw_warn(rtwdev, "should be 4-byte aligned\n");
602 return -EINVAL;
603 }
604
605 offset += fifo->rsvd_boundary << TX_PAGE_SIZE_SHIFT;
606 residue = offset & (FIFO_PAGE_SIZE - 1);
607 start_pg = offset >> FIFO_PAGE_SIZE_SHIFT;
608 start_pg += RSVD_PAGE_START_ADDR;
609
610 rcr = rtw_read8(rtwdev, REG_RCR + 2);
611 ctl = rtw_read16(rtwdev, REG_PKTBUF_DBG_CTRL) & 0xf000;
612
613
614 rtw_write8(rtwdev, REG_RCR, rcr | BIT(3));
615
616 do {
617 rtw_write16(rtwdev, REG_PKTBUF_DBG_CTRL, start_pg | ctl);
618
619 for (i = FIFO_DUMP_ADDR + residue;
620 i < FIFO_DUMP_ADDR + FIFO_PAGE_SIZE; i += 4) {
621 buf[idx++] = rtw_read32(rtwdev, i);
622 size -= 4;
623 if (size == 0)
624 goto out;
625 }
626
627 residue = 0;
628 start_pg++;
629 } while (size);
630
631out:
632 rtw_write16(rtwdev, REG_PKTBUF_DBG_CTRL, ctl);
633 rtw_write8(rtwdev, REG_RCR + 2, rcr);
634 return 0;
635}
636