diff options
Diffstat (limited to 'drivers/net/wireless/iwlwifi/iwl-rx.c')
-rw-r--r-- | drivers/net/wireless/iwlwifi/iwl-rx.c | 758 |
1 files changed, 4 insertions, 754 deletions
diff --git a/drivers/net/wireless/iwlwifi/iwl-rx.c b/drivers/net/wireless/iwlwifi/iwl-rx.c index d3b2fb389e58..267eb8935902 100644 --- a/drivers/net/wireless/iwlwifi/iwl-rx.c +++ b/drivers/net/wireless/iwlwifi/iwl-rx.c | |||
@@ -163,197 +163,6 @@ void iwl_rx_queue_update_write_ptr(struct iwl_priv *priv, struct iwl_rx_queue *q | |||
163 | spin_unlock_irqrestore(&q->lock, flags); | 163 | spin_unlock_irqrestore(&q->lock, flags); |
164 | } | 164 | } |
165 | EXPORT_SYMBOL(iwl_rx_queue_update_write_ptr); | 165 | EXPORT_SYMBOL(iwl_rx_queue_update_write_ptr); |
166 | /** | ||
167 | * iwl_dma_addr2rbd_ptr - convert a DMA address to a uCode read buffer ptr | ||
168 | */ | ||
169 | static inline __le32 iwl_dma_addr2rbd_ptr(struct iwl_priv *priv, | ||
170 | dma_addr_t dma_addr) | ||
171 | { | ||
172 | return cpu_to_le32((u32)(dma_addr >> 8)); | ||
173 | } | ||
174 | |||
175 | /** | ||
176 | * iwl_rx_queue_restock - refill RX queue from pre-allocated pool | ||
177 | * | ||
178 | * If there are slots in the RX queue that need to be restocked, | ||
179 | * and we have free pre-allocated buffers, fill the ranks as much | ||
180 | * as we can, pulling from rx_free. | ||
181 | * | ||
182 | * This moves the 'write' index forward to catch up with 'processed', and | ||
183 | * also updates the memory address in the firmware to reference the new | ||
184 | * target buffer. | ||
185 | */ | ||
186 | void iwl_rx_queue_restock(struct iwl_priv *priv) | ||
187 | { | ||
188 | struct iwl_rx_queue *rxq = &priv->rxq; | ||
189 | struct list_head *element; | ||
190 | struct iwl_rx_mem_buffer *rxb; | ||
191 | unsigned long flags; | ||
192 | int write; | ||
193 | |||
194 | spin_lock_irqsave(&rxq->lock, flags); | ||
195 | write = rxq->write & ~0x7; | ||
196 | while ((iwl_rx_queue_space(rxq) > 0) && (rxq->free_count)) { | ||
197 | /* Get next free Rx buffer, remove from free list */ | ||
198 | element = rxq->rx_free.next; | ||
199 | rxb = list_entry(element, struct iwl_rx_mem_buffer, list); | ||
200 | list_del(element); | ||
201 | |||
202 | /* Point to Rx buffer via next RBD in circular buffer */ | ||
203 | rxq->bd[rxq->write] = iwl_dma_addr2rbd_ptr(priv, rxb->page_dma); | ||
204 | rxq->queue[rxq->write] = rxb; | ||
205 | rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; | ||
206 | rxq->free_count--; | ||
207 | } | ||
208 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
209 | /* If the pre-allocated buffer pool is dropping low, schedule to | ||
210 | * refill it */ | ||
211 | if (rxq->free_count <= RX_LOW_WATERMARK) | ||
212 | queue_work(priv->workqueue, &priv->rx_replenish); | ||
213 | |||
214 | |||
215 | /* If we've added more space for the firmware to place data, tell it. | ||
216 | * Increment device's write pointer in multiples of 8. */ | ||
217 | if (rxq->write_actual != (rxq->write & ~0x7)) { | ||
218 | spin_lock_irqsave(&rxq->lock, flags); | ||
219 | rxq->need_update = 1; | ||
220 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
221 | iwl_rx_queue_update_write_ptr(priv, rxq); | ||
222 | } | ||
223 | } | ||
224 | EXPORT_SYMBOL(iwl_rx_queue_restock); | ||
225 | |||
226 | |||
227 | /** | ||
228 | * iwl_rx_replenish - Move all used packet from rx_used to rx_free | ||
229 | * | ||
230 | * When moving to rx_free an SKB is allocated for the slot. | ||
231 | * | ||
232 | * Also restock the Rx queue via iwl_rx_queue_restock. | ||
233 | * This is called as a scheduled work item (except for during initialization) | ||
234 | */ | ||
235 | void iwl_rx_allocate(struct iwl_priv *priv, gfp_t priority) | ||
236 | { | ||
237 | struct iwl_rx_queue *rxq = &priv->rxq; | ||
238 | struct list_head *element; | ||
239 | struct iwl_rx_mem_buffer *rxb; | ||
240 | struct page *page; | ||
241 | unsigned long flags; | ||
242 | gfp_t gfp_mask = priority; | ||
243 | |||
244 | while (1) { | ||
245 | spin_lock_irqsave(&rxq->lock, flags); | ||
246 | if (list_empty(&rxq->rx_used)) { | ||
247 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
248 | return; | ||
249 | } | ||
250 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
251 | |||
252 | if (rxq->free_count > RX_LOW_WATERMARK) | ||
253 | gfp_mask |= __GFP_NOWARN; | ||
254 | |||
255 | if (priv->hw_params.rx_page_order > 0) | ||
256 | gfp_mask |= __GFP_COMP; | ||
257 | |||
258 | /* Alloc a new receive buffer */ | ||
259 | page = alloc_pages(gfp_mask, priv->hw_params.rx_page_order); | ||
260 | if (!page) { | ||
261 | if (net_ratelimit()) | ||
262 | IWL_DEBUG_INFO(priv, "alloc_pages failed, " | ||
263 | "order: %d\n", | ||
264 | priv->hw_params.rx_page_order); | ||
265 | |||
266 | if ((rxq->free_count <= RX_LOW_WATERMARK) && | ||
267 | net_ratelimit()) | ||
268 | IWL_CRIT(priv, "Failed to alloc_pages with %s. Only %u free buffers remaining.\n", | ||
269 | priority == GFP_ATOMIC ? "GFP_ATOMIC" : "GFP_KERNEL", | ||
270 | rxq->free_count); | ||
271 | /* We don't reschedule replenish work here -- we will | ||
272 | * call the restock method and if it still needs | ||
273 | * more buffers it will schedule replenish */ | ||
274 | return; | ||
275 | } | ||
276 | |||
277 | spin_lock_irqsave(&rxq->lock, flags); | ||
278 | |||
279 | if (list_empty(&rxq->rx_used)) { | ||
280 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
281 | __free_pages(page, priv->hw_params.rx_page_order); | ||
282 | return; | ||
283 | } | ||
284 | element = rxq->rx_used.next; | ||
285 | rxb = list_entry(element, struct iwl_rx_mem_buffer, list); | ||
286 | list_del(element); | ||
287 | |||
288 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
289 | |||
290 | rxb->page = page; | ||
291 | /* Get physical address of the RB */ | ||
292 | rxb->page_dma = pci_map_page(priv->pci_dev, page, 0, | ||
293 | PAGE_SIZE << priv->hw_params.rx_page_order, | ||
294 | PCI_DMA_FROMDEVICE); | ||
295 | /* dma address must be no more than 36 bits */ | ||
296 | BUG_ON(rxb->page_dma & ~DMA_BIT_MASK(36)); | ||
297 | /* and also 256 byte aligned! */ | ||
298 | BUG_ON(rxb->page_dma & DMA_BIT_MASK(8)); | ||
299 | |||
300 | spin_lock_irqsave(&rxq->lock, flags); | ||
301 | |||
302 | list_add_tail(&rxb->list, &rxq->rx_free); | ||
303 | rxq->free_count++; | ||
304 | priv->alloc_rxb_page++; | ||
305 | |||
306 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
307 | } | ||
308 | } | ||
309 | |||
310 | void iwl_rx_replenish(struct iwl_priv *priv) | ||
311 | { | ||
312 | unsigned long flags; | ||
313 | |||
314 | iwl_rx_allocate(priv, GFP_KERNEL); | ||
315 | |||
316 | spin_lock_irqsave(&priv->lock, flags); | ||
317 | iwl_rx_queue_restock(priv); | ||
318 | spin_unlock_irqrestore(&priv->lock, flags); | ||
319 | } | ||
320 | EXPORT_SYMBOL(iwl_rx_replenish); | ||
321 | |||
322 | void iwl_rx_replenish_now(struct iwl_priv *priv) | ||
323 | { | ||
324 | iwl_rx_allocate(priv, GFP_ATOMIC); | ||
325 | |||
326 | iwl_rx_queue_restock(priv); | ||
327 | } | ||
328 | EXPORT_SYMBOL(iwl_rx_replenish_now); | ||
329 | |||
330 | |||
331 | /* Assumes that the skb field of the buffers in 'pool' is kept accurate. | ||
332 | * If an SKB has been detached, the POOL needs to have its SKB set to NULL | ||
333 | * This free routine walks the list of POOL entries and if SKB is set to | ||
334 | * non NULL it is unmapped and freed | ||
335 | */ | ||
336 | void iwl_rx_queue_free(struct iwl_priv *priv, struct iwl_rx_queue *rxq) | ||
337 | { | ||
338 | int i; | ||
339 | for (i = 0; i < RX_QUEUE_SIZE + RX_FREE_BUFFERS; i++) { | ||
340 | if (rxq->pool[i].page != NULL) { | ||
341 | pci_unmap_page(priv->pci_dev, rxq->pool[i].page_dma, | ||
342 | PAGE_SIZE << priv->hw_params.rx_page_order, | ||
343 | PCI_DMA_FROMDEVICE); | ||
344 | __iwl_free_pages(priv, rxq->pool[i].page); | ||
345 | rxq->pool[i].page = NULL; | ||
346 | } | ||
347 | } | ||
348 | |||
349 | dma_free_coherent(&priv->pci_dev->dev, 4 * RX_QUEUE_SIZE, rxq->bd, | ||
350 | rxq->dma_addr); | ||
351 | dma_free_coherent(&priv->pci_dev->dev, sizeof(struct iwl_rb_status), | ||
352 | rxq->rb_stts, rxq->rb_stts_dma); | ||
353 | rxq->bd = NULL; | ||
354 | rxq->rb_stts = NULL; | ||
355 | } | ||
356 | EXPORT_SYMBOL(iwl_rx_queue_free); | ||
357 | 166 | ||
358 | int iwl_rx_queue_alloc(struct iwl_priv *priv) | 167 | int iwl_rx_queue_alloc(struct iwl_priv *priv) |
359 | { | 168 | { |
@@ -396,98 +205,6 @@ err_bd: | |||
396 | } | 205 | } |
397 | EXPORT_SYMBOL(iwl_rx_queue_alloc); | 206 | EXPORT_SYMBOL(iwl_rx_queue_alloc); |
398 | 207 | ||
399 | void iwl_rx_queue_reset(struct iwl_priv *priv, struct iwl_rx_queue *rxq) | ||
400 | { | ||
401 | unsigned long flags; | ||
402 | int i; | ||
403 | spin_lock_irqsave(&rxq->lock, flags); | ||
404 | INIT_LIST_HEAD(&rxq->rx_free); | ||
405 | INIT_LIST_HEAD(&rxq->rx_used); | ||
406 | /* Fill the rx_used queue with _all_ of the Rx buffers */ | ||
407 | for (i = 0; i < RX_FREE_BUFFERS + RX_QUEUE_SIZE; i++) { | ||
408 | /* In the reset function, these buffers may have been allocated | ||
409 | * to an SKB, so we need to unmap and free potential storage */ | ||
410 | if (rxq->pool[i].page != NULL) { | ||
411 | pci_unmap_page(priv->pci_dev, rxq->pool[i].page_dma, | ||
412 | PAGE_SIZE << priv->hw_params.rx_page_order, | ||
413 | PCI_DMA_FROMDEVICE); | ||
414 | __iwl_free_pages(priv, rxq->pool[i].page); | ||
415 | rxq->pool[i].page = NULL; | ||
416 | } | ||
417 | list_add_tail(&rxq->pool[i].list, &rxq->rx_used); | ||
418 | } | ||
419 | |||
420 | /* Set us so that we have processed and used all buffers, but have | ||
421 | * not restocked the Rx queue with fresh buffers */ | ||
422 | rxq->read = rxq->write = 0; | ||
423 | rxq->write_actual = 0; | ||
424 | rxq->free_count = 0; | ||
425 | spin_unlock_irqrestore(&rxq->lock, flags); | ||
426 | } | ||
427 | |||
428 | int iwl_rx_init(struct iwl_priv *priv, struct iwl_rx_queue *rxq) | ||
429 | { | ||
430 | u32 rb_size; | ||
431 | const u32 rfdnlog = RX_QUEUE_SIZE_LOG; /* 256 RBDs */ | ||
432 | u32 rb_timeout = 0; /* FIXME: RX_RB_TIMEOUT for all devices? */ | ||
433 | |||
434 | if (!priv->cfg->use_isr_legacy) | ||
435 | rb_timeout = RX_RB_TIMEOUT; | ||
436 | |||
437 | if (priv->cfg->mod_params->amsdu_size_8K) | ||
438 | rb_size = FH_RCSR_RX_CONFIG_REG_VAL_RB_SIZE_8K; | ||
439 | else | ||
440 | rb_size = FH_RCSR_RX_CONFIG_REG_VAL_RB_SIZE_4K; | ||
441 | |||
442 | /* Stop Rx DMA */ | ||
443 | iwl_write_direct32(priv, FH_MEM_RCSR_CHNL0_CONFIG_REG, 0); | ||
444 | |||
445 | /* Reset driver's Rx queue write index */ | ||
446 | iwl_write_direct32(priv, FH_RSCSR_CHNL0_RBDCB_WPTR_REG, 0); | ||
447 | |||
448 | /* Tell device where to find RBD circular buffer in DRAM */ | ||
449 | iwl_write_direct32(priv, FH_RSCSR_CHNL0_RBDCB_BASE_REG, | ||
450 | (u32)(rxq->dma_addr >> 8)); | ||
451 | |||
452 | /* Tell device where in DRAM to update its Rx status */ | ||
453 | iwl_write_direct32(priv, FH_RSCSR_CHNL0_STTS_WPTR_REG, | ||
454 | rxq->rb_stts_dma >> 4); | ||
455 | |||
456 | /* Enable Rx DMA | ||
457 | * FH_RCSR_CHNL0_RX_IGNORE_RXF_EMPTY is set because of HW bug in | ||
458 | * the credit mechanism in 5000 HW RX FIFO | ||
459 | * Direct rx interrupts to hosts | ||
460 | * Rx buffer size 4 or 8k | ||
461 | * RB timeout 0x10 | ||
462 | * 256 RBDs | ||
463 | */ | ||
464 | iwl_write_direct32(priv, FH_MEM_RCSR_CHNL0_CONFIG_REG, | ||
465 | FH_RCSR_RX_CONFIG_CHNL_EN_ENABLE_VAL | | ||
466 | FH_RCSR_CHNL0_RX_IGNORE_RXF_EMPTY | | ||
467 | FH_RCSR_CHNL0_RX_CONFIG_IRQ_DEST_INT_HOST_VAL | | ||
468 | FH_RCSR_CHNL0_RX_CONFIG_SINGLE_FRAME_MSK | | ||
469 | rb_size| | ||
470 | (rb_timeout << FH_RCSR_RX_CONFIG_REG_IRQ_RBTH_POS)| | ||
471 | (rfdnlog << FH_RCSR_RX_CONFIG_RBDCB_SIZE_POS)); | ||
472 | |||
473 | /* Set interrupt coalescing timer to default (2048 usecs) */ | ||
474 | iwl_write8(priv, CSR_INT_COALESCING, IWL_HOST_INT_TIMEOUT_DEF); | ||
475 | |||
476 | return 0; | ||
477 | } | ||
478 | |||
479 | int iwl_rxq_stop(struct iwl_priv *priv) | ||
480 | { | ||
481 | |||
482 | /* stop Rx DMA */ | ||
483 | iwl_write_direct32(priv, FH_MEM_RCSR_CHNL0_CONFIG_REG, 0); | ||
484 | iwl_poll_direct_bit(priv, FH_MEM_RSSR_RX_STATUS_REG, | ||
485 | FH_RSSR_CHNL0_RX_STATUS_CHNL_IDLE, 1000); | ||
486 | |||
487 | return 0; | ||
488 | } | ||
489 | EXPORT_SYMBOL(iwl_rxq_stop); | ||
490 | |||
491 | void iwl_rx_missed_beacon_notif(struct iwl_priv *priv, | 208 | void iwl_rx_missed_beacon_notif(struct iwl_priv *priv, |
492 | struct iwl_rx_mem_buffer *rxb) | 209 | struct iwl_rx_mem_buffer *rxb) |
493 | 210 | ||
@@ -543,6 +260,7 @@ static void iwl_rx_calc_noise(struct iwl_priv *priv) | |||
543 | le32_to_cpu(rx_info->beacon_silence_rssi_b) & IN_BAND_FILTER; | 260 | le32_to_cpu(rx_info->beacon_silence_rssi_b) & IN_BAND_FILTER; |
544 | int bcn_silence_c = | 261 | int bcn_silence_c = |
545 | le32_to_cpu(rx_info->beacon_silence_rssi_c) & IN_BAND_FILTER; | 262 | le32_to_cpu(rx_info->beacon_silence_rssi_c) & IN_BAND_FILTER; |
263 | int last_rx_noise; | ||
546 | 264 | ||
547 | if (bcn_silence_a) { | 265 | if (bcn_silence_a) { |
548 | total_silence += bcn_silence_a; | 266 | total_silence += bcn_silence_a; |
@@ -559,13 +277,13 @@ static void iwl_rx_calc_noise(struct iwl_priv *priv) | |||
559 | 277 | ||
560 | /* Average among active antennas */ | 278 | /* Average among active antennas */ |
561 | if (num_active_rx) | 279 | if (num_active_rx) |
562 | priv->last_rx_noise = (total_silence / num_active_rx) - 107; | 280 | last_rx_noise = (total_silence / num_active_rx) - 107; |
563 | else | 281 | else |
564 | priv->last_rx_noise = IWL_NOISE_MEAS_NOT_AVAILABLE; | 282 | last_rx_noise = IWL_NOISE_MEAS_NOT_AVAILABLE; |
565 | 283 | ||
566 | IWL_DEBUG_CALIB(priv, "inband silence a %u, b %u, c %u, dBm %d\n", | 284 | IWL_DEBUG_CALIB(priv, "inband silence a %u, b %u, c %u, dBm %d\n", |
567 | bcn_silence_a, bcn_silence_b, bcn_silence_c, | 285 | bcn_silence_a, bcn_silence_b, bcn_silence_c, |
568 | priv->last_rx_noise); | 286 | last_rx_noise); |
569 | } | 287 | } |
570 | 288 | ||
571 | #ifdef CONFIG_IWLWIFI_DEBUG | 289 | #ifdef CONFIG_IWLWIFI_DEBUG |
@@ -617,63 +335,6 @@ static void iwl_accumulative_statistics(struct iwl_priv *priv, | |||
617 | 335 | ||
618 | #define REG_RECALIB_PERIOD (60) | 336 | #define REG_RECALIB_PERIOD (60) |
619 | 337 | ||
620 | /* the threshold ratio of actual_ack_cnt to expected_ack_cnt in percent */ | ||
621 | #define ACK_CNT_RATIO (50) | ||
622 | #define BA_TIMEOUT_CNT (5) | ||
623 | #define BA_TIMEOUT_MAX (16) | ||
624 | |||
625 | #if defined(CONFIG_IWLAGN) || defined(CONFIG_IWLAGN_MODULE) | ||
626 | /** | ||
627 | * iwl_good_ack_health - checks for ACK count ratios, BA timeout retries. | ||
628 | * | ||
629 | * When the ACK count ratio is 0 and aggregated BA timeout retries exceeding | ||
630 | * the BA_TIMEOUT_MAX, reload firmware and bring system back to normal | ||
631 | * operation state. | ||
632 | */ | ||
633 | bool iwl_good_ack_health(struct iwl_priv *priv, | ||
634 | struct iwl_rx_packet *pkt) | ||
635 | { | ||
636 | bool rc = true; | ||
637 | int actual_ack_cnt_delta, expected_ack_cnt_delta; | ||
638 | int ba_timeout_delta; | ||
639 | |||
640 | actual_ack_cnt_delta = | ||
641 | le32_to_cpu(pkt->u.stats.tx.actual_ack_cnt) - | ||
642 | le32_to_cpu(priv->statistics.tx.actual_ack_cnt); | ||
643 | expected_ack_cnt_delta = | ||
644 | le32_to_cpu(pkt->u.stats.tx.expected_ack_cnt) - | ||
645 | le32_to_cpu(priv->statistics.tx.expected_ack_cnt); | ||
646 | ba_timeout_delta = | ||
647 | le32_to_cpu(pkt->u.stats.tx.agg.ba_timeout) - | ||
648 | le32_to_cpu(priv->statistics.tx.agg.ba_timeout); | ||
649 | if ((priv->_agn.agg_tids_count > 0) && | ||
650 | (expected_ack_cnt_delta > 0) && | ||
651 | (((actual_ack_cnt_delta * 100) / expected_ack_cnt_delta) | ||
652 | < ACK_CNT_RATIO) && | ||
653 | (ba_timeout_delta > BA_TIMEOUT_CNT)) { | ||
654 | IWL_DEBUG_RADIO(priv, "actual_ack_cnt delta = %d," | ||
655 | " expected_ack_cnt = %d\n", | ||
656 | actual_ack_cnt_delta, expected_ack_cnt_delta); | ||
657 | |||
658 | #ifdef CONFIG_IWLWIFI_DEBUG | ||
659 | IWL_DEBUG_RADIO(priv, "rx_detected_cnt delta = %d\n", | ||
660 | priv->delta_statistics.tx.rx_detected_cnt); | ||
661 | IWL_DEBUG_RADIO(priv, | ||
662 | "ack_or_ba_timeout_collision delta = %d\n", | ||
663 | priv->delta_statistics.tx. | ||
664 | ack_or_ba_timeout_collision); | ||
665 | #endif | ||
666 | IWL_DEBUG_RADIO(priv, "agg ba_timeout delta = %d\n", | ||
667 | ba_timeout_delta); | ||
668 | if (!actual_ack_cnt_delta && | ||
669 | (ba_timeout_delta >= BA_TIMEOUT_MAX)) | ||
670 | rc = false; | ||
671 | } | ||
672 | return rc; | ||
673 | } | ||
674 | EXPORT_SYMBOL(iwl_good_ack_health); | ||
675 | #endif | ||
676 | |||
677 | /** | 338 | /** |
678 | * iwl_good_plcp_health - checks for plcp error. | 339 | * iwl_good_plcp_health - checks for plcp error. |
679 | * | 340 | * |
@@ -830,139 +491,6 @@ void iwl_reply_statistics(struct iwl_priv *priv, | |||
830 | } | 491 | } |
831 | EXPORT_SYMBOL(iwl_reply_statistics); | 492 | EXPORT_SYMBOL(iwl_reply_statistics); |
832 | 493 | ||
833 | /* Calc max signal level (dBm) among 3 possible receivers */ | ||
834 | static inline int iwl_calc_rssi(struct iwl_priv *priv, | ||
835 | struct iwl_rx_phy_res *rx_resp) | ||
836 | { | ||
837 | return priv->cfg->ops->utils->calc_rssi(priv, rx_resp); | ||
838 | } | ||
839 | |||
840 | #ifdef CONFIG_IWLWIFI_DEBUG | ||
841 | /** | ||
842 | * iwl_dbg_report_frame - dump frame to syslog during debug sessions | ||
843 | * | ||
844 | * You may hack this function to show different aspects of received frames, | ||
845 | * including selective frame dumps. | ||
846 | * group100 parameter selects whether to show 1 out of 100 good data frames. | ||
847 | * All beacon and probe response frames are printed. | ||
848 | */ | ||
849 | static void iwl_dbg_report_frame(struct iwl_priv *priv, | ||
850 | struct iwl_rx_phy_res *phy_res, u16 length, | ||
851 | struct ieee80211_hdr *header, int group100) | ||
852 | { | ||
853 | u32 to_us; | ||
854 | u32 print_summary = 0; | ||
855 | u32 print_dump = 0; /* set to 1 to dump all frames' contents */ | ||
856 | u32 hundred = 0; | ||
857 | u32 dataframe = 0; | ||
858 | __le16 fc; | ||
859 | u16 seq_ctl; | ||
860 | u16 channel; | ||
861 | u16 phy_flags; | ||
862 | u32 rate_n_flags; | ||
863 | u32 tsf_low; | ||
864 | int rssi; | ||
865 | |||
866 | if (likely(!(iwl_get_debug_level(priv) & IWL_DL_RX))) | ||
867 | return; | ||
868 | |||
869 | /* MAC header */ | ||
870 | fc = header->frame_control; | ||
871 | seq_ctl = le16_to_cpu(header->seq_ctrl); | ||
872 | |||
873 | /* metadata */ | ||
874 | channel = le16_to_cpu(phy_res->channel); | ||
875 | phy_flags = le16_to_cpu(phy_res->phy_flags); | ||
876 | rate_n_flags = le32_to_cpu(phy_res->rate_n_flags); | ||
877 | |||
878 | /* signal statistics */ | ||
879 | rssi = iwl_calc_rssi(priv, phy_res); | ||
880 | tsf_low = le64_to_cpu(phy_res->timestamp) & 0x0ffffffff; | ||
881 | |||
882 | to_us = !compare_ether_addr(header->addr1, priv->mac_addr); | ||
883 | |||
884 | /* if data frame is to us and all is good, | ||
885 | * (optionally) print summary for only 1 out of every 100 */ | ||
886 | if (to_us && (fc & ~cpu_to_le16(IEEE80211_FCTL_PROTECTED)) == | ||
887 | cpu_to_le16(IEEE80211_FCTL_FROMDS | IEEE80211_FTYPE_DATA)) { | ||
888 | dataframe = 1; | ||
889 | if (!group100) | ||
890 | print_summary = 1; /* print each frame */ | ||
891 | else if (priv->framecnt_to_us < 100) { | ||
892 | priv->framecnt_to_us++; | ||
893 | print_summary = 0; | ||
894 | } else { | ||
895 | priv->framecnt_to_us = 0; | ||
896 | print_summary = 1; | ||
897 | hundred = 1; | ||
898 | } | ||
899 | } else { | ||
900 | /* print summary for all other frames */ | ||
901 | print_summary = 1; | ||
902 | } | ||
903 | |||
904 | if (print_summary) { | ||
905 | char *title; | ||
906 | int rate_idx; | ||
907 | u32 bitrate; | ||
908 | |||
909 | if (hundred) | ||
910 | title = "100Frames"; | ||
911 | else if (ieee80211_has_retry(fc)) | ||
912 | title = "Retry"; | ||
913 | else if (ieee80211_is_assoc_resp(fc)) | ||
914 | title = "AscRsp"; | ||
915 | else if (ieee80211_is_reassoc_resp(fc)) | ||
916 | title = "RasRsp"; | ||
917 | else if (ieee80211_is_probe_resp(fc)) { | ||
918 | title = "PrbRsp"; | ||
919 | print_dump = 1; /* dump frame contents */ | ||
920 | } else if (ieee80211_is_beacon(fc)) { | ||
921 | title = "Beacon"; | ||
922 | print_dump = 1; /* dump frame contents */ | ||
923 | } else if (ieee80211_is_atim(fc)) | ||
924 | title = "ATIM"; | ||
925 | else if (ieee80211_is_auth(fc)) | ||
926 | title = "Auth"; | ||
927 | else if (ieee80211_is_deauth(fc)) | ||
928 | title = "DeAuth"; | ||
929 | else if (ieee80211_is_disassoc(fc)) | ||
930 | title = "DisAssoc"; | ||
931 | else | ||
932 | title = "Frame"; | ||
933 | |||
934 | rate_idx = iwl_hwrate_to_plcp_idx(rate_n_flags); | ||
935 | if (unlikely((rate_idx < 0) || (rate_idx >= IWL_RATE_COUNT))) { | ||
936 | bitrate = 0; | ||
937 | WARN_ON_ONCE(1); | ||
938 | } else { | ||
939 | bitrate = iwl_rates[rate_idx].ieee / 2; | ||
940 | } | ||
941 | |||
942 | /* print frame summary. | ||
943 | * MAC addresses show just the last byte (for brevity), | ||
944 | * but you can hack it to show more, if you'd like to. */ | ||
945 | if (dataframe) | ||
946 | IWL_DEBUG_RX(priv, "%s: mhd=0x%04x, dst=0x%02x, " | ||
947 | "len=%u, rssi=%d, chnl=%d, rate=%u, \n", | ||
948 | title, le16_to_cpu(fc), header->addr1[5], | ||
949 | length, rssi, channel, bitrate); | ||
950 | else { | ||
951 | /* src/dst addresses assume managed mode */ | ||
952 | IWL_DEBUG_RX(priv, "%s: 0x%04x, dst=0x%02x, src=0x%02x, " | ||
953 | "len=%u, rssi=%d, tim=%lu usec, " | ||
954 | "phy=0x%02x, chnl=%d\n", | ||
955 | title, le16_to_cpu(fc), header->addr1[5], | ||
956 | header->addr3[5], length, rssi, | ||
957 | tsf_low - priv->scan_start_tsf, | ||
958 | phy_flags, channel); | ||
959 | } | ||
960 | } | ||
961 | if (print_dump) | ||
962 | iwl_print_hex_dump(priv, IWL_DL_RX, header, length); | ||
963 | } | ||
964 | #endif | ||
965 | |||
966 | /* | 494 | /* |
967 | * returns non-zero if packet should be dropped | 495 | * returns non-zero if packet should be dropped |
968 | */ | 496 | */ |
@@ -1010,281 +538,3 @@ int iwl_set_decrypted_flag(struct iwl_priv *priv, | |||
1010 | return 0; | 538 | return 0; |
1011 | } | 539 | } |
1012 | EXPORT_SYMBOL(iwl_set_decrypted_flag); | 540 | EXPORT_SYMBOL(iwl_set_decrypted_flag); |
1013 | |||
1014 | static u32 iwl_translate_rx_status(struct iwl_priv *priv, u32 decrypt_in) | ||
1015 | { | ||
1016 | u32 decrypt_out = 0; | ||
1017 | |||
1018 | if ((decrypt_in & RX_RES_STATUS_STATION_FOUND) == | ||
1019 | RX_RES_STATUS_STATION_FOUND) | ||
1020 | decrypt_out |= (RX_RES_STATUS_STATION_FOUND | | ||
1021 | RX_RES_STATUS_NO_STATION_INFO_MISMATCH); | ||
1022 | |||
1023 | decrypt_out |= (decrypt_in & RX_RES_STATUS_SEC_TYPE_MSK); | ||
1024 | |||
1025 | /* packet was not encrypted */ | ||
1026 | if ((decrypt_in & RX_RES_STATUS_SEC_TYPE_MSK) == | ||
1027 | RX_RES_STATUS_SEC_TYPE_NONE) | ||
1028 | return decrypt_out; | ||
1029 | |||
1030 | /* packet was encrypted with unknown alg */ | ||
1031 | if ((decrypt_in & RX_RES_STATUS_SEC_TYPE_MSK) == | ||
1032 | RX_RES_STATUS_SEC_TYPE_ERR) | ||
1033 | return decrypt_out; | ||
1034 | |||
1035 | /* decryption was not done in HW */ | ||
1036 | if ((decrypt_in & RX_MPDU_RES_STATUS_DEC_DONE_MSK) != | ||
1037 | RX_MPDU_RES_STATUS_DEC_DONE_MSK) | ||
1038 | return decrypt_out; | ||
1039 | |||
1040 | switch (decrypt_in & RX_RES_STATUS_SEC_TYPE_MSK) { | ||
1041 | |||
1042 | case RX_RES_STATUS_SEC_TYPE_CCMP: | ||
1043 | /* alg is CCM: check MIC only */ | ||
1044 | if (!(decrypt_in & RX_MPDU_RES_STATUS_MIC_OK)) | ||
1045 | /* Bad MIC */ | ||
1046 | decrypt_out |= RX_RES_STATUS_BAD_ICV_MIC; | ||
1047 | else | ||
1048 | decrypt_out |= RX_RES_STATUS_DECRYPT_OK; | ||
1049 | |||
1050 | break; | ||
1051 | |||
1052 | case RX_RES_STATUS_SEC_TYPE_TKIP: | ||
1053 | if (!(decrypt_in & RX_MPDU_RES_STATUS_TTAK_OK)) { | ||
1054 | /* Bad TTAK */ | ||
1055 | decrypt_out |= RX_RES_STATUS_BAD_KEY_TTAK; | ||
1056 | break; | ||
1057 | } | ||
1058 | /* fall through if TTAK OK */ | ||
1059 | default: | ||
1060 | if (!(decrypt_in & RX_MPDU_RES_STATUS_ICV_OK)) | ||
1061 | decrypt_out |= RX_RES_STATUS_BAD_ICV_MIC; | ||
1062 | else | ||
1063 | decrypt_out |= RX_RES_STATUS_DECRYPT_OK; | ||
1064 | break; | ||
1065 | }; | ||
1066 | |||
1067 | IWL_DEBUG_RX(priv, "decrypt_in:0x%x decrypt_out = 0x%x\n", | ||
1068 | decrypt_in, decrypt_out); | ||
1069 | |||
1070 | return decrypt_out; | ||
1071 | } | ||
1072 | |||
1073 | static void iwl_pass_packet_to_mac80211(struct iwl_priv *priv, | ||
1074 | struct ieee80211_hdr *hdr, | ||
1075 | u16 len, | ||
1076 | u32 ampdu_status, | ||
1077 | struct iwl_rx_mem_buffer *rxb, | ||
1078 | struct ieee80211_rx_status *stats) | ||
1079 | { | ||
1080 | struct sk_buff *skb; | ||
1081 | int ret = 0; | ||
1082 | __le16 fc = hdr->frame_control; | ||
1083 | |||
1084 | /* We only process data packets if the interface is open */ | ||
1085 | if (unlikely(!priv->is_open)) { | ||
1086 | IWL_DEBUG_DROP_LIMIT(priv, | ||
1087 | "Dropping packet while interface is not open.\n"); | ||
1088 | return; | ||
1089 | } | ||
1090 | |||
1091 | /* In case of HW accelerated crypto and bad decryption, drop */ | ||
1092 | if (!priv->cfg->mod_params->sw_crypto && | ||
1093 | iwl_set_decrypted_flag(priv, hdr, ampdu_status, stats)) | ||
1094 | return; | ||
1095 | |||
1096 | skb = alloc_skb(IWL_LINK_HDR_MAX * 2, GFP_ATOMIC); | ||
1097 | if (!skb) { | ||
1098 | IWL_ERR(priv, "alloc_skb failed\n"); | ||
1099 | return; | ||
1100 | } | ||
1101 | |||
1102 | skb_reserve(skb, IWL_LINK_HDR_MAX); | ||
1103 | skb_add_rx_frag(skb, 0, rxb->page, (void *)hdr - rxb_addr(rxb), len); | ||
1104 | |||
1105 | /* mac80211 currently doesn't support paged SKB. Convert it to | ||
1106 | * linear SKB for management frame and data frame requires | ||
1107 | * software decryption or software defragementation. */ | ||
1108 | if (ieee80211_is_mgmt(fc) || | ||
1109 | ieee80211_has_protected(fc) || | ||
1110 | ieee80211_has_morefrags(fc) || | ||
1111 | le16_to_cpu(hdr->seq_ctrl) & IEEE80211_SCTL_FRAG || | ||
1112 | (ieee80211_is_data_qos(fc) && | ||
1113 | *ieee80211_get_qos_ctl(hdr) & | ||
1114 | IEEE80211_QOS_CONTROL_A_MSDU_PRESENT)) | ||
1115 | ret = skb_linearize(skb); | ||
1116 | else | ||
1117 | ret = __pskb_pull_tail(skb, min_t(u16, IWL_LINK_HDR_MAX, len)) ? | ||
1118 | 0 : -ENOMEM; | ||
1119 | |||
1120 | if (ret) { | ||
1121 | kfree_skb(skb); | ||
1122 | goto out; | ||
1123 | } | ||
1124 | |||
1125 | /* | ||
1126 | * XXX: We cannot touch the page and its virtual memory (hdr) after | ||
1127 | * here. It might have already been freed by the above skb change. | ||
1128 | */ | ||
1129 | |||
1130 | iwl_update_stats(priv, false, fc, len); | ||
1131 | memcpy(IEEE80211_SKB_RXCB(skb), stats, sizeof(*stats)); | ||
1132 | |||
1133 | ieee80211_rx(priv->hw, skb); | ||
1134 | out: | ||
1135 | priv->alloc_rxb_page--; | ||
1136 | rxb->page = NULL; | ||
1137 | } | ||
1138 | |||
1139 | /* Called for REPLY_RX (legacy ABG frames), or | ||
1140 | * REPLY_RX_MPDU_CMD (HT high-throughput N frames). */ | ||
1141 | void iwl_rx_reply_rx(struct iwl_priv *priv, | ||
1142 | struct iwl_rx_mem_buffer *rxb) | ||
1143 | { | ||
1144 | struct ieee80211_hdr *header; | ||
1145 | struct ieee80211_rx_status rx_status; | ||
1146 | struct iwl_rx_packet *pkt = rxb_addr(rxb); | ||
1147 | struct iwl_rx_phy_res *phy_res; | ||
1148 | __le32 rx_pkt_status; | ||
1149 | struct iwl4965_rx_mpdu_res_start *amsdu; | ||
1150 | u32 len; | ||
1151 | u32 ampdu_status; | ||
1152 | u32 rate_n_flags; | ||
1153 | |||
1154 | /** | ||
1155 | * REPLY_RX and REPLY_RX_MPDU_CMD are handled differently. | ||
1156 | * REPLY_RX: physical layer info is in this buffer | ||
1157 | * REPLY_RX_MPDU_CMD: physical layer info was sent in separate | ||
1158 | * command and cached in priv->last_phy_res | ||
1159 | * | ||
1160 | * Here we set up local variables depending on which command is | ||
1161 | * received. | ||
1162 | */ | ||
1163 | if (pkt->hdr.cmd == REPLY_RX) { | ||
1164 | phy_res = (struct iwl_rx_phy_res *)pkt->u.raw; | ||
1165 | header = (struct ieee80211_hdr *)(pkt->u.raw + sizeof(*phy_res) | ||
1166 | + phy_res->cfg_phy_cnt); | ||
1167 | |||
1168 | len = le16_to_cpu(phy_res->byte_count); | ||
1169 | rx_pkt_status = *(__le32 *)(pkt->u.raw + sizeof(*phy_res) + | ||
1170 | phy_res->cfg_phy_cnt + len); | ||
1171 | ampdu_status = le32_to_cpu(rx_pkt_status); | ||
1172 | } else { | ||
1173 | if (!priv->last_phy_res[0]) { | ||
1174 | IWL_ERR(priv, "MPDU frame without cached PHY data\n"); | ||
1175 | return; | ||
1176 | } | ||
1177 | phy_res = (struct iwl_rx_phy_res *)&priv->last_phy_res[1]; | ||
1178 | amsdu = (struct iwl4965_rx_mpdu_res_start *)pkt->u.raw; | ||
1179 | header = (struct ieee80211_hdr *)(pkt->u.raw + sizeof(*amsdu)); | ||
1180 | len = le16_to_cpu(amsdu->byte_count); | ||
1181 | rx_pkt_status = *(__le32 *)(pkt->u.raw + sizeof(*amsdu) + len); | ||
1182 | ampdu_status = iwl_translate_rx_status(priv, | ||
1183 | le32_to_cpu(rx_pkt_status)); | ||
1184 | } | ||
1185 | |||
1186 | if ((unlikely(phy_res->cfg_phy_cnt > 20))) { | ||
1187 | IWL_DEBUG_DROP(priv, "dsp size out of range [0,20]: %d/n", | ||
1188 | phy_res->cfg_phy_cnt); | ||
1189 | return; | ||
1190 | } | ||
1191 | |||
1192 | if (!(rx_pkt_status & RX_RES_STATUS_NO_CRC32_ERROR) || | ||
1193 | !(rx_pkt_status & RX_RES_STATUS_NO_RXE_OVERFLOW)) { | ||
1194 | IWL_DEBUG_RX(priv, "Bad CRC or FIFO: 0x%08X.\n", | ||
1195 | le32_to_cpu(rx_pkt_status)); | ||
1196 | return; | ||
1197 | } | ||
1198 | |||
1199 | /* This will be used in several places later */ | ||
1200 | rate_n_flags = le32_to_cpu(phy_res->rate_n_flags); | ||
1201 | |||
1202 | /* rx_status carries information about the packet to mac80211 */ | ||
1203 | rx_status.mactime = le64_to_cpu(phy_res->timestamp); | ||
1204 | rx_status.freq = | ||
1205 | ieee80211_channel_to_frequency(le16_to_cpu(phy_res->channel)); | ||
1206 | rx_status.band = (phy_res->phy_flags & RX_RES_PHY_FLAGS_BAND_24_MSK) ? | ||
1207 | IEEE80211_BAND_2GHZ : IEEE80211_BAND_5GHZ; | ||
1208 | rx_status.rate_idx = | ||
1209 | iwl_hwrate_to_mac80211_idx(rate_n_flags, rx_status.band); | ||
1210 | rx_status.flag = 0; | ||
1211 | |||
1212 | /* TSF isn't reliable. In order to allow smooth user experience, | ||
1213 | * this W/A doesn't propagate it to the mac80211 */ | ||
1214 | /*rx_status.flag |= RX_FLAG_TSFT;*/ | ||
1215 | |||
1216 | priv->ucode_beacon_time = le32_to_cpu(phy_res->beacon_time_stamp); | ||
1217 | |||
1218 | /* Find max signal strength (dBm) among 3 antenna/receiver chains */ | ||
1219 | rx_status.signal = iwl_calc_rssi(priv, phy_res); | ||
1220 | |||
1221 | /* Meaningful noise values are available only from beacon statistics, | ||
1222 | * which are gathered only when associated, and indicate noise | ||
1223 | * only for the associated network channel ... | ||
1224 | * Ignore these noise values while scanning (other channels) */ | ||
1225 | if (iwl_is_associated(priv) && | ||
1226 | !test_bit(STATUS_SCANNING, &priv->status)) { | ||
1227 | rx_status.noise = priv->last_rx_noise; | ||
1228 | } else { | ||
1229 | rx_status.noise = IWL_NOISE_MEAS_NOT_AVAILABLE; | ||
1230 | } | ||
1231 | |||
1232 | /* Reset beacon noise level if not associated. */ | ||
1233 | if (!iwl_is_associated(priv)) | ||
1234 | priv->last_rx_noise = IWL_NOISE_MEAS_NOT_AVAILABLE; | ||
1235 | |||
1236 | #ifdef CONFIG_IWLWIFI_DEBUG | ||
1237 | /* Set "1" to report good data frames in groups of 100 */ | ||
1238 | if (unlikely(iwl_get_debug_level(priv) & IWL_DL_RX)) | ||
1239 | iwl_dbg_report_frame(priv, phy_res, len, header, 1); | ||
1240 | #endif | ||
1241 | iwl_dbg_log_rx_data_frame(priv, len, header); | ||
1242 | IWL_DEBUG_STATS_LIMIT(priv, "Rssi %d, noise %d, TSF %llu\n", | ||
1243 | rx_status.signal, rx_status.noise, | ||
1244 | (unsigned long long)rx_status.mactime); | ||
1245 | |||
1246 | /* | ||
1247 | * "antenna number" | ||
1248 | * | ||
1249 | * It seems that the antenna field in the phy flags value | ||
1250 | * is actually a bit field. This is undefined by radiotap, | ||
1251 | * it wants an actual antenna number but I always get "7" | ||
1252 | * for most legacy frames I receive indicating that the | ||
1253 | * same frame was received on all three RX chains. | ||
1254 | * | ||
1255 | * I think this field should be removed in favor of a | ||
1256 | * new 802.11n radiotap field "RX chains" that is defined | ||
1257 | * as a bitmask. | ||
1258 | */ | ||
1259 | rx_status.antenna = | ||
1260 | (le16_to_cpu(phy_res->phy_flags) & RX_RES_PHY_FLAGS_ANTENNA_MSK) | ||
1261 | >> RX_RES_PHY_FLAGS_ANTENNA_POS; | ||
1262 | |||
1263 | /* set the preamble flag if appropriate */ | ||
1264 | if (phy_res->phy_flags & RX_RES_PHY_FLAGS_SHORT_PREAMBLE_MSK) | ||
1265 | rx_status.flag |= RX_FLAG_SHORTPRE; | ||
1266 | |||
1267 | /* Set up the HT phy flags */ | ||
1268 | if (rate_n_flags & RATE_MCS_HT_MSK) | ||
1269 | rx_status.flag |= RX_FLAG_HT; | ||
1270 | if (rate_n_flags & RATE_MCS_HT40_MSK) | ||
1271 | rx_status.flag |= RX_FLAG_40MHZ; | ||
1272 | if (rate_n_flags & RATE_MCS_SGI_MSK) | ||
1273 | rx_status.flag |= RX_FLAG_SHORT_GI; | ||
1274 | |||
1275 | iwl_pass_packet_to_mac80211(priv, header, len, ampdu_status, | ||
1276 | rxb, &rx_status); | ||
1277 | } | ||
1278 | EXPORT_SYMBOL(iwl_rx_reply_rx); | ||
1279 | |||
1280 | /* Cache phy data (Rx signal strength, etc) for HT frame (REPLY_RX_PHY_CMD). | ||
1281 | * This will be used later in iwl_rx_reply_rx() for REPLY_RX_MPDU_CMD. */ | ||
1282 | void iwl_rx_reply_rx_phy(struct iwl_priv *priv, | ||
1283 | struct iwl_rx_mem_buffer *rxb) | ||
1284 | { | ||
1285 | struct iwl_rx_packet *pkt = rxb_addr(rxb); | ||
1286 | priv->last_phy_res[0] = 1; | ||
1287 | memcpy(&priv->last_phy_res[1], &(pkt->u.raw[0]), | ||
1288 | sizeof(struct iwl_rx_phy_res)); | ||
1289 | } | ||
1290 | EXPORT_SYMBOL(iwl_rx_reply_rx_phy); | ||