Lines Matching refs:rx_small

175 	struct myri10ge_rx_buf rx_small;  member
1056 ss->rx_small.cnt = 0; in myri10ge_reset()
1319 rx = &ss->rx_small; in myri10ge_rx_done()
1476 if (ss->rx_small.fill_cnt - ss->rx_small.cnt < myri10ge_fill_thresh) in myri10ge_clean_rx_done()
1477 myri10ge_alloc_rx_pages(mgp, &ss->rx_small, in myri10ge_clean_rx_done()
1713 ring->rx_mini_max_pending = mgp->ss[0].rx_small.mask + 1; in myri10ge_get_ringparam()
1844 data[i++] = (unsigned int)ss->rx_small.cnt; in myri10ge_get_ethtool_stats()
1960 ss->rx_small.mask = ss->rx_big.mask = rx_ring_entries - 1; in myri10ge_allocate_rings()
1977 bytes = rx_ring_entries * sizeof(*ss->rx_small.shadow); in myri10ge_allocate_rings()
1978 ss->rx_small.shadow = kzalloc(bytes, GFP_KERNEL); in myri10ge_allocate_rings()
1979 if (ss->rx_small.shadow == NULL) in myri10ge_allocate_rings()
1994 bytes = rx_ring_entries * sizeof(*ss->rx_small.info); in myri10ge_allocate_rings()
1995 ss->rx_small.info = kzalloc(bytes, GFP_KERNEL); in myri10ge_allocate_rings()
1996 if (ss->rx_small.info == NULL) in myri10ge_allocate_rings()
2006 ss->rx_small.cnt = 0; in myri10ge_allocate_rings()
2008 ss->rx_small.fill_cnt = 0; in myri10ge_allocate_rings()
2009 ss->rx_small.page_offset = MYRI10GE_ALLOC_SIZE; in myri10ge_allocate_rings()
2011 ss->rx_small.watchdog_needed = 0; in myri10ge_allocate_rings()
2014 ss->rx_small.fill_cnt = ss->rx_small.mask + 1; in myri10ge_allocate_rings()
2016 myri10ge_alloc_rx_pages(mgp, &ss->rx_small, in myri10ge_allocate_rings()
2020 if (ss->rx_small.fill_cnt < ss->rx_small.mask + 1) { in myri10ge_allocate_rings()
2022 slice, ss->rx_small.fill_cnt); in myri10ge_allocate_rings()
2045 ss->rx_small.fill_cnt = ss->rx_small.cnt; in myri10ge_allocate_rings()
2046 for (i = ss->rx_small.cnt; i < ss->rx_small.fill_cnt; i++) { in myri10ge_allocate_rings()
2047 int idx = i & ss->rx_small.mask; in myri10ge_allocate_rings()
2048 myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_small.info[idx], in myri10ge_allocate_rings()
2050 put_page(ss->rx_small.info[idx].page); in myri10ge_allocate_rings()
2056 kfree(ss->rx_small.info); in myri10ge_allocate_rings()
2065 kfree(ss->rx_small.shadow); in myri10ge_allocate_rings()
2097 ss->rx_small.fill_cnt = ss->rx_small.cnt; in myri10ge_free_rings()
2098 for (i = ss->rx_small.cnt; i < ss->rx_small.fill_cnt; i++) { in myri10ge_free_rings()
2099 idx = i & ss->rx_small.mask; in myri10ge_free_rings()
2100 if (i == ss->rx_small.fill_cnt - 1) in myri10ge_free_rings()
2101 ss->rx_small.info[idx].page_offset = in myri10ge_free_rings()
2103 myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_small.info[idx], in myri10ge_free_rings()
2105 put_page(ss->rx_small.info[idx].page); in myri10ge_free_rings()
2135 kfree(ss->rx_small.info); in myri10ge_free_rings()
2141 kfree(ss->rx_small.shadow); in myri10ge_free_rings()
2249 ss->rx_small.lanai = (struct mcp_kreq_ether_recv __iomem *) in myri10ge_get_txrx()
3489 if (ss->rx_small.watchdog_needed) { in myri10ge_watchdog_timer()
3490 myri10ge_alloc_rx_pages(mgp, &ss->rx_small, in myri10ge_watchdog_timer()
3493 if (ss->rx_small.fill_cnt - ss->rx_small.cnt >= in myri10ge_watchdog_timer()
3495 ss->rx_small.watchdog_needed = 0; in myri10ge_watchdog_timer()