[go: nahoru, domu]

Lines Matching defs:unmap_q

112 /* Caller should ensure that the entry at unmap_q[index] is valid */
115 struct bnad_tx_unmap *unmap_q,
122 unmap = &unmap_q[index];
140 unmap = &unmap_q[index];
164 struct bnad_tx_unmap *unmap_q = tcb->unmap_q;
169 skb = unmap_q[i].skb;
172 bnad_tx_buff_unmap(bnad, unmap_q, tcb->q_depth, i);
188 struct bnad_tx_unmap *unmap_q = tcb->unmap_q;
204 unmap = &unmap_q[cons];
214 cons = bnad_tx_buff_unmap(bnad, unmap_q, q_depth, cons);
273 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
275 unmap_q->reuse_pi = -1;
276 unmap_q->alloc_order = -1;
277 unmap_q->map_size = 0;
278 unmap_q->type = BNAD_RXBUF_NONE;
285 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
292 unmap_q->type = BNAD_RXBUF_PAGE;
295 unmap_q->alloc_order = 0;
296 unmap_q->map_size = rcb->rxq->buffer_size;
299 unmap_q->alloc_order = 0;
300 unmap_q->map_size = rcb->rxq->buffer_size;
301 unmap_q->type = BNAD_RXBUF_MULTI_BUFF;
303 unmap_q->alloc_order = order;
304 unmap_q->map_size =
310 BUG_ON(((PAGE_SIZE << order) % unmap_q->map_size));
348 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
352 struct bnad_rx_unmap *unmap = &unmap_q->unmap[i];
354 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type))
366 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
376 alloc_size = PAGE_SIZE << unmap_q->alloc_order;
380 unmap = &unmap_q->unmap[prod];
382 if (unmap_q->reuse_pi < 0) {
384 unmap_q->alloc_order);
387 prev = &unmap_q->unmap[unmap_q->reuse_pi];
389 page_offset = prev->page_offset + unmap_q->map_size;
400 unmap_q->map_size, DMA_FROM_DEVICE);
405 unmap->vector.len = unmap_q->map_size;
406 page_offset += unmap_q->map_size;
409 unmap_q->reuse_pi = prod;
411 unmap_q->reuse_pi = -1;
434 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
446 unmap = &unmap_q->unmap[prod];
482 struct bnad_rx_unmap_q *unmap_q = rcb->unmap_q;
489 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type))
513 struct bnad_rx_unmap_q *unmap_q;
517 unmap_q = rcb->unmap_q;
519 unmap = &unmap_q->unmap[ci];
522 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type))
535 struct bnad_rx_unmap_q *unmap_q;
538 unmap_q = rcb->unmap_q;
542 prefetch(page_address(unmap_q->unmap[sop_ci].page) +
543 unmap_q->unmap[sop_ci].page_offset);
546 unmap = &unmap_q->unmap[ci];
591 struct bnad_rx_unmap_q *unmap_q;
624 unmap_q = rcb->unmap_q;
629 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type)) {
630 unmap = &unmap_q->unmap[sop_ci];
647 if (BNAD_RXBUF_IS_MULTI_BUFF(unmap_q->type) &&
688 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type))
714 if (BNAD_RXBUF_IS_SK_BUFF(unmap_q->type))
2936 struct bnad_tx_unmap *unmap_q, *unmap, *head_unmap;
2977 unmap_q = tcb->unmap_q;
3019 head_unmap = &unmap_q[prod];
3047 bnad_tx_buff_unmap(bnad, unmap_q, q_depth,
3062 unmap = &unmap_q[prod];
3077 bnad_tx_buff_unmap(bnad, unmap_q, q_depth, tcb->producer_index);