Searched refs:process_queue (Results 1 – 5 of 5) sorted by relevance
40 self.process_queue()44 def process_queue(self): member in MtreeCommand
229 struct sk_buff_head process_queue; in tun_napi_receive() local233 __skb_queue_head_init(&process_queue); in tun_napi_receive()236 skb_queue_splice_tail_init(queue, &process_queue); in tun_napi_receive()239 while (received < budget && (skb = __skb_dequeue(&process_queue))) { in tun_napi_receive()244 if (!skb_queue_empty(&process_queue)) { in tun_napi_receive()246 skb_queue_splice(&process_queue, queue); in tun_napi_receive()1549 struct sk_buff_head process_queue; in tun_rx_batched() local1563 __skb_queue_head_init(&process_queue); in tun_rx_batched()1564 skb_queue_splice_tail_init(queue, &process_queue); in tun_rx_batched()1575 while ((nskb = __skb_dequeue(&process_queue))) { in tun_rx_batched()
125 return skb_queue_len_lockless(&sd->process_queue); in softnet_process_queue_len()
5908 skb_queue_walk_safe(&sd->process_queue, skb, tmp) { in flush_backlog() 5910 __skb_unlink(skb, &sd->process_queue); in flush_backlog() 5926 /* as insertion into process_queue happens with the rps lock held, in flush_required() 5927 * process_queue access may race only with dequeue in flush_required() 5930 !skb_queue_empty_lockless(&sd->process_queue); in flush_required() 5937 * input_pkt_queue and process_queue even if the latter could end-up in flush_required() 6036 while ((skb = __skb_dequeue(&sd->process_queue))) { in process_backlog() 6060 &sd->process_queue); in process_backlog() 11338 * We properly handle process_queue & input_pkt_queue later. in dev_cpu_dead() 11363 while ((skb = __skb_dequeue(&oldsd->process_queue))) { in dev_cpu_dead() [all...]
3218 struct sk_buff_head process_queue; member