Lines Matching refs:ctx_data

381 	unsigned int state = s->ctx_data.rx.data_block_state;  in pool_ideal_nonblocking_data_blocks()
415 s->ctx_data.rx.data_block_state = state; in pool_ideal_nonblocking_data_blocks()
461 unsigned int last = s->ctx_data.rx.last_syt_offset; in pool_ideal_syt_offsets()
462 unsigned int state = s->ctx_data.rx.syt_offset_state; in pool_ideal_syt_offsets()
473 s->ctx_data.rx.last_syt_offset = last; in pool_ideal_syt_offsets()
474 s->ctx_data.rx.syt_offset_state = state; in pool_ideal_syt_offsets()
504 const unsigned int cache_size = s->ctx_data.tx.cache.size; in calculate_cached_cycle_count()
505 unsigned int cycles = s->ctx_data.tx.cache.pos; in calculate_cached_cycle_count()
517 const unsigned int cache_size = s->ctx_data.tx.cache.size; in cache_seq()
518 struct seq_desc *cache = s->ctx_data.tx.cache.descs; in cache_seq()
519 unsigned int cache_pos = s->ctx_data.tx.cache.pos; in cache_seq()
536 s->ctx_data.tx.cache.pos = cache_pos; in cache_seq()
553 struct amdtp_stream *target = s->ctx_data.rx.replay_target; in pool_replayed_seq()
554 const struct seq_desc *cache = target->ctx_data.tx.cache.descs; in pool_replayed_seq()
555 const unsigned int cache_size = target->ctx_data.tx.cache.size; in pool_replayed_seq()
556 unsigned int cache_pos = s->ctx_data.rx.cache_pos; in pool_replayed_seq()
565 s->ctx_data.rx.cache_pos = cache_pos; in pool_replayed_seq()
575 if (!d->replay.enable || !s->ctx_data.rx.replay_target) { in pool_seq_descs()
581 struct amdtp_stream *tx = s->ctx_data.rx.replay_target; in pool_seq_descs()
582 const unsigned int cache_size = tx->ctx_data.tx.cache.size; in pool_seq_descs()
583 const unsigned int cache_pos = s->ctx_data.rx.cache_pos; in pool_seq_descs()
661 params->header_length = s->ctx_data.tx.ctx_header_size; in queue_in_packet()
662 params->payload_length = s->ctx_data.tx.max_ctx_payload_length; in queue_in_packet()
676 ((s->ctx_data.rx.fdf << CIP_FDF_SHIFT) & CIP_FDF_MASK) | in generate_cip_header()
777 if (*data_blocks > 0 && s->ctx_data.tx.dbc_interval > 0) in check_cip_header()
778 dbc_interval = s->ctx_data.tx.dbc_interval; in check_cip_header()
821 if (payload_length > cip_header_size + s->ctx_data.tx.max_ctx_payload_length) { in parse_ir_ctx_header()
824 payload_length, cip_header_size + s->ctx_data.tx.max_ctx_payload_length); in parse_ir_ctx_header()
984 ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header); in generate_tx_packet_descs()
1008 struct seq_desc *seq_descs = s->ctx_data.rx.seq.descs; in generate_rx_packet_descs()
1009 unsigned int seq_size = s->ctx_data.rx.seq.size; in generate_rx_packet_descs()
1010 unsigned int seq_pos = s->ctx_data.rx.seq.pos; in generate_rx_packet_descs()
1047 s->ctx_data.rx.seq.pos = seq_pos; in generate_rx_packet_descs()
1146 unsigned int event_count = s->ctx_data.rx.event_count; in process_rx_packets()
1209 s->ctx_data.rx.event_count = event_count; in process_rx_packets()
1289 s->ctx_data.rx.cache_pos = 0; in process_rx_packets_intermediately()
1317 packet_count = header_length / s->ctx_data.tx.ctx_header_size; in process_tx_packets()
1361 packets = header_length / s->ctx_data.tx.ctx_header_size; in drop_tx_packets()
1363 ctx_header += (packets - 1) * s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header); in drop_tx_packets()
1389 packets = header_length / s->ctx_data.tx.ctx_header_size; in process_tx_packets_intermediately()
1399 ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32); in process_tx_packets_intermediately()
1406 size_t length = s->ctx_data.tx.ctx_header_size * offset; in process_tx_packets_intermediately()
1441 count = header_length / s->ctx_data.tx.ctx_header_size; in drop_tx_packets_initially()
1478 ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32); in drop_tx_packets_initially()
1484 s->ctx_data.tx.event_starts = true; in drop_tx_packets_initially()
1495 if (s->ctx_data.tx.event_starts) in drop_tx_packets_initially()
1587 tx = rx->ctx_data.rx.replay_target; in irq_target_callback_skip()
1589 if (cached_cycles > tx->ctx_data.tx.cache.size / 2) in irq_target_callback_skip()
1716 s->ctx_data.tx.max_ctx_payload_length = max_ctx_payload_size; in amdtp_stream_start()
1717 s->ctx_data.tx.ctx_header_size = ctx_header_size; in amdtp_stream_start()
1718 s->ctx_data.tx.event_starts = false; in amdtp_stream_start()
1723 s->ctx_data.tx.cache.size = max_t(unsigned int, s->syt_interval * 2, in amdtp_stream_start()
1725 s->ctx_data.tx.cache.pos = 0; in amdtp_stream_start()
1726 s->ctx_data.tx.cache.descs = kcalloc(s->ctx_data.tx.cache.size, in amdtp_stream_start()
1727 sizeof(*s->ctx_data.tx.cache.descs), GFP_KERNEL); in amdtp_stream_start()
1728 if (!s->ctx_data.tx.cache.descs) { in amdtp_stream_start()
1747 s->ctx_data.rx.seq.descs = kcalloc(queue_size, sizeof(*s->ctx_data.rx.seq.descs), GFP_KERNEL); in amdtp_stream_start()
1748 if (!s->ctx_data.rx.seq.descs) { in amdtp_stream_start()
1752 s->ctx_data.rx.seq.size = queue_size; in amdtp_stream_start()
1753 s->ctx_data.rx.seq.pos = 0; in amdtp_stream_start()
1756 s->ctx_data.rx.data_block_state = entry->data_block; in amdtp_stream_start()
1757 s->ctx_data.rx.syt_offset_state = entry->syt_offset; in amdtp_stream_start()
1758 s->ctx_data.rx.last_syt_offset = TICKS_PER_CYCLE; in amdtp_stream_start()
1760 s->ctx_data.rx.event_count = 0; in amdtp_stream_start()
1828 kfree(s->ctx_data.rx.seq.descs); in amdtp_stream_start()
1831 kfree(s->ctx_data.tx.cache.descs); in amdtp_stream_start()
1923 kfree(s->ctx_data.rx.seq.descs); in amdtp_stream_stop()
1926 kfree(s->ctx_data.tx.cache.descs); in amdtp_stream_stop()
2038 rx->ctx_data.rx.replay_target = tx; in make_association()