1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * Qualcomm Technologies HIDMA data structures 4 * 5 * Copyright (c) 2014-2016, The Linux Foundation. All rights reserved. 6 */ 7 8 #ifndef QCOM_HIDMA_H 9 #define QCOM_HIDMA_H 10 11 #include <linux/kfifo.h> 12 #include <linux/interrupt.h> 13 #include <linux/dmaengine.h> 14 15 #define HIDMA_TRE_SIZE 32 /* each TRE is 32 bytes */ 16 #define HIDMA_TRE_CFG_IDX 0 17 #define HIDMA_TRE_LEN_IDX 1 18 #define HIDMA_TRE_SRC_LOW_IDX 2 19 #define HIDMA_TRE_SRC_HI_IDX 3 20 #define HIDMA_TRE_DEST_LOW_IDX 4 21 #define HIDMA_TRE_DEST_HI_IDX 5 22 23 enum tre_type { 24 HIDMA_TRE_MEMCPY = 3, 25 HIDMA_TRE_MEMSET = 4, 26 }; 27 28 struct hidma_tre { 29 atomic_t allocated; /* if this channel is allocated */ 30 bool queued; /* flag whether this is pending */ 31 u16 status; /* status */ 32 u32 idx; /* index of the tre */ 33 u32 dma_sig; /* signature of the tre */ 34 const char *dev_name; /* name of the device */ 35 void (*callback)(void *data); /* requester callback */ 36 void *data; /* Data associated with this channel*/ 37 struct hidma_lldev *lldev; /* lldma device pointer */ 38 u32 tre_local[HIDMA_TRE_SIZE / sizeof(u32) + 1]; /* TRE local copy */ 39 u32 tre_index; /* the offset where this was written*/ 40 u32 int_flags; /* interrupt flags */ 41 u8 err_info; /* error record in this transfer */ 42 u8 err_code; /* completion code */ 43 }; 44 45 struct hidma_lldev { 46 bool msi_support; /* flag indicating MSI support */ 47 bool initialized; /* initialized flag */ 48 u8 trch_state; /* trch_state of the device */ 49 u8 evch_state; /* evch_state of the device */ 50 u8 chidx; /* channel index in the core */ 51 u32 nr_tres; /* max number of configs */ 52 spinlock_t lock; /* reentrancy */ 53 struct hidma_tre *trepool; /* trepool of user configs */ 54 struct device *dev; /* device */ 55 void __iomem *trca; /* Transfer Channel address */ 56 void __iomem *evca; /* Event Channel address */ 57 struct hidma_tre 58 **pending_tre_list; /* Pointers to pending TREs */ 59 atomic_t pending_tre_count; /* Number of TREs pending */ 60 61 void *tre_ring; /* TRE ring */ 62 dma_addr_t tre_dma; /* TRE ring to be shared with HW */ 63 u32 tre_ring_size; /* Byte size of the ring */ 64 u32 tre_processed_off; /* last processed TRE */ 65 66 void *evre_ring; /* EVRE ring */ 67 dma_addr_t evre_dma; /* EVRE ring to be shared with HW */ 68 u32 evre_ring_size; /* Byte size of the ring */ 69 u32 evre_processed_off; /* last processed EVRE */ 70 71 u32 tre_write_offset; /* TRE write location */ 72 struct tasklet_struct task; /* task delivering notifications */ 73 DECLARE_KFIFO_PTR(handoff_fifo, 74 struct hidma_tre *); /* pending TREs FIFO */ 75 }; 76 77 struct hidma_desc { 78 struct dma_async_tx_descriptor desc; 79 /* link list node for this channel*/ 80 struct list_head node; 81 u32 tre_ch; 82 }; 83 84 struct hidma_chan { 85 bool paused; 86 bool allocated; 87 char dbg_name[16]; 88 u32 dma_sig; 89 dma_cookie_t last_success; 90 91 /* 92 * active descriptor on this channel 93 * It is used by the DMA complete notification to 94 * locate the descriptor that initiated the transfer. 95 */ 96 struct hidma_dev *dmadev; 97 struct hidma_desc *running; 98 99 struct dma_chan chan; 100 struct list_head free; 101 struct list_head prepared; 102 struct list_head queued; 103 struct list_head active; 104 struct list_head completed; 105 106 /* Lock for this structure */ 107 spinlock_t lock; 108 }; 109 110 struct hidma_dev { 111 int irq; 112 int chidx; 113 u32 nr_descriptors; 114 int msi_virqbase; 115 116 struct hidma_lldev *lldev; 117 void __iomem *dev_trca; 118 struct resource *trca_resource; 119 void __iomem *dev_evca; 120 struct resource *evca_resource; 121 122 /* used to protect the pending channel list*/ 123 spinlock_t lock; 124 struct dma_device ddev; 125 126 struct dentry *debugfs; 127 128 /* sysfs entry for the channel id */ 129 struct device_attribute *chid_attrs; 130 131 /* Task delivering issue_pending */ 132 struct tasklet_struct task; 133 }; 134 135 int hidma_ll_request(struct hidma_lldev *llhndl, u32 dev_id, 136 const char *dev_name, 137 void (*callback)(void *data), void *data, u32 *tre_ch); 138 139 void hidma_ll_free(struct hidma_lldev *llhndl, u32 tre_ch); 140 enum dma_status hidma_ll_status(struct hidma_lldev *llhndl, u32 tre_ch); 141 bool hidma_ll_isenabled(struct hidma_lldev *llhndl); 142 void hidma_ll_queue_request(struct hidma_lldev *llhndl, u32 tre_ch); 143 void hidma_ll_start(struct hidma_lldev *llhndl); 144 int hidma_ll_disable(struct hidma_lldev *lldev); 145 int hidma_ll_enable(struct hidma_lldev *llhndl); 146 void hidma_ll_set_transfer_params(struct hidma_lldev *llhndl, u32 tre_ch, 147 dma_addr_t src, dma_addr_t dest, u32 len, u32 flags, u32 txntype); 148 void hidma_ll_setup_irq(struct hidma_lldev *lldev, bool msi); 149 int hidma_ll_setup(struct hidma_lldev *lldev); 150 struct hidma_lldev *hidma_ll_init(struct device *dev, u32 max_channels, 151 void __iomem *trca, void __iomem *evca, 152 u8 chidx); 153 int hidma_ll_uninit(struct hidma_lldev *llhndl); 154 irqreturn_t hidma_ll_inthandler(int irq, void *arg); 155 irqreturn_t hidma_ll_inthandler_msi(int irq, void *arg, int cause); 156 void hidma_cleanup_pending_tre(struct hidma_lldev *llhndl, u8 err_info, 157 u8 err_code); 158 void hidma_debug_init(struct hidma_dev *dmadev); 159 void hidma_debug_uninit(struct hidma_dev *dmadev); 160 #endif 161