Lines Matching refs:rpm_msg
81 struct rpmh_request *rpm_msg = container_of(msg, struct rpmh_request, in rpmh_tx_done() local
83 struct completion *compl = rpm_msg->completion; in rpmh_tx_done()
84 bool free = rpm_msg->needs_free; in rpmh_tx_done()
94 kfree(rpm_msg); in rpmh_tx_done()
171 struct rpmh_request *rpm_msg) in __rpmh_write() argument
179 for (i = 0; i < rpm_msg->msg.num_cmds; i++) { in __rpmh_write()
180 req = cache_rpm_request(ctrlr, state, &rpm_msg->msg.cmds[i]); in __rpmh_write()
186 ret = rpmh_rsc_send_data(ctrlr_to_drv(ctrlr), &rpm_msg->msg); in __rpmh_write()
190 rpmh_tx_done(&rpm_msg->msg); in __rpmh_write()
225 struct rpmh_request *rpm_msg; in rpmh_write_async() local
228 rpm_msg = kzalloc(sizeof(*rpm_msg), GFP_ATOMIC); in rpmh_write_async()
229 if (!rpm_msg) in rpmh_write_async()
231 rpm_msg->needs_free = true; in rpmh_write_async()
233 ret = __fill_rpmh_msg(rpm_msg, state, cmd, n); in rpmh_write_async()
235 kfree(rpm_msg); in rpmh_write_async()
239 return __rpmh_write(dev, state, rpm_msg); in rpmh_write_async()
257 DEFINE_RPMH_MSG_ONSTACK(dev, state, &compl, rpm_msg); in rpmh_write()
260 ret = __fill_rpmh_msg(&rpm_msg, state, cmd, n); in rpmh_write()
264 ret = __rpmh_write(dev, state, &rpm_msg); in rpmh_write()
287 const struct rpmh_request *rpm_msg; in flush_batch() local
294 rpm_msg = req->rpm_msgs + i; in flush_batch()
296 &rpm_msg->msg); in flush_batch()
409 DEFINE_RPMH_MSG_ONSTACK(NULL, state, NULL, rpm_msg); in send_single()
412 rpm_msg.msg.wait_for_compl = (state == RPMH_WAKE_ONLY_STATE); in send_single()
413 rpm_msg.cmd[0].addr = addr; in send_single()
414 rpm_msg.cmd[0].data = data; in send_single()
415 rpm_msg.msg.num_cmds = 1; in send_single()
417 return rpmh_rsc_write_ctrl_data(ctrlr_to_drv(ctrlr), &rpm_msg.msg); in send_single()