Lines Matching refs:cl

341 	cb->cl->tx_cb_queued++;  in mei_tx_cb_enqueue()
353 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue()
354 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue()
367 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument
372 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp()
389 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument
401 cb->cl = cl; in mei_io_cb_init()
417 const struct mei_cl *cl) in mei_io_list_flush_cl() argument
422 if (cl == cb->cl) { in mei_io_list_flush_cl()
438 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument
444 if (cl == cb->cl && (!fp || fp == cb->fp)) in mei_io_tx_list_free_cl()
469 static void mei_cl_free_pending(struct mei_cl *cl) in mei_cl_free_pending() argument
473 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_free_pending()
487 struct mei_cl_cb *mei_cl_alloc_cb(struct mei_cl *cl, size_t length, in mei_cl_alloc_cb() argument
493 cb = mei_io_cb_init(cl, fop_type, fp); in mei_cl_alloc_cb()
522 struct mei_cl_cb *mei_cl_enqueue_ctrl_wr_cb(struct mei_cl *cl, size_t length, in mei_cl_enqueue_ctrl_wr_cb() argument
530 length = max_t(size_t, length, mei_cl_mtu(cl)); in mei_cl_enqueue_ctrl_wr_cb()
532 cb = mei_cl_alloc_cb(cl, length, fop_type, fp); in mei_cl_enqueue_ctrl_wr_cb()
536 list_add_tail(&cb->list, &cl->dev->ctrl_wr_list); in mei_cl_enqueue_ctrl_wr_cb()
549 struct mei_cl_cb *mei_cl_read_cb(struct mei_cl *cl, const struct file *fp) in mei_cl_read_cb() argument
554 spin_lock(&cl->rd_completed_lock); in mei_cl_read_cb()
555 list_for_each_entry(cb, &cl->rd_completed, list) in mei_cl_read_cb()
560 spin_unlock(&cl->rd_completed_lock); in mei_cl_read_cb()
572 int mei_cl_flush_queues(struct mei_cl *cl, const struct file *fp) in mei_cl_flush_queues() argument
576 if (WARN_ON(!cl || !cl->dev)) in mei_cl_flush_queues()
579 dev = cl->dev; in mei_cl_flush_queues()
581 cl_dbg(dev, cl, "remove list entry belonging to cl\n"); in mei_cl_flush_queues()
582 mei_io_tx_list_free_cl(&cl->dev->write_list, cl, fp); in mei_cl_flush_queues()
583 mei_io_tx_list_free_cl(&cl->dev->write_waiting_list, cl, fp); in mei_cl_flush_queues()
586 mei_io_list_flush_cl(&cl->dev->ctrl_wr_list, cl); in mei_cl_flush_queues()
587 mei_io_list_flush_cl(&cl->dev->ctrl_rd_list, cl); in mei_cl_flush_queues()
588 mei_cl_free_pending(cl); in mei_cl_flush_queues()
590 spin_lock(&cl->rd_completed_lock); in mei_cl_flush_queues()
591 mei_io_list_free_fp(&cl->rd_completed, fp); in mei_cl_flush_queues()
592 spin_unlock(&cl->rd_completed_lock); in mei_cl_flush_queues()
603 static void mei_cl_init(struct mei_cl *cl, struct mei_device *dev) in mei_cl_init() argument
605 memset(cl, 0, sizeof(*cl)); in mei_cl_init()
606 init_waitqueue_head(&cl->wait); in mei_cl_init()
607 init_waitqueue_head(&cl->rx_wait); in mei_cl_init()
608 init_waitqueue_head(&cl->tx_wait); in mei_cl_init()
609 init_waitqueue_head(&cl->ev_wait); in mei_cl_init()
610 INIT_LIST_HEAD(&cl->vtag_map); in mei_cl_init()
611 spin_lock_init(&cl->rd_completed_lock); in mei_cl_init()
612 INIT_LIST_HEAD(&cl->rd_completed); in mei_cl_init()
613 INIT_LIST_HEAD(&cl->rd_pending); in mei_cl_init()
614 INIT_LIST_HEAD(&cl->link); in mei_cl_init()
615 cl->writing_state = MEI_IDLE; in mei_cl_init()
616 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_init()
617 cl->dev = dev; in mei_cl_init()
628 struct mei_cl *cl; in mei_cl_allocate() local
630 cl = kmalloc(sizeof(*cl), GFP_KERNEL); in mei_cl_allocate()
631 if (!cl) in mei_cl_allocate()
634 mei_cl_init(cl, dev); in mei_cl_allocate()
636 return cl; in mei_cl_allocate()
648 int mei_cl_link(struct mei_cl *cl) in mei_cl_link() argument
653 if (WARN_ON(!cl || !cl->dev)) in mei_cl_link()
656 dev = cl->dev; in mei_cl_link()
672 cl->host_client_id = id; in mei_cl_link()
673 list_add_tail(&cl->link, &dev->file_list); in mei_cl_link()
677 cl->state = MEI_FILE_INITIALIZING; in mei_cl_link()
679 cl_dbg(dev, cl, "link cl\n"); in mei_cl_link()
690 int mei_cl_unlink(struct mei_cl *cl) in mei_cl_unlink() argument
695 if (!cl) in mei_cl_unlink()
698 if (WARN_ON(!cl->dev)) in mei_cl_unlink()
701 dev = cl->dev; in mei_cl_unlink()
703 cl_dbg(dev, cl, "unlink client"); in mei_cl_unlink()
705 if (cl->state == MEI_FILE_UNINITIALIZED) in mei_cl_unlink()
712 if (cl->host_client_id) in mei_cl_unlink()
713 clear_bit(cl->host_client_id, dev->host_clients_map); in mei_cl_unlink()
715 list_del_init(&cl->link); in mei_cl_unlink()
717 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_unlink()
718 cl->writing_state = MEI_IDLE; in mei_cl_unlink()
720 WARN_ON(!list_empty(&cl->rd_completed) || in mei_cl_unlink()
721 !list_empty(&cl->rd_pending) || in mei_cl_unlink()
722 !list_empty(&cl->link)); in mei_cl_unlink()
769 static void mei_cl_wake_all(struct mei_cl *cl) in mei_cl_wake_all() argument
771 struct mei_device *dev = cl->dev; in mei_cl_wake_all()
774 if (waitqueue_active(&cl->rx_wait)) { in mei_cl_wake_all()
775 cl_dbg(dev, cl, "Waking up reading client!\n"); in mei_cl_wake_all()
776 wake_up_interruptible(&cl->rx_wait); in mei_cl_wake_all()
779 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_wake_all()
780 cl_dbg(dev, cl, "Waking up writing client!\n"); in mei_cl_wake_all()
781 wake_up_interruptible(&cl->tx_wait); in mei_cl_wake_all()
784 if (waitqueue_active(&cl->ev_wait)) { in mei_cl_wake_all()
785 cl_dbg(dev, cl, "Waking up waiting for event clients!\n"); in mei_cl_wake_all()
786 wake_up_interruptible(&cl->ev_wait); in mei_cl_wake_all()
789 if (waitqueue_active(&cl->wait)) { in mei_cl_wake_all()
790 cl_dbg(dev, cl, "Waking up ctrl write clients!\n"); in mei_cl_wake_all()
791 wake_up(&cl->wait); in mei_cl_wake_all()
801 static void mei_cl_set_disconnected(struct mei_cl *cl) in mei_cl_set_disconnected() argument
803 struct mei_device *dev = cl->dev; in mei_cl_set_disconnected()
805 if (cl->state == MEI_FILE_DISCONNECTED || in mei_cl_set_disconnected()
806 cl->state <= MEI_FILE_INITIALIZING) in mei_cl_set_disconnected()
809 cl->state = MEI_FILE_DISCONNECTED; in mei_cl_set_disconnected()
810 mei_io_tx_list_free_cl(&dev->write_list, cl, NULL); in mei_cl_set_disconnected()
811 mei_io_tx_list_free_cl(&dev->write_waiting_list, cl, NULL); in mei_cl_set_disconnected()
812 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_set_disconnected()
813 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_set_disconnected()
814 mei_cl_wake_all(cl); in mei_cl_set_disconnected()
815 cl->rx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
816 cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
817 cl->timer_count = 0; in mei_cl_set_disconnected()
819 if (!cl->me_cl) in mei_cl_set_disconnected()
822 if (!WARN_ON(cl->me_cl->connect_count == 0)) in mei_cl_set_disconnected()
823 cl->me_cl->connect_count--; in mei_cl_set_disconnected()
825 if (cl->me_cl->connect_count == 0) in mei_cl_set_disconnected()
826 cl->me_cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
828 mei_me_cl_put(cl->me_cl); in mei_cl_set_disconnected()
829 cl->me_cl = NULL; in mei_cl_set_disconnected()
832 static int mei_cl_set_connecting(struct mei_cl *cl, struct mei_me_client *me_cl) in mei_cl_set_connecting() argument
845 cl->me_cl = me_cl; in mei_cl_set_connecting()
846 cl->state = MEI_FILE_CONNECTING; in mei_cl_set_connecting()
847 cl->me_cl->connect_count++; in mei_cl_set_connecting()
860 static int mei_cl_send_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_disconnect() argument
865 dev = cl->dev; in mei_cl_send_disconnect()
867 ret = mei_hbm_cl_disconnect_req(dev, cl); in mei_cl_send_disconnect()
868 cl->status = ret; in mei_cl_send_disconnect()
870 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_disconnect()
875 cl->timer_count = dev->timeouts.connect; in mei_cl_send_disconnect()
891 int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_disconnect() argument
894 struct mei_device *dev = cl->dev; in mei_cl_irq_disconnect()
907 ret = mei_cl_send_disconnect(cl, cb); in mei_cl_irq_disconnect()
922 static int __mei_cl_disconnect(struct mei_cl *cl) in __mei_cl_disconnect() argument
928 dev = cl->dev; in __mei_cl_disconnect()
930 cl->state = MEI_FILE_DISCONNECTING; in __mei_cl_disconnect()
932 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT, NULL); in __mei_cl_disconnect()
939 rets = mei_cl_send_disconnect(cl, cb); in __mei_cl_disconnect()
941 cl_err(dev, cl, "failed to disconnect.\n"); in __mei_cl_disconnect()
947 wait_event_timeout(cl->wait, in __mei_cl_disconnect()
948 cl->state == MEI_FILE_DISCONNECT_REPLY || in __mei_cl_disconnect()
949 cl->state == MEI_FILE_DISCONNECTED, in __mei_cl_disconnect()
953 rets = cl->status; in __mei_cl_disconnect()
954 if (cl->state != MEI_FILE_DISCONNECT_REPLY && in __mei_cl_disconnect()
955 cl->state != MEI_FILE_DISCONNECTED) { in __mei_cl_disconnect()
956 cl_dbg(dev, cl, "timeout on disconnect from FW client.\n"); in __mei_cl_disconnect()
962 mei_cl_set_disconnected(cl); in __mei_cl_disconnect()
964 cl_dbg(dev, cl, "successfully disconnected from FW client.\n"); in __mei_cl_disconnect()
979 int mei_cl_disconnect(struct mei_cl *cl) in mei_cl_disconnect() argument
984 if (WARN_ON(!cl || !cl->dev)) in mei_cl_disconnect()
987 dev = cl->dev; in mei_cl_disconnect()
989 cl_dbg(dev, cl, "disconnecting"); in mei_cl_disconnect()
991 if (!mei_cl_is_connected(cl)) in mei_cl_disconnect()
994 if (mei_cl_is_fixed_address(cl)) { in mei_cl_disconnect()
995 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
1001 cl_dbg(dev, cl, "Device is powering down, don't bother with disconnection\n"); in mei_cl_disconnect()
1002 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
1009 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_disconnect()
1013 rets = __mei_cl_disconnect(cl); in mei_cl_disconnect()
1015 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_disconnect()
1031 static bool mei_cl_is_other_connecting(struct mei_cl *cl) in mei_cl_is_other_connecting() argument
1036 dev = cl->dev; in mei_cl_is_other_connecting()
1040 mei_cl_me_id(cl) == mei_cl_me_id(cb->cl)) in mei_cl_is_other_connecting()
1055 static int mei_cl_send_connect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_connect() argument
1060 dev = cl->dev; in mei_cl_send_connect()
1062 ret = mei_hbm_cl_connect_req(dev, cl); in mei_cl_send_connect()
1063 cl->status = ret; in mei_cl_send_connect()
1065 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_connect()
1070 cl->timer_count = dev->timeouts.connect; in mei_cl_send_connect()
1084 int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_connect() argument
1087 struct mei_device *dev = cl->dev; in mei_cl_irq_connect()
1092 if (mei_cl_is_other_connecting(cl)) in mei_cl_irq_connect()
1103 rets = mei_cl_send_connect(cl, cb); in mei_cl_irq_connect()
1121 int mei_cl_connect(struct mei_cl *cl, struct mei_me_client *me_cl, in mei_cl_connect() argument
1128 if (WARN_ON(!cl || !cl->dev || !me_cl)) in mei_cl_connect()
1131 dev = cl->dev; in mei_cl_connect()
1133 rets = mei_cl_set_connecting(cl, me_cl); in mei_cl_connect()
1137 if (mei_cl_is_fixed_address(cl)) { in mei_cl_connect()
1138 cl->state = MEI_FILE_CONNECTED; in mei_cl_connect()
1146 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_connect()
1150 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_CONNECT, fp); in mei_cl_connect()
1157 if (!mei_cl_is_other_connecting(cl) && mei_hbuf_acquire(dev)) { in mei_cl_connect()
1158 rets = mei_cl_send_connect(cl, cb); in mei_cl_connect()
1164 wait_event_timeout(cl->wait, in mei_cl_connect()
1165 (cl->state == MEI_FILE_CONNECTED || in mei_cl_connect()
1166 cl->state == MEI_FILE_DISCONNECTED || in mei_cl_connect()
1167 cl->state == MEI_FILE_DISCONNECT_REQUIRED || in mei_cl_connect()
1168 cl->state == MEI_FILE_DISCONNECT_REPLY), in mei_cl_connect()
1172 if (!mei_cl_is_connected(cl)) { in mei_cl_connect()
1173 if (cl->state == MEI_FILE_DISCONNECT_REQUIRED) { in mei_cl_connect()
1174 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_connect()
1175 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_connect()
1179 __mei_cl_disconnect(cl); in mei_cl_connect()
1185 if (!cl->status) in mei_cl_connect()
1186 cl->status = -EFAULT; in mei_cl_connect()
1189 rets = cl->status; in mei_cl_connect()
1191 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_connect()
1198 if (!mei_cl_is_connected(cl)) in mei_cl_connect()
1199 mei_cl_set_disconnected(cl); in mei_cl_connect()
1213 struct mei_cl *cl; in mei_cl_alloc_linked() local
1216 cl = mei_cl_allocate(dev); in mei_cl_alloc_linked()
1217 if (!cl) { in mei_cl_alloc_linked()
1222 ret = mei_cl_link(cl); in mei_cl_alloc_linked()
1226 return cl; in mei_cl_alloc_linked()
1228 kfree(cl); in mei_cl_alloc_linked()
1239 static int mei_cl_tx_flow_ctrl_creds(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds() argument
1241 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds()
1244 if (cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1247 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds()
1250 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds()
1251 if (cl->me_cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1267 static int mei_cl_tx_flow_ctrl_creds_reduce(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds_reduce() argument
1269 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1272 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1275 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds_reduce()
1276 if (WARN_ON(cl->me_cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1278 cl->me_cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1280 if (WARN_ON(cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1282 cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1322 const struct file *mei_cl_fp_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_fp_by_vtag() argument
1326 list_for_each_entry(vtag_l, &cl->vtag_map, list) in mei_cl_fp_by_vtag()
1328 if ((cl->cldev && mei_cldev_enabled(cl->cldev)) || in mei_cl_fp_by_vtag()
1341 static void mei_cl_reset_read_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_reset_read_by_vtag() argument
1345 list_for_each_entry(vtag_l, &cl->vtag_map, list) { in mei_cl_reset_read_by_vtag()
1347 if ((cl->cldev && mei_cldev_enabled(cl->cldev)) || in mei_cl_reset_read_by_vtag()
1361 static void mei_cl_read_vtag_add_fc(struct mei_cl *cl) in mei_cl_read_vtag_add_fc() argument
1365 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_read_vtag_add_fc()
1367 if (mei_cl_enqueue_ctrl_wr_cb(cl, in mei_cl_read_vtag_add_fc()
1368 mei_cl_mtu(cl), in mei_cl_read_vtag_add_fc()
1371 cl->rx_flow_ctrl_creds++; in mei_cl_read_vtag_add_fc()
1386 int mei_cl_vt_support_check(const struct mei_cl *cl) in mei_cl_vt_support_check() argument
1388 struct mei_device *dev = cl->dev; in mei_cl_vt_support_check()
1393 if (!cl->me_cl) in mei_cl_vt_support_check()
1396 return cl->me_cl->props.vt_supported ? 0 : -EOPNOTSUPP; in mei_cl_vt_support_check()
1407 void mei_cl_add_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_add_rd_completed() argument
1411 if (!mei_cl_vt_support_check(cl)) { in mei_cl_add_rd_completed()
1412 fp = mei_cl_fp_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1419 mei_cl_reset_read_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1420 mei_cl_read_vtag_add_fc(cl); in mei_cl_add_rd_completed()
1423 spin_lock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1424 list_add_tail(&cb->list, &cl->rd_completed); in mei_cl_add_rd_completed()
1425 spin_unlock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1435 void mei_cl_del_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_del_rd_completed() argument
1437 spin_lock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1439 spin_unlock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1481 int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_notify() argument
1484 struct mei_device *dev = cl->dev; in mei_cl_irq_notify()
1499 ret = mei_hbm_cl_notify_req(dev, cl, request); in mei_cl_irq_notify()
1501 cl->status = ret; in mei_cl_irq_notify()
1521 int mei_cl_notify_request(struct mei_cl *cl, in mei_cl_notify_request() argument
1529 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_request()
1532 dev = cl->dev; in mei_cl_notify_request()
1535 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_request()
1539 if (!mei_cl_is_connected(cl)) in mei_cl_notify_request()
1545 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_notify_request()
1550 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, fop_type, fp); in mei_cl_notify_request()
1557 if (mei_hbm_cl_notify_req(dev, cl, request)) { in mei_cl_notify_request()
1565 wait_event_timeout(cl->wait, in mei_cl_notify_request()
1566 cl->notify_en == request || in mei_cl_notify_request()
1567 cl->status || in mei_cl_notify_request()
1568 !mei_cl_is_connected(cl), in mei_cl_notify_request()
1572 if (cl->notify_en != request && !cl->status) in mei_cl_notify_request()
1573 cl->status = -EFAULT; in mei_cl_notify_request()
1575 rets = cl->status; in mei_cl_notify_request()
1578 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_notify_request()
1593 void mei_cl_notify(struct mei_cl *cl) in mei_cl_notify() argument
1597 if (!cl || !cl->dev) in mei_cl_notify()
1600 dev = cl->dev; in mei_cl_notify()
1602 if (!cl->notify_en) in mei_cl_notify()
1605 cl_dbg(dev, cl, "notify event"); in mei_cl_notify()
1606 cl->notify_ev = true; in mei_cl_notify()
1607 if (!mei_cl_bus_notify_event(cl)) in mei_cl_notify()
1608 wake_up_interruptible(&cl->ev_wait); in mei_cl_notify()
1610 if (cl->ev_async) in mei_cl_notify()
1611 kill_fasync(&cl->ev_async, SIGIO, POLL_PRI); in mei_cl_notify()
1626 int mei_cl_notify_get(struct mei_cl *cl, bool block, bool *notify_ev) in mei_cl_notify_get() argument
1633 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_get()
1636 dev = cl->dev; in mei_cl_notify_get()
1639 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_get()
1643 if (!mei_cl_is_connected(cl)) in mei_cl_notify_get()
1646 if (cl->notify_ev) in mei_cl_notify_get()
1653 rets = wait_event_interruptible(cl->ev_wait, cl->notify_ev); in mei_cl_notify_get()
1660 *notify_ev = cl->notify_ev; in mei_cl_notify_get()
1661 cl->notify_ev = false; in mei_cl_notify_get()
1674 int mei_cl_read_start(struct mei_cl *cl, size_t length, const struct file *fp) in mei_cl_read_start() argument
1680 if (WARN_ON(!cl || !cl->dev)) in mei_cl_read_start()
1683 dev = cl->dev; in mei_cl_read_start()
1685 if (!mei_cl_is_connected(cl)) in mei_cl_read_start()
1688 if (!mei_me_cl_is_active(cl->me_cl)) { in mei_cl_read_start()
1689 cl_err(dev, cl, "no such me client\n"); in mei_cl_read_start()
1693 if (mei_cl_is_fixed_address(cl)) in mei_cl_read_start()
1697 if (cl->rx_flow_ctrl_creds) { in mei_cl_read_start()
1698 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1702 cb = mei_cl_enqueue_ctrl_wr_cb(cl, length, MEI_FOP_READ, fp); in mei_cl_read_start()
1706 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1711 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_read_start()
1717 rets = mei_hbm_cl_flow_control_req(dev, cl); in mei_cl_read_start()
1721 list_move_tail(&cb->list, &cl->rd_pending); in mei_cl_read_start()
1723 cl->rx_flow_ctrl_creds++; in mei_cl_read_start()
1726 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_read_start()
1778 is_hbm = cb->cl->me_cl->client_id == 0; in mei_msg_hdr_init()
1779 is_gsc = ((!is_hbm) && cb->cl->dev->hbm_f_gsc_supported && mei_ext_hdr_is_gsc(cb->ext_hdr)); in mei_msg_hdr_init()
1800 mei_hdr->host_addr = mei_cl_host_addr(cb->cl); in mei_msg_hdr_init()
1801 mei_hdr->me_addr = mei_cl_me_id(cb->cl); in mei_msg_hdr_init()
1838 int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_write() argument
1855 if (WARN_ON(!cl || !cl->dev)) in mei_cl_irq_write()
1858 dev = cl->dev; in mei_cl_irq_write()
1864 rets = first_chunk ? mei_cl_tx_flow_ctrl_creds(cl) : 1; in mei_cl_irq_write()
1869 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_irq_write()
1929 cl->status = 0; in mei_cl_irq_write()
1930 cl->writing_state = MEI_WRITING; in mei_cl_irq_write()
1934 if (mei_cl_tx_flow_ctrl_creds_reduce(cl)) { in mei_cl_irq_write()
1948 cl->status = rets; in mei_cl_irq_write()
1965 ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb, unsigned long timeout) in mei_cl_write() argument
1981 if (WARN_ON(!cl || !cl->dev)) in mei_cl_write()
1987 dev = cl->dev; in mei_cl_write()
1992 cl_dbg(dev, cl, "buf_len=%zd\n", buf_len); in mei_cl_write()
2000 cl_err(dev, cl, "rpm: get failed %zd\n", rets); in mei_cl_write()
2005 cl->writing_state = MEI_IDLE; in mei_cl_write()
2008 rets = mei_cl_tx_flow_ctrl_creds(cl); in mei_cl_write()
2022 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_write()
2028 cl_dbg(dev, cl, "Cannot acquire the host buffer: not sending.\n"); in mei_cl_write()
2070 rets = mei_cl_tx_flow_ctrl_creds_reduce(cl); in mei_cl_write()
2074 cl->writing_state = MEI_WRITING; in mei_cl_write()
2086 if (blocking && cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2089 rets = wait_event_interruptible_timeout(cl->tx_wait, in mei_cl_write()
2090 cl->writing_state == MEI_WRITE_COMPLETE || in mei_cl_write()
2091 (!mei_cl_is_connected(cl)), in mei_cl_write()
2097 mei_io_tx_list_free_cl(&dev->write_list, cl, NULL); in mei_cl_write()
2098 mei_io_tx_list_free_cl(&dev->write_waiting_list, cl, NULL); in mei_cl_write()
2108 if (cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2116 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_write()
2133 void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_complete() argument
2135 struct mei_device *dev = cl->dev; in mei_cl_complete()
2140 cl->writing_state = MEI_WRITE_COMPLETE; in mei_cl_complete()
2141 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_complete()
2142 wake_up_interruptible(&cl->tx_wait); in mei_cl_complete()
2150 mei_cl_add_rd_completed(cl, cb); in mei_cl_complete()
2151 if (!mei_cl_is_fixed_address(cl) && in mei_cl_complete()
2152 !WARN_ON(!cl->rx_flow_ctrl_creds)) in mei_cl_complete()
2153 cl->rx_flow_ctrl_creds--; in mei_cl_complete()
2154 if (!mei_cl_bus_rx_event(cl)) in mei_cl_complete()
2155 wake_up_interruptible(&cl->rx_wait); in mei_cl_complete()
2164 if (waitqueue_active(&cl->wait)) in mei_cl_complete()
2165 wake_up(&cl->wait); in mei_cl_complete()
2170 mei_cl_set_disconnected(cl); in mei_cl_complete()
2185 struct mei_cl *cl; in mei_cl_all_disconnect() local
2187 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_all_disconnect()
2188 mei_cl_set_disconnected(cl); in mei_cl_all_disconnect()
2194 struct mei_cl *cl; in mei_cl_dma_map_find() local
2196 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_dma_map_find()
2197 if (cl->dma.buffer_id == buffer_id) in mei_cl_dma_map_find()
2198 return cl; in mei_cl_dma_map_find()
2211 int mei_cl_irq_dma_map(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_map() argument
2214 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_map()
2227 ret = mei_hbm_cl_dma_map_req(dev, cl); in mei_cl_irq_dma_map()
2229 cl->status = ret; in mei_cl_irq_dma_map()
2247 int mei_cl_irq_dma_unmap(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_unmap() argument
2250 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_unmap()
2263 ret = mei_hbm_cl_dma_unmap_req(dev, cl); in mei_cl_irq_dma_unmap()
2265 cl->status = ret; in mei_cl_irq_dma_unmap()
2274 static int mei_cl_dma_alloc(struct mei_cl *cl, u8 buf_id, size_t size) in mei_cl_dma_alloc() argument
2276 cl->dma.vaddr = dmam_alloc_coherent(cl->dev->dev, size, in mei_cl_dma_alloc()
2277 &cl->dma.daddr, GFP_KERNEL); in mei_cl_dma_alloc()
2278 if (!cl->dma.vaddr) in mei_cl_dma_alloc()
2281 cl->dma.buffer_id = buf_id; in mei_cl_dma_alloc()
2282 cl->dma.size = size; in mei_cl_dma_alloc()
2287 static void mei_cl_dma_free(struct mei_cl *cl) in mei_cl_dma_free() argument
2289 cl->dma.buffer_id = 0; in mei_cl_dma_free()
2290 dmam_free_coherent(cl->dev->dev, in mei_cl_dma_free()
2291 cl->dma.size, cl->dma.vaddr, cl->dma.daddr); in mei_cl_dma_free()
2292 cl->dma.size = 0; in mei_cl_dma_free()
2293 cl->dma.vaddr = NULL; in mei_cl_dma_free()
2294 cl->dma.daddr = 0; in mei_cl_dma_free()
2314 int mei_cl_dma_alloc_and_map(struct mei_cl *cl, const struct file *fp, in mei_cl_dma_alloc_and_map() argument
2321 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_alloc_and_map()
2324 dev = cl->dev; in mei_cl_dma_alloc_and_map()
2327 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_alloc_and_map()
2334 if (mei_cl_is_connected(cl)) in mei_cl_dma_alloc_and_map()
2337 if (cl->dma_mapped) in mei_cl_dma_alloc_and_map()
2341 cl_dbg(dev, cl, "client dma with id %d is already allocated\n", in mei_cl_dma_alloc_and_map()
2342 cl->dma.buffer_id); in mei_cl_dma_alloc_and_map()
2349 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_alloc_and_map()
2353 rets = mei_cl_dma_alloc(cl, buffer_id, size); in mei_cl_dma_alloc_and_map()
2359 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_MAP, fp); in mei_cl_dma_alloc_and_map()
2366 if (mei_hbm_cl_dma_map_req(dev, cl)) { in mei_cl_dma_alloc_and_map()
2373 cl->status = 0; in mei_cl_dma_alloc_and_map()
2376 wait_event_timeout(cl->wait, in mei_cl_dma_alloc_and_map()
2377 cl->dma_mapped || cl->status, in mei_cl_dma_alloc_and_map()
2381 if (!cl->dma_mapped && !cl->status) in mei_cl_dma_alloc_and_map()
2382 cl->status = -EFAULT; in mei_cl_dma_alloc_and_map()
2384 rets = cl->status; in mei_cl_dma_alloc_and_map()
2388 mei_cl_dma_free(cl); in mei_cl_dma_alloc_and_map()
2390 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_alloc_and_map()
2408 int mei_cl_dma_unmap(struct mei_cl *cl, const struct file *fp) in mei_cl_dma_unmap() argument
2414 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_unmap()
2417 dev = cl->dev; in mei_cl_dma_unmap()
2420 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_unmap()
2425 if (mei_cl_is_connected(cl)) in mei_cl_dma_unmap()
2428 if (!cl->dma_mapped) in mei_cl_dma_unmap()
2434 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_unmap()
2438 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_UNMAP, fp); in mei_cl_dma_unmap()
2445 if (mei_hbm_cl_dma_unmap_req(dev, cl)) { in mei_cl_dma_unmap()
2452 cl->status = 0; in mei_cl_dma_unmap()
2455 wait_event_timeout(cl->wait, in mei_cl_dma_unmap()
2456 !cl->dma_mapped || cl->status, in mei_cl_dma_unmap()
2460 if (cl->dma_mapped && !cl->status) in mei_cl_dma_unmap()
2461 cl->status = -EFAULT; in mei_cl_dma_unmap()
2463 rets = cl->status; in mei_cl_dma_unmap()
2466 mei_cl_dma_free(cl); in mei_cl_dma_unmap()
2468 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_unmap()