1 // SPDX-License-Identifier: GPL-2.0 2 // Copyright (c) 2010-2011 EIA Electronics, 3 // Pieter Beyens <pieter.beyens@eia.be> 4 // Copyright (c) 2010-2011 EIA Electronics, 5 // Kurt Van Dijck <kurt.van.dijck@eia.be> 6 // Copyright (c) 2018 Protonic, 7 // Robin van der Gracht <robin@protonic.nl> 8 // Copyright (c) 2017-2019 Pengutronix, 9 // Marc Kleine-Budde <kernel@pengutronix.de> 10 // Copyright (c) 2017-2019 Pengutronix, 11 // Oleksij Rempel <kernel@pengutronix.de> 12 13 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 14 15 #include <linux/can/core.h> 16 #include <linux/can/skb.h> 17 #include <linux/errqueue.h> 18 #include <linux/if_arp.h> 19 20 #include "j1939-priv.h" 21 22 #define J1939_MIN_NAMELEN CAN_REQUIRED_SIZE(struct sockaddr_can, can_addr.j1939) 23 24 /* conversion function between struct sock::sk_priority from linux and 25 * j1939 priority field 26 */ 27 static inline priority_t j1939_prio(u32 sk_priority) 28 { 29 sk_priority = min(sk_priority, 7U); 30 31 return 7 - sk_priority; 32 } 33 34 static inline u32 j1939_to_sk_priority(priority_t prio) 35 { 36 return 7 - prio; 37 } 38 39 /* function to see if pgn is to be evaluated */ 40 static inline bool j1939_pgn_is_valid(pgn_t pgn) 41 { 42 return pgn <= J1939_PGN_MAX; 43 } 44 45 /* test function to avoid non-zero DA placeholder for pdu1 pgn's */ 46 static inline bool j1939_pgn_is_clean_pdu(pgn_t pgn) 47 { 48 if (j1939_pgn_is_pdu1(pgn)) 49 return !(pgn & 0xff); 50 else 51 return true; 52 } 53 54 static inline void j1939_sock_pending_add(struct sock *sk) 55 { 56 struct j1939_sock *jsk = j1939_sk(sk); 57 58 atomic_inc(&jsk->skb_pending); 59 } 60 61 static int j1939_sock_pending_get(struct sock *sk) 62 { 63 struct j1939_sock *jsk = j1939_sk(sk); 64 65 return atomic_read(&jsk->skb_pending); 66 } 67 68 void j1939_sock_pending_del(struct sock *sk) 69 { 70 struct j1939_sock *jsk = j1939_sk(sk); 71 72 /* atomic_dec_return returns the new value */ 73 if (!atomic_dec_return(&jsk->skb_pending)) 74 wake_up(&jsk->waitq); /* no pending SKB's */ 75 } 76 77 static void j1939_jsk_add(struct j1939_priv *priv, struct j1939_sock *jsk) 78 { 79 jsk->state |= J1939_SOCK_BOUND; 80 j1939_priv_get(priv); 81 jsk->priv = priv; 82 83 spin_lock_bh(&priv->j1939_socks_lock); 84 list_add_tail(&jsk->list, &priv->j1939_socks); 85 spin_unlock_bh(&priv->j1939_socks_lock); 86 } 87 88 static void j1939_jsk_del(struct j1939_priv *priv, struct j1939_sock *jsk) 89 { 90 spin_lock_bh(&priv->j1939_socks_lock); 91 list_del_init(&jsk->list); 92 spin_unlock_bh(&priv->j1939_socks_lock); 93 94 jsk->priv = NULL; 95 j1939_priv_put(priv); 96 jsk->state &= ~J1939_SOCK_BOUND; 97 } 98 99 static bool j1939_sk_queue_session(struct j1939_session *session) 100 { 101 struct j1939_sock *jsk = j1939_sk(session->sk); 102 bool empty; 103 104 spin_lock_bh(&jsk->sk_session_queue_lock); 105 empty = list_empty(&jsk->sk_session_queue); 106 j1939_session_get(session); 107 list_add_tail(&session->sk_session_queue_entry, &jsk->sk_session_queue); 108 spin_unlock_bh(&jsk->sk_session_queue_lock); 109 j1939_sock_pending_add(&jsk->sk); 110 111 return empty; 112 } 113 114 static struct 115 j1939_session *j1939_sk_get_incomplete_session(struct j1939_sock *jsk) 116 { 117 struct j1939_session *session = NULL; 118 119 spin_lock_bh(&jsk->sk_session_queue_lock); 120 if (!list_empty(&jsk->sk_session_queue)) { 121 session = list_last_entry(&jsk->sk_session_queue, 122 struct j1939_session, 123 sk_session_queue_entry); 124 if (session->total_queued_size == session->total_message_size) 125 session = NULL; 126 else 127 j1939_session_get(session); 128 } 129 spin_unlock_bh(&jsk->sk_session_queue_lock); 130 131 return session; 132 } 133 134 static void j1939_sk_queue_drop_all(struct j1939_priv *priv, 135 struct j1939_sock *jsk, int err) 136 { 137 struct j1939_session *session, *tmp; 138 139 netdev_dbg(priv->ndev, "%s: err: %i\n", __func__, err); 140 spin_lock_bh(&jsk->sk_session_queue_lock); 141 list_for_each_entry_safe(session, tmp, &jsk->sk_session_queue, 142 sk_session_queue_entry) { 143 list_del_init(&session->sk_session_queue_entry); 144 session->err = err; 145 j1939_session_put(session); 146 } 147 spin_unlock_bh(&jsk->sk_session_queue_lock); 148 } 149 150 static void j1939_sk_queue_activate_next_locked(struct j1939_session *session) 151 { 152 struct j1939_sock *jsk; 153 struct j1939_session *first; 154 int err; 155 156 /* RX-Session don't have a socket (yet) */ 157 if (!session->sk) 158 return; 159 160 jsk = j1939_sk(session->sk); 161 lockdep_assert_held(&jsk->sk_session_queue_lock); 162 163 err = session->err; 164 165 first = list_first_entry_or_null(&jsk->sk_session_queue, 166 struct j1939_session, 167 sk_session_queue_entry); 168 169 /* Some else has already activated the next session */ 170 if (first != session) 171 return; 172 173 activate_next: 174 list_del_init(&first->sk_session_queue_entry); 175 j1939_session_put(first); 176 first = list_first_entry_or_null(&jsk->sk_session_queue, 177 struct j1939_session, 178 sk_session_queue_entry); 179 if (!first) 180 return; 181 182 if (WARN_ON_ONCE(j1939_session_activate(first))) { 183 first->err = -EBUSY; 184 goto activate_next; 185 } else { 186 /* Give receiver some time (arbitrary chosen) to recover */ 187 int time_ms = 0; 188 189 if (err) 190 time_ms = 10 + prandom_u32_max(16); 191 192 j1939_tp_schedule_txtimer(first, time_ms); 193 } 194 } 195 196 void j1939_sk_queue_activate_next(struct j1939_session *session) 197 { 198 struct j1939_sock *jsk; 199 200 if (!session->sk) 201 return; 202 203 jsk = j1939_sk(session->sk); 204 205 spin_lock_bh(&jsk->sk_session_queue_lock); 206 j1939_sk_queue_activate_next_locked(session); 207 spin_unlock_bh(&jsk->sk_session_queue_lock); 208 } 209 210 static bool j1939_sk_match_dst(struct j1939_sock *jsk, 211 const struct j1939_sk_buff_cb *skcb) 212 { 213 if ((jsk->state & J1939_SOCK_PROMISC)) 214 return true; 215 216 /* Destination address filter */ 217 if (jsk->addr.src_name && skcb->addr.dst_name) { 218 if (jsk->addr.src_name != skcb->addr.dst_name) 219 return false; 220 } else { 221 /* receive (all sockets) if 222 * - all packages that match our bind() address 223 * - all broadcast on a socket if SO_BROADCAST 224 * is set 225 */ 226 if (j1939_address_is_unicast(skcb->addr.da)) { 227 if (jsk->addr.sa != skcb->addr.da) 228 return false; 229 } else if (!sock_flag(&jsk->sk, SOCK_BROADCAST)) { 230 /* receiving broadcast without SO_BROADCAST 231 * flag is not allowed 232 */ 233 return false; 234 } 235 } 236 237 /* Source address filter */ 238 if (jsk->state & J1939_SOCK_CONNECTED) { 239 /* receive (all sockets) if 240 * - all packages that match our connect() name or address 241 */ 242 if (jsk->addr.dst_name && skcb->addr.src_name) { 243 if (jsk->addr.dst_name != skcb->addr.src_name) 244 return false; 245 } else { 246 if (jsk->addr.da != skcb->addr.sa) 247 return false; 248 } 249 } 250 251 /* PGN filter */ 252 if (j1939_pgn_is_valid(jsk->pgn_rx_filter) && 253 jsk->pgn_rx_filter != skcb->addr.pgn) 254 return false; 255 256 return true; 257 } 258 259 /* matches skb control buffer (addr) with a j1939 filter */ 260 static bool j1939_sk_match_filter(struct j1939_sock *jsk, 261 const struct j1939_sk_buff_cb *skcb) 262 { 263 const struct j1939_filter *f = jsk->filters; 264 int nfilter = jsk->nfilters; 265 266 if (!nfilter) 267 /* receive all when no filters are assigned */ 268 return true; 269 270 for (; nfilter; ++f, --nfilter) { 271 if ((skcb->addr.pgn & f->pgn_mask) != f->pgn) 272 continue; 273 if ((skcb->addr.sa & f->addr_mask) != f->addr) 274 continue; 275 if ((skcb->addr.src_name & f->name_mask) != f->name) 276 continue; 277 return true; 278 } 279 return false; 280 } 281 282 static bool j1939_sk_recv_match_one(struct j1939_sock *jsk, 283 const struct j1939_sk_buff_cb *skcb) 284 { 285 if (!(jsk->state & J1939_SOCK_BOUND)) 286 return false; 287 288 if (!j1939_sk_match_dst(jsk, skcb)) 289 return false; 290 291 if (!j1939_sk_match_filter(jsk, skcb)) 292 return false; 293 294 return true; 295 } 296 297 static void j1939_sk_recv_one(struct j1939_sock *jsk, struct sk_buff *oskb) 298 { 299 const struct j1939_sk_buff_cb *oskcb = j1939_skb_to_cb(oskb); 300 struct j1939_sk_buff_cb *skcb; 301 struct sk_buff *skb; 302 303 if (oskb->sk == &jsk->sk) 304 return; 305 306 if (!j1939_sk_recv_match_one(jsk, oskcb)) 307 return; 308 309 skb = skb_clone(oskb, GFP_ATOMIC); 310 if (!skb) { 311 pr_warn("skb clone failed\n"); 312 return; 313 } 314 can_skb_set_owner(skb, oskb->sk); 315 316 skcb = j1939_skb_to_cb(skb); 317 skcb->msg_flags &= ~(MSG_DONTROUTE); 318 if (skb->sk) 319 skcb->msg_flags |= MSG_DONTROUTE; 320 321 if (sock_queue_rcv_skb(&jsk->sk, skb) < 0) 322 kfree_skb(skb); 323 } 324 325 bool j1939_sk_recv_match(struct j1939_priv *priv, struct j1939_sk_buff_cb *skcb) 326 { 327 struct j1939_sock *jsk; 328 bool match = false; 329 330 spin_lock_bh(&priv->j1939_socks_lock); 331 list_for_each_entry(jsk, &priv->j1939_socks, list) { 332 match = j1939_sk_recv_match_one(jsk, skcb); 333 if (match) 334 break; 335 } 336 spin_unlock_bh(&priv->j1939_socks_lock); 337 338 return match; 339 } 340 341 void j1939_sk_recv(struct j1939_priv *priv, struct sk_buff *skb) 342 { 343 struct j1939_sock *jsk; 344 345 spin_lock_bh(&priv->j1939_socks_lock); 346 list_for_each_entry(jsk, &priv->j1939_socks, list) { 347 j1939_sk_recv_one(jsk, skb); 348 } 349 spin_unlock_bh(&priv->j1939_socks_lock); 350 } 351 352 static int j1939_sk_init(struct sock *sk) 353 { 354 struct j1939_sock *jsk = j1939_sk(sk); 355 356 /* Ensure that "sk" is first member in "struct j1939_sock", so that we 357 * can skip it during memset(). 358 */ 359 BUILD_BUG_ON(offsetof(struct j1939_sock, sk) != 0); 360 memset((void *)jsk + sizeof(jsk->sk), 0x0, 361 sizeof(*jsk) - sizeof(jsk->sk)); 362 363 INIT_LIST_HEAD(&jsk->list); 364 init_waitqueue_head(&jsk->waitq); 365 jsk->sk.sk_priority = j1939_to_sk_priority(6); 366 jsk->sk.sk_reuse = 1; /* per default */ 367 jsk->addr.sa = J1939_NO_ADDR; 368 jsk->addr.da = J1939_NO_ADDR; 369 jsk->addr.pgn = J1939_NO_PGN; 370 jsk->pgn_rx_filter = J1939_NO_PGN; 371 atomic_set(&jsk->skb_pending, 0); 372 spin_lock_init(&jsk->sk_session_queue_lock); 373 INIT_LIST_HEAD(&jsk->sk_session_queue); 374 375 return 0; 376 } 377 378 static int j1939_sk_sanity_check(struct sockaddr_can *addr, int len) 379 { 380 if (!addr) 381 return -EDESTADDRREQ; 382 if (len < J1939_MIN_NAMELEN) 383 return -EINVAL; 384 if (addr->can_family != AF_CAN) 385 return -EINVAL; 386 if (!addr->can_ifindex) 387 return -ENODEV; 388 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) && 389 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn)) 390 return -EINVAL; 391 392 return 0; 393 } 394 395 static int j1939_sk_bind(struct socket *sock, struct sockaddr *uaddr, int len) 396 { 397 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr; 398 struct j1939_sock *jsk = j1939_sk(sock->sk); 399 struct j1939_priv *priv = jsk->priv; 400 struct sock *sk = sock->sk; 401 struct net *net = sock_net(sk); 402 int ret = 0; 403 404 ret = j1939_sk_sanity_check(addr, len); 405 if (ret) 406 return ret; 407 408 lock_sock(sock->sk); 409 410 /* Already bound to an interface? */ 411 if (jsk->state & J1939_SOCK_BOUND) { 412 /* A re-bind() to a different interface is not 413 * supported. 414 */ 415 if (jsk->ifindex != addr->can_ifindex) { 416 ret = -EINVAL; 417 goto out_release_sock; 418 } 419 420 /* drop old references */ 421 j1939_jsk_del(priv, jsk); 422 j1939_local_ecu_put(priv, jsk->addr.src_name, jsk->addr.sa); 423 } else { 424 struct net_device *ndev; 425 426 ndev = dev_get_by_index(net, addr->can_ifindex); 427 if (!ndev) { 428 ret = -ENODEV; 429 goto out_release_sock; 430 } 431 432 if (ndev->type != ARPHRD_CAN) { 433 dev_put(ndev); 434 ret = -ENODEV; 435 goto out_release_sock; 436 } 437 438 priv = j1939_netdev_start(ndev); 439 dev_put(ndev); 440 if (IS_ERR(priv)) { 441 ret = PTR_ERR(priv); 442 goto out_release_sock; 443 } 444 445 jsk->ifindex = addr->can_ifindex; 446 } 447 448 /* set default transmit pgn */ 449 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn)) 450 jsk->pgn_rx_filter = addr->can_addr.j1939.pgn; 451 jsk->addr.src_name = addr->can_addr.j1939.name; 452 jsk->addr.sa = addr->can_addr.j1939.addr; 453 454 /* get new references */ 455 ret = j1939_local_ecu_get(priv, jsk->addr.src_name, jsk->addr.sa); 456 if (ret) { 457 j1939_netdev_stop(priv); 458 goto out_release_sock; 459 } 460 461 j1939_jsk_add(priv, jsk); 462 463 out_release_sock: /* fall through */ 464 release_sock(sock->sk); 465 466 return ret; 467 } 468 469 static int j1939_sk_connect(struct socket *sock, struct sockaddr *uaddr, 470 int len, int flags) 471 { 472 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr; 473 struct j1939_sock *jsk = j1939_sk(sock->sk); 474 int ret = 0; 475 476 ret = j1939_sk_sanity_check(addr, len); 477 if (ret) 478 return ret; 479 480 lock_sock(sock->sk); 481 482 /* bind() before connect() is mandatory */ 483 if (!(jsk->state & J1939_SOCK_BOUND)) { 484 ret = -EINVAL; 485 goto out_release_sock; 486 } 487 488 /* A connect() to a different interface is not supported. */ 489 if (jsk->ifindex != addr->can_ifindex) { 490 ret = -EINVAL; 491 goto out_release_sock; 492 } 493 494 if (!addr->can_addr.j1939.name && 495 addr->can_addr.j1939.addr == J1939_NO_ADDR && 496 !sock_flag(&jsk->sk, SOCK_BROADCAST)) { 497 /* broadcast, but SO_BROADCAST not set */ 498 ret = -EACCES; 499 goto out_release_sock; 500 } 501 502 jsk->addr.dst_name = addr->can_addr.j1939.name; 503 jsk->addr.da = addr->can_addr.j1939.addr; 504 505 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn)) 506 jsk->addr.pgn = addr->can_addr.j1939.pgn; 507 508 jsk->state |= J1939_SOCK_CONNECTED; 509 510 out_release_sock: /* fall through */ 511 release_sock(sock->sk); 512 513 return ret; 514 } 515 516 static void j1939_sk_sock2sockaddr_can(struct sockaddr_can *addr, 517 const struct j1939_sock *jsk, int peer) 518 { 519 addr->can_family = AF_CAN; 520 addr->can_ifindex = jsk->ifindex; 521 addr->can_addr.j1939.pgn = jsk->addr.pgn; 522 if (peer) { 523 addr->can_addr.j1939.name = jsk->addr.dst_name; 524 addr->can_addr.j1939.addr = jsk->addr.da; 525 } else { 526 addr->can_addr.j1939.name = jsk->addr.src_name; 527 addr->can_addr.j1939.addr = jsk->addr.sa; 528 } 529 } 530 531 static int j1939_sk_getname(struct socket *sock, struct sockaddr *uaddr, 532 int peer) 533 { 534 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr; 535 struct sock *sk = sock->sk; 536 struct j1939_sock *jsk = j1939_sk(sk); 537 int ret = 0; 538 539 lock_sock(sk); 540 541 if (peer && !(jsk->state & J1939_SOCK_CONNECTED)) { 542 ret = -EADDRNOTAVAIL; 543 goto failure; 544 } 545 546 j1939_sk_sock2sockaddr_can(addr, jsk, peer); 547 ret = J1939_MIN_NAMELEN; 548 549 failure: 550 release_sock(sk); 551 552 return ret; 553 } 554 555 static int j1939_sk_release(struct socket *sock) 556 { 557 struct sock *sk = sock->sk; 558 struct j1939_sock *jsk; 559 560 if (!sk) 561 return 0; 562 563 jsk = j1939_sk(sk); 564 lock_sock(sk); 565 566 if (jsk->state & J1939_SOCK_BOUND) { 567 struct j1939_priv *priv = jsk->priv; 568 569 if (wait_event_interruptible(jsk->waitq, 570 !j1939_sock_pending_get(&jsk->sk))) { 571 j1939_cancel_active_session(priv, sk); 572 j1939_sk_queue_drop_all(priv, jsk, ESHUTDOWN); 573 } 574 575 j1939_jsk_del(priv, jsk); 576 577 j1939_local_ecu_put(priv, jsk->addr.src_name, 578 jsk->addr.sa); 579 580 j1939_netdev_stop(priv); 581 } 582 583 kfree(jsk->filters); 584 sock_orphan(sk); 585 sock->sk = NULL; 586 587 release_sock(sk); 588 sock_put(sk); 589 590 return 0; 591 } 592 593 static int j1939_sk_setsockopt_flag(struct j1939_sock *jsk, char __user *optval, 594 unsigned int optlen, int flag) 595 { 596 int tmp; 597 598 if (optlen != sizeof(tmp)) 599 return -EINVAL; 600 if (copy_from_user(&tmp, optval, optlen)) 601 return -EFAULT; 602 lock_sock(&jsk->sk); 603 if (tmp) 604 jsk->state |= flag; 605 else 606 jsk->state &= ~flag; 607 release_sock(&jsk->sk); 608 return tmp; 609 } 610 611 static int j1939_sk_setsockopt(struct socket *sock, int level, int optname, 612 char __user *optval, unsigned int optlen) 613 { 614 struct sock *sk = sock->sk; 615 struct j1939_sock *jsk = j1939_sk(sk); 616 int tmp, count = 0, ret = 0; 617 struct j1939_filter *filters = NULL, *ofilters; 618 619 if (level != SOL_CAN_J1939) 620 return -EINVAL; 621 622 switch (optname) { 623 case SO_J1939_FILTER: 624 if (optval) { 625 struct j1939_filter *f; 626 int c; 627 628 if (optlen % sizeof(*filters) != 0) 629 return -EINVAL; 630 631 if (optlen > J1939_FILTER_MAX * 632 sizeof(struct j1939_filter)) 633 return -EINVAL; 634 635 count = optlen / sizeof(*filters); 636 filters = memdup_user(optval, optlen); 637 if (IS_ERR(filters)) 638 return PTR_ERR(filters); 639 640 for (f = filters, c = count; c; f++, c--) { 641 f->name &= f->name_mask; 642 f->pgn &= f->pgn_mask; 643 f->addr &= f->addr_mask; 644 } 645 } 646 647 lock_sock(&jsk->sk); 648 ofilters = jsk->filters; 649 jsk->filters = filters; 650 jsk->nfilters = count; 651 release_sock(&jsk->sk); 652 kfree(ofilters); 653 return 0; 654 case SO_J1939_PROMISC: 655 return j1939_sk_setsockopt_flag(jsk, optval, optlen, 656 J1939_SOCK_PROMISC); 657 case SO_J1939_ERRQUEUE: 658 ret = j1939_sk_setsockopt_flag(jsk, optval, optlen, 659 J1939_SOCK_ERRQUEUE); 660 if (ret < 0) 661 return ret; 662 663 if (!(jsk->state & J1939_SOCK_ERRQUEUE)) 664 skb_queue_purge(&sk->sk_error_queue); 665 return ret; 666 case SO_J1939_SEND_PRIO: 667 if (optlen != sizeof(tmp)) 668 return -EINVAL; 669 if (copy_from_user(&tmp, optval, optlen)) 670 return -EFAULT; 671 if (tmp < 0 || tmp > 7) 672 return -EDOM; 673 if (tmp < 2 && !capable(CAP_NET_ADMIN)) 674 return -EPERM; 675 lock_sock(&jsk->sk); 676 jsk->sk.sk_priority = j1939_to_sk_priority(tmp); 677 release_sock(&jsk->sk); 678 return 0; 679 default: 680 return -ENOPROTOOPT; 681 } 682 } 683 684 static int j1939_sk_getsockopt(struct socket *sock, int level, int optname, 685 char __user *optval, int __user *optlen) 686 { 687 struct sock *sk = sock->sk; 688 struct j1939_sock *jsk = j1939_sk(sk); 689 int ret, ulen; 690 /* set defaults for using 'int' properties */ 691 int tmp = 0; 692 int len = sizeof(tmp); 693 void *val = &tmp; 694 695 if (level != SOL_CAN_J1939) 696 return -EINVAL; 697 if (get_user(ulen, optlen)) 698 return -EFAULT; 699 if (ulen < 0) 700 return -EINVAL; 701 702 lock_sock(&jsk->sk); 703 switch (optname) { 704 case SO_J1939_PROMISC: 705 tmp = (jsk->state & J1939_SOCK_PROMISC) ? 1 : 0; 706 break; 707 case SO_J1939_ERRQUEUE: 708 tmp = (jsk->state & J1939_SOCK_ERRQUEUE) ? 1 : 0; 709 break; 710 case SO_J1939_SEND_PRIO: 711 tmp = j1939_prio(jsk->sk.sk_priority); 712 break; 713 default: 714 ret = -ENOPROTOOPT; 715 goto no_copy; 716 } 717 718 /* copy to user, based on 'len' & 'val' 719 * but most sockopt's are 'int' properties, and have 'len' & 'val' 720 * left unchanged, but instead modified 'tmp' 721 */ 722 if (len > ulen) 723 ret = -EFAULT; 724 else if (put_user(len, optlen)) 725 ret = -EFAULT; 726 else if (copy_to_user(optval, val, len)) 727 ret = -EFAULT; 728 else 729 ret = 0; 730 no_copy: 731 release_sock(&jsk->sk); 732 return ret; 733 } 734 735 static int j1939_sk_recvmsg(struct socket *sock, struct msghdr *msg, 736 size_t size, int flags) 737 { 738 struct sock *sk = sock->sk; 739 struct sk_buff *skb; 740 struct j1939_sk_buff_cb *skcb; 741 int ret = 0; 742 743 if (flags & ~(MSG_DONTWAIT | MSG_ERRQUEUE)) 744 return -EINVAL; 745 746 if (flags & MSG_ERRQUEUE) 747 return sock_recv_errqueue(sock->sk, msg, size, SOL_CAN_J1939, 748 SCM_J1939_ERRQUEUE); 749 750 skb = skb_recv_datagram(sk, flags, 0, &ret); 751 if (!skb) 752 return ret; 753 754 if (size < skb->len) 755 msg->msg_flags |= MSG_TRUNC; 756 else 757 size = skb->len; 758 759 ret = memcpy_to_msg(msg, skb->data, size); 760 if (ret < 0) { 761 skb_free_datagram(sk, skb); 762 return ret; 763 } 764 765 skcb = j1939_skb_to_cb(skb); 766 if (j1939_address_is_valid(skcb->addr.da)) 767 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_ADDR, 768 sizeof(skcb->addr.da), &skcb->addr.da); 769 770 if (skcb->addr.dst_name) 771 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_NAME, 772 sizeof(skcb->addr.dst_name), &skcb->addr.dst_name); 773 774 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_PRIO, 775 sizeof(skcb->priority), &skcb->priority); 776 777 if (msg->msg_name) { 778 struct sockaddr_can *paddr = msg->msg_name; 779 780 msg->msg_namelen = J1939_MIN_NAMELEN; 781 memset(msg->msg_name, 0, msg->msg_namelen); 782 paddr->can_family = AF_CAN; 783 paddr->can_ifindex = skb->skb_iif; 784 paddr->can_addr.j1939.name = skcb->addr.src_name; 785 paddr->can_addr.j1939.addr = skcb->addr.sa; 786 paddr->can_addr.j1939.pgn = skcb->addr.pgn; 787 } 788 789 sock_recv_ts_and_drops(msg, sk, skb); 790 msg->msg_flags |= skcb->msg_flags; 791 skb_free_datagram(sk, skb); 792 793 return size; 794 } 795 796 static struct sk_buff *j1939_sk_alloc_skb(struct net_device *ndev, 797 struct sock *sk, 798 struct msghdr *msg, size_t size, 799 int *errcode) 800 { 801 struct j1939_sock *jsk = j1939_sk(sk); 802 struct j1939_sk_buff_cb *skcb; 803 struct sk_buff *skb; 804 int ret; 805 806 skb = sock_alloc_send_skb(sk, 807 size + 808 sizeof(struct can_frame) - 809 sizeof(((struct can_frame *)NULL)->data) + 810 sizeof(struct can_skb_priv), 811 msg->msg_flags & MSG_DONTWAIT, &ret); 812 if (!skb) 813 goto failure; 814 815 can_skb_reserve(skb); 816 can_skb_prv(skb)->ifindex = ndev->ifindex; 817 can_skb_prv(skb)->skbcnt = 0; 818 skb_reserve(skb, offsetof(struct can_frame, data)); 819 820 ret = memcpy_from_msg(skb_put(skb, size), msg, size); 821 if (ret < 0) 822 goto free_skb; 823 824 skb->dev = ndev; 825 826 skcb = j1939_skb_to_cb(skb); 827 memset(skcb, 0, sizeof(*skcb)); 828 skcb->addr = jsk->addr; 829 skcb->priority = j1939_prio(sk->sk_priority); 830 831 if (msg->msg_name) { 832 struct sockaddr_can *addr = msg->msg_name; 833 834 if (addr->can_addr.j1939.name || 835 addr->can_addr.j1939.addr != J1939_NO_ADDR) { 836 skcb->addr.dst_name = addr->can_addr.j1939.name; 837 skcb->addr.da = addr->can_addr.j1939.addr; 838 } 839 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn)) 840 skcb->addr.pgn = addr->can_addr.j1939.pgn; 841 } 842 843 *errcode = ret; 844 return skb; 845 846 free_skb: 847 kfree_skb(skb); 848 failure: 849 *errcode = ret; 850 return NULL; 851 } 852 853 static size_t j1939_sk_opt_stats_get_size(void) 854 { 855 return 856 nla_total_size(sizeof(u32)) + /* J1939_NLA_BYTES_ACKED */ 857 0; 858 } 859 860 static struct sk_buff * 861 j1939_sk_get_timestamping_opt_stats(struct j1939_session *session) 862 { 863 struct sk_buff *stats; 864 u32 size; 865 866 stats = alloc_skb(j1939_sk_opt_stats_get_size(), GFP_ATOMIC); 867 if (!stats) 868 return NULL; 869 870 if (session->skcb.addr.type == J1939_SIMPLE) 871 size = session->total_message_size; 872 else 873 size = min(session->pkt.tx_acked * 7, 874 session->total_message_size); 875 876 nla_put_u32(stats, J1939_NLA_BYTES_ACKED, size); 877 878 return stats; 879 } 880 881 void j1939_sk_errqueue(struct j1939_session *session, 882 enum j1939_sk_errqueue_type type) 883 { 884 struct j1939_priv *priv = session->priv; 885 struct sock *sk = session->sk; 886 struct j1939_sock *jsk; 887 struct sock_exterr_skb *serr; 888 struct sk_buff *skb; 889 char *state = "UNK"; 890 int err; 891 892 /* currently we have no sk for the RX session */ 893 if (!sk) 894 return; 895 896 jsk = j1939_sk(sk); 897 898 if (!(jsk->state & J1939_SOCK_ERRQUEUE)) 899 return; 900 901 skb = j1939_sk_get_timestamping_opt_stats(session); 902 if (!skb) 903 return; 904 905 skb->tstamp = ktime_get_real(); 906 907 BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb)); 908 909 serr = SKB_EXT_ERR(skb); 910 memset(serr, 0, sizeof(*serr)); 911 switch (type) { 912 case J1939_ERRQUEUE_ACK: 913 if (!(sk->sk_tsflags & SOF_TIMESTAMPING_TX_ACK)) { 914 kfree_skb(skb); 915 return; 916 } 917 918 serr->ee.ee_errno = ENOMSG; 919 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING; 920 serr->ee.ee_info = SCM_TSTAMP_ACK; 921 state = "ACK"; 922 break; 923 case J1939_ERRQUEUE_SCHED: 924 if (!(sk->sk_tsflags & SOF_TIMESTAMPING_TX_SCHED)) { 925 kfree_skb(skb); 926 return; 927 } 928 929 serr->ee.ee_errno = ENOMSG; 930 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING; 931 serr->ee.ee_info = SCM_TSTAMP_SCHED; 932 state = "SCH"; 933 break; 934 case J1939_ERRQUEUE_ABORT: 935 serr->ee.ee_errno = session->err; 936 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL; 937 serr->ee.ee_info = J1939_EE_INFO_TX_ABORT; 938 state = "ABT"; 939 break; 940 default: 941 netdev_err(priv->ndev, "Unknown errqueue type %i\n", type); 942 } 943 944 serr->opt_stats = true; 945 if (sk->sk_tsflags & SOF_TIMESTAMPING_OPT_ID) 946 serr->ee.ee_data = session->tskey; 947 948 netdev_dbg(session->priv->ndev, "%s: 0x%p tskey: %i, state: %s\n", 949 __func__, session, session->tskey, state); 950 err = sock_queue_err_skb(sk, skb); 951 952 if (err) 953 kfree_skb(skb); 954 }; 955 956 void j1939_sk_send_loop_abort(struct sock *sk, int err) 957 { 958 sk->sk_err = err; 959 960 sk->sk_error_report(sk); 961 } 962 963 static int j1939_sk_send_loop(struct j1939_priv *priv, struct sock *sk, 964 struct msghdr *msg, size_t size) 965 966 { 967 struct j1939_sock *jsk = j1939_sk(sk); 968 struct j1939_session *session = j1939_sk_get_incomplete_session(jsk); 969 struct sk_buff *skb; 970 size_t segment_size, todo_size; 971 int ret = 0; 972 973 if (session && 974 session->total_message_size != session->total_queued_size + size) { 975 j1939_session_put(session); 976 return -EIO; 977 } 978 979 todo_size = size; 980 981 while (todo_size) { 982 struct j1939_sk_buff_cb *skcb; 983 984 segment_size = min_t(size_t, J1939_MAX_TP_PACKET_SIZE, 985 todo_size); 986 987 /* Allocate skb for one segment */ 988 skb = j1939_sk_alloc_skb(priv->ndev, sk, msg, segment_size, 989 &ret); 990 if (ret) 991 break; 992 993 skcb = j1939_skb_to_cb(skb); 994 995 if (!session) { 996 /* at this point the size should be full size 997 * of the session 998 */ 999 skcb->offset = 0; 1000 session = j1939_tp_send(priv, skb, size); 1001 if (IS_ERR(session)) { 1002 ret = PTR_ERR(session); 1003 goto kfree_skb; 1004 } 1005 if (j1939_sk_queue_session(session)) { 1006 /* try to activate session if we a 1007 * fist in the queue 1008 */ 1009 if (!j1939_session_activate(session)) { 1010 j1939_tp_schedule_txtimer(session, 0); 1011 } else { 1012 ret = -EBUSY; 1013 session->err = ret; 1014 j1939_sk_queue_drop_all(priv, jsk, 1015 EBUSY); 1016 break; 1017 } 1018 } 1019 } else { 1020 skcb->offset = session->total_queued_size; 1021 j1939_session_skb_queue(session, skb); 1022 } 1023 1024 todo_size -= segment_size; 1025 session->total_queued_size += segment_size; 1026 } 1027 1028 switch (ret) { 1029 case 0: /* OK */ 1030 if (todo_size) 1031 netdev_warn(priv->ndev, 1032 "no error found and not completely queued?! %zu\n", 1033 todo_size); 1034 ret = size; 1035 break; 1036 case -ERESTARTSYS: 1037 ret = -EINTR; 1038 /* fall through */ 1039 case -EAGAIN: /* OK */ 1040 if (todo_size != size) 1041 ret = size - todo_size; 1042 break; 1043 default: /* ERROR */ 1044 break; 1045 } 1046 1047 if (session) 1048 j1939_session_put(session); 1049 1050 return ret; 1051 1052 kfree_skb: 1053 kfree_skb(skb); 1054 return ret; 1055 } 1056 1057 static int j1939_sk_sendmsg(struct socket *sock, struct msghdr *msg, 1058 size_t size) 1059 { 1060 struct sock *sk = sock->sk; 1061 struct j1939_sock *jsk = j1939_sk(sk); 1062 struct j1939_priv *priv = jsk->priv; 1063 int ifindex; 1064 int ret; 1065 1066 /* various socket state tests */ 1067 if (!(jsk->state & J1939_SOCK_BOUND)) 1068 return -EBADFD; 1069 1070 ifindex = jsk->ifindex; 1071 1072 if (!jsk->addr.src_name && jsk->addr.sa == J1939_NO_ADDR) 1073 /* no source address assigned yet */ 1074 return -EBADFD; 1075 1076 /* deal with provided destination address info */ 1077 if (msg->msg_name) { 1078 struct sockaddr_can *addr = msg->msg_name; 1079 1080 if (msg->msg_namelen < J1939_MIN_NAMELEN) 1081 return -EINVAL; 1082 1083 if (addr->can_family != AF_CAN) 1084 return -EINVAL; 1085 1086 if (addr->can_ifindex && addr->can_ifindex != ifindex) 1087 return -EBADFD; 1088 1089 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) && 1090 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn)) 1091 return -EINVAL; 1092 1093 if (!addr->can_addr.j1939.name && 1094 addr->can_addr.j1939.addr == J1939_NO_ADDR && 1095 !sock_flag(sk, SOCK_BROADCAST)) 1096 /* broadcast, but SO_BROADCAST not set */ 1097 return -EACCES; 1098 } else { 1099 if (!jsk->addr.dst_name && jsk->addr.da == J1939_NO_ADDR && 1100 !sock_flag(sk, SOCK_BROADCAST)) 1101 /* broadcast, but SO_BROADCAST not set */ 1102 return -EACCES; 1103 } 1104 1105 ret = j1939_sk_send_loop(priv, sk, msg, size); 1106 1107 return ret; 1108 } 1109 1110 void j1939_sk_netdev_event_netdown(struct j1939_priv *priv) 1111 { 1112 struct j1939_sock *jsk; 1113 int error_code = ENETDOWN; 1114 1115 spin_lock_bh(&priv->j1939_socks_lock); 1116 list_for_each_entry(jsk, &priv->j1939_socks, list) { 1117 jsk->sk.sk_err = error_code; 1118 if (!sock_flag(&jsk->sk, SOCK_DEAD)) 1119 jsk->sk.sk_error_report(&jsk->sk); 1120 1121 j1939_sk_queue_drop_all(priv, jsk, error_code); 1122 } 1123 spin_unlock_bh(&priv->j1939_socks_lock); 1124 } 1125 1126 static int j1939_sk_no_ioctlcmd(struct socket *sock, unsigned int cmd, 1127 unsigned long arg) 1128 { 1129 /* no ioctls for socket layer -> hand it down to NIC layer */ 1130 return -ENOIOCTLCMD; 1131 } 1132 1133 static const struct proto_ops j1939_ops = { 1134 .family = PF_CAN, 1135 .release = j1939_sk_release, 1136 .bind = j1939_sk_bind, 1137 .connect = j1939_sk_connect, 1138 .socketpair = sock_no_socketpair, 1139 .accept = sock_no_accept, 1140 .getname = j1939_sk_getname, 1141 .poll = datagram_poll, 1142 .ioctl = j1939_sk_no_ioctlcmd, 1143 .listen = sock_no_listen, 1144 .shutdown = sock_no_shutdown, 1145 .setsockopt = j1939_sk_setsockopt, 1146 .getsockopt = j1939_sk_getsockopt, 1147 .sendmsg = j1939_sk_sendmsg, 1148 .recvmsg = j1939_sk_recvmsg, 1149 .mmap = sock_no_mmap, 1150 .sendpage = sock_no_sendpage, 1151 }; 1152 1153 static struct proto j1939_proto __read_mostly = { 1154 .name = "CAN_J1939", 1155 .owner = THIS_MODULE, 1156 .obj_size = sizeof(struct j1939_sock), 1157 .init = j1939_sk_init, 1158 }; 1159 1160 const struct can_proto j1939_can_proto = { 1161 .type = SOCK_DGRAM, 1162 .protocol = CAN_J1939, 1163 .ops = &j1939_ops, 1164 .prot = &j1939_proto, 1165 }; 1166