1#!/usr/bin/env python3 2# 3# Build the required docker image to run package unit tests 4# 5# Script Variables: 6# DOCKER_IMG_NAME: <optional, the name of the docker image to generate> 7# default is openbmc/ubuntu-unit-test 8# DISTRO: <optional, the distro to build a docker image against> 9# FORCE_DOCKER_BUILD: <optional, a non-zero value with force all Docker 10# images to be rebuilt rather than reusing caches.> 11# BUILD_URL: <optional, used to detect running under CI context 12# (ex. Jenkins)> 13# BRANCH: <optional, branch to build from each of the openbmc/ 14# repositories> 15# default is master, which will be used if input branch not 16# provided or not found 17# UBUNTU_MIRROR: <optional, the URL of a mirror of Ubuntu to override the 18# default ones in /etc/apt/sources.list> 19# default is empty, and no mirror is used. 20# http_proxy The HTTP address of the proxy server to connect to. 21# Default: "", proxy is not setup if this is not set 22 23import os 24import sys 25import threading 26from datetime import date 27from hashlib import sha256 28 29# typing.Dict is used for type-hints. 30from typing import Any, Callable, Dict, Iterable, Optional # noqa: F401 31 32from sh import docker, git, nproc, uname # type: ignore 33 34try: 35 # Python before 3.8 doesn't have TypedDict, so reroute to standard 'dict'. 36 from typing import TypedDict 37except Exception: 38 39 class TypedDict(dict): # type: ignore 40 # We need to do this to eat the 'total' argument. 41 def __init_subclass__(cls, **kwargs: Any) -> None: 42 super().__init_subclass__() 43 44 45# Declare some variables used in package definitions. 46prefix = "/usr/local" 47proc_count = nproc().strip() 48 49 50class PackageDef(TypedDict, total=False): 51 """Package Definition for packages dictionary.""" 52 53 # rev [optional]: Revision of package to use. 54 rev: str 55 # url [optional]: lambda function to create URL: (package, rev) -> url. 56 url: Callable[[str, str], str] 57 # depends [optional]: List of package dependencies. 58 depends: Iterable[str] 59 # build_type [required]: Build type used for package. 60 # Currently supported: autoconf, cmake, custom, make, meson 61 build_type: str 62 # build_steps [optional]: Steps to run for 'custom' build_type. 63 build_steps: Iterable[str] 64 # config_flags [optional]: List of options to pass configuration tool. 65 config_flags: Iterable[str] 66 # config_env [optional]: List of environment variables to set for config. 67 config_env: Iterable[str] 68 # custom_post_dl [optional]: List of steps to run after download, but 69 # before config / build / install. 70 custom_post_dl: Iterable[str] 71 # custom_post_install [optional]: List of steps to run after install. 72 custom_post_install: Iterable[str] 73 74 # __tag [private]: Generated Docker tag name for package stage. 75 __tag: str 76 # __package [private]: Package object associated with this package. 77 __package: Any # Type is Package, but not defined yet. 78 79 80# Packages to include in image. 81packages = { 82 "boost": PackageDef( 83 rev="1.80.0", 84 url=( 85 lambda pkg, rev: f"https://downloads.yoctoproject.org/mirror/sources/{pkg}_{rev.replace('.', '_')}.tar.bz2" # noqa: E501 86 ), 87 build_type="custom", 88 build_steps=[ 89 ( 90 "./bootstrap.sh" 91 f" --prefix={prefix} --with-libraries=context,coroutine" 92 ), 93 "./b2", 94 f"./b2 install --prefix={prefix}", 95 ], 96 ), 97 "USCiLab/cereal": PackageDef( 98 rev="v1.3.2", 99 build_type="custom", 100 build_steps=[f"cp -a include/cereal/ {prefix}/include/"], 101 ), 102 "danmar/cppcheck": PackageDef( 103 rev="2.9", 104 build_type="cmake", 105 ), 106 "CLIUtils/CLI11": PackageDef( 107 rev="v1.9.1", 108 build_type="cmake", 109 config_flags=[ 110 "-DBUILD_TESTING=OFF", 111 "-DCLI11_BUILD_DOCS=OFF", 112 "-DCLI11_BUILD_EXAMPLES=OFF", 113 ], 114 ), 115 "fmtlib/fmt": PackageDef( 116 rev="9.1.0", 117 build_type="cmake", 118 config_flags=[ 119 "-DFMT_DOC=OFF", 120 "-DFMT_TEST=OFF", 121 ], 122 ), 123 "Naios/function2": PackageDef( 124 rev="4.2.1", 125 build_type="custom", 126 build_steps=[ 127 f"mkdir {prefix}/include/function2", 128 f"cp include/function2/function2.hpp {prefix}/include/function2/", 129 ], 130 ), 131 # release-1.12.1 132 "google/googletest": PackageDef( 133 rev="58d77fa8070e8cec2dc1ed015d66b454c8d78850", 134 build_type="cmake", 135 config_env=["CXXFLAGS=-std=c++20"], 136 config_flags=["-DTHREADS_PREFER_PTHREAD_FLAG=ON"], 137 ), 138 "nlohmann/json": PackageDef( 139 rev="v3.11.2", 140 build_type="cmake", 141 config_flags=["-DJSON_BuildTests=OFF"], 142 custom_post_install=[ 143 ( 144 f"ln -s {prefix}/include/nlohmann/json.hpp" 145 f" {prefix}/include/json.hpp" 146 ), 147 ], 148 ), 149 # Snapshot from 2019-05-24 150 "linux-test-project/lcov": PackageDef( 151 rev="v1.15", 152 build_type="make", 153 ), 154 # dev-6.0 2022-11-28 155 "openbmc/linux": PackageDef( 156 rev="1b16243b004ce4d977a9f3b9d9e715cf5028f867", 157 build_type="custom", 158 build_steps=[ 159 f"make -j{proc_count} defconfig", 160 f"make INSTALL_HDR_PATH={prefix} headers_install", 161 ], 162 ), 163 "LibVNC/libvncserver": PackageDef( 164 rev="LibVNCServer-0.9.13", 165 build_type="cmake", 166 ), 167 "leethomason/tinyxml2": PackageDef( 168 rev="9.0.0", 169 build_type="cmake", 170 ), 171 # version from /meta-openembedded/meta-oe/recipes-devtools/boost-url/boost-url_git.bb # noqa: E501 172 "CPPAlliance/url": PackageDef( 173 rev="d740a92d38e3a8f4d5b2153f53b82f1c98e312ab", 174 build_type="custom", 175 build_steps=[f"cp -a include/** {prefix}/include/"], 176 ), 177 "tristanpenman/valijson": PackageDef( 178 rev="v0.7", 179 build_type="cmake", 180 config_flags=[ 181 "-Dvalijson_BUILD_TESTS=0", 182 "-Dvalijson_INSTALL_HEADERS=1", 183 ], 184 ), 185 "open-power/pdbg": PackageDef(build_type="autoconf"), 186 "openbmc/gpioplus": PackageDef( 187 depends=["openbmc/stdplus"], 188 build_type="meson", 189 config_flags=[ 190 "-Dexamples=false", 191 "-Dtests=disabled", 192 ], 193 ), 194 "openbmc/phosphor-dbus-interfaces": PackageDef( 195 depends=["openbmc/sdbusplus"], 196 build_type="meson", 197 config_flags=["-Dgenerate_md=false"], 198 ), 199 "openbmc/phosphor-logging": PackageDef( 200 depends=[ 201 "USCiLab/cereal", 202 "openbmc/phosphor-dbus-interfaces", 203 "openbmc/sdbusplus", 204 "openbmc/sdeventplus", 205 ], 206 build_type="meson", 207 config_flags=[ 208 "-Dlibonly=true", 209 "-Dtests=disabled", 210 f"-Dyamldir={prefix}/share/phosphor-dbus-yaml/yaml", 211 ], 212 ), 213 "openbmc/phosphor-objmgr": PackageDef( 214 depends=[ 215 "CLIUtils/CLI11", 216 "boost", 217 "leethomason/tinyxml2", 218 "openbmc/phosphor-dbus-interfaces", 219 "openbmc/phosphor-logging", 220 "openbmc/sdbusplus", 221 ], 222 build_type="meson", 223 config_flags=[ 224 "-Dtests=disabled", 225 ], 226 ), 227 "openbmc/libpldm": PackageDef( 228 build_type="meson", 229 config_flags=[ 230 "-Doem-ibm=enabled", 231 "-Dtests=disabled", 232 ], 233 ), 234 "openbmc/sdbusplus": PackageDef( 235 build_type="meson", 236 custom_post_dl=[ 237 "cd tools", 238 f"./setup.py install --root=/ --prefix={prefix}", 239 "cd ..", 240 ], 241 config_flags=[ 242 "-Dexamples=disabled", 243 "-Dtests=disabled", 244 ], 245 ), 246 "openbmc/sdeventplus": PackageDef( 247 depends=[ 248 "Naios/function2", 249 "openbmc/stdplus", 250 ], 251 build_type="meson", 252 config_flags=[ 253 "-Dexamples=false", 254 "-Dtests=disabled", 255 ], 256 ), 257 "openbmc/stdplus": PackageDef( 258 depends=[ 259 "fmtlib/fmt", 260 "google/googletest", 261 "Naios/function2", 262 ], 263 build_type="meson", 264 config_flags=[ 265 "-Dexamples=false", 266 "-Dtests=disabled", 267 "-Dgtest=enabled", 268 ], 269 ), 270} # type: Dict[str, PackageDef] 271 272# Define common flags used for builds 273configure_flags = " ".join( 274 [ 275 f"--prefix={prefix}", 276 ] 277) 278cmake_flags = " ".join( 279 [ 280 "-DBUILD_SHARED_LIBS=ON", 281 "-DCMAKE_BUILD_TYPE=RelWithDebInfo", 282 f"-DCMAKE_INSTALL_PREFIX:PATH={prefix}", 283 "-GNinja", 284 "-DCMAKE_MAKE_PROGRAM=ninja", 285 ] 286) 287meson_flags = " ".join( 288 [ 289 "--wrap-mode=nodownload", 290 f"-Dprefix={prefix}", 291 ] 292) 293 294 295class Package(threading.Thread): 296 """Class used to build the Docker stages for each package. 297 298 Generally, this class should not be instantiated directly but through 299 Package.generate_all(). 300 """ 301 302 # Copy the packages dictionary. 303 packages = packages.copy() 304 305 # Lock used for thread-safety. 306 lock = threading.Lock() 307 308 def __init__(self, pkg: str): 309 """pkg - The name of this package (ex. foo/bar )""" 310 super(Package, self).__init__() 311 312 self.package = pkg 313 self.exception = None # type: Optional[Exception] 314 315 # Reference to this package's 316 self.pkg_def = Package.packages[pkg] 317 self.pkg_def["__package"] = self 318 319 def run(self) -> None: 320 """Thread 'run' function. Builds the Docker stage.""" 321 322 # In case this package has no rev, fetch it from Github. 323 self._update_rev() 324 325 # Find all the Package objects that this package depends on. 326 # This section is locked because we are looking into another 327 # package's PackageDef dict, which could be being modified. 328 Package.lock.acquire() 329 deps: Iterable[Package] = [ 330 Package.packages[deppkg]["__package"] 331 for deppkg in self.pkg_def.get("depends", []) 332 ] 333 Package.lock.release() 334 335 # Wait until all the depends finish building. We need them complete 336 # for the "COPY" commands. 337 for deppkg in deps: 338 deppkg.join() 339 340 # Generate this package's Dockerfile. 341 dockerfile = f""" 342FROM {docker_base_img_name} 343{self._df_copycmds()} 344{self._df_build()} 345""" 346 347 # Generate the resulting tag name and save it to the PackageDef. 348 # This section is locked because we are modifying the PackageDef, 349 # which can be accessed by other threads. 350 Package.lock.acquire() 351 tag = Docker.tagname(self._stagename(), dockerfile) 352 self.pkg_def["__tag"] = tag 353 Package.lock.release() 354 355 # Do the build / save any exceptions. 356 try: 357 Docker.build(self.package, tag, dockerfile) 358 except Exception as e: 359 self.exception = e 360 361 @classmethod 362 def generate_all(cls) -> None: 363 """Ensure a Docker stage is created for all defined packages. 364 365 These are done in parallel but with appropriate blocking per 366 package 'depends' specifications. 367 """ 368 369 # Create a Package for each defined package. 370 pkg_threads = [Package(p) for p in cls.packages.keys()] 371 372 # Start building them all. 373 # This section is locked because threads depend on each other, 374 # based on the packages, and they cannot 'join' on a thread 375 # which is not yet started. Adding a lock here allows all the 376 # threads to start before they 'join' their dependencies. 377 Package.lock.acquire() 378 for t in pkg_threads: 379 t.start() 380 Package.lock.release() 381 382 # Wait for completion. 383 for t in pkg_threads: 384 t.join() 385 # Check if the thread saved off its own exception. 386 if t.exception: 387 print(f"Package {t.package} failed!", file=sys.stderr) 388 raise t.exception 389 390 @staticmethod 391 def df_all_copycmds() -> str: 392 """Formulate the Dockerfile snippet necessary to copy all packages 393 into the final image. 394 """ 395 return Package.df_copycmds_set(Package.packages.keys()) 396 397 @classmethod 398 def depcache(cls) -> str: 399 """Create the contents of the '/tmp/depcache'. 400 This file is a comma-separated list of "<pkg>:<rev>". 401 """ 402 403 # This needs to be sorted for consistency. 404 depcache = "" 405 for pkg in sorted(cls.packages.keys()): 406 depcache += "%s:%s," % (pkg, cls.packages[pkg]["rev"]) 407 return depcache 408 409 def _update_rev(self) -> None: 410 """Look up the HEAD for missing a static rev.""" 411 412 if "rev" in self.pkg_def: 413 return 414 415 # Check if Jenkins/Gerrit gave us a revision and use it. 416 if gerrit_project == self.package and gerrit_rev: 417 print( 418 f"Found Gerrit revision for {self.package}: {gerrit_rev}", 419 file=sys.stderr, 420 ) 421 self.pkg_def["rev"] = gerrit_rev 422 return 423 424 # Ask Github for all the branches. 425 lookup = git( 426 "ls-remote", "--heads", f"https://github.com/{self.package}" 427 ) 428 429 # Find the branch matching {branch} (or fallback to master). 430 # This section is locked because we are modifying the PackageDef. 431 Package.lock.acquire() 432 for line in lookup.split("\n"): 433 if f"refs/heads/{branch}" in line: 434 self.pkg_def["rev"] = line.split()[0] 435 elif ( 436 "refs/heads/master" in line or "refs/heads/main" in line 437 ) and "rev" not in self.pkg_def: 438 self.pkg_def["rev"] = line.split()[0] 439 Package.lock.release() 440 441 def _stagename(self) -> str: 442 """Create a name for the Docker stage associated with this pkg.""" 443 return self.package.replace("/", "-").lower() 444 445 def _url(self) -> str: 446 """Get the URL for this package.""" 447 rev = self.pkg_def["rev"] 448 449 # If the lambda exists, call it. 450 if "url" in self.pkg_def: 451 return self.pkg_def["url"](self.package, rev) 452 453 # Default to the github archive URL. 454 return f"https://github.com/{self.package}/archive/{rev}.tar.gz" 455 456 def _cmd_download(self) -> str: 457 """Formulate the command necessary to download and unpack to source.""" 458 459 url = self._url() 460 if ".tar." not in url: 461 raise NotImplementedError( 462 f"Unhandled download type for {self.package}: {url}" 463 ) 464 465 cmd = f"curl -L {url} | tar -x" 466 467 if url.endswith(".bz2"): 468 cmd += "j" 469 elif url.endswith(".gz"): 470 cmd += "z" 471 else: 472 raise NotImplementedError( 473 f"Unknown tar flags needed for {self.package}: {url}" 474 ) 475 476 return cmd 477 478 def _cmd_cd_srcdir(self) -> str: 479 """Formulate the command necessary to 'cd' into the source dir.""" 480 return f"cd {self.package.split('/')[-1]}*" 481 482 def _df_copycmds(self) -> str: 483 """Formulate the dockerfile snippet necessary to COPY all depends.""" 484 485 if "depends" not in self.pkg_def: 486 return "" 487 return Package.df_copycmds_set(self.pkg_def["depends"]) 488 489 @staticmethod 490 def df_copycmds_set(pkgs: Iterable[str]) -> str: 491 """Formulate the Dockerfile snippet necessary to COPY a set of 492 packages into a Docker stage. 493 """ 494 495 copy_cmds = "" 496 497 # Sort the packages for consistency. 498 for p in sorted(pkgs): 499 tag = Package.packages[p]["__tag"] 500 copy_cmds += f"COPY --from={tag} {prefix} {prefix}\n" 501 # Workaround for upstream docker bug and multiple COPY cmds 502 # https://github.com/moby/moby/issues/37965 503 copy_cmds += "RUN true\n" 504 505 return copy_cmds 506 507 def _df_build(self) -> str: 508 """Formulate the Dockerfile snippet necessary to download, build, and 509 install a package into a Docker stage. 510 """ 511 512 # Download and extract source. 513 result = f"RUN {self._cmd_download()} && {self._cmd_cd_srcdir()} && " 514 515 # Handle 'custom_post_dl' commands. 516 custom_post_dl = self.pkg_def.get("custom_post_dl") 517 if custom_post_dl: 518 result += " && ".join(custom_post_dl) + " && " 519 520 # Build and install package based on 'build_type'. 521 build_type = self.pkg_def["build_type"] 522 if build_type == "autoconf": 523 result += self._cmd_build_autoconf() 524 elif build_type == "cmake": 525 result += self._cmd_build_cmake() 526 elif build_type == "custom": 527 result += self._cmd_build_custom() 528 elif build_type == "make": 529 result += self._cmd_build_make() 530 elif build_type == "meson": 531 result += self._cmd_build_meson() 532 else: 533 raise NotImplementedError( 534 f"Unhandled build type for {self.package}: {build_type}" 535 ) 536 537 # Handle 'custom_post_install' commands. 538 custom_post_install = self.pkg_def.get("custom_post_install") 539 if custom_post_install: 540 result += " && " + " && ".join(custom_post_install) 541 542 return result 543 544 def _cmd_build_autoconf(self) -> str: 545 options = " ".join(self.pkg_def.get("config_flags", [])) 546 env = " ".join(self.pkg_def.get("config_env", [])) 547 result = "./bootstrap.sh && " 548 result += f"{env} ./configure {configure_flags} {options} && " 549 result += f"make -j{proc_count} && make install" 550 return result 551 552 def _cmd_build_cmake(self) -> str: 553 options = " ".join(self.pkg_def.get("config_flags", [])) 554 env = " ".join(self.pkg_def.get("config_env", [])) 555 result = "mkdir builddir && cd builddir && " 556 result += f"{env} cmake {cmake_flags} {options} .. && " 557 result += "cmake --build . --target all && " 558 result += "cmake --build . --target install && " 559 result += "cd .." 560 return result 561 562 def _cmd_build_custom(self) -> str: 563 return " && ".join(self.pkg_def.get("build_steps", [])) 564 565 def _cmd_build_make(self) -> str: 566 return f"make -j{proc_count} && make install" 567 568 def _cmd_build_meson(self) -> str: 569 options = " ".join(self.pkg_def.get("config_flags", [])) 570 env = " ".join(self.pkg_def.get("config_env", [])) 571 result = f"{env} meson builddir {meson_flags} {options} && " 572 result += "ninja -C builddir && ninja -C builddir install" 573 return result 574 575 576class Docker: 577 """Class to assist with Docker interactions. All methods are static.""" 578 579 @staticmethod 580 def timestamp() -> str: 581 """Generate a timestamp for today using the ISO week.""" 582 today = date.today().isocalendar() 583 return f"{today[0]}-W{today[1]:02}" 584 585 @staticmethod 586 def tagname(pkgname: Optional[str], dockerfile: str) -> str: 587 """Generate a tag name for a package using a hash of the Dockerfile.""" 588 result = docker_image_name 589 if pkgname: 590 result += "-" + pkgname 591 592 result += ":" + Docker.timestamp() 593 result += "-" + sha256(dockerfile.encode()).hexdigest()[0:16] 594 595 return result 596 597 @staticmethod 598 def build(pkg: str, tag: str, dockerfile: str) -> None: 599 """Build a docker image using the Dockerfile and tagging it with 'tag'. 600 """ 601 602 # If we're not forcing builds, check if it already exists and skip. 603 if not force_build: 604 if docker.image.ls(tag, "--format", '"{{.Repository}}:{{.Tag}}"'): 605 print( 606 f"Image {tag} already exists. Skipping.", file=sys.stderr 607 ) 608 return 609 610 # Build it. 611 # Capture the output of the 'docker build' command and send it to 612 # stderr (prefixed with the package name). This allows us to see 613 # progress but not polute stdout. Later on we output the final 614 # docker tag to stdout and we want to keep that pristine. 615 # 616 # Other unusual flags: 617 # --no-cache: Bypass the Docker cache if 'force_build'. 618 # --force-rm: Clean up Docker processes if they fail. 619 docker.build( 620 proxy_args, 621 "--network=host", 622 "--force-rm", 623 "--no-cache=true" if force_build else "--no-cache=false", 624 "-t", 625 tag, 626 "-", 627 _in=dockerfile, 628 _out=( 629 lambda line: print( 630 pkg + ":", line, end="", file=sys.stderr, flush=True 631 ) 632 ), 633 ) 634 635 636# Read a bunch of environment variables. 637docker_image_name = os.environ.get( 638 "DOCKER_IMAGE_NAME", "openbmc/ubuntu-unit-test" 639) 640force_build = os.environ.get("FORCE_DOCKER_BUILD") 641is_automated_ci_build = os.environ.get("BUILD_URL", False) 642distro = os.environ.get("DISTRO", "ubuntu:kinetic") 643branch = os.environ.get("BRANCH", "master") 644ubuntu_mirror = os.environ.get("UBUNTU_MIRROR") 645http_proxy = os.environ.get("http_proxy") 646 647gerrit_project = os.environ.get("GERRIT_PROJECT") 648gerrit_rev = os.environ.get("GERRIT_PATCHSET_REVISION") 649 650# Set up some common variables. 651username = os.environ.get("USER", "root") 652homedir = os.environ.get("HOME", "/root") 653gid = os.getgid() 654uid = os.getuid() 655 656# Use well-known constants if user is root 657if username == "root": 658 homedir = "/root" 659 gid = 0 660 uid = 0 661 662# Determine the architecture for Docker. 663arch = uname("-m").strip() 664if arch == "ppc64le": 665 docker_base = "ppc64le/" 666elif arch == "x86_64": 667 docker_base = "" 668elif arch == "aarch64": 669 docker_base = "arm64v8/" 670else: 671 print( 672 f"Unsupported system architecture({arch}) found for docker image", 673 file=sys.stderr, 674 ) 675 sys.exit(1) 676 677# Special flags if setting up a deb mirror. 678mirror = "" 679if "ubuntu" in distro and ubuntu_mirror: 680 mirror = f""" 681RUN echo "deb {ubuntu_mirror} \ 682 $(. /etc/os-release && echo $VERSION_CODENAME) \ 683 main restricted universe multiverse" > /etc/apt/sources.list && \\ 684 echo "deb {ubuntu_mirror} \ 685 $(. /etc/os-release && echo $VERSION_CODENAME)-updates \ 686 main restricted universe multiverse" >> /etc/apt/sources.list && \\ 687 echo "deb {ubuntu_mirror} \ 688 $(. /etc/os-release && echo $VERSION_CODENAME)-security \ 689 main restricted universe multiverse" >> /etc/apt/sources.list && \\ 690 echo "deb {ubuntu_mirror} \ 691 $(. /etc/os-release && echo $VERSION_CODENAME)-proposed \ 692 main restricted universe multiverse" >> /etc/apt/sources.list && \\ 693 echo "deb {ubuntu_mirror} \ 694 $(. /etc/os-release && echo $VERSION_CODENAME)-backports \ 695 main restricted universe multiverse" >> /etc/apt/sources.list 696""" 697 698# Special flags for proxying. 699proxy_cmd = "" 700proxy_keyserver = "" 701proxy_args = [] 702if http_proxy: 703 proxy_cmd = f""" 704RUN echo "[http]" >> {homedir}/.gitconfig && \ 705 echo "proxy = {http_proxy}" >> {homedir}/.gitconfig 706""" 707 proxy_keyserver = f"--keyserver-options http-proxy={http_proxy}" 708 709 proxy_args.extend( 710 [ 711 "--build-arg", 712 f"http_proxy={http_proxy}", 713 "--build-arg", 714 f"https_proxy={http_proxy}", 715 ] 716 ) 717 718# Create base Dockerfile. 719dockerfile_base = f""" 720FROM {docker_base}{distro} 721 722{mirror} 723 724ENV DEBIAN_FRONTEND noninteractive 725 726ENV PYTHONPATH "/usr/local/lib/python3.10/site-packages/" 727 728# Sometimes the ubuntu key expires and we need a way to force an execution 729# of the apt-get commands for the dbgsym-keyring. When this happens we see 730# an error like: "Release: The following signatures were invalid:" 731# Insert a bogus echo that we can change here when we get this error to force 732# the update. 733RUN echo "ubuntu keyserver rev as of 2021-04-21" 734 735# We need the keys to be imported for dbgsym repos 736# New releases have a package, older ones fall back to manual fetching 737# https://wiki.ubuntu.com/Debug%20Symbol%20Packages 738RUN apt-get update && apt-get dist-upgrade -yy && \ 739 ( apt-get install gpgv ubuntu-dbgsym-keyring || \ 740 ( apt-get install -yy dirmngr && \ 741 apt-key adv --keyserver keyserver.ubuntu.com \ 742 {proxy_keyserver} \ 743 --recv-keys F2EDC64DC5AEE1F6B9C621F0C8CAB6595FDFF622 ) ) 744 745# Parse the current repo list into a debug repo list 746RUN sed -n '/^deb /s,^deb [^ ]* ,deb http://ddebs.ubuntu.com ,p' \ 747 /etc/apt/sources.list >/etc/apt/sources.list.d/debug.list 748 749# Remove non-existent debug repos 750RUN sed -i '/-\\(backports\\|security\\) /d' /etc/apt/sources.list.d/debug.list 751 752RUN cat /etc/apt/sources.list.d/debug.list 753 754RUN apt-get update && apt-get dist-upgrade -yy && apt-get install -yy \ 755 gcc-12 \ 756 g++-12 \ 757 libc6-dbg \ 758 libc6-dev \ 759 libtool \ 760 bison \ 761 libdbus-1-dev \ 762 flex \ 763 cmake \ 764 python3 \ 765 python3-dev\ 766 python3-yaml \ 767 python3-mako \ 768 python3-pip \ 769 python3-setuptools \ 770 python3-git \ 771 python3-socks \ 772 pkg-config \ 773 autoconf \ 774 autoconf-archive \ 775 libsystemd-dev \ 776 systemd \ 777 libssl-dev \ 778 libevdev-dev \ 779 libjpeg-dev \ 780 libpng-dev \ 781 ninja-build \ 782 sudo \ 783 curl \ 784 git \ 785 dbus \ 786 iputils-ping \ 787 clang-15 \ 788 clang-format-15 \ 789 clang-tidy-15 \ 790 clang-tools-15 \ 791 shellcheck \ 792 npm \ 793 iproute2 \ 794 libnl-3-dev \ 795 libnl-genl-3-dev \ 796 libconfig++-dev \ 797 libsnmp-dev \ 798 valgrind \ 799 valgrind-dbg \ 800 libpam0g-dev \ 801 xxd \ 802 libi2c-dev \ 803 wget \ 804 libldap2-dev \ 805 libprotobuf-dev \ 806 liburing-dev \ 807 liburing2-dbgsym \ 808 libperlio-gzip-perl \ 809 libjson-perl \ 810 protobuf-compiler \ 811 libgpiod-dev \ 812 device-tree-compiler \ 813 libpciaccess-dev \ 814 libmimetic-dev \ 815 libxml2-utils \ 816 libxml-simple-perl \ 817 rsync \ 818 libcryptsetup-dev 819 820RUN npm install -g eslint@latest eslint-plugin-json@latest 821 822# Kinetic comes with GCC-12, so skip this. 823#RUN update-alternatives --install /usr/bin/gcc gcc /usr/bin/gcc-12 12 \ 824# --slave /usr/bin/g++ g++ /usr/bin/g++-12 \ 825# --slave /usr/bin/gcov gcov /usr/bin/gcov-12 \ 826# --slave /usr/bin/gcov-dump gcov-dump /usr/bin/gcov-dump-12 \ 827# --slave /usr/bin/gcov-tool gcov-tool /usr/bin/gcov-tool-12 828#RUN update-alternatives --install /usr/bin/cpp cpp /usr/bin/cpp-12 12 829 830RUN update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 1000 \ 831 --slave /usr/bin/clang++ clang++ /usr/bin/clang++-15 \ 832 --slave /usr/bin/clang-tidy clang-tidy /usr/bin/clang-tidy-15 \ 833 --slave /usr/bin/clang-format clang-format /usr/bin/clang-format-15 \ 834 --slave /usr/bin/run-clang-tidy run-clang-tidy.py \ 835 /usr/bin/run-clang-tidy-15 \ 836 --slave /usr/bin/scan-build scan-build /usr/bin/scan-build-15 837 838""" 839 840if is_automated_ci_build: 841 dockerfile_base += f""" 842# Run an arbitrary command to polute the docker cache regularly force us 843# to re-run `apt-get update` daily. 844RUN echo {Docker.timestamp()} 845RUN apt-get update && apt-get dist-upgrade -yy 846 847""" 848 849dockerfile_base += """ 850RUN pip3 install inflection 851RUN pip3 install pycodestyle 852RUN pip3 install jsonschema 853RUN pip3 install meson==0.63.0 854RUN pip3 install packaging 855RUN pip3 install protobuf 856RUN pip3 install codespell 857RUN pip3 install requests 858RUN pip3 install gitlint 859""" 860 861# Build the base and stage docker images. 862docker_base_img_name = Docker.tagname("base", dockerfile_base) 863Docker.build("base", docker_base_img_name, dockerfile_base) 864Package.generate_all() 865 866# Create the final Dockerfile. 867dockerfile = f""" 868# Build the final output image 869FROM {docker_base_img_name} 870{Package.df_all_copycmds()} 871 872# Some of our infrastructure still relies on the presence of this file 873# even though it is no longer needed to rebuild the docker environment 874# NOTE: The file is sorted to ensure the ordering is stable. 875RUN echo '{Package.depcache()}' > /tmp/depcache 876 877# Final configuration for the workspace 878RUN grep -q {gid} /etc/group || groupadd -f -g {gid} {username} 879RUN mkdir -p "{os.path.dirname(homedir)}" 880RUN grep -q {uid} /etc/passwd || \ 881 useradd -d {homedir} -m -u {uid} -g {gid} {username} 882RUN sed -i '1iDefaults umask=000' /etc/sudoers 883RUN echo "{username} ALL=(ALL) NOPASSWD: ALL" >>/etc/sudoers 884 885# Ensure user has ability to write to /usr/local for different tool 886# and data installs 887RUN chown -R {username}:{username} /usr/local/share 888 889{proxy_cmd} 890 891RUN /bin/bash 892""" 893 894# Do the final docker build 895docker_final_img_name = Docker.tagname(None, dockerfile) 896Docker.build("final", docker_final_img_name, dockerfile) 897 898# Print the tag of the final image. 899print(docker_final_img_name) 900