/openbmc/qemu/.gitlab-ci.d/ |
H A D | base.yml | 37 # want jobs to run 40 # Never run jobs upstream on stable branch, staging branch jobs already ran 44 # Never run jobs upstream on tags, staging branch jobs already ran 52 # Cirrus jobs can't run unless the creds / target repo are set 56 # Publishing jobs should only run on the default branch in upstream 60 # Non-publishing jobs should only run on staging branches in upstream 64 # Jobs only intended for forks should always be skipped on upstream 72 # Avocado jobs don't run in forks unless $QEMU_CI_AVOCADO_TESTING is set 78 # Stage 2: fine tune execution of jobs in specific scenarios 82 # Optional jobs should not be run unless manually triggered [all …]
|
H A D | crossbuild-template.yml | 14 - JOBS=$(expr $(nproc) + 1) 34 - make -j"$JOBS" all check-build 39 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ; 67 - JOBS=$(expr $(nproc) + 1) 80 - make -j"$JOBS" all check-build 85 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ; 99 - JOBS=$(expr $(nproc) + 1) 114 - make -j"$JOBS" all check-build 119 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ; 123 # We can still run some tests on some of our cross build jobs. They can add this
|
H A D | opensbi.yml | 1 # All jobs needing docker-opensbi must use the same rules it uses. 83 - export JOBS=$(($(getconf _NPROCESSORS_ONLN) + 1)) 84 - echo "=== Using ${JOBS} simultaneous jobs ===" 85 - make -j${JOBS} -C roms/opensbi clean 86 …- make -j${JOBS} -C roms opensbi32-generic 2>&1 1>opensbi32-generic-stdout.log | tee -a opensbi32-… 87 - make -j${JOBS} -C roms/opensbi clean 88 …- make -j${JOBS} -C roms opensbi64-generic 2>&1 1>opensbi64-generic-stdout.log | tee -a opensbi64-…
|
/openbmc/qemu/tests/qemu-iotests/ |
H A D | 264 | 66 jobs = self.vm.qmp('query-block-jobs')['return'] 67 if jobs and jobs[0]['offset'] > 0: 72 self.assertTrue(jobs and jobs[0]['offset'] > 0) # job started 74 jobs = self.vm.qmp('query-block-jobs')['return'] 76 self.assertTrue(jobs) 77 self.assertTrue(jobs[0]['offset'] < jobs[0]['len'])
|
H A D | 219 | 21 # Check using the job-* QMP commands with block jobs 32 result = vm.qmp('query-jobs') 48 result = vm.qmp('query-jobs') 59 result = vm.qmp('query-jobs') 63 # any further; therefore, the query-jobs result can be 65 iotests.log(vm.qmp('query-jobs')) 82 result = vm.qmp('query-jobs') 93 while vm.qmp('query-jobs')['return'][0]['total-progress'] < img_size: 117 iotests.log(vm.qmp('query-jobs')) 146 iotests.log(vm.qmp('query-jobs')) [all …]
|
H A D | 056 | 216 res = self.vm.qmp('query-block-jobs') 221 res = self.vm.qmp('query-block-jobs') 225 res = self.vm.qmp('query-block-jobs') 230 res = self.vm.qmp('query-block-jobs') 233 res = self.vm.qmp('query-block-jobs') 237 res = self.vm.qmp('query-block-jobs') 243 res = self.vm.qmp('query-block-jobs') 248 res = self.vm.qmp('query-block-jobs') 258 res = self.vm.qmp('query-block-jobs') 266 res = self.vm.qmp('query-block-jobs') [all …]
|
/openbmc/linux/scripts/ |
H A D | jobserver-exec | 6 # with PARALLELISM environment variable set, and releases the jobs back again. 15 jobs = b"" variable 48 jobs += slot 53 # If something went wrong, give back the jobs. 54 if len(jobs): 55 os.write(writer, jobs) 59 claim = len(jobs) + 1 74 if len(jobs): 75 os.write(writer, jobs)
|
/openbmc/linux/Documentation/core-api/ |
H A D | padata.rst | 9 Padata is a mechanism by which the kernel can farm jobs out to be done in 16 Padata also supports multithreaded jobs, splitting up the job evenly while load 19 Running Serialized Jobs 25 The first step in using padata to run serialized jobs is to set up a 26 padata_instance structure for overall control of how jobs are to be run:: 39 jobs to be serialized independently. A padata_instance may have one or more 40 padata_shells associated with it, each allowing a separate series of jobs. 45 The CPUs used to run jobs can be changed in two ways, programmatically with 52 parallel cpumask describes which processors will be used to execute jobs 116 true parallelism is achieved by submitting multiple jobs. parallel() runs with [all …]
|
/openbmc/linux/tools/perf/ |
H A D | Makefile | 22 # Do a parallel build with multiple jobs, based on the number of CPUs online 25 # (To override it, run 'make JOBS=1' and similar.) 27 ifeq ($(JOBS),) 28 …JOBS := $(shell (getconf _NPROCESSORS_ONLN || grep -E -c '^processor|^CPU[0-9]' /proc/cpuinfo) 2>/… macro 29 ifeq ($(JOBS),0) 30 JOBS := 1 macro 55 @printf ' BUILD: Doing '\''make \033[33m-j'$(JOBS)'\033[m'\'' parallel build\n' 59 @$(MAKE) -f Makefile.perf --no-print-directory -j$(JOBS) O=$(FULL_O) $(SET_DEBUG) $@ 93 # The build-test target is not really parallel, don't print the jobs info,
|
/openbmc/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_job.c | 108 /* JS0: fragment jobs. in panfrost_job_get_slot() 109 * JS1: vertex/tiler jobs in panfrost_job_get_slot() 110 * JS2: compute jobs in panfrost_job_get_slot() 136 * Eventually we may need to support tiler only jobs and h/w with in panfrost_job_write_affinity() 159 struct panfrost_job *job = pfdev->jobs[slot][0]; in panfrost_dequeue_job() 162 pfdev->jobs[slot][0] = pfdev->jobs[slot][1]; in panfrost_dequeue_job() 163 pfdev->jobs[slot][1] = NULL; in panfrost_dequeue_job() 175 if (!pfdev->jobs[slot][0]) { in panfrost_enqueue_job() 176 pfdev->jobs[slot][0] = job; in panfrost_enqueue_job() 180 WARN_ON(pfdev->jobs[slot][1]); in panfrost_enqueue_job() [all …]
|
/openbmc/linux/drivers/md/ |
H A D | dm-kcopyd.c | 80 * We maintain four lists of jobs: 82 * i) jobs waiting for pages 83 * ii) jobs that have pages, and are waiting for the io to be issued. 84 * iii) jobs that don't need to do any IO and just run a callback 85 * iv) jobs that have completed. 417 static struct kcopyd_job *pop_io_job(struct list_head *jobs, in pop_io_job() argument 423 * For I/O jobs, pop any read, any write without sequential write in pop_io_job() 426 list_for_each_entry(job, jobs, list) { in pop_io_job() 443 static struct kcopyd_job *pop(struct list_head *jobs, in pop() argument 450 if (!list_empty(jobs)) { in pop() [all …]
|
/openbmc/qemu/qapi/ |
H A D | job.json | 5 # = Background jobs 56 # completion. This is used for long-running jobs like mirror that 62 # @waiting: The job is waiting for other jobs in the transaction to 199 # needs to be run explicitly for jobs that don't have automatic 203 # reached its terminal state, JOB_STATUS_CONCLUDED. For jobs that 216 # Instructs all jobs in a transaction (or a single job if it is not 218 # necessary cleanup. This command requires that all involved jobs are 221 # For jobs in a transaction, instructing one job to finalize will 222 # force ALL jobs in the transaction to finalize, so it is only 268 # @query-jobs: [all …]
|
/openbmc/qemu/docs/devel/testing/ |
H A D | ci-runners.rst.inc | 1 Jobs on Custom Runners 4 Besides the jobs run under the various CI systems listed before, there 5 are a number additional jobs that will run before an actual merge. 7 other GitLab based CI jobs, but rely on additional systems, not the 12 care of running jobs created by events such as a push to a branch. 16 The GitLab CI jobs definition for the custom runners are located under:: 67 will run jobs. The association between a machine and a GitLab project 97 Tags are very important as they are used to route specific jobs to
|
H A D | ci-jobs.rst.inc | 65 * QEMU_JOB_nnnn - variables to be defined in individual jobs 71 which jobs get run in a pipeline 74 in stage 1, for use by build jobs in stage 2. Defaults to 147 the jobs to be manually started from the UI 150 the jobs immediately, as was the historical behaviour 156 these artifacts are not already cached, downloading them make the jobs 163 These variables are primarily to control execution of jobs on 187 The jobs are configured to use "ccache" by default since this typically
|
/openbmc/qemu/include/qemu/ |
H A D | job.h | 2 * Declarations for background jobs 46 /** The ID of the job. May be NULL for internal jobs. */ 177 /** Element of the list of jobs */ 183 /** Element of the list of jobs in a job transaction */ 211 * is the last job to complete in its transaction, all jobs in the 227 * into the paused state. Paused jobs must not perform any asynchronous 228 * I/O or event loop activity. This callback is used to quiesce jobs. 259 * If the callback is not NULL, prepare will be invoked when all the jobs 269 * If the callback is not NULL, it will be invoked when all the jobs 273 * All jobs will complete with a call to either .commit() or .abort() but [all …]
|
/openbmc/u-boot/tools/ |
H A D | genboardscfg.py | 226 def scan_defconfigs(jobs=1): argument 232 jobs: The number of jobs to run simultaneously 244 for i in range(jobs): 245 defconfigs = all_defconfigs[total_boards * i / jobs : 246 total_boards * (i + 1) / jobs] 410 def gen_boards_cfg(output, jobs=1, force=False): argument 415 jobs: The number of jobs to run simultaneously 424 params_list = scan_defconfigs(jobs) 438 parser.add_option('-j', '--jobs', type='int', default=cpu_count, 439 help='the number of jobs to run simultaneously') [all …]
|
/openbmc/openbmc-test-automation/redfish/extended/ |
H A D | test_basic_ci.robot | 36 # root@witherspoon:~# systemctl list-jobs --no-pager | cat 39 # 1 jobs listed. 42 # root@witherspoon:~# systemctl list-jobs --no-pager | cat 43 # No jobs running. 47 ... systemctl list-jobs --no-pager | cat 48 Should Be Equal As Strings ${stdout} No jobs running.
|
/openbmc/linux/include/drm/ |
H A D | gpu_scheduler.h | 38 * Setting this flag on a scheduler fence prevents pipelining of jobs depending 40 * dependen jobs are pushed to the hw queue. 74 /* Used to chose between FIFO and RR jobs scheduling */ 84 * Entities will emit jobs in order to their corresponding hardware 113 * A list of schedulers (struct drm_gpu_scheduler). Jobs from this entity can 151 * @job_queue: the list of jobs of this entity. 318 * @queue_node: used to append this struct to the queue of jobs in an entity. 403 * necessary. Skipped when jobs are killed instead of run. 428 * 2. Try to gracefully stop non-faulty jobs (optional) 430 * 4. Re-submit jobs using drm_sched_resubmit_jobs() [all …]
|
/openbmc/qemu/ |
H A D | job.c | 2 * Background jobs (long-running operations) 50 * job_mutex protects the jobs list, but also makes the 56 static QLIST_HEAD(, Job) jobs = QLIST_HEAD_INITIALIZER(jobs); 86 /* Transactional group of jobs */ 92 /* List of jobs */ 93 QLIST_HEAD(, Job) jobs; 117 QLIST_INIT(&txn->jobs); in job_txn_new() 162 QLIST_INSERT_HEAD(&txn->jobs, job, txn_list); in job_txn_add_job_locked() 191 QLIST_FOREACH_SAFE(other_job, &txn->jobs, txn_list, next) { in job_txn_apply_locked() 342 return QLIST_FIRST(&jobs); in job_next_locked() [all …]
|
H A D | .travis.yml | 47 # We no longer use nproc to calculate jobs: 51 - export JOBS=3 52 - echo "=== Using ${JOBS} simultaneous jobs ===" 61 - BUILD_RC=0 && make -j${JOBS} || BUILD_RC=$? 79 jobs: 188 - BUILD_RC=0 && make -j${JOBS} || BUILD_RC=$?
|
/openbmc/linux/tools/memory-model/scripts/ |
H A D | parseargs.sh | 42 echo " --jobs N (number of jobs, default one per CPU)" 45 …STDIR_DEF' --herdopts '$LKMM_HERD_OPTIONS_DEF' --hw '$LKMM_HW_MAP_FILE' --jobs '$LKMM_JOBS_DEF' --… 115 --jobs|--job|-j) 116 checkarg --jobs "(number)" "$#" "$2" '^[1-9][0-9]*$' '^--'
|
/openbmc/bmcweb/redfish-core/schema/dmtf/json-schema/ |
H A D | JobService.v1_0_6.json | 36 …s, represents the properties for the job service itself, and has links to jobs managed by the job … 95 "Jobs": { object 97 "description": "The links to the jobs collection.", 164 "description": "The maximum number of jobs supported.", 165 …"longDescription": "This property shall contain the maximum number of jobs supported by the implem… 182 "description": "An indication of whether scheduling of jobs is supported.", 183 …operty shall indicate whether the `Schedule` property within the job supports scheduling of jobs.",
|
/openbmc/qemu/ui/ |
H A D | vnc-jobs.c | 31 #include "vnc-jobs.h" 41 * - jobs queue lock: for each operation on the queue (push, pop, isEmpty?) 61 QTAILQ_HEAD(, VncJob) jobs; 117 QTAILQ_INSERT_TAIL(&queue->jobs, job, next); in vnc_job_push() 127 QTAILQ_FOREACH(job, &queue->jobs, next) { in vnc_has_job_locked() 247 while (QTAILQ_EMPTY(&queue->jobs) && !queue->exit) { in vnc_worker_thread_loop() 251 job = QTAILQ_FIRST(&queue->jobs); in vnc_worker_thread_loop() 330 QTAILQ_REMOVE(&queue->jobs, job, next); in vnc_worker_thread_loop() 344 QTAILQ_INIT(&queue->jobs); in vnc_queue_init()
|
/openbmc/qemu/tests/qemu-iotests/tests/ |
H A D | mirror-change-copy-mode | 127 result = self.vm.cmd('query-block-jobs') 132 result = self.vm.cmd('query-block-jobs') 146 result = self.vm.cmd('query-block-jobs') 162 result = self.vm.cmd('query-block-jobs') 168 result = self.vm.cmd('query-block-jobs') 188 while len(self.vm.cmd('query-block-jobs')) > 0:
|
/openbmc/linux/tools/testing/selftests/net/ |
H A D | udpgro_bench.sh | 13 local -r jobs="$(jobs -p)" 16 [ -n "${jobs}" ] && kill -INT ${jobs} 2>/dev/null
|