Home
last modified time | relevance | path

Searched full:jobs (Results 1 – 25 of 389) sorted by relevance

12345678910>>...16

/openbmc/qemu/.gitlab-ci.d/
H A Dbase.yml37 # want jobs to run
40 # Never run jobs upstream on stable branch, staging branch jobs already ran
44 # Never run jobs upstream on tags, staging branch jobs already ran
52 # Cirrus jobs can't run unless the creds / target repo are set
56 # Publishing jobs should only run on the default branch in upstream
60 # Non-publishing jobs should only run on staging branches in upstream
64 # Jobs only intended for forks should always be skipped on upstream
72 # Avocado jobs don't run in forks unless $QEMU_CI_AVOCADO_TESTING is set
78 # Stage 2: fine tune execution of jobs in specific scenarios
82 # Optional jobs should not be run unless manually triggered
[all …]
H A Dcrossbuild-template.yml14 - JOBS=$(expr $(nproc) + 1)
34 - make -j"$JOBS" all check-build
39 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ;
67 - JOBS=$(expr $(nproc) + 1)
80 - make -j"$JOBS" all check-build
85 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ;
99 - JOBS=$(expr $(nproc) + 1)
114 - make -j"$JOBS" all check-build
119 $MAKE -j"$JOBS" $MAKE_CHECK_ARGS ;
123 # We can still run some tests on some of our cross build jobs. They can add this
H A Dopensbi.yml1 # All jobs needing docker-opensbi must use the same rules it uses.
83 - export JOBS=$(($(getconf _NPROCESSORS_ONLN) + 1))
84 - echo "=== Using ${JOBS} simultaneous jobs ==="
85 - make -j${JOBS} -C roms/opensbi clean
86 …- make -j${JOBS} -C roms opensbi32-generic 2>&1 1>opensbi32-generic-stdout.log | tee -a opensbi32-…
87 - make -j${JOBS} -C roms/opensbi clean
88 …- make -j${JOBS} -C roms opensbi64-generic 2>&1 1>opensbi64-generic-stdout.log | tee -a opensbi64-…
/openbmc/qemu/tests/qemu-iotests/
H A D26466 jobs = self.vm.qmp('query-block-jobs')['return']
67 if jobs and jobs[0]['offset'] > 0:
72 self.assertTrue(jobs and jobs[0]['offset'] > 0) # job started
74 jobs = self.vm.qmp('query-block-jobs')['return']
76 self.assertTrue(jobs)
77 self.assertTrue(jobs[0]['offset'] < jobs[0]['len'])
H A D21921 # Check using the job-* QMP commands with block jobs
32 result = vm.qmp('query-jobs')
48 result = vm.qmp('query-jobs')
59 result = vm.qmp('query-jobs')
63 # any further; therefore, the query-jobs result can be
65 iotests.log(vm.qmp('query-jobs'))
82 result = vm.qmp('query-jobs')
93 while vm.qmp('query-jobs')['return'][0]['total-progress'] < img_size:
117 iotests.log(vm.qmp('query-jobs'))
146 iotests.log(vm.qmp('query-jobs'))
[all …]
H A D056216 res = self.vm.qmp('query-block-jobs')
221 res = self.vm.qmp('query-block-jobs')
225 res = self.vm.qmp('query-block-jobs')
230 res = self.vm.qmp('query-block-jobs')
233 res = self.vm.qmp('query-block-jobs')
237 res = self.vm.qmp('query-block-jobs')
243 res = self.vm.qmp('query-block-jobs')
248 res = self.vm.qmp('query-block-jobs')
258 res = self.vm.qmp('query-block-jobs')
266 res = self.vm.qmp('query-block-jobs')
[all …]
/openbmc/linux/scripts/
H A Djobserver-exec6 # with PARALLELISM environment variable set, and releases the jobs back again.
15 jobs = b"" variable
48 jobs += slot
53 # If something went wrong, give back the jobs.
54 if len(jobs):
55 os.write(writer, jobs)
59 claim = len(jobs) + 1
74 if len(jobs):
75 os.write(writer, jobs)
/openbmc/linux/Documentation/core-api/
H A Dpadata.rst9 Padata is a mechanism by which the kernel can farm jobs out to be done in
16 Padata also supports multithreaded jobs, splitting up the job evenly while load
19 Running Serialized Jobs
25 The first step in using padata to run serialized jobs is to set up a
26 padata_instance structure for overall control of how jobs are to be run::
39 jobs to be serialized independently. A padata_instance may have one or more
40 padata_shells associated with it, each allowing a separate series of jobs.
45 The CPUs used to run jobs can be changed in two ways, programmatically with
52 parallel cpumask describes which processors will be used to execute jobs
116 true parallelism is achieved by submitting multiple jobs. parallel() runs with
[all …]
/openbmc/linux/tools/perf/
H A DMakefile22 # Do a parallel build with multiple jobs, based on the number of CPUs online
25 # (To override it, run 'make JOBS=1' and similar.)
27 ifeq ($(JOBS),)
28JOBS := $(shell (getconf _NPROCESSORS_ONLN || grep -E -c '^processor|^CPU[0-9]' /proc/cpuinfo) 2>/… macro
29 ifeq ($(JOBS),0)
30 JOBS := 1 macro
55 @printf ' BUILD: Doing '\''make \033[33m-j'$(JOBS)'\033[m'\'' parallel build\n'
59 @$(MAKE) -f Makefile.perf --no-print-directory -j$(JOBS) O=$(FULL_O) $(SET_DEBUG) $@
93 # The build-test target is not really parallel, don't print the jobs info,
/openbmc/linux/drivers/gpu/drm/panfrost/
H A Dpanfrost_job.c108 /* JS0: fragment jobs. in panfrost_job_get_slot()
109 * JS1: vertex/tiler jobs in panfrost_job_get_slot()
110 * JS2: compute jobs in panfrost_job_get_slot()
136 * Eventually we may need to support tiler only jobs and h/w with in panfrost_job_write_affinity()
159 struct panfrost_job *job = pfdev->jobs[slot][0]; in panfrost_dequeue_job()
162 pfdev->jobs[slot][0] = pfdev->jobs[slot][1]; in panfrost_dequeue_job()
163 pfdev->jobs[slot][1] = NULL; in panfrost_dequeue_job()
175 if (!pfdev->jobs[slot][0]) { in panfrost_enqueue_job()
176 pfdev->jobs[slot][0] = job; in panfrost_enqueue_job()
180 WARN_ON(pfdev->jobs[slot][1]); in panfrost_enqueue_job()
[all …]
/openbmc/linux/drivers/md/
H A Ddm-kcopyd.c80 * We maintain four lists of jobs:
82 * i) jobs waiting for pages
83 * ii) jobs that have pages, and are waiting for the io to be issued.
84 * iii) jobs that don't need to do any IO and just run a callback
85 * iv) jobs that have completed.
417 static struct kcopyd_job *pop_io_job(struct list_head *jobs, in pop_io_job() argument
423 * For I/O jobs, pop any read, any write without sequential write in pop_io_job()
426 list_for_each_entry(job, jobs, list) { in pop_io_job()
443 static struct kcopyd_job *pop(struct list_head *jobs, in pop() argument
450 if (!list_empty(jobs)) { in pop()
[all …]
/openbmc/qemu/qapi/
H A Djob.json5 # = Background jobs
56 # completion. This is used for long-running jobs like mirror that
62 # @waiting: The job is waiting for other jobs in the transaction to
199 # needs to be run explicitly for jobs that don't have automatic
203 # reached its terminal state, JOB_STATUS_CONCLUDED. For jobs that
216 # Instructs all jobs in a transaction (or a single job if it is not
218 # necessary cleanup. This command requires that all involved jobs are
221 # For jobs in a transaction, instructing one job to finalize will
222 # force ALL jobs in the transaction to finalize, so it is only
268 # @query-jobs:
[all …]
/openbmc/qemu/docs/devel/testing/
H A Dci-runners.rst.inc1 Jobs on Custom Runners
4 Besides the jobs run under the various CI systems listed before, there
5 are a number additional jobs that will run before an actual merge.
7 other GitLab based CI jobs, but rely on additional systems, not the
12 care of running jobs created by events such as a push to a branch.
16 The GitLab CI jobs definition for the custom runners are located under::
67 will run jobs. The association between a machine and a GitLab project
97 Tags are very important as they are used to route specific jobs to
H A Dci-jobs.rst.inc65 * QEMU_JOB_nnnn - variables to be defined in individual jobs
71 which jobs get run in a pipeline
74 in stage 1, for use by build jobs in stage 2. Defaults to
147 the jobs to be manually started from the UI
150 the jobs immediately, as was the historical behaviour
156 these artifacts are not already cached, downloading them make the jobs
163 These variables are primarily to control execution of jobs on
187 The jobs are configured to use "ccache" by default since this typically
/openbmc/qemu/include/qemu/
H A Djob.h2 * Declarations for background jobs
46 /** The ID of the job. May be NULL for internal jobs. */
177 /** Element of the list of jobs */
183 /** Element of the list of jobs in a job transaction */
211 * is the last job to complete in its transaction, all jobs in the
227 * into the paused state. Paused jobs must not perform any asynchronous
228 * I/O or event loop activity. This callback is used to quiesce jobs.
259 * If the callback is not NULL, prepare will be invoked when all the jobs
269 * If the callback is not NULL, it will be invoked when all the jobs
273 * All jobs will complete with a call to either .commit() or .abort() but
[all …]
/openbmc/u-boot/tools/
H A Dgenboardscfg.py226 def scan_defconfigs(jobs=1): argument
232 jobs: The number of jobs to run simultaneously
244 for i in range(jobs):
245 defconfigs = all_defconfigs[total_boards * i / jobs :
246 total_boards * (i + 1) / jobs]
410 def gen_boards_cfg(output, jobs=1, force=False): argument
415 jobs: The number of jobs to run simultaneously
424 params_list = scan_defconfigs(jobs)
438 parser.add_option('-j', '--jobs', type='int', default=cpu_count,
439 help='the number of jobs to run simultaneously')
[all …]
/openbmc/openbmc-test-automation/redfish/extended/
H A Dtest_basic_ci.robot36 # root@witherspoon:~# systemctl list-jobs --no-pager | cat
39 # 1 jobs listed.
42 # root@witherspoon:~# systemctl list-jobs --no-pager | cat
43 # No jobs running.
47 ... systemctl list-jobs --no-pager | cat
48 Should Be Equal As Strings ${stdout} No jobs running.
/openbmc/linux/include/drm/
H A Dgpu_scheduler.h38 * Setting this flag on a scheduler fence prevents pipelining of jobs depending
40 * dependen jobs are pushed to the hw queue.
74 /* Used to chose between FIFO and RR jobs scheduling */
84 * Entities will emit jobs in order to their corresponding hardware
113 * A list of schedulers (struct drm_gpu_scheduler). Jobs from this entity can
151 * @job_queue: the list of jobs of this entity.
318 * @queue_node: used to append this struct to the queue of jobs in an entity.
403 * necessary. Skipped when jobs are killed instead of run.
428 * 2. Try to gracefully stop non-faulty jobs (optional)
430 * 4. Re-submit jobs using drm_sched_resubmit_jobs()
[all …]
/openbmc/qemu/
H A Djob.c2 * Background jobs (long-running operations)
50 * job_mutex protects the jobs list, but also makes the
56 static QLIST_HEAD(, Job) jobs = QLIST_HEAD_INITIALIZER(jobs);
86 /* Transactional group of jobs */
92 /* List of jobs */
93 QLIST_HEAD(, Job) jobs;
117 QLIST_INIT(&txn->jobs); in job_txn_new()
162 QLIST_INSERT_HEAD(&txn->jobs, job, txn_list); in job_txn_add_job_locked()
191 QLIST_FOREACH_SAFE(other_job, &txn->jobs, txn_list, next) { in job_txn_apply_locked()
342 return QLIST_FIRST(&jobs); in job_next_locked()
[all …]
H A D.travis.yml47 # We no longer use nproc to calculate jobs:
51 - export JOBS=3
52 - echo "=== Using ${JOBS} simultaneous jobs ==="
61 - BUILD_RC=0 && make -j${JOBS} || BUILD_RC=$?
79 jobs:
188 - BUILD_RC=0 && make -j${JOBS} || BUILD_RC=$?
/openbmc/linux/tools/memory-model/scripts/
H A Dparseargs.sh42 echo " --jobs N (number of jobs, default one per CPU)"
45 …STDIR_DEF' --herdopts '$LKMM_HERD_OPTIONS_DEF' --hw '$LKMM_HW_MAP_FILE' --jobs '$LKMM_JOBS_DEF' --…
115 --jobs|--job|-j)
116 checkarg --jobs "(number)" "$#" "$2" '^[1-9][0-9]*$' '^--'
/openbmc/bmcweb/redfish-core/schema/dmtf/json-schema/
H A DJobService.v1_0_6.json36 …s, represents the properties for the job service itself, and has links to jobs managed by the job …
95 "Jobs": { object
97 "description": "The links to the jobs collection.",
164 "description": "The maximum number of jobs supported.",
165 …"longDescription": "This property shall contain the maximum number of jobs supported by the implem…
182 "description": "An indication of whether scheduling of jobs is supported.",
183 …operty shall indicate whether the `Schedule` property within the job supports scheduling of jobs.",
/openbmc/qemu/ui/
H A Dvnc-jobs.c31 #include "vnc-jobs.h"
41 * - jobs queue lock: for each operation on the queue (push, pop, isEmpty?)
61 QTAILQ_HEAD(, VncJob) jobs;
117 QTAILQ_INSERT_TAIL(&queue->jobs, job, next); in vnc_job_push()
127 QTAILQ_FOREACH(job, &queue->jobs, next) { in vnc_has_job_locked()
247 while (QTAILQ_EMPTY(&queue->jobs) && !queue->exit) { in vnc_worker_thread_loop()
251 job = QTAILQ_FIRST(&queue->jobs); in vnc_worker_thread_loop()
330 QTAILQ_REMOVE(&queue->jobs, job, next); in vnc_worker_thread_loop()
344 QTAILQ_INIT(&queue->jobs); in vnc_queue_init()
/openbmc/qemu/tests/qemu-iotests/tests/
H A Dmirror-change-copy-mode127 result = self.vm.cmd('query-block-jobs')
132 result = self.vm.cmd('query-block-jobs')
146 result = self.vm.cmd('query-block-jobs')
162 result = self.vm.cmd('query-block-jobs')
168 result = self.vm.cmd('query-block-jobs')
188 while len(self.vm.cmd('query-block-jobs')) > 0:
/openbmc/linux/tools/testing/selftests/net/
H A Dudpgro_bench.sh13 local -r jobs="$(jobs -p)"
16 [ -n "${jobs}" ] && kill -INT ${jobs} 2>/dev/null

12345678910>>...16