Home
last modified time | relevance | path

Searched full:schedule (Results 1 – 25 of 1672) sorted by relevance

12345678910>>...67

/openbmc/bmcweb/redfish-core/schema/dmtf/csdl/
H A DSchedule_v1.xml4 <!--# Redfish Schema: Schedule v1.2.5 -->
25 <Schema xmlns="http://docs.oasis-open.org/odata/ns/edm" Namespace="Schedule">
29 <ComplexType Name="Schedule" Abstract="true">
32 …<Annotation Term="OData.LongDescription" String="The properties of this type shall schedule a seri…
36 <Schema xmlns="http://docs.oasis-open.org/odata/ns/edm" Namespace="Schedule.v1_0_0">
39 <ComplexType Name="Schedule" BaseType="Schedule.Schedule">
40 <Annotation Term="OData.Description" String="Schedule a series of occurrences."/>
41 …<Annotation Term="OData.LongDescription" String="The properties of this type shall schedule a seri…
44 <Annotation Term="OData.Description" String="The schedule name."/>
45 …<Annotation Term="OData.LongDescription" String="The name of the schedule, which is constructed as…
[all …]
/openbmc/linux/tools/testing/selftests/ftrace/test.d/ftrace/
H A Dfunc_profiler.tc31 echo "Now filter on just schedule"
32 echo '*schedule' > set_ftrace_filter
40 # make sure only schedule functions exist
42 echo "testing if only schedule is being traced"
43 if grep -v -e '^#' -e 'schedule' trace; then
44 fail "more than schedule was found"
47 echo "Make sure schedule was traced"
48 if ! grep -e 'schedule' trace > /dev/null; then
50 fail "can not find schedule in trace"
59 if ! grep -v -e '^#' -e 'schedule' trace > /dev/null; then
[all …]
H A Dfunc-filter-glob.tc20 ftrace_filter_check '*schedule' '^.*schedule$'
23 ftrace_filter_check '*schedule*' '^.*schedule.*$'
26 ftrace_filter_check 'schedule*' '^schedule.*$'
44 # filter by ?, schedule is always good
48 fail "can not enable schedule filter"
50 cat set_ftrace_filter | grep '^.schedule$'
52 cat set_ftrace_filter | grep '^schedule$'
H A Dfgraph-filter-stack.tc24 # filter something, schedule is always good
25 if ! echo "schedule" > set_ftrace_filter; then
27 if ! echo ".schedule" > set_ftrace_filter; then
28 fail "can not enable schedule filter"
43 count=`cat trace | grep '()' | grep -v schedule | wc -l`
50 count=`cat trace | grep 'schedule()' | wc -l`
52 fail "No schedule traces found?"
60 count=`cat trace | grep '()' | grep -v schedule | wc -l`
66 count=`cat trace | grep 'schedule()' | wc -l`
68 fail "No schedule traces found?"
H A Dfgraph-filter.tc16 # filter something, schedule is always good
17 if ! echo "schedule" > set_ftrace_filter; then
19 if ! echo ".schedule" > set_ftrace_filter; then
20 fail "can not enable schedule filter"
28 # that only the schedule function was found
29 count=`cat trace | grep '()' | grep -v schedule | wc -l`
35 count=`cat trace | grep 'schedule()' | wc -l`
37 fail "No schedule traces found?"
H A Dfunc_traceonoff_triggers.tc51 # powerpc uses .schedule
52 func="schedule"
57 x=`grep '^\.schedule$' available_filter_functions | wc -l`
59 func=".schedule"
75 cnt=`grep schedule set_ftrace_filter | wc -l`
110 cnt=`grep schedule set_ftrace_filter | wc -l`
122 cnt=`grep schedule set_ftrace_filter | wc -l`
140 cnt=`grep schedule set_ftrace_filter | wc -l`
/openbmc/linux/drivers/net/dsa/sja1105/
H A Dsja1105_tas.c96 * all schedule entries for all ports. These are the Gate Control List (GCL)
101 * iterate cyclically through the "schedule". Each "cycle" has an entry point
102 * and an exit point, both being timeslot indices in the schedule table. The
112 * For each schedule entry (timeslot) k, the engine executes the gate control
151 * - cycle 0: iterates the schedule table from 0 to 2 (and back)
152 * - cycle 1: iterates the schedule table from 3 to 5 (and back)
166 struct sja1105_schedule_entry *schedule; in sja1105_init_scheduling() local
182 /* Discard previous Schedule Table */ in sja1105_init_scheduling()
189 /* Discard previous Schedule Entry Points Parameters Table */ in sja1105_init_scheduling()
196 /* Discard previous Schedule Parameters Table */ in sja1105_init_scheduling()
[all …]
/openbmc/linux/drivers/usb/dwc2/
H A Dhcd_queue.c69 * for the specified QH in the periodic schedule
77 * periodic schedule may occur in the same (micro)frame
112 * pmap_schedule() - Schedule time in a periodic bitmap (pmap).
114 * @map: The bitmap representing the schedule; will be updated
116 * @bits_per_period: The schedule represents several periods. This is how many
118 * of the schedule will repeat after its end.
119 * @periods_in_map: The number of periods in the schedule.
133 * The idea here is that we want to schedule time for repeating events that all
135 * and the events want to repeat every "interval" periods. The schedule
138 * To keep things "simple", we'll represent our schedule with a bitmap that
[all …]
/openbmc/sdbusplus/include/sdbusplus/async/stdexec/__detail/
H A D__schedulers.hpp30 // [execution.senders.schedule]
38 noexcept(noexcept(static_cast<_Scheduler&&>(__sched).schedule())) in tag_invoke()
39 -> decltype(static_cast<_Scheduler&&>(__sched).schedule()) in tag_invoke()
42 sender<decltype(static_cast<_Scheduler&&>(__sched).schedule())>, in tag_invoke()
43 "schedule() member functions must return a sender"); in tag_invoke()
44 return static_cast<_Scheduler&&>(__sched).schedule(); in tag_invoke()
65 inline constexpr schedule_t schedule{}; variable
70 { schedule(static_cast<_Scheduler&&>(__sched)) } -> sender;
78 get_env(schedule(static_cast<_Scheduler&&>(__sched)))))
/openbmc/bmcweb/redfish-core/schema/dmtf/json-schema/
H A DSchedule.v1_2_5.json2 "$id": "http://redfish.dmtf.org/schemas/v1/Schedule.v1_2_5.json",
72 "Schedule": { object
74 "description": "Schedule a series of occurrences.",
75 … "longDescription": "The properties of this type shall schedule a series of occurrences.",
161 … "description": "The time after provisioning when the schedule as a whole expires.",
162 … contain a Redfish Duration that describes the time after provisioning when the schedule expires.",
180 "description": "The schedule name.",
181 …"longDescription": "The name of the schedule, which is constructed as OrgID:ScheduleName. Example…
205 "title": "#Schedule.v1_2_5"
/openbmc/linux/tools/perf/Documentation/
H A Dperf-probe.txt234 …or code changes. For example, actual 10th line of schedule() can be moved easily by modifying sche…
247 Display which lines in schedule() can be probed:
249 ./perf probe --line schedule
251 Add a probe on schedule() function 12th line with recording cpu local variable:
253 ./perf probe schedule:12 cpu
255 ./perf probe --add='schedule:12 cpu'
257 Add one or more probes which has the name start with "schedule".
259 ./perf probe schedule*
261 ./perf probe --add='schedule*'
263 Add probes on lines in schedule() function which calls update_rq_clock().
[all …]
/openbmc/linux/include/crypto/
H A Ddes.h34 * des_expand_key - Expand a DES input key into a key schedule
35 * @ctx: the key schedule
45 * des3_ede_expand_key - Expand a triple DES input key into a key schedule
46 * @ctx: the key schedule
/openbmc/linux/drivers/gpu/drm/i915/
H A Di915_scheduler_types.h97 * A schedule engine represents a submission queue with different priority
107 * @ref: reference count of schedule engine object
118 * @requests: list of requests inflight on this schedule engine
167 * @destroy: destroy schedule engine / cleanup in backend
195 * @schedule: adjust priority of request
201 void (*schedule)(struct i915_request *request, member
/openbmc/openbmc/meta-openembedded/meta-oe/recipes-support/poppler/poppler/
H A D0001-Do-not-overwrite-all-our-build-flags.patch28 set(CMAKE_CXX_FLAGS_DEBUG "-g -O2 -fno-reorder-blocks -fno-schedule-insns -fno-inline")
35 set(CMAKE_C_FLAGS_DEBUG "-g -O2 -fno-reorder-blocks -fno-schedule-insns -fno-inline")
44 # clang does not support -fno-reorder-blocks -fno-schedule-insns, so do not use -O2
52 # clang does not support -fno-reorder-blocks -fno-schedule-insns, so do not use -O2
/openbmc/linux/tools/testing/selftests/bpf/progs/
H A Dtest_snprintf.c35 extern const void schedule __ksym;
58 &schedule, &schedule, &schedule); in handler()
/openbmc/linux/drivers/net/ethernet/microchip/sparx5/
H A Dsparx5_calendar.c41 u32 schedule[SPX5_DSM_CAL_LEN]; member
297 data->schedule[idx] = SPX5_DSM_CAL_EMPTY; in sparx5_dsm_calendar_calc()
301 data->schedule[0] = SPX5_DSM_CAL_MAX_DEVS_PER_TAXI; in sparx5_dsm_calendar_calc()
390 data->schedule[idx] = SPX5_DSM_CAL_MAX_DEVS_PER_TAXI; in sparx5_dsm_calendar_calc()
413 if (data->schedule[slot] == SPX5_DSM_CAL_EMPTY) in sparx5_dsm_calendar_calc()
423 memcpy(data->short_list, data->schedule, in sparx5_dsm_calendar_calc()
432 memcpy(data->long_list, data->schedule, in sparx5_dsm_calendar_calc()
469 data->schedule[slot] = data->temp_sched[slot]; in sparx5_dsm_calendar_calc()
483 u32 cal_length = sparx5_dsm_cal_len(data->schedule); in sparx5_dsm_calendar_check()
494 if (data->schedule[idx] == port) { in sparx5_dsm_calendar_check()
[all …]
/openbmc/linux/drivers/usb/host/
H A Duhci-hcd.h35 * the schedule is buggy */
132 * can easily splice a QH for some endpoint into the schedule at the right
135 * In the schedule, qh->link maintains a list of QHs seen by the HC:
140 * state are on the software list but not the hardware schedule.
143 * but they never get added to the hardware schedule.
147 * schedule but the hardware may
149 #define QH_STATE_ACTIVE 3 /* QH is on the schedule */
153 __hc32 link; /* Next QH in the schedule */
290 * automatic queuing. To make it easy to insert entries into the schedule,
297 * appropriate skeleton QH. For instance, the schedule can look like this:
[all …]
/openbmc/linux/Documentation/scheduler/
H A Dsched-arch.rst27 be enabled to call schedule() then disabled again.
30 be cleared until the running task has called schedule(). Idle
35 schedule(). It should not call schedule() otherwise.
/openbmc/linux/tools/perf/pmu-events/arch/x86/tigerlake/
H A Dother.json3 … where the core was running in a manner where Turbo may be clipped to the Non-AVX turbo schedule.",
11 …les where the core was running in a manner where Turbo may be clipped to the AVX2 turbo schedule.",
19 …s where the core was running in a manner where Turbo may be clipped to the AVX512 turbo schedule.",
/openbmc/linux/tools/perf/tests/
H A Dhists_cumulate.c27 /* perf [kernel] schedule() */
54 /* schedule => run_command => main */
277 * 10.00% perf [kernel] [k] schedule in test1()
288 { 0, 1000, "perf", "[kernel]", "schedule" }, in test1()
353 * 10.00% perf [kernel] [k] schedule in test2()
355 * --- schedule in test2()
386 { 0, 1000, "perf", "[kernel]", "schedule" }, in test2()
418 3, { { "[kernel]", "schedule" }, in test2()
480 * 10.00% 10.00% perf [kernel] [k] schedule in test3()
493 { 1000, 1000, "perf", "[kernel]", "schedule" }, in test3()
[all …]
H A Dhists_common.h49 * perf: 100 [kernel] schedule
58 * perf: 200 [kernel] schedule
67 * bash: 300 [kernel] schedule
/openbmc/qemu/include/hw/usb/
H A Dehci-regs.h18 #define USBCMD_PSE (1 << 4) // Periodic Schedule Enable
19 #define USBCMD_ASE (1 << 5) // Asynch Schedule Enable
37 #define USBSTS_PSS (1 << 14) // Periodic Schedule Status
38 #define USBSTS_ASS (1 << 15) // Asynchronous Schedule Status
/openbmc/linux/arch/powerpc/lib/
H A Dvmx-helper.c20 * We need to disable page faults as they can call schedule and in enter_vmx_usercopy()
41 * Must never explicitly call schedule (including preempt_enable()) in exit_vmx_usercopy()
45 * set and we are preemptible. The hack here is to schedule a in exit_vmx_usercopy()
/openbmc/linux/include/linux/
H A Dwait.h116 * schedule();
330 schedule())
369 0, 0, schedule())
445 cmd1; schedule(); cmd2)
458 cmd1; schedule(); cmd2)
483 schedule())
562 schedule()); \
623 schedule())
636 schedule())
650 schedule())
[all …]
/openbmc/linux/net/atm/
H A Dsvc.c64 schedule(); in svc_disconnect()
140 schedule(); in svc_bind()
221 schedule(); in svc_connect()
244 schedule(); in svc_connect()
251 schedule(); in svc_connect()
311 schedule(); in svc_listen()
363 schedule(); in svc_accept()
405 schedule(); in svc_accept()
450 schedule(); in svc_change_qos()
552 schedule(); in svc_addparty()
[all …]

12345678910>>...67