/openbmc/bmcweb/redfish-core/schema/dmtf/csdl/ |
H A D | Schedule_v1.xml | 4 <!--# Redfish Schema: Schedule v1.2.5 --> 25 <Schema xmlns="http://docs.oasis-open.org/odata/ns/edm" Namespace="Schedule"> 29 <ComplexType Name="Schedule" Abstract="true"> 32 …<Annotation Term="OData.LongDescription" String="The properties of this type shall schedule a seri… 36 <Schema xmlns="http://docs.oasis-open.org/odata/ns/edm" Namespace="Schedule.v1_0_0"> 39 <ComplexType Name="Schedule" BaseType="Schedule.Schedule"> 40 <Annotation Term="OData.Description" String="Schedule a series of occurrences."/> 41 …<Annotation Term="OData.LongDescription" String="The properties of this type shall schedule a seri… 44 <Annotation Term="OData.Description" String="The schedule name."/> 45 …<Annotation Term="OData.LongDescription" String="The name of the schedule, which is constructed as… [all …]
|
/openbmc/linux/tools/testing/selftests/ftrace/test.d/ftrace/ |
H A D | func_profiler.tc | 31 echo "Now filter on just schedule" 32 echo '*schedule' > set_ftrace_filter 40 # make sure only schedule functions exist 42 echo "testing if only schedule is being traced" 43 if grep -v -e '^#' -e 'schedule' trace; then 44 fail "more than schedule was found" 47 echo "Make sure schedule was traced" 48 if ! grep -e 'schedule' trace > /dev/null; then 50 fail "can not find schedule in trace" 59 if ! grep -v -e '^#' -e 'schedule' trace > /dev/null; then [all …]
|
H A D | func-filter-glob.tc | 20 ftrace_filter_check '*schedule' '^.*schedule$' 23 ftrace_filter_check '*schedule*' '^.*schedule.*$' 26 ftrace_filter_check 'schedule*' '^schedule.*$' 44 # filter by ?, schedule is always good 48 fail "can not enable schedule filter" 50 cat set_ftrace_filter | grep '^.schedule$' 52 cat set_ftrace_filter | grep '^schedule$'
|
H A D | fgraph-filter-stack.tc | 24 # filter something, schedule is always good 25 if ! echo "schedule" > set_ftrace_filter; then 27 if ! echo ".schedule" > set_ftrace_filter; then 28 fail "can not enable schedule filter" 43 count=`cat trace | grep '()' | grep -v schedule | wc -l` 50 count=`cat trace | grep 'schedule()' | wc -l` 52 fail "No schedule traces found?" 60 count=`cat trace | grep '()' | grep -v schedule | wc -l` 66 count=`cat trace | grep 'schedule()' | wc -l` 68 fail "No schedule traces found?"
|
H A D | fgraph-filter.tc | 16 # filter something, schedule is always good 17 if ! echo "schedule" > set_ftrace_filter; then 19 if ! echo ".schedule" > set_ftrace_filter; then 20 fail "can not enable schedule filter" 28 # that only the schedule function was found 29 count=`cat trace | grep '()' | grep -v schedule | wc -l` 35 count=`cat trace | grep 'schedule()' | wc -l` 37 fail "No schedule traces found?"
|
H A D | func_traceonoff_triggers.tc | 51 # powerpc uses .schedule 52 func="schedule" 57 x=`grep '^\.schedule$' available_filter_functions | wc -l` 59 func=".schedule" 75 cnt=`grep schedule set_ftrace_filter | wc -l` 110 cnt=`grep schedule set_ftrace_filter | wc -l` 122 cnt=`grep schedule set_ftrace_filter | wc -l` 140 cnt=`grep schedule set_ftrace_filter | wc -l`
|
/openbmc/linux/drivers/net/dsa/sja1105/ |
H A D | sja1105_tas.c | 96 * all schedule entries for all ports. These are the Gate Control List (GCL) 101 * iterate cyclically through the "schedule". Each "cycle" has an entry point 102 * and an exit point, both being timeslot indices in the schedule table. The 112 * For each schedule entry (timeslot) k, the engine executes the gate control 151 * - cycle 0: iterates the schedule table from 0 to 2 (and back) 152 * - cycle 1: iterates the schedule table from 3 to 5 (and back) 166 struct sja1105_schedule_entry *schedule; in sja1105_init_scheduling() local 182 /* Discard previous Schedule Table */ in sja1105_init_scheduling() 189 /* Discard previous Schedule Entry Points Parameters Table */ in sja1105_init_scheduling() 196 /* Discard previous Schedule Parameters Table */ in sja1105_init_scheduling() [all …]
|
/openbmc/linux/drivers/usb/dwc2/ |
H A D | hcd_queue.c | 69 * for the specified QH in the periodic schedule 77 * periodic schedule may occur in the same (micro)frame 112 * pmap_schedule() - Schedule time in a periodic bitmap (pmap). 114 * @map: The bitmap representing the schedule; will be updated 116 * @bits_per_period: The schedule represents several periods. This is how many 118 * of the schedule will repeat after its end. 119 * @periods_in_map: The number of periods in the schedule. 133 * The idea here is that we want to schedule time for repeating events that all 135 * and the events want to repeat every "interval" periods. The schedule 138 * To keep things "simple", we'll represent our schedule with a bitmap that [all …]
|
/openbmc/sdbusplus/include/sdbusplus/async/stdexec/__detail/ |
H A D | __schedulers.hpp | 30 // [execution.senders.schedule] 38 noexcept(noexcept(static_cast<_Scheduler&&>(__sched).schedule())) in tag_invoke() 39 -> decltype(static_cast<_Scheduler&&>(__sched).schedule()) in tag_invoke() 42 sender<decltype(static_cast<_Scheduler&&>(__sched).schedule())>, in tag_invoke() 43 "schedule() member functions must return a sender"); in tag_invoke() 44 return static_cast<_Scheduler&&>(__sched).schedule(); in tag_invoke() 65 inline constexpr schedule_t schedule{}; variable 70 { schedule(static_cast<_Scheduler&&>(__sched)) } -> sender; 78 get_env(schedule(static_cast<_Scheduler&&>(__sched)))))
|
/openbmc/bmcweb/redfish-core/schema/dmtf/json-schema/ |
H A D | Schedule.v1_2_5.json | 2 "$id": "http://redfish.dmtf.org/schemas/v1/Schedule.v1_2_5.json", 72 "Schedule": { object 74 "description": "Schedule a series of occurrences.", 75 … "longDescription": "The properties of this type shall schedule a series of occurrences.", 161 … "description": "The time after provisioning when the schedule as a whole expires.", 162 … contain a Redfish Duration that describes the time after provisioning when the schedule expires.", 180 "description": "The schedule name.", 181 …"longDescription": "The name of the schedule, which is constructed as OrgID:ScheduleName. Example… 205 "title": "#Schedule.v1_2_5"
|
/openbmc/linux/tools/perf/Documentation/ |
H A D | perf-probe.txt | 234 …or code changes. For example, actual 10th line of schedule() can be moved easily by modifying sche… 247 Display which lines in schedule() can be probed: 249 ./perf probe --line schedule 251 Add a probe on schedule() function 12th line with recording cpu local variable: 253 ./perf probe schedule:12 cpu 255 ./perf probe --add='schedule:12 cpu' 257 Add one or more probes which has the name start with "schedule". 259 ./perf probe schedule* 261 ./perf probe --add='schedule*' 263 Add probes on lines in schedule() function which calls update_rq_clock(). [all …]
|
/openbmc/linux/include/crypto/ |
H A D | des.h | 34 * des_expand_key - Expand a DES input key into a key schedule 35 * @ctx: the key schedule 45 * des3_ede_expand_key - Expand a triple DES input key into a key schedule 46 * @ctx: the key schedule
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_scheduler_types.h | 97 * A schedule engine represents a submission queue with different priority 107 * @ref: reference count of schedule engine object 118 * @requests: list of requests inflight on this schedule engine 167 * @destroy: destroy schedule engine / cleanup in backend 195 * @schedule: adjust priority of request 201 void (*schedule)(struct i915_request *request, member
|
/openbmc/openbmc/meta-openembedded/meta-oe/recipes-support/poppler/poppler/ |
H A D | 0001-Do-not-overwrite-all-our-build-flags.patch | 28 set(CMAKE_CXX_FLAGS_DEBUG "-g -O2 -fno-reorder-blocks -fno-schedule-insns -fno-inline") 35 set(CMAKE_C_FLAGS_DEBUG "-g -O2 -fno-reorder-blocks -fno-schedule-insns -fno-inline") 44 # clang does not support -fno-reorder-blocks -fno-schedule-insns, so do not use -O2 52 # clang does not support -fno-reorder-blocks -fno-schedule-insns, so do not use -O2
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_snprintf.c | 35 extern const void schedule __ksym; 58 &schedule, &schedule, &schedule); in handler()
|
/openbmc/linux/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_calendar.c | 41 u32 schedule[SPX5_DSM_CAL_LEN]; member 297 data->schedule[idx] = SPX5_DSM_CAL_EMPTY; in sparx5_dsm_calendar_calc() 301 data->schedule[0] = SPX5_DSM_CAL_MAX_DEVS_PER_TAXI; in sparx5_dsm_calendar_calc() 390 data->schedule[idx] = SPX5_DSM_CAL_MAX_DEVS_PER_TAXI; in sparx5_dsm_calendar_calc() 413 if (data->schedule[slot] == SPX5_DSM_CAL_EMPTY) in sparx5_dsm_calendar_calc() 423 memcpy(data->short_list, data->schedule, in sparx5_dsm_calendar_calc() 432 memcpy(data->long_list, data->schedule, in sparx5_dsm_calendar_calc() 469 data->schedule[slot] = data->temp_sched[slot]; in sparx5_dsm_calendar_calc() 483 u32 cal_length = sparx5_dsm_cal_len(data->schedule); in sparx5_dsm_calendar_check() 494 if (data->schedule[idx] == port) { in sparx5_dsm_calendar_check() [all …]
|
/openbmc/linux/drivers/usb/host/ |
H A D | uhci-hcd.h | 35 * the schedule is buggy */ 132 * can easily splice a QH for some endpoint into the schedule at the right 135 * In the schedule, qh->link maintains a list of QHs seen by the HC: 140 * state are on the software list but not the hardware schedule. 143 * but they never get added to the hardware schedule. 147 * schedule but the hardware may 149 #define QH_STATE_ACTIVE 3 /* QH is on the schedule */ 153 __hc32 link; /* Next QH in the schedule */ 290 * automatic queuing. To make it easy to insert entries into the schedule, 297 * appropriate skeleton QH. For instance, the schedule can look like this: [all …]
|
/openbmc/linux/Documentation/scheduler/ |
H A D | sched-arch.rst | 27 be enabled to call schedule() then disabled again. 30 be cleared until the running task has called schedule(). Idle 35 schedule(). It should not call schedule() otherwise.
|
/openbmc/linux/tools/perf/pmu-events/arch/x86/tigerlake/ |
H A D | other.json | 3 … where the core was running in a manner where Turbo may be clipped to the Non-AVX turbo schedule.", 11 …les where the core was running in a manner where Turbo may be clipped to the AVX2 turbo schedule.", 19 …s where the core was running in a manner where Turbo may be clipped to the AVX512 turbo schedule.",
|
/openbmc/linux/tools/perf/tests/ |
H A D | hists_cumulate.c | 27 /* perf [kernel] schedule() */ 54 /* schedule => run_command => main */ 277 * 10.00% perf [kernel] [k] schedule in test1() 288 { 0, 1000, "perf", "[kernel]", "schedule" }, in test1() 353 * 10.00% perf [kernel] [k] schedule in test2() 355 * --- schedule in test2() 386 { 0, 1000, "perf", "[kernel]", "schedule" }, in test2() 418 3, { { "[kernel]", "schedule" }, in test2() 480 * 10.00% 10.00% perf [kernel] [k] schedule in test3() 493 { 1000, 1000, "perf", "[kernel]", "schedule" }, in test3() [all …]
|
H A D | hists_common.h | 49 * perf: 100 [kernel] schedule 58 * perf: 200 [kernel] schedule 67 * bash: 300 [kernel] schedule
|
/openbmc/qemu/include/hw/usb/ |
H A D | ehci-regs.h | 18 #define USBCMD_PSE (1 << 4) // Periodic Schedule Enable 19 #define USBCMD_ASE (1 << 5) // Asynch Schedule Enable 37 #define USBSTS_PSS (1 << 14) // Periodic Schedule Status 38 #define USBSTS_ASS (1 << 15) // Asynchronous Schedule Status
|
/openbmc/linux/arch/powerpc/lib/ |
H A D | vmx-helper.c | 20 * We need to disable page faults as they can call schedule and in enter_vmx_usercopy() 41 * Must never explicitly call schedule (including preempt_enable()) in exit_vmx_usercopy() 45 * set and we are preemptible. The hack here is to schedule a in exit_vmx_usercopy()
|
/openbmc/linux/include/linux/ |
H A D | wait.h | 116 * schedule(); 330 schedule()) 369 0, 0, schedule()) 445 cmd1; schedule(); cmd2) 458 cmd1; schedule(); cmd2) 483 schedule()) 562 schedule()); \ 623 schedule()) 636 schedule()) 650 schedule()) [all …]
|
/openbmc/linux/net/atm/ |
H A D | svc.c | 64 schedule(); in svc_disconnect() 140 schedule(); in svc_bind() 221 schedule(); in svc_connect() 244 schedule(); in svc_connect() 251 schedule(); in svc_connect() 311 schedule(); in svc_listen() 363 schedule(); in svc_accept() 405 schedule(); in svc_accept() 450 schedule(); in svc_change_qos() 552 schedule(); in svc_addparty() [all …]
|