/linux/tools/testing/selftests/cgroup/ |
H A D | memcg_protection.m | 6 % hierarchy to illustrate how overcommitted protection spreads among siblings 9 % Simulation assumes siblings consumed the initial amount of memory (w/out 16 % n vector nominal protection of siblings set at the given level (memory.low) 48 siblings = sum(u); variable 52 e = protected * min(1, E / siblings); % normalize overcommit 55 unclaimed = max(0, E - siblings); 56 parent_overuse = sum(c) - siblings; 79 % XXX here I do parallel reclaim of all siblings
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | ws.c | 133 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node() 139 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node() 214 list_del(&tc_node->siblings); in irdma_remove_leaf() 219 list_del(&vsi_node->siblings); in irdma_remove_leaf() 295 list_add(&vsi_node->siblings, &ws_tree_root->child_list_head); in irdma_ws_add() 322 list_add(&tc_node->siblings, &vsi_node->child_list_head); in irdma_ws_add() 356 list_del(&tc_node->siblings); in irdma_ws_add() 362 list_del(&vsi_node->siblings); in irdma_ws_add()
|
H A D | ws.h | 19 struct list_head siblings; member
|
/linux/drivers/gpu/drm/i915/gt/uc/ |
H A D | selftest_guc_multi_lrc.c | 34 struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; in multi_lrc_create_parent() local 43 siblings[i++] = engine; in multi_lrc_create_parent() 49 logical_sort(siblings, i); in multi_lrc_create_parent() 51 return intel_engine_create_parallel(siblings, 1, i); in multi_lrc_create_parent()
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.c | 444 siblings = kmalloc_array(num_siblings, sizeof(*siblings), GFP_KERNEL); in set_proto_ctx_engines_balance() 445 if (!siblings) in set_proto_ctx_engines_balance() 459 if (!siblings[n]) { in set_proto_ctx_engines_balance() 471 kfree(siblings); in set_proto_ctx_engines_balance() 475 set->engines[idx].siblings = siblings; in set_proto_ctx_engines_balance() 481 kfree(siblings); in set_proto_ctx_engines_balance() 649 if (!siblings) in set_proto_ctx_engines_parallel_submit() 665 siblings[n] = in set_proto_ctx_engines_parallel_submit() 668 if (!siblings[n]) { in set_proto_ctx_engines_parallel_submit() 717 set->engines[slot].siblings = siblings; in set_proto_ctx_engines_parallel_submit() [all …]
|
H A D | i915_gem_context_types.h | 122 struct intel_engine_cs **siblings; member
|
/linux/drivers/gpu/drm/i915/gt/ |
H A D | selftest_execlists.c | 3712 struct intel_engine_cs **siblings, in nop_virtual_engine() argument 3840 struct intel_engine_cs **siblings, in __select_siblings() argument 3862 struct intel_engine_cs **siblings) in select_siblings() argument 4022 struct intel_engine_cs **siblings, in slicein_virtual_engine() argument 4040 ce = intel_context_create(siblings[n]); in slicein_virtual_engine() 4089 struct intel_engine_cs **siblings, in sliceout_virtual_engine() argument 4125 ce = intel_context_create(siblings[n]); in sliceout_virtual_engine() 4142 __func__, siblings[n]->name); in sliceout_virtual_engine() 4189 struct intel_engine_cs **siblings, in preserved_virtual_engine() argument 4331 struct intel_engine_cs **siblings, in reset_virtual_engine() argument [all …]
|
H A D | intel_execlists_submission.c | 198 struct intel_engine_cs *siblings[]; member 1054 if (likely(engine == ve->siblings[0])) in virtual_xfer_context() 1068 if (ve->siblings[n] == engine) { in virtual_xfer_context() 1069 swap(ve->siblings[n], ve->siblings[0]); in virtual_xfer_context() 3709 swap(ve->siblings[swp], ve->siblings[0]); in virtual_engine_initial_hint() 3716 return lrc_alloc(ce, ve->siblings[0]); in virtual_context_alloc() 3742 intel_engine_pm_get(ve->siblings[n]); in virtual_context_enter() 3755 intel_engine_pm_put(ve->siblings[n]); in virtual_context_exit() 3766 return ve->siblings[sibling]; in virtual_get_sibling() 3803 mask = ve->siblings[0]->mask; in virtual_submission_mask() [all …]
|
/linux/drivers/gpio/ |
H A D | gpio-sim.c | 623 struct list_head siblings; member 657 struct list_head siblings; member 801 list_for_each_entry(bank, &dev->bank_list, siblings) { in gpio_sim_add_hogs() 819 list_for_each_entry(bank, &dev->bank_list, siblings) { in gpio_sim_add_hogs() 919 list_for_each_entry(this, &dev->bank_list, siblings) { in gpio_sim_bank_labels_non_unique() 920 list_for_each_entry(pos, &dev->bank_list, siblings) { in gpio_sim_bank_labels_non_unique() 959 list_for_each_entry(bank, &dev->bank_list, siblings) { in gpio_sim_device_activate() 1365 list_del(&line->siblings); in gpio_sim_line_config_group_release() 1414 list_add_tail(&line->siblings, &bank->line_list); in gpio_sim_bank_config_make_line_group() 1425 list_del(&bank->siblings); in gpio_sim_bank_config_group_release() [all …]
|
/linux/Documentation/admin-guide/hw-vuln/ |
H A D | core-scheduling.rst | 100 siblings of a core such that all the selected tasks running on a core are 107 the sibling has the task enqueued. For rest of the siblings in the core, 112 Once a task has been selected for all the siblings in the core, an IPI is sent to 113 siblings for whom a new task was selected. Siblings on receiving the IPI will 125 siblings could be forced to select a lower priority task if the highest 157 and are considered system-wide trusted. The forced-idling of siblings running 174 the siblings to switch to the new task. But there could be hardware delays in 176 cause an attacker task to start running on a CPU before its siblings receive the 177 IPI. Even though cache is flushed on entry to user mode, victim tasks on siblings 185 Core scheduling cannot protect against MDS attacks between the siblings [all …]
|
/linux/arch/x86/kernel/cpu/microcode/ |
H A D | core.c | 534 unsigned int cpu, updated = 0, failed = 0, timedout = 0, siblings = 0; in load_late_stop_cpus() local 567 case UCODE_OK: siblings++; break; in load_late_stop_cpus() 595 pr_info("load: updated on %u primary CPUs with %u siblings\n", updated, siblings); in load_late_stop_cpus() 598 num_online_cpus() - (updated + siblings)); in load_late_stop_cpus() 603 return updated + siblings == num_online_cpus() ? 0 : -EIO; in load_late_stop_cpus()
|
/linux/drivers/nvme/host/ |
H A D | multipath.c | 216 list_for_each_entry_rcu(ns, &head->list, siblings) { in nvme_mpath_revalidate_paths() 249 list_for_each_entry_rcu(ns, &head->list, siblings) { in __nvme_find_path() 287 ns = list_next_or_null_rcu(&head->list, &ns->siblings, struct nvme_ns, in nvme_next_ns() 288 siblings); in nvme_next_ns() 291 return list_first_or_null_rcu(&head->list, struct nvme_ns, siblings); in nvme_next_ns() 363 list_for_each_entry_rcu(ns, &head->list, siblings) { in nvme_available_path()
|
/linux/Documentation/devicetree/bindings/net/dsa/ |
H A D | marvell,mv88e6xxx.yaml | 28 This switch and its siblings will be autodetected from 34 This switch and its siblings will be autodetected from 39 This switch and its siblings will be autodetected from
|
/linux/include/scsi/ |
H A D | scsi_device.h | 108 struct list_head siblings; /* list of all devices on this host */ member 337 struct list_head siblings; member 445 list_for_each_entry((sdev), &((shost)->__devices), siblings)
|
/linux/Documentation/devicetree/bindings/display/mediatek/ |
H A D | mediatek,od.yaml | 16 OD device node must be siblings to the central MMSYS_CONFIG node.
|
H A D | mediatek,ufoe.yaml | 17 UFOe device node must be siblings to the central MMSYS_CONFIG node.
|
H A D | mediatek,ccorr.yaml | 16 CCORR device node must be siblings to the central MMSYS_CONFIG node.
|
H A D | mediatek,dither.yaml | 17 DITHER device node must be siblings to the central MMSYS_CONFIG node.
|
H A D | mediatek,postmask.yaml | 16 POSTMASK device node must be siblings to the central MMSYS_CONFIG node.
|
H A D | mediatek,split.yaml | 16 SPLIT device node must be siblings to the central MMSYS_CONFIG node.
|
/linux/drivers/scsi/ |
H A D | scsi_scan.c | 303 INIT_LIST_HEAD(&sdev->siblings); in scsi_alloc_sdev() 402 list_del_init(&starget->siblings); in scsi_target_destroy() 435 list_for_each_entry(starget, &shost->__targets, siblings) { in __scsi_find_target() 520 INIT_LIST_HEAD(&starget->siblings); in scsi_alloc_target() 532 list_add_tail(&starget->siblings, &shost->__targets); in scsi_alloc_target() 2065 list_for_each_entry(sdev, &shost->__devices, siblings) { in scsi_forget_host()
|
H A D | scsi.c | 801 struct list_head *list = (prev ? &prev->siblings : &shost->__devices); in __scsi_iterate_devices() 807 next = list_entry(list->next, struct scsi_device, siblings); in __scsi_iterate_devices() 952 list_for_each_entry(sdev, &shost->__devices, siblings) { in __scsi_device_lookup()
|
/linux/kernel/ |
H A D | Kconfig.preempt | 123 selection across SMT siblings. When enabled -- see 124 prctl(PR_SCHED_CORE) -- task selection ensures that all SMT siblings
|
/linux/Documentation/admin-guide/pm/ |
H A D | intel_epb.rst | 40 example, SMT siblings or cores in one package). For this reason, updating the
|
/linux/drivers/scsi/libsas/ |
H A D | sas_expander.c | 856 list_add_tail(&child->siblings, &parent_ex->children); in sas_ex_discover_end_dev() 964 list_add_tail(&child->siblings, &parent->ex_dev.children); in sas_ex_discover_expander() 1092 list_for_each_entry(child, &ex->children, siblings) { in sas_check_level_subtractive_boundary() 1820 list_for_each_entry(ch, &ex->children, siblings) { in sas_find_bcast_dev() 1836 list_for_each_entry_safe(child, n, &ex->children, siblings) { in sas_unregister_ex_tree() 1854 &ex_dev->children, siblings) { in sas_unregister_devs_sas_addr() 1888 list_for_each_entry(child, &ex_root->children, siblings) { in sas_discover_bfs_by_root_level() 1939 list_for_each_entry(child, &dev->ex_dev.children, siblings) { in sas_discover_new()
|