8e7fbcbc22
It's been broken forever (i.e. it's not scheduling in a power aware fashion), as reported by Suresh and others sending patches, and nobody cares enough to fix it properly ... so remove it to make space free for something better. There's various problems with the code as it stands today, first and foremost the user interface which is bound to topology levels and has multiple values per level. This results in a state explosion which the administrator or distro needs to master and almost nobody does. Furthermore large configuration state spaces aren't good, it means the thing doesn't just work right because it's either under so many impossibe to meet constraints, or even if there's an achievable state workloads have to be aware of it precisely and can never meet it for dynamic workloads. So pushing this kind of decision to user-space was a bad idea even with a single knob - it's exponentially worse with knobs on every node of the topology. There is a proposal to replace the user interface with a single 3 state knob: sched_balance_policy := { performance, power, auto } where 'auto' would be the preferred default which looks at things like Battery/AC mode and possible cpufreq state or whatever the hw exposes to show us power use expectations - but there's been no progress on it in the past many months. Aside from that, the actual implementation of the various knobs is known to be broken. There have been sporadic attempts at fixing things but these always stop short of reaching a mergable state. Therefore this wholesale removal with the hopes of spurring people who care to come forward once again and work on a coherent replacement. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Suresh Siddha <suresh.b.siddha@intel.com> Cc: Arjan van de Ven <arjan@linux.intel.com> Cc: Vincent Guittot <vincent.guittot@linaro.org> Cc: Vaidyanathan Srinivasan <svaidy@linux.vnet.ibm.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Andrew Morton <akpm@linux-foundation.org> Link: http://lkml.kernel.org/r/1326104915.2442.53.camel@twins Signed-off-by: Ingo Molnar <mingo@kernel.org>
178 lines
6.1 KiB
Text
178 lines
6.1 KiB
Text
What: /sys/devices/system/cpu/
|
|
Date: pre-git history
|
|
Contact: Linux kernel mailing list <linux-kernel@vger.kernel.org>
|
|
Description:
|
|
A collection of both global and individual CPU attributes
|
|
|
|
Individual CPU attributes are contained in subdirectories
|
|
named by the kernel's logical CPU number, e.g.:
|
|
|
|
/sys/devices/system/cpu/cpu#/
|
|
|
|
What: /sys/devices/system/cpu/kernel_max
|
|
/sys/devices/system/cpu/offline
|
|
/sys/devices/system/cpu/online
|
|
/sys/devices/system/cpu/possible
|
|
/sys/devices/system/cpu/present
|
|
Date: December 2008
|
|
Contact: Linux kernel mailing list <linux-kernel@vger.kernel.org>
|
|
Description: CPU topology files that describe kernel limits related to
|
|
hotplug. Briefly:
|
|
|
|
kernel_max: the maximum cpu index allowed by the kernel
|
|
configuration.
|
|
|
|
offline: cpus that are not online because they have been
|
|
HOTPLUGGED off or exceed the limit of cpus allowed by the
|
|
kernel configuration (kernel_max above).
|
|
|
|
online: cpus that are online and being scheduled.
|
|
|
|
possible: cpus that have been allocated resources and can be
|
|
brought online if they are present.
|
|
|
|
present: cpus that have been identified as being present in
|
|
the system.
|
|
|
|
See Documentation/cputopology.txt for more information.
|
|
|
|
|
|
What: /sys/devices/system/cpu/probe
|
|
/sys/devices/system/cpu/release
|
|
Date: November 2009
|
|
Contact: Linux kernel mailing list <linux-kernel@vger.kernel.org>
|
|
Description: Dynamic addition and removal of CPU's. This is not hotplug
|
|
removal, this is meant complete removal/addition of the CPU
|
|
from the system.
|
|
|
|
probe: writes to this file will dynamically add a CPU to the
|
|
system. Information written to the file to add CPU's is
|
|
architecture specific.
|
|
|
|
release: writes to this file dynamically remove a CPU from
|
|
the system. Information writtento the file to remove CPU's
|
|
is architecture specific.
|
|
|
|
What: /sys/devices/system/cpu/cpu#/node
|
|
Date: October 2009
|
|
Contact: Linux memory management mailing list <linux-mm@kvack.org>
|
|
Description: Discover NUMA node a CPU belongs to
|
|
|
|
When CONFIG_NUMA is enabled, a symbolic link that points
|
|
to the corresponding NUMA node directory.
|
|
|
|
For example, the following symlink is created for cpu42
|
|
in NUMA node 2:
|
|
|
|
/sys/devices/system/cpu/cpu42/node2 -> ../../node/node2
|
|
|
|
|
|
What: /sys/devices/system/cpu/cpu#/node
|
|
Date: October 2009
|
|
Contact: Linux memory management mailing list <linux-mm@kvack.org>
|
|
Description: Discover NUMA node a CPU belongs to
|
|
|
|
When CONFIG_NUMA is enabled, a symbolic link that points
|
|
to the corresponding NUMA node directory.
|
|
|
|
For example, the following symlink is created for cpu42
|
|
in NUMA node 2:
|
|
|
|
/sys/devices/system/cpu/cpu42/node2 -> ../../node/node2
|
|
|
|
|
|
What: /sys/devices/system/cpu/cpu#/topology/core_id
|
|
/sys/devices/system/cpu/cpu#/topology/core_siblings
|
|
/sys/devices/system/cpu/cpu#/topology/core_siblings_list
|
|
/sys/devices/system/cpu/cpu#/topology/physical_package_id
|
|
/sys/devices/system/cpu/cpu#/topology/thread_siblings
|
|
/sys/devices/system/cpu/cpu#/topology/thread_siblings_list
|
|
Date: December 2008
|
|
Contact: Linux kernel mailing list <linux-kernel@vger.kernel.org>
|
|
Description: CPU topology files that describe a logical CPU's relationship
|
|
to other cores and threads in the same physical package.
|
|
|
|
One cpu# directory is created per logical CPU in the system,
|
|
e.g. /sys/devices/system/cpu/cpu42/.
|
|
|
|
Briefly, the files above are:
|
|
|
|
core_id: the CPU core ID of cpu#. Typically it is the
|
|
hardware platform's identifier (rather than the kernel's).
|
|
The actual value is architecture and platform dependent.
|
|
|
|
core_siblings: internal kernel map of cpu#'s hardware threads
|
|
within the same physical_package_id.
|
|
|
|
core_siblings_list: human-readable list of the logical CPU
|
|
numbers within the same physical_package_id as cpu#.
|
|
|
|
physical_package_id: physical package id of cpu#. Typically
|
|
corresponds to a physical socket number, but the actual value
|
|
is architecture and platform dependent.
|
|
|
|
thread_siblings: internel kernel map of cpu#'s hardware
|
|
threads within the same core as cpu#
|
|
|
|
thread_siblings_list: human-readable list of cpu#'s hardware
|
|
threads within the same core as cpu#
|
|
|
|
See Documentation/cputopology.txt for more information.
|
|
|
|
|
|
What: /sys/devices/system/cpu/cpuidle/current_driver
|
|
/sys/devices/system/cpu/cpuidle/current_governer_ro
|
|
Date: September 2007
|
|
Contact: Linux kernel mailing list <linux-kernel@vger.kernel.org>
|
|
Description: Discover cpuidle policy and mechanism
|
|
|
|
Various CPUs today support multiple idle levels that are
|
|
differentiated by varying exit latencies and power
|
|
consumption during idle.
|
|
|
|
Idle policy (governor) is differentiated from idle mechanism
|
|
(driver)
|
|
|
|
current_driver: displays current idle mechanism
|
|
|
|
current_governor_ro: displays current idle policy
|
|
|
|
See files in Documentation/cpuidle/ for more information.
|
|
|
|
|
|
What: /sys/devices/system/cpu/cpu#/cpufreq/*
|
|
Date: pre-git history
|
|
Contact: cpufreq@vger.kernel.org
|
|
Description: Discover and change clock speed of CPUs
|
|
|
|
Clock scaling allows you to change the clock speed of the
|
|
CPUs on the fly. This is a nice method to save battery
|
|
power, because the lower the clock speed, the less power
|
|
the CPU consumes.
|
|
|
|
There are many knobs to tweak in this directory.
|
|
|
|
See files in Documentation/cpu-freq/ for more information.
|
|
|
|
In particular, read Documentation/cpu-freq/user-guide.txt
|
|
to learn how to control the knobs.
|
|
|
|
|
|
What: /sys/devices/system/cpu/cpu*/cache/index3/cache_disable_{0,1}
|
|
Date: August 2008
|
|
KernelVersion: 2.6.27
|
|
Contact: discuss@x86-64.org
|
|
Description: Disable L3 cache indices
|
|
|
|
These files exist in every CPU's cache/index3 directory. Each
|
|
cache_disable_{0,1} file corresponds to one disable slot which
|
|
can be used to disable a cache index. Reading from these files
|
|
on a processor with this functionality will return the currently
|
|
disabled index for that node. There is one L3 structure per
|
|
node, or per internal node on MCM machines. Writing a valid
|
|
index to one of these files will cause the specificed cache
|
|
index to be disabled.
|
|
|
|
All AMD processors with L3 caches provide this functionality.
|
|
For details, see BKDGs at
|
|
http://developer.amd.com/documentation/guides/Pages/default.aspx
|