Formatting of Kconfig files doesn't look so pretty, so let the Great White Handkerchief come around and clean it up.
Signed-off-by: Enrico Weigelt, metux IT consult <i...@metux.net> --- kernel/irq/Kconfig | 26 +++---- kernel/power/Kconfig | 12 +-- kernel/rcu/Kconfig | 2 +- kernel/time/Kconfig | 20 ++--- kernel/trace/Kconfig | 205 +++++++++++++++++++++++++-------------------------- 5 files changed, 132 insertions(+), 133 deletions(-) diff --git a/kernel/irq/Kconfig b/kernel/irq/Kconfig index 5f3e2ba..c595a48 100644 --- a/kernel/irq/Kconfig +++ b/kernel/irq/Kconfig @@ -3,11 +3,11 @@ menu "IRQ subsystem" # Make sparse irq Kconfig switch below available config MAY_HAVE_SPARSE_IRQ - bool + bool # Legacy support, required for itanic config GENERIC_IRQ_LEGACY - bool + bool # Enable the generic irq autoprobe mechanism config GENERIC_IRQ_PROBE @@ -15,20 +15,20 @@ config GENERIC_IRQ_PROBE # Use the generic /proc/interrupts implementation config GENERIC_IRQ_SHOW - bool + bool # Print level/edge extra information config GENERIC_IRQ_SHOW_LEVEL - bool + bool # Supports effective affinity mask config GENERIC_IRQ_EFFECTIVE_AFF_MASK - bool + bool # Facility to allocate a hardware interrupt. This is legacy support # and should not be used in new code. Use irq domains instead. config GENERIC_IRQ_LEGACY_ALLOC_HWIRQ - bool + bool # Support for delayed migration from interrupt context config GENERIC_PENDING_IRQ @@ -40,24 +40,24 @@ config GENERIC_IRQ_MIGRATION # Alpha specific irq affinity mechanism config AUTO_IRQ_AFFINITY - bool + bool # Tasklet based software resend for pending interrupts on enable_irq() config HARDIRQS_SW_RESEND - bool + bool # Preflow handler support for fasteoi (sparc64) config IRQ_PREFLOW_FASTEOI - bool + bool # Edge style eoi based handler (cell) config IRQ_EDGE_EOI_HANDLER - bool + bool # Generic configurable interrupt chip implementation config GENERIC_IRQ_CHIP - bool - select IRQ_DOMAIN + bool + select IRQ_DOMAIN # Generic irq_domain hw <--> linux irq number translation config IRQ_DOMAIN @@ -105,7 +105,7 @@ config GENERIC_IRQ_RESERVATION_MODE # Support forced irq threading config IRQ_FORCED_THREADING - bool + bool config SPARSE_IRQ bool "Support sparse irq numbering" if MAY_HAVE_SPARSE_IRQ diff --git a/kernel/power/Kconfig b/kernel/power/Kconfig index f8fe57d..a646337 100644 --- a/kernel/power/Kconfig +++ b/kernel/power/Kconfig @@ -85,20 +85,20 @@ config PM_STD_PARTITION default "" ---help--- The default resume partition is the partition that the suspend- - to-disk implementation will look for a suspended disk image. + to-disk implementation will look for a suspended disk image. - The partition specified here will be different for almost every user. + The partition specified here will be different for almost every user. It should be a valid swap partition (at least for now) that is turned - on before suspending. + on before suspending. The partition specified can be overridden by specifying: - resume=/dev/<other device> + resume=/dev/<other device> - which will set the resume partition to the device specified. + which will set the resume partition to the device specified. Note there is currently not a way to specify which device to save the - suspended image to. It will simply pick the first available swap + suspended image to. It will simply pick the first available swap device. config PM_SLEEP diff --git a/kernel/rcu/Kconfig b/kernel/rcu/Kconfig index 3730143..2f213fe 100644 --- a/kernel/rcu/Kconfig +++ b/kernel/rcu/Kconfig @@ -154,7 +154,7 @@ config RCU_FAST_NO_HZ for example, slowing down synchronize_rcu(). Say Y if energy efficiency is critically important, and you - don't care about increased grace-period durations. + don't care about increased grace-period durations. Say N if you are unsure. diff --git a/kernel/time/Kconfig b/kernel/time/Kconfig index e2c038d..8e8e263 100644 --- a/kernel/time/Kconfig +++ b/kernel/time/Kconfig @@ -101,19 +101,19 @@ config NO_HZ_FULL select IRQ_WORK select CPU_ISOLATION help - Adaptively try to shutdown the tick whenever possible, even when - the CPU is running tasks. Typically this requires running a single - task on the CPU. Chances for running tickless are maximized when - the task mostly runs in userspace and has few kernel activity. + Adaptively try to shutdown the tick whenever possible, even when + the CPU is running tasks. Typically this requires running a single + task on the CPU. Chances for running tickless are maximized when + the task mostly runs in userspace and has few kernel activity. - You need to fill up the nohz_full boot parameter with the - desired range of dynticks CPUs. + You need to fill up the nohz_full boot parameter with the + desired range of dynticks CPUs. - This is implemented at the expense of some overhead in user <-> kernel - transitions: syscalls, exceptions and interrupts. Even when it's - dynamically off. + This is implemented at the expense of some overhead in user <-> kernel + transitions: syscalls, exceptions and interrupts. Even when it's + dynamically off. - Say N. + Say N. endchoice diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig index fa8b1fe..05f29e2 100644 --- a/kernel/trace/Kconfig +++ b/kernel/trace/Kconfig @@ -69,13 +69,13 @@ config RING_BUFFER select IRQ_WORK config FTRACE_NMI_ENTER - bool - depends on HAVE_FTRACE_NMI_ENTER - default y + bool + depends on HAVE_FTRACE_NMI_ENTER + default y config EVENT_TRACING select CONTEXT_SWITCH_TRACER - select GLOB + select GLOB bool config CONTEXT_SWITCH_TRACER @@ -84,8 +84,8 @@ config CONTEXT_SWITCH_TRACER config RING_BUFFER_ALLOW_SWAP bool help - Allow the use of ring_buffer_swap_cpu. - Adds a very slight overhead to tracing when enabled. + Allow the use of ring_buffer_swap_cpu. + Adds a very slight overhead to tracing when enabled. config PREEMPTIRQ_TRACEPOINTS bool @@ -248,36 +248,36 @@ config HWLAT_TRACER bool "Tracer to detect hardware latencies (like SMIs)" select GENERIC_TRACER help - This tracer, when enabled will create one or more kernel threads, - depending on what the cpumask file is set to, which each thread - spinning in a loop looking for interruptions caused by - something other than the kernel. For example, if a - System Management Interrupt (SMI) takes a noticeable amount of - time, this tracer will detect it. This is useful for testing - if a system is reliable for Real Time tasks. + This tracer, when enabled will create one or more kernel threads, + depending on what the cpumask file is set to, which each thread + spinning in a loop looking for interruptions caused by + something other than the kernel. For example, if a + System Management Interrupt (SMI) takes a noticeable amount of + time, this tracer will detect it. This is useful for testing + if a system is reliable for Real Time tasks. - Some files are created in the tracing directory when this - is enabled: + Some files are created in the tracing directory when this + is enabled: hwlat_detector/width - time in usecs for how long to spin for hwlat_detector/window - time in usecs between the start of each iteration - A kernel thread is created that will spin with interrupts disabled - for "width" microseconds in every "window" cycle. It will not spin - for "window - width" microseconds, where the system can - continue to operate. + A kernel thread is created that will spin with interrupts disabled + for "width" microseconds in every "window" cycle. It will not spin + for "window - width" microseconds, where the system can + continue to operate. - The output will appear in the trace and trace_pipe files. + The output will appear in the trace and trace_pipe files. - When the tracer is not running, it has no affect on the system, - but when it is running, it can cause the system to be - periodically non responsive. Do not run this tracer on a - production system. + When the tracer is not running, it has no affect on the system, + but when it is running, it can cause the system to be + periodically non responsive. Do not run this tracer on a + production system. - To enable this tracer, echo in "hwlat" into the current_tracer - file. Every time a latency is greater than tracing_thresh, it will - be recorded into the ring buffer. + To enable this tracer, echo in "hwlat" into the current_tracer + file. Every time a latency is greater than tracing_thresh, it will + be recorded into the ring buffer. config ENABLE_DEFAULT_TRACERS bool "Trace process context switches and events" @@ -307,7 +307,7 @@ config TRACER_SNAPSHOT cat snapshot config TRACER_SNAPSHOT_PER_CPU_SWAP - bool "Allow snapshot to swap per CPU" + bool "Allow snapshot to swap per CPU" depends on TRACER_SNAPSHOT select RING_BUFFER_ALLOW_SWAP help @@ -335,18 +335,18 @@ choice prompt "Branch Profiling" default BRANCH_PROFILE_NONE help - The branch profiling is a software profiler. It will add hooks - into the C conditionals to test which path a branch takes. + The branch profiling is a software profiler. It will add hooks + into the C conditionals to test which path a branch takes. - The likely/unlikely profiler only looks at the conditions that - are annotated with a likely or unlikely macro. + The likely/unlikely profiler only looks at the conditions that + are annotated with a likely or unlikely macro. - The "all branch" profiler will profile every if-statement in the - kernel. This profiler will also enable the likely/unlikely - profiler. + The "all branch" profiler will profile every if-statement in the + kernel. This profiler will also enable the likely/unlikely + profiler. - Either of the above profilers adds a bit of overhead to the system. - If unsure, choose "No branch profiling". + Either of the above profilers adds a bit of overhead to the system. + If unsure, choose "No branch profiling". config BRANCH_PROFILE_NONE bool "No branch profiling" @@ -574,8 +574,8 @@ config BPF_KPROBE_OVERRIDE depends on FUNCTION_ERROR_INJECTION default n help - Allows BPF to override the execution of a probed function and - set a different return value. This is used for error injection. + Allows BPF to override the execution of a probed function and + set a different return value. This is used for error injection. config FTRACE_MCOUNT_RECORD def_bool y @@ -599,13 +599,13 @@ config EVENT_TRACE_TEST_SYSCALLS bool "Run selftest on syscall events" depends on FTRACE_STARTUP_TEST help - This option will also enable testing every syscall event. - It only enables the event and disables it and runs various loads - with the event enabled. This adds a bit more time for kernel boot - up since it runs this on every system call defined. + This option will also enable testing every syscall event. + It only enables the event and disables it and runs various loads + with the event enabled. This adds a bit more time for kernel boot + up since it runs this on every system call defined. - TBD - enable a way to actually call the syscalls as we test their - events + TBD - enable a way to actually call the syscalls as we test their + events config MMIOTRACE bool "Memory mapped IO tracing" @@ -662,24 +662,24 @@ config MMIOTRACE_TEST Say N, unless you absolutely know what you are doing. config TRACEPOINT_BENCHMARK - bool "Add tracepoint that benchmarks tracepoints" + bool "Add tracepoint that benchmarks tracepoints" help - This option creates the tracepoint "benchmark:benchmark_event". - When the tracepoint is enabled, it kicks off a kernel thread that - goes into an infinite loop (calling cond_sched() to let other tasks - run), and calls the tracepoint. Each iteration will record the time - it took to write to the tracepoint and the next iteration that - data will be passed to the tracepoint itself. That is, the tracepoint - will report the time it took to do the previous tracepoint. - The string written to the tracepoint is a static string of 128 bytes - to keep the time the same. The initial string is simply a write of - "START". The second string records the cold cache time of the first - write which is not added to the rest of the calculations. + This option creates the tracepoint "benchmark:benchmark_event". + When the tracepoint is enabled, it kicks off a kernel thread that + goes into an infinite loop (calling cond_sched() to let other tasks + run), and calls the tracepoint. Each iteration will record the time + it took to write to the tracepoint and the next iteration that + data will be passed to the tracepoint itself. That is, the tracepoint + will report the time it took to do the previous tracepoint. + The string written to the tracepoint is a static string of 128 bytes + to keep the time the same. The initial string is simply a write of + "START". The second string records the cold cache time of the first + write which is not added to the rest of the calculations. - As it is a tight loop, it benchmarks as hot cache. That's fine because - we care most about hot paths that are probably in cache already. + As it is a tight loop, it benchmarks as hot cache. That's fine because + we care most about hot paths that are probably in cache already. - An example of the output: + An example of the output: START first=3672 [COLD CACHED] @@ -708,27 +708,27 @@ config RING_BUFFER_BENCHMARK If unsure, say N. config RING_BUFFER_STARTUP_TEST - bool "Ring buffer startup self test" - depends on RING_BUFFER - help - Run a simple self test on the ring buffer on boot up. Late in the - kernel boot sequence, the test will start that kicks off - a thread per cpu. Each thread will write various size events - into the ring buffer. Another thread is created to send IPIs - to each of the threads, where the IPI handler will also write - to the ring buffer, to test/stress the nesting ability. - If any anomalies are discovered, a warning will be displayed - and all ring buffers will be disabled. - - The test runs for 10 seconds. This will slow your boot time - by at least 10 more seconds. - - At the end of the test, statics and more checks are done. - It will output the stats of each per cpu buffer. What - was written, the sizes, what was read, what was lost, and - other similar details. - - If unsure, say N + bool "Ring buffer startup self test" + depends on RING_BUFFER + help + Run a simple self test on the ring buffer on boot up. Late in the + kernel boot sequence, the test will start that kicks off + a thread per cpu. Each thread will write various size events + into the ring buffer. Another thread is created to send IPIs + to each of the threads, where the IPI handler will also write + to the ring buffer, to test/stress the nesting ability. + If any anomalies are discovered, a warning will be displayed + and all ring buffers will be disabled. + + The test runs for 10 seconds. This will slow your boot time + by at least 10 more seconds. + + At the end of the test, statics and more checks are done. + It will output the stats of each per cpu buffer. What + was written, the sizes, what was read, what was lost, and + other similar details. + + If unsure, say N config PREEMPTIRQ_DELAY_TEST tristate "Preempt / IRQ disable delay thread to test latency tracers" @@ -746,32 +746,32 @@ config PREEMPTIRQ_DELAY_TEST If unsure, say N config TRACE_EVAL_MAP_FILE - bool "Show eval mappings for trace events" - depends on TRACING - help - The "print fmt" of the trace events will show the enum/sizeof names - instead of their values. This can cause problems for user space tools - that use this string to parse the raw data as user space does not know - how to convert the string to its value. + bool "Show eval mappings for trace events" + depends on TRACING + help + The "print fmt" of the trace events will show the enum/sizeof names + instead of their values. This can cause problems for user space tools + that use this string to parse the raw data as user space does not know + how to convert the string to its value. - To fix this, there's a special macro in the kernel that can be used - to convert an enum/sizeof into its value. If this macro is used, then - the print fmt strings will be converted to their values. + To fix this, there's a special macro in the kernel that can be used + to convert an enum/sizeof into its value. If this macro is used, then + the print fmt strings will be converted to their values. - If something does not get converted properly, this option can be - used to show what enums/sizeof the kernel tried to convert. + If something does not get converted properly, this option can be + used to show what enums/sizeof the kernel tried to convert. - This option is for debugging the conversions. A file is created - in the tracing directory called "eval_map" that will show the - names matched with their values and what trace event system they - belong too. + This option is for debugging the conversions. A file is created + in the tracing directory called "eval_map" that will show the + names matched with their values and what trace event system they + belong too. - Normally, the mapping of the strings to values will be freed after - boot up or module load. With this option, they will not be freed, as - they are needed for the "eval_map" file. Enabling this option will - increase the memory footprint of the running kernel. + Normally, the mapping of the strings to values will be freed after + boot up or module load. With this option, they will not be freed, as + they are needed for the "eval_map" file. Enabling this option will + increase the memory footprint of the running kernel. - If unsure, say N + If unsure, say N config TRACING_EVENTS_GPIO bool "Trace gpio events" @@ -795,4 +795,3 @@ config GCOV_PROFILE_FTRACE endif # FTRACE endif # TRACING_SUPPORT - -- 1.9.1