|
@@ -304,19 +304,6 @@ static inline void lockup_detector_init(void)
|
|
|
}
|
|
|
#endif
|
|
|
|
|
|
-#ifdef CONFIG_DETECT_HUNG_TASK
|
|
|
-extern unsigned int sysctl_hung_task_panic;
|
|
|
-extern unsigned long sysctl_hung_task_check_count;
|
|
|
-extern unsigned long sysctl_hung_task_timeout_secs;
|
|
|
-extern unsigned long sysctl_hung_task_warnings;
|
|
|
-extern int proc_dohung_task_timeout_secs(struct ctl_table *table, int write,
|
|
|
- void __user *buffer,
|
|
|
- size_t *lenp, loff_t *ppos);
|
|
|
-#else
|
|
|
-/* Avoid need for ifdefs elsewhere in the code */
|
|
|
-enum { sysctl_hung_task_timeout_secs = 0 };
|
|
|
-#endif
|
|
|
-
|
|
|
/* Attach to any functions which should be ignored in wchan output. */
|
|
|
#define __sched __attribute__((__section__(".sched.text")))
|
|
|
|
|
@@ -338,23 +325,6 @@ extern int mutex_spin_on_owner(struct mutex *lock, struct task_struct *owner);
|
|
|
struct nsproxy;
|
|
|
struct user_namespace;
|
|
|
|
|
|
-/*
|
|
|
- * Default maximum number of active map areas, this limits the number of vmas
|
|
|
- * per mm struct. Users can overwrite this number by sysctl but there is a
|
|
|
- * problem.
|
|
|
- *
|
|
|
- * When a program's coredump is generated as ELF format, a section is created
|
|
|
- * per a vma. In ELF, the number of sections is represented in unsigned short.
|
|
|
- * This means the number of sections should be smaller than 65535 at coredump.
|
|
|
- * Because the kernel adds some informative sections to a image of program at
|
|
|
- * generating coredump, we need some margin. The number of extra sections is
|
|
|
- * 1-3 now and depends on arch. We use "5" as safe margin, here.
|
|
|
- */
|
|
|
-#define MAPCOUNT_ELF_CORE_MARGIN (5)
|
|
|
-#define DEFAULT_MAX_MAP_COUNT (USHRT_MAX - MAPCOUNT_ELF_CORE_MARGIN)
|
|
|
-
|
|
|
-extern int sysctl_max_map_count;
|
|
|
-
|
|
|
#include <linux/aio.h>
|
|
|
|
|
|
#ifdef CONFIG_MMU
|
|
@@ -1221,12 +1191,6 @@ struct sched_rt_entity {
|
|
|
#endif
|
|
|
};
|
|
|
|
|
|
-/*
|
|
|
- * default timeslice is 100 msecs (used only for SCHED_RR tasks).
|
|
|
- * Timeslices get refilled after they expire.
|
|
|
- */
|
|
|
-#define RR_TIMESLICE (100 * HZ / 1000)
|
|
|
-
|
|
|
struct rcu_node;
|
|
|
|
|
|
enum perf_event_task_context {
|
|
@@ -2074,58 +2038,7 @@ extern void wake_up_idle_cpu(int cpu);
|
|
|
static inline void wake_up_idle_cpu(int cpu) { }
|
|
|
#endif
|
|
|
|
|
|
-extern unsigned int sysctl_sched_latency;
|
|
|
-extern unsigned int sysctl_sched_min_granularity;
|
|
|
-extern unsigned int sysctl_sched_wakeup_granularity;
|
|
|
-extern unsigned int sysctl_sched_child_runs_first;
|
|
|
-
|
|
|
-enum sched_tunable_scaling {
|
|
|
- SCHED_TUNABLESCALING_NONE,
|
|
|
- SCHED_TUNABLESCALING_LOG,
|
|
|
- SCHED_TUNABLESCALING_LINEAR,
|
|
|
- SCHED_TUNABLESCALING_END,
|
|
|
-};
|
|
|
-extern enum sched_tunable_scaling sysctl_sched_tunable_scaling;
|
|
|
-
|
|
|
-extern unsigned int sysctl_numa_balancing_scan_delay;
|
|
|
-extern unsigned int sysctl_numa_balancing_scan_period_min;
|
|
|
-extern unsigned int sysctl_numa_balancing_scan_period_max;
|
|
|
-extern unsigned int sysctl_numa_balancing_scan_period_reset;
|
|
|
-extern unsigned int sysctl_numa_balancing_scan_size;
|
|
|
-extern unsigned int sysctl_numa_balancing_settle_count;
|
|
|
-
|
|
|
-#ifdef CONFIG_SCHED_DEBUG
|
|
|
-extern unsigned int sysctl_sched_migration_cost;
|
|
|
-extern unsigned int sysctl_sched_nr_migrate;
|
|
|
-extern unsigned int sysctl_sched_time_avg;
|
|
|
-extern unsigned int sysctl_timer_migration;
|
|
|
-extern unsigned int sysctl_sched_shares_window;
|
|
|
-
|
|
|
-int sched_proc_update_handler(struct ctl_table *table, int write,
|
|
|
- void __user *buffer, size_t *length,
|
|
|
- loff_t *ppos);
|
|
|
-#endif
|
|
|
-#ifdef CONFIG_SCHED_DEBUG
|
|
|
-static inline unsigned int get_sysctl_timer_migration(void)
|
|
|
-{
|
|
|
- return sysctl_timer_migration;
|
|
|
-}
|
|
|
-#else
|
|
|
-static inline unsigned int get_sysctl_timer_migration(void)
|
|
|
-{
|
|
|
- return 1;
|
|
|
-}
|
|
|
-#endif
|
|
|
-extern unsigned int sysctl_sched_rt_period;
|
|
|
-extern int sysctl_sched_rt_runtime;
|
|
|
-
|
|
|
-int sched_rt_handler(struct ctl_table *table, int write,
|
|
|
- void __user *buffer, size_t *lenp,
|
|
|
- loff_t *ppos);
|
|
|
-
|
|
|
#ifdef CONFIG_SCHED_AUTOGROUP
|
|
|
-extern unsigned int sysctl_sched_autogroup_enabled;
|
|
|
-
|
|
|
extern void sched_autogroup_create_attach(struct task_struct *p);
|
|
|
extern void sched_autogroup_detach(struct task_struct *p);
|
|
|
extern void sched_autogroup_fork(struct signal_struct *sig);
|
|
@@ -2141,10 +2054,6 @@ static inline void sched_autogroup_fork(struct signal_struct *sig) { }
|
|
|
static inline void sched_autogroup_exit(struct signal_struct *sig) { }
|
|
|
#endif
|
|
|
|
|
|
-#ifdef CONFIG_CFS_BANDWIDTH
|
|
|
-extern unsigned int sysctl_sched_cfs_bandwidth_slice;
|
|
|
-#endif
|
|
|
-
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
extern int rt_mutex_getprio(struct task_struct *p);
|
|
|
extern void rt_mutex_setprio(struct task_struct *p, int prio);
|