--- zzzz-none-000/linux-2.6.39.4/include/linux/sched.h 2011-08-03 19:43:28.000000000 +0000 +++ puma6-atom-6490-729/linux-2.6.39.4/include/linux/sched.h 2021-11-10 13:38:17.000000000 +0000 @@ -157,14 +157,14 @@ print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq); #else static inline void -proc_sched_show_task(struct task_struct *p, struct seq_file *m) +proc_sched_show_task(struct task_struct *p __attribute__((unused)), struct seq_file *m __attribute__((unused))) { } -static inline void proc_sched_set_task(struct task_struct *p) +static inline void proc_sched_set_task(struct task_struct *p __attribute__((unused))) { } static inline void -print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq) +print_cfs_rq(struct seq_file *m __attribute__((unused)), int cpu __attribute__((unused)), struct cfs_rq *cfs_rq __attribute__((unused))) { } #endif @@ -275,7 +275,7 @@ extern void select_nohz_load_balancer(int stop_tick); extern int get_nohz_timer_target(void); #else -static inline void select_nohz_load_balancer(int stop_tick) { } +static inline void select_nohz_load_balancer(int stop_tick __attribute__((unused))) { } #endif /* @@ -1017,8 +1017,8 @@ struct sched_domain_attr; static inline void -partition_sched_domains(int ndoms_new, cpumask_var_t doms_new[], - struct sched_domain_attr *dattr_new) +partition_sched_domains(int ndoms_new __attribute__((unused)), cpumask_var_t doms_new[] __attribute__((unused)), + struct sched_domain_attr *dattr_new __attribute__((unused))) { } #endif /* !CONFIG_SMP */ @@ -1030,7 +1030,7 @@ #ifdef ARCH_HAS_PREFETCH_SWITCH_STACK extern void prefetch_stack(struct task_struct *t); #else -static inline void prefetch_stack(struct task_struct *t) { } +static inline void prefetch_stack(struct task_struct *t __attribute__((unused))) { } #endif struct audit_context; /* See audit.c */ @@ -1508,6 +1508,8 @@ unsigned long default_timer_slack_ns; struct list_head *scm_work_list; + + __u32 netmark; /* AVM calle, generic network mark */ #ifdef CONFIG_FUNCTION_GRAPH_TRACER /* Index of current stored address in ret_stack */ int curr_ret_stack; @@ -1804,7 +1806,7 @@ #else -static inline void rcu_copy_process(struct task_struct *p) +static inline void rcu_copy_process(struct task_struct *p __attribute__((unused))) { } @@ -1814,7 +1816,7 @@ extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask); #else -static inline int set_cpus_allowed_ptr(struct task_struct *p, +static inline int set_cpus_allowed_ptr(struct task_struct *p __attribute__((unused)), const struct cpumask *new_mask) { if (!cpumask_test_cpu(0, new_mask)) @@ -1858,7 +1860,7 @@ { } -static inline void sched_clock_idle_wakeup_event(u64 delta_ns) +static inline void sched_clock_idle_wakeup_event(u64 delta_ns __attribute__((unused))) { } #else @@ -1911,7 +1913,7 @@ #if defined(CONFIG_NO_HZ) && defined(CONFIG_SMP) extern void wake_up_idle_cpu(int cpu); #else -static inline void wake_up_idle_cpu(int cpu) { } +static inline void wake_up_idle_cpu(int cpu __attribute__((unused))) { } #endif extern unsigned int sysctl_sched_latency; @@ -1968,10 +1970,10 @@ extern int proc_sched_autogroup_set_nice(struct task_struct *p, int *nice); #endif #else -static inline void sched_autogroup_create_attach(struct task_struct *p) { } -static inline void sched_autogroup_detach(struct task_struct *p) { } -static inline void sched_autogroup_fork(struct signal_struct *sig) { } -static inline void sched_autogroup_exit(struct signal_struct *sig) { } +static inline void sched_autogroup_create_attach(struct task_struct *p __attribute__((unused))) { } +static inline void sched_autogroup_detach(struct task_struct *p __attribute__((unused))) { } +static inline void sched_autogroup_fork(struct signal_struct *sig __attribute__((unused))) { } +static inline void sched_autogroup_exit(struct signal_struct *sig __attribute__((unused))) { } #endif #ifdef CONFIG_RT_MUTEXES @@ -2068,7 +2070,7 @@ #ifdef CONFIG_SMP extern void kick_process(struct task_struct *tsk); #else - static inline void kick_process(struct task_struct *tsk) { } + static inline void kick_process(struct task_struct *tsk __attribute__((unused))) { } #endif extern void sched_fork(struct task_struct *p, int clone_flags); extern void sched_dead(struct task_struct *p); @@ -2197,8 +2199,8 @@ #ifdef CONFIG_SMP extern unsigned long wait_task_inactive(struct task_struct *, long match_state); #else -static inline unsigned long wait_task_inactive(struct task_struct *p, - long match_state) +static inline unsigned long wait_task_inactive(struct task_struct *p __attribute__((unused)), + long match_state __attribute__((unused))) { return 1; } @@ -2460,7 +2462,7 @@ * task waiting?: (technically does not depend on CONFIG_PREEMPT, * but a general need for low latency) */ -static inline int spin_needbreak(spinlock_t *lock) +static inline int spin_needbreak(spinlock_t *lock __attribute__((unused))) { #ifdef CONFIG_PREEMPT return spin_is_contended(lock); @@ -2505,12 +2507,12 @@ #else -static inline unsigned int task_cpu(const struct task_struct *p) +static inline unsigned int task_cpu(const struct task_struct *p __attribute__((unused))) { return 0; } -static inline void set_task_cpu(struct task_struct *p, unsigned int cpu) +static inline void set_task_cpu(struct task_struct *p __attribute__((unused)), unsigned int cpu __attribute__((unused))) { } @@ -2567,19 +2569,19 @@ tsk->ioac.syscw++; } #else -static inline void add_rchar(struct task_struct *tsk, ssize_t amt) + static inline void add_rchar(struct task_struct *tsk __attribute__((unused)), ssize_t amt __attribute__((unused))) { } -static inline void add_wchar(struct task_struct *tsk, ssize_t amt) +static inline void add_wchar(struct task_struct *tsk __attribute__((unused)), ssize_t amt __attribute__((unused))) { } -static inline void inc_syscr(struct task_struct *tsk) +static inline void inc_syscr(struct task_struct *tsk __attribute__((unused))) { } -static inline void inc_syscw(struct task_struct *tsk) +static inline void inc_syscw(struct task_struct *tsk __attribute__((unused))) { } #endif @@ -2592,11 +2594,11 @@ extern void mm_update_next_owner(struct mm_struct *mm); extern void mm_init_owner(struct mm_struct *mm, struct task_struct *p); #else -static inline void mm_update_next_owner(struct mm_struct *mm) +static inline void mm_update_next_owner(struct mm_struct *mm __attribute__((unused))) { } -static inline void mm_init_owner(struct mm_struct *mm, struct task_struct *p) +static inline void mm_init_owner(struct mm_struct *mm __attribute__((unused)), struct task_struct *p __attribute__((unused))) { } #endif /* CONFIG_MM_OWNER */