aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched.c2
-rw-r--r--kernel/sched_fair.c3
-rw-r--r--kernel/sched_rt.c3
3 files changed, 3 insertions, 5 deletions
diff --git a/kernel/sched.c b/kernel/sched.c
index 49a5fb0..43ae156 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -852,7 +852,7 @@ static void
enqueue_task(struct rq *rq, struct task_struct *p, int wakeup, u64 now)
{
sched_info_queued(p);
- p->sched_class->enqueue_task(rq, p, wakeup, now);
+ p->sched_class->enqueue_task(rq, p, wakeup);
p->se.on_rq = 1;
}
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index a11d188..81db962 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -782,8 +782,7 @@ static inline int is_same_group(struct task_struct *curr, struct task_struct *p)
* increased. Here we update the fair scheduling stats and
* then put the task into the rbtree:
*/
-static void
-enqueue_task_fair(struct rq *rq, struct task_struct *p, int wakeup, u64 now)
+static void enqueue_task_fair(struct rq *rq, struct task_struct *p, int wakeup)
{
struct cfs_rq *cfs_rq;
struct sched_entity *se = &p->se;
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c
index fa5a462..1edaa99 100644
--- a/kernel/sched_rt.c
+++ b/kernel/sched_rt.c
@@ -25,8 +25,7 @@ static inline void update_curr_rt(struct rq *rq)
curr->se.exec_start = rq->clock;
}
-static void
-enqueue_task_rt(struct rq *rq, struct task_struct *p, int wakeup, u64 now)
+static void enqueue_task_rt(struct rq *rq, struct task_struct *p, int wakeup)
{
struct rt_prio_array *array = &rq->rt.active;