[v6,1/4] sched/fair: Fix attaching task sched avgs twice when switching to fair or changing task group

Message ID 20160616163013.GA32169@vingu-laptop
State New
Headers show

Commit Message

Vincent Guittot June 16, 2016, 4:30 p.m.
Le Wednesday 15 Jun 2016 à 17:22:17 (+0200), Peter Zijlstra a écrit :
> On Wed, Jun 15, 2016 at 09:46:53AM +0200, Vincent Guittot wrote:

> > I still have concerned with this change of the behavior that  attaches

> > the task only when it is enqueued. The load avg of the task will not

> > be decayed between the time we move it into its new group until its

> > enqueue. With this change, a task's load can stay high whereas it has

> > slept for the last couple of seconds. Then, its load and utilization

> > is no more accounted anywhere in the mean time just because we have

> > moved the task which will be enqueued on the same rq.

> > A task should always be attached to a cfs_rq and its load/utilization

> > should always be accounted on a cfs_rq and decayed for its sleep

> > period

> 

> OK; so I think I agree with that. Does the below (completely untested,

> hasn't even been near a compiler) look reasonable?

> 

> The general idea is to always attach to a cfs_rq -- through

> post_init_entity_util_avg(). This covers both the new task isn't

> attached yet and was never in the fair class to begin with issues.


Your patch ensures that a task will be attached to a cfs_rq and fix the issue raised by Yuyang because of se->avg.last_update_time = 0 at init. During the test the following message has raised  "BUG: using smp_processor_id() in preemptible [00000000] code: systemd/1" because of cfs_rq_util_change that is called in attach_entity_load_avg

With patch [1] for the init of cfs_rq side, all use cases will be covered regarding the issue linked to a last_update_time set to 0 at init
[1] https://lkml.org/lkml/2016/5/30/508

> 

> That only leaves a tiny hole in fork() where the task is hashed but

> hasn't yet passed through wake_up_new_task() in which someone can do

> cgroup move on it. That is closed with TASK_NEW and can_attach()

> refusing those tasks.

> 


But a new fair task is still detached and attached from/to task_group with cgroup_post_fork()-->ss->fork(child)-->cpu_cgroup_fork()-->sched_move_task()-->task_move_group_fair().
cpu_cgroup_can_attach is not used in this path and sched_move_task do the move unconditionally for fair task.

With your patch, we still have the sequence

sched_fork()
    set_task_cpu()
cgroup_post_fork()--> ... --> task_move_group_fair()
    detach_task_cfs_rq()
    set_task_rq()
    attach_task_cfs_rq()
wake_up_new_task()
    select_task_rq() can select a new cpu
    set_task_cpu()
        migrate_task_rq_fair if the new_cpu != task_cpu
             remove_load()
        __set_task_cpu
    post_init_entity_util_avg
        attach_task_cfs_rq()
    activate_task
        enqueue_task

In fact, cpu_cgroup_fork needs a small part of sched_move_task so we can just call this small part directly instead sched_move_task. And the task doesn't really migrate because it is not yet attached so we need the sequence:
sched_fork()
    __set_task_cpu()
cgroup_post_fork()--> ... --> task_move_group_fair()
    set_task_rq() to set task group and runqueue
wake_up_new_task()
    select_task_rq() can select a new cpu
    __set_task_cpu
    post_init_entity_util_avg
        attach_task_cfs_rq()
    activate_task
        enqueue_task

The patch below on top of your patch, ensures that we follow the right sequence :

---
 kernel/sched/core.c | 60 +++++++++++++++++++++++++++++++++++------------------
 1 file changed, 40 insertions(+), 20 deletions(-)

-- 
1.9.1

Comments

Vincent Guittot June 16, 2016, 6:57 p.m. | #1
On 16 June 2016 at 19:17, Peter Zijlstra <peterz@infradead.org> wrote:
> On Thu, Jun 16, 2016 at 06:30:13PM +0200, Vincent Guittot wrote:

>> Le Wednesday 15 Jun 2016 à 17:22:17 (+0200), Peter Zijlstra a écrit :

>> > On Wed, Jun 15, 2016 at 09:46:53AM +0200, Vincent Guittot wrote:

>> > > I still have concerned with this change of the behavior that  attaches

>> > > the task only when it is enqueued. The load avg of the task will not

>> > > be decayed between the time we move it into its new group until its

>> > > enqueue. With this change, a task's load can stay high whereas it has

>> > > slept for the last couple of seconds. Then, its load and utilization

>> > > is no more accounted anywhere in the mean time just because we have

>> > > moved the task which will be enqueued on the same rq.

>> > > A task should always be attached to a cfs_rq and its load/utilization

>> > > should always be accounted on a cfs_rq and decayed for its sleep

>> > > period

>> >

>> > OK; so I think I agree with that. Does the below (completely untested,

>> > hasn't even been near a compiler) look reasonable?

>> >

>> > The general idea is to always attach to a cfs_rq -- through

>> > post_init_entity_util_avg(). This covers both the new task isn't

>> > attached yet and was never in the fair class to begin with issues.

>>

>> Your patch ensures that a task will be attached to a cfs_rq and fix

>> the issue raised by Yuyang because of se->avg.last_update_time = 0 at

>> init.

>

>> During the test the following message has raised  "BUG: using

>> smp_processor_id() in preemptible [00000000] code: systemd/1" because

>> of cfs_rq_util_change that is called in attach_entity_load_avg

>

> But per commit:

>

>   b7fa30c9cc48 ("sched/fair: Fix post_init_entity_util_avg() serialization")

>

> that should be with rq->lock held !?


yes, you're right, i forgot to fetch latest commits

>

>> With patch [1] for the init of cfs_rq side, all use cases will be

>> covered regarding the issue linked to a last_update_time set to 0 at

>> init [1] https://lkml.org/lkml/2016/5/30/508

>

> Yes, I saw that patch. However, it felt strange to me to set

> last_update_time to 1, would that not result in a massive aging because

> now - last_update_time ends up as a giant delta?


yes, it will but it's similar to what can also happen when the other
tasks will be enqueued in the cfs_rq. The cfs_rq->avg.last_update_time
can be really far from now for cfs_rq that is idle.

>

> By doing attach_entity_load_avg() we actually add the initial values to

> the cfs_rq avg and set last_update_time to the current time. Ensuring

> 'regular' contribution and progress.

>

>> > That only leaves a tiny hole in fork() where the task is hashed but

>> > hasn't yet passed through wake_up_new_task() in which someone can do

>> > cgroup move on it. That is closed with TASK_NEW and can_attach()

>> > refusing those tasks.

>> >

>>

>> But a new fair task is still detached and attached from/to task_group

>> with

>

>> cgroup_post_fork()-->

>     ss->fork(child)-->

>       cpu_cgroup_fork()-->

>         sched_move_task()-->

>           task_move_group_fair().

>

> Blergh, I knew I missed something in there..

>

>> cpu_cgroup_can_attach is not used in this path and sched_move_task do

>> the move unconditionally for fair task.

>>

>> With your patch, we still have the sequence

>>

>> sched_fork()

>>     set_task_cpu()

>> cgroup_post_fork()--> ... --> task_move_group_fair()

>>     detach_task_cfs_rq()

>>     set_task_rq()

>>     attach_task_cfs_rq()

>> wake_up_new_task()

>>     select_task_rq() can select a new cpu

>>     set_task_cpu()

>>         migrate_task_rq_fair if the new_cpu != task_cpu

>>              remove_load()

>>         __set_task_cpu

>>     post_init_entity_util_avg

>>         attach_task_cfs_rq()

>>     activate_task

>>         enqueue_task

>>

>> In fact, cpu_cgroup_fork needs a small part of sched_move_task so we

>> can just call this small part directly instead sched_move_task. And

>> the task doesn't really migrate because it is not yet attached so we

>> need the sequence:

>

>> sched_fork()

>>     __set_task_cpu()

>> cgroup_post_fork()--> ... --> task_move_group_fair()

>>     set_task_rq() to set task group and runqueue

>> wake_up_new_task()

>>     select_task_rq() can select a new cpu

>>     __set_task_cpu

>>     post_init_entity_util_avg

>>         attach_task_cfs_rq()

>>     activate_task

>>         enqueue_task

>>

>> The patch below on top of your patch, ensures that we follow the right sequence :

>>

>> ---

>>  kernel/sched/core.c | 60 +++++++++++++++++++++++++++++++++++------------------

>>  1 file changed, 40 insertions(+), 20 deletions(-)

>>

>> diff --git a/kernel/sched/core.c b/kernel/sched/core.c

>> index 7895689a..a21e3dc 100644

>> --- a/kernel/sched/core.c

>> +++ b/kernel/sched/core.c

>> @@ -2373,7 +2373,7 @@ int sched_fork(unsigned long clone_flags, struct task_struct *p)

>>        * Silence PROVE_RCU.

>>        */

>>       raw_spin_lock_irqsave(&p->pi_lock, flags);

>> -     set_task_cpu(p, cpu);

>> +     __set_task_cpu(p, cpu);

>

> Indeed, this should not be calling migrate, we're setting the CPU for

> the first time.

>

>>       if (p->sched_class->task_fork)

>>               p->sched_class->task_fork(p);

>>       raw_spin_unlock_irqrestore(&p->pi_lock, flags);

>> @@ -2515,7 +2515,7 @@ void wake_up_new_task(struct task_struct *p)

>>        *  - cpus_allowed can change in the fork path

>>        *  - any previously selected cpu might disappear through hotplug

>>        */

>> -     set_task_cpu(p, select_task_rq(p, task_cpu(p), SD_BALANCE_FORK, 0));

>> +     __set_task_cpu(p, select_task_rq(p, task_cpu(p), SD_BALANCE_FORK, 0));

>

> Similarly here I suppose, since we're not yet properly attached as such.

>

>>  #endif

>>       /* Post initialize new task's util average when its cfs_rq is set */

>>       post_init_entity_util_avg(&p->se);

>

> You were indeed running an 'old' kernel, as this

> post_init_entity_util_avg() call should be _after_ the __task_rq_lock().

>

>> @@ -7715,6 +7715,35 @@ void sched_offline_group(struct task_group *tg)

>>       spin_unlock_irqrestore(&task_group_lock, flags);

>>  }

>>

>> +/* Set task's runqueue and group

>> + *     In case of a move between group, we update src and dst group

>> + *     thanks to sched_class->task_move_group. Otherwise, we just need to set

>> + *     runqueue and group pointers. The task will be attached to the runqueue

>> + *     during its wake up.

>

> Broken comment style.

>

>> + */

>> +static void sched_set_group(struct task_struct *tsk, bool move)

>> +{

>> +     struct task_group *tg;

>> +

>> +     /*

>> +      * All callers are synchronized by task_rq_lock(); we do not use RCU

>> +      * which is pointless here. Thus, we pass "true" to task_css_check()

>> +      * to prevent lockdep warnings.

>> +      */

>> +     tg = container_of(task_css_check(tsk, cpu_cgrp_id, true),

>> +                       struct task_group, css);

>> +     tg = autogroup_task_group(tsk, tg);

>> +     tsk->sched_task_group = tg;

>> +

>> +#ifdef CONFIG_FAIR_GROUP_SCHED

>> +     if (move && tsk->sched_class->task_move_group)

>> +             tsk->sched_class->task_move_group(tsk);

>> +     else

>> +#endif

>> +             set_task_rq(tsk, task_cpu(tsk));

>> +

>> +}

>> +

>>  /* change task's runqueue when it moves between groups.

>>   *   The caller of this function should have put the task in its new group

>>   *   by now. This function just updates tsk->se.cfs_rq and tsk->se.parent to

>> @@ -7722,7 +7751,6 @@ void sched_offline_group(struct task_group *tg)

>>   */

>>  void sched_move_task(struct task_struct *tsk)

>>  {

>> -     struct task_group *tg;

>>       int queued, running;

>>       struct rq_flags rf;

>>       struct rq *rq;

>> @@ -7737,22 +7765,7 @@ void sched_move_task(struct task_struct *tsk)

>>       if (unlikely(running))

>>               put_prev_task(rq, tsk);

>>

>> +     sched_set_group(tsk, true);

>>

>>       if (unlikely(running))

>>               tsk->sched_class->set_curr_task(rq);

>> @@ -8182,7 +8195,14 @@ static void cpu_cgroup_css_free(struct cgroup_subsys_state *css)

>>

>>  static void cpu_cgroup_fork(struct task_struct *task)

>>  {

>> -     sched_move_task(task);

>> +     struct rq_flags rf;

>> +     struct rq *rq;

>> +

>> +     rq = task_rq_lock(task, &rf);

>> +

>> +     sched_set_group(task, false);

>> +

>> +     task_rq_unlock(rq, task, &rf);

>>  }

>

> Hmm, yeah, I think you're right.

>

> Let me fold that.
Vincent Guittot June 16, 2016, 7 p.m. | #2
On 16 June 2016 at 20:51, Peter Zijlstra <peterz@infradead.org> wrote:
> On Thu, Jun 16, 2016 at 06:30:13PM +0200, Vincent Guittot wrote:

>> With patch [1] for the init of cfs_rq side, all use cases will be

>> covered regarding the issue linked to a last_update_time set to 0 at

>> init

>> [1] https://lkml.org/lkml/2016/5/30/508

>

> Aah, wait, now I get it :-)

>

> Still, we should put cfs_rq_clock_task(cfs_rq) in it, not 1. And since

> we now acquire rq->lock on init this should well be possible. Lemme sort

> that.


yes with the rq->lock we can use cfs_rq_clock_task which is make more
sense than 1.
But the delta can be still significant between the creation of the
task group and the 1st task that will be attach to the cfs_rq
Vincent Guittot June 16, 2016, 9:21 p.m. | #3
On 16 June 2016 at 22:07, Peter Zijlstra <peterz@infradead.org> wrote:
> On Thu, Jun 16, 2016 at 09:00:57PM +0200, Vincent Guittot wrote:

>> On 16 June 2016 at 20:51, Peter Zijlstra <peterz@infradead.org> wrote:

>> > On Thu, Jun 16, 2016 at 06:30:13PM +0200, Vincent Guittot wrote:

>> >> With patch [1] for the init of cfs_rq side, all use cases will be

>> >> covered regarding the issue linked to a last_update_time set to 0 at

>> >> init

>> >> [1] https://lkml.org/lkml/2016/5/30/508

>> >

>> > Aah, wait, now I get it :-)

>> >

>> > Still, we should put cfs_rq_clock_task(cfs_rq) in it, not 1. And since

>> > we now acquire rq->lock on init this should well be possible. Lemme sort

>> > that.

>>

>> yes with the rq->lock we can use cfs_rq_clock_task which is make more

>> sense than 1.

>> But the delta can be still significant between the creation of the

>> task group and the 1st task that will be attach to the cfs_rq

>

> Ah, I think I've spotted more fail.

>

> And I think you're right, it doesn't matter, in fact, 0 should have been

> fine too!

>

> enqueue_entity()

>   enqueue_entity_load_avg()

>     update_cfs_rq_load_avg()

>       now = clock()

>       __update_load_avg(&cfs_rq->avg)

>         cfs_rq->avg.last_load_update = now

>         // ages 0 load/util for: now - 0

>     if (migrated)

>       attach_entity_load_avg()

>         se->avg.last_load_update = cfs_rq->avg.last_load_update; // now != 0

>

> So I don't see how it can end up being attached again.


In fact it has already been attached during the sched_move_task. The
sequence for the 1st task that is attached to a cfs_rq is :

sched_move_task()
  task_move_group_fair()
    detach_task_cfs_rq()
    set_task_rq()
    attach_task_cfs_rq()
      attach_entity_load_avg()
        se->avg.last_load_update = cfs_rq->avg.last_load_update == 0

Then we enqueue the task but se->avg.last_load_update is still 0 so
migrated is set and we attach the task one more time

>

>

> Now I do see another problem, and that is that we're forgetting to

> update_cfs_rq_load_avg() in all detach_entity_load_avg() callers and all

> but the enqueue caller of attach_entity_load_avg().


Yes, calling update_cfs_rq_load_avg before all attach_entity_load_avg
will ensure that cfs_rq->avg.last_load_update will never be 0 when
attaching a task
And doing that before the detach will ensure that we move an up-to-date load

Your proposal below looks good to me

>

> Something like the below.

>

>

>

> ---

>  kernel/sched/fair.c | 4 ++++

>  1 file changed, 4 insertions(+)

>

> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c

> index f75930bdd326..5d8fa135bbc5 100644

> --- a/kernel/sched/fair.c

> +++ b/kernel/sched/fair.c

> @@ -8349,6 +8349,7 @@ static void detach_task_cfs_rq(struct task_struct *p)

>  {

>         struct sched_entity *se = &p->se;

>         struct cfs_rq *cfs_rq = cfs_rq_of(se);

> +       u64 now = cfs_rq_clock_task(cfs_rq);

>

>         if (!vruntime_normalized(p)) {

>                 /*

> @@ -8360,6 +8361,7 @@ static void detach_task_cfs_rq(struct task_struct *p)

>         }

>

>         /* Catch up with the cfs_rq and remove our load when we leave */

> +       update_cfs_rq_load_avg(now, cfs_rq, false);

>         detach_entity_load_avg(cfs_rq, se);

>  }

>

> @@ -8367,6 +8369,7 @@ static void attach_task_cfs_rq(struct task_struct *p)

>  {

>         struct sched_entity *se = &p->se;

>         struct cfs_rq *cfs_rq = cfs_rq_of(se);

> +       u64 now = cfs_rq_clock_task(cfs_rq);

>

>  #ifdef CONFIG_FAIR_GROUP_SCHED

>         /*

> @@ -8377,6 +8380,7 @@ static void attach_task_cfs_rq(struct task_struct *p)

>  #endif

>

>         /* Synchronize task with its cfs_rq */

> +       update_cfs_rq_load_avg(now, cfs_rq, false);

>         attach_entity_load_avg(cfs_rq, se);

>

>         if (!vruntime_normalized(p))
Vincent Guittot June 17, 2016, noon | #4
On 17 June 2016 at 04:12, Yuyang Du <yuyang.du@intel.com> wrote:
>

> On Thu, Jun 16, 2016 at 11:21:55PM +0200, Vincent Guittot wrote:

> > On 16 June 2016 at 22:07, Peter Zijlstra <peterz@infradead.org> wrote:

> > > On Thu, Jun 16, 2016 at 09:00:57PM +0200, Vincent Guittot wrote:

> > >> On 16 June 2016 at 20:51, Peter Zijlstra <peterz@infradead.org> wrote:

> > >> > On Thu, Jun 16, 2016 at 06:30:13PM +0200, Vincent Guittot wrote:

> > >> >> With patch [1] for the init of cfs_rq side, all use cases will be

> > >> >> covered regarding the issue linked to a last_update_time set to 0 at

> > >> >> init

> > >> >> [1] https://lkml.org/lkml/2016/5/30/508

> > >> >

> > >> > Aah, wait, now I get it :-)

> > >> >

> > >> > Still, we should put cfs_rq_clock_task(cfs_rq) in it, not 1. And since

> > >> > we now acquire rq->lock on init this should well be possible. Lemme sort

> > >> > that.

> > >>

> > >> yes with the rq->lock we can use cfs_rq_clock_task which is make more

> > >> sense than 1.

> > >> But the delta can be still significant between the creation of the

> > >> task group and the 1st task that will be attach to the cfs_rq

> > >

> > > Ah, I think I've spotted more fail.

> > >

> > > And I think you're right, it doesn't matter, in fact, 0 should have been

> > > fine too!

> > >

> > > enqueue_entity()

> > >   enqueue_entity_load_avg()

> > >     update_cfs_rq_load_avg()

> > >       now = clock()

> > >       __update_load_avg(&cfs_rq->avg)

> > >         cfs_rq->avg.last_load_update = now

> > >         // ages 0 load/util for: now - 0

> > >     if (migrated)

> > >       attach_entity_load_avg()

> > >         se->avg.last_load_update = cfs_rq->avg.last_load_update; // now != 0

> > >

> > > So I don't see how it can end up being attached again.

> >

> > In fact it has already been attached during the sched_move_task. The

> > sequence for the 1st task that is attached to a cfs_rq is :

> >

> > sched_move_task()

> >   task_move_group_fair()

> >     detach_task_cfs_rq()

> >     set_task_rq()

> >     attach_task_cfs_rq()

> >       attach_entity_load_avg()

> >         se->avg.last_load_update = cfs_rq->avg.last_load_update == 0

> >

>

> Then again, does this fix it?


Peter's proposal to update the cfs_rq before attaching/detaching the
task,  fixes the problem

>

>

> static void task_move_group_fair(struct task_struct *p)

> {

>         detach_task_cfs_rq(p);

>         set_task_rq(p, task_cpu(p));

>         attach_task_cfs_rq(p);

>         /*

>          * If the cfs_rq's last_update_time is 0, attach the sched avgs

>          * won't be anything useful, as it will be decayed to 0 when any

>          * sched_entity is enqueued to that cfs_rq.

>          *

>          * On the other hand, if the cfs_rq's last_update_time is 0, we

>          * must reset the task's last_update_time to ensure we will attach

>          * the sched avgs when the task is enqueued.

>          */

>         if (!cfs_rq_of(&p->se)->avg.last_update_time)

>                 reset_task_last_update_time(p);

>         else

>                 attach_entity_load_avg(cfs_rq_of(&p->se), &p->se);

> }

Patch

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 7895689a..a21e3dc 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -2373,7 +2373,7 @@  int sched_fork(unsigned long clone_flags, struct task_struct *p)
 	 * Silence PROVE_RCU.
 	 */
 	raw_spin_lock_irqsave(&p->pi_lock, flags);
-	set_task_cpu(p, cpu);
+	__set_task_cpu(p, cpu);
 	if (p->sched_class->task_fork)
 		p->sched_class->task_fork(p);
 	raw_spin_unlock_irqrestore(&p->pi_lock, flags);
@@ -2515,7 +2515,7 @@  void wake_up_new_task(struct task_struct *p)
 	 *  - cpus_allowed can change in the fork path
 	 *  - any previously selected cpu might disappear through hotplug
 	 */
-	set_task_cpu(p, select_task_rq(p, task_cpu(p), SD_BALANCE_FORK, 0));
+	__set_task_cpu(p, select_task_rq(p, task_cpu(p), SD_BALANCE_FORK, 0));
 #endif
 	/* Post initialize new task's util average when its cfs_rq is set */
 	post_init_entity_util_avg(&p->se);
@@ -7715,6 +7715,35 @@  void sched_offline_group(struct task_group *tg)
 	spin_unlock_irqrestore(&task_group_lock, flags);
 }
 
+/* Set task's runqueue and group
+ *     In case of a move between group, we update src and dst group
+ *     thanks to sched_class->task_move_group. Otherwise, we just need to set
+ *     runqueue and group pointers. The task will be attached to the runqueue
+ *     during its wake up.
+ */
+static void sched_set_group(struct task_struct *tsk, bool move)
+{
+	struct task_group *tg;
+
+	/*
+	 * All callers are synchronized by task_rq_lock(); we do not use RCU
+	 * which is pointless here. Thus, we pass "true" to task_css_check()
+	 * to prevent lockdep warnings.
+	 */
+	tg = container_of(task_css_check(tsk, cpu_cgrp_id, true),
+			  struct task_group, css);
+	tg = autogroup_task_group(tsk, tg);
+	tsk->sched_task_group = tg;
+
+#ifdef CONFIG_FAIR_GROUP_SCHED
+	if (move && tsk->sched_class->task_move_group)
+		tsk->sched_class->task_move_group(tsk);
+	else
+#endif
+		set_task_rq(tsk, task_cpu(tsk));
+
+}
+
 /* change task's runqueue when it moves between groups.
  *	The caller of this function should have put the task in its new group
  *	by now. This function just updates tsk->se.cfs_rq and tsk->se.parent to
@@ -7722,7 +7751,6 @@  void sched_offline_group(struct task_group *tg)
  */
 void sched_move_task(struct task_struct *tsk)
 {
-	struct task_group *tg;
 	int queued, running;
 	struct rq_flags rf;
 	struct rq *rq;
@@ -7737,22 +7765,7 @@  void sched_move_task(struct task_struct *tsk)
 	if (unlikely(running))
 		put_prev_task(rq, tsk);
 
-	/*
-	 * All callers are synchronized by task_rq_lock(); we do not use RCU
-	 * which is pointless here. Thus, we pass "true" to task_css_check()
-	 * to prevent lockdep warnings.
-	 */
-	tg = container_of(task_css_check(tsk, cpu_cgrp_id, true),
-			  struct task_group, css);
-	tg = autogroup_task_group(tsk, tg);
-	tsk->sched_task_group = tg;
-
-#ifdef CONFIG_FAIR_GROUP_SCHED
-	if (tsk->sched_class->task_move_group)
-		tsk->sched_class->task_move_group(tsk);
-	else
-#endif
-		set_task_rq(tsk, task_cpu(tsk));
+	sched_set_group(tsk, true);
 
 	if (unlikely(running))
 		tsk->sched_class->set_curr_task(rq);
@@ -8182,7 +8195,14 @@  static void cpu_cgroup_css_free(struct cgroup_subsys_state *css)
 
 static void cpu_cgroup_fork(struct task_struct *task)
 {
-	sched_move_task(task);
+	struct rq_flags rf;
+	struct rq *rq;
+
+	rq = task_rq_lock(task, &rf);
+
+	sched_set_group(task, false);
+
+	task_rq_unlock(rq, task, &rf);
 }
 
 static int cpu_cgroup_can_attach(struct cgroup_taskset *tset)