From patchwork Thu Apr 9 03:27:18 2015 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Xunlei Pang X-Patchwork-Id: 46900 Return-Path: X-Original-To: linaro@patches.linaro.org Delivered-To: linaro@patches.linaro.org Received: from mail-wg0-f72.google.com (mail-wg0-f72.google.com [74.125.82.72]) by ip-10-151-82-157.ec2.internal (Postfix) with ESMTPS id A544521416 for ; Thu, 9 Apr 2015 03:29:46 +0000 (UTC) Received: by wghm4 with SMTP id m4sf23512206wgh.2 for ; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:delivered-to:from:to:cc:subject :date:message-id:in-reply-to:references:sender:precedence:list-id :x-original-sender:x-original-authentication-results:mailing-list :list-post:list-help:list-archive:list-unsubscribe; bh=c+KLE8Oox9Typr9OQk81g2wTlyWoCLLhoIw5FK3xd64=; b=ds0fWP//EgO4cnv9KerASEESwr2KPJrfeY2zrgdaf5OKxoyY66Ig03hdeisLgWOWMc nggjMzzAWu/McppCrxttwlgDdIfdHAgDbbV5ZJ8aojTrrPo8gJ8u0NNztkWDvzvk0wxO UxmIquE0L4NWP2SbcIH21PSs4EBPDtO+SF1E1WNGcVC/uk0r2ybK9sXyfpt5jyBAEeTZ 5UJpVqaEjJRJdFS8pjbcwTIJQ2GtRqdl9FaKN9P5m5SfxrP1b8iLhbgH3b+BHJf/ydQD M+4c2nVEkCLr9DEF6wjXL9al0x3R7HgX2rKfUyyTwlyT85dmWFWFceW+4qmyZUqwqTB/ jkzg== X-Gm-Message-State: ALoCoQmvH/MdNiYBn1AwnvSdE0otKwMn/wxa1yPVJYEUls3/uhM984QqCw7kTawkEhegYMEqhWXW X-Received: by 10.152.3.70 with SMTP id a6mr71081laa.0.1428550185888; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) MIME-Version: 1.0 X-BeenThere: patchwork-forward@linaro.org Received: by 10.152.1.3 with SMTP id 3ls239027lai.36.gmail; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) X-Received: by 10.152.26.34 with SMTP id i2mr2621726lag.117.1428550185708; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) Received: from mail-la0-x236.google.com (mail-la0-x236.google.com. [2a00:1450:4010:c03::236]) by mx.google.com with ESMTPS id tp6si1139927lbb.97.2015.04.08.20.29.45 for (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 08 Apr 2015 20:29:45 -0700 (PDT) Received-SPF: pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 2a00:1450:4010:c03::236 as permitted sender) client-ip=2a00:1450:4010:c03::236; Received: by laat2 with SMTP id t2so73103425laa.1 for ; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) X-Received: by 10.152.163.35 with SMTP id yf3mr2471829lab.86.1428550185565; Wed, 08 Apr 2015 20:29:45 -0700 (PDT) X-Forwarded-To: patchwork-forward@linaro.org X-Forwarded-For: patch@linaro.org patchwork-forward@linaro.org Delivered-To: patch@linaro.org Received: by 10.112.67.65 with SMTP id l1csp221216lbt; Wed, 8 Apr 2015 20:29:44 -0700 (PDT) X-Received: by 10.68.195.2 with SMTP id ia2mr34201303pbc.33.1428550183684; Wed, 08 Apr 2015 20:29:43 -0700 (PDT) Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id h4si19409222pat.146.2015.04.08.20.29.42; Wed, 08 Apr 2015 20:29:43 -0700 (PDT) Received-SPF: none (google.com: linux-kernel-owner@vger.kernel.org does not designate permitted sender hosts) client-ip=209.132.180.67; Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754307AbbDID3g (ORCPT + 27 others); Wed, 8 Apr 2015 23:29:36 -0400 Received: from m15-112.126.com ([220.181.15.112]:39775 "EHLO m15-112.126.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752889AbbDID3e (ORCPT ); Wed, 8 Apr 2015 23:29:34 -0400 Received: from localhost.localdomain (unknown [210.21.223.3]) by smtp2 (Coremail) with SMTP id DMmowEDJvlOa8SVVTV0aAQ--.926S4; Thu, 09 Apr 2015 11:27:30 +0800 (CST) From: Xunlei Pang To: linux-kernel@vger.kernel.org Cc: Peter Zijlstra , Steven Rostedt , Juri Lelli , Dan Streetman , Xunlei Pang Subject: [PATCH v5 3/3] sched/rt: Check to push the task when changing its affinity Date: Thu, 9 Apr 2015 11:27:18 +0800 Message-Id: <1428550038-13619-3-git-send-email-xlpang@126.com> X-Mailer: git-send-email 1.9.1 In-Reply-To: <1428550038-13619-1-git-send-email-xlpang@126.com> References: <1428550038-13619-1-git-send-email-xlpang@126.com> X-CM-TRANSID: DMmowEDJvlOa8SVVTV0aAQ--.926S4 X-Coremail-Antispam: 1Uf129KBjvJXoWxXFy3Jr48KFy5ur4fAF1UJrb_yoWrCF13pa 1vk390gr4DJa1SgF1fZw4DZr45K3sYv34rJrnxtw1FkFs8tr4Fv3W5tF1ayF95ur1YgF4a qr4DtrsrKF1DZaDanT9S1TB71UUUUUUqnTZGkaVYY2UrUUUUjbIjqfuFe4nvWSU5nxnvy2 9KBjDUYxBIdaVFxhVjvjDU0xZFpf9x07b59a9UUUUU= X-Originating-IP: [210.21.223.3] X-CM-SenderInfo: p0ost0bj6rjloofrz/1tbimgLWv1GfV0Y-GgAAsL Sender: linux-kernel-owner@vger.kernel.org Precedence: list List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Original-Sender: xlpang@126.com X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 2a00:1450:4010:c03::236 as permitted sender) smtp.mail=patch+caf_=patchwork-forward=linaro.org@linaro.org; dkim=neutral (body hash did not verify) header.i=@; dmarc=fail (p=NONE dis=NONE) header.from=126.com Mailing-list: list patchwork-forward@linaro.org; contact patchwork-forward+owners@linaro.org X-Google-Group-Id: 836684582541 List-Post: , List-Help: , List-Archive: List-Unsubscribe: , From: Xunlei Pang We may suffer from extra rt overload rq due to the affinity, so when the affinity of any runnable rt task is changed, we should check to trigger balancing, otherwise it will cause some unnecessary delayed real-time response. Unfortunately, current RT global scheduler doesn't trigger anything. For example: a 2-cpu system with two runnable FIFO tasks(same rt_priority) bound on CPU0, let's name them rt1(running) and rt2(runnable) respectively; CPU1 has no RTs. Then, someone sets the affinity of rt2 to 0x3(i.e. CPU0 and CPU1), but after this, rt2 still can't be scheduled until rt1 enters schedule(), this definitely causes some/big response latency for rt2. So, when doing set_cpus_allowed_rt(), if detecting such cases, check to trigger a push behaviour. Signed-off-by: Xunlei Pang --- kernel/sched/rt.c | 81 ++++++++++++++++++++++++++++++++++++++++++++++++------- 1 file changed, 71 insertions(+), 10 deletions(-) diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c index 402162a..9704ed3 100644 --- a/kernel/sched/rt.c +++ b/kernel/sched/rt.c @@ -1445,10 +1445,9 @@ static struct sched_rt_entity *pick_next_rt_entity(struct rq *rq, return next; } -static struct task_struct *_pick_next_task_rt(struct rq *rq) +static struct task_struct *peek_next_task_rt(struct rq *rq) { struct sched_rt_entity *rt_se; - struct task_struct *p; struct rt_rq *rt_rq = &rq->rt; do { @@ -1457,7 +1456,14 @@ static struct task_struct *_pick_next_task_rt(struct rq *rq) rt_rq = group_rt_rq(rt_se); } while (rt_rq); - p = rt_task_of(rt_se); + return rt_task_of(rt_se); +} + +static inline struct task_struct *_pick_next_task_rt(struct rq *rq) +{ + struct task_struct *p; + + p = peek_next_task_rt(rq); p->se.exec_start = rq_clock_task(rq); return p; @@ -2077,28 +2083,77 @@ static void set_cpus_allowed_rt(struct task_struct *p, const struct cpumask *new_mask) { struct rq *rq; - int weight; + int old_weight, new_weight; + int preempt_push = 0, direct_push = 0; BUG_ON(!rt_task(p)); if (!task_on_rq_queued(p)) return; - weight = cpumask_weight(new_mask); + old_weight = p->nr_cpus_allowed; + new_weight = cpumask_weight(new_mask); + rq = task_rq(p); + + if (new_weight > 1 && + rt_task(rq->curr) && + rq->rt.rt_nr_total > 1 && + !test_tsk_need_resched(rq->curr)) { + /* + * We own p->pi_lock and rq->lock. rq->lock might + * get released when doing direct pushing, however + * p->pi_lock is always held, so it's safe to assign + * new_mask and new_weight to p below. + */ + if (!task_running(rq, p)) { + cpumask_copy(&p->cpus_allowed, new_mask); + p->nr_cpus_allowed = new_weight; + direct_push = 1; + } else if (cpumask_test_cpu(task_cpu(p), new_mask)) { + struct task_struct *next; + + cpumask_copy(&p->cpus_allowed, new_mask); + p->nr_cpus_allowed = new_weight; + if (!cpupri_find(&rq->rd->cpupri, p, NULL)) + goto update; + + /* + * At this point, current task gets migratable most + * likely due to the change of its affinity, let's + * figure out if we can migrate it. + * + * Can we find any task with the same priority as + * current? To accomplish this, firstly we requeue + * current to the tail and peek next, then restore + * current to the head. + */ + requeue_task_rt(rq, p, 0); + next = peek_next_task_rt(rq); + requeue_task_rt(rq, p, 1); + if (next != p && next->prio == p->prio) { + /* + * Target found, so let's reschedule to try + * and push current away. + */ + requeue_task_rt(rq, next, 1); + preempt_push = 1; + } + } + } + +update: /* * Only update if the process changes its state from whether it * can migrate or not. */ - if ((p->nr_cpus_allowed > 1) == (weight > 1)) - return; - - rq = task_rq(p); + if ((old_weight > 1) == (new_weight > 1)) + goto out; /* * The process used to be able to migrate OR it can now migrate */ - if (weight <= 1) { + if (new_weight <= 1) { if (!task_current(rq, p)) dequeue_pushable_task(rq, p); BUG_ON(!rq->rt.rt_nr_migratory); @@ -2110,6 +2165,12 @@ static void set_cpus_allowed_rt(struct task_struct *p, } update_rt_migration(&rq->rt); + +out: + if (direct_push) + push_rt_tasks(rq); + else if (preempt_push) + resched_curr(rq); } /* Assumes rq->lock is held */