From patchwork Wed Aug 24 07:44:50 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Leizhen \(ThunderTown\)" X-Patchwork-Id: 74555 Delivered-To: patch@linaro.org Received: by 10.140.29.52 with SMTP id a49csp180041qga; Wed, 24 Aug 2016 00:47:48 -0700 (PDT) X-Received: by 10.66.51.98 with SMTP id j2mr3112403pao.87.1472024868778; Wed, 24 Aug 2016 00:47:48 -0700 (PDT) Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id a5si8363799pfe.203.2016.08.24.00.47.47; Wed, 24 Aug 2016 00:47:48 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932218AbcHXHrg (ORCPT + 27 others); Wed, 24 Aug 2016 03:47:36 -0400 Received: from szxga02-in.huawei.com ([119.145.14.65]:9211 "EHLO szxga02-in.huawei.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752902AbcHXHqf (ORCPT ); Wed, 24 Aug 2016 03:46:35 -0400 Received: from 172.24.1.47 (EHLO szxeml425-hub.china.huawei.com) ([172.24.1.47]) by szxrg02-dlp.huawei.com (MOS 4.3.7-GA FastPath queued) with ESMTP id DMC31325; Wed, 24 Aug 2016 15:46:31 +0800 (CST) Received: from localhost (10.177.23.164) by szxeml425-hub.china.huawei.com (10.82.67.180) with Microsoft SMTP Server id 14.3.235.1; Wed, 24 Aug 2016 15:46:21 +0800 From: Zhen Lei To: Catalin Marinas , Will Deacon , linux-arm-kernel , linux-kernel , Rob Herring , "Frank Rowand" , devicetree CC: Zefan Li , Xinwei Hu , "Tianhong Ding" , Hanjun Guo , Zhen Lei Subject: [PATCH v7 11/14] arm64/numa: support HAVE_MEMORYLESS_NODES Date: Wed, 24 Aug 2016 15:44:50 +0800 Message-ID: <1472024693-12912-12-git-send-email-thunder.leizhen@huawei.com> X-Mailer: git-send-email 1.9.5.msysgit.1 In-Reply-To: <1472024693-12912-1-git-send-email-thunder.leizhen@huawei.com> References: <1472024693-12912-1-git-send-email-thunder.leizhen@huawei.com> MIME-Version: 1.0 X-Originating-IP: [10.177.23.164] X-CFilter-Loop: Reflected X-Mirapoint-Virus-RAPID-Raw: score=unknown(0), refid=str=0001.0A020202.57BD50D8.003C, ss=1, re=0.000, recu=0.000, reip=0.000, cl=1, cld=1, fgs=0, ip=0.0.0.0, so=2013-06-18 04:22:30, dmn=2013-03-21 17:37:32 X-Mirapoint-Loop-Id: 43e6bcfdf7315b2ddd10b39927d78bd0 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Some numa nodes may have no memory. For example: 1. cpu0 on node0 2. cpu1 on node1 3. device0 access the momory from node0 and node1 take the same time. So, we can not simply classify device0 to node0 or node1, but we can define a node2 which distances to node0 and node1 are the same. Signed-off-by: Zhen Lei --- arch/arm64/Kconfig | 4 ++++ arch/arm64/kernel/smp.c | 1 + arch/arm64/mm/numa.c | 43 +++++++++++++++++++++++++++++++++++++++++-- 3 files changed, 46 insertions(+), 2 deletions(-) -- 2.5.0 diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig index 2815af6..3a2b6ed 100644 --- a/arch/arm64/Kconfig +++ b/arch/arm64/Kconfig @@ -611,6 +611,10 @@ config NEED_PER_CPU_EMBED_FIRST_CHUNK def_bool y depends on NUMA +config HAVE_MEMORYLESS_NODES + def_bool y + depends on NUMA + source kernel/Kconfig.preempt source kernel/Kconfig.hz diff --git a/arch/arm64/kernel/smp.c b/arch/arm64/kernel/smp.c index d93d433..4879085 100644 --- a/arch/arm64/kernel/smp.c +++ b/arch/arm64/kernel/smp.c @@ -619,6 +619,7 @@ static void __init of_parse_and_init_cpus(void) } bootcpu_valid = true; + early_map_cpu_to_node(0, of_node_to_nid(dn)); /* * cpu_logical_map has already been diff --git a/arch/arm64/mm/numa.c b/arch/arm64/mm/numa.c index 6853db7..114180f 100644 --- a/arch/arm64/mm/numa.c +++ b/arch/arm64/mm/numa.c @@ -129,6 +129,14 @@ void __init early_map_cpu_to_node(unsigned int cpu, int nid) nid = 0; cpu_to_node_map[cpu] = nid; + + /* + * We should set the numa node of cpu0 as soon as possible, because it + * has already been set up online before. cpu_to_node(0) will soon be + * called. + */ + if (!cpu) + set_cpu_numa_node(cpu, nid); } #ifdef CONFIG_HAVE_SETUP_PER_CPU_AREA @@ -211,6 +219,35 @@ int __init numa_add_memblk(int nid, u64 start, u64 end) return ret; } +static u64 __init alloc_node_data_from_nearest_node(int nid, const size_t size) +{ + int i, best_nid, distance; + u64 pa; + DECLARE_BITMAP(nodes_map, MAX_NUMNODES); + + bitmap_zero(nodes_map, MAX_NUMNODES); + bitmap_set(nodes_map, nid, 1); + +find_nearest_node: + best_nid = NUMA_NO_NODE; + distance = INT_MAX; + + for_each_clear_bit(i, nodes_map, MAX_NUMNODES) + if (numa_distance[nid][i] < distance) { + best_nid = i; + distance = numa_distance[nid][i]; + } + + pa = memblock_alloc_nid(size, SMP_CACHE_BYTES, best_nid); + if (!pa) { + BUG_ON(best_nid == NUMA_NO_NODE); + bitmap_set(nodes_map, best_nid, 1); + goto find_nearest_node; + } + + return pa; +} + /** * Initialize NODE_DATA for a node on the local memory */ @@ -224,7 +261,9 @@ static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) pr_info("Initmem setup node %d [mem %#010Lx-%#010Lx]\n", nid, start_pfn << PAGE_SHIFT, (end_pfn << PAGE_SHIFT) - 1); - nd_pa = memblock_alloc_try_nid(nd_size, SMP_CACHE_BYTES, nid); + nd_pa = memblock_alloc_nid(nd_size, SMP_CACHE_BYTES, nid); + if (!nd_pa) + nd_pa = alloc_node_data_from_nearest_node(nid, nd_size); nd = __va(nd_pa); /* report and initialize */ @@ -234,7 +273,7 @@ static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) if (tnid != nid) pr_info(" NODE_DATA(%d) on node %d\n", nid, tnid); - node_data[nid] = nd; + NODE_DATA(nid) = nd; memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); NODE_DATA(nid)->node_id = nid; NODE_DATA(nid)->node_start_pfn = start_pfn;