Message ID | 1403717444-23559-8-git-send-email-sudeep.holla@arm.com |
---|---|
State | New |
Headers | show |
Hi Sudeep, On Wed, Jun 25, 2014 at 06:30:42PM +0100, Sudeep Holla wrote: > From: Sudeep Holla <sudeep.holla@arm.com> > > This patch adds support for cacheinfo on ARM64. > > On ARMv8, the cache hierarchy can be identified through Cache Level ID > (CLIDR) register while the cache geometry is provided by Cache Size ID > (CCSIDR) register. > > Since the architecture doesn't provide any way of detecting the cpus > sharing particular cache, device tree is used for the same purpose. > > Signed-off-by: Sudeep Holla <sudeep.holla@arm.com> > Cc: Catalin Marinas <catalin.marinas@arm.com> > Cc: Will Deacon <will.deacon@arm.com> > Cc: Lorenzo Pieralisi <lorenzo.pieralisi@arm.com> > Cc: linux-arm-kernel@lists.infradead.org > --- > arch/arm64/kernel/Makefile | 3 +- > arch/arm64/kernel/cacheinfo.c | 135 ++++++++++++++++++++++++++++++++++++++++++ > 2 files changed, 137 insertions(+), 1 deletion(-) > create mode 100644 arch/arm64/kernel/cacheinfo.c [...] > +static inline enum cache_type get_cache_type(int level) > +{ > + unsigned int clidr; > + > + if (level > MAX_CACHE_LEVEL) > + return CACHE_TYPE_NOCACHE; > + asm volatile ("mrs %0, clidr_el1" : "=r" (clidr)); Can't that allocate a w register? You can make clidr a u64 to avoid that. > + return CLIDR_CTYPE(clidr, level); > +} > + > +/* > + * NumSets, bits[27:13] - (Number of sets in cache) - 1 > + * Associativity, bits[12:3] - (Associativity of cache) - 1 > + * LineSize, bits[2:0] - (Log2(Number of words in cache line)) - 2 > + */ > +#define CCSIDR_WRITE_THROUGH BIT(31) > +#define CCSIDR_WRITE_BACK BIT(30) > +#define CCSIDR_READ_ALLOCATE BIT(29) > +#define CCSIDR_WRITE_ALLOCATE BIT(28) > +#define CCSIDR_LINESIZE_MASK 0x7 > +#define CCSIDR_ASSOCIAT_SHIFT 3 > +#define CCSIDR_ASSOCIAT_MASK 0x3FF ASSOCIAT doesn't quite roll off of the tongue... > +#define CCSIDR_NUMSETS_SHIFT 13 > +#define CCSIDR_NUMSETS_MASK 0x7FF > + > +/* > + * Which cache CCSIDR represents depends on CSSELR value > + * Make sure no one else changes CSSELR during this > + * smp_call_function_single prevents preemption for us > + */ > +static inline u32 get_ccsidr(u32 csselr) > +{ > + u32 ccsidr; > + > + /* Put value into CSSELR */ > + asm volatile("msr csselr_el1, %x0" : : "r" (csselr)); This looks a little dodgy. I think GCC can leave the upper 32 bits in a random state. Why not cast csselr to a u64 here? > + isb(); > + /* Read result out of CCSIDR */ > + asm volatile("mrs %0, ccsidr_el1" : "=r" (ccsidr)); > + > + return ccsidr; Similarly it might make sense to make the temporary variable a u64. [...] > +int init_cache_level(unsigned int cpu) > +{ > + unsigned int ctype, level = 1, leaves = 0; > + struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); > + > + if (!this_cpu_ci) > + return -EINVAL; > + > + do { > + ctype = get_cache_type(level); > + if (ctype == CACHE_TYPE_NOCACHE) > + break; > + /* Separate instruction and data caches */ > + leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1; > + } while (++level <= MAX_CACHE_LEVEL); I think this would be clearer with: for (level = 1; level <= MAX_CACHE_LEVEL; level++) We do something like that in populate_cache_leaves below. > + > + this_cpu_ci->num_levels = level - 1; > + this_cpu_ci->num_leaves = leaves; > + return 0; > +} > + > +int populate_cache_leaves(unsigned int cpu) > +{ > + unsigned int level, idx; > + enum cache_type type; > + struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); > + struct cacheinfo *this_leaf = this_cpu_ci->info_list; > + > + for (idx = 0, level = 1; level <= this_cpu_ci->num_levels && > + idx < this_cpu_ci->num_leaves; idx++, level++) { > + if (!this_leaf) > + return -EINVAL; > + > + type = get_cache_type(level); > + if (type == CACHE_TYPE_SEPARATE) { > + ci_leaf_init(this_leaf++, CACHE_TYPE_DATA, level); > + ci_leaf_init(this_leaf++, CACHE_TYPE_INST, level); > + } else { > + ci_leaf_init(this_leaf++, type, level); > + } > + } > + return 0; > +} Cheers, Mark. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Hi Mark, Thanks for the review. On 27/06/14 11:36, Mark Rutland wrote: > Hi Sudeep, > > On Wed, Jun 25, 2014 at 06:30:42PM +0100, Sudeep Holla wrote: >> From: Sudeep Holla <sudeep.holla@arm.com> >> >> This patch adds support for cacheinfo on ARM64. >> >> On ARMv8, the cache hierarchy can be identified through Cache Level ID >> (CLIDR) register while the cache geometry is provided by Cache Size ID >> (CCSIDR) register. >> >> Since the architecture doesn't provide any way of detecting the cpus >> sharing particular cache, device tree is used for the same purpose. >> >> Signed-off-by: Sudeep Holla <sudeep.holla@arm.com> >> Cc: Catalin Marinas <catalin.marinas@arm.com> >> Cc: Will Deacon <will.deacon@arm.com> >> Cc: Lorenzo Pieralisi <lorenzo.pieralisi@arm.com> >> Cc: linux-arm-kernel@lists.infradead.org >> --- >> arch/arm64/kernel/Makefile | 3 +- >> arch/arm64/kernel/cacheinfo.c | 135 ++++++++++++++++++++++++++++++++++++++++++ >> 2 files changed, 137 insertions(+), 1 deletion(-) >> create mode 100644 arch/arm64/kernel/cacheinfo.c > > [...] > >> +static inline enum cache_type get_cache_type(int level) >> +{ >> + unsigned int clidr; >> + >> + if (level > MAX_CACHE_LEVEL) >> + return CACHE_TYPE_NOCACHE; >> + asm volatile ("mrs %0, clidr_el1" : "=r" (clidr)); > > Can't that allocate a w register? > That should be fine, as all of these cache info registers are 32-bit. > You can make clidr a u64 to avoid that. > What would be the preference ? Using w registers for all these cache registers or using u64 with x registers? >> + return CLIDR_CTYPE(clidr, level); >> +} >> + >> +/* >> + * NumSets, bits[27:13] - (Number of sets in cache) - 1 >> + * Associativity, bits[12:3] - (Associativity of cache) - 1 >> + * LineSize, bits[2:0] - (Log2(Number of words in cache line)) - 2 >> + */ >> +#define CCSIDR_WRITE_THROUGH BIT(31) >> +#define CCSIDR_WRITE_BACK BIT(30) >> +#define CCSIDR_READ_ALLOCATE BIT(29) >> +#define CCSIDR_WRITE_ALLOCATE BIT(28) >> +#define CCSIDR_LINESIZE_MASK 0x7 >> +#define CCSIDR_ASSOCIAT_SHIFT 3 >> +#define CCSIDR_ASSOCIAT_MASK 0x3FF > > ASSOCIAT doesn't quite roll off of the tongue... > I have no idea why I chose that incomplete name :( >> +#define CCSIDR_NUMSETS_SHIFT 13 >> +#define CCSIDR_NUMSETS_MASK 0x7FF >> + >> +/* >> + * Which cache CCSIDR represents depends on CSSELR value >> + * Make sure no one else changes CSSELR during this >> + * smp_call_function_single prevents preemption for us >> + */ >> +static inline u32 get_ccsidr(u32 csselr) >> +{ >> + u32 ccsidr; >> + >> + /* Put value into CSSELR */ >> + asm volatile("msr csselr_el1, %x0" : : "r" (csselr)); > > This looks a little dodgy. I think GCC can leave the upper 32 bits in a > random state. Why not cast csselr to a u64 here? > >> + isb(); >> + /* Read result out of CCSIDR */ >> + asm volatile("mrs %0, ccsidr_el1" : "=r" (ccsidr)); >> + >> + return ccsidr; > > Similarly it might make sense to make the temporary variable a u64. > > [...] > >> +int init_cache_level(unsigned int cpu) >> +{ >> + unsigned int ctype, level = 1, leaves = 0; >> + struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); >> + >> + if (!this_cpu_ci) >> + return -EINVAL; >> + >> + do { >> + ctype = get_cache_type(level); >> + if (ctype == CACHE_TYPE_NOCACHE) >> + break; >> + /* Separate instruction and data caches */ >> + leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1; >> + } while (++level <= MAX_CACHE_LEVEL); > > I think this would be clearer with: > > for (level = 1; level <= MAX_CACHE_LEVEL; level++) > > We do something like that in populate_cache_leaves below. > Right will change it. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
On Fri, Jun 27, 2014 at 12:22:17PM +0100, Sudeep Holla wrote: > Hi Mark, > > Thanks for the review. > > On 27/06/14 11:36, Mark Rutland wrote: > > Hi Sudeep, > > > > On Wed, Jun 25, 2014 at 06:30:42PM +0100, Sudeep Holla wrote: > >> From: Sudeep Holla <sudeep.holla@arm.com> > >> > >> This patch adds support for cacheinfo on ARM64. > >> > >> On ARMv8, the cache hierarchy can be identified through Cache Level ID > >> (CLIDR) register while the cache geometry is provided by Cache Size ID > >> (CCSIDR) register. > >> > >> Since the architecture doesn't provide any way of detecting the cpus > >> sharing particular cache, device tree is used for the same purpose. > >> > >> Signed-off-by: Sudeep Holla <sudeep.holla@arm.com> > >> Cc: Catalin Marinas <catalin.marinas@arm.com> > >> Cc: Will Deacon <will.deacon@arm.com> > >> Cc: Lorenzo Pieralisi <lorenzo.pieralisi@arm.com> > >> Cc: linux-arm-kernel@lists.infradead.org > >> --- > >> arch/arm64/kernel/Makefile | 3 +- > >> arch/arm64/kernel/cacheinfo.c | 135 ++++++++++++++++++++++++++++++++++++++++++ > >> 2 files changed, 137 insertions(+), 1 deletion(-) > >> create mode 100644 arch/arm64/kernel/cacheinfo.c > > > > [...] > > > >> +static inline enum cache_type get_cache_type(int level) > >> +{ > >> + unsigned int clidr; > >> + > >> + if (level > MAX_CACHE_LEVEL) > >> + return CACHE_TYPE_NOCACHE; > >> + asm volatile ("mrs %0, clidr_el1" : "=r" (clidr)); > > > > Can't that allocate a w register? > > > > That should be fine, as all of these cache info registers are 32-bit. In A64 mrs/msr only works for x registers, and gas will barf for w registers: [mark@leverpostej:~]% echo "mrs x0, clidr_el1" | aarch64-linux-gnu-as - [mark@leverpostej:~]% echo "mrs w0, clidr_el1" | aarch64-linux-gnu-as - {standard input}: Assembler messages: {standard input}:1: Error: operand mismatch -- `mrs w0,clidr_el1' [mark@leverpostej:~]% echo "msr clidr_el1, x0" | aarch64-linux-gnu-as - [mark@leverpostej:~]% echo "msr clidr_el1, w0" | aarch64-linux-gnu-as - {standard input}: Assembler messages: {standard input}:1: Error: operand mismatch -- `msr clidr_el1,w0' [mark@leverpostej:~]% > > You can make clidr a u64 to avoid that. > > > > What would be the preference ? > Using w registers for all these cache registers or using u64 with x registers? You must use x registers. To prevent GCC from making the assumption that the upper 32-bits are irrelevant, it's better to cast to a u64 than use %xN in the asm. > >> + return CLIDR_CTYPE(clidr, level); > >> +} > >> + > >> +/* > >> + * NumSets, bits[27:13] - (Number of sets in cache) - 1 > >> + * Associativity, bits[12:3] - (Associativity of cache) - 1 > >> + * LineSize, bits[2:0] - (Log2(Number of words in cache line)) - 2 > >> + */ > >> +#define CCSIDR_WRITE_THROUGH BIT(31) > >> +#define CCSIDR_WRITE_BACK BIT(30) > >> +#define CCSIDR_READ_ALLOCATE BIT(29) > >> +#define CCSIDR_WRITE_ALLOCATE BIT(28) > >> +#define CCSIDR_LINESIZE_MASK 0x7 > >> +#define CCSIDR_ASSOCIAT_SHIFT 3 > >> +#define CCSIDR_ASSOCIAT_MASK 0x3FF > > > > ASSOCIAT doesn't quite roll off of the tongue... > > > > I have no idea why I chose that incomplete name :( At least we can fix it :) Cheers, Mark. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
diff --git a/arch/arm64/kernel/Makefile b/arch/arm64/kernel/Makefile index cdaedad..754a3d0 100644 --- a/arch/arm64/kernel/Makefile +++ b/arch/arm64/kernel/Makefile @@ -15,7 +15,8 @@ CFLAGS_REMOVE_return_address.o = -pg arm64-obj-y := cputable.o debug-monitors.o entry.o irq.o fpsimd.o \ entry-fpsimd.o process.o ptrace.o setup.o signal.o \ sys.o stacktrace.o time.o traps.o io.o vdso.o \ - hyp-stub.o psci.o cpu_ops.o insn.o return_address.o + hyp-stub.o psci.o cpu_ops.o insn.o return_address.o \ + cacheinfo.o arm64-obj-$(CONFIG_COMPAT) += sys32.o kuser32.o signal32.o \ sys_compat.o diff --git a/arch/arm64/kernel/cacheinfo.c b/arch/arm64/kernel/cacheinfo.c new file mode 100644 index 0000000..a6f81cc --- /dev/null +++ b/arch/arm64/kernel/cacheinfo.c @@ -0,0 +1,135 @@ +/* + * ARM64 cacheinfo support + * + * Copyright (C) 2014 ARM Ltd. + * All Rights Reserved + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 as + * published by the Free Software Foundation. + */ + +#include <linux/bitops.h> +#include <linux/cacheinfo.h> +#include <linux/cpu.h> +#include <linux/compiler.h> +#include <linux/of.h> + +#include <asm/processor.h> + +#define MAX_CACHE_LEVEL 7 /* Max 7 level supported */ +/* Ctypen, bits[3(n - 1) + 2 : 3(n - 1)], for n = 1 to 7 */ +#define CLIDR_CTYPE_SHIFT(level) (3 * (level - 1)) +#define CLIDR_CTYPE_MASK(level) (7 << CLIDR_CTYPE_SHIFT(level)) +#define CLIDR_CTYPE(clidr, level) \ + (((clidr) & CLIDR_CTYPE_MASK(level)) >> CLIDR_CTYPE_SHIFT(level)) + +static inline enum cache_type get_cache_type(int level) +{ + unsigned int clidr; + + if (level > MAX_CACHE_LEVEL) + return CACHE_TYPE_NOCACHE; + asm volatile ("mrs %0, clidr_el1" : "=r" (clidr)); + return CLIDR_CTYPE(clidr, level); +} + +/* + * NumSets, bits[27:13] - (Number of sets in cache) - 1 + * Associativity, bits[12:3] - (Associativity of cache) - 1 + * LineSize, bits[2:0] - (Log2(Number of words in cache line)) - 2 + */ +#define CCSIDR_WRITE_THROUGH BIT(31) +#define CCSIDR_WRITE_BACK BIT(30) +#define CCSIDR_READ_ALLOCATE BIT(29) +#define CCSIDR_WRITE_ALLOCATE BIT(28) +#define CCSIDR_LINESIZE_MASK 0x7 +#define CCSIDR_ASSOCIAT_SHIFT 3 +#define CCSIDR_ASSOCIAT_MASK 0x3FF +#define CCSIDR_NUMSETS_SHIFT 13 +#define CCSIDR_NUMSETS_MASK 0x7FF + +/* + * Which cache CCSIDR represents depends on CSSELR value + * Make sure no one else changes CSSELR during this + * smp_call_function_single prevents preemption for us + */ +static inline u32 get_ccsidr(u32 csselr) +{ + u32 ccsidr; + + /* Put value into CSSELR */ + asm volatile("msr csselr_el1, %x0" : : "r" (csselr)); + isb(); + /* Read result out of CCSIDR */ + asm volatile("mrs %0, ccsidr_el1" : "=r" (ccsidr)); + + return ccsidr; +} + +static void ci_leaf_init(struct cacheinfo *this_leaf, + enum cache_type type, unsigned int level) +{ + bool is_instr_cache = type & CACHE_TYPE_INST; + u32 tmp = get_ccsidr((level - 1) << 1 | is_instr_cache); + + this_leaf->level = level; + this_leaf->type = type; + this_leaf->coherency_line_size = + (1 << ((tmp & CCSIDR_LINESIZE_MASK) + 2)) * 4; + this_leaf->number_of_sets = + ((tmp >> CCSIDR_NUMSETS_SHIFT) & CCSIDR_NUMSETS_MASK) + 1; + this_leaf->ways_of_associativity = + ((tmp >> CCSIDR_ASSOCIAT_SHIFT) & CCSIDR_ASSOCIAT_MASK) + 1; + this_leaf->size = this_leaf->number_of_sets * + this_leaf->coherency_line_size * this_leaf->ways_of_associativity; + this_leaf->attributes = + ((tmp & CCSIDR_WRITE_THROUGH) ? CACHE_WRITE_THROUGH : 0) | + ((tmp & CCSIDR_WRITE_BACK) ? CACHE_WRITE_BACK : 0) | + ((tmp & CCSIDR_READ_ALLOCATE) ? CACHE_READ_ALLOCATE : 0) | + ((tmp & CCSIDR_WRITE_ALLOCATE) ? CACHE_WRITE_ALLOCATE : 0); +} + +int init_cache_level(unsigned int cpu) +{ + unsigned int ctype, level = 1, leaves = 0; + struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); + + if (!this_cpu_ci) + return -EINVAL; + + do { + ctype = get_cache_type(level); + if (ctype == CACHE_TYPE_NOCACHE) + break; + /* Separate instruction and data caches */ + leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1; + } while (++level <= MAX_CACHE_LEVEL); + + this_cpu_ci->num_levels = level - 1; + this_cpu_ci->num_leaves = leaves; + return 0; +} + +int populate_cache_leaves(unsigned int cpu) +{ + unsigned int level, idx; + enum cache_type type; + struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); + struct cacheinfo *this_leaf = this_cpu_ci->info_list; + + for (idx = 0, level = 1; level <= this_cpu_ci->num_levels && + idx < this_cpu_ci->num_leaves; idx++, level++) { + if (!this_leaf) + return -EINVAL; + + type = get_cache_type(level); + if (type == CACHE_TYPE_SEPARATE) { + ci_leaf_init(this_leaf++, CACHE_TYPE_DATA, level); + ci_leaf_init(this_leaf++, CACHE_TYPE_INST, level); + } else { + ci_leaf_init(this_leaf++, type, level); + } + } + return 0; +}