From patchwork Wed Mar 4 13:25:45 2015 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Daniel Thompson X-Patchwork-Id: 45413 Return-Path: X-Original-To: linaro@patches.linaro.org Delivered-To: linaro@patches.linaro.org Received: from mail-wi0-f200.google.com (mail-wi0-f200.google.com [209.85.212.200]) by ip-10-151-82-157.ec2.internal (Postfix) with ESMTPS id 393C721486 for ; Wed, 4 Mar 2015 13:26:05 +0000 (UTC) Received: by widfb4 with SMTP id fb4sf7529723wid.2 for ; Wed, 04 Mar 2015 05:26:04 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:delivered-to:from:to:cc:subject :date:message-id:in-reply-to:references:x-original-sender :x-original-authentication-results:precedence:mailing-list:list-id :list-post:list-help:list-archive:list-unsubscribe; bh=4kg+9XNCxijuAwwuM6BZrQaWOzfqwKcwbsKJXuGE820=; b=d6wdaidmGDbsiUdz6zkTvGUrg05onpMm3jcFOxVyERBix0Ez4qiOwslhsbeniBkDiM JhxDuj09HZMSFUZM/rwVXSvgm7vcs5dCuCjIoevai+oYMDNmTJqOL5J5Q8H69p2VAJTh hiQfTKc48VPI1ZLCE/sgWiGtxnFsyJkOEG6KFwdvDXGK+A39nmbntknJNxsu4TG6C7i+ CTYAqMdwjgYXdvfZbNYqBXXi1PTXbJb17wv6UZSRBXGri4MSlhsaqiutWo19A++PtRsh 1tLt0Q6bxMK1ipp0S8wKUMRzgcBT9mBaV0sF8iFB4Kn+8JMzV53Xi7Ot4dOBz7rutuEC 7Onw== X-Gm-Message-State: ALoCoQnVEQJaHb3kpWi2cGFBkfSLuZhDfa75o2+8c9CC+Bo9rEnYv4TIOuZY5wO6G85YJ3gyAwSa X-Received: by 10.152.6.41 with SMTP id x9mr673681lax.10.1425475564417; Wed, 04 Mar 2015 05:26:04 -0800 (PST) MIME-Version: 1.0 X-BeenThere: patchwork-forward@linaro.org Received: by 10.152.8.102 with SMTP id q6ls177664laa.24.gmail; Wed, 04 Mar 2015 05:26:04 -0800 (PST) X-Received: by 10.112.51.114 with SMTP id j18mr3150745lbo.97.1425475564089; Wed, 04 Mar 2015 05:26:04 -0800 (PST) Received: from mail-la0-f41.google.com (mail-la0-f41.google.com. [209.85.215.41]) by mx.google.com with ESMTPS id bl10si2519753lbc.68.2015.03.04.05.26.03 for (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 04 Mar 2015 05:26:04 -0800 (PST) Received-SPF: pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 209.85.215.41 as permitted sender) client-ip=209.85.215.41; Received: by lamq1 with SMTP id q1so21304902lam.9 for ; Wed, 04 Mar 2015 05:26:03 -0800 (PST) X-Received: by 10.152.28.5 with SMTP id x5mr3291307lag.112.1425475563865; Wed, 04 Mar 2015 05:26:03 -0800 (PST) X-Forwarded-To: patchwork-forward@linaro.org X-Forwarded-For: patch@linaro.org patchwork-forward@linaro.org Delivered-To: patches@linaro.org Received: by 10.112.35.133 with SMTP id h5csp1217016lbj; Wed, 4 Mar 2015 05:26:02 -0800 (PST) X-Received: by 10.180.219.102 with SMTP id pn6mr12464707wic.50.1425475562442; Wed, 04 Mar 2015 05:26:02 -0800 (PST) Received: from mail-we0-f173.google.com (mail-we0-f173.google.com. [74.125.82.173]) by mx.google.com with ESMTPS id yq1si6818525wjc.153.2015.03.04.05.26.02 for (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 04 Mar 2015 05:26:02 -0800 (PST) Received-SPF: pass (google.com: domain of daniel.thompson@linaro.org designates 74.125.82.173 as permitted sender) client-ip=74.125.82.173; Received: by wevm14 with SMTP id m14so46327049wev.8 for ; Wed, 04 Mar 2015 05:26:02 -0800 (PST) X-Received: by 10.180.108.13 with SMTP id hg13mr12583585wib.7.1425475562027; Wed, 04 Mar 2015 05:26:02 -0800 (PST) Received: from wychelm.lan (cpc4-aztw19-0-0-cust71.18-1.cable.virginm.net. [82.33.25.72]) by mx.google.com with ESMTPSA id s5sm25862832wia.1.2015.03.04.05.26.00 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 04 Mar 2015 05:26:01 -0800 (PST) From: Daniel Thompson To: Russell King , Will Deacon Cc: Daniel Thompson , linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, Shawn Guo , Sascha Hauer , Peter Zijlstra , Paul Mackerras , Ingo Molnar , Arnaldo Carvalho de Melo , Thomas Gleixner , Lucas Stach , Linus Walleij , Mark Rutland , patches@linaro.org, linaro-kernel@lists.linaro.org, John Stultz , Sumit Semwal Subject: [PATCH v6] arm: perf: Directly handle SMP platforms with one SPI Date: Wed, 4 Mar 2015 13:25:45 +0000 Message-Id: <1425475545-4323-1-git-send-email-daniel.thompson@linaro.org> X-Mailer: git-send-email 2.1.0 In-Reply-To: <1416581603-30557-1-git-send-email-daniel.thompson@linaro.org> References: <1416581603-30557-1-git-send-email-daniel.thompson@linaro.org> X-Removed-Original-Auth: Dkim didn't pass. X-Original-Sender: daniel.thompson@linaro.org X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 209.85.215.41 as permitted sender) smtp.mail=patch+caf_=patchwork-forward=linaro.org@linaro.org Precedence: list Mailing-list: list patchwork-forward@linaro.org; contact patchwork-forward+owners@linaro.org List-ID: X-Google-Group-Id: 836684582541 List-Post: , List-Help: , List-Archive: List-Unsubscribe: , Some ARM platforms mux the PMU interrupt of every core into a single SPI. On such platforms if the PMU of any core except 0 raises an interrupt then it cannot be serviced and eventually, if you are lucky, the spurious irq detection might forcefully disable the interrupt. On these SoCs it is not possible to determine which core raised the interrupt so workaround this issue by queuing irqwork on the other cores whenever the primary interrupt handler is unable to service the interrupt. The u8500 platform has an alternative workaround that dynamically alters the affinity of the PMU interrupt. This workaround logic is no longer required so the original code is removed as is the hook it relied upon. Tested on imx6q (which has fours cores/PMUs all muxed to a single SPI) using a simple soak, combined perf and CPU hotplug soak and using perf fuzzer's fast_repro.sh. Signed-off-by: Daniel Thompson --- Notes: v2 was tested on u8500 (thanks to Linus Walleij). The latest patch doesn't change the nature of the workaround itself but there has been substantial churn in the logic to decide when it can safely be deployed. Thus the changes were sufficient for me not to preserve the Tested-By:. v6: * Redesigned the code that decides if it is safe to deploy the workaround (acting on the review by Mark Rutland). Code should no longer race during hot unplug; previous patches sought to make the hot unplug race benign and the old approach had flaws and even if it could be made correct was tortuously hard to review. v5: * Removed the work queue nonsense; being completely race-free requires us to take a mutex or avoid dispatch from interrupt (Will Deacon). Replacement code can potentially race with a CPU hot unplug however it is careful to minimise exposure, to mitigate harmful effects and has fairly prominent comments. v4: * Ripped out the logic that tried to preserve the operation of the spurious interrupt detector. It was complex and not really needed (Will Deacon). * Removed a redundant memory barrier and added a comment explaining why it is not needed (Will Deacon). * Made fully safe w.r.t. hotplug by falling back to a work queue if there is a hotplug operation in flight when the PMU interrupt comes in (Will Deacon). The work queue code paths have been tested synthetically (by changing the if condition). * Posted the correct, as in compilable and tested, version of the code (Will Deacon). v3: * Removed function pointer indirection when deploying workaround code and reorganise the code accordingly (Mark Rutland). * Move the workaround state tracking into the existing percpu data structure (Mark Rutland). * Renamed cret to percpu_ret and rewrote the comment describing the purpose of this variable (Mark Rutland). * Copy the cpu_online_mask and use that to act on a consistent set of cpus throughout the workaround (Mark Rutland). * Changed "single_irq" to "muxed_spi" to more explicitly describe the problem. v2: * Fixed build problems on systems without SMP. v1: * Thanks to Lucas Stach, Russell King and Thomas Gleixner for critiquing an older, completely different way to tackle the same problem. arch/arm/include/asm/pmu.h | 12 +++ arch/arm/kernel/perf_event.c | 9 +- arch/arm/kernel/perf_event_cpu.c | 179 +++++++++++++++++++++++++++++++++++---- arch/arm/kernel/perf_event_v7.c | 2 +- arch/arm/mach-ux500/cpu-db8500.c | 29 ------- 5 files changed, 178 insertions(+), 53 deletions(-) -- 2.1.0 diff --git a/arch/arm/include/asm/pmu.h b/arch/arm/include/asm/pmu.h index b1596bd59129..dfef7904b790 100644 --- a/arch/arm/include/asm/pmu.h +++ b/arch/arm/include/asm/pmu.h @@ -87,6 +87,14 @@ struct pmu_hw_events { * already have to allocate this struct per cpu. */ struct arm_pmu *percpu_pmu; + +#ifdef CONFIG_SMP + /* + * This is used to schedule workaround logic on platforms where all + * the PMUs are attached to a single SPI. + */ + struct irq_work work; +#endif }; struct arm_pmu { @@ -117,6 +125,10 @@ struct arm_pmu { struct platform_device *plat_device; struct pmu_hw_events __percpu *hw_events; struct notifier_block hotplug_nb; +#ifdef CONFIG_SMP + int muxed_spi_workaround_irq; + atomic_t remaining_irq_work; +#endif }; #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu)) diff --git a/arch/arm/kernel/perf_event.c b/arch/arm/kernel/perf_event.c index 557e128e4df0..e3fc1a0ce969 100644 --- a/arch/arm/kernel/perf_event.c +++ b/arch/arm/kernel/perf_event.c @@ -305,8 +305,6 @@ validate_group(struct perf_event *event) static irqreturn_t armpmu_dispatch_irq(int irq, void *dev) { struct arm_pmu *armpmu; - struct platform_device *plat_device; - struct arm_pmu_platdata *plat; int ret; u64 start_clock, finish_clock; @@ -317,14 +315,9 @@ static irqreturn_t armpmu_dispatch_irq(int irq, void *dev) * dereference. */ armpmu = *(void **)dev; - plat_device = armpmu->plat_device; - plat = dev_get_platdata(&plat_device->dev); start_clock = sched_clock(); - if (plat && plat->handle_irq) - ret = plat->handle_irq(irq, armpmu, armpmu->handle_irq); - else - ret = armpmu->handle_irq(irq, armpmu); + ret = armpmu->handle_irq(irq, armpmu); finish_clock = sched_clock(); perf_sample_event_took(finish_clock - start_clock); diff --git a/arch/arm/kernel/perf_event_cpu.c b/arch/arm/kernel/perf_event_cpu.c index 61b53c46edfa..d5bbd79abd4c 100644 --- a/arch/arm/kernel/perf_event_cpu.c +++ b/arch/arm/kernel/perf_event_cpu.c @@ -59,6 +59,116 @@ int perf_num_counters(void) } EXPORT_SYMBOL_GPL(perf_num_counters); +#ifdef CONFIG_SMP + +static cpumask_t down_prepare_cpu_mask; +static DEFINE_SPINLOCK(down_prepare_cpu_lock); + +/* + * Workaround logic that is distributed to all cores if the PMU has only + * a single IRQ and the CPU receiving that IRQ cannot handle it. Its + * job is to try to service the interrupt on the current CPU. It will + * also enable the IRQ again if all the other CPUs have already tried to + * service it. + */ +static void cpu_pmu_do_percpu_work(struct irq_work *w) +{ + struct pmu_hw_events *hw_events = + container_of(w, struct pmu_hw_events, work); + struct arm_pmu *cpu_pmu = hw_events->percpu_pmu; + + /* Ignore the return code, we can do nothing useful with it */ + (void) cpu_pmu->handle_irq(0, cpu_pmu); + + if (atomic_dec_and_test(&cpu_pmu->remaining_irq_work)) + enable_irq(cpu_pmu->muxed_spi_workaround_irq); +} + +/* + * Workaround for systems where all PMU interrupts are targeting a + * single SPI. + * + * The workaround will disable the interrupt and distribute irqwork to all + * the other processors in the system. Hopefully one of them will clear the + * interrupt... + * + * The workaround is only deployed when all PMU interrupts are aimed + * at a single core. As a result the workaround is never re-entered + * making it safe for us to use static data to maintain state. + */ +static void cpu_pmu_deploy_muxed_spi_workaround(struct arm_pmu *cpu_pmu) +{ + static cpumask_t irqwork_mask; + int cpu; + + disable_irq_nosync(cpu_pmu->muxed_spi_workaround_irq); + spin_lock(&down_prepare_cpu_lock); + + /* + * Combining cpu_online_mask and down_prepare_cpu_mask gives + * us the CPUs that are currently online and cannot die until + * we release down_prepare_cpu_lock. + */ + cpumask_andnot(&irqwork_mask, cpu_online_mask, &down_prepare_cpu_mask); + cpumask_clear_cpu(smp_processor_id(), &irqwork_mask); + atomic_add(cpumask_weight(&irqwork_mask), &cpu_pmu->remaining_irq_work); + + for_each_cpu(cpu, &irqwork_mask) { + struct pmu_hw_events *hw_events = + per_cpu_ptr(cpu_pmu->hw_events, cpu); + + if (!irq_work_queue_on(&hw_events->work, cpu)) + if (atomic_dec_and_test(&cpu_pmu->remaining_irq_work)) + enable_irq(cpu_pmu->muxed_spi_workaround_irq); + } + + spin_unlock(&down_prepare_cpu_lock); +} + +/* + * Called when the main interrupt handler cannot determine the source + * of interrupt. It will deploy a workaround if we are running on an SMP + * platform with only a single muxed SPI. + */ +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) +{ + if (irq_num != cpu_pmu->muxed_spi_workaround_irq) + return IRQ_NONE; + + cpu_pmu_deploy_muxed_spi_workaround(cpu_pmu); + return IRQ_HANDLED; +} + +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) +{ + struct platform_device *pmu_device = cpu_pmu->plat_device; + + atomic_set(&cpu_pmu->remaining_irq_work, 0); + cpu_pmu->muxed_spi_workaround_irq = platform_get_irq(pmu_device, 0); + + return 0; +} + +static void cpu_pmu_muxed_spi_workaround_term(struct arm_pmu *cpu_pmu) +{ + cpu_pmu->muxed_spi_workaround_irq = 0; +} +#else /* CONFIG_SMP */ +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) +{ + return 0; +} + +static void cpu_pmu_muxed_spi_workaround_term(struct arm_pmu *cpu_pmu) +{ +} + +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) +{ + return IRQ_NONE; +} +#endif /* CONFIG_SMP */ + /* Include the PMU-specific implementations. */ #include "perf_event_xscale.c" #include "perf_event_v6.c" @@ -98,6 +208,8 @@ static void cpu_pmu_free_irq(struct arm_pmu *cpu_pmu) if (irq >= 0) free_irq(irq, per_cpu_ptr(&hw_events->percpu_pmu, i)); } + + cpu_pmu_muxed_spi_workaround_term(cpu_pmu); } } @@ -155,31 +267,65 @@ static int cpu_pmu_request_irq(struct arm_pmu *cpu_pmu, irq_handler_t handler) cpumask_set_cpu(i, &cpu_pmu->active_irqs); } + + /* + * If we are running SMP and have only one interrupt source + * then get ready to share that single irq among the cores. + */ + if (nr_cpu_ids > 1 && irqs == 1) { + err = cpu_pmu_muxed_spi_workaround_init(cpu_pmu); + if (err) + return err; + } } return 0; } -/* - * PMU hardware loses all context when a CPU goes offline. - * When a CPU is hotplugged back in, since some hardware registers are - * UNKNOWN at reset, the PMU must be explicitly reset to avoid reading - * junk values out of them. - */ static int cpu_pmu_notify(struct notifier_block *b, unsigned long action, void *hcpu) { struct arm_pmu *pmu = container_of(b, struct arm_pmu, hotplug_nb); + long cpu = (long)hcpu; + unsigned long flags; + + switch (action & ~CPU_TASKS_FROZEN) { + /* + * PMU hardware loses all context when a CPU goes offline. + * When a CPU is hotplugged back in, since some hardware registers are + * UNKNOWN at reset, the PMU must be explicitly reset to avoid reading + * junk values out of them. + */ + case CPU_STARTING: + if (pmu->reset) + pmu->reset(pmu); + else + return NOTIFY_DONE; + + return NOTIFY_OK; + +#ifdef CONFIG_SMP + /* + * The workaround logic in cpu_pmu_handle_irq_none requires us + * to keep track of CPUs that may be about to die so we can + * avoid sending them irqwork. + */ + case CPU_DOWN_PREPARE: + spin_lock_irqsave(&down_prepare_cpu_lock, flags); + cpumask_set_cpu(cpu, &down_prepare_cpu_mask); + spin_unlock_irqrestore(&down_prepare_cpu_lock, flags); + return NOTIFY_OK; + + case CPU_DOWN_FAILED: + case CPU_DEAD: + spin_lock_irqsave(&down_prepare_cpu_lock, flags); + cpumask_clear_cpu(cpu, &down_prepare_cpu_mask); + spin_unlock_irqrestore(&down_prepare_cpu_lock, flags); + return NOTIFY_OK; +#endif + } - if ((action & ~CPU_TASKS_FROZEN) != CPU_STARTING) - return NOTIFY_DONE; - - if (pmu->reset) - pmu->reset(pmu); - else - return NOTIFY_DONE; - - return NOTIFY_OK; + return NOTIFY_DONE; } static int cpu_pmu_init(struct arm_pmu *cpu_pmu) @@ -201,6 +347,9 @@ static int cpu_pmu_init(struct arm_pmu *cpu_pmu) struct pmu_hw_events *events = per_cpu_ptr(cpu_hw_events, cpu); raw_spin_lock_init(&events->pmu_lock); events->percpu_pmu = cpu_pmu; +#ifdef CONFIG_SMP + init_irq_work(&events->work, cpu_pmu_do_percpu_work); +#endif } cpu_pmu->hw_events = cpu_hw_events; diff --git a/arch/arm/kernel/perf_event_v7.c b/arch/arm/kernel/perf_event_v7.c index 8993770c47de..0dd914c10803 100644 --- a/arch/arm/kernel/perf_event_v7.c +++ b/arch/arm/kernel/perf_event_v7.c @@ -792,7 +792,7 @@ static irqreturn_t armv7pmu_handle_irq(int irq_num, void *dev) * Did an overflow occur? */ if (!armv7_pmnc_has_overflowed(pmnc)) - return IRQ_NONE; + return cpu_pmu_handle_irq_none(irq_num, cpu_pmu); /* * Handle the counter(s) overflow(s) diff --git a/arch/arm/mach-ux500/cpu-db8500.c b/arch/arm/mach-ux500/cpu-db8500.c index 6f63954c8bde..917774999c5c 100644 --- a/arch/arm/mach-ux500/cpu-db8500.c +++ b/arch/arm/mach-ux500/cpu-db8500.c @@ -12,8 +12,6 @@ #include #include #include -#include -#include #include #include #include @@ -23,7 +21,6 @@ #include #include -#include #include #include "setup.h" @@ -99,30 +96,6 @@ static void __init u8500_map_io(void) iotable_init(u8500_io_desc, ARRAY_SIZE(u8500_io_desc)); } -/* - * The PMU IRQ lines of two cores are wired together into a single interrupt. - * Bounce the interrupt to the other core if it's not ours. - */ -static irqreturn_t db8500_pmu_handler(int irq, void *dev, irq_handler_t handler) -{ - irqreturn_t ret = handler(irq, dev); - int other = !smp_processor_id(); - - if (ret == IRQ_NONE && cpu_online(other)) - irq_set_affinity(irq, cpumask_of(other)); - - /* - * We should be able to get away with the amount of IRQ_NONEs we give, - * while still having the spurious IRQ detection code kick in if the - * interrupt really starts hitting spuriously. - */ - return ret; -} - -static struct arm_pmu_platdata db8500_pmu_platdata = { - .handle_irq = db8500_pmu_handler, -}; - static const char *db8500_read_soc_id(void) { void __iomem *uid = __io_address(U8500_BB_UID_BASE); @@ -143,8 +116,6 @@ static struct device * __init db8500_soc_device_init(void) } static struct of_dev_auxdata u8500_auxdata_lookup[] __initdata = { - /* Requires call-back bindings. */ - OF_DEV_AUXDATA("arm,cortex-a9-pmu", 0, "arm-pmu", &db8500_pmu_platdata), /* Requires DMA bindings. */ OF_DEV_AUXDATA("stericsson,ux500-msp-i2s", 0x80123000, "ux500-msp-i2s.0", &msp0_platform_data),