Message ID | 1519226968-19821-7-git-send-email-daniel.lezcano@linaro.org |
---|---|
State | New |
Headers | show |
Series | None | expand |
On 21-02-18, 16:29, Daniel Lezcano wrote: > diff --git a/drivers/thermal/cpu_cooling.c b/drivers/thermal/cpu_cooling.c > index 5c219dc..9340216 100644 > --- a/drivers/thermal/cpu_cooling.c > +++ b/drivers/thermal/cpu_cooling.c > @@ -10,18 +10,32 @@ > * Viresh Kumar <viresh.kumar@linaro.org> > * > */ > +#undef DEBUG Why is this required ? > +#define pr_fmt(fmt) "CPU cooling: " fmt I think you can use the dev_***() routines instead, as you can directly the CPU device from anywhere. > #include <linux/module.h> > #include <linux/thermal.h> > #include <linux/cpufreq.h> > +#include <linux/cpuidle.h> > #include <linux/err.h> > +#include <linux/freezer.h> > #include <linux/idr.h> > +#include <linux/kthread.h> > #include <linux/pm_opp.h> > #include <linux/slab.h> > +#include <linux/sched/prio.h> > +#include <linux/sched/rt.h> > #include <linux/cpu.h> > #include <linux/cpu_cooling.h> > +#include <linux/wait.h> > + > +#include <linux/platform_device.h> > +#include <linux/of_platform.h> > > #include <trace/events/thermal.h> > > +#include <uapi/linux/sched/types.h> > + > #ifdef CONFIG_CPU_FREQ_THERMAL > /* > * Cooling state <-> CPUFreq frequency > @@ -928,3 +942,440 @@ void cpufreq_cooling_unregister(struct thermal_cooling_device *cdev) > } > EXPORT_SYMBOL_GPL(cpufreq_cooling_unregister); > #endif /* CONFIG_CPU_FREQ_THERMAL */ > + > +#ifdef CONFIG_CPU_IDLE_THERMAL > +/* > + * The idle duration injection. As we don't have yet a way to specify > + * from the DT configuration, let's default to a tick duration. > + */ > +#define DEFAULT_IDLE_TIME_US TICK_USEC I think this macro is a bit unnecessary here. Its used only at initialization and so the same comment can be present there and you can use TICK_USEC there. Else, Keep it as it is and remove the "idle_cycle" field from the below structure, as it holds a constant forever. > + > +/** > + * struct cpuidle_cooling_device - data for the idle cooling device > + * @cdev: a pointer to a struct thermal_cooling_device > + * @cpumask: a cpumask containing the CPU managed by the cooling device Missed @node here. > + * @timer: a hrtimer giving the tempo for the idle injection cycles > + * @kref: a kernel refcount on this structure > + * @count: an atomic to keep track of the last task exiting the idle cycle > + * @idle_cycle: an integer defining the duration of the idle injection > + * @state: an normalized integer giving the state of the cooling device > + */ > +struct cpuidle_cooling_device { > + struct thermal_cooling_device *cdev; > + struct cpumask *cpumask; > + struct list_head node; > + struct hrtimer timer; > + struct kref kref; > + atomic_t count; > + unsigned int idle_cycle; > + unsigned int state; > +}; > + > +/** > + * @tsk: an array of pointer to the idle injection tasks > + * @waitq: the waiq for the idle injection tasks > + */ > +struct cpuidle_cooling_tsk { > + struct task_struct *tsk; > + wait_queue_head_t waitq; > +}; > + > +DEFINE_PER_CPU(struct cpuidle_cooling_tsk, cpuidle_cooling_tsk); static ? > + > +static LIST_HEAD(cpuidle_cdev_list); > + > +/** > + * cpuidle_cooling_wakeup - Wake up all idle injection threads > + * @idle_cdev: the idle cooling device > + * > + * Every idle injection task belonging to the idle cooling device and > + * running on an online cpu will be wake up by this call. > + */ > +static void cpuidle_cooling_wakeup(struct cpuidle_cooling_device *idle_cdev) > +{ > + int cpu; > + struct cpuidle_cooling_tsk *cct; > + > + for_each_cpu_and(cpu, idle_cdev->cpumask, cpu_online_mask) { > + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); > + wake_up_process(cct->tsk); > + } > +} > + > +/** > + * cpuidle_cooling_wakeup_fn - Running cycle timer callback > + * @timer: a hrtimer structure > + * > + * When the mitigation is acting, the CPU is allowed to run an amount > + * of time, then the idle injection happens for the specified delay > + * and the idle task injection schedules itself until the timer event > + * wakes the idle injection tasks again for a new idle injection > + * cycle. The time between the end of the idle injection and the timer > + * expiration is the allocated running time for the CPU. > + * > + * Returns always HRTIMER_NORESTART > + */ > +static enum hrtimer_restart cpuidle_cooling_wakeup_fn(struct hrtimer *timer) > +{ > + struct cpuidle_cooling_device *idle_cdev = > + container_of(timer, struct cpuidle_cooling_device, timer); > + > + cpuidle_cooling_wakeup(idle_cdev); > + > + return HRTIMER_NORESTART; > +} > + > +/** > + * cpuidle_cooling_runtime - Running time computation > + * @idle_cdev: the idle cooling device > + * > + * The running duration is computed from the idle injection duration > + * which is fixed. If we reach 100% of idle injection ratio, that > + * means the running duration is zero. If we have a 50% ratio > + * injection, that means we have equal duration for idle and for > + * running duration. > + * > + * The formula is deduced as the following: > + * > + * running = idle x ((100 / ratio) - 1) > + * > + * For precision purpose for integer math, we use the following: > + * > + * running = (idle x 100) / ratio - idle > + * > + * For example, if we have an injected duration of 50%, then we end up > + * with 10ms of idle injection and 10ms of running duration. > + * > + * Returns a s64 nanosecond based > + */ > +static s64 cpuidle_cooling_runtime(struct cpuidle_cooling_device *idle_cdev) > +{ > + s64 next_wakeup; > + int state = idle_cdev->state; > + > + /* > + * The function must never be called when there is no > + * mitigation because: > + * - that does not make sense > + * - we end up with a division by zero > + */ > + BUG_ON(!state); As there is no locking in place, we can surely hit this case. What if the state changed to 0 right before this routine was called ? I would suggest we should just return 0 in that case and get away with the BUG_ON(). > + > + next_wakeup = (s64)((idle_cdev->idle_cycle * 100) / state) - > + idle_cdev->idle_cycle; > + > + return next_wakeup * NSEC_PER_USEC; > +} > + > +/** > + * cpuidle_cooling_injection_thread - Idle injection mainloop thread function > + * @arg: a void pointer containing the idle cooling device address > + * > + * This main function does basically two operations: > + * > + * - Goes idle for a specific amount of time > + * > + * - Sets a timer to wake up all the idle injection threads after a > + * running period > + * > + * That happens only when the mitigation is enabled, otherwise the > + * task is scheduled out. > + * > + * In order to keep the tasks synchronized together, it is the last > + * task exiting the idle period which is in charge of setting the > + * timer. > + * > + * This function never returns. > + */ > +static int cpuidle_cooling_injection_thread(void *arg) > +{ > + struct sched_param param = { .sched_priority = MAX_USER_RT_PRIO/2 }; > + struct cpuidle_cooling_device *idle_cdev = arg; > + struct cpuidle_cooling_tsk *cct = per_cpu_ptr(&cpuidle_cooling_tsk, > + smp_processor_id()); this_cpu_ptr ? > + DEFINE_WAIT(wait); > + > + set_freezable(); > + > + sched_setscheduler(current, SCHED_FIFO, ¶m); > + > + while (1) { > + s64 next_wakeup; > + > + prepare_to_wait(&cct->waitq, &wait, TASK_INTERRUPTIBLE); > + > + schedule(); > + > + atomic_inc(&idle_cdev->count); > + > + play_idle(idle_cdev->idle_cycle / USEC_PER_MSEC); > + > + /* > + * The last CPU waking up is in charge of setting the > + * timer. If the CPU is hotplugged, the timer will > + * move to another CPU (which may not belong to the > + * same cluster) but that is not a problem as the > + * timer will be set again by another CPU belonging to > + * the cluster, so this mechanism is self adaptive and > + * does not require any hotplugging dance. > + */ Well this depends on how CPU hotplug really happens. What happens to the per-cpu-tasks which are in the middle of something when hotplug happens? Does hotplug wait for those per-cpu-tasks to finish ? > + if (!atomic_dec_and_test(&idle_cdev->count)) > + continue; > + > + if (!idle_cdev->state) > + continue; > + > + next_wakeup = cpuidle_cooling_runtime(idle_cdev); > + > + hrtimer_start(&idle_cdev->timer, ns_to_ktime(next_wakeup), > + HRTIMER_MODE_REL_PINNED); > + } > + > + finish_wait(&cct->waitq, &wait); > + > + return 0; > +} > + > +/** > + * cpuidle_cooling_release - Kref based release helper > + * @kref: a pointer to the kref structure > + * > + * This function is automatically called by the kref_put function when > + * the idle cooling device refcount reaches zero. At this point, we > + * have the guarantee the structure is no longer in use and we can > + * safely release all the ressources. > + */ > +static void __init cpuidle_cooling_release(struct kref *kref) > +{ > + struct cpuidle_cooling_device *idle_cdev = > + container_of(kref, struct cpuidle_cooling_device, kref); > + > + thermal_cooling_device_unregister(idle_cdev->cdev); > + kfree(idle_cdev); > +} Maybe just move it closer to the only function that calls it? > + > +/** > + * cpuidle_cooling_get_max_state - Get the maximum state > + * @cdev : the thermal cooling device > + * @state : a pointer to the state variable to be filled > + * > + * The function gives always 100 as the injection ratio is percentile > + * based for consistency accros different platforms. > + * > + * The function can not fail, it returns always zero. > + */ > +static int cpuidle_cooling_get_max_state(struct thermal_cooling_device *cdev, > + unsigned long *state) > +{ > + /* > + * Depending on the configuration or the hardware, the running > + * cycle and the idle cycle could be different. We want unify > + * that to an 0..100 interval, so the set state interface will > + * be the same whatever the platform is. > + * > + * The state 100% will make the cluster 100% ... idle. A 0% > + * injection ratio means no idle injection at all and 50% > + * means for 10ms of idle injection, we have 10ms of running > + * time. > + */ > + *state = 100; > + > + return 0; > +} > + > +/** > + * cpuidle_cooling_get_cur_state - Get the current cooling state > + * @cdev: the thermal cooling device > + * @state: a pointer to the state > + * > + * The function just copy the state value from the private thermal > + * cooling device structure, the mapping is 1 <-> 1. > + * > + * The function can not fail, it returns always zero. > + */ > +static int cpuidle_cooling_get_cur_state(struct thermal_cooling_device *cdev, > + unsigned long *state) > +{ > + struct cpuidle_cooling_device *idle_cdev = cdev->devdata; This line isn't aligned properly. Spaces are present instead of a tab. > + > + *state = idle_cdev->state; > + > + return 0; > +} > + > +/** > + * cpuidle_cooling_set_cur_state - Set the current cooling state > + * @cdev: the thermal cooling device > + * @state: the target state > + * > + * The function checks first if we are initiating the mitigation which > + * in turn wakes up all the idle injection tasks belonging to the idle > + * cooling device. In any case, it updates the internal state for the > + * cooling device. > + * > + * The function can not fail, it returns always zero. it always returns zero. Same at other places as well. > + */ > +static int cpuidle_cooling_set_cur_state(struct thermal_cooling_device *cdev, > + unsigned long state) > +{ > + struct cpuidle_cooling_device *idle_cdev = cdev->devdata; > + unsigned long current_state = idle_cdev->state; > + > + idle_cdev->state = state; > + > + if (current_state == 0 && state > 0) { > + pr_debug("Starting cooling cpus '%*pbl'\n", > + cpumask_pr_args(idle_cdev->cpumask)); > + cpuidle_cooling_wakeup(idle_cdev); > + } else if (current_state > 0 && !state) { > + pr_debug("Stopping cooling cpus '%*pbl'\n", > + cpumask_pr_args(idle_cdev->cpumask)); > + } > + > + return 0; > +} > + > +/** > + * cpuidle_cooling_ops - thermal cooling device ops > + */ > +static struct thermal_cooling_device_ops cpuidle_cooling_ops = { > + .get_max_state = cpuidle_cooling_get_max_state, > + .get_cur_state = cpuidle_cooling_get_cur_state, > + .set_cur_state = cpuidle_cooling_set_cur_state, > +}; > + > +/** > + * cpuilde_cooling_unregister - Idle cooling device exit function > + * > + * This function unregisters the cpuidle cooling device and frees the > + * ressources previously allocated by the init function. This function > + * is called when the initialization fails. > + */ > +static void cpuidle_cooling_unregister(void) > +{ > + struct cpuidle_cooling_device *tmp, *idle_cdev = NULL; > + struct cpuidle_cooling_tsk *cct; > + int cpu; > + > + list_for_each_entry_safe(idle_cdev, tmp, &cpuidle_cdev_list, node) { > + for_each_cpu(cpu, idle_cdev->cpumask) { > + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); > + if (cct->tsk) > + kthread_stop(cct->tsk); What about hrtimer ? Shouldn't that be stopped as well ? > + kref_put(&idle_cdev->kref, cpuidle_cooling_release); > + } > + } > +} > + > +/** > + * cpuidle_cooling_register - Idle cooling device initialization function > + * > + * This function is in charge of creating a cooling device per cluster > + * and register it to thermal framework. For this we rely on the > + * topology as there is nothing yet describing better the idle state > + * power domains. > + * > + * For each first CPU of the cluster's cpumask, we allocate the idle > + * cooling device, initialize the general fields and then we initialze > + * the rest in a per cpu basis. > + * > + * Returns zero on success, < 0 otherwise. This wouldn't get shown in the doc properly, it should be written as: * Return: zero on success, < 0 otherwise. > + */ > +int cpuidle_cooling_register(void) > +{ > + struct cpuidle_cooling_device *idle_cdev = NULL; > + struct thermal_cooling_device *cdev; > + struct cpuidle_cooling_tsk *cct; > + struct task_struct *tsk; > + struct device_node *np; > + cpumask_t *cpumask; > + char dev_name[THERMAL_NAME_LENGTH]; > + int ret = -ENOMEM, cpu; > + int index = 0; > + > + for_each_possible_cpu(cpu) { > + cpumask = topology_core_cpumask(cpu); > + > + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); > + > + /* > + * This condition makes the first cpu belonging to the > + * cluster to create a cooling device and allocates > + * the structure. Others CPUs belonging to the same > + * cluster will just increment the refcount on the > + * cooling device structure and initialize it. > + */ > + if (cpu == cpumask_first(cpumask)) { Your function still have few assumptions of cpu numbering and it will break in few cases. What if the CPUs on a big Little system (4x4) are present in this order: B L L L L B B B ?? This configuration can happen if CPUs in DT are marked as: 0-3 LITTLE, 4-7 big and a big CPU is used by the boot loader to bring up Linux. > + np = of_cpu_device_node_get(cpu); > + > + idle_cdev = kzalloc(sizeof(*idle_cdev), GFP_KERNEL); > + if (!idle_cdev) > + goto out_fail; > + > + idle_cdev->idle_cycle = DEFAULT_IDLE_TIME_US; > + > + atomic_set(&idle_cdev->count, 0); This should already be 0, isn't it ? > + > + kref_init(&idle_cdev->kref); > + > + /* > + * Initialize the timer to wakeup all the idle > + * injection tasks > + */ > + hrtimer_init(&idle_cdev->timer, > + CLOCK_MONOTONIC, HRTIMER_MODE_REL); > + > + /* > + * The wakeup function callback which is in > + * charge of waking up all CPUs belonging to > + * the same cluster > + */ > + idle_cdev->timer.function = cpuidle_cooling_wakeup_fn; > + > + /* > + * The thermal cooling device name > + */ > + snprintf(dev_name, sizeof(dev_name), "thermal-idle-%d", index++); > + cdev = thermal_of_cooling_device_register(np, dev_name, > + idle_cdev, > + &cpuidle_cooling_ops); You registered the cooling device, while the idle_cdev is still getting initialized. Ideally, we should register with the thermal core (or any other framework) when we are fully ready. For example, any of the callbacks present in cpuidle_cooling_ops() can get called by the core after this point and you should be ready to handle them. It will result in kernel crash in your case as idle_cdev isn't fully initialized yet. For example, with the set-state callback you will end up calling wake_up_task(NULL). > + if (IS_ERR(cdev)) { > + ret = PTR_ERR(cdev); > + goto out_fail; > + } > + > + idle_cdev->cdev = cdev; > + > + idle_cdev->cpumask = cpumask; > + > + list_add(&idle_cdev->node, &cpuidle_cdev_list); I would have removed the above two blank lines as these can all go together. > + > + pr_info("Created idle cooling device for cpus '%*pbl'\n", > + cpumask_pr_args(cpumask)); I am not sure if it makes sense to print the message right here. We can still fail while creating the cooling devices. Maybe a single print at the very end of the function is more than enough? > + } > + > + kref_get(&idle_cdev->kref); Did you check if the kref thing is really working here or not? I think your structure will never get freed on errors because kref_init() initializes the count by 1 and then you do an additional kref_get() here for the first CPU of the cluster. > + > + init_waitqueue_head(&cct->waitq); > + > + tsk = kthread_create_on_cpu(cpuidle_cooling_injection_thread, > + idle_cdev, cpu, "kidle_inject/%u"); > + if (IS_ERR(tsk)) { > + ret = PTR_ERR(tsk); > + goto out_fail; > + } > + > + cct->tsk = tsk; > + > + wake_up_process(tsk); > + } > + > + return 0; > + > +out_fail: > + cpuidle_cooling_unregister(); > + pr_err("Failed to create idle cooling device (%d)\n", ret); So you are already printing the error message here, just make the return type void as the caller of this can't do anything anyway. > + > + return ret; > +} > +#endif /* CONFIG_CPU_IDLE_THERMAL */ -- viresh
On 23/02/2018 16:26, Vincent Guittot wrote: > Hi Daniel, > > On 21 February 2018 at 16:29, Daniel Lezcano <daniel.lezcano@linaro.org> wrote: >> + >> +/** >> + * struct cpuidle_cooling_device - data for the idle cooling device >> + * @cdev: a pointer to a struct thermal_cooling_device >> + * @cpumask: a cpumask containing the CPU managed by the cooling device >> + * @timer: a hrtimer giving the tempo for the idle injection cycles >> + * @kref: a kernel refcount on this structure >> + * @count: an atomic to keep track of the last task exiting the idle cycle >> + * @idle_cycle: an integer defining the duration of the idle injection >> + * @state: an normalized integer giving the state of the cooling device >> + */ >> +struct cpuidle_cooling_device { >> + struct thermal_cooling_device *cdev; >> + struct cpumask *cpumask; >> + struct list_head node; >> + struct hrtimer timer; >> + struct kref kref; >> + atomic_t count; >> + unsigned int idle_cycle; >> + unsigned int state; >> +}; >> + >> +/** >> + * @tsk: an array of pointer to the idle injection tasks >> + * @waitq: the waiq for the idle injection tasks >> + */ >> +struct cpuidle_cooling_tsk { >> + struct task_struct *tsk; >> + wait_queue_head_t waitq; > > Why are you creating one wait_queue_head_t per cpu instead of one per > cooling device and then save a pointer in the per cpu struct > cpuidle_cooling_tsk ? > Then, you can use wake_up_interruptible_all() to wake up all threads > instead of using for_each_cpu ... wake_up_process() loop in > cpuidle_cooling_wakeup() ? Yes, that should do the trick. I will give it a try. <http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook | <http://twitter.com/#!/linaroorg> Twitter | <http://www.linaro.org/linaro-blog/> Blog
On 23-02-18, 12:28, Daniel Lezcano wrote: > On 23/02/2018 08:34, Viresh Kumar wrote: > > On 21-02-18, 16:29, Daniel Lezcano wrote: > >> diff --git a/drivers/thermal/cpu_cooling.c b/drivers/thermal/cpu_cooling.c > >> index 5c219dc..9340216 100644 > >> --- a/drivers/thermal/cpu_cooling.c > >> +++ b/drivers/thermal/cpu_cooling.c > >> @@ -10,18 +10,32 @@ > >> * Viresh Kumar <viresh.kumar@linaro.org> > >> * > >> */ > >> +#undef DEBUG > > > > Why is this required ? > > It is usually added, so if you set the -DDEBUG flag when compiling, you > don't get all the pr_debug traces for all files, but the just the ones > where you commented the #undef above. pr_debug is a no-op otherwise. Yeah, but this is a mess as you need to go edit the files before enabling debug with it. Everyone prefers the dynamic debug thing now, where we don't need such stuff. Just drop it. > >> +#define pr_fmt(fmt) "CPU cooling: " fmt > > > > I think you can use the dev_***() routines instead, as you can > > directly the CPU device from anywhere. > > Can we postpone this change for later ? All the file is using pr_* > (cpufreq_cooling included). There is only one place where dev_err is > used but it is removed by the patch 3/7. okay. > >> + while (1) { > >> + s64 next_wakeup; > >> + > >> + prepare_to_wait(&cct->waitq, &wait, TASK_INTERRUPTIBLE); > >> + > >> + schedule(); > >> + > >> + atomic_inc(&idle_cdev->count); > >> + > >> + play_idle(idle_cdev->idle_cycle / USEC_PER_MSEC); > >> + > >> + /* > >> + * The last CPU waking up is in charge of setting the > >> + * timer. If the CPU is hotplugged, the timer will > >> + * move to another CPU (which may not belong to the > >> + * same cluster) but that is not a problem as the > >> + * timer will be set again by another CPU belonging to > >> + * the cluster, so this mechanism is self adaptive and > >> + * does not require any hotplugging dance. > >> + */ > > > > Well this depends on how CPU hotplug really happens. What happens to > > the per-cpu-tasks which are in the middle of something when hotplug > > happens? Does hotplug wait for those per-cpu-tasks to finish ? Missed this one ? > >> +int cpuidle_cooling_register(void) > >> +{ > >> + struct cpuidle_cooling_device *idle_cdev = NULL; > >> + struct thermal_cooling_device *cdev; > >> + struct cpuidle_cooling_tsk *cct; > >> + struct task_struct *tsk; > >> + struct device_node *np; > >> + cpumask_t *cpumask; > >> + char dev_name[THERMAL_NAME_LENGTH]; > >> + int ret = -ENOMEM, cpu; > >> + int index = 0; > >> + > >> + for_each_possible_cpu(cpu) { > >> + cpumask = topology_core_cpumask(cpu); > >> + > >> + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); > >> + > >> + /* > >> + * This condition makes the first cpu belonging to the > >> + * cluster to create a cooling device and allocates > >> + * the structure. Others CPUs belonging to the same > >> + * cluster will just increment the refcount on the > >> + * cooling device structure and initialize it. > >> + */ > >> + if (cpu == cpumask_first(cpumask)) { > > > > Your function still have few assumptions of cpu numbering and it will > > break in few cases. What if the CPUs on a big Little system (4x4) are > > present in this order: B L L L L B B B ?? > > > > This configuration can happen if CPUs in DT are marked as: 0-3 LITTLE, > > 4-7 big and a big CPU is used by the boot loader to bring up Linux. > > Ok, how can I sort it out ? I would do something like this: cpumask_copy(possible, cpu_possible_mask); while (!cpumask_empty(possible)) { first = cpumask_first(possible); cpumask = topology_core_cpumask(first); cpumask_andnot(possible, possible, cpumask); allocate_cooling_dev(first); //This is most of this function in your patch. while (!cpumask_empty(cpumask)) { temp = cpumask_first(possible); //rest init "temp" cpumask_clear_cpu(temp, cpumask); } //Everything done, register cooling device for cpumask. } > >> + np = of_cpu_device_node_get(cpu); > >> + > >> + idle_cdev = kzalloc(sizeof(*idle_cdev), GFP_KERNEL); > >> + if (!idle_cdev) > >> + goto out_fail; > >> + > >> + idle_cdev->idle_cycle = DEFAULT_IDLE_TIME_US; > >> + > >> + atomic_set(&idle_cdev->count, 0); > > > > This should already be 0, isn't it ? > > Yes. I read it as, "I will drop it" :) -- viresh
On 26/02/2018 05:30, Viresh Kumar wrote: [ ... ] >>>> + /* >>>> + * The last CPU waking up is in charge of setting the >>>> + * timer. If the CPU is hotplugged, the timer will >>>> + * move to another CPU (which may not belong to the >>>> + * same cluster) but that is not a problem as the >>>> + * timer will be set again by another CPU belonging to >>>> + * the cluster, so this mechanism is self adaptive and >>>> + * does not require any hotplugging dance. >>>> + */ >>> >>> Well this depends on how CPU hotplug really happens. What happens to >>> the per-cpu-tasks which are in the middle of something when hotplug >>> happens? Does hotplug wait for those per-cpu-tasks to finish ? > > Missed this one ? To be frank, I don't know. I have been through the hp code and didn't find the answer. AFAICT, the sched/core.c sched_cpu_dying() disables the rq and then migrates the tasks. I assume this kthread is not migrated and stays in sleeping state until the rq is online again. As the wakeup callback discards offline cpus, the kthread is not wakeup at any moment. I created a situation where that happens: mitigation (raised with dhrystone) + cpu hotplugging (offline/online loop) and never hit any issue. However, I'm wondering if using the 'struct smp_hotplug_thread' infra-stucture (git show f97f8f06) shouldn't be used. >>>> +int cpuidle_cooling_register(void) >>>> +{ >>>> + struct cpuidle_cooling_device *idle_cdev = NULL; >>>> + struct thermal_cooling_device *cdev; >>>> + struct cpuidle_cooling_tsk *cct; >>>> + struct task_struct *tsk; >>>> + struct device_node *np; >>>> + cpumask_t *cpumask; >>>> + char dev_name[THERMAL_NAME_LENGTH]; >>>> + int ret = -ENOMEM, cpu; >>>> + int index = 0; >>>> + >>>> + for_each_possible_cpu(cpu) { >>>> + cpumask = topology_core_cpumask(cpu); >>>> + >>>> + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); >>>> + >>>> + /* >>>> + * This condition makes the first cpu belonging to the >>>> + * cluster to create a cooling device and allocates >>>> + * the structure. Others CPUs belonging to the same >>>> + * cluster will just increment the refcount on the >>>> + * cooling device structure and initialize it. >>>> + */ >>>> + if (cpu == cpumask_first(cpumask)) { >>> >>> Your function still have few assumptions of cpu numbering and it will >>> break in few cases. What if the CPUs on a big Little system (4x4) are >>> present in this order: B L L L L B B B ?? >>> >>> This configuration can happen if CPUs in DT are marked as: 0-3 LITTLE, >>> 4-7 big and a big CPU is used by the boot loader to bring up Linux. >> >> Ok, how can I sort it out ? > > I would do something like this: > > cpumask_copy(possible, cpu_possible_mask); > > while (!cpumask_empty(possible)) { > first = cpumask_first(possible); > cpumask = topology_core_cpumask(first); > cpumask_andnot(possible, possible, cpumask); > > allocate_cooling_dev(first); //This is most of this function in your patch. > > while (!cpumask_empty(cpumask)) { > temp = cpumask_first(possible); > //rest init "temp" > cpumask_clear_cpu(temp, cpumask); > } > > //Everything done, register cooling device for cpumask. > } > >>>> + np = of_cpu_device_node_get(cpu); >>>> + >>>> + idle_cdev = kzalloc(sizeof(*idle_cdev), GFP_KERNEL); >>>> + if (!idle_cdev) >>>> + goto out_fail; >>>> + >>>> + idle_cdev->idle_cycle = DEFAULT_IDLE_TIME_US; >>>> + >>>> + atomic_set(&idle_cdev->count, 0); >>> >>> This should already be 0, isn't it ? >> >> Yes. > > I read it as, "I will drop it" :) > -- <http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook | <http://twitter.com/#!/linaroorg> Twitter | <http://www.linaro.org/linaro-blog/> Blog
On Wed, Feb 21, 2018 at 04:29:27PM +0100, Daniel Lezcano wrote: [...] > +/** > + * cpuidle_cooling_injection_thread - Idle injection mainloop thread function > + * @arg: a void pointer containing the idle cooling device address > + * > + * This main function does basically two operations: > + * > + * - Goes idle for a specific amount of time > + * > + * - Sets a timer to wake up all the idle injection threads after a > + * running period > + * > + * That happens only when the mitigation is enabled, otherwise the > + * task is scheduled out. > + * > + * In order to keep the tasks synchronized together, it is the last > + * task exiting the idle period which is in charge of setting the > + * timer. > + * > + * This function never returns. > + */ > +static int cpuidle_cooling_injection_thread(void *arg) > +{ > + struct sched_param param = { .sched_priority = MAX_USER_RT_PRIO/2 }; I am just wandering if should set priority to (MAX_RT_PRIO - 1)? Otherwise I am concern it might be cannot enter deep idle state when any CPU idle injection thread is preempted by other higher priority RT threads so all CPUs have no alignment for idle state entering/exiting. > + struct cpuidle_cooling_device *idle_cdev = arg; > + struct cpuidle_cooling_tsk *cct = per_cpu_ptr(&cpuidle_cooling_tsk, > + smp_processor_id()); > + DEFINE_WAIT(wait); > + > + set_freezable(); > + > + sched_setscheduler(current, SCHED_FIFO, ¶m); > + > + while (1) { > + s64 next_wakeup; > + > + prepare_to_wait(&cct->waitq, &wait, TASK_INTERRUPTIBLE); > + > + schedule(); > + > + atomic_inc(&idle_cdev->count); > + > + play_idle(idle_cdev->idle_cycle / USEC_PER_MSEC); > + > + /* > + * The last CPU waking up is in charge of setting the > + * timer. If the CPU is hotplugged, the timer will > + * move to another CPU (which may not belong to the > + * same cluster) but that is not a problem as the > + * timer will be set again by another CPU belonging to > + * the cluster, so this mechanism is self adaptive and > + * does not require any hotplugging dance. > + */ > + if (!atomic_dec_and_test(&idle_cdev->count)) > + continue; > + > + if (!idle_cdev->state) > + continue; > + > + next_wakeup = cpuidle_cooling_runtime(idle_cdev); > + > + hrtimer_start(&idle_cdev->timer, ns_to_ktime(next_wakeup), > + HRTIMER_MODE_REL_PINNED); If SoC temperature descreases under tipping point, will the timer be disabled for this case? Or will here set next timer event with big value from next_wakeup? [...] Thanks, Leo Yan
On Fri, Feb 23, 2018 at 12:28:51PM +0100, Daniel Lezcano wrote: > On 23/02/2018 08:34, Viresh Kumar wrote: > > On 21-02-18, 16:29, Daniel Lezcano wrote: > [ ... ] > > >> +static s64 cpuidle_cooling_runtime(struct cpuidle_cooling_device *idle_cdev) > >> +{ > >> + s64 next_wakeup; > >> + int state = idle_cdev->state; > >> + > >> + /* > >> + * The function must never be called when there is no > >> + * mitigation because: > >> + * - that does not make sense > >> + * - we end up with a division by zero > >> + */ > >> + BUG_ON(!state); > > > > As there is no locking in place, we can surely hit this case. What if > > the state changed to 0 right before this routine was called ? > > > > I would suggest we should just return 0 in that case and get away with > > the BUG_ON(). Here if 'state' equals to 0 and we return 0, then the return value will be same with when 'state' = 100; this lets the return value confused. I think for 'state' = 0, should we return -1 so indicate the hrtimer will not be set for this case? Thanks, Leo Yan
On 27/03/2018 04:03, Leo Yan wrote: > Hi Daniel, > > On Wed, Feb 21, 2018 at 04:29:27PM +0100, Daniel Lezcano wrote: >> The cpu idle cooling driver performs synchronized idle injection across all >> cpus belonging to the same cluster and offers a new method to cool down a SoC. >> >> Each cluster has its own idle cooling device, each core has its own idle >> injection thread, each idle injection thread uses play_idle to enter idle. In >> order to reach the deepest idle state, each cooling device has the idle >> injection threads synchronized together. >> >> It has some similarity with the intel power clamp driver but it is actually >> designed to work on the ARM architecture via the DT with a mathematical proof >> with the power model which comes with the Documentation. >> >> The idle injection cycle is fixed while the running cycle is variable. That >> allows to have control on the device reactivity for the user experience. At >> the mitigation point the idle threads are unparked, they play idle the >> specified amount of time and they schedule themselves. The last thread sets >> the next idle injection deadline and when the timer expires it wakes up all >> the threads which in turn play idle again. Meanwhile the running cycle is >> changed by set_cur_state. When the mitigation ends, the threads are parked. >> The algorithm is self adaptive, so there is no need to handle hotplugging. > > The idle injection threads are RT threads (FIFO) and I saw in > play_idle() set/clear flag PF_IDLE for it. Will these idle injection > threads utilization be accounted into RT utilization? > > If idle injection threads utilization is accounted as RT tasks > utilization, will this impact CPUFreq governor 'schedutil' for OPP > selection? Hi Leo, The idle injection task has a very low utilization when it is not in the play_idle function, basically it wakes up, sets a timer and play_idle(). Regarding the use case, the idle injection is the base brick for an combo cooling device with cpufreq + cpuidle. When the idle injection is used alone, it is because there is no cpufreq driver for the platform. If there is a cpufreq driver, then we should endup with the cpu cooling device where we have control of the OPP (and there is no idle injection threads) or the combo cooling device. Except I'm missing something, the idle injection threads won't impact the OPP selection. -- <http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook | <http://twitter.com/#!/linaroorg> Twitter | <http://www.linaro.org/linaro-blog/> Blog
Hi Daniel, On 27/03/18 12:26, Daniel Lezcano wrote: > On 27/03/2018 04:03, Leo Yan wrote: > > Hi Daniel, > > > > On Wed, Feb 21, 2018 at 04:29:27PM +0100, Daniel Lezcano wrote: > >> The cpu idle cooling driver performs synchronized idle injection across all > >> cpus belonging to the same cluster and offers a new method to cool down a SoC. > >> > >> Each cluster has its own idle cooling device, each core has its own idle > >> injection thread, each idle injection thread uses play_idle to enter idle. In > >> order to reach the deepest idle state, each cooling device has the idle > >> injection threads synchronized together. > >> > >> It has some similarity with the intel power clamp driver but it is actually > >> designed to work on the ARM architecture via the DT with a mathematical proof > >> with the power model which comes with the Documentation. > >> > >> The idle injection cycle is fixed while the running cycle is variable. That > >> allows to have control on the device reactivity for the user experience. At > >> the mitigation point the idle threads are unparked, they play idle the > >> specified amount of time and they schedule themselves. The last thread sets > >> the next idle injection deadline and when the timer expires it wakes up all > >> the threads which in turn play idle again. Meanwhile the running cycle is > >> changed by set_cur_state. When the mitigation ends, the threads are parked. > >> The algorithm is self adaptive, so there is no need to handle hotplugging. > > > > The idle injection threads are RT threads (FIFO) and I saw in > > play_idle() set/clear flag PF_IDLE for it. Will these idle injection > > threads utilization be accounted into RT utilization? > > > > If idle injection threads utilization is accounted as RT tasks > > utilization, will this impact CPUFreq governor 'schedutil' for OPP > > selection? > > Hi Leo, > > The idle injection task has a very low utilization when it is not in the > play_idle function, basically it wakes up, sets a timer and play_idle(). > > Regarding the use case, the idle injection is the base brick for an > combo cooling device with cpufreq + cpuidle. When the idle injection is > used alone, it is because there is no cpufreq driver for the platform. > If there is a cpufreq driver, then we should endup with the cpu cooling > device where we have control of the OPP (and there is no idle injection > threads) or the combo cooling device. > > Except I'm missing something, the idle injection threads won't impact > the OPP selection. Mmm, they actually might. schedutil selects max OPP as soon as it sees an RT thread. I fear these guys might generate unwanted spikes. Maybe you can filter them out? Best, - Juri
On 27/03/2018 14:28, Juri Lelli wrote: > Hi Daniel, > > On 27/03/18 12:26, Daniel Lezcano wrote: >> On 27/03/2018 04:03, Leo Yan wrote: >>> Hi Daniel, >>> >>> On Wed, Feb 21, 2018 at 04:29:27PM +0100, Daniel Lezcano wrote: >>>> The cpu idle cooling driver performs synchronized idle injection across all >>>> cpus belonging to the same cluster and offers a new method to cool down a SoC. >>>> >>>> Each cluster has its own idle cooling device, each core has its own idle >>>> injection thread, each idle injection thread uses play_idle to enter idle. In >>>> order to reach the deepest idle state, each cooling device has the idle >>>> injection threads synchronized together. >>>> >>>> It has some similarity with the intel power clamp driver but it is actually >>>> designed to work on the ARM architecture via the DT with a mathematical proof >>>> with the power model which comes with the Documentation. >>>> >>>> The idle injection cycle is fixed while the running cycle is variable. That >>>> allows to have control on the device reactivity for the user experience. At >>>> the mitigation point the idle threads are unparked, they play idle the >>>> specified amount of time and they schedule themselves. The last thread sets >>>> the next idle injection deadline and when the timer expires it wakes up all >>>> the threads which in turn play idle again. Meanwhile the running cycle is >>>> changed by set_cur_state. When the mitigation ends, the threads are parked. >>>> The algorithm is self adaptive, so there is no need to handle hotplugging. >>> >>> The idle injection threads are RT threads (FIFO) and I saw in >>> play_idle() set/clear flag PF_IDLE for it. Will these idle injection >>> threads utilization be accounted into RT utilization? >>> >>> If idle injection threads utilization is accounted as RT tasks >>> utilization, will this impact CPUFreq governor 'schedutil' for OPP >>> selection? >> >> Hi Leo, >> >> The idle injection task has a very low utilization when it is not in the >> play_idle function, basically it wakes up, sets a timer and play_idle(). >> >> Regarding the use case, the idle injection is the base brick for an >> combo cooling device with cpufreq + cpuidle. When the idle injection is >> used alone, it is because there is no cpufreq driver for the platform. >> If there is a cpufreq driver, then we should endup with the cpu cooling >> device where we have control of the OPP (and there is no idle injection >> threads) or the combo cooling device. >> >> Except I'm missing something, the idle injection threads won't impact >> the OPP selection. > > Mmm, they actually might. schedutil selects max OPP as soon as it sees > an RT thread. I fear these guys might generate unwanted spikes. Maybe > you can filter them out? Yes, absolutely. Leo pointed it also. We might want to change the check at: https://elixir.bootlin.com/linux/v4.16-rc7/source/kernel/sched/cpufreq_schedutil.c#L364 in order to ignore PF_IDLE tagged tasks. -- <http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook | <http://twitter.com/#!/linaroorg> Twitter | <http://www.linaro.org/linaro-blog/> Blog
On 27/03/18 14:31, Daniel Lezcano wrote: > On 27/03/2018 14:28, Juri Lelli wrote: > > Hi Daniel, > > > > On 27/03/18 12:26, Daniel Lezcano wrote: > >> On 27/03/2018 04:03, Leo Yan wrote: > >>> Hi Daniel, > >>> > >>> On Wed, Feb 21, 2018 at 04:29:27PM +0100, Daniel Lezcano wrote: > >>>> The cpu idle cooling driver performs synchronized idle injection across all > >>>> cpus belonging to the same cluster and offers a new method to cool down a SoC. > >>>> > >>>> Each cluster has its own idle cooling device, each core has its own idle > >>>> injection thread, each idle injection thread uses play_idle to enter idle. In > >>>> order to reach the deepest idle state, each cooling device has the idle > >>>> injection threads synchronized together. > >>>> > >>>> It has some similarity with the intel power clamp driver but it is actually > >>>> designed to work on the ARM architecture via the DT with a mathematical proof > >>>> with the power model which comes with the Documentation. > >>>> > >>>> The idle injection cycle is fixed while the running cycle is variable. That > >>>> allows to have control on the device reactivity for the user experience. At > >>>> the mitigation point the idle threads are unparked, they play idle the > >>>> specified amount of time and they schedule themselves. The last thread sets > >>>> the next idle injection deadline and when the timer expires it wakes up all > >>>> the threads which in turn play idle again. Meanwhile the running cycle is > >>>> changed by set_cur_state. When the mitigation ends, the threads are parked. > >>>> The algorithm is self adaptive, so there is no need to handle hotplugging. > >>> > >>> The idle injection threads are RT threads (FIFO) and I saw in > >>> play_idle() set/clear flag PF_IDLE for it. Will these idle injection > >>> threads utilization be accounted into RT utilization? > >>> > >>> If idle injection threads utilization is accounted as RT tasks > >>> utilization, will this impact CPUFreq governor 'schedutil' for OPP > >>> selection? > >> > >> Hi Leo, > >> > >> The idle injection task has a very low utilization when it is not in the > >> play_idle function, basically it wakes up, sets a timer and play_idle(). > >> > >> Regarding the use case, the idle injection is the base brick for an > >> combo cooling device with cpufreq + cpuidle. When the idle injection is > >> used alone, it is because there is no cpufreq driver for the platform. > >> If there is a cpufreq driver, then we should endup with the cpu cooling > >> device where we have control of the OPP (and there is no idle injection > >> threads) or the combo cooling device. > >> > >> Except I'm missing something, the idle injection threads won't impact > >> the OPP selection. > > > > Mmm, they actually might. schedutil selects max OPP as soon as it sees > > an RT thread. I fear these guys might generate unwanted spikes. Maybe > > you can filter them out? > > Yes, absolutely. Leo pointed it also. > > We might want to change the check at: > > https://elixir.bootlin.com/linux/v4.16-rc7/source/kernel/sched/cpufreq_schedutil.c#L364 > > in order to ignore PF_IDLE tagged tasks. We might yes. And also for the update_single cases, I guess. Best, - Juri
diff --git a/drivers/thermal/Kconfig b/drivers/thermal/Kconfig index 5aaae1b..6c34117 100644 --- a/drivers/thermal/Kconfig +++ b/drivers/thermal/Kconfig @@ -166,6 +166,16 @@ config CPU_FREQ_THERMAL This will be useful for platforms using the generic thermal interface and not the ACPI interface. +config CPU_IDLE_THERMAL + bool "CPU idle cooling strategy" + depends on CPU_IDLE + help + This implements the generic CPU cooling mechanism through + idle injection. This will throttle the CPU by injecting + fixed idle cycle. All CPUs belonging to the same cluster + will enter idle synchronously to reach the deepest idle + state. + endchoice config CLOCK_THERMAL diff --git a/drivers/thermal/cpu_cooling.c b/drivers/thermal/cpu_cooling.c index 5c219dc..9340216 100644 --- a/drivers/thermal/cpu_cooling.c +++ b/drivers/thermal/cpu_cooling.c @@ -10,18 +10,32 @@ * Viresh Kumar <viresh.kumar@linaro.org> * */ +#undef DEBUG +#define pr_fmt(fmt) "CPU cooling: " fmt + #include <linux/module.h> #include <linux/thermal.h> #include <linux/cpufreq.h> +#include <linux/cpuidle.h> #include <linux/err.h> +#include <linux/freezer.h> #include <linux/idr.h> +#include <linux/kthread.h> #include <linux/pm_opp.h> #include <linux/slab.h> +#include <linux/sched/prio.h> +#include <linux/sched/rt.h> #include <linux/cpu.h> #include <linux/cpu_cooling.h> +#include <linux/wait.h> + +#include <linux/platform_device.h> +#include <linux/of_platform.h> #include <trace/events/thermal.h> +#include <uapi/linux/sched/types.h> + #ifdef CONFIG_CPU_FREQ_THERMAL /* * Cooling state <-> CPUFreq frequency @@ -928,3 +942,440 @@ void cpufreq_cooling_unregister(struct thermal_cooling_device *cdev) } EXPORT_SYMBOL_GPL(cpufreq_cooling_unregister); #endif /* CONFIG_CPU_FREQ_THERMAL */ + +#ifdef CONFIG_CPU_IDLE_THERMAL +/* + * The idle duration injection. As we don't have yet a way to specify + * from the DT configuration, let's default to a tick duration. + */ +#define DEFAULT_IDLE_TIME_US TICK_USEC + +/** + * struct cpuidle_cooling_device - data for the idle cooling device + * @cdev: a pointer to a struct thermal_cooling_device + * @cpumask: a cpumask containing the CPU managed by the cooling device + * @timer: a hrtimer giving the tempo for the idle injection cycles + * @kref: a kernel refcount on this structure + * @count: an atomic to keep track of the last task exiting the idle cycle + * @idle_cycle: an integer defining the duration of the idle injection + * @state: an normalized integer giving the state of the cooling device + */ +struct cpuidle_cooling_device { + struct thermal_cooling_device *cdev; + struct cpumask *cpumask; + struct list_head node; + struct hrtimer timer; + struct kref kref; + atomic_t count; + unsigned int idle_cycle; + unsigned int state; +}; + +/** + * @tsk: an array of pointer to the idle injection tasks + * @waitq: the waiq for the idle injection tasks + */ +struct cpuidle_cooling_tsk { + struct task_struct *tsk; + wait_queue_head_t waitq; +}; + +DEFINE_PER_CPU(struct cpuidle_cooling_tsk, cpuidle_cooling_tsk); + +static LIST_HEAD(cpuidle_cdev_list); + +/** + * cpuidle_cooling_wakeup - Wake up all idle injection threads + * @idle_cdev: the idle cooling device + * + * Every idle injection task belonging to the idle cooling device and + * running on an online cpu will be wake up by this call. + */ +static void cpuidle_cooling_wakeup(struct cpuidle_cooling_device *idle_cdev) +{ + int cpu; + struct cpuidle_cooling_tsk *cct; + + for_each_cpu_and(cpu, idle_cdev->cpumask, cpu_online_mask) { + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); + wake_up_process(cct->tsk); + } +} + +/** + * cpuidle_cooling_wakeup_fn - Running cycle timer callback + * @timer: a hrtimer structure + * + * When the mitigation is acting, the CPU is allowed to run an amount + * of time, then the idle injection happens for the specified delay + * and the idle task injection schedules itself until the timer event + * wakes the idle injection tasks again for a new idle injection + * cycle. The time between the end of the idle injection and the timer + * expiration is the allocated running time for the CPU. + * + * Returns always HRTIMER_NORESTART + */ +static enum hrtimer_restart cpuidle_cooling_wakeup_fn(struct hrtimer *timer) +{ + struct cpuidle_cooling_device *idle_cdev = + container_of(timer, struct cpuidle_cooling_device, timer); + + cpuidle_cooling_wakeup(idle_cdev); + + return HRTIMER_NORESTART; +} + +/** + * cpuidle_cooling_runtime - Running time computation + * @idle_cdev: the idle cooling device + * + * The running duration is computed from the idle injection duration + * which is fixed. If we reach 100% of idle injection ratio, that + * means the running duration is zero. If we have a 50% ratio + * injection, that means we have equal duration for idle and for + * running duration. + * + * The formula is deduced as the following: + * + * running = idle x ((100 / ratio) - 1) + * + * For precision purpose for integer math, we use the following: + * + * running = (idle x 100) / ratio - idle + * + * For example, if we have an injected duration of 50%, then we end up + * with 10ms of idle injection and 10ms of running duration. + * + * Returns a s64 nanosecond based + */ +static s64 cpuidle_cooling_runtime(struct cpuidle_cooling_device *idle_cdev) +{ + s64 next_wakeup; + int state = idle_cdev->state; + + /* + * The function must never be called when there is no + * mitigation because: + * - that does not make sense + * - we end up with a division by zero + */ + BUG_ON(!state); + + next_wakeup = (s64)((idle_cdev->idle_cycle * 100) / state) - + idle_cdev->idle_cycle; + + return next_wakeup * NSEC_PER_USEC; +} + +/** + * cpuidle_cooling_injection_thread - Idle injection mainloop thread function + * @arg: a void pointer containing the idle cooling device address + * + * This main function does basically two operations: + * + * - Goes idle for a specific amount of time + * + * - Sets a timer to wake up all the idle injection threads after a + * running period + * + * That happens only when the mitigation is enabled, otherwise the + * task is scheduled out. + * + * In order to keep the tasks synchronized together, it is the last + * task exiting the idle period which is in charge of setting the + * timer. + * + * This function never returns. + */ +static int cpuidle_cooling_injection_thread(void *arg) +{ + struct sched_param param = { .sched_priority = MAX_USER_RT_PRIO/2 }; + struct cpuidle_cooling_device *idle_cdev = arg; + struct cpuidle_cooling_tsk *cct = per_cpu_ptr(&cpuidle_cooling_tsk, + smp_processor_id()); + DEFINE_WAIT(wait); + + set_freezable(); + + sched_setscheduler(current, SCHED_FIFO, ¶m); + + while (1) { + s64 next_wakeup; + + prepare_to_wait(&cct->waitq, &wait, TASK_INTERRUPTIBLE); + + schedule(); + + atomic_inc(&idle_cdev->count); + + play_idle(idle_cdev->idle_cycle / USEC_PER_MSEC); + + /* + * The last CPU waking up is in charge of setting the + * timer. If the CPU is hotplugged, the timer will + * move to another CPU (which may not belong to the + * same cluster) but that is not a problem as the + * timer will be set again by another CPU belonging to + * the cluster, so this mechanism is self adaptive and + * does not require any hotplugging dance. + */ + if (!atomic_dec_and_test(&idle_cdev->count)) + continue; + + if (!idle_cdev->state) + continue; + + next_wakeup = cpuidle_cooling_runtime(idle_cdev); + + hrtimer_start(&idle_cdev->timer, ns_to_ktime(next_wakeup), + HRTIMER_MODE_REL_PINNED); + } + + finish_wait(&cct->waitq, &wait); + + return 0; +} + +/** + * cpuidle_cooling_release - Kref based release helper + * @kref: a pointer to the kref structure + * + * This function is automatically called by the kref_put function when + * the idle cooling device refcount reaches zero. At this point, we + * have the guarantee the structure is no longer in use and we can + * safely release all the ressources. + */ +static void __init cpuidle_cooling_release(struct kref *kref) +{ + struct cpuidle_cooling_device *idle_cdev = + container_of(kref, struct cpuidle_cooling_device, kref); + + thermal_cooling_device_unregister(idle_cdev->cdev); + kfree(idle_cdev); +} + +/** + * cpuidle_cooling_get_max_state - Get the maximum state + * @cdev : the thermal cooling device + * @state : a pointer to the state variable to be filled + * + * The function gives always 100 as the injection ratio is percentile + * based for consistency accros different platforms. + * + * The function can not fail, it returns always zero. + */ +static int cpuidle_cooling_get_max_state(struct thermal_cooling_device *cdev, + unsigned long *state) +{ + /* + * Depending on the configuration or the hardware, the running + * cycle and the idle cycle could be different. We want unify + * that to an 0..100 interval, so the set state interface will + * be the same whatever the platform is. + * + * The state 100% will make the cluster 100% ... idle. A 0% + * injection ratio means no idle injection at all and 50% + * means for 10ms of idle injection, we have 10ms of running + * time. + */ + *state = 100; + + return 0; +} + +/** + * cpuidle_cooling_get_cur_state - Get the current cooling state + * @cdev: the thermal cooling device + * @state: a pointer to the state + * + * The function just copy the state value from the private thermal + * cooling device structure, the mapping is 1 <-> 1. + * + * The function can not fail, it returns always zero. + */ +static int cpuidle_cooling_get_cur_state(struct thermal_cooling_device *cdev, + unsigned long *state) +{ + struct cpuidle_cooling_device *idle_cdev = cdev->devdata; + + *state = idle_cdev->state; + + return 0; +} + +/** + * cpuidle_cooling_set_cur_state - Set the current cooling state + * @cdev: the thermal cooling device + * @state: the target state + * + * The function checks first if we are initiating the mitigation which + * in turn wakes up all the idle injection tasks belonging to the idle + * cooling device. In any case, it updates the internal state for the + * cooling device. + * + * The function can not fail, it returns always zero. + */ +static int cpuidle_cooling_set_cur_state(struct thermal_cooling_device *cdev, + unsigned long state) +{ + struct cpuidle_cooling_device *idle_cdev = cdev->devdata; + unsigned long current_state = idle_cdev->state; + + idle_cdev->state = state; + + if (current_state == 0 && state > 0) { + pr_debug("Starting cooling cpus '%*pbl'\n", + cpumask_pr_args(idle_cdev->cpumask)); + cpuidle_cooling_wakeup(idle_cdev); + } else if (current_state > 0 && !state) { + pr_debug("Stopping cooling cpus '%*pbl'\n", + cpumask_pr_args(idle_cdev->cpumask)); + } + + return 0; +} + +/** + * cpuidle_cooling_ops - thermal cooling device ops + */ +static struct thermal_cooling_device_ops cpuidle_cooling_ops = { + .get_max_state = cpuidle_cooling_get_max_state, + .get_cur_state = cpuidle_cooling_get_cur_state, + .set_cur_state = cpuidle_cooling_set_cur_state, +}; + +/** + * cpuilde_cooling_unregister - Idle cooling device exit function + * + * This function unregisters the cpuidle cooling device and frees the + * ressources previously allocated by the init function. This function + * is called when the initialization fails. + */ +static void cpuidle_cooling_unregister(void) +{ + struct cpuidle_cooling_device *tmp, *idle_cdev = NULL; + struct cpuidle_cooling_tsk *cct; + int cpu; + + list_for_each_entry_safe(idle_cdev, tmp, &cpuidle_cdev_list, node) { + for_each_cpu(cpu, idle_cdev->cpumask) { + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); + if (cct->tsk) + kthread_stop(cct->tsk); + kref_put(&idle_cdev->kref, cpuidle_cooling_release); + } + } +} + +/** + * cpuidle_cooling_register - Idle cooling device initialization function + * + * This function is in charge of creating a cooling device per cluster + * and register it to thermal framework. For this we rely on the + * topology as there is nothing yet describing better the idle state + * power domains. + * + * For each first CPU of the cluster's cpumask, we allocate the idle + * cooling device, initialize the general fields and then we initialze + * the rest in a per cpu basis. + * + * Returns zero on success, < 0 otherwise. + */ +int cpuidle_cooling_register(void) +{ + struct cpuidle_cooling_device *idle_cdev = NULL; + struct thermal_cooling_device *cdev; + struct cpuidle_cooling_tsk *cct; + struct task_struct *tsk; + struct device_node *np; + cpumask_t *cpumask; + char dev_name[THERMAL_NAME_LENGTH]; + int ret = -ENOMEM, cpu; + int index = 0; + + for_each_possible_cpu(cpu) { + cpumask = topology_core_cpumask(cpu); + + cct = per_cpu_ptr(&cpuidle_cooling_tsk, cpu); + + /* + * This condition makes the first cpu belonging to the + * cluster to create a cooling device and allocates + * the structure. Others CPUs belonging to the same + * cluster will just increment the refcount on the + * cooling device structure and initialize it. + */ + if (cpu == cpumask_first(cpumask)) { + np = of_cpu_device_node_get(cpu); + + idle_cdev = kzalloc(sizeof(*idle_cdev), GFP_KERNEL); + if (!idle_cdev) + goto out_fail; + + idle_cdev->idle_cycle = DEFAULT_IDLE_TIME_US; + + atomic_set(&idle_cdev->count, 0); + + kref_init(&idle_cdev->kref); + + /* + * Initialize the timer to wakeup all the idle + * injection tasks + */ + hrtimer_init(&idle_cdev->timer, + CLOCK_MONOTONIC, HRTIMER_MODE_REL); + + /* + * The wakeup function callback which is in + * charge of waking up all CPUs belonging to + * the same cluster + */ + idle_cdev->timer.function = cpuidle_cooling_wakeup_fn; + + /* + * The thermal cooling device name + */ + snprintf(dev_name, sizeof(dev_name), "thermal-idle-%d", index++); + cdev = thermal_of_cooling_device_register(np, dev_name, + idle_cdev, + &cpuidle_cooling_ops); + if (IS_ERR(cdev)) { + ret = PTR_ERR(cdev); + goto out_fail; + } + + idle_cdev->cdev = cdev; + + idle_cdev->cpumask = cpumask; + + list_add(&idle_cdev->node, &cpuidle_cdev_list); + + pr_info("Created idle cooling device for cpus '%*pbl'\n", + cpumask_pr_args(cpumask)); + } + + kref_get(&idle_cdev->kref); + + init_waitqueue_head(&cct->waitq); + + tsk = kthread_create_on_cpu(cpuidle_cooling_injection_thread, + idle_cdev, cpu, "kidle_inject/%u"); + if (IS_ERR(tsk)) { + ret = PTR_ERR(tsk); + goto out_fail; + } + + cct->tsk = tsk; + + wake_up_process(tsk); + } + + return 0; + +out_fail: + cpuidle_cooling_unregister(); + pr_err("Failed to create idle cooling device (%d)\n", ret); + + return ret; +} +#endif /* CONFIG_CPU_IDLE_THERMAL */ diff --git a/include/linux/cpu_cooling.h b/include/linux/cpu_cooling.h index c0accc7..fee2038 100644 --- a/include/linux/cpu_cooling.h +++ b/include/linux/cpu_cooling.h @@ -120,4 +120,13 @@ void cpufreq_cooling_unregister(struct thermal_cooling_device *cdev) } #endif /* CONFIG_CPU_FREQ_THERMAL */ +#ifdef CONFIG_CPU_IDLE_THERMAL +extern int cpuidle_cooling_register(void); +#else /* CONFIG_CPU_IDLE_THERMAL */ +static inline int cpuidle_cooling_register(void) +{ + return 0; +} +#endif /* CONFIG_CPU_IDLE_THERMAL */ + #endif /* __CPU_COOLING_H__ */