diff mbox series

[1/1] Revert "scsi: ufs: core: Only suspend clock scaling if scaling down"

Message ID 20240228053421.19700-1-quic_rampraka@quicinc.com
State New
Headers show
Series [1/1] Revert "scsi: ufs: core: Only suspend clock scaling if scaling down" | expand

Commit Message

Ram Prakash Gupta Feb. 28, 2024, 5:34 a.m. UTC
This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
Approx 28% random perf IO degradation is observed by suspending clk
scaling only when clks are scaled down. Concern for original fix was
power consumption, which is already taken care by clk gating by putting
the link into hibern8 state.

Signed-off-by: Ram Prakash Gupta <quic_rampraka@quicinc.com>
---
 drivers/ufs/core/ufshcd.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

Comments

Bart Van Assche Feb. 28, 2024, 6:52 p.m. UTC | #1
On 2/27/24 21:34, Ram Prakash Gupta wrote:
> This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
> Approx 28% random perf IO degradation is observed by suspending clk
> scaling only when clks are scaled down. Concern for original fix was
> power consumption, which is already taken care by clk gating by putting
> the link into hibern8 state.
> 
> Signed-off-by: Ram Prakash Gupta <quic_rampraka@quicinc.com>
> ---
>   drivers/ufs/core/ufshcd.c | 2 +-
>   1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
> index c416826762e9..f6be18db031c 100644
> --- a/drivers/ufs/core/ufshcd.c
> +++ b/drivers/ufs/core/ufshcd.c
> @@ -1586,7 +1586,7 @@ static int ufshcd_devfreq_target(struct device *dev,
>   		ktime_to_us(ktime_sub(ktime_get(), start)), ret);
>   
>   out:
> -	if (sched_clk_scaling_suspend_work && !scale_up)
> +	if (sched_clk_scaling_suspend_work)
>   		queue_work(hba->clk_scaling.workq,
>   			   &hba->clk_scaling.suspend_work);

What is the base kernel version for your tests? Was this patch series
included in your kernel tree: "[PATCH V6 0/2] Add CPU latency QoS 
support for ufs driver" 
(https://lore.kernel.org/all/20231219123706.6463-1-quic_mnaresh@quicinc.com/)?

Thanks,

Bart.
Peter Wang (王信友) Feb. 29, 2024, 7:51 a.m. UTC | #2
On Wed, 2024-02-28 at 11:04 +0530, Ram Prakash Gupta wrote:
>  	 
> External email : Please do not click links or open attachments until
> you have verified the sender or the content.
>  This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
> Approx 28% random perf IO degradation is observed by suspending clk
> scaling only when clks are scaled down. Concern for original fix was
> power consumption, which is already taken care by clk gating by
> putting
> the link into hibern8 state.
> 
> Signed-off-by: Ram Prakash Gupta <quic_rampraka@quicinc.com>
> ---
>  drivers/ufs/core/ufshcd.c | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
> index c416826762e9..f6be18db031c 100644
> --- a/drivers/ufs/core/ufshcd.c
> +++ b/drivers/ufs/core/ufshcd.c
> @@ -1586,7 +1586,7 @@ static int ufshcd_devfreq_target(struct device
> *dev,
>  		ktime_to_us(ktime_sub(ktime_get(), start)), ret);
>  
>  out:
> -	if (sched_clk_scaling_suspend_work && !scale_up)
> +	if (sched_clk_scaling_suspend_work)
>  		queue_work(hba->clk_scaling.workq,
>  			   &hba->clk_scaling.suspend_work);
>  
> -- 
> 2.17.1

Hi Ram,

It is logic wrong to keep high gear when no read/write traffic.
Even high gear turn off clock and enter hibernate, there still have
other power consume hardhware to keep IO in high gear, ex. CPU latency,
CPU power.

Besides, clock scaling is designed for power concern, not for
performance. If you want to keep high performance, you can just turn
off clock scaling and keep in highest gear.

Finally, mediatek dosen't suffer performance drop with this patch.
Could you help list the test procedure and performance drop data more
detail? I am curious that in what scenario your random drop 28%.
And I think your dvfs parameter could be the drop reason. 

Thanks
Peter
Thorsten Leemhuis March 3, 2024, 11:53 a.m. UTC | #3
[TLDR: I'm adding this report to the list of tracked Linux kernel
regressions; the text you find below is based on a few templates
paragraphs you might have encountered already in similar form.
See link in footer if these mails annoy you.]

On 28.02.24 06:34, Ram Prakash Gupta wrote:
> This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
> Approx 28% random perf IO degradation is observed by suspending clk
> scaling only when clks are scaled down. Concern for original fix was
> power consumption, which is already taken care by clk gating by putting
> the link into hibern8 state.

Thanks for the report. To be sure the issue doesn't fall through the
cracks unnoticed, I'm adding it to regzbot, the Linux kernel regression
tracking bot:

#regzbot ^introduced 1d969731b87f122108c50a64acfdbaa63486296e
#regzbot title scsi: ufs: core: approx 28% random perf IO degradation
#regzbot ignore-activity

Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat)
--
Everything you wanna know about Linux kernel regression tracking:
https://linux-regtracking.leemhuis.info/about/#tldr
That page also explains what to do if mails like this annoy you.
Ram Prakash Gupta March 5, 2024, 6:23 a.m. UTC | #4
On 2/29/2024 12:22 AM, Bart Van Assche wrote:
> On 2/27/24 21:34, Ram Prakash Gupta wrote:
>> This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
>> Approx 28% random perf IO degradation is observed by suspending clk
>> scaling only when clks are scaled down. Concern for original fix was
>> power consumption, which is already taken care by clk gating by putting
>> the link into hibern8 state.
>>
>> Signed-off-by: Ram Prakash Gupta <quic_rampraka@quicinc.com>
>> ---
>>   drivers/ufs/core/ufshcd.c | 2 +-
>>   1 file changed, 1 insertion(+), 1 deletion(-)
>>
>> diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
>> index c416826762e9..f6be18db031c 100644
>> --- a/drivers/ufs/core/ufshcd.c
>> +++ b/drivers/ufs/core/ufshcd.c
>> @@ -1586,7 +1586,7 @@ static int ufshcd_devfreq_target(struct device 
>> *dev,
>>           ktime_to_us(ktime_sub(ktime_get(), start)), ret);
>>   out:
>> -    if (sched_clk_scaling_suspend_work && !scale_up)
>> +    if (sched_clk_scaling_suspend_work)
>>           queue_work(hba->clk_scaling.workq,
>>                  &hba->clk_scaling.suspend_work);
> 
> What is the base kernel version for your tests? Was this patch series
> included in your kernel tree: "[PATCH V6 0/2] Add CPU latency QoS 
> support for ufs driver" 
> (https://lore.kernel.org/all/20231219123706.6463-1-quic_mnaresh@quicinc.com/)?
> 
> Thanks,
> 
> Bart.

Hi Bart,

kernel version used is 6.1 and QoS support for CPU latency is present.

Thanks,
Ram
Peter Wang (王信友) March 5, 2024, 12:55 p.m. UTC | #5
On Tue, 2024-03-05 at 12:59 +0530, Ram Prakash Gupta wrote:
>  	 
> External email : Please do not click links or open attachments until
> you have verified the sender or the content.
>  
> 
> On 2/29/2024 1:21 PM, Peter Wang (王信友) wrote:
> > On Wed, 2024-02-28 at 11:04 +0530, Ram Prakash Gupta wrote:
> >>   
> >> External email : Please do not click links or open attachments
> until
> >> you have verified the sender or the content.
> >>   This reverts commit 1d969731b87f122108c50a64acfdbaa63486296e.
> >> Approx 28% random perf IO degradation is observed by suspending
> clk
> >> scaling only when clks are scaled down. Concern for original fix
> was
> >> power consumption, which is already taken care by clk gating by
> >> putting
> >> the link into hibern8 state.
> >>
> >> Signed-off-by: Ram Prakash Gupta <quic_rampraka@quicinc.com>
> >> ---
> >>   drivers/ufs/core/ufshcd.c | 2 +-
> >>   1 file changed, 1 insertion(+), 1 deletion(-)
> >>
> >> diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
> >> index c416826762e9..f6be18db031c 100644
> >> --- a/drivers/ufs/core/ufshcd.c
> >> +++ b/drivers/ufs/core/ufshcd.c
> >> @@ -1586,7 +1586,7 @@ static int ufshcd_devfreq_target(struct
> device
> >> *dev,
> >>   ktime_to_us(ktime_sub(ktime_get(), start)), ret);
> >>   
> >>   out:
> >> -if (sched_clk_scaling_suspend_work && !scale_up)
> >> +if (sched_clk_scaling_suspend_work)
> >>   queue_work(hba->clk_scaling.workq,
> >>      &hba->clk_scaling.suspend_work);
> >>   
> >> -- 
> >> 2.17.1
> > 
> > Hi Ram,
> > 
> > It is logic wrong to keep high gear when no read/write traffic.
> > Even high gear turn off clock and enter hibernate, there still have
> > other power consume hardhware to keep IO in high gear, ex. CPU
> latency,
> > CPU power.
> > 
> By CPU latency and power, do you mean the QoS/bus vote from ufs
> driver?
> if yes, in that case if clk gating kicks in, it means ufs is already
> in 
> hibernate state and there is no point keeping the votes (pm qos &
> bus 
> votes) from ufs driver. And as part of clk gating, qos and other
> votes 
> on SoC can be removed then no power concern would be there.
> 

Not only pm qos vote, but also vcore raise is needed for mediatek hw
design. And this core voltage is used for our entire SoC. Which means
The power impact is very huge for mediatek.


> Now with your implementation, since scaling is not suspended, in
> next 
> window of devfreq polling will get lowest possible load when active 
> request count is zero, because of which devfreq will scale down the 
> clock. So next request would always be completed with low clock 
> frequency, which is not desirable from performance point of view 

If a polling period without any IO traffic, it should scale down right?
If IO is busy then scale up. 
If IO is not busy then scale down.
It is basic logic.


> when 
> power consumption is already taken care in clk gating.
> 
> > Besides, clock scaling is designed for power concern, not for
> > performance. If you want to keep high performance, you can just
> turn
> > off clock scaling and keep in highest gear.
> > 
> I think its about striking a right balance. And if there is really a
> big
> power concern on mediatek boards, clock gating can be made bit more 
> aggressive there by removing the all votes (qos, bus) from ufs driver
> on 
> SoC as part of clock gating. Also is clock scaling disabled on
> mediatek 
> platforms?
> 

Mediatek has use clock gating.
But, there still have a window after enter auto-hibernate and clock
gati
ng. This window power waste is not reasonalbe.


> > Finally, mediatek dosen't suffer performance drop with this patch.
> > Could you help list the test procedure and performance drop data
> more
> > detail? I am curious that in what scenario your random drop 28%.
> > And I think your dvfs parameter could be the drop reason.
> > 
> There is no specific environment or procedure used, I am just using 
> antutu benchmark to get the numbers. And random IO numbers are
> degraded 
> by approx 28%.
> 

Have you try another dvfs setting? 
For example, enlarge polling period to make scale down harder?

Anyway, I think it is not correct to benefit from performance gain
through a kernel bug (scale up when no IO on-going)


Thanks.
Peter
Peter Wang (王信友) April 25, 2024, 1:15 p.m. UTC | #6
On Wed, 2024-04-24 at 14:44 +0530, Ram Prakash Gupta wrote:
> 
> Hi Peter,
> 
> I tried different dvfs settings, none is helping including enlarged 
> polling period time, its degrading perf numbers as its taking longer 
> time to scale up when the load is high and clk is low.
> 
> I checked from power side on qualcomm boards, suspending with zero 
> request is not impacting power hence I am consider a vops to add
> which 
> can help your use case too, I tested this vops and it works fine on 
> qualcomm boards.
> 
> here is a small snippet of a different approach using vops, which I
> am 
> planning to push under a separate mail subject to remove this
> deadlock 
> between mediatek and qualcomm, scaling config.
> 
> -       if (sched_clk_scaling_suspend_work && !scale_up)
> +       if (sched_clk_scaling_suspend_work && 
> hba->clk_scaling.no_req_suspend)
> +               queue_work(hba->clk_scaling.workq,
> +                          &hba->clk_scaling.suspend_work);
> 

Hi Ram,

It is weird for me that if no_req_suspend is true, queue suspend work?
Dosen't "no_req_suspend" simply mean "do not suspend"? 

Thanks
Peter



> +       else if (sched_clk_scaling_suspend_work && !scale_up)
> 
> Here no_req_suspend would be false by default, so would hit else if 
> case, which is desirable for mediatek boards. For qualcomm, 
> no_req_suspend would set it to true via vops. please let me know if
> this 
> is ok for you.
> 
> Thanks,
> Ram
diff mbox series

Patch

diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
index c416826762e9..f6be18db031c 100644
--- a/drivers/ufs/core/ufshcd.c
+++ b/drivers/ufs/core/ufshcd.c
@@ -1586,7 +1586,7 @@  static int ufshcd_devfreq_target(struct device *dev,
 		ktime_to_us(ktime_sub(ktime_get(), start)), ret);
 
 out:
-	if (sched_clk_scaling_suspend_work && !scale_up)
+	if (sched_clk_scaling_suspend_work)
 		queue_work(hba->clk_scaling.workq,
 			   &hba->clk_scaling.suspend_work);