Message ID | 1486650171-20598-4-git-send-email-m.szyprowski@samsung.com |
---|---|
State | New |
Headers | show |
Series | DMA Engine: switch PL330 driver to non-irq-safe runtime PM | expand |
On Thu, Feb 09, 2017 at 03:22:51PM +0100, Marek Szyprowski wrote: > +static int pl330_set_slave(struct dma_chan *chan, struct device *slave) > +{ > + struct dma_pl330_chan *pch = to_pchan(chan); > + struct pl330_dmac *pl330 = pch->dmac; > + int i; > + > + mutex_lock(&pl330->rpm_lock); > + > + for (i = 0; i < pl330->num_peripherals; i++) { > + if (pl330->peripherals[i].chan.slave == slave && > + pl330->peripherals[i].slave_link) { > + pch->slave_link = pl330->peripherals[i].slave_link; > + goto done; > + } > + } > + > + pch->slave_link = device_link_add(slave, pl330->ddma.dev, > + DL_FLAG_PM_RUNTIME | DL_FLAG_RPM_ACTIVE); So you are going to add the link on channel allocation and tear down on the freeup. I am not sure I really like the idea here. First, these thing shouldn't be handled in the drivers. These things should be set in core and each driver setting the links doesn't sound great to me. Second, should the link be always there and we only mange the state? Here it seems that we have link being created and destroyed, so why not mark it ACTIVE and DORMANT instead... Lastly, looking at th description of the issue here, am perceiving (maybe my understanding is not quite right here) that you have an IP block in SoC which has multiple things and share common stuff and doing right PM is a challenge for you, right? -- ~Vinod
On 10 February 2017 at 12:51, Marek Szyprowski <m.szyprowski@samsung.com> wrote: > Hi Vinod, > > On 2017-02-10 05:50, Vinod Koul wrote: >> >> On Thu, Feb 09, 2017 at 03:22:51PM +0100, Marek Szyprowski wrote: >> >>> +static int pl330_set_slave(struct dma_chan *chan, struct device *slave) >>> +{ >>> + struct dma_pl330_chan *pch = to_pchan(chan); >>> + struct pl330_dmac *pl330 = pch->dmac; >>> + int i; >>> + >>> + mutex_lock(&pl330->rpm_lock); >>> + >>> + for (i = 0; i < pl330->num_peripherals; i++) { >>> + if (pl330->peripherals[i].chan.slave == slave && >>> + pl330->peripherals[i].slave_link) { >>> + pch->slave_link = >>> pl330->peripherals[i].slave_link; >>> + goto done; >>> + } >>> + } >>> + >>> + pch->slave_link = device_link_add(slave, pl330->ddma.dev, >>> + DL_FLAG_PM_RUNTIME | >>> DL_FLAG_RPM_ACTIVE); >> >> So you are going to add the link on channel allocation and tear down on >> the >> freeup. > > > Right. Channel allocation is typically done once per driver operation and it > won't hurt system performance. > >> I am not sure I really like the idea here. > > > Could you point what's wrong with it? > >> First, these thing shouldn't be handled in the drivers. These things >> should >> be set in core and each driver setting the links doesn't sound great to >> me. > > > Which core? And what's wrong with the device links? They have been > introduced to > model relations between devices that are behind the usual parent/child/bus > topology. I think Vinod mean the dmaengine core. Which also would make perfect sense to me as it would benefit all dma drivers. The only related PM thing, that shall be the decision of the driver, is whether it wants to enable runtime PM or not, during ->probe(). > >> Second, should the link be always there and we only mange the state? Here >> it >> seems that we have link being created and destroyed, so why not mark it >> ACTIVE and DORMANT instead... > > > Link state is managed by device core and should not be touched by the > drivers. > It is related to both provider and consumer drivers states (probed/not > probed/etc). > > Second we would need to create those links first. The question is where to > create them then. Just to fill in, to me this is really also the key question. If we could set up the device link already at device initialization, it should also be possible to avoid getting -EPROBE_DEFER for dma client drivers when requesting their dma channels. > >> Lastly, looking at th description of the issue here, am perceiving (maybe >> my >> understanding is not quite right here) that you have an IP block in SoC >> which has multiple things and share common stuff and doing right PM is a >> challenge for you, right? > > > Nope. Doing right PM in my SoC is not that complex and I would say it is > rather > typical for any embedded stuff. It works fine (in terms of the power > consumption reduction) when all drivers simply properly manage their runtime > PM state, thus if device is not in use, the state is set to suspended and > finally, the power domain gets turned off. > > I've used device links for PM only because the current DMA engine API is > simply insufficient to implement it in the other way. > > I want to let a power domain, which contains a few devices, among those a > PL330 > device, to get turned off when there is no activity. Handling power domain > power > on / off requires non-atomic context, what is typical for runtime pm calls. > For > that I need to have non-irq-safe runtime pm implemented for all devices that > belongs to that domains. Again, allow me to fill in. This issue exists for all ARM SoC which has a dma controller residing in a PM domain. I think that is quite many. Currently the only solution I have seen for this problem, but which I really dislike. That is, each dma client driver requests/releases their dma channel from their respective ->runtime_suspend|resume() callbacks - then the dma driver can use the dma request/release hooks, to do pm_runtime_get|put() which then becomes non-irq-safe. > > The problem with PL330 driver is that it use irq-safe runtime pm, which like > it > was stated in the patch description doesn't bring much benefits. To switch > to > standard (non-irq-safe) runtime pm, the pm_runtime calls have to be done > from > a context which permits sleeping. The problem with DMA engine driver API is > that > most of its callbacks have to be IRQ-safe and frankly only > device_{alloc,release}_chan_resources() what more or less maps to > dma_request_chan()/dma_release_channel() and friends. There are DMA engine > drivers which do runtime PM calls there (tegra20-apb-dma, sirf-dma, cppi41, > rcar-dmac), but this is not really efficient. DMA engine clients usually > allocate > dma channel during their probe() and keep them for the whole driver life. In > turn > this very similar to calling pm_runtime_get() in the DMA engine driver > probe(). > The result of both approaches is that DMA engine device keeps its power > domain > enabled almost all the time. This problem is also mentioned in the DMA > engine > TODO list, you have pointed me yesterday. > > To avoid such situation that DMA engine driver blocks turning off the power > domain and avoid changing DMA engine client API I came up with the device > links > pm based approach. I don't want to duplicate the description here, the > details > were in the patch description, however if you have any particular question > about > the details, let me know and I will try to clarify it more. So besides solving the irq-safe issue for dma driver, using the device-links has additionally two advantages. I already mentioned the -EPROBE_DEFER issue above. The second thing, is the runtime/system PM relations we get for free by using the links. In other words, the dma driver/core don't need to care about dealing with pm_runtime_get|put() as that would be managed by the dma client driver. Kind regards Uffe
On Mon, Feb 13, 2017 at 12:11:54PM +0100, Ulf Hansson wrote: > >> > >> If we could set up the device link already at device initialization, > >> it should also be possible to avoid getting -EPROBE_DEFER for dma > >> client drivers when requesting their dma channels. > > > > Well if we defer then driver will regiser with dmaengine after it is > > probed, so a client will either get a channel or not. IOW we won't get > > -EPROBE_DEFER. > > I didn't quite get this. What do you mean by "if we defer..."? > > Defer into *what* and defer of *what*? Could you please elaborate. Nevermind I think below is much interesting now.. > >> Again, allow me to fill in. This issue exists for all ARM SoC which > >> has a dma controller residing in a PM domain. I think that is quite > >> many. > >> > >> Currently the only solution I have seen for this problem, but which I > >> really dislike. That is, each dma client driver requests/releases > >> their dma channel from their respective ->runtime_suspend|resume() > >> callbacks - then the dma driver can use the dma request/release hooks, > >> to do pm_runtime_get|put() which then becomes non-irq-safe. > > > > Yeah that is not the best way to do. But looking at it current one doesnt > > seem best fit either. > > > > So on seeing the device_link_add() I was thinking that this is some SoC > > dependent problem being solved whereas the problem statmement is non-atomic > > channel prepare. > > You may be right. > > Although, I don't know of other examples, besides the runtime PM use > case, where non-atomic channel prepare/unprepare would make sense. Do > you? The primary ask for that has been to enable runtime_pm for drivers. It's not a new ask, but we somehow haven't gotten around to do it. > > As I said earlier, if we want to solve that problem a better idea is to > > actually split the prepare as we discussed in [1] > > > > This way we can get a non atomic descriptor allocate/prepare and release. > > Yes we need to redesign the APIs to solve this, but if you guys are up for > > it, I think we can do it and avoid any further round abouts :) > > Adding/re-designing dma APIs is a viable option to solve the runtime PM case. > > Changes would be needed for all related dma client drivers as well, > although if that's what we need to do - let's do it. Yes, but do bear in mind that some cases do need atomic prepare. The primary cases for DMA had that in mind and also submitting next transaction from the callback (tasklet) context, so that won't go away. It would help in other cases where clients know that they will not be in atomic context so we provide additional non-atomic "allocation" followed by prepare, so that drivers can split the work among these and people can do runtime_pm and other things.. > >> So besides solving the irq-safe issue for dma driver, using the > >> device-links has additionally two advantages. I already mentioned the > >> -EPROBE_DEFER issue above. > >> > >> The second thing, is the runtime/system PM relations we get for free by > >> using the links. In other words, the dma driver/core don't need to care > >> about dealing with pm_runtime_get|put() as that would be managed by the > >> dma client driver. > > > > Yeah sorry took me a while to figure that out :), If we do a different > > API then dmaengine core can call pm_runtime_get|put() from non-atomic > > context. > > Yes, it can and this works from runtime PM point of view. But the > following issues would remain unsolved. > > 1) Dependencies between dma drivers and dma client drivers during system > PM. For example, a dma client driver needs the dma controller to be > operational (remain system resumed), until the dma client driver itself > becomes system suspended. > > The *only* currently available solution for this, is to try to system > suspend the dma controller later than the dma client, via using the *late > or the *noirq system PM callbacks. This works for most cases, but it > becomes a problem when the dma client also needs to be system suspended at > the *late or the *noirq phase. Clearly this solution that doesn't scale. > > Using device links explicitly solves this problem as it allows to specify > this dependency between devices. Yes this is an interesting point. Yes till now people have been doing above to workaround this problem, but hey this is not a unique to dmaengine. Any subsystem which provides services to others has this issue, so the solution much be driver or pm framework and not unique to dmaengine. > 2) We won't avoid dma clients from getting -EPROBE_DEFER when requesting > their dma channels in their ->probe() routines. This would be possible, if > we can set up the device links at device initialization. Well setting those links is not practical at initialization time. Most modern dma controllers feature a SW mux, with multiple clients connecting and requesting, would we link all of them? Most of times dmaengine driver wont know about those.. Thanks -- ~Vinod
[...] >> The only related PM thing, that shall be the decision of the driver, >> is whether it wants to enable runtime PM or not, during ->probe(). > > > So do you want to create the links during the DMAengine driver probe? How do > you > plan to find all the client devices? Please note that you really want to > create > links to devices which will really use the DMA engine calls. Some client > drivers might decide in runtime weather to use DMA engine or not, depending > on > other data. I don't have great plan, just wanted to share my thoughts around the problems we want to solve. [...] >> >> If we could set up the device link already at device initialization, >> it should also be possible to avoid getting -EPROBE_DEFER for dma >> client drivers when requesting their dma channels. > > > At the first glance this sounds like an ultimate solution for all problems, > but I don't think that device links can be used this way. If I get it right, > you would like to create links on client device initialization, preferably > somewhere in the kernel driver core. This will be handled somehow by a > completely generic code, which will create a link each pair of devices, > which are connected by a phandle. Is this what you meant? Please note that > that time no driver for both client and provider are probed. IMHO that > doesn't look like a right generic approach > > How that code will know get following information: > 1. is it really needed to create a link for given device pair? > 2. what link flags should it use? > 3. what about circular dependencies? > 4. what about runtime optional dependencies? > 5. what about non-dt platforms? acpi? > To give a good answer of these questions, I need to spend more time investigating. However, from a top-level point of view, I think the device links seems like the perfect match for solving the runtime/system PM problems. No matter whether we can set up the links at device initialization time, driver probe or whatever time. > This looks like another newer ending story of "how can we avoid deferred > probe > in a generic way". IMHO we should first solve the problem of irq-safe > runtime > PM in DMA engine drivers first. I proposed how it can be done with device > links. > With no changes in the client API. Later if one decide to extend the client > API > in a way it will allow other runtime PM implementation - I see no problem to > convert pl330 driver to the new approach, but for the time being - this > would > be the easiest way to get it really functional. Agree, let's drop the deferred probe topic from the discussions - it's just going to be overwhelming. :-) [...] >> So besides solving the irq-safe issue for dma driver, using the >> device-links has additionally two advantages. I already mentioned the >> -EPROBE_DEFER issue above. > > > Not really. IMHO device links can be properly established once both drivers > are probed... Okay. > >> >> The second thing, is the runtime/system PM relations we get for free >> by using the links. In other words, the dma driver/core don't need to >> care about dealing with pm_runtime_get|put() as that would be managed >> by the dma client driver. > > > IMHO there might be drivers which don't want to use device links based > runtime > PM in favor of irq-safe PM or something else. This should be really left to > drivers. Okay. Kind regards Uffe
[...] >> Although, I don't know of other examples, besides the runtime PM use >> case, where non-atomic channel prepare/unprepare would make sense. Do >> you? > > The primary ask for that has been to enable runtime_pm for drivers. It's not > a new ask, but we somehow haven't gotten around to do it. Okay, I see. > >> > As I said earlier, if we want to solve that problem a better idea is to >> > actually split the prepare as we discussed in [1] >> > >> > This way we can get a non atomic descriptor allocate/prepare and release. >> > Yes we need to redesign the APIs to solve this, but if you guys are up for >> > it, I think we can do it and avoid any further round abouts :) >> >> Adding/re-designing dma APIs is a viable option to solve the runtime PM case. >> >> Changes would be needed for all related dma client drivers as well, >> although if that's what we need to do - let's do it. > > Yes, but do bear in mind that some cases do need atomic prepare. The primary > cases for DMA had that in mind and also submitting next transaction from the > callback (tasklet) context, so that won't go away. > > It would help in other cases where clients know that they will not be in > atomic context so we provide additional non-atomic "allocation" followed by > prepare, so that drivers can split the work among these and people can do > runtime_pm and other things.. That for sharing the details. It seems like some dma expert really need to be heavily involved if we ever are going to complete this work. :-) [...] >> >> 1) Dependencies between dma drivers and dma client drivers during system >> PM. For example, a dma client driver needs the dma controller to be >> operational (remain system resumed), until the dma client driver itself >> becomes system suspended. >> >> The *only* currently available solution for this, is to try to system >> suspend the dma controller later than the dma client, via using the *late >> or the *noirq system PM callbacks. This works for most cases, but it >> becomes a problem when the dma client also needs to be system suspended at >> the *late or the *noirq phase. Clearly this solution that doesn't scale. >> >> Using device links explicitly solves this problem as it allows to specify >> this dependency between devices. > > Yes this is an interesting point. Yes till now people have been doing above > to workaround this problem, but hey this is not a unique to dmaengine. Any > subsystem which provides services to others has this issue, so the solution > much be driver or pm framework and not unique to dmaengine. I definitely agree, these problems aren't unique to the dmaengine subsystem. Exactly how/where to manage them, that I guess, is the key question. However, I can't resist from finding the device links useful, as those really do address and solve our issues from a runtime/system PM point of view. > >> 2) We won't avoid dma clients from getting -EPROBE_DEFER when requesting >> their dma channels in their ->probe() routines. This would be possible, if >> we can set up the device links at device initialization. > > Well setting those links is not practical at initialization time. Most > modern dma controllers feature a SW mux, with multiple clients connecting > and requesting, would we link all of them? Most of times dmaengine driver > wont know about those.. Okay, I see! Kind regards Uffe
On Mon, Feb 13, 2017 at 04:32:32PM +0100, Ulf Hansson wrote: > [...] > > >> Although, I don't know of other examples, besides the runtime PM use > >> case, where non-atomic channel prepare/unprepare would make sense. Do > >> you? > > > > The primary ask for that has been to enable runtime_pm for drivers. It's not > > a new ask, but we somehow haven't gotten around to do it. > > Okay, I see. > > > > >> > As I said earlier, if we want to solve that problem a better idea is to > >> > actually split the prepare as we discussed in [1] > >> > > >> > This way we can get a non atomic descriptor allocate/prepare and release. > >> > Yes we need to redesign the APIs to solve this, but if you guys are up for > >> > it, I think we can do it and avoid any further round abouts :) > >> > >> Adding/re-designing dma APIs is a viable option to solve the runtime PM case. > >> > >> Changes would be needed for all related dma client drivers as well, > >> although if that's what we need to do - let's do it. > > > > Yes, but do bear in mind that some cases do need atomic prepare. The primary > > cases for DMA had that in mind and also submitting next transaction from the > > callback (tasklet) context, so that won't go away. > > > > It would help in other cases where clients know that they will not be in > > atomic context so we provide additional non-atomic "allocation" followed by > > prepare, so that drivers can split the work among these and people can do > > runtime_pm and other things.. > > That for sharing the details. > > It seems like some dma expert really need to be heavily involved if we > ever are going to complete this work. :-) Sure, I will help out :) If anyone of you are in Portland next week, then we can discuss these f2f. I will try taking a stab at the new API design next week. > > [...] > > >> > >> 1) Dependencies between dma drivers and dma client drivers during system > >> PM. For example, a dma client driver needs the dma controller to be > >> operational (remain system resumed), until the dma client driver itself > >> becomes system suspended. > >> > >> The *only* currently available solution for this, is to try to system > >> suspend the dma controller later than the dma client, via using the *late > >> or the *noirq system PM callbacks. This works for most cases, but it > >> becomes a problem when the dma client also needs to be system suspended at > >> the *late or the *noirq phase. Clearly this solution that doesn't scale. > >> > >> Using device links explicitly solves this problem as it allows to specify > >> this dependency between devices. > > > > Yes this is an interesting point. Yes till now people have been doing above > > to workaround this problem, but hey this is not a unique to dmaengine. Any > > subsystem which provides services to others has this issue, so the solution > > much be driver or pm framework and not unique to dmaengine. > > I definitely agree, these problems aren't unique to the dmaengine > subsystem. Exactly how/where to manage them, that I guess, is the key > question. > > However, I can't resist from finding the device links useful, as those > really do address and solve our issues from a runtime/system PM point > of view. > > > > >> 2) We won't avoid dma clients from getting -EPROBE_DEFER when requesting > >> their dma channels in their ->probe() routines. This would be possible, if > >> we can set up the device links at device initialization. > > > > Well setting those links is not practical at initialization time. Most > > modern dma controllers feature a SW mux, with multiple clients connecting > > and requesting, would we link all of them? Most of times dmaengine driver > > wont know about those.. > > Okay, I see! > > Kind regards > Uffe -- ~Vinod
On 13 February 2017 at 16:47, Vinod Koul <vinod.koul@intel.com> wrote: > On Mon, Feb 13, 2017 at 04:32:32PM +0100, Ulf Hansson wrote: >> [...] >> >> >> Although, I don't know of other examples, besides the runtime PM use >> >> case, where non-atomic channel prepare/unprepare would make sense. Do >> >> you? >> > >> > The primary ask for that has been to enable runtime_pm for drivers. It's not >> > a new ask, but we somehow haven't gotten around to do it. >> >> Okay, I see. >> >> > >> >> > As I said earlier, if we want to solve that problem a better idea is to >> >> > actually split the prepare as we discussed in [1] >> >> > >> >> > This way we can get a non atomic descriptor allocate/prepare and release. >> >> > Yes we need to redesign the APIs to solve this, but if you guys are up for >> >> > it, I think we can do it and avoid any further round abouts :) >> >> >> >> Adding/re-designing dma APIs is a viable option to solve the runtime PM case. >> >> >> >> Changes would be needed for all related dma client drivers as well, >> >> although if that's what we need to do - let's do it. >> > >> > Yes, but do bear in mind that some cases do need atomic prepare. The primary >> > cases for DMA had that in mind and also submitting next transaction from the >> > callback (tasklet) context, so that won't go away. >> > >> > It would help in other cases where clients know that they will not be in >> > atomic context so we provide additional non-atomic "allocation" followed by >> > prepare, so that drivers can split the work among these and people can do >> > runtime_pm and other things.. >> >> That for sharing the details. >> >> It seems like some dma expert really need to be heavily involved if we >> ever are going to complete this work. :-) > > Sure, I will help out :) That sounds great! :-) > > If anyone of you are in Portland next week, then we can discuss these f2f. I > will try taking a stab at the new API design next week. > Unfortunate not. We will have to meet some other time. Anyway, please keep me posted on any related topics. Kind regards Uffe
diff --git a/drivers/dma/pl330.c b/drivers/dma/pl330.c index 8b0da7fa520d..17efad418faa 100644 --- a/drivers/dma/pl330.c +++ b/drivers/dma/pl330.c @@ -22,6 +22,7 @@ #include <linux/dma-mapping.h> #include <linux/dmaengine.h> #include <linux/amba/bus.h> +#include <linux/mutex.h> #include <linux/scatterlist.h> #include <linux/of.h> #include <linux/of_dma.h> @@ -268,9 +269,6 @@ enum pl330_byteswap { #define NR_DEFAULT_DESC 16 -/* Delay for runtime PM autosuspend, ms */ -#define PL330_AUTOSUSPEND_DELAY 20 - /* Populated by the PL330 core driver for DMA API driver's info */ struct pl330_config { u32 periph_id; @@ -449,7 +447,7 @@ struct dma_pl330_chan { bool cyclic; /* for runtime pm tracking */ - bool active; + struct device_link *slave_link; }; struct pl330_dmac { @@ -463,6 +461,8 @@ struct pl330_dmac { struct list_head desc_pool; /* To protect desc_pool manipulation */ spinlock_t pool_lock; + /* For management of slave PM links */ + struct mutex rpm_lock; /* Size of MicroCode buffers for each channel. */ unsigned mcbufsz; @@ -2008,7 +2008,6 @@ static void pl330_tasklet(unsigned long data) struct dma_pl330_chan *pch = (struct dma_pl330_chan *)data; struct dma_pl330_desc *desc, *_dt; unsigned long flags; - bool power_down = false; spin_lock_irqsave(&pch->lock, flags); @@ -2023,18 +2022,10 @@ static void pl330_tasklet(unsigned long data) /* Try to submit a req imm. next to the last completed cookie */ fill_queue(pch); - if (list_empty(&pch->work_list)) { - spin_lock(&pch->thread->dmac->lock); - _stop(pch->thread); - spin_unlock(&pch->thread->dmac->lock); - power_down = true; - pch->active = false; - } else { - /* Make sure the PL330 Channel thread is active */ - spin_lock(&pch->thread->dmac->lock); - _start(pch->thread); - spin_unlock(&pch->thread->dmac->lock); - } + /* Make sure the PL330 Channel thread is active */ + spin_lock(&pch->thread->dmac->lock); + _start(pch->thread); + spin_unlock(&pch->thread->dmac->lock); while (!list_empty(&pch->completed_list)) { struct dmaengine_desc_callback cb; @@ -2047,13 +2038,6 @@ static void pl330_tasklet(unsigned long data) if (pch->cyclic) { desc->status = PREP; list_move_tail(&desc->node, &pch->work_list); - if (power_down) { - pch->active = true; - spin_lock(&pch->thread->dmac->lock); - _start(pch->thread); - spin_unlock(&pch->thread->dmac->lock); - power_down = false; - } } else { desc->status = FREE; list_move_tail(&desc->node, &pch->dmac->desc_pool); @@ -2068,12 +2052,6 @@ static void pl330_tasklet(unsigned long data) } } spin_unlock_irqrestore(&pch->lock, flags); - - /* If work list empty, power down */ - if (power_down) { - pm_runtime_mark_last_busy(pch->dmac->ddma.dev); - pm_runtime_put_autosuspend(pch->dmac->ddma.dev); - } } static struct dma_chan *of_dma_pl330_xlate(struct of_phandle_args *dma_spec, @@ -2096,11 +2074,68 @@ static struct dma_chan *of_dma_pl330_xlate(struct of_phandle_args *dma_spec, return dma_get_slave_channel(&pl330->peripherals[chan_id].chan); } +static int pl330_set_slave(struct dma_chan *chan, struct device *slave) +{ + struct dma_pl330_chan *pch = to_pchan(chan); + struct pl330_dmac *pl330 = pch->dmac; + int i; + + mutex_lock(&pl330->rpm_lock); + + for (i = 0; i < pl330->num_peripherals; i++) { + if (pl330->peripherals[i].chan.slave == slave && + pl330->peripherals[i].slave_link) { + pch->slave_link = pl330->peripherals[i].slave_link; + goto done; + } + } + + pch->slave_link = device_link_add(slave, pl330->ddma.dev, + DL_FLAG_PM_RUNTIME | DL_FLAG_RPM_ACTIVE); + if (!pch->slave_link) { + mutex_unlock(&pl330->rpm_lock); + return -ENODEV; + } +done: + mutex_unlock(&pl330->rpm_lock); + + pm_runtime_put(pl330->ddma.dev); + + return 0; +} + +static void pl330_release_slave(struct dma_chan *chan) +{ + struct dma_pl330_chan *pch = to_pchan(chan); + struct pl330_dmac *pl330 = pch->dmac; + struct device_link *link = pch->slave_link; + int i, count = 0; + + pm_runtime_get_sync(pl330->ddma.dev); + + mutex_lock(&pl330->rpm_lock); + + for (i = 0; i < pl330->num_peripherals; i++) + if (pl330->peripherals[i].slave_link == link) + count++; + + pch->slave_link = NULL; + if (count == 1) + device_link_del(link); + + mutex_unlock(&pl330->rpm_lock); +} + static int pl330_alloc_chan_resources(struct dma_chan *chan) { struct dma_pl330_chan *pch = to_pchan(chan); struct pl330_dmac *pl330 = pch->dmac; unsigned long flags; + int ret; + + ret = pm_runtime_get_sync(pl330->ddma.dev); + if (ret < 0) + return ret; spin_lock_irqsave(&pl330->lock, flags); @@ -2110,6 +2145,7 @@ static int pl330_alloc_chan_resources(struct dma_chan *chan) pch->thread = pl330_request_channel(pl330); if (!pch->thread) { spin_unlock_irqrestore(&pl330->lock, flags); + pm_runtime_put(pl330->ddma.dev); return -ENOMEM; } @@ -2151,9 +2187,7 @@ static int pl330_terminate_all(struct dma_chan *chan) unsigned long flags; struct pl330_dmac *pl330 = pch->dmac; LIST_HEAD(list); - bool power_down = false; - pm_runtime_get_sync(pl330->ddma.dev); spin_lock_irqsave(&pch->lock, flags); spin_lock(&pl330->lock); _stop(pch->thread); @@ -2162,8 +2196,6 @@ static int pl330_terminate_all(struct dma_chan *chan) pch->thread->req[0].desc = NULL; pch->thread->req[1].desc = NULL; pch->thread->req_running = -1; - power_down = pch->active; - pch->active = false; /* Mark all desc done */ list_for_each_entry(desc, &pch->submitted_list, node) { @@ -2180,10 +2212,6 @@ static int pl330_terminate_all(struct dma_chan *chan) list_splice_tail_init(&pch->work_list, &pl330->desc_pool); list_splice_tail_init(&pch->completed_list, &pl330->desc_pool); spin_unlock_irqrestore(&pch->lock, flags); - pm_runtime_mark_last_busy(pl330->ddma.dev); - if (power_down) - pm_runtime_put_autosuspend(pl330->ddma.dev); - pm_runtime_put_autosuspend(pl330->ddma.dev); return 0; } @@ -2201,7 +2229,6 @@ static int pl330_pause(struct dma_chan *chan) struct pl330_dmac *pl330 = pch->dmac; unsigned long flags; - pm_runtime_get_sync(pl330->ddma.dev); spin_lock_irqsave(&pch->lock, flags); spin_lock(&pl330->lock); @@ -2209,8 +2236,6 @@ static int pl330_pause(struct dma_chan *chan) spin_unlock(&pl330->lock); spin_unlock_irqrestore(&pch->lock, flags); - pm_runtime_mark_last_busy(pl330->ddma.dev); - pm_runtime_put_autosuspend(pl330->ddma.dev); return 0; } @@ -2223,7 +2248,6 @@ static void pl330_free_chan_resources(struct dma_chan *chan) tasklet_kill(&pch->task); - pm_runtime_get_sync(pch->dmac->ddma.dev); spin_lock_irqsave(&pl330->lock, flags); pl330_release_channel(pch->thread); @@ -2233,19 +2257,17 @@ static void pl330_free_chan_resources(struct dma_chan *chan) list_splice_tail_init(&pch->work_list, &pch->dmac->desc_pool); spin_unlock_irqrestore(&pl330->lock, flags); - pm_runtime_mark_last_busy(pch->dmac->ddma.dev); - pm_runtime_put_autosuspend(pch->dmac->ddma.dev); + + pm_runtime_put(pl330->ddma.dev); } static int pl330_get_current_xferred_count(struct dma_pl330_chan *pch, struct dma_pl330_desc *desc) { struct pl330_thread *thrd = pch->thread; - struct pl330_dmac *pl330 = pch->dmac; void __iomem *regs = thrd->dmac->base; u32 val, addr; - pm_runtime_get_sync(pl330->ddma.dev); val = addr = 0; if (desc->rqcfg.src_inc) { val = readl(regs + SA(thrd->id)); @@ -2254,8 +2276,6 @@ static int pl330_get_current_xferred_count(struct dma_pl330_chan *pch, val = readl(regs + DA(thrd->id)); addr = desc->px.dst_addr; } - pm_runtime_mark_last_busy(pch->dmac->ddma.dev); - pm_runtime_put_autosuspend(pl330->ddma.dev); /* If DMAMOV hasn't finished yet, SAR/DAR can be zero */ if (!val) @@ -2341,16 +2361,6 @@ static void pl330_issue_pending(struct dma_chan *chan) unsigned long flags; spin_lock_irqsave(&pch->lock, flags); - if (list_empty(&pch->work_list)) { - /* - * Warn on nothing pending. Empty submitted_list may - * break our pm_runtime usage counter as it is - * updated on work_list emptiness status. - */ - WARN_ON(list_empty(&pch->submitted_list)); - pch->active = true; - pm_runtime_get_sync(pch->dmac->ddma.dev); - } list_splice_tail_init(&pch->submitted_list, &pch->work_list); spin_unlock_irqrestore(&pch->lock, flags); @@ -2778,44 +2788,12 @@ static irqreturn_t pl330_irq_handler(int irq, void *data) BIT(DMA_SLAVE_BUSWIDTH_8_BYTES) /* - * Runtime PM callbacks are provided by amba/bus.c driver. - * - * It is assumed here that IRQ safe runtime PM is chosen in probe and amba - * bus driver will only disable/enable the clock in runtime PM callbacks. + * Runtime PM callbacks are provided by amba/bus.c driver, system sleep + * suspend/resume is implemented by generic helpers, which use existing + * runtime PM callbacks. */ -static int __maybe_unused pl330_suspend(struct device *dev) -{ - struct amba_device *pcdev = to_amba_device(dev); - - pm_runtime_disable(dev); - - if (!pm_runtime_status_suspended(dev)) { - /* amba did not disable the clock */ - amba_pclk_disable(pcdev); - } - amba_pclk_unprepare(pcdev); - - return 0; -} - -static int __maybe_unused pl330_resume(struct device *dev) -{ - struct amba_device *pcdev = to_amba_device(dev); - int ret; - - ret = amba_pclk_prepare(pcdev); - if (ret) - return ret; - - if (!pm_runtime_status_suspended(dev)) - ret = amba_pclk_enable(pcdev); - - pm_runtime_enable(dev); - - return ret; -} - -static SIMPLE_DEV_PM_OPS(pl330_pm, pl330_suspend, pl330_resume); +static SIMPLE_DEV_PM_OPS(pl330_pm, pm_runtime_force_suspend, + pm_runtime_force_resume); static int pl330_probe(struct amba_device *adev, const struct amba_id *id) @@ -2877,6 +2855,7 @@ static int __maybe_unused pl330_resume(struct device *dev) INIT_LIST_HEAD(&pl330->desc_pool); spin_lock_init(&pl330->pool_lock); + mutex_init(&pl330->rpm_lock); /* Create a descriptor pool of default size */ if (!add_desc(pl330, GFP_KERNEL, NR_DEFAULT_DESC)) @@ -2920,6 +2899,8 @@ static int __maybe_unused pl330_resume(struct device *dev) pd->device_alloc_chan_resources = pl330_alloc_chan_resources; pd->device_free_chan_resources = pl330_free_chan_resources; + pd->device_set_slave = pl330_set_slave; + pd->device_release_slave = pl330_release_slave; pd->device_prep_dma_memcpy = pl330_prep_dma_memcpy; pd->device_prep_dma_cyclic = pl330_prep_dma_cyclic; pd->device_tx_status = pl330_tx_status; @@ -2968,11 +2949,7 @@ static int __maybe_unused pl330_resume(struct device *dev) pcfg->data_buf_dep, pcfg->data_bus_width / 8, pcfg->num_chan, pcfg->num_peri, pcfg->num_events); - pm_runtime_irq_safe(&adev->dev); - pm_runtime_use_autosuspend(&adev->dev); - pm_runtime_set_autosuspend_delay(&adev->dev, PL330_AUTOSUSPEND_DELAY); - pm_runtime_mark_last_busy(&adev->dev); - pm_runtime_put_autosuspend(&adev->dev); + pm_runtime_put(&adev->dev); return 0; probe_err3:
This patch replaces irq-safe runtime PM with non-irq-safe version based on the new approach. Existing, irq-safe runtime PM implementation for PL330 was not bringing much benefits of its own - only clocks were enabled/disabled. Another limitation of irq-safe runtime PM is a fact, that it may prevent the generic PM domain (genpd) from being powered off, particularly in cases when the genpd doesn't have the GENPD_FLAG_IRQ_SAFE set. Till now non-irq-safe runtime PM implementation was only possible by calling pm_runtime_get/put functions from alloc/free_chan_resources. All other DMA engine API functions cannot be called from a context, which permits sleeping. Such implementation, in practice would result in keeping DMA controller's device active almost all the time, because most of the slave device drivers (DMA engine clients) acquire DMA channel in their probe() function and released it during driver removal. This patch provides a new, different approach. It is based on an observation that there can be only one slave device using each DMA channel. PL330 hardware always has dedicated channels for each peripheral device. Using recently introduced device dependencies (links) infrastructure one can ensure proper runtime PM state of PL330 DMA controller basing on the runtime PM state of the slave device. In this approach in pl330_set_slave() function a new dependency is being created between PL330 DMA controller device (as a supplier) and given slave device (as a consumer). This way PL330 DMA controller device runtime active counter is increased when the slave device is resumed and decreased the same time when given slave device is put to suspend. This way it has been ensured to keep PL330 DMA controller runtime active if there is an active used of any of its DMA channels. This is similar to what has been already implemented in Exynos IOMMU driver in commit 2f5f44f205cc95 ("iommu/exynos: Use device dependency links to control runtime pm"). If slave device doesn't implement runtime PM or keeps device runtime active all the time, then PL330 DMA controller will be runtime active all the time when channel is being allocated. If one requests memory-to-memory channel, runtime active counter is increased unconditionally. This might be a drawback of this approach, but PL330 is not really used for memory-to-memory operations due to poor performance in such operations compared to the CPU. Removal of irq-safe runtime PM is based on the revert of the following commits: 1. commit 5c9e6c2b2ba3 "dmaengine: pl330: fix runtime pm support" 2. commit 81cc6edc0870 "dmaengine: pl330: Fix hang on dmaengine_terminate_all on certain boards" 3. commit ae43b3289186 "ARM: 8202/1: dmaengine: pl330: Add runtime Power Management support v12" Introducing non-irq-safe runtime power management finally allows to turn off audio power domain on Exynos5 SoCs. Signed-off-by: Marek Szyprowski <m.szyprowski@samsung.com> --- drivers/dma/pl330.c | 177 +++++++++++++++++++++++----------------------------- 1 file changed, 77 insertions(+), 100 deletions(-) -- 1.9.1