Message ID | 20220304032106.2866043-2-dmitry.baryshkov@linaro.org |
---|---|
State | Superseded |
Headers | show |
Series | drm/msm: rework MDSS drivers | expand |
On 3/3/2022 7:21 PM, Dmitry Baryshkov wrote: > MDP5 and DPU1 both provide the driver handling the MDSS region, which > handles the irq domain and (incase of DPU1) adds some init for the UBWC > controller. Unify those two pieces of code into a common driver. > > Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> Reviewed-by: Abhinav Kumar <quic_abhinavk@quicinc.com> > --- > drivers/gpu/drm/msm/Makefile | 3 +- > drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c | 252 ------------------ > drivers/gpu/drm/msm/msm_drv.c | 4 +- > drivers/gpu/drm/msm/msm_kms.h | 3 +- > .../msm/{disp/dpu1/dpu_mdss.c => msm_mdss.c} | 145 +++++----- > 5 files changed, 83 insertions(+), 324 deletions(-) > delete mode 100644 drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c > rename drivers/gpu/drm/msm/{disp/dpu1/dpu_mdss.c => msm_mdss.c} (63%) > > diff --git a/drivers/gpu/drm/msm/Makefile b/drivers/gpu/drm/msm/Makefile > index e9cc7d8ac301..e76927b42033 100644 > --- a/drivers/gpu/drm/msm/Makefile > +++ b/drivers/gpu/drm/msm/Makefile > @@ -42,7 +42,6 @@ msm-y := \ > disp/mdp5/mdp5_crtc.o \ > disp/mdp5/mdp5_encoder.o \ > disp/mdp5/mdp5_irq.o \ > - disp/mdp5/mdp5_mdss.o \ > disp/mdp5/mdp5_kms.o \ > disp/mdp5/mdp5_pipe.o \ > disp/mdp5/mdp5_mixer.o \ > @@ -67,7 +66,6 @@ msm-y := \ > disp/dpu1/dpu_hw_util.o \ > disp/dpu1/dpu_hw_vbif.o \ > disp/dpu1/dpu_kms.o \ > - disp/dpu1/dpu_mdss.o \ > disp/dpu1/dpu_plane.o \ > disp/dpu1/dpu_rm.o \ > disp/dpu1/dpu_vbif.o \ > @@ -88,6 +86,7 @@ msm-y := \ > msm_gpu_devfreq.o \ > msm_io_utils.o \ > msm_iommu.o \ > + msm_mdss.o \ > msm_perf.o \ > msm_rd.o \ > msm_ringbuffer.o \ > diff --git a/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c b/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c > deleted file mode 100644 > index 049c6784a531..000000000000 > --- a/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c > +++ /dev/null > @@ -1,252 +0,0 @@ > -// SPDX-License-Identifier: GPL-2.0-only > -/* > - * Copyright (c) 2016, The Linux Foundation. All rights reserved. > - */ > - > -#include <linux/irqdomain.h> > -#include <linux/irq.h> > - > -#include "msm_drv.h" > -#include "mdp5_kms.h" > - > -#define to_mdp5_mdss(x) container_of(x, struct mdp5_mdss, base) > - > -struct mdp5_mdss { > - struct msm_mdss base; > - > - void __iomem *mmio, *vbif; > - > - struct clk *ahb_clk; > - struct clk *axi_clk; > - struct clk *vsync_clk; > - > - struct { > - volatile unsigned long enabled_mask; > - struct irq_domain *domain; > - } irqcontroller; > -}; > - > -static inline void mdss_write(struct mdp5_mdss *mdp5_mdss, u32 reg, u32 data) > -{ > - msm_writel(data, mdp5_mdss->mmio + reg); > -} > - > -static inline u32 mdss_read(struct mdp5_mdss *mdp5_mdss, u32 reg) > -{ > - return msm_readl(mdp5_mdss->mmio + reg); > -} > - > -static irqreturn_t mdss_irq(int irq, void *arg) > -{ > - struct mdp5_mdss *mdp5_mdss = arg; > - u32 intr; > - > - intr = mdss_read(mdp5_mdss, REG_MDSS_HW_INTR_STATUS); > - > - VERB("intr=%08x", intr); > - > - while (intr) { > - irq_hw_number_t hwirq = fls(intr) - 1; > - > - generic_handle_domain_irq(mdp5_mdss->irqcontroller.domain, hwirq); > - intr &= ~(1 << hwirq); > - } > - > - return IRQ_HANDLED; > -} > - > -/* > - * interrupt-controller implementation, so sub-blocks (MDP/HDMI/eDP/DSI/etc) > - * can register to get their irq's delivered > - */ > - > -#define VALID_IRQS (MDSS_HW_INTR_STATUS_INTR_MDP | \ > - MDSS_HW_INTR_STATUS_INTR_DSI0 | \ > - MDSS_HW_INTR_STATUS_INTR_DSI1 | \ > - MDSS_HW_INTR_STATUS_INTR_HDMI | \ > - MDSS_HW_INTR_STATUS_INTR_EDP) > - > -static void mdss_hw_mask_irq(struct irq_data *irqd) > -{ > - struct mdp5_mdss *mdp5_mdss = irq_data_get_irq_chip_data(irqd); > - > - smp_mb__before_atomic(); > - clear_bit(irqd->hwirq, &mdp5_mdss->irqcontroller.enabled_mask); > - smp_mb__after_atomic(); > -} > - > -static void mdss_hw_unmask_irq(struct irq_data *irqd) > -{ > - struct mdp5_mdss *mdp5_mdss = irq_data_get_irq_chip_data(irqd); > - > - smp_mb__before_atomic(); > - set_bit(irqd->hwirq, &mdp5_mdss->irqcontroller.enabled_mask); > - smp_mb__after_atomic(); > -} > - > -static struct irq_chip mdss_hw_irq_chip = { > - .name = "mdss", > - .irq_mask = mdss_hw_mask_irq, > - .irq_unmask = mdss_hw_unmask_irq, > -}; > - > -static int mdss_hw_irqdomain_map(struct irq_domain *d, unsigned int irq, > - irq_hw_number_t hwirq) > -{ > - struct mdp5_mdss *mdp5_mdss = d->host_data; > - > - if (!(VALID_IRQS & (1 << hwirq))) > - return -EPERM; > - > - irq_set_chip_and_handler(irq, &mdss_hw_irq_chip, handle_level_irq); > - irq_set_chip_data(irq, mdp5_mdss); > - > - return 0; > -} > - > -static const struct irq_domain_ops mdss_hw_irqdomain_ops = { > - .map = mdss_hw_irqdomain_map, > - .xlate = irq_domain_xlate_onecell, > -}; > - > - > -static int mdss_irq_domain_init(struct mdp5_mdss *mdp5_mdss) > -{ > - struct device *dev = mdp5_mdss->base.dev; > - struct irq_domain *d; > - > - d = irq_domain_add_linear(dev->of_node, 32, &mdss_hw_irqdomain_ops, > - mdp5_mdss); > - if (!d) { > - DRM_DEV_ERROR(dev, "mdss irq domain add failed\n"); > - return -ENXIO; > - } > - > - mdp5_mdss->irqcontroller.enabled_mask = 0; > - mdp5_mdss->irqcontroller.domain = d; > - > - return 0; > -} > - > -static int mdp5_mdss_enable(struct msm_mdss *mdss) > -{ > - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); > - DBG(""); > - > - clk_prepare_enable(mdp5_mdss->ahb_clk); > - clk_prepare_enable(mdp5_mdss->axi_clk); > - clk_prepare_enable(mdp5_mdss->vsync_clk); > - > - return 0; > -} > - > -static int mdp5_mdss_disable(struct msm_mdss *mdss) > -{ > - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); > - DBG(""); > - > - clk_disable_unprepare(mdp5_mdss->vsync_clk); > - clk_disable_unprepare(mdp5_mdss->axi_clk); > - clk_disable_unprepare(mdp5_mdss->ahb_clk); > - > - return 0; > -} > - > -static int msm_mdss_get_clocks(struct mdp5_mdss *mdp5_mdss) > -{ > - struct platform_device *pdev = > - to_platform_device(mdp5_mdss->base.dev); > - > - mdp5_mdss->ahb_clk = msm_clk_get(pdev, "iface"); > - if (IS_ERR(mdp5_mdss->ahb_clk)) > - mdp5_mdss->ahb_clk = NULL; > - > - mdp5_mdss->axi_clk = msm_clk_get(pdev, "bus"); > - if (IS_ERR(mdp5_mdss->axi_clk)) > - mdp5_mdss->axi_clk = NULL; > - > - mdp5_mdss->vsync_clk = msm_clk_get(pdev, "vsync"); > - if (IS_ERR(mdp5_mdss->vsync_clk)) > - mdp5_mdss->vsync_clk = NULL; > - > - return 0; > -} > - > -static void mdp5_mdss_destroy(struct msm_mdss *mdss) > -{ > - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); > - > - if (!mdp5_mdss) > - return; > - > - irq_domain_remove(mdp5_mdss->irqcontroller.domain); > - mdp5_mdss->irqcontroller.domain = NULL; > - > - pm_runtime_disable(mdss->dev); > -} > - > -static const struct msm_mdss_funcs mdss_funcs = { > - .enable = mdp5_mdss_enable, > - .disable = mdp5_mdss_disable, > - .destroy = mdp5_mdss_destroy, > -}; > - > -int mdp5_mdss_init(struct platform_device *pdev) > -{ > - struct msm_drm_private *priv = platform_get_drvdata(pdev); > - struct mdp5_mdss *mdp5_mdss; > - int ret; > - > - DBG(""); > - > - if (!of_device_is_compatible(pdev->dev.of_node, "qcom,mdss")) > - return 0; > - > - mdp5_mdss = devm_kzalloc(&pdev->dev, sizeof(*mdp5_mdss), GFP_KERNEL); > - if (!mdp5_mdss) { > - ret = -ENOMEM; > - goto fail; > - } > - > - mdp5_mdss->base.dev = &pdev->dev; > - > - mdp5_mdss->mmio = msm_ioremap(pdev, "mdss_phys"); > - if (IS_ERR(mdp5_mdss->mmio)) { > - ret = PTR_ERR(mdp5_mdss->mmio); > - goto fail; > - } > - > - mdp5_mdss->vbif = msm_ioremap(pdev, "vbif_phys"); > - if (IS_ERR(mdp5_mdss->vbif)) { > - ret = PTR_ERR(mdp5_mdss->vbif); > - goto fail; > - } > - > - ret = msm_mdss_get_clocks(mdp5_mdss); > - if (ret) { > - DRM_DEV_ERROR(&pdev->dev, "failed to get clocks: %d\n", ret); > - goto fail; > - } > - > - ret = devm_request_irq(&pdev->dev, platform_get_irq(pdev, 0), > - mdss_irq, 0, "mdss_isr", mdp5_mdss); > - if (ret) { > - DRM_DEV_ERROR(&pdev->dev, "failed to init irq: %d\n", ret); > - goto fail; > - } > - > - ret = mdss_irq_domain_init(mdp5_mdss); > - if (ret) { > - DRM_DEV_ERROR(&pdev->dev, "failed to init sub-block irqs: %d\n", ret); > - goto fail; > - } > - > - mdp5_mdss->base.funcs = &mdss_funcs; > - priv->mdss = &mdp5_mdss->base; > - > - pm_runtime_enable(&pdev->dev); > - > - return 0; > -fail: > - return ret; > -} > diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c > index 30fd18ca88c4..078c7e951a6e 100644 > --- a/drivers/gpu/drm/msm/msm_drv.c > +++ b/drivers/gpu/drm/msm/msm_drv.c > @@ -1208,10 +1208,10 @@ static int msm_pdev_probe(struct platform_device *pdev) > > switch (get_mdp_ver(pdev)) { > case KMS_MDP5: > - ret = mdp5_mdss_init(pdev); > + ret = msm_mdss_init(pdev, true); > break; > case KMS_DPU: > - ret = dpu_mdss_init(pdev); > + ret = msm_mdss_init(pdev, false); > break; > default: > ret = 0; > diff --git a/drivers/gpu/drm/msm/msm_kms.h b/drivers/gpu/drm/msm/msm_kms.h > index 2a4f0526cb98..10d5ae3e76df 100644 > --- a/drivers/gpu/drm/msm/msm_kms.h > +++ b/drivers/gpu/drm/msm/msm_kms.h > @@ -212,8 +212,7 @@ struct msm_mdss { > const struct msm_mdss_funcs *funcs; > }; > > -int mdp5_mdss_init(struct platform_device *dev); > -int dpu_mdss_init(struct platform_device *dev); > +int msm_mdss_init(struct platform_device *pdev, bool is_mdp5); > > #define for_each_crtc_mask(dev, crtc, crtc_mask) \ > drm_for_each_crtc(crtc, dev) \ > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c b/drivers/gpu/drm/msm/msm_mdss.c > similarity index 63% > rename from drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c > rename to drivers/gpu/drm/msm/msm_mdss.c > index b10ca505f9ac..71f3277bde32 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c > +++ b/drivers/gpu/drm/msm/msm_mdss.c > @@ -7,7 +7,12 @@ > #include <linux/irqchip.h> > #include <linux/irqdesc.h> > #include <linux/irqchip/chained_irq.h> > -#include "dpu_kms.h" > + > +#include "msm_drv.h" > +#include "msm_kms.h" > + > +/* for DPU_HW_* defines */ > +#include "disp/dpu1/dpu_hw_catalog.h" > > #define to_dpu_mdss(x) container_of(x, struct dpu_mdss, base) > > @@ -18,23 +23,18 @@ > #define UBWC_CTRL_2 0x150 > #define UBWC_PREDICTION_MODE 0x154 > > -/* Max BW defined in KBps */ > -#define MAX_BW 6800000 > - > -struct dpu_irq_controller { > - unsigned long enabled_mask; > - struct irq_domain *domain; > -}; > - > struct dpu_mdss { > struct msm_mdss base; > void __iomem *mmio; > struct clk_bulk_data *clocks; > size_t num_clocks; > - struct dpu_irq_controller irq_controller; > + struct { > + unsigned long enabled_mask; > + struct irq_domain *domain; > + } irq_controller; > }; > > -static void dpu_mdss_irq(struct irq_desc *desc) > +static void msm_mdss_irq(struct irq_desc *desc) > { > struct dpu_mdss *dpu_mdss = irq_desc_get_handler_data(desc); > struct irq_chip *chip = irq_desc_get_chip(desc); > @@ -62,7 +62,7 @@ static void dpu_mdss_irq(struct irq_desc *desc) > chained_irq_exit(chip, desc); > } > > -static void dpu_mdss_irq_mask(struct irq_data *irqd) > +static void msm_mdss_irq_mask(struct irq_data *irqd) > { > struct dpu_mdss *dpu_mdss = irq_data_get_irq_chip_data(irqd); > > @@ -73,7 +73,7 @@ static void dpu_mdss_irq_mask(struct irq_data *irqd) > smp_mb__after_atomic(); > } > > -static void dpu_mdss_irq_unmask(struct irq_data *irqd) > +static void msm_mdss_irq_unmask(struct irq_data *irqd) > { > struct dpu_mdss *dpu_mdss = irq_data_get_irq_chip_data(irqd); > > @@ -84,30 +84,31 @@ static void dpu_mdss_irq_unmask(struct irq_data *irqd) > smp_mb__after_atomic(); > } > > -static struct irq_chip dpu_mdss_irq_chip = { > +static struct irq_chip msm_mdss_irq_chip = { > .name = "dpu_mdss", > - .irq_mask = dpu_mdss_irq_mask, > - .irq_unmask = dpu_mdss_irq_unmask, > + .irq_mask = msm_mdss_irq_mask, > + .irq_unmask = msm_mdss_irq_unmask, > }; > > -static struct lock_class_key dpu_mdss_lock_key, dpu_mdss_request_key; > +static struct lock_class_key msm_mdss_lock_key, msm_mdss_request_key; > > -static int dpu_mdss_irqdomain_map(struct irq_domain *domain, > +static int msm_mdss_irqdomain_map(struct irq_domain *domain, > unsigned int irq, irq_hw_number_t hwirq) > { > struct dpu_mdss *dpu_mdss = domain->host_data; > > - irq_set_lockdep_class(irq, &dpu_mdss_lock_key, &dpu_mdss_request_key); > - irq_set_chip_and_handler(irq, &dpu_mdss_irq_chip, handle_level_irq); > + irq_set_lockdep_class(irq, &msm_mdss_lock_key, &msm_mdss_request_key); > + irq_set_chip_and_handler(irq, &msm_mdss_irq_chip, handle_level_irq); > + > return irq_set_chip_data(irq, dpu_mdss); > } > > -static const struct irq_domain_ops dpu_mdss_irqdomain_ops = { > - .map = dpu_mdss_irqdomain_map, > +static const struct irq_domain_ops msm_mdss_irqdomain_ops = { > + .map = msm_mdss_irqdomain_map, > .xlate = irq_domain_xlate_onecell, > }; > > -static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) > +static int _msm_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) > { > struct device *dev; > struct irq_domain *domain; > @@ -115,9 +116,9 @@ static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) > dev = dpu_mdss->base.dev; > > domain = irq_domain_add_linear(dev->of_node, 32, > - &dpu_mdss_irqdomain_ops, dpu_mdss); > + &msm_mdss_irqdomain_ops, dpu_mdss); > if (!domain) { > - DPU_ERROR("failed to add irq_domain\n"); > + DRM_ERROR("failed to add irq_domain\n"); > return -EINVAL; > } > > @@ -127,21 +128,14 @@ static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) > return 0; > } > > -static void _dpu_mdss_irq_domain_fini(struct dpu_mdss *dpu_mdss) > -{ > - if (dpu_mdss->irq_controller.domain) { > - irq_domain_remove(dpu_mdss->irq_controller.domain); > - dpu_mdss->irq_controller.domain = NULL; > - } > -} > -static int dpu_mdss_enable(struct msm_mdss *mdss) > +static int msm_mdss_enable(struct msm_mdss *mdss) > { > struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); > int ret; > > ret = clk_bulk_prepare_enable(dpu_mdss->num_clocks, dpu_mdss->clocks); > if (ret) { > - DPU_ERROR("clock enable failed, ret:%d\n", ret); > + DRM_ERROR("clock enable failed, ret:%d\n", ret); > return ret; > } > > @@ -171,7 +165,7 @@ static int dpu_mdss_enable(struct msm_mdss *mdss) > return ret; > } > > -static int dpu_mdss_disable(struct msm_mdss *mdss) > +static int msm_mdss_disable(struct msm_mdss *mdss) > { > struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); > > @@ -180,7 +174,7 @@ static int dpu_mdss_disable(struct msm_mdss *mdss) > return 0; > } > > -static void dpu_mdss_destroy(struct msm_mdss *mdss) > +static void msm_mdss_destroy(struct msm_mdss *mdss) > { > struct platform_device *pdev = to_platform_device(mdss->dev); > struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); > @@ -188,22 +182,49 @@ static void dpu_mdss_destroy(struct msm_mdss *mdss) > > pm_runtime_suspend(mdss->dev); > pm_runtime_disable(mdss->dev); > - _dpu_mdss_irq_domain_fini(dpu_mdss); > + irq_domain_remove(dpu_mdss->irq_controller.domain); > + dpu_mdss->irq_controller.domain = NULL; > irq = platform_get_irq(pdev, 0); > irq_set_chained_handler_and_data(irq, NULL, NULL); > - > - if (dpu_mdss->mmio) > - devm_iounmap(&pdev->dev, dpu_mdss->mmio); > - dpu_mdss->mmio = NULL; > } > > static const struct msm_mdss_funcs mdss_funcs = { > - .enable = dpu_mdss_enable, > - .disable = dpu_mdss_disable, > - .destroy = dpu_mdss_destroy, > + .enable = msm_mdss_enable, > + .disable = msm_mdss_disable, > + .destroy = msm_mdss_destroy, > }; > > -int dpu_mdss_init(struct platform_device *pdev) > +/* > + * MDP5 MDSS uses at most three specified clocks. > + */ > +#define MDP5_MDSS_NUM_CLOCKS 3 > +static int mdp5_mdss_parse_clock(struct platform_device *pdev, struct clk_bulk_data **clocks) > +{ > + struct clk_bulk_data *bulk; > + int num_clocks = 0; > + int ret; > + > + if (!pdev) > + return -EINVAL; > + > + bulk = devm_kcalloc(&pdev->dev, MDP5_MDSS_NUM_CLOCKS, sizeof(struct clk_bulk_data), GFP_KERNEL); > + if (!bulk) > + return -ENOMEM; > + > + bulk[num_clocks++].id = "iface"; > + bulk[num_clocks++].id = "bus"; > + bulk[num_clocks++].id = "vsync"; > + > + ret = devm_clk_bulk_get_optional(&pdev->dev, num_clocks, bulk); > + if (ret) > + return ret; > + > + *clocks = bulk; > + > + return num_clocks; > +} > + > +int msm_mdss_init(struct platform_device *pdev, bool is_mdp5) > { > struct msm_drm_private *priv = platform_get_drvdata(pdev); > struct dpu_mdss *dpu_mdss; > @@ -220,27 +241,28 @@ int dpu_mdss_init(struct platform_device *pdev) > > DRM_DEBUG("mapped mdss address space @%pK\n", dpu_mdss->mmio); > > - ret = devm_clk_bulk_get_all(&pdev->dev, &dpu_mdss->clocks); > + if (is_mdp5) > + ret = mdp5_mdss_parse_clock(pdev, &dpu_mdss->clocks); > + else > + ret = devm_clk_bulk_get_all(&pdev->dev, &dpu_mdss->clocks); > if (ret < 0) { > - DPU_ERROR("failed to parse clocks, ret=%d\n", ret); > - goto clk_parse_err; > + DRM_ERROR("failed to parse clocks, ret=%d\n", ret); > + return ret; > } > dpu_mdss->num_clocks = ret; > > dpu_mdss->base.dev = &pdev->dev; > dpu_mdss->base.funcs = &mdss_funcs; > > - ret = _dpu_mdss_irq_domain_add(dpu_mdss); > - if (ret) > - goto irq_domain_error; > - > irq = platform_get_irq(pdev, 0); > - if (irq < 0) { > - ret = irq; > - goto irq_error; > - } > + if (irq < 0) > + return irq; > > - irq_set_chained_handler_and_data(irq, dpu_mdss_irq, > + ret = _msm_mdss_irq_domain_add(dpu_mdss); > + if (ret) > + return ret; > + > + irq_set_chained_handler_and_data(irq, msm_mdss_irq, > dpu_mdss); > > priv->mdss = &dpu_mdss->base; > @@ -248,13 +270,4 @@ int dpu_mdss_init(struct platform_device *pdev) > pm_runtime_enable(&pdev->dev); > > return 0; > - > -irq_error: > - _dpu_mdss_irq_domain_fini(dpu_mdss); > -irq_domain_error: > -clk_parse_err: > - if (dpu_mdss->mmio) > - devm_iounmap(&pdev->dev, dpu_mdss->mmio); > - dpu_mdss->mmio = NULL; > - return ret; > }
Quoting Dmitry Baryshkov (2022-03-03 19:21:02) > MDP5 and DPU1 both provide the driver handling the MDSS region, which > handles the irq domain and (incase of DPU1) adds some init for the UBWC > controller. Unify those two pieces of code into a common driver. > > Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> > --- Reviewed-by: Stephen Boyd <swboyd@chromium.org>
diff --git a/drivers/gpu/drm/msm/Makefile b/drivers/gpu/drm/msm/Makefile index e9cc7d8ac301..e76927b42033 100644 --- a/drivers/gpu/drm/msm/Makefile +++ b/drivers/gpu/drm/msm/Makefile @@ -42,7 +42,6 @@ msm-y := \ disp/mdp5/mdp5_crtc.o \ disp/mdp5/mdp5_encoder.o \ disp/mdp5/mdp5_irq.o \ - disp/mdp5/mdp5_mdss.o \ disp/mdp5/mdp5_kms.o \ disp/mdp5/mdp5_pipe.o \ disp/mdp5/mdp5_mixer.o \ @@ -67,7 +66,6 @@ msm-y := \ disp/dpu1/dpu_hw_util.o \ disp/dpu1/dpu_hw_vbif.o \ disp/dpu1/dpu_kms.o \ - disp/dpu1/dpu_mdss.o \ disp/dpu1/dpu_plane.o \ disp/dpu1/dpu_rm.o \ disp/dpu1/dpu_vbif.o \ @@ -88,6 +86,7 @@ msm-y := \ msm_gpu_devfreq.o \ msm_io_utils.o \ msm_iommu.o \ + msm_mdss.o \ msm_perf.o \ msm_rd.o \ msm_ringbuffer.o \ diff --git a/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c b/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c deleted file mode 100644 index 049c6784a531..000000000000 --- a/drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c +++ /dev/null @@ -1,252 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0-only -/* - * Copyright (c) 2016, The Linux Foundation. All rights reserved. - */ - -#include <linux/irqdomain.h> -#include <linux/irq.h> - -#include "msm_drv.h" -#include "mdp5_kms.h" - -#define to_mdp5_mdss(x) container_of(x, struct mdp5_mdss, base) - -struct mdp5_mdss { - struct msm_mdss base; - - void __iomem *mmio, *vbif; - - struct clk *ahb_clk; - struct clk *axi_clk; - struct clk *vsync_clk; - - struct { - volatile unsigned long enabled_mask; - struct irq_domain *domain; - } irqcontroller; -}; - -static inline void mdss_write(struct mdp5_mdss *mdp5_mdss, u32 reg, u32 data) -{ - msm_writel(data, mdp5_mdss->mmio + reg); -} - -static inline u32 mdss_read(struct mdp5_mdss *mdp5_mdss, u32 reg) -{ - return msm_readl(mdp5_mdss->mmio + reg); -} - -static irqreturn_t mdss_irq(int irq, void *arg) -{ - struct mdp5_mdss *mdp5_mdss = arg; - u32 intr; - - intr = mdss_read(mdp5_mdss, REG_MDSS_HW_INTR_STATUS); - - VERB("intr=%08x", intr); - - while (intr) { - irq_hw_number_t hwirq = fls(intr) - 1; - - generic_handle_domain_irq(mdp5_mdss->irqcontroller.domain, hwirq); - intr &= ~(1 << hwirq); - } - - return IRQ_HANDLED; -} - -/* - * interrupt-controller implementation, so sub-blocks (MDP/HDMI/eDP/DSI/etc) - * can register to get their irq's delivered - */ - -#define VALID_IRQS (MDSS_HW_INTR_STATUS_INTR_MDP | \ - MDSS_HW_INTR_STATUS_INTR_DSI0 | \ - MDSS_HW_INTR_STATUS_INTR_DSI1 | \ - MDSS_HW_INTR_STATUS_INTR_HDMI | \ - MDSS_HW_INTR_STATUS_INTR_EDP) - -static void mdss_hw_mask_irq(struct irq_data *irqd) -{ - struct mdp5_mdss *mdp5_mdss = irq_data_get_irq_chip_data(irqd); - - smp_mb__before_atomic(); - clear_bit(irqd->hwirq, &mdp5_mdss->irqcontroller.enabled_mask); - smp_mb__after_atomic(); -} - -static void mdss_hw_unmask_irq(struct irq_data *irqd) -{ - struct mdp5_mdss *mdp5_mdss = irq_data_get_irq_chip_data(irqd); - - smp_mb__before_atomic(); - set_bit(irqd->hwirq, &mdp5_mdss->irqcontroller.enabled_mask); - smp_mb__after_atomic(); -} - -static struct irq_chip mdss_hw_irq_chip = { - .name = "mdss", - .irq_mask = mdss_hw_mask_irq, - .irq_unmask = mdss_hw_unmask_irq, -}; - -static int mdss_hw_irqdomain_map(struct irq_domain *d, unsigned int irq, - irq_hw_number_t hwirq) -{ - struct mdp5_mdss *mdp5_mdss = d->host_data; - - if (!(VALID_IRQS & (1 << hwirq))) - return -EPERM; - - irq_set_chip_and_handler(irq, &mdss_hw_irq_chip, handle_level_irq); - irq_set_chip_data(irq, mdp5_mdss); - - return 0; -} - -static const struct irq_domain_ops mdss_hw_irqdomain_ops = { - .map = mdss_hw_irqdomain_map, - .xlate = irq_domain_xlate_onecell, -}; - - -static int mdss_irq_domain_init(struct mdp5_mdss *mdp5_mdss) -{ - struct device *dev = mdp5_mdss->base.dev; - struct irq_domain *d; - - d = irq_domain_add_linear(dev->of_node, 32, &mdss_hw_irqdomain_ops, - mdp5_mdss); - if (!d) { - DRM_DEV_ERROR(dev, "mdss irq domain add failed\n"); - return -ENXIO; - } - - mdp5_mdss->irqcontroller.enabled_mask = 0; - mdp5_mdss->irqcontroller.domain = d; - - return 0; -} - -static int mdp5_mdss_enable(struct msm_mdss *mdss) -{ - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); - DBG(""); - - clk_prepare_enable(mdp5_mdss->ahb_clk); - clk_prepare_enable(mdp5_mdss->axi_clk); - clk_prepare_enable(mdp5_mdss->vsync_clk); - - return 0; -} - -static int mdp5_mdss_disable(struct msm_mdss *mdss) -{ - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); - DBG(""); - - clk_disable_unprepare(mdp5_mdss->vsync_clk); - clk_disable_unprepare(mdp5_mdss->axi_clk); - clk_disable_unprepare(mdp5_mdss->ahb_clk); - - return 0; -} - -static int msm_mdss_get_clocks(struct mdp5_mdss *mdp5_mdss) -{ - struct platform_device *pdev = - to_platform_device(mdp5_mdss->base.dev); - - mdp5_mdss->ahb_clk = msm_clk_get(pdev, "iface"); - if (IS_ERR(mdp5_mdss->ahb_clk)) - mdp5_mdss->ahb_clk = NULL; - - mdp5_mdss->axi_clk = msm_clk_get(pdev, "bus"); - if (IS_ERR(mdp5_mdss->axi_clk)) - mdp5_mdss->axi_clk = NULL; - - mdp5_mdss->vsync_clk = msm_clk_get(pdev, "vsync"); - if (IS_ERR(mdp5_mdss->vsync_clk)) - mdp5_mdss->vsync_clk = NULL; - - return 0; -} - -static void mdp5_mdss_destroy(struct msm_mdss *mdss) -{ - struct mdp5_mdss *mdp5_mdss = to_mdp5_mdss(mdss); - - if (!mdp5_mdss) - return; - - irq_domain_remove(mdp5_mdss->irqcontroller.domain); - mdp5_mdss->irqcontroller.domain = NULL; - - pm_runtime_disable(mdss->dev); -} - -static const struct msm_mdss_funcs mdss_funcs = { - .enable = mdp5_mdss_enable, - .disable = mdp5_mdss_disable, - .destroy = mdp5_mdss_destroy, -}; - -int mdp5_mdss_init(struct platform_device *pdev) -{ - struct msm_drm_private *priv = platform_get_drvdata(pdev); - struct mdp5_mdss *mdp5_mdss; - int ret; - - DBG(""); - - if (!of_device_is_compatible(pdev->dev.of_node, "qcom,mdss")) - return 0; - - mdp5_mdss = devm_kzalloc(&pdev->dev, sizeof(*mdp5_mdss), GFP_KERNEL); - if (!mdp5_mdss) { - ret = -ENOMEM; - goto fail; - } - - mdp5_mdss->base.dev = &pdev->dev; - - mdp5_mdss->mmio = msm_ioremap(pdev, "mdss_phys"); - if (IS_ERR(mdp5_mdss->mmio)) { - ret = PTR_ERR(mdp5_mdss->mmio); - goto fail; - } - - mdp5_mdss->vbif = msm_ioremap(pdev, "vbif_phys"); - if (IS_ERR(mdp5_mdss->vbif)) { - ret = PTR_ERR(mdp5_mdss->vbif); - goto fail; - } - - ret = msm_mdss_get_clocks(mdp5_mdss); - if (ret) { - DRM_DEV_ERROR(&pdev->dev, "failed to get clocks: %d\n", ret); - goto fail; - } - - ret = devm_request_irq(&pdev->dev, platform_get_irq(pdev, 0), - mdss_irq, 0, "mdss_isr", mdp5_mdss); - if (ret) { - DRM_DEV_ERROR(&pdev->dev, "failed to init irq: %d\n", ret); - goto fail; - } - - ret = mdss_irq_domain_init(mdp5_mdss); - if (ret) { - DRM_DEV_ERROR(&pdev->dev, "failed to init sub-block irqs: %d\n", ret); - goto fail; - } - - mdp5_mdss->base.funcs = &mdss_funcs; - priv->mdss = &mdp5_mdss->base; - - pm_runtime_enable(&pdev->dev); - - return 0; -fail: - return ret; -} diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c index 30fd18ca88c4..078c7e951a6e 100644 --- a/drivers/gpu/drm/msm/msm_drv.c +++ b/drivers/gpu/drm/msm/msm_drv.c @@ -1208,10 +1208,10 @@ static int msm_pdev_probe(struct platform_device *pdev) switch (get_mdp_ver(pdev)) { case KMS_MDP5: - ret = mdp5_mdss_init(pdev); + ret = msm_mdss_init(pdev, true); break; case KMS_DPU: - ret = dpu_mdss_init(pdev); + ret = msm_mdss_init(pdev, false); break; default: ret = 0; diff --git a/drivers/gpu/drm/msm/msm_kms.h b/drivers/gpu/drm/msm/msm_kms.h index 2a4f0526cb98..10d5ae3e76df 100644 --- a/drivers/gpu/drm/msm/msm_kms.h +++ b/drivers/gpu/drm/msm/msm_kms.h @@ -212,8 +212,7 @@ struct msm_mdss { const struct msm_mdss_funcs *funcs; }; -int mdp5_mdss_init(struct platform_device *dev); -int dpu_mdss_init(struct platform_device *dev); +int msm_mdss_init(struct platform_device *pdev, bool is_mdp5); #define for_each_crtc_mask(dev, crtc, crtc_mask) \ drm_for_each_crtc(crtc, dev) \ diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c b/drivers/gpu/drm/msm/msm_mdss.c similarity index 63% rename from drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c rename to drivers/gpu/drm/msm/msm_mdss.c index b10ca505f9ac..71f3277bde32 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_mdss.c +++ b/drivers/gpu/drm/msm/msm_mdss.c @@ -7,7 +7,12 @@ #include <linux/irqchip.h> #include <linux/irqdesc.h> #include <linux/irqchip/chained_irq.h> -#include "dpu_kms.h" + +#include "msm_drv.h" +#include "msm_kms.h" + +/* for DPU_HW_* defines */ +#include "disp/dpu1/dpu_hw_catalog.h" #define to_dpu_mdss(x) container_of(x, struct dpu_mdss, base) @@ -18,23 +23,18 @@ #define UBWC_CTRL_2 0x150 #define UBWC_PREDICTION_MODE 0x154 -/* Max BW defined in KBps */ -#define MAX_BW 6800000 - -struct dpu_irq_controller { - unsigned long enabled_mask; - struct irq_domain *domain; -}; - struct dpu_mdss { struct msm_mdss base; void __iomem *mmio; struct clk_bulk_data *clocks; size_t num_clocks; - struct dpu_irq_controller irq_controller; + struct { + unsigned long enabled_mask; + struct irq_domain *domain; + } irq_controller; }; -static void dpu_mdss_irq(struct irq_desc *desc) +static void msm_mdss_irq(struct irq_desc *desc) { struct dpu_mdss *dpu_mdss = irq_desc_get_handler_data(desc); struct irq_chip *chip = irq_desc_get_chip(desc); @@ -62,7 +62,7 @@ static void dpu_mdss_irq(struct irq_desc *desc) chained_irq_exit(chip, desc); } -static void dpu_mdss_irq_mask(struct irq_data *irqd) +static void msm_mdss_irq_mask(struct irq_data *irqd) { struct dpu_mdss *dpu_mdss = irq_data_get_irq_chip_data(irqd); @@ -73,7 +73,7 @@ static void dpu_mdss_irq_mask(struct irq_data *irqd) smp_mb__after_atomic(); } -static void dpu_mdss_irq_unmask(struct irq_data *irqd) +static void msm_mdss_irq_unmask(struct irq_data *irqd) { struct dpu_mdss *dpu_mdss = irq_data_get_irq_chip_data(irqd); @@ -84,30 +84,31 @@ static void dpu_mdss_irq_unmask(struct irq_data *irqd) smp_mb__after_atomic(); } -static struct irq_chip dpu_mdss_irq_chip = { +static struct irq_chip msm_mdss_irq_chip = { .name = "dpu_mdss", - .irq_mask = dpu_mdss_irq_mask, - .irq_unmask = dpu_mdss_irq_unmask, + .irq_mask = msm_mdss_irq_mask, + .irq_unmask = msm_mdss_irq_unmask, }; -static struct lock_class_key dpu_mdss_lock_key, dpu_mdss_request_key; +static struct lock_class_key msm_mdss_lock_key, msm_mdss_request_key; -static int dpu_mdss_irqdomain_map(struct irq_domain *domain, +static int msm_mdss_irqdomain_map(struct irq_domain *domain, unsigned int irq, irq_hw_number_t hwirq) { struct dpu_mdss *dpu_mdss = domain->host_data; - irq_set_lockdep_class(irq, &dpu_mdss_lock_key, &dpu_mdss_request_key); - irq_set_chip_and_handler(irq, &dpu_mdss_irq_chip, handle_level_irq); + irq_set_lockdep_class(irq, &msm_mdss_lock_key, &msm_mdss_request_key); + irq_set_chip_and_handler(irq, &msm_mdss_irq_chip, handle_level_irq); + return irq_set_chip_data(irq, dpu_mdss); } -static const struct irq_domain_ops dpu_mdss_irqdomain_ops = { - .map = dpu_mdss_irqdomain_map, +static const struct irq_domain_ops msm_mdss_irqdomain_ops = { + .map = msm_mdss_irqdomain_map, .xlate = irq_domain_xlate_onecell, }; -static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) +static int _msm_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) { struct device *dev; struct irq_domain *domain; @@ -115,9 +116,9 @@ static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) dev = dpu_mdss->base.dev; domain = irq_domain_add_linear(dev->of_node, 32, - &dpu_mdss_irqdomain_ops, dpu_mdss); + &msm_mdss_irqdomain_ops, dpu_mdss); if (!domain) { - DPU_ERROR("failed to add irq_domain\n"); + DRM_ERROR("failed to add irq_domain\n"); return -EINVAL; } @@ -127,21 +128,14 @@ static int _dpu_mdss_irq_domain_add(struct dpu_mdss *dpu_mdss) return 0; } -static void _dpu_mdss_irq_domain_fini(struct dpu_mdss *dpu_mdss) -{ - if (dpu_mdss->irq_controller.domain) { - irq_domain_remove(dpu_mdss->irq_controller.domain); - dpu_mdss->irq_controller.domain = NULL; - } -} -static int dpu_mdss_enable(struct msm_mdss *mdss) +static int msm_mdss_enable(struct msm_mdss *mdss) { struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); int ret; ret = clk_bulk_prepare_enable(dpu_mdss->num_clocks, dpu_mdss->clocks); if (ret) { - DPU_ERROR("clock enable failed, ret:%d\n", ret); + DRM_ERROR("clock enable failed, ret:%d\n", ret); return ret; } @@ -171,7 +165,7 @@ static int dpu_mdss_enable(struct msm_mdss *mdss) return ret; } -static int dpu_mdss_disable(struct msm_mdss *mdss) +static int msm_mdss_disable(struct msm_mdss *mdss) { struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); @@ -180,7 +174,7 @@ static int dpu_mdss_disable(struct msm_mdss *mdss) return 0; } -static void dpu_mdss_destroy(struct msm_mdss *mdss) +static void msm_mdss_destroy(struct msm_mdss *mdss) { struct platform_device *pdev = to_platform_device(mdss->dev); struct dpu_mdss *dpu_mdss = to_dpu_mdss(mdss); @@ -188,22 +182,49 @@ static void dpu_mdss_destroy(struct msm_mdss *mdss) pm_runtime_suspend(mdss->dev); pm_runtime_disable(mdss->dev); - _dpu_mdss_irq_domain_fini(dpu_mdss); + irq_domain_remove(dpu_mdss->irq_controller.domain); + dpu_mdss->irq_controller.domain = NULL; irq = platform_get_irq(pdev, 0); irq_set_chained_handler_and_data(irq, NULL, NULL); - - if (dpu_mdss->mmio) - devm_iounmap(&pdev->dev, dpu_mdss->mmio); - dpu_mdss->mmio = NULL; } static const struct msm_mdss_funcs mdss_funcs = { - .enable = dpu_mdss_enable, - .disable = dpu_mdss_disable, - .destroy = dpu_mdss_destroy, + .enable = msm_mdss_enable, + .disable = msm_mdss_disable, + .destroy = msm_mdss_destroy, }; -int dpu_mdss_init(struct platform_device *pdev) +/* + * MDP5 MDSS uses at most three specified clocks. + */ +#define MDP5_MDSS_NUM_CLOCKS 3 +static int mdp5_mdss_parse_clock(struct platform_device *pdev, struct clk_bulk_data **clocks) +{ + struct clk_bulk_data *bulk; + int num_clocks = 0; + int ret; + + if (!pdev) + return -EINVAL; + + bulk = devm_kcalloc(&pdev->dev, MDP5_MDSS_NUM_CLOCKS, sizeof(struct clk_bulk_data), GFP_KERNEL); + if (!bulk) + return -ENOMEM; + + bulk[num_clocks++].id = "iface"; + bulk[num_clocks++].id = "bus"; + bulk[num_clocks++].id = "vsync"; + + ret = devm_clk_bulk_get_optional(&pdev->dev, num_clocks, bulk); + if (ret) + return ret; + + *clocks = bulk; + + return num_clocks; +} + +int msm_mdss_init(struct platform_device *pdev, bool is_mdp5) { struct msm_drm_private *priv = platform_get_drvdata(pdev); struct dpu_mdss *dpu_mdss; @@ -220,27 +241,28 @@ int dpu_mdss_init(struct platform_device *pdev) DRM_DEBUG("mapped mdss address space @%pK\n", dpu_mdss->mmio); - ret = devm_clk_bulk_get_all(&pdev->dev, &dpu_mdss->clocks); + if (is_mdp5) + ret = mdp5_mdss_parse_clock(pdev, &dpu_mdss->clocks); + else + ret = devm_clk_bulk_get_all(&pdev->dev, &dpu_mdss->clocks); if (ret < 0) { - DPU_ERROR("failed to parse clocks, ret=%d\n", ret); - goto clk_parse_err; + DRM_ERROR("failed to parse clocks, ret=%d\n", ret); + return ret; } dpu_mdss->num_clocks = ret; dpu_mdss->base.dev = &pdev->dev; dpu_mdss->base.funcs = &mdss_funcs; - ret = _dpu_mdss_irq_domain_add(dpu_mdss); - if (ret) - goto irq_domain_error; - irq = platform_get_irq(pdev, 0); - if (irq < 0) { - ret = irq; - goto irq_error; - } + if (irq < 0) + return irq; - irq_set_chained_handler_and_data(irq, dpu_mdss_irq, + ret = _msm_mdss_irq_domain_add(dpu_mdss); + if (ret) + return ret; + + irq_set_chained_handler_and_data(irq, msm_mdss_irq, dpu_mdss); priv->mdss = &dpu_mdss->base; @@ -248,13 +270,4 @@ int dpu_mdss_init(struct platform_device *pdev) pm_runtime_enable(&pdev->dev); return 0; - -irq_error: - _dpu_mdss_irq_domain_fini(dpu_mdss); -irq_domain_error: -clk_parse_err: - if (dpu_mdss->mmio) - devm_iounmap(&pdev->dev, dpu_mdss->mmio); - dpu_mdss->mmio = NULL; - return ret; }
MDP5 and DPU1 both provide the driver handling the MDSS region, which handles the irq domain and (incase of DPU1) adds some init for the UBWC controller. Unify those two pieces of code into a common driver. Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> --- drivers/gpu/drm/msm/Makefile | 3 +- drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c | 252 ------------------ drivers/gpu/drm/msm/msm_drv.c | 4 +- drivers/gpu/drm/msm/msm_kms.h | 3 +- .../msm/{disp/dpu1/dpu_mdss.c => msm_mdss.c} | 145 +++++----- 5 files changed, 83 insertions(+), 324 deletions(-) delete mode 100644 drivers/gpu/drm/msm/disp/mdp5/mdp5_mdss.c rename drivers/gpu/drm/msm/{disp/dpu1/dpu_mdss.c => msm_mdss.c} (63%)