From: Jie Zhang <[email protected]>

Add support for Adreno 612 GPU found in SM6150/QCS615 chipsets.
A612 falls under ADRENO_6XX_GEN1 family and is a cut down version
of A615 GPU.

A612 has a new IP called Reduced Graphics Management Unit or RGMU
which is a small state machine which helps to toggle GX GDSC
(connected to CX rail) to implement IFPC feature. It doesn't support
any other features of a full fledged GMU like clock control, resource
voting to rpmh etc. So we need linux clock driver support like other
gmu-wrapper implementations to control gpu core clock and gpu GX gdsc.
This patch skips RGMU core initialization and act more like a
gmu-wrapper case.

Co-developed-by: Akhil P Oommen <[email protected]>
Signed-off-by: Jie Zhang <[email protected]>
Signed-off-by: Akhil P Oommen <[email protected]>
---
 drivers/gpu/drm/msm/adreno/a6xx_catalog.c   |  16 ++++
 drivers/gpu/drm/msm/adreno/a6xx_gmu.c       |  87 ++++++++++++++++-
 drivers/gpu/drm/msm/adreno/a6xx_gpu.c       | 143 ++++++++++++++++++++++++++--
 drivers/gpu/drm/msm/adreno/a6xx_gpu.h       |   1 +
 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c |   3 +-
 drivers/gpu/drm/msm/adreno/adreno_gpu.c     |   1 +
 drivers/gpu/drm/msm/adreno/adreno_gpu.h     |  16 +++-
 7 files changed, 256 insertions(+), 11 deletions(-)

diff --git a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c 
b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
index 
44df6410bce17613702d7d04906469de4dd021b5..5db01fa2ed441d3fc33b27f7c9868c591e08604e
 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_catalog.c
@@ -705,6 +705,22 @@ static const struct adreno_info a6xx_gpus[] = {
                        { 157, 3 },
                        { 127, 4 },
                ),
+       }, {
+               .chip_ids = ADRENO_CHIP_IDS(0x06010200),
+               .family = ADRENO_6XX_GEN1,
+               .fw = {
+                       [ADRENO_FW_SQE] = "a630_sqe.fw",
+                       [ADRENO_FW_GMU] = "a612_rgmu.bin",
+               },
+               .gmem = (SZ_128K + SZ_4K),
+               .inactive_period = DRM_MSM_INACTIVE_PERIOD,
+               .init = a6xx_gpu_init,
+               .a6xx = &(const struct a6xx_info) {
+                       .hwcg = a612_hwcg,
+                       .protect = &a630_protect,
+                       .gmu_cgc_mode = 0x00000022,
+                       .prim_fifo_threshold = 0x00080000,
+               },
        }, {
                .chip_ids = ADRENO_CHIP_IDS(0x06010500),
                .family = ADRENO_6XX_GEN1,
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c 
b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
index 
fc62fef2fed87f065cb8fa4e997abefe4ff11cd5..9a11171ec8ad8252d9c7b89d752d57b53651456e
 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
@@ -350,12 +350,18 @@ static const struct a6xx_gmu_oob_bits a6xx_gmu_oob_bits[] 
= {
 /* Trigger a OOB (out of band) request to the GMU */
 int a6xx_gmu_set_oob(struct a6xx_gmu *gmu, enum a6xx_gmu_oob_state state)
 {
+       struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
+       struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
        int ret;
        u32 val;
        int request, ack;
 
        WARN_ON_ONCE(!mutex_is_locked(&gmu->lock));
 
+       /* Skip OOB calls since RGMU is not enabled */
+       if (adreno_has_rgmu(adreno_gpu))
+               return 0;
+
        if (state >= ARRAY_SIZE(a6xx_gmu_oob_bits))
                return -EINVAL;
 
@@ -395,10 +401,16 @@ int a6xx_gmu_set_oob(struct a6xx_gmu *gmu, enum 
a6xx_gmu_oob_state state)
 /* Clear a pending OOB state in the GMU */
 void a6xx_gmu_clear_oob(struct a6xx_gmu *gmu, enum a6xx_gmu_oob_state state)
 {
+       struct a6xx_gpu *a6xx_gpu = container_of(gmu, struct a6xx_gpu, gmu);
+       struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
        int bit;
 
        WARN_ON_ONCE(!mutex_is_locked(&gmu->lock));
 
+       /* Skip OOB calls since RGMU is not enabled */
+       if (adreno_has_rgmu(adreno_gpu))
+               return;
+
        if (state >= ARRAY_SIZE(a6xx_gmu_oob_bits))
                return;
 
@@ -1897,7 +1909,8 @@ void a6xx_gmu_remove(struct a6xx_gpu *a6xx_gpu)
        gmu->mmio = NULL;
        gmu->rscc = NULL;
 
-       if (!adreno_has_gmu_wrapper(adreno_gpu)) {
+       if (!adreno_has_gmu_wrapper(adreno_gpu) &&
+           !adreno_has_rgmu(adreno_gpu)) {
                a6xx_gmu_memory_free(gmu);
 
                free_irq(gmu->gmu_irq, gmu);
@@ -1984,6 +1997,78 @@ int a6xx_gmu_wrapper_init(struct a6xx_gpu *a6xx_gpu, 
struct device_node *node)
        return ret;
 }
 
+int a6xx_rgmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node)
+{
+       struct platform_device *pdev = of_find_device_by_node(node);
+       struct a6xx_gmu *gmu = &a6xx_gpu->gmu;
+       int ret;
+
+       if (!pdev)
+               return -ENODEV;
+
+       gmu->dev = &pdev->dev;
+
+       ret = of_dma_configure(gmu->dev, node, true);
+       if (ret)
+               return ret;
+
+       pm_runtime_enable(gmu->dev);
+
+       /* Mark legacy for manual SPTPRAC control */
+       gmu->legacy = true;
+
+       /* RGMU requires clocks */
+       ret = devm_clk_bulk_get_all(gmu->dev, &gmu->clocks);
+       if (ret < 1)
+               return ret;
+
+       gmu->nr_clocks = ret;
+
+       /* Map the GMU registers */
+       gmu->mmio = a6xx_gmu_get_mmio(pdev, "gmu");
+       if (IS_ERR(gmu->mmio)) {
+               ret = PTR_ERR(gmu->mmio);
+               goto err_mmio;
+       }
+
+       gmu->cxpd = dev_pm_domain_attach_by_name(gmu->dev, "cx");
+       if (IS_ERR(gmu->cxpd)) {
+               ret = PTR_ERR(gmu->cxpd);
+               goto err_mmio;
+       }
+
+       if (!device_link_add(gmu->dev, gmu->cxpd, DL_FLAG_PM_RUNTIME)) {
+               ret = -ENODEV;
+               goto detach_cxpd;
+       }
+
+       init_completion(&gmu->pd_gate);
+       complete_all(&gmu->pd_gate);
+       gmu->pd_nb.notifier_call = cxpd_notifier_cb;
+
+       /* Get a link to the GX power domain to reset the GPU */
+       gmu->gxpd = dev_pm_domain_attach_by_name(gmu->dev, "gx");
+       if (IS_ERR(gmu->gxpd)) {
+               ret = PTR_ERR(gmu->gxpd);
+               goto err_mmio;
+       }
+
+       gmu->initialized = true;
+
+       return 0;
+
+detach_cxpd:
+       dev_pm_domain_detach(gmu->cxpd, false);
+
+err_mmio:
+       iounmap(gmu->mmio);
+
+       /* Drop reference taken in of_find_device_by_node */
+       put_device(gmu->dev);
+
+       return ret;
+}
+
 int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node)
 {
        struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c 
b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
index 
b8f8ae940b55f5578abdbdec6bf1e90a53e721a5..b0494f1e331cfec3141922a0f0c6596934acf63e
 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
@@ -612,15 +612,26 @@ static void a6xx_set_hwcg(struct msm_gpu *gpu, bool state)
 
        if (adreno_is_a630(adreno_gpu))
                clock_cntl_on = 0x8aa8aa02;
-       else if (adreno_is_a610(adreno_gpu))
+       else if (adreno_is_a610(adreno_gpu) || adreno_is_a612(adreno_gpu))
                clock_cntl_on = 0xaaa8aa82;
        else if (adreno_is_a702(adreno_gpu))
                clock_cntl_on = 0xaaaaaa82;
        else
                clock_cntl_on = 0x8aa8aa82;
 
-       cgc_delay = adreno_is_a615_family(adreno_gpu) ? 0x111 : 0x10111;
-       cgc_hyst = adreno_is_a615_family(adreno_gpu) ? 0x555 : 0x5555;
+       if (adreno_is_a612(adreno_gpu))
+               cgc_delay = 0x11;
+       else if (adreno_is_a615_family(adreno_gpu))
+               cgc_delay = 0x111;
+       else
+               cgc_delay = 0x10111;
+
+       if (adreno_is_a612(adreno_gpu))
+               cgc_hyst = 0x55;
+       else if (adreno_is_a615_family(adreno_gpu))
+               cgc_delay = 0x555;
+       else
+               cgc_delay = 0x5555;
 
        gmu_write(&a6xx_gpu->gmu, REG_A6XX_GPU_GMU_AO_GMU_CGC_MODE_CNTL,
                        state ? adreno_gpu->info->a6xx->gmu_cgc_mode : 0);
@@ -714,6 +725,9 @@ static int a6xx_calc_ubwc_config(struct adreno_gpu *gpu)
                cfg->ubwc_swizzle = 0x7;
        }
 
+       if (adreno_is_a612(gpu))
+               cfg->highest_bank_bit = 14;
+
        if (adreno_is_a618(gpu))
                cfg->highest_bank_bit = 14;
 
@@ -1288,7 +1302,7 @@ static int hw_init(struct msm_gpu *gpu)
                gpu_write(gpu, REG_A6XX_CP_LPAC_PROG_FIFO_SIZE, 0x00000020);
 
        /* Setting the mem pool size */
-       if (adreno_is_a610(adreno_gpu)) {
+       if (adreno_is_a610(adreno_gpu) || adreno_is_a612(adreno_gpu)) {
                gpu_write(gpu, REG_A6XX_CP_MEM_POOL_SIZE, 48);
                gpu_write(gpu, REG_A6XX_CP_MEM_POOL_DBG_ADDR, 47);
        } else if (adreno_is_a702(adreno_gpu)) {
@@ -1322,7 +1336,7 @@ static int hw_init(struct msm_gpu *gpu)
 
        /* Enable fault detection */
        if (adreno_is_a730(adreno_gpu) ||
-           adreno_is_a740_family(adreno_gpu))
+           adreno_is_a740_family(adreno_gpu) || adreno_is_a612(adreno_gpu))
                gpu_write(gpu, REG_A6XX_RBBM_INTERFACE_HANG_INT_CNTL, (1 << 30) 
| 0xcfffff);
        else if (adreno_is_a690(adreno_gpu))
                gpu_write(gpu, REG_A6XX_RBBM_INTERFACE_HANG_INT_CNTL, (1 << 30) 
| 0x4fffff);
@@ -1576,7 +1590,7 @@ static void a6xx_recover(struct msm_gpu *gpu)
         */
        gpu->active_submits = 0;
 
-       if (adreno_has_gmu_wrapper(adreno_gpu)) {
+       if (adreno_has_gmu_wrapper(adreno_gpu) || adreno_has_rgmu(adreno_gpu)) {
                /* Drain the outstanding traffic on memory buses */
                a6xx_bus_clear_pending_transactions(adreno_gpu, true);
 
@@ -2248,6 +2262,55 @@ static int a6xx_pm_resume(struct msm_gpu *gpu)
        return ret;
 }
 
+static int a6xx_rgmu_pm_resume(struct msm_gpu *gpu)
+{
+       struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
+       struct a6xx_gpu *a6xx_gpu = to_a6xx_gpu(adreno_gpu);
+       struct a6xx_gmu *gmu = &a6xx_gpu->gmu;
+       unsigned long freq = gpu->fast_rate;
+       struct dev_pm_opp *opp;
+       int ret;
+
+       gpu->needs_hw_init = true;
+
+       trace_msm_gpu_resume(0);
+
+       opp = dev_pm_opp_find_freq_ceil(&gpu->pdev->dev, &freq);
+       if (IS_ERR(opp))
+               return PTR_ERR(opp);
+
+       dev_pm_opp_put(opp);
+
+       /* Set the core clock and bus bw, having VDD scaling in mind */
+       dev_pm_opp_set_opp(&gpu->pdev->dev, opp);
+
+       pm_runtime_resume_and_get(gmu->dev);
+       pm_runtime_resume_and_get(gmu->gxpd);
+
+       ret = clk_bulk_prepare_enable(gmu->nr_clocks, gmu->clocks);
+       if (ret)
+               goto err_rpm_put;
+
+       ret = clk_bulk_prepare_enable(gpu->nr_clocks, gpu->grp_clks);
+       if (ret)
+               goto err_bulk_clk;
+
+       msm_devfreq_resume(gpu);
+
+       a6xx_llc_activate(a6xx_gpu);
+
+       return 0;
+
+err_bulk_clk:
+       clk_bulk_disable_unprepare(gmu->nr_clocks, gmu->clocks);
+err_rpm_put:
+       pm_runtime_put(gmu->gxpd);
+       pm_runtime_put(gmu->dev);
+       dev_pm_opp_set_opp(&gpu->pdev->dev, NULL);
+
+       return ret;
+}
+
 static int a6xx_gmu_pm_suspend(struct msm_gpu *gpu)
 {
        struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
@@ -2311,6 +2374,41 @@ static int a6xx_pm_suspend(struct msm_gpu *gpu)
        return 0;
 }
 
+static int a6xx_rgmu_pm_suspend(struct msm_gpu *gpu)
+{
+       struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
+       struct a6xx_gpu *a6xx_gpu = to_a6xx_gpu(adreno_gpu);
+       struct a6xx_gmu *gmu = &a6xx_gpu->gmu;
+
+       trace_msm_gpu_suspend(0);
+
+       a6xx_llc_deactivate(a6xx_gpu);
+
+       msm_devfreq_suspend(gpu);
+
+       mutex_lock(&a6xx_gpu->gmu.lock);
+
+       /* Drain the outstanding traffic on memory buses */
+       a6xx_bus_clear_pending_transactions(adreno_gpu, true);
+
+       clk_bulk_disable_unprepare(gpu->nr_clocks, gpu->grp_clks);
+       clk_bulk_disable_unprepare(gmu->nr_clocks, gmu->clocks);
+
+       pm_runtime_put_sync(gmu->gxpd);
+       dev_pm_opp_set_opp(&gpu->pdev->dev, NULL);
+       pm_runtime_put_sync(gmu->dev);
+
+       mutex_unlock(&a6xx_gpu->gmu.lock);
+
+       if (a6xx_gpu->shadow_bo)
+               for (int i = 0; i < gpu->nr_rings; i++)
+                       a6xx_gpu->shadow[i] = 0;
+
+       gpu->suspend_count++;
+
+       return 0;
+}
+
 static int a6xx_gmu_get_timestamp(struct msm_gpu *gpu, uint64_t *value)
 {
        struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
@@ -2588,6 +2686,35 @@ static const struct adreno_gpu_funcs funcs_gmuwrapper = {
        .get_timestamp = a6xx_get_timestamp,
 };
 
+static const struct adreno_gpu_funcs funcs_rgmu = {
+       .base = {
+               .get_param = adreno_get_param,
+               .set_param = adreno_set_param,
+               .hw_init = a6xx_hw_init,
+               .ucode_load = a6xx_ucode_load,
+               .pm_suspend = a6xx_rgmu_pm_suspend,
+               .pm_resume = a6xx_rgmu_pm_resume,
+               .recover = a6xx_recover,
+               .submit = a6xx_submit,
+               .active_ring = a6xx_active_ring,
+               .irq = a6xx_irq,
+               .destroy = a6xx_destroy,
+#if defined(CONFIG_DRM_MSM_GPU_STATE)
+               .show = a6xx_show,
+#endif
+               .gpu_busy = a6xx_gpu_busy,
+#if defined(CONFIG_DRM_MSM_GPU_STATE)
+               .gpu_state_get = a6xx_gpu_state_get,
+               .gpu_state_put = a6xx_gpu_state_put,
+#endif
+               .create_vm = a6xx_create_vm,
+               .create_private_vm = a6xx_create_private_vm,
+               .get_rptr = a6xx_get_rptr,
+               .progress = a6xx_progress,
+       },
+       .get_timestamp = a6xx_gmu_get_timestamp,
+};
+
 static const struct adreno_gpu_funcs funcs_a7xx = {
        .base = {
                .get_param = adreno_get_param,
@@ -2675,6 +2802,8 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
                ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_a7xx, 1);
        else if (adreno_has_gmu_wrapper(adreno_gpu))
                ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_gmuwrapper, 
1);
+       else if (of_device_is_compatible(node, "qcom,adreno-rgmu"))
+               ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs_rgmu, 1);
        else
                ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs, 1);
        if (ret) {
@@ -2691,6 +2820,8 @@ struct msm_gpu *a6xx_gpu_init(struct drm_device *dev)
 
        if (adreno_has_gmu_wrapper(adreno_gpu))
                ret = a6xx_gmu_wrapper_init(a6xx_gpu, node);
+       else if (of_device_is_compatible(node, "qcom,adreno-rgmu"))
+               ret = a6xx_rgmu_init(a6xx_gpu, node);
        else
                ret = a6xx_gmu_init(a6xx_gpu, node);
        of_node_put(node);
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h 
b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
index 
0b17d36c36a9567e6afa4269ae7783ed3578e40e..073e5766198d0ea96c71df34963a354961f21b95
 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.h
@@ -254,6 +254,7 @@ void a6xx_gmu_clear_oob(struct a6xx_gmu *gmu, enum 
a6xx_gmu_oob_state state);
 
 int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node);
 int a6xx_gmu_wrapper_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node);
+int a6xx_rgmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node);
 void a6xx_gmu_remove(struct a6xx_gpu *a6xx_gpu);
 void a6xx_gmu_sysprof_setup(struct msm_gpu *gpu);
 
diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c 
b/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
index 
4c7f3c642f6ac0a397ac3b282966d25cd4488d3e..838150ff49ab1877da97eecc63a5bc1ea5f1edfe
 100644
--- a/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
+++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
@@ -1596,7 +1596,8 @@ struct msm_gpu_state *a6xx_gpu_state_get(struct msm_gpu 
*gpu)
        /* Get the generic state from the adreno core */
        adreno_gpu_state_get(gpu, &a6xx_state->base);
 
-       if (!adreno_has_gmu_wrapper(adreno_gpu)) {
+       if (!adreno_has_gmu_wrapper(adreno_gpu) &&
+           !adreno_has_rgmu(adreno_gpu)) {
                a6xx_get_gmu_registers(gpu, a6xx_state);
 
                a6xx_state->gmu_log = a6xx_snapshot_gmu_bo(a6xx_state, 
&a6xx_gpu->gmu.log);
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c 
b/drivers/gpu/drm/msm/adreno/adreno_gpu.c
index 
afaa3cfefd357dc0230994c8b5830a14c6d7a352..60973c3ec021bcd15b7a180c93c05b171b21ebe0
 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c
+++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c
@@ -1196,6 +1196,7 @@ int adreno_gpu_init(struct drm_device *drm, struct 
platform_device *pdev,
 
        /* Only handle the core clock when GMU is not in use (or is absent). */
        if (adreno_has_gmu_wrapper(adreno_gpu) ||
+           adreno_has_rgmu(adreno_gpu) ||
            adreno_gpu->info->family < ADRENO_6XX_GEN1) {
                /*
                 * This can only be done before devm_pm_opp_of_add_table(), or
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h 
b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
index 
390fa6720d9b096f4fa7d1639645d453d43b153a..25ee6b277fe2db5950a057a92d33244d76de299c
 100644
--- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
+++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
@@ -392,6 +392,16 @@ static inline int adreno_is_a610(const struct adreno_gpu 
*gpu)
        return adreno_is_revn(gpu, 610);
 }
 
+static inline int adreno_is_a612(const struct adreno_gpu *gpu)
+{
+       return gpu->info->chip_ids[0] == 0x06010200;
+}
+
+static inline bool adreno_has_rgmu(const struct adreno_gpu *gpu)
+{
+       return adreno_is_a612(gpu);
+}
+
 static inline int adreno_is_a618(const struct adreno_gpu *gpu)
 {
        return adreno_is_revn(gpu, 618);
@@ -466,9 +476,9 @@ static inline int adreno_is_a610_family(const struct 
adreno_gpu *gpu)
 {
        if (WARN_ON_ONCE(!gpu->info))
                return false;
-
-       /* TODO: A612 */
-       return adreno_is_a610(gpu) || adreno_is_a702(gpu);
+       return adreno_is_a610(gpu) ||
+              adreno_is_a612(gpu) ||
+              adreno_is_a702(gpu);
 }
 
 /* TODO: 615/616 */

-- 
2.51.0

Reply via email to