Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 | /* SPDX-License-Identifier: GPL-2.0 */ /* Copyright (c) 2017 The Linux Foundation. All rights reserved. */ #ifndef _A6XX_GMU_H_ #define _A6XX_GMU_H_ #include <linux/completion.h> #include <linux/iopoll.h> #include <linux/interrupt.h> #include <linux/notifier.h> #include "msm_drv.h" #include "a6xx_hfi.h" struct a6xx_gmu_bo { struct drm_gem_object *obj; void *virt; size_t size; u64 iova; }; /* * These define the different GMU wake up options - these define how both the * CPU and the GMU bring up the hardware */ /* THe GMU has already been booted and the rentention registers are active */ #define GMU_WARM_BOOT 0 /* the GMU is coming up for the first time or back from a power collapse */ #define GMU_COLD_BOOT 1 /* * These define the level of control that the GMU has - the higher the number * the more things that the GMU hardware controls on its own. */ /* The GMU does not do any idle state management */ #define GMU_IDLE_STATE_ACTIVE 0 /* The GMU manages SPTP power collapse */ #define GMU_IDLE_STATE_SPTP 2 /* The GMU does automatic IFPC (intra-frame power collapse) */ #define GMU_IDLE_STATE_IFPC 3 struct a6xx_gmu { struct device *dev; /* For serializing communication with the GMU: */ struct mutex lock; struct msm_gem_address_space *aspace; void __iomem *mmio; void __iomem *rscc; int hfi_irq; int gmu_irq; struct device *gxpd; struct device *cxpd; int idle_level; struct a6xx_gmu_bo hfi; struct a6xx_gmu_bo debug; struct a6xx_gmu_bo icache; struct a6xx_gmu_bo dcache; struct a6xx_gmu_bo dummy; struct a6xx_gmu_bo log; int nr_clocks; struct clk_bulk_data *clocks; struct clk *core_clk; struct clk *hub_clk; /* current performance index set externally */ int current_perf_index; int nr_gpu_freqs; unsigned long gpu_freqs[16]; u32 gx_arc_votes[16]; int nr_gmu_freqs; unsigned long gmu_freqs[4]; u32 cx_arc_votes[4]; unsigned long freq; struct a6xx_hfi_queue queues[2]; bool initialized; bool hung; bool legacy; /* a618 or a630 */ /* For power domain callback */ struct notifier_block pd_nb; struct completion pd_gate; }; static inline u32 gmu_read(struct a6xx_gmu *gmu, u32 offset) { return msm_readl(gmu->mmio + (offset << 2)); } static inline void gmu_write(struct a6xx_gmu *gmu, u32 offset, u32 value) { msm_writel(value, gmu->mmio + (offset << 2)); } static inline void gmu_write_bulk(struct a6xx_gmu *gmu, u32 offset, const u32 *data, u32 size) { memcpy_toio(gmu->mmio + (offset << 2), data, size); wmb(); } static inline void gmu_rmw(struct a6xx_gmu *gmu, u32 reg, u32 mask, u32 or) { u32 val = gmu_read(gmu, reg); val &= ~mask; gmu_write(gmu, reg, val | or); } static inline u64 gmu_read64(struct a6xx_gmu *gmu, u32 lo, u32 hi) { u64 val; val = (u64) msm_readl(gmu->mmio + (lo << 2)); val |= ((u64) msm_readl(gmu->mmio + (hi << 2)) << 32); return val; } #define gmu_poll_timeout(gmu, addr, val, cond, interval, timeout) \ readl_poll_timeout((gmu)->mmio + ((addr) << 2), val, cond, \ interval, timeout) static inline u32 gmu_read_rscc(struct a6xx_gmu *gmu, u32 offset) { return msm_readl(gmu->rscc + (offset << 2)); } static inline void gmu_write_rscc(struct a6xx_gmu *gmu, u32 offset, u32 value) { msm_writel(value, gmu->rscc + (offset << 2)); } #define gmu_poll_timeout_rscc(gmu, addr, val, cond, interval, timeout) \ readl_poll_timeout((gmu)->rscc + ((addr) << 2), val, cond, \ interval, timeout) /* * These are the available OOB (out of band requests) to the GMU where "out of * band" means that the CPU talks to the GMU directly and not through HFI. * Normally this works by writing a ITCM/DTCM register and then triggering a * interrupt (the "request" bit) and waiting for an acknowledgment (the "ack" * bit). The state is cleared by writing the "clear' bit to the GMU interrupt. * * These are used to force the GMU/GPU to stay on during a critical sequence or * for hardware workarounds. */ enum a6xx_gmu_oob_state { /* * Let the GMU know that a boot or slumber operation has started. The value in * REG_A6XX_GMU_BOOT_SLUMBER_OPTION lets the GMU know which operation we are * doing */ GMU_OOB_BOOT_SLUMBER = 0, /* * Let the GMU know to not turn off any GPU registers while the CPU is in a * critical section */ GMU_OOB_GPU_SET, /* * Set a new power level for the GPU when the CPU is doing frequency scaling */ GMU_OOB_DCVS_SET, /* * Used to keep the GPU on for CPU-side reads of performance counters. */ GMU_OOB_PERFCOUNTER_SET, }; void a6xx_hfi_init(struct a6xx_gmu *gmu); int a6xx_hfi_start(struct a6xx_gmu *gmu, int boot_state); void a6xx_hfi_stop(struct a6xx_gmu *gmu); int a6xx_hfi_send_prep_slumber(struct a6xx_gmu *gmu); int a6xx_hfi_set_freq(struct a6xx_gmu *gmu, int index); bool a6xx_gmu_gx_is_on(struct a6xx_gmu *gmu); bool a6xx_gmu_sptprac_is_on(struct a6xx_gmu *gmu); void a6xx_sptprac_disable(struct a6xx_gmu *gmu); int a6xx_sptprac_enable(struct a6xx_gmu *gmu); #endif |