diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2014-12-15 18:52:01 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2014-12-15 18:52:01 -0500 |
commit | 988adfdffdd43cfd841df734664727993076d7cb (patch) | |
tree | 6794f7bba8f595500c2b7d33376ad6614adcfaf2 /drivers/gpu/drm/msm | |
parent | 26178ec11ef3c6c814bf16a0a2b9c2f7242e3c64 (diff) | |
parent | 4e0cd68115620bc3236ff4e58e4c073948629b41 (diff) |
Merge branch 'drm-next' of git://people.freedesktop.org/~airlied/linux
Pull drm updates from Dave Airlie:
"Highlights:
- AMD KFD driver merge
This is the AMD HSA interface for exposing a lowlevel interface for
GPGPU use. They have an open source userspace built on top of this
interface, and the code looks as good as it was going to get out of
tree.
- Initial atomic modesetting work
The need for an atomic modesetting interface to allow userspace to
try and send a complete set of modesetting state to the driver has
arisen, and been suffering from neglect this past year. No more,
the start of the common code and changes for msm driver to use it
are in this tree. Ongoing work to get the userspace ioctl finished
and the code clean will probably wait until next kernel.
- DisplayID 1.3 and tiled monitor exposed to userspace.
Tiled monitor property is now exposed for userspace to make use of.
- Rockchip drm driver merged.
- imx gpu driver moved out of staging
Other stuff:
- core:
panel - MIPI DSI + new panels.
expose suggested x/y properties for virtual GPUs
- i915:
Initial Skylake (SKL) support
gen3/4 reset work
start of dri1/ums removal
infoframe tracking
fixes for lots of things.
- nouveau:
tegra k1 voltage support
GM204 modesetting support
GT21x memory reclocking work
- radeon:
CI dpm fixes
GPUVM improvements
Initial DPM fan control
- rcar-du:
HDMI support added
removed some support for old boards
slave encoder driver for Analog Devices adv7511
- exynos:
Exynos4415 SoC support
- msm:
a4xx gpu support
atomic helper conversion
- tegra:
iommu support
universal plane support
ganged-mode DSI support
- sti:
HDMI i2c improvements
- vmwgfx:
some late fixes.
- qxl:
use suggested x/y properties"
* 'drm-next' of git://people.freedesktop.org/~airlied/linux: (969 commits)
drm: sti: fix module compilation issue
drm/i915: save/restore GMBUS freq across suspend/resume on gen4
drm: sti: correctly cleanup CRTC and planes
drm: sti: add HQVDP plane
drm: sti: add cursor plane
drm: sti: enable auxiliary CRTC
drm: sti: fix delay in VTG programming
drm: sti: prepare sti_tvout to support auxiliary crtc
drm: sti: use drm_crtc_vblank_{on/off} instead of drm_vblank_{on/off}
drm: sti: fix hdmi avi infoframe
drm: sti: remove event lock while disabling vblank
drm: sti: simplify gdp code
drm: sti: clear all mixer control
drm: sti: remove gpio for HDMI hot plug detection
drm: sti: allow to change hdmi ddc i2c adapter
drm/doc: Document drm_add_modes_noedid() usage
drm/i915: Remove '& 0xffff' from the mask given to WA_REG()
drm/i915: Invert the mask and val arguments in wa_add() and WA_REG()
drm: Zero out DRM object memory upon cleanup
drm/i915/bdw: Fix the write setting up the WIZ hashing mode
...
Diffstat (limited to 'drivers/gpu/drm/msm')
50 files changed, 5578 insertions, 1230 deletions
diff --git a/drivers/gpu/drm/msm/Kconfig b/drivers/gpu/drm/msm/Kconfig index 9d907c526c94..5b2a1ff95d3d 100644 --- a/drivers/gpu/drm/msm/Kconfig +++ b/drivers/gpu/drm/msm/Kconfig | |||
@@ -3,6 +3,7 @@ config DRM_MSM | |||
3 | tristate "MSM DRM" | 3 | tristate "MSM DRM" |
4 | depends on DRM | 4 | depends on DRM |
5 | depends on ARCH_QCOM || (ARM && COMPILE_TEST) | 5 | depends on ARCH_QCOM || (ARM && COMPILE_TEST) |
6 | select REGULATOR | ||
6 | select DRM_KMS_HELPER | 7 | select DRM_KMS_HELPER |
7 | select DRM_PANEL | 8 | select DRM_PANEL |
8 | select SHMEM | 9 | select SHMEM |
diff --git a/drivers/gpu/drm/msm/Makefile b/drivers/gpu/drm/msm/Makefile index 6283dcb96af5..143d988f8add 100644 --- a/drivers/gpu/drm/msm/Makefile +++ b/drivers/gpu/drm/msm/Makefile | |||
@@ -7,6 +7,7 @@ msm-y := \ | |||
7 | adreno/adreno_device.o \ | 7 | adreno/adreno_device.o \ |
8 | adreno/adreno_gpu.o \ | 8 | adreno/adreno_gpu.o \ |
9 | adreno/a3xx_gpu.o \ | 9 | adreno/a3xx_gpu.o \ |
10 | adreno/a4xx_gpu.o \ | ||
10 | hdmi/hdmi.o \ | 11 | hdmi/hdmi.o \ |
11 | hdmi/hdmi_audio.o \ | 12 | hdmi/hdmi_audio.o \ |
12 | hdmi/hdmi_bridge.o \ | 13 | hdmi/hdmi_bridge.o \ |
@@ -24,12 +25,15 @@ msm-y := \ | |||
24 | mdp/mdp4/mdp4_irq.o \ | 25 | mdp/mdp4/mdp4_irq.o \ |
25 | mdp/mdp4/mdp4_kms.o \ | 26 | mdp/mdp4/mdp4_kms.o \ |
26 | mdp/mdp4/mdp4_plane.o \ | 27 | mdp/mdp4/mdp4_plane.o \ |
28 | mdp/mdp5/mdp5_cfg.o \ | ||
29 | mdp/mdp5/mdp5_ctl.o \ | ||
27 | mdp/mdp5/mdp5_crtc.o \ | 30 | mdp/mdp5/mdp5_crtc.o \ |
28 | mdp/mdp5/mdp5_encoder.o \ | 31 | mdp/mdp5/mdp5_encoder.o \ |
29 | mdp/mdp5/mdp5_irq.o \ | 32 | mdp/mdp5/mdp5_irq.o \ |
30 | mdp/mdp5/mdp5_kms.o \ | 33 | mdp/mdp5/mdp5_kms.o \ |
31 | mdp/mdp5/mdp5_plane.o \ | 34 | mdp/mdp5/mdp5_plane.o \ |
32 | mdp/mdp5/mdp5_smp.o \ | 35 | mdp/mdp5/mdp5_smp.o \ |
36 | msm_atomic.o \ | ||
33 | msm_drv.o \ | 37 | msm_drv.o \ |
34 | msm_fb.o \ | 38 | msm_fb.o \ |
35 | msm_gem.o \ | 39 | msm_gem.o \ |
diff --git a/drivers/gpu/drm/msm/adreno/a2xx.xml.h b/drivers/gpu/drm/msm/adreno/a2xx.xml.h index a3104598c27f..22882cc0a573 100644 --- a/drivers/gpu/drm/msm/adreno/a2xx.xml.h +++ b/drivers/gpu/drm/msm/adreno/a2xx.xml.h | |||
@@ -11,10 +11,10 @@ The rules-ng-ng source files this header was generated from are: | |||
11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 9859 bytes, from 2014-06-02 15:21:30) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 10551 bytes, from 2014-11-13 22:44:30) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 14960 bytes, from 2014-07-27 17:22:13) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 15053 bytes, from 2014-11-09 15:45:47) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 58020 bytes, from 2014-08-01 12:22:48) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 63169 bytes, from 2014-11-13 22:44:18) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 41068 bytes, from 2014-08-01 12:22:48) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 49097 bytes, from 2014-11-14 15:38:00) |
18 | 18 | ||
19 | Copyright (C) 2013-2014 by the following authors: | 19 | Copyright (C) 2013-2014 by the following authors: |
20 | - Rob Clark <robdclark@gmail.com> (robclark) | 20 | - Rob Clark <robdclark@gmail.com> (robclark) |
@@ -926,11 +926,11 @@ static inline uint32_t A2XX_VGT_DRAW_INITIATOR_INDEX_SIZE(enum pc_di_index_size | |||
926 | #define A2XX_VGT_DRAW_INITIATOR_NOT_EOP 0x00001000 | 926 | #define A2XX_VGT_DRAW_INITIATOR_NOT_EOP 0x00001000 |
927 | #define A2XX_VGT_DRAW_INITIATOR_SMALL_INDEX 0x00002000 | 927 | #define A2XX_VGT_DRAW_INITIATOR_SMALL_INDEX 0x00002000 |
928 | #define A2XX_VGT_DRAW_INITIATOR_PRE_DRAW_INITIATOR_ENABLE 0x00004000 | 928 | #define A2XX_VGT_DRAW_INITIATOR_PRE_DRAW_INITIATOR_ENABLE 0x00004000 |
929 | #define A2XX_VGT_DRAW_INITIATOR_NUM_INDICES__MASK 0xffff0000 | 929 | #define A2XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__MASK 0xff000000 |
930 | #define A2XX_VGT_DRAW_INITIATOR_NUM_INDICES__SHIFT 16 | 930 | #define A2XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__SHIFT 24 |
931 | static inline uint32_t A2XX_VGT_DRAW_INITIATOR_NUM_INDICES(uint32_t val) | 931 | static inline uint32_t A2XX_VGT_DRAW_INITIATOR_NUM_INSTANCES(uint32_t val) |
932 | { | 932 | { |
933 | return ((val) << A2XX_VGT_DRAW_INITIATOR_NUM_INDICES__SHIFT) & A2XX_VGT_DRAW_INITIATOR_NUM_INDICES__MASK; | 933 | return ((val) << A2XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__SHIFT) & A2XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__MASK; |
934 | } | 934 | } |
935 | 935 | ||
936 | #define REG_A2XX_VGT_IMMED_DATA 0x000021fd | 936 | #define REG_A2XX_VGT_IMMED_DATA 0x000021fd |
@@ -1243,13 +1243,13 @@ static inline uint32_t A2XX_CLEAR_COLOR_ALPHA(uint32_t val) | |||
1243 | #define A2XX_PA_SU_POINT_SIZE_HEIGHT__SHIFT 0 | 1243 | #define A2XX_PA_SU_POINT_SIZE_HEIGHT__SHIFT 0 |
1244 | static inline uint32_t A2XX_PA_SU_POINT_SIZE_HEIGHT(float val) | 1244 | static inline uint32_t A2XX_PA_SU_POINT_SIZE_HEIGHT(float val) |
1245 | { | 1245 | { |
1246 | return ((((uint32_t)(val * 8.0))) << A2XX_PA_SU_POINT_SIZE_HEIGHT__SHIFT) & A2XX_PA_SU_POINT_SIZE_HEIGHT__MASK; | 1246 | return ((((uint32_t)(val * 16.0))) << A2XX_PA_SU_POINT_SIZE_HEIGHT__SHIFT) & A2XX_PA_SU_POINT_SIZE_HEIGHT__MASK; |
1247 | } | 1247 | } |
1248 | #define A2XX_PA_SU_POINT_SIZE_WIDTH__MASK 0xffff0000 | 1248 | #define A2XX_PA_SU_POINT_SIZE_WIDTH__MASK 0xffff0000 |
1249 | #define A2XX_PA_SU_POINT_SIZE_WIDTH__SHIFT 16 | 1249 | #define A2XX_PA_SU_POINT_SIZE_WIDTH__SHIFT 16 |
1250 | static inline uint32_t A2XX_PA_SU_POINT_SIZE_WIDTH(float val) | 1250 | static inline uint32_t A2XX_PA_SU_POINT_SIZE_WIDTH(float val) |
1251 | { | 1251 | { |
1252 | return ((((uint32_t)(val * 8.0))) << A2XX_PA_SU_POINT_SIZE_WIDTH__SHIFT) & A2XX_PA_SU_POINT_SIZE_WIDTH__MASK; | 1252 | return ((((uint32_t)(val * 16.0))) << A2XX_PA_SU_POINT_SIZE_WIDTH__SHIFT) & A2XX_PA_SU_POINT_SIZE_WIDTH__MASK; |
1253 | } | 1253 | } |
1254 | 1254 | ||
1255 | #define REG_A2XX_PA_SU_POINT_MINMAX 0x00002281 | 1255 | #define REG_A2XX_PA_SU_POINT_MINMAX 0x00002281 |
@@ -1257,13 +1257,13 @@ static inline uint32_t A2XX_PA_SU_POINT_SIZE_WIDTH(float val) | |||
1257 | #define A2XX_PA_SU_POINT_MINMAX_MIN__SHIFT 0 | 1257 | #define A2XX_PA_SU_POINT_MINMAX_MIN__SHIFT 0 |
1258 | static inline uint32_t A2XX_PA_SU_POINT_MINMAX_MIN(float val) | 1258 | static inline uint32_t A2XX_PA_SU_POINT_MINMAX_MIN(float val) |
1259 | { | 1259 | { |
1260 | return ((((uint32_t)(val * 8.0))) << A2XX_PA_SU_POINT_MINMAX_MIN__SHIFT) & A2XX_PA_SU_POINT_MINMAX_MIN__MASK; | 1260 | return ((((uint32_t)(val * 16.0))) << A2XX_PA_SU_POINT_MINMAX_MIN__SHIFT) & A2XX_PA_SU_POINT_MINMAX_MIN__MASK; |
1261 | } | 1261 | } |
1262 | #define A2XX_PA_SU_POINT_MINMAX_MAX__MASK 0xffff0000 | 1262 | #define A2XX_PA_SU_POINT_MINMAX_MAX__MASK 0xffff0000 |
1263 | #define A2XX_PA_SU_POINT_MINMAX_MAX__SHIFT 16 | 1263 | #define A2XX_PA_SU_POINT_MINMAX_MAX__SHIFT 16 |
1264 | static inline uint32_t A2XX_PA_SU_POINT_MINMAX_MAX(float val) | 1264 | static inline uint32_t A2XX_PA_SU_POINT_MINMAX_MAX(float val) |
1265 | { | 1265 | { |
1266 | return ((((uint32_t)(val * 8.0))) << A2XX_PA_SU_POINT_MINMAX_MAX__SHIFT) & A2XX_PA_SU_POINT_MINMAX_MAX__MASK; | 1266 | return ((((uint32_t)(val * 16.0))) << A2XX_PA_SU_POINT_MINMAX_MAX__SHIFT) & A2XX_PA_SU_POINT_MINMAX_MAX__MASK; |
1267 | } | 1267 | } |
1268 | 1268 | ||
1269 | #define REG_A2XX_PA_SU_LINE_CNTL 0x00002282 | 1269 | #define REG_A2XX_PA_SU_LINE_CNTL 0x00002282 |
@@ -1271,7 +1271,7 @@ static inline uint32_t A2XX_PA_SU_POINT_MINMAX_MAX(float val) | |||
1271 | #define A2XX_PA_SU_LINE_CNTL_WIDTH__SHIFT 0 | 1271 | #define A2XX_PA_SU_LINE_CNTL_WIDTH__SHIFT 0 |
1272 | static inline uint32_t A2XX_PA_SU_LINE_CNTL_WIDTH(float val) | 1272 | static inline uint32_t A2XX_PA_SU_LINE_CNTL_WIDTH(float val) |
1273 | { | 1273 | { |
1274 | return ((((uint32_t)(val * 8.0))) << A2XX_PA_SU_LINE_CNTL_WIDTH__SHIFT) & A2XX_PA_SU_LINE_CNTL_WIDTH__MASK; | 1274 | return ((((uint32_t)(val * 16.0))) << A2XX_PA_SU_LINE_CNTL_WIDTH__SHIFT) & A2XX_PA_SU_LINE_CNTL_WIDTH__MASK; |
1275 | } | 1275 | } |
1276 | 1276 | ||
1277 | #define REG_A2XX_PA_SC_LINE_STIPPLE 0x00002283 | 1277 | #define REG_A2XX_PA_SC_LINE_STIPPLE 0x00002283 |
diff --git a/drivers/gpu/drm/msm/adreno/a3xx.xml.h b/drivers/gpu/drm/msm/adreno/a3xx.xml.h index 82d015279b47..109e9a263daf 100644 --- a/drivers/gpu/drm/msm/adreno/a3xx.xml.h +++ b/drivers/gpu/drm/msm/adreno/a3xx.xml.h | |||
@@ -11,10 +11,10 @@ The rules-ng-ng source files this header was generated from are: | |||
11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 9859 bytes, from 2014-06-02 15:21:30) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 10551 bytes, from 2014-11-13 22:44:30) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 14960 bytes, from 2014-07-27 17:22:13) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 15053 bytes, from 2014-11-09 15:45:47) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 58020 bytes, from 2014-08-01 12:22:48) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 63169 bytes, from 2014-11-13 22:44:18) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 41068 bytes, from 2014-08-01 12:22:48) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 49097 bytes, from 2014-11-14 15:38:00) |
18 | 18 | ||
19 | Copyright (C) 2013-2014 by the following authors: | 19 | Copyright (C) 2013-2014 by the following authors: |
20 | - Rob Clark <robdclark@gmail.com> (robclark) | 20 | - Rob Clark <robdclark@gmail.com> (robclark) |
@@ -86,6 +86,14 @@ enum a3xx_vtx_fmt { | |||
86 | VFMT_NORM_USHORT_16_16 = 29, | 86 | VFMT_NORM_USHORT_16_16 = 29, |
87 | VFMT_NORM_USHORT_16_16_16 = 30, | 87 | VFMT_NORM_USHORT_16_16_16 = 30, |
88 | VFMT_NORM_USHORT_16_16_16_16 = 31, | 88 | VFMT_NORM_USHORT_16_16_16_16 = 31, |
89 | VFMT_UINT_32 = 32, | ||
90 | VFMT_UINT_32_32 = 33, | ||
91 | VFMT_UINT_32_32_32 = 34, | ||
92 | VFMT_UINT_32_32_32_32 = 35, | ||
93 | VFMT_INT_32 = 36, | ||
94 | VFMT_INT_32_32 = 37, | ||
95 | VFMT_INT_32_32_32 = 38, | ||
96 | VFMT_INT_32_32_32_32 = 39, | ||
89 | VFMT_UBYTE_8 = 40, | 97 | VFMT_UBYTE_8 = 40, |
90 | VFMT_UBYTE_8_8 = 41, | 98 | VFMT_UBYTE_8_8 = 41, |
91 | VFMT_UBYTE_8_8_8 = 42, | 99 | VFMT_UBYTE_8_8_8 = 42, |
@@ -112,7 +120,9 @@ enum a3xx_tex_fmt { | |||
112 | TFMT_NORM_USHORT_565 = 4, | 120 | TFMT_NORM_USHORT_565 = 4, |
113 | TFMT_NORM_USHORT_5551 = 6, | 121 | TFMT_NORM_USHORT_5551 = 6, |
114 | TFMT_NORM_USHORT_4444 = 7, | 122 | TFMT_NORM_USHORT_4444 = 7, |
123 | TFMT_NORM_USHORT_Z16 = 9, | ||
115 | TFMT_NORM_UINT_X8Z24 = 10, | 124 | TFMT_NORM_UINT_X8Z24 = 10, |
125 | TFMT_FLOAT_Z32 = 11, | ||
116 | TFMT_NORM_UINT_NV12_UV_TILED = 17, | 126 | TFMT_NORM_UINT_NV12_UV_TILED = 17, |
117 | TFMT_NORM_UINT_NV12_Y_TILED = 19, | 127 | TFMT_NORM_UINT_NV12_Y_TILED = 19, |
118 | TFMT_NORM_UINT_NV12_UV = 21, | 128 | TFMT_NORM_UINT_NV12_UV = 21, |
@@ -121,18 +131,38 @@ enum a3xx_tex_fmt { | |||
121 | TFMT_NORM_UINT_I420_U = 26, | 131 | TFMT_NORM_UINT_I420_U = 26, |
122 | TFMT_NORM_UINT_I420_V = 27, | 132 | TFMT_NORM_UINT_I420_V = 27, |
123 | TFMT_NORM_UINT_2_10_10_10 = 41, | 133 | TFMT_NORM_UINT_2_10_10_10 = 41, |
134 | TFMT_FLOAT_9_9_9_E5 = 42, | ||
135 | TFMT_FLOAT_10_11_11 = 43, | ||
124 | TFMT_NORM_UINT_A8 = 44, | 136 | TFMT_NORM_UINT_A8 = 44, |
125 | TFMT_NORM_UINT_L8_A8 = 47, | 137 | TFMT_NORM_UINT_L8_A8 = 47, |
126 | TFMT_NORM_UINT_8 = 48, | 138 | TFMT_NORM_UINT_8 = 48, |
127 | TFMT_NORM_UINT_8_8 = 49, | 139 | TFMT_NORM_UINT_8_8 = 49, |
128 | TFMT_NORM_UINT_8_8_8 = 50, | 140 | TFMT_NORM_UINT_8_8_8 = 50, |
129 | TFMT_NORM_UINT_8_8_8_8 = 51, | 141 | TFMT_NORM_UINT_8_8_8_8 = 51, |
142 | TFMT_NORM_SINT_8_8 = 53, | ||
143 | TFMT_NORM_SINT_8_8_8_8 = 55, | ||
144 | TFMT_UINT_8_8 = 57, | ||
145 | TFMT_UINT_8_8_8_8 = 59, | ||
146 | TFMT_SINT_8_8 = 61, | ||
147 | TFMT_SINT_8_8_8_8 = 63, | ||
130 | TFMT_FLOAT_16 = 64, | 148 | TFMT_FLOAT_16 = 64, |
131 | TFMT_FLOAT_16_16 = 65, | 149 | TFMT_FLOAT_16_16 = 65, |
132 | TFMT_FLOAT_16_16_16_16 = 67, | 150 | TFMT_FLOAT_16_16_16_16 = 67, |
151 | TFMT_UINT_16 = 68, | ||
152 | TFMT_UINT_16_16 = 69, | ||
153 | TFMT_UINT_16_16_16_16 = 71, | ||
154 | TFMT_SINT_16 = 72, | ||
155 | TFMT_SINT_16_16 = 73, | ||
156 | TFMT_SINT_16_16_16_16 = 75, | ||
133 | TFMT_FLOAT_32 = 84, | 157 | TFMT_FLOAT_32 = 84, |
134 | TFMT_FLOAT_32_32 = 85, | 158 | TFMT_FLOAT_32_32 = 85, |
135 | TFMT_FLOAT_32_32_32_32 = 87, | 159 | TFMT_FLOAT_32_32_32_32 = 87, |
160 | TFMT_UINT_32 = 88, | ||
161 | TFMT_UINT_32_32 = 89, | ||
162 | TFMT_UINT_32_32_32_32 = 91, | ||
163 | TFMT_SINT_32 = 92, | ||
164 | TFMT_SINT_32_32 = 93, | ||
165 | TFMT_SINT_32_32_32_32 = 95, | ||
136 | }; | 166 | }; |
137 | 167 | ||
138 | enum a3xx_tex_fetchsize { | 168 | enum a3xx_tex_fetchsize { |
@@ -145,19 +175,34 @@ enum a3xx_tex_fetchsize { | |||
145 | }; | 175 | }; |
146 | 176 | ||
147 | enum a3xx_color_fmt { | 177 | enum a3xx_color_fmt { |
178 | RB_R5G6B5_UNORM = 0, | ||
179 | RB_R5G5B5A1_UNORM = 1, | ||
180 | RB_R4G4B4A4_UNORM = 3, | ||
148 | RB_R8G8B8_UNORM = 4, | 181 | RB_R8G8B8_UNORM = 4, |
149 | RB_R8G8B8A8_UNORM = 8, | 182 | RB_R8G8B8A8_UNORM = 8, |
150 | RB_Z16_UNORM = 12, | 183 | RB_R8G8B8A8_UINT = 10, |
184 | RB_R8G8B8A8_SINT = 11, | ||
185 | RB_R8G8_UNORM = 12, | ||
186 | RB_R8_UINT = 14, | ||
187 | RB_R8_SINT = 15, | ||
188 | RB_R10G10B10A2_UNORM = 16, | ||
151 | RB_A8_UNORM = 20, | 189 | RB_A8_UNORM = 20, |
190 | RB_R8_UNORM = 21, | ||
152 | RB_R16G16B16A16_FLOAT = 27, | 191 | RB_R16G16B16A16_FLOAT = 27, |
192 | RB_R11G11B10_FLOAT = 28, | ||
193 | RB_R16_SINT = 40, | ||
194 | RB_R16G16_SINT = 41, | ||
195 | RB_R16G16B16A16_SINT = 43, | ||
196 | RB_R16_UINT = 44, | ||
197 | RB_R16G16_UINT = 45, | ||
198 | RB_R16G16B16A16_UINT = 47, | ||
153 | RB_R32G32B32A32_FLOAT = 51, | 199 | RB_R32G32B32A32_FLOAT = 51, |
154 | }; | 200 | RB_R32_SINT = 52, |
155 | 201 | RB_R32G32_SINT = 53, | |
156 | enum a3xx_color_swap { | 202 | RB_R32G32B32A32_SINT = 55, |
157 | WZYX = 0, | 203 | RB_R32_UINT = 56, |
158 | WXYZ = 1, | 204 | RB_R32G32_UINT = 57, |
159 | ZYXW = 2, | 205 | RB_R32G32B32A32_UINT = 59, |
160 | XYZW = 3, | ||
161 | }; | 206 | }; |
162 | 207 | ||
163 | enum a3xx_sp_perfcounter_select { | 208 | enum a3xx_sp_perfcounter_select { |
@@ -194,6 +239,11 @@ enum a3xx_rb_blend_opcode { | |||
194 | BLEND_MAX_DST_SRC = 4, | 239 | BLEND_MAX_DST_SRC = 4, |
195 | }; | 240 | }; |
196 | 241 | ||
242 | enum a3xx_intp_mode { | ||
243 | SMOOTH = 0, | ||
244 | FLAT = 1, | ||
245 | }; | ||
246 | |||
197 | enum a3xx_tex_filter { | 247 | enum a3xx_tex_filter { |
198 | A3XX_TEX_NEAREST = 0, | 248 | A3XX_TEX_NEAREST = 0, |
199 | A3XX_TEX_LINEAR = 1, | 249 | A3XX_TEX_LINEAR = 1, |
@@ -536,6 +586,10 @@ enum a3xx_tex_type { | |||
536 | 586 | ||
537 | #define REG_A3XX_CP_MEQ_DATA 0x000001db | 587 | #define REG_A3XX_CP_MEQ_DATA 0x000001db |
538 | 588 | ||
589 | #define REG_A3XX_CP_WFI_PEND_CTR 0x000001f5 | ||
590 | |||
591 | #define REG_A3XX_RBBM_PM_OVERRIDE2 0x0000039d | ||
592 | |||
539 | #define REG_A3XX_CP_PERFCOUNTER_SELECT 0x00000445 | 593 | #define REG_A3XX_CP_PERFCOUNTER_SELECT 0x00000445 |
540 | 594 | ||
541 | #define REG_A3XX_CP_HW_FAULT 0x0000045c | 595 | #define REG_A3XX_CP_HW_FAULT 0x0000045c |
@@ -550,6 +604,12 @@ static inline uint32_t REG_A3XX_CP_PROTECT_REG(uint32_t i0) { return 0x00000460 | |||
550 | 604 | ||
551 | #define REG_A3XX_CP_AHB_FAULT 0x0000054d | 605 | #define REG_A3XX_CP_AHB_FAULT 0x0000054d |
552 | 606 | ||
607 | #define REG_A3XX_SQ_GPR_MANAGEMENT 0x00000d00 | ||
608 | |||
609 | #define REG_A3XX_SQ_INST_STORE_MANAGMENT 0x00000d02 | ||
610 | |||
611 | #define REG_A3XX_TP0_CHICKEN 0x00000e1e | ||
612 | |||
553 | #define REG_A3XX_SP_GLOBAL_MEM_SIZE 0x00000e22 | 613 | #define REG_A3XX_SP_GLOBAL_MEM_SIZE 0x00000e22 |
554 | 614 | ||
555 | #define REG_A3XX_SP_GLOBAL_MEM_ADDR 0x00000e23 | 615 | #define REG_A3XX_SP_GLOBAL_MEM_ADDR 0x00000e23 |
@@ -632,13 +692,13 @@ static inline uint32_t A3XX_GRAS_CL_VPORT_ZSCALE(float val) | |||
632 | #define A3XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT 0 | 692 | #define A3XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT 0 |
633 | static inline uint32_t A3XX_GRAS_SU_POINT_MINMAX_MIN(float val) | 693 | static inline uint32_t A3XX_GRAS_SU_POINT_MINMAX_MIN(float val) |
634 | { | 694 | { |
635 | return ((((uint32_t)(val * 8.0))) << A3XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT) & A3XX_GRAS_SU_POINT_MINMAX_MIN__MASK; | 695 | return ((((uint32_t)(val * 16.0))) << A3XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT) & A3XX_GRAS_SU_POINT_MINMAX_MIN__MASK; |
636 | } | 696 | } |
637 | #define A3XX_GRAS_SU_POINT_MINMAX_MAX__MASK 0xffff0000 | 697 | #define A3XX_GRAS_SU_POINT_MINMAX_MAX__MASK 0xffff0000 |
638 | #define A3XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT 16 | 698 | #define A3XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT 16 |
639 | static inline uint32_t A3XX_GRAS_SU_POINT_MINMAX_MAX(float val) | 699 | static inline uint32_t A3XX_GRAS_SU_POINT_MINMAX_MAX(float val) |
640 | { | 700 | { |
641 | return ((((uint32_t)(val * 8.0))) << A3XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT) & A3XX_GRAS_SU_POINT_MINMAX_MAX__MASK; | 701 | return ((((uint32_t)(val * 16.0))) << A3XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT) & A3XX_GRAS_SU_POINT_MINMAX_MAX__MASK; |
642 | } | 702 | } |
643 | 703 | ||
644 | #define REG_A3XX_GRAS_SU_POINT_SIZE 0x00002069 | 704 | #define REG_A3XX_GRAS_SU_POINT_SIZE 0x00002069 |
@@ -646,7 +706,7 @@ static inline uint32_t A3XX_GRAS_SU_POINT_MINMAX_MAX(float val) | |||
646 | #define A3XX_GRAS_SU_POINT_SIZE__SHIFT 0 | 706 | #define A3XX_GRAS_SU_POINT_SIZE__SHIFT 0 |
647 | static inline uint32_t A3XX_GRAS_SU_POINT_SIZE(float val) | 707 | static inline uint32_t A3XX_GRAS_SU_POINT_SIZE(float val) |
648 | { | 708 | { |
649 | return ((((uint32_t)(val * 8.0))) << A3XX_GRAS_SU_POINT_SIZE__SHIFT) & A3XX_GRAS_SU_POINT_SIZE__MASK; | 709 | return ((((int32_t)(val * 16.0))) << A3XX_GRAS_SU_POINT_SIZE__SHIFT) & A3XX_GRAS_SU_POINT_SIZE__MASK; |
650 | } | 710 | } |
651 | 711 | ||
652 | #define REG_A3XX_GRAS_SU_POLY_OFFSET_SCALE 0x0000206c | 712 | #define REG_A3XX_GRAS_SU_POLY_OFFSET_SCALE 0x0000206c |
@@ -654,7 +714,7 @@ static inline uint32_t A3XX_GRAS_SU_POINT_SIZE(float val) | |||
654 | #define A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__SHIFT 0 | 714 | #define A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__SHIFT 0 |
655 | static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL(float val) | 715 | static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL(float val) |
656 | { | 716 | { |
657 | return ((((uint32_t)(val * 28.0))) << A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__SHIFT) & A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__MASK; | 717 | return ((((int32_t)(val * 16384.0))) << A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__SHIFT) & A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL__MASK; |
658 | } | 718 | } |
659 | 719 | ||
660 | #define REG_A3XX_GRAS_SU_POLY_OFFSET_OFFSET 0x0000206d | 720 | #define REG_A3XX_GRAS_SU_POLY_OFFSET_OFFSET 0x0000206d |
@@ -662,7 +722,7 @@ static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_SCALE_VAL(float val) | |||
662 | #define A3XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT 0 | 722 | #define A3XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT 0 |
663 | static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_OFFSET(float val) | 723 | static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_OFFSET(float val) |
664 | { | 724 | { |
665 | return ((((uint32_t)(val * 28.0))) << A3XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT) & A3XX_GRAS_SU_POLY_OFFSET_OFFSET__MASK; | 725 | return ((((int32_t)(val * 16384.0))) << A3XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT) & A3XX_GRAS_SU_POLY_OFFSET_OFFSET__MASK; |
666 | } | 726 | } |
667 | 727 | ||
668 | #define REG_A3XX_GRAS_SU_MODE_CONTROL 0x00002070 | 728 | #define REG_A3XX_GRAS_SU_MODE_CONTROL 0x00002070 |
@@ -673,7 +733,7 @@ static inline uint32_t A3XX_GRAS_SU_POLY_OFFSET_OFFSET(float val) | |||
673 | #define A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT 3 | 733 | #define A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT 3 |
674 | static inline uint32_t A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH(float val) | 734 | static inline uint32_t A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH(float val) |
675 | { | 735 | { |
676 | return ((((uint32_t)(val * 4.0))) << A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT) & A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__MASK; | 736 | return ((((int32_t)(val * 4.0))) << A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT) & A3XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__MASK; |
677 | } | 737 | } |
678 | #define A3XX_GRAS_SU_MODE_CONTROL_POLY_OFFSET 0x00000800 | 738 | #define A3XX_GRAS_SU_MODE_CONTROL_POLY_OFFSET 0x00000800 |
679 | 739 | ||
@@ -863,6 +923,7 @@ static inline uint32_t A3XX_RB_MRT_BUF_INFO_COLOR_SWAP(enum a3xx_color_swap val) | |||
863 | { | 923 | { |
864 | return ((val) << A3XX_RB_MRT_BUF_INFO_COLOR_SWAP__SHIFT) & A3XX_RB_MRT_BUF_INFO_COLOR_SWAP__MASK; | 924 | return ((val) << A3XX_RB_MRT_BUF_INFO_COLOR_SWAP__SHIFT) & A3XX_RB_MRT_BUF_INFO_COLOR_SWAP__MASK; |
865 | } | 925 | } |
926 | #define A3XX_RB_MRT_BUF_INFO_COLOR_SRGB 0x00004000 | ||
866 | #define A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__MASK 0xfffe0000 | 927 | #define A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__MASK 0xfffe0000 |
867 | #define A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__SHIFT 17 | 928 | #define A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__SHIFT 17 |
868 | static inline uint32_t A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH(uint32_t val) | 929 | static inline uint32_t A3XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH(uint32_t val) |
@@ -1001,6 +1062,7 @@ static inline uint32_t A3XX_RB_COPY_CONTROL_FASTCLEAR(uint32_t val) | |||
1001 | { | 1062 | { |
1002 | return ((val) << A3XX_RB_COPY_CONTROL_FASTCLEAR__SHIFT) & A3XX_RB_COPY_CONTROL_FASTCLEAR__MASK; | 1063 | return ((val) << A3XX_RB_COPY_CONTROL_FASTCLEAR__SHIFT) & A3XX_RB_COPY_CONTROL_FASTCLEAR__MASK; |
1003 | } | 1064 | } |
1065 | #define A3XX_RB_COPY_CONTROL_UNK12 0x00001000 | ||
1004 | #define A3XX_RB_COPY_CONTROL_GMEM_BASE__MASK 0xffffc000 | 1066 | #define A3XX_RB_COPY_CONTROL_GMEM_BASE__MASK 0xffffc000 |
1005 | #define A3XX_RB_COPY_CONTROL_GMEM_BASE__SHIFT 14 | 1067 | #define A3XX_RB_COPY_CONTROL_GMEM_BASE__SHIFT 14 |
1006 | static inline uint32_t A3XX_RB_COPY_CONTROL_GMEM_BASE(uint32_t val) | 1068 | static inline uint32_t A3XX_RB_COPY_CONTROL_GMEM_BASE(uint32_t val) |
@@ -1079,7 +1141,7 @@ static inline uint32_t A3XX_RB_DEPTH_CONTROL_ZFUNC(enum adreno_compare_func val) | |||
1079 | #define REG_A3XX_RB_DEPTH_CLEAR 0x00002101 | 1141 | #define REG_A3XX_RB_DEPTH_CLEAR 0x00002101 |
1080 | 1142 | ||
1081 | #define REG_A3XX_RB_DEPTH_INFO 0x00002102 | 1143 | #define REG_A3XX_RB_DEPTH_INFO 0x00002102 |
1082 | #define A3XX_RB_DEPTH_INFO_DEPTH_FORMAT__MASK 0x00000001 | 1144 | #define A3XX_RB_DEPTH_INFO_DEPTH_FORMAT__MASK 0x00000003 |
1083 | #define A3XX_RB_DEPTH_INFO_DEPTH_FORMAT__SHIFT 0 | 1145 | #define A3XX_RB_DEPTH_INFO_DEPTH_FORMAT__SHIFT 0 |
1084 | static inline uint32_t A3XX_RB_DEPTH_INFO_DEPTH_FORMAT(enum adreno_rb_depth_format val) | 1146 | static inline uint32_t A3XX_RB_DEPTH_INFO_DEPTH_FORMAT(enum adreno_rb_depth_format val) |
1085 | { | 1147 | { |
@@ -1265,6 +1327,7 @@ static inline uint32_t A3XX_PC_PRIM_VTX_CNTL_POLYMODE_BACK_PTYPE(enum adreno_pa_ | |||
1265 | { | 1327 | { |
1266 | return ((val) << A3XX_PC_PRIM_VTX_CNTL_POLYMODE_BACK_PTYPE__SHIFT) & A3XX_PC_PRIM_VTX_CNTL_POLYMODE_BACK_PTYPE__MASK; | 1328 | return ((val) << A3XX_PC_PRIM_VTX_CNTL_POLYMODE_BACK_PTYPE__SHIFT) & A3XX_PC_PRIM_VTX_CNTL_POLYMODE_BACK_PTYPE__MASK; |
1267 | } | 1329 | } |
1330 | #define A3XX_PC_PRIM_VTX_CNTL_PRIMITIVE_RESTART 0x00100000 | ||
1268 | #define A3XX_PC_PRIM_VTX_CNTL_PROVOKING_VTX_LAST 0x02000000 | 1331 | #define A3XX_PC_PRIM_VTX_CNTL_PROVOKING_VTX_LAST 0x02000000 |
1269 | #define A3XX_PC_PRIM_VTX_CNTL_PSIZE 0x04000000 | 1332 | #define A3XX_PC_PRIM_VTX_CNTL_PSIZE 0x04000000 |
1270 | 1333 | ||
@@ -1281,7 +1344,12 @@ static inline uint32_t A3XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE(enum a3xx_threadsize | |||
1281 | #define A3XX_HLSQ_CONTROL_0_REG_SPSHADERRESTART 0x00000200 | 1344 | #define A3XX_HLSQ_CONTROL_0_REG_SPSHADERRESTART 0x00000200 |
1282 | #define A3XX_HLSQ_CONTROL_0_REG_RESERVED2 0x00000400 | 1345 | #define A3XX_HLSQ_CONTROL_0_REG_RESERVED2 0x00000400 |
1283 | #define A3XX_HLSQ_CONTROL_0_REG_CHUNKDISABLE 0x04000000 | 1346 | #define A3XX_HLSQ_CONTROL_0_REG_CHUNKDISABLE 0x04000000 |
1284 | #define A3XX_HLSQ_CONTROL_0_REG_CONSTSWITCHMODE 0x08000000 | 1347 | #define A3XX_HLSQ_CONTROL_0_REG_CONSTMODE__MASK 0x08000000 |
1348 | #define A3XX_HLSQ_CONTROL_0_REG_CONSTMODE__SHIFT 27 | ||
1349 | static inline uint32_t A3XX_HLSQ_CONTROL_0_REG_CONSTMODE(uint32_t val) | ||
1350 | { | ||
1351 | return ((val) << A3XX_HLSQ_CONTROL_0_REG_CONSTMODE__SHIFT) & A3XX_HLSQ_CONTROL_0_REG_CONSTMODE__MASK; | ||
1352 | } | ||
1285 | #define A3XX_HLSQ_CONTROL_0_REG_LAZYUPDATEDISABLE 0x10000000 | 1353 | #define A3XX_HLSQ_CONTROL_0_REG_LAZYUPDATEDISABLE 0x10000000 |
1286 | #define A3XX_HLSQ_CONTROL_0_REG_SPCONSTFULLUPDATE 0x20000000 | 1354 | #define A3XX_HLSQ_CONTROL_0_REG_SPCONSTFULLUPDATE 0x20000000 |
1287 | #define A3XX_HLSQ_CONTROL_0_REG_TPFULLUPDATE 0x40000000 | 1355 | #define A3XX_HLSQ_CONTROL_0_REG_TPFULLUPDATE 0x40000000 |
@@ -1484,6 +1552,8 @@ static inline uint32_t A3XX_VFD_CONTROL_1_REGID4INST(uint32_t val) | |||
1484 | 1552 | ||
1485 | #define REG_A3XX_VFD_INDEX_OFFSET 0x00002245 | 1553 | #define REG_A3XX_VFD_INDEX_OFFSET 0x00002245 |
1486 | 1554 | ||
1555 | #define REG_A3XX_VFD_INDEX_OFFSET 0x00002245 | ||
1556 | |||
1487 | static inline uint32_t REG_A3XX_VFD_FETCH(uint32_t i0) { return 0x00002246 + 0x2*i0; } | 1557 | static inline uint32_t REG_A3XX_VFD_FETCH(uint32_t i0) { return 0x00002246 + 0x2*i0; } |
1488 | 1558 | ||
1489 | static inline uint32_t REG_A3XX_VFD_FETCH_INSTR_0(uint32_t i0) { return 0x00002246 + 0x2*i0; } | 1559 | static inline uint32_t REG_A3XX_VFD_FETCH_INSTR_0(uint32_t i0) { return 0x00002246 + 0x2*i0; } |
@@ -1537,6 +1607,7 @@ static inline uint32_t A3XX_VFD_DECODE_INSTR_REGID(uint32_t val) | |||
1537 | { | 1607 | { |
1538 | return ((val) << A3XX_VFD_DECODE_INSTR_REGID__SHIFT) & A3XX_VFD_DECODE_INSTR_REGID__MASK; | 1608 | return ((val) << A3XX_VFD_DECODE_INSTR_REGID__SHIFT) & A3XX_VFD_DECODE_INSTR_REGID__MASK; |
1539 | } | 1609 | } |
1610 | #define A3XX_VFD_DECODE_INSTR_INT 0x00100000 | ||
1540 | #define A3XX_VFD_DECODE_INSTR_SWAP__MASK 0x00c00000 | 1611 | #define A3XX_VFD_DECODE_INSTR_SWAP__MASK 0x00c00000 |
1541 | #define A3XX_VFD_DECODE_INSTR_SWAP__SHIFT 22 | 1612 | #define A3XX_VFD_DECODE_INSTR_SWAP__SHIFT 22 |
1542 | static inline uint32_t A3XX_VFD_DECODE_INSTR_SWAP(enum a3xx_color_swap val) | 1613 | static inline uint32_t A3XX_VFD_DECODE_INSTR_SWAP(enum a3xx_color_swap val) |
@@ -1604,6 +1675,102 @@ static inline uint32_t A3XX_VPC_PACK_NUMNONPOSVSVAR(uint32_t val) | |||
1604 | static inline uint32_t REG_A3XX_VPC_VARYING_INTERP(uint32_t i0) { return 0x00002282 + 0x1*i0; } | 1675 | static inline uint32_t REG_A3XX_VPC_VARYING_INTERP(uint32_t i0) { return 0x00002282 + 0x1*i0; } |
1605 | 1676 | ||
1606 | static inline uint32_t REG_A3XX_VPC_VARYING_INTERP_MODE(uint32_t i0) { return 0x00002282 + 0x1*i0; } | 1677 | static inline uint32_t REG_A3XX_VPC_VARYING_INTERP_MODE(uint32_t i0) { return 0x00002282 + 0x1*i0; } |
1678 | #define A3XX_VPC_VARYING_INTERP_MODE_C0__MASK 0x00000003 | ||
1679 | #define A3XX_VPC_VARYING_INTERP_MODE_C0__SHIFT 0 | ||
1680 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C0(enum a3xx_intp_mode val) | ||
1681 | { | ||
1682 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C0__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C0__MASK; | ||
1683 | } | ||
1684 | #define A3XX_VPC_VARYING_INTERP_MODE_C1__MASK 0x0000000c | ||
1685 | #define A3XX_VPC_VARYING_INTERP_MODE_C1__SHIFT 2 | ||
1686 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C1(enum a3xx_intp_mode val) | ||
1687 | { | ||
1688 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C1__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C1__MASK; | ||
1689 | } | ||
1690 | #define A3XX_VPC_VARYING_INTERP_MODE_C2__MASK 0x00000030 | ||
1691 | #define A3XX_VPC_VARYING_INTERP_MODE_C2__SHIFT 4 | ||
1692 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C2(enum a3xx_intp_mode val) | ||
1693 | { | ||
1694 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C2__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C2__MASK; | ||
1695 | } | ||
1696 | #define A3XX_VPC_VARYING_INTERP_MODE_C3__MASK 0x000000c0 | ||
1697 | #define A3XX_VPC_VARYING_INTERP_MODE_C3__SHIFT 6 | ||
1698 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C3(enum a3xx_intp_mode val) | ||
1699 | { | ||
1700 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C3__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C3__MASK; | ||
1701 | } | ||
1702 | #define A3XX_VPC_VARYING_INTERP_MODE_C4__MASK 0x00000300 | ||
1703 | #define A3XX_VPC_VARYING_INTERP_MODE_C4__SHIFT 8 | ||
1704 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C4(enum a3xx_intp_mode val) | ||
1705 | { | ||
1706 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C4__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C4__MASK; | ||
1707 | } | ||
1708 | #define A3XX_VPC_VARYING_INTERP_MODE_C5__MASK 0x00000c00 | ||
1709 | #define A3XX_VPC_VARYING_INTERP_MODE_C5__SHIFT 10 | ||
1710 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C5(enum a3xx_intp_mode val) | ||
1711 | { | ||
1712 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C5__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C5__MASK; | ||
1713 | } | ||
1714 | #define A3XX_VPC_VARYING_INTERP_MODE_C6__MASK 0x00003000 | ||
1715 | #define A3XX_VPC_VARYING_INTERP_MODE_C6__SHIFT 12 | ||
1716 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C6(enum a3xx_intp_mode val) | ||
1717 | { | ||
1718 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C6__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C6__MASK; | ||
1719 | } | ||
1720 | #define A3XX_VPC_VARYING_INTERP_MODE_C7__MASK 0x0000c000 | ||
1721 | #define A3XX_VPC_VARYING_INTERP_MODE_C7__SHIFT 14 | ||
1722 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C7(enum a3xx_intp_mode val) | ||
1723 | { | ||
1724 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C7__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C7__MASK; | ||
1725 | } | ||
1726 | #define A3XX_VPC_VARYING_INTERP_MODE_C8__MASK 0x00030000 | ||
1727 | #define A3XX_VPC_VARYING_INTERP_MODE_C8__SHIFT 16 | ||
1728 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C8(enum a3xx_intp_mode val) | ||
1729 | { | ||
1730 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C8__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C8__MASK; | ||
1731 | } | ||
1732 | #define A3XX_VPC_VARYING_INTERP_MODE_C9__MASK 0x000c0000 | ||
1733 | #define A3XX_VPC_VARYING_INTERP_MODE_C9__SHIFT 18 | ||
1734 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_C9(enum a3xx_intp_mode val) | ||
1735 | { | ||
1736 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_C9__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_C9__MASK; | ||
1737 | } | ||
1738 | #define A3XX_VPC_VARYING_INTERP_MODE_CA__MASK 0x00300000 | ||
1739 | #define A3XX_VPC_VARYING_INTERP_MODE_CA__SHIFT 20 | ||
1740 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CA(enum a3xx_intp_mode val) | ||
1741 | { | ||
1742 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CA__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CA__MASK; | ||
1743 | } | ||
1744 | #define A3XX_VPC_VARYING_INTERP_MODE_CB__MASK 0x00c00000 | ||
1745 | #define A3XX_VPC_VARYING_INTERP_MODE_CB__SHIFT 22 | ||
1746 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CB(enum a3xx_intp_mode val) | ||
1747 | { | ||
1748 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CB__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CB__MASK; | ||
1749 | } | ||
1750 | #define A3XX_VPC_VARYING_INTERP_MODE_CC__MASK 0x03000000 | ||
1751 | #define A3XX_VPC_VARYING_INTERP_MODE_CC__SHIFT 24 | ||
1752 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CC(enum a3xx_intp_mode val) | ||
1753 | { | ||
1754 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CC__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CC__MASK; | ||
1755 | } | ||
1756 | #define A3XX_VPC_VARYING_INTERP_MODE_CD__MASK 0x0c000000 | ||
1757 | #define A3XX_VPC_VARYING_INTERP_MODE_CD__SHIFT 26 | ||
1758 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CD(enum a3xx_intp_mode val) | ||
1759 | { | ||
1760 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CD__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CD__MASK; | ||
1761 | } | ||
1762 | #define A3XX_VPC_VARYING_INTERP_MODE_CE__MASK 0x30000000 | ||
1763 | #define A3XX_VPC_VARYING_INTERP_MODE_CE__SHIFT 28 | ||
1764 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CE(enum a3xx_intp_mode val) | ||
1765 | { | ||
1766 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CE__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CE__MASK; | ||
1767 | } | ||
1768 | #define A3XX_VPC_VARYING_INTERP_MODE_CF__MASK 0xc0000000 | ||
1769 | #define A3XX_VPC_VARYING_INTERP_MODE_CF__SHIFT 30 | ||
1770 | static inline uint32_t A3XX_VPC_VARYING_INTERP_MODE_CF(enum a3xx_intp_mode val) | ||
1771 | { | ||
1772 | return ((val) << A3XX_VPC_VARYING_INTERP_MODE_CF__SHIFT) & A3XX_VPC_VARYING_INTERP_MODE_CF__MASK; | ||
1773 | } | ||
1607 | 1774 | ||
1608 | static inline uint32_t REG_A3XX_VPC_VARYING_PS_REPL(uint32_t i0) { return 0x00002286 + 0x1*i0; } | 1775 | static inline uint32_t REG_A3XX_VPC_VARYING_PS_REPL(uint32_t i0) { return 0x00002286 + 0x1*i0; } |
1609 | 1776 | ||
@@ -1928,6 +2095,8 @@ static inline uint32_t A3XX_SP_FS_MRT_REG_REGID(uint32_t val) | |||
1928 | return ((val) << A3XX_SP_FS_MRT_REG_REGID__SHIFT) & A3XX_SP_FS_MRT_REG_REGID__MASK; | 2095 | return ((val) << A3XX_SP_FS_MRT_REG_REGID__SHIFT) & A3XX_SP_FS_MRT_REG_REGID__MASK; |
1929 | } | 2096 | } |
1930 | #define A3XX_SP_FS_MRT_REG_HALF_PRECISION 0x00000100 | 2097 | #define A3XX_SP_FS_MRT_REG_HALF_PRECISION 0x00000100 |
2098 | #define A3XX_SP_FS_MRT_REG_SINT 0x00000400 | ||
2099 | #define A3XX_SP_FS_MRT_REG_UINT 0x00000800 | ||
1931 | 2100 | ||
1932 | static inline uint32_t REG_A3XX_SP_FS_IMAGE_OUTPUT(uint32_t i0) { return 0x000022f4 + 0x1*i0; } | 2101 | static inline uint32_t REG_A3XX_SP_FS_IMAGE_OUTPUT(uint32_t i0) { return 0x000022f4 + 0x1*i0; } |
1933 | 2102 | ||
@@ -1947,6 +2116,8 @@ static inline uint32_t A3XX_SP_FS_LENGTH_REG_SHADERLENGTH(uint32_t val) | |||
1947 | return ((val) << A3XX_SP_FS_LENGTH_REG_SHADERLENGTH__SHIFT) & A3XX_SP_FS_LENGTH_REG_SHADERLENGTH__MASK; | 2116 | return ((val) << A3XX_SP_FS_LENGTH_REG_SHADERLENGTH__SHIFT) & A3XX_SP_FS_LENGTH_REG_SHADERLENGTH__MASK; |
1948 | } | 2117 | } |
1949 | 2118 | ||
2119 | #define REG_A3XX_PA_SC_AA_CONFIG 0x00002301 | ||
2120 | |||
1950 | #define REG_A3XX_TPL1_TP_VS_TEX_OFFSET 0x00002340 | 2121 | #define REG_A3XX_TPL1_TP_VS_TEX_OFFSET 0x00002340 |
1951 | #define A3XX_TPL1_TP_VS_TEX_OFFSET_SAMPLEROFFSET__MASK 0x000000ff | 2122 | #define A3XX_TPL1_TP_VS_TEX_OFFSET_SAMPLEROFFSET__MASK 0x000000ff |
1952 | #define A3XX_TPL1_TP_VS_TEX_OFFSET_SAMPLEROFFSET__SHIFT 0 | 2123 | #define A3XX_TPL1_TP_VS_TEX_OFFSET_SAMPLEROFFSET__SHIFT 0 |
@@ -2297,11 +2468,11 @@ static inline uint32_t A3XX_VGT_DRAW_INITIATOR_INDEX_SIZE(enum pc_di_index_size | |||
2297 | #define A3XX_VGT_DRAW_INITIATOR_NOT_EOP 0x00001000 | 2468 | #define A3XX_VGT_DRAW_INITIATOR_NOT_EOP 0x00001000 |
2298 | #define A3XX_VGT_DRAW_INITIATOR_SMALL_INDEX 0x00002000 | 2469 | #define A3XX_VGT_DRAW_INITIATOR_SMALL_INDEX 0x00002000 |
2299 | #define A3XX_VGT_DRAW_INITIATOR_PRE_DRAW_INITIATOR_ENABLE 0x00004000 | 2470 | #define A3XX_VGT_DRAW_INITIATOR_PRE_DRAW_INITIATOR_ENABLE 0x00004000 |
2300 | #define A3XX_VGT_DRAW_INITIATOR_NUM_INDICES__MASK 0xffff0000 | 2471 | #define A3XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__MASK 0xff000000 |
2301 | #define A3XX_VGT_DRAW_INITIATOR_NUM_INDICES__SHIFT 16 | 2472 | #define A3XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__SHIFT 24 |
2302 | static inline uint32_t A3XX_VGT_DRAW_INITIATOR_NUM_INDICES(uint32_t val) | 2473 | static inline uint32_t A3XX_VGT_DRAW_INITIATOR_NUM_INSTANCES(uint32_t val) |
2303 | { | 2474 | { |
2304 | return ((val) << A3XX_VGT_DRAW_INITIATOR_NUM_INDICES__SHIFT) & A3XX_VGT_DRAW_INITIATOR_NUM_INDICES__MASK; | 2475 | return ((val) << A3XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__SHIFT) & A3XX_VGT_DRAW_INITIATOR_NUM_INSTANCES__MASK; |
2305 | } | 2476 | } |
2306 | 2477 | ||
2307 | #define REG_A3XX_VGT_IMMED_DATA 0x000021fd | 2478 | #define REG_A3XX_VGT_IMMED_DATA 0x000021fd |
@@ -2347,17 +2518,23 @@ static inline uint32_t A3XX_TEX_SAMP_0_COMPARE_FUNC(enum adreno_compare_func val | |||
2347 | #define A3XX_TEX_SAMP_0_UNNORM_COORDS 0x80000000 | 2518 | #define A3XX_TEX_SAMP_0_UNNORM_COORDS 0x80000000 |
2348 | 2519 | ||
2349 | #define REG_A3XX_TEX_SAMP_1 0x00000001 | 2520 | #define REG_A3XX_TEX_SAMP_1 0x00000001 |
2521 | #define A3XX_TEX_SAMP_1_LOD_BIAS__MASK 0x000007ff | ||
2522 | #define A3XX_TEX_SAMP_1_LOD_BIAS__SHIFT 0 | ||
2523 | static inline uint32_t A3XX_TEX_SAMP_1_LOD_BIAS(float val) | ||
2524 | { | ||
2525 | return ((((int32_t)(val * 64.0))) << A3XX_TEX_SAMP_1_LOD_BIAS__SHIFT) & A3XX_TEX_SAMP_1_LOD_BIAS__MASK; | ||
2526 | } | ||
2350 | #define A3XX_TEX_SAMP_1_MAX_LOD__MASK 0x003ff000 | 2527 | #define A3XX_TEX_SAMP_1_MAX_LOD__MASK 0x003ff000 |
2351 | #define A3XX_TEX_SAMP_1_MAX_LOD__SHIFT 12 | 2528 | #define A3XX_TEX_SAMP_1_MAX_LOD__SHIFT 12 |
2352 | static inline uint32_t A3XX_TEX_SAMP_1_MAX_LOD(float val) | 2529 | static inline uint32_t A3XX_TEX_SAMP_1_MAX_LOD(float val) |
2353 | { | 2530 | { |
2354 | return ((((uint32_t)(val * 12.0))) << A3XX_TEX_SAMP_1_MAX_LOD__SHIFT) & A3XX_TEX_SAMP_1_MAX_LOD__MASK; | 2531 | return ((((uint32_t)(val * 64.0))) << A3XX_TEX_SAMP_1_MAX_LOD__SHIFT) & A3XX_TEX_SAMP_1_MAX_LOD__MASK; |
2355 | } | 2532 | } |
2356 | #define A3XX_TEX_SAMP_1_MIN_LOD__MASK 0xffc00000 | 2533 | #define A3XX_TEX_SAMP_1_MIN_LOD__MASK 0xffc00000 |
2357 | #define A3XX_TEX_SAMP_1_MIN_LOD__SHIFT 22 | 2534 | #define A3XX_TEX_SAMP_1_MIN_LOD__SHIFT 22 |
2358 | static inline uint32_t A3XX_TEX_SAMP_1_MIN_LOD(float val) | 2535 | static inline uint32_t A3XX_TEX_SAMP_1_MIN_LOD(float val) |
2359 | { | 2536 | { |
2360 | return ((((uint32_t)(val * 12.0))) << A3XX_TEX_SAMP_1_MIN_LOD__SHIFT) & A3XX_TEX_SAMP_1_MIN_LOD__MASK; | 2537 | return ((((uint32_t)(val * 64.0))) << A3XX_TEX_SAMP_1_MIN_LOD__SHIFT) & A3XX_TEX_SAMP_1_MIN_LOD__MASK; |
2361 | } | 2538 | } |
2362 | 2539 | ||
2363 | #define REG_A3XX_TEX_CONST_0 0x00000000 | 2540 | #define REG_A3XX_TEX_CONST_0 0x00000000 |
@@ -2448,6 +2625,24 @@ static inline uint32_t A3XX_TEX_CONST_2_SWAP(enum a3xx_color_swap val) | |||
2448 | } | 2625 | } |
2449 | 2626 | ||
2450 | #define REG_A3XX_TEX_CONST_3 0x00000003 | 2627 | #define REG_A3XX_TEX_CONST_3 0x00000003 |
2628 | #define A3XX_TEX_CONST_3_LAYERSZ1__MASK 0x0000000f | ||
2629 | #define A3XX_TEX_CONST_3_LAYERSZ1__SHIFT 0 | ||
2630 | static inline uint32_t A3XX_TEX_CONST_3_LAYERSZ1(uint32_t val) | ||
2631 | { | ||
2632 | return ((val >> 12) << A3XX_TEX_CONST_3_LAYERSZ1__SHIFT) & A3XX_TEX_CONST_3_LAYERSZ1__MASK; | ||
2633 | } | ||
2634 | #define A3XX_TEX_CONST_3_DEPTH__MASK 0x0ffe0000 | ||
2635 | #define A3XX_TEX_CONST_3_DEPTH__SHIFT 17 | ||
2636 | static inline uint32_t A3XX_TEX_CONST_3_DEPTH(uint32_t val) | ||
2637 | { | ||
2638 | return ((val) << A3XX_TEX_CONST_3_DEPTH__SHIFT) & A3XX_TEX_CONST_3_DEPTH__MASK; | ||
2639 | } | ||
2640 | #define A3XX_TEX_CONST_3_LAYERSZ2__MASK 0xf0000000 | ||
2641 | #define A3XX_TEX_CONST_3_LAYERSZ2__SHIFT 28 | ||
2642 | static inline uint32_t A3XX_TEX_CONST_3_LAYERSZ2(uint32_t val) | ||
2643 | { | ||
2644 | return ((val >> 12) << A3XX_TEX_CONST_3_LAYERSZ2__SHIFT) & A3XX_TEX_CONST_3_LAYERSZ2__MASK; | ||
2645 | } | ||
2451 | 2646 | ||
2452 | 2647 | ||
2453 | #endif /* A3XX_XML */ | 2648 | #endif /* A3XX_XML */ |
diff --git a/drivers/gpu/drm/msm/adreno/a3xx_gpu.c b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c index 218c5b060398..b66c53bdc039 100644 --- a/drivers/gpu/drm/msm/adreno/a3xx_gpu.c +++ b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c | |||
@@ -2,6 +2,8 @@ | |||
2 | * Copyright (C) 2013 Red Hat | 2 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 3 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 4 | * |
5 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
6 | * | ||
5 | * This program is free software; you can redistribute it and/or modify it | 7 | * This program is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 as published by | 8 | * under the terms of the GNU General Public License version 2 as published by |
7 | * the Free Software Foundation. | 9 | * the Free Software Foundation. |
@@ -406,6 +408,94 @@ static void a3xx_dump(struct msm_gpu *gpu) | |||
406 | gpu_read(gpu, REG_A3XX_RBBM_STATUS)); | 408 | gpu_read(gpu, REG_A3XX_RBBM_STATUS)); |
407 | adreno_dump(gpu); | 409 | adreno_dump(gpu); |
408 | } | 410 | } |
411 | /* Register offset defines for A3XX */ | ||
412 | static const unsigned int a3xx_register_offsets[REG_ADRENO_REGISTER_MAX] = { | ||
413 | REG_ADRENO_DEFINE(REG_ADRENO_CP_DEBUG, REG_AXXX_CP_DEBUG), | ||
414 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_WADDR, REG_AXXX_CP_ME_RAM_WADDR), | ||
415 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_DATA, REG_AXXX_CP_ME_RAM_DATA), | ||
416 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PFP_UCODE_DATA, | ||
417 | REG_A3XX_CP_PFP_UCODE_DATA), | ||
418 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PFP_UCODE_ADDR, | ||
419 | REG_A3XX_CP_PFP_UCODE_ADDR), | ||
420 | REG_ADRENO_DEFINE(REG_ADRENO_CP_WFI_PEND_CTR, REG_A3XX_CP_WFI_PEND_CTR), | ||
421 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_BASE, REG_AXXX_CP_RB_BASE), | ||
422 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_RPTR_ADDR, REG_AXXX_CP_RB_RPTR_ADDR), | ||
423 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_RPTR, REG_AXXX_CP_RB_RPTR), | ||
424 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_WPTR, REG_AXXX_CP_RB_WPTR), | ||
425 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PROTECT_CTRL, REG_A3XX_CP_PROTECT_CTRL), | ||
426 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_CNTL, REG_AXXX_CP_ME_CNTL), | ||
427 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_CNTL, REG_AXXX_CP_RB_CNTL), | ||
428 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB1_BASE, REG_AXXX_CP_IB1_BASE), | ||
429 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB1_BUFSZ, REG_AXXX_CP_IB1_BUFSZ), | ||
430 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB2_BASE, REG_AXXX_CP_IB2_BASE), | ||
431 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB2_BUFSZ, REG_AXXX_CP_IB2_BUFSZ), | ||
432 | REG_ADRENO_DEFINE(REG_ADRENO_CP_TIMESTAMP, REG_AXXX_CP_SCRATCH_REG0), | ||
433 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_RADDR, REG_AXXX_CP_ME_RAM_RADDR), | ||
434 | REG_ADRENO_DEFINE(REG_ADRENO_SCRATCH_ADDR, REG_AXXX_SCRATCH_ADDR), | ||
435 | REG_ADRENO_DEFINE(REG_ADRENO_SCRATCH_UMSK, REG_AXXX_SCRATCH_UMSK), | ||
436 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ROQ_ADDR, REG_A3XX_CP_ROQ_ADDR), | ||
437 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ROQ_DATA, REG_A3XX_CP_ROQ_DATA), | ||
438 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_ADDR, REG_A3XX_CP_MERCIU_ADDR), | ||
439 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_DATA, REG_A3XX_CP_MERCIU_DATA), | ||
440 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_DATA2, REG_A3XX_CP_MERCIU_DATA2), | ||
441 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MEQ_ADDR, REG_A3XX_CP_MEQ_ADDR), | ||
442 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MEQ_DATA, REG_A3XX_CP_MEQ_DATA), | ||
443 | REG_ADRENO_DEFINE(REG_ADRENO_CP_HW_FAULT, REG_A3XX_CP_HW_FAULT), | ||
444 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PROTECT_STATUS, | ||
445 | REG_A3XX_CP_PROTECT_STATUS), | ||
446 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_STATUS, REG_A3XX_RBBM_STATUS), | ||
447 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_CTL, | ||
448 | REG_A3XX_RBBM_PERFCTR_CTL), | ||
449 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_CMD0, | ||
450 | REG_A3XX_RBBM_PERFCTR_LOAD_CMD0), | ||
451 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_CMD1, | ||
452 | REG_A3XX_RBBM_PERFCTR_LOAD_CMD1), | ||
453 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_PWR_1_LO, | ||
454 | REG_A3XX_RBBM_PERFCTR_PWR_1_LO), | ||
455 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_0_MASK, REG_A3XX_RBBM_INT_0_MASK), | ||
456 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_0_STATUS, | ||
457 | REG_A3XX_RBBM_INT_0_STATUS), | ||
458 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_ERROR_STATUS, | ||
459 | REG_A3XX_RBBM_AHB_ERROR_STATUS), | ||
460 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_CMD, REG_A3XX_RBBM_AHB_CMD), | ||
461 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_CLEAR_CMD, | ||
462 | REG_A3XX_RBBM_INT_CLEAR_CMD), | ||
463 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_CLOCK_CTL, REG_A3XX_RBBM_CLOCK_CTL), | ||
464 | REG_ADRENO_DEFINE(REG_ADRENO_VPC_DEBUG_RAM_SEL, | ||
465 | REG_A3XX_VPC_VPC_DEBUG_RAM_SEL), | ||
466 | REG_ADRENO_DEFINE(REG_ADRENO_VPC_DEBUG_RAM_READ, | ||
467 | REG_A3XX_VPC_VPC_DEBUG_RAM_READ), | ||
468 | REG_ADRENO_DEFINE(REG_ADRENO_VSC_SIZE_ADDRESS, | ||
469 | REG_A3XX_VSC_SIZE_ADDRESS), | ||
470 | REG_ADRENO_DEFINE(REG_ADRENO_VFD_CONTROL_0, REG_A3XX_VFD_CONTROL_0), | ||
471 | REG_ADRENO_DEFINE(REG_ADRENO_VFD_INDEX_MAX, REG_A3XX_VFD_INDEX_MAX), | ||
472 | REG_ADRENO_DEFINE(REG_ADRENO_SP_VS_PVT_MEM_ADDR_REG, | ||
473 | REG_A3XX_SP_VS_PVT_MEM_ADDR_REG), | ||
474 | REG_ADRENO_DEFINE(REG_ADRENO_SP_FS_PVT_MEM_ADDR_REG, | ||
475 | REG_A3XX_SP_FS_PVT_MEM_ADDR_REG), | ||
476 | REG_ADRENO_DEFINE(REG_ADRENO_SP_VS_OBJ_START_REG, | ||
477 | REG_A3XX_SP_VS_OBJ_START_REG), | ||
478 | REG_ADRENO_DEFINE(REG_ADRENO_SP_FS_OBJ_START_REG, | ||
479 | REG_A3XX_SP_FS_OBJ_START_REG), | ||
480 | REG_ADRENO_DEFINE(REG_ADRENO_PA_SC_AA_CONFIG, REG_A3XX_PA_SC_AA_CONFIG), | ||
481 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PM_OVERRIDE2, | ||
482 | REG_A3XX_RBBM_PM_OVERRIDE2), | ||
483 | REG_ADRENO_DEFINE(REG_ADRENO_SCRATCH_REG2, REG_AXXX_CP_SCRATCH_REG2), | ||
484 | REG_ADRENO_DEFINE(REG_ADRENO_SQ_GPR_MANAGEMENT, | ||
485 | REG_A3XX_SQ_GPR_MANAGEMENT), | ||
486 | REG_ADRENO_DEFINE(REG_ADRENO_SQ_INST_STORE_MANAGMENT, | ||
487 | REG_A3XX_SQ_INST_STORE_MANAGMENT), | ||
488 | REG_ADRENO_DEFINE(REG_ADRENO_TP0_CHICKEN, REG_A3XX_TP0_CHICKEN), | ||
489 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_RBBM_CTL, REG_A3XX_RBBM_RBBM_CTL), | ||
490 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_SW_RESET_CMD, | ||
491 | REG_A3XX_RBBM_SW_RESET_CMD), | ||
492 | REG_ADRENO_DEFINE(REG_ADRENO_UCHE_INVALIDATE0, | ||
493 | REG_A3XX_UCHE_CACHE_INVALIDATE0_REG), | ||
494 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_LO, | ||
495 | REG_A3XX_RBBM_PERFCTR_LOAD_VALUE_LO), | ||
496 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_HI, | ||
497 | REG_A3XX_RBBM_PERFCTR_LOAD_VALUE_HI), | ||
498 | }; | ||
409 | 499 | ||
410 | static const struct adreno_gpu_funcs funcs = { | 500 | static const struct adreno_gpu_funcs funcs = { |
411 | .base = { | 501 | .base = { |
@@ -463,6 +553,7 @@ struct msm_gpu *a3xx_gpu_init(struct drm_device *dev) | |||
463 | gpu->num_perfcntrs = ARRAY_SIZE(perfcntrs); | 553 | gpu->num_perfcntrs = ARRAY_SIZE(perfcntrs); |
464 | 554 | ||
465 | adreno_gpu->registers = a3xx_registers; | 555 | adreno_gpu->registers = a3xx_registers; |
556 | adreno_gpu->reg_offsets = a3xx_register_offsets; | ||
466 | 557 | ||
467 | ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs); | 558 | ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs); |
468 | if (ret) | 559 | if (ret) |
diff --git a/drivers/gpu/drm/msm/adreno/a4xx.xml.h b/drivers/gpu/drm/msm/adreno/a4xx.xml.h new file mode 100644 index 000000000000..5a24c416d2dd --- /dev/null +++ b/drivers/gpu/drm/msm/adreno/a4xx.xml.h | |||
@@ -0,0 +1,2144 @@ | |||
1 | #ifndef A4XX_XML | ||
2 | #define A4XX_XML | ||
3 | |||
4 | /* Autogenerated file, DO NOT EDIT manually! | ||
5 | |||
6 | This file was generated by the rules-ng-ng headergen tool in this git repository: | ||
7 | http://github.com/freedreno/envytools/ | ||
8 | git clone https://github.com/freedreno/envytools.git | ||
9 | |||
10 | The rules-ng-ng source files this header was generated from are: | ||
11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) | ||
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | ||
13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) | ||
14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 10551 bytes, from 2014-11-13 22:44:30) | ||
15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 15053 bytes, from 2014-11-09 15:45:47) | ||
16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 63169 bytes, from 2014-11-13 22:44:18) | ||
17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 49097 bytes, from 2014-11-14 15:38:00) | ||
18 | |||
19 | Copyright (C) 2013-2014 by the following authors: | ||
20 | - Rob Clark <robdclark@gmail.com> (robclark) | ||
21 | |||
22 | Permission is hereby granted, free of charge, to any person obtaining | ||
23 | a copy of this software and associated documentation files (the | ||
24 | "Software"), to deal in the Software without restriction, including | ||
25 | without limitation the rights to use, copy, modify, merge, publish, | ||
26 | distribute, sublicense, and/or sell copies of the Software, and to | ||
27 | permit persons to whom the Software is furnished to do so, subject to | ||
28 | the following conditions: | ||
29 | |||
30 | The above copyright notice and this permission notice (including the | ||
31 | next paragraph) shall be included in all copies or substantial | ||
32 | portions of the Software. | ||
33 | |||
34 | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | ||
35 | EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | ||
36 | MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. | ||
37 | IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE | ||
38 | LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION | ||
39 | OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION | ||
40 | WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
41 | */ | ||
42 | |||
43 | |||
44 | enum a4xx_color_fmt { | ||
45 | RB4_A8_UNORM = 1, | ||
46 | RB4_R5G6R5_UNORM = 14, | ||
47 | RB4_Z16_UNORM = 15, | ||
48 | RB4_R8G8B8_UNORM = 25, | ||
49 | RB4_R8G8B8A8_UNORM = 26, | ||
50 | }; | ||
51 | |||
52 | enum a4xx_tile_mode { | ||
53 | TILE4_LINEAR = 0, | ||
54 | TILE4_3 = 3, | ||
55 | }; | ||
56 | |||
57 | enum a4xx_rb_blend_opcode { | ||
58 | BLEND_DST_PLUS_SRC = 0, | ||
59 | BLEND_SRC_MINUS_DST = 1, | ||
60 | BLEND_DST_MINUS_SRC = 2, | ||
61 | BLEND_MIN_DST_SRC = 3, | ||
62 | BLEND_MAX_DST_SRC = 4, | ||
63 | }; | ||
64 | |||
65 | enum a4xx_vtx_fmt { | ||
66 | VFMT4_FLOAT_32 = 1, | ||
67 | VFMT4_FLOAT_32_32 = 2, | ||
68 | VFMT4_FLOAT_32_32_32 = 3, | ||
69 | VFMT4_FLOAT_32_32_32_32 = 4, | ||
70 | VFMT4_FLOAT_16 = 5, | ||
71 | VFMT4_FLOAT_16_16 = 6, | ||
72 | VFMT4_FLOAT_16_16_16 = 7, | ||
73 | VFMT4_FLOAT_16_16_16_16 = 8, | ||
74 | VFMT4_FIXED_32 = 9, | ||
75 | VFMT4_FIXED_32_32 = 10, | ||
76 | VFMT4_FIXED_32_32_32 = 11, | ||
77 | VFMT4_FIXED_32_32_32_32 = 12, | ||
78 | VFMT4_SHORT_16 = 16, | ||
79 | VFMT4_SHORT_16_16 = 17, | ||
80 | VFMT4_SHORT_16_16_16 = 18, | ||
81 | VFMT4_SHORT_16_16_16_16 = 19, | ||
82 | VFMT4_USHORT_16 = 20, | ||
83 | VFMT4_USHORT_16_16 = 21, | ||
84 | VFMT4_USHORT_16_16_16 = 22, | ||
85 | VFMT4_USHORT_16_16_16_16 = 23, | ||
86 | VFMT4_NORM_SHORT_16 = 24, | ||
87 | VFMT4_NORM_SHORT_16_16 = 25, | ||
88 | VFMT4_NORM_SHORT_16_16_16 = 26, | ||
89 | VFMT4_NORM_SHORT_16_16_16_16 = 27, | ||
90 | VFMT4_NORM_USHORT_16 = 28, | ||
91 | VFMT4_NORM_USHORT_16_16 = 29, | ||
92 | VFMT4_NORM_USHORT_16_16_16 = 30, | ||
93 | VFMT4_NORM_USHORT_16_16_16_16 = 31, | ||
94 | VFMT4_UBYTE_8 = 40, | ||
95 | VFMT4_UBYTE_8_8 = 41, | ||
96 | VFMT4_UBYTE_8_8_8 = 42, | ||
97 | VFMT4_UBYTE_8_8_8_8 = 43, | ||
98 | VFMT4_NORM_UBYTE_8 = 44, | ||
99 | VFMT4_NORM_UBYTE_8_8 = 45, | ||
100 | VFMT4_NORM_UBYTE_8_8_8 = 46, | ||
101 | VFMT4_NORM_UBYTE_8_8_8_8 = 47, | ||
102 | VFMT4_BYTE_8 = 48, | ||
103 | VFMT4_BYTE_8_8 = 49, | ||
104 | VFMT4_BYTE_8_8_8 = 50, | ||
105 | VFMT4_BYTE_8_8_8_8 = 51, | ||
106 | VFMT4_NORM_BYTE_8 = 52, | ||
107 | VFMT4_NORM_BYTE_8_8 = 53, | ||
108 | VFMT4_NORM_BYTE_8_8_8 = 54, | ||
109 | VFMT4_NORM_BYTE_8_8_8_8 = 55, | ||
110 | VFMT4_UINT_10_10_10_2 = 60, | ||
111 | VFMT4_NORM_UINT_10_10_10_2 = 61, | ||
112 | VFMT4_INT_10_10_10_2 = 62, | ||
113 | VFMT4_NORM_INT_10_10_10_2 = 63, | ||
114 | }; | ||
115 | |||
116 | enum a4xx_tex_fmt { | ||
117 | TFMT4_NORM_USHORT_565 = 11, | ||
118 | TFMT4_NORM_USHORT_5551 = 10, | ||
119 | TFMT4_NORM_USHORT_4444 = 8, | ||
120 | TFMT4_NORM_UINT_X8Z24 = 71, | ||
121 | TFMT4_NORM_UINT_2_10_10_10 = 33, | ||
122 | TFMT4_NORM_UINT_A8 = 3, | ||
123 | TFMT4_NORM_UINT_L8_A8 = 13, | ||
124 | TFMT4_NORM_UINT_8 = 4, | ||
125 | TFMT4_NORM_UINT_8_8_8_8 = 28, | ||
126 | TFMT4_FLOAT_16 = 20, | ||
127 | TFMT4_FLOAT_16_16 = 40, | ||
128 | TFMT4_FLOAT_16_16_16_16 = 53, | ||
129 | TFMT4_FLOAT_32 = 43, | ||
130 | TFMT4_FLOAT_32_32 = 56, | ||
131 | TFMT4_FLOAT_32_32_32_32 = 63, | ||
132 | }; | ||
133 | |||
134 | enum a4xx_depth_format { | ||
135 | DEPTH4_NONE = 0, | ||
136 | DEPTH4_16 = 1, | ||
137 | DEPTH4_24_8 = 2, | ||
138 | }; | ||
139 | |||
140 | enum a4xx_tex_filter { | ||
141 | A4XX_TEX_NEAREST = 0, | ||
142 | A4XX_TEX_LINEAR = 1, | ||
143 | }; | ||
144 | |||
145 | enum a4xx_tex_clamp { | ||
146 | A4XX_TEX_REPEAT = 0, | ||
147 | A4XX_TEX_CLAMP_TO_EDGE = 1, | ||
148 | A4XX_TEX_MIRROR_REPEAT = 2, | ||
149 | A4XX_TEX_CLAMP_NONE = 3, | ||
150 | }; | ||
151 | |||
152 | enum a4xx_tex_swiz { | ||
153 | A4XX_TEX_X = 0, | ||
154 | A4XX_TEX_Y = 1, | ||
155 | A4XX_TEX_Z = 2, | ||
156 | A4XX_TEX_W = 3, | ||
157 | A4XX_TEX_ZERO = 4, | ||
158 | A4XX_TEX_ONE = 5, | ||
159 | }; | ||
160 | |||
161 | enum a4xx_tex_type { | ||
162 | A4XX_TEX_1D = 0, | ||
163 | A4XX_TEX_2D = 1, | ||
164 | A4XX_TEX_CUBE = 2, | ||
165 | A4XX_TEX_3D = 3, | ||
166 | }; | ||
167 | |||
168 | #define A4XX_CGC_HLSQ_EARLY_CYC__MASK 0x00700000 | ||
169 | #define A4XX_CGC_HLSQ_EARLY_CYC__SHIFT 20 | ||
170 | static inline uint32_t A4XX_CGC_HLSQ_EARLY_CYC(uint32_t val) | ||
171 | { | ||
172 | return ((val) << A4XX_CGC_HLSQ_EARLY_CYC__SHIFT) & A4XX_CGC_HLSQ_EARLY_CYC__MASK; | ||
173 | } | ||
174 | #define A4XX_INT0_RBBM_GPU_IDLE 0x00000001 | ||
175 | #define A4XX_INT0_RBBM_AHB_ERROR 0x00000002 | ||
176 | #define A4XX_INT0_RBBM_REG_TIMEOUT 0x00000004 | ||
177 | #define A4XX_INT0_RBBM_ME_MS_TIMEOUT 0x00000008 | ||
178 | #define A4XX_INT0_RBBM_PFP_MS_TIMEOUT 0x00000010 | ||
179 | #define A4XX_INT0_RBBM_ATB_BUS_OVERFLOW 0x00000020 | ||
180 | #define A4XX_INT0_VFD_ERROR 0x00000040 | ||
181 | #define A4XX_INT0_CP_SW_INT 0x00000080 | ||
182 | #define A4XX_INT0_CP_T0_PACKET_IN_IB 0x00000100 | ||
183 | #define A4XX_INT0_CP_OPCODE_ERROR 0x00000200 | ||
184 | #define A4XX_INT0_CP_RESERVED_BIT_ERROR 0x00000400 | ||
185 | #define A4XX_INT0_CP_HW_FAULT 0x00000800 | ||
186 | #define A4XX_INT0_CP_DMA 0x00001000 | ||
187 | #define A4XX_INT0_CP_IB2_INT 0x00002000 | ||
188 | #define A4XX_INT0_CP_IB1_INT 0x00004000 | ||
189 | #define A4XX_INT0_CP_RB_INT 0x00008000 | ||
190 | #define A4XX_INT0_CP_REG_PROTECT_FAULT 0x00010000 | ||
191 | #define A4XX_INT0_CP_RB_DONE_TS 0x00020000 | ||
192 | #define A4XX_INT0_CP_VS_DONE_TS 0x00040000 | ||
193 | #define A4XX_INT0_CP_PS_DONE_TS 0x00080000 | ||
194 | #define A4XX_INT0_CACHE_FLUSH_TS 0x00100000 | ||
195 | #define A4XX_INT0_CP_AHB_ERROR_HALT 0x00200000 | ||
196 | #define A4XX_INT0_MISC_HANG_DETECT 0x01000000 | ||
197 | #define A4XX_INT0_UCHE_OOB_ACCESS 0x02000000 | ||
198 | #define REG_A4XX_RB_GMEM_BASE_ADDR 0x00000cc0 | ||
199 | |||
200 | #define REG_A4XX_RB_PERFCTR_RB_SEL_0 0x00000cc7 | ||
201 | |||
202 | #define REG_A4XX_RB_PERFCTR_RB_SEL_1 0x00000cc8 | ||
203 | |||
204 | #define REG_A4XX_RB_PERFCTR_RB_SEL_2 0x00000cc9 | ||
205 | |||
206 | #define REG_A4XX_RB_PERFCTR_RB_SEL_3 0x00000cca | ||
207 | |||
208 | #define REG_A4XX_RB_PERFCTR_RB_SEL_4 0x00000ccb | ||
209 | |||
210 | #define REG_A4XX_RB_PERFCTR_RB_SEL_5 0x00000ccc | ||
211 | |||
212 | #define REG_A4XX_RB_PERFCTR_RB_SEL_6 0x00000ccd | ||
213 | |||
214 | #define REG_A4XX_RB_PERFCTR_RB_SEL_7 0x00000cce | ||
215 | |||
216 | #define REG_A4XX_RB_PERFCTR_CCU_SEL_3 0x00000cd2 | ||
217 | |||
218 | #define REG_A4XX_RB_FRAME_BUFFER_DIMENSION 0x00000ce0 | ||
219 | #define A4XX_RB_FRAME_BUFFER_DIMENSION_WIDTH__MASK 0x00003fff | ||
220 | #define A4XX_RB_FRAME_BUFFER_DIMENSION_WIDTH__SHIFT 0 | ||
221 | static inline uint32_t A4XX_RB_FRAME_BUFFER_DIMENSION_WIDTH(uint32_t val) | ||
222 | { | ||
223 | return ((val) << A4XX_RB_FRAME_BUFFER_DIMENSION_WIDTH__SHIFT) & A4XX_RB_FRAME_BUFFER_DIMENSION_WIDTH__MASK; | ||
224 | } | ||
225 | #define A4XX_RB_FRAME_BUFFER_DIMENSION_HEIGHT__MASK 0x3fff0000 | ||
226 | #define A4XX_RB_FRAME_BUFFER_DIMENSION_HEIGHT__SHIFT 16 | ||
227 | static inline uint32_t A4XX_RB_FRAME_BUFFER_DIMENSION_HEIGHT(uint32_t val) | ||
228 | { | ||
229 | return ((val) << A4XX_RB_FRAME_BUFFER_DIMENSION_HEIGHT__SHIFT) & A4XX_RB_FRAME_BUFFER_DIMENSION_HEIGHT__MASK; | ||
230 | } | ||
231 | |||
232 | #define REG_A4XX_RB_CLEAR_COLOR_DW0 0x000020cc | ||
233 | |||
234 | #define REG_A4XX_RB_CLEAR_COLOR_DW1 0x000020cd | ||
235 | |||
236 | #define REG_A4XX_RB_CLEAR_COLOR_DW2 0x000020ce | ||
237 | |||
238 | #define REG_A4XX_RB_CLEAR_COLOR_DW3 0x000020cf | ||
239 | |||
240 | #define REG_A4XX_RB_MODE_CONTROL 0x000020a0 | ||
241 | #define A4XX_RB_MODE_CONTROL_WIDTH__MASK 0x0000003f | ||
242 | #define A4XX_RB_MODE_CONTROL_WIDTH__SHIFT 0 | ||
243 | static inline uint32_t A4XX_RB_MODE_CONTROL_WIDTH(uint32_t val) | ||
244 | { | ||
245 | return ((val >> 5) << A4XX_RB_MODE_CONTROL_WIDTH__SHIFT) & A4XX_RB_MODE_CONTROL_WIDTH__MASK; | ||
246 | } | ||
247 | #define A4XX_RB_MODE_CONTROL_HEIGHT__MASK 0x00003f00 | ||
248 | #define A4XX_RB_MODE_CONTROL_HEIGHT__SHIFT 8 | ||
249 | static inline uint32_t A4XX_RB_MODE_CONTROL_HEIGHT(uint32_t val) | ||
250 | { | ||
251 | return ((val >> 5) << A4XX_RB_MODE_CONTROL_HEIGHT__SHIFT) & A4XX_RB_MODE_CONTROL_HEIGHT__MASK; | ||
252 | } | ||
253 | |||
254 | #define REG_A4XX_RB_RENDER_CONTROL 0x000020a1 | ||
255 | #define A4XX_RB_RENDER_CONTROL_BINNING_PASS 0x00000001 | ||
256 | #define A4XX_RB_RENDER_CONTROL_DISABLE_COLOR_PIPE 0x00000020 | ||
257 | |||
258 | #define REG_A4XX_RB_MSAA_CONTROL 0x000020a2 | ||
259 | #define A4XX_RB_MSAA_CONTROL_DISABLE 0x00001000 | ||
260 | #define A4XX_RB_MSAA_CONTROL_SAMPLES__MASK 0x0000e000 | ||
261 | #define A4XX_RB_MSAA_CONTROL_SAMPLES__SHIFT 13 | ||
262 | static inline uint32_t A4XX_RB_MSAA_CONTROL_SAMPLES(uint32_t val) | ||
263 | { | ||
264 | return ((val) << A4XX_RB_MSAA_CONTROL_SAMPLES__SHIFT) & A4XX_RB_MSAA_CONTROL_SAMPLES__MASK; | ||
265 | } | ||
266 | |||
267 | #define REG_A4XX_RB_MSAA_CONTROL2 0x000020a3 | ||
268 | #define A4XX_RB_MSAA_CONTROL2_MSAA_SAMPLES__MASK 0x00000380 | ||
269 | #define A4XX_RB_MSAA_CONTROL2_MSAA_SAMPLES__SHIFT 7 | ||
270 | static inline uint32_t A4XX_RB_MSAA_CONTROL2_MSAA_SAMPLES(uint32_t val) | ||
271 | { | ||
272 | return ((val) << A4XX_RB_MSAA_CONTROL2_MSAA_SAMPLES__SHIFT) & A4XX_RB_MSAA_CONTROL2_MSAA_SAMPLES__MASK; | ||
273 | } | ||
274 | #define A4XX_RB_MSAA_CONTROL2_VARYING 0x00001000 | ||
275 | |||
276 | static inline uint32_t REG_A4XX_RB_MRT(uint32_t i0) { return 0x000020a4 + 0x5*i0; } | ||
277 | |||
278 | static inline uint32_t REG_A4XX_RB_MRT_CONTROL(uint32_t i0) { return 0x000020a4 + 0x5*i0; } | ||
279 | #define A4XX_RB_MRT_CONTROL_READ_DEST_ENABLE 0x00000008 | ||
280 | #define A4XX_RB_MRT_CONTROL_BLEND 0x00000010 | ||
281 | #define A4XX_RB_MRT_CONTROL_BLEND2 0x00000020 | ||
282 | #define A4XX_RB_MRT_CONTROL_FASTCLEAR 0x00000400 | ||
283 | #define A4XX_RB_MRT_CONTROL_B11 0x00000800 | ||
284 | #define A4XX_RB_MRT_CONTROL_COMPONENT_ENABLE__MASK 0x0f000000 | ||
285 | #define A4XX_RB_MRT_CONTROL_COMPONENT_ENABLE__SHIFT 24 | ||
286 | static inline uint32_t A4XX_RB_MRT_CONTROL_COMPONENT_ENABLE(uint32_t val) | ||
287 | { | ||
288 | return ((val) << A4XX_RB_MRT_CONTROL_COMPONENT_ENABLE__SHIFT) & A4XX_RB_MRT_CONTROL_COMPONENT_ENABLE__MASK; | ||
289 | } | ||
290 | |||
291 | static inline uint32_t REG_A4XX_RB_MRT_BUF_INFO(uint32_t i0) { return 0x000020a5 + 0x5*i0; } | ||
292 | #define A4XX_RB_MRT_BUF_INFO_COLOR_FORMAT__MASK 0x0000003f | ||
293 | #define A4XX_RB_MRT_BUF_INFO_COLOR_FORMAT__SHIFT 0 | ||
294 | static inline uint32_t A4XX_RB_MRT_BUF_INFO_COLOR_FORMAT(enum a4xx_color_fmt val) | ||
295 | { | ||
296 | return ((val) << A4XX_RB_MRT_BUF_INFO_COLOR_FORMAT__SHIFT) & A4XX_RB_MRT_BUF_INFO_COLOR_FORMAT__MASK; | ||
297 | } | ||
298 | #define A4XX_RB_MRT_BUF_INFO_DITHER_MODE__MASK 0x00000600 | ||
299 | #define A4XX_RB_MRT_BUF_INFO_DITHER_MODE__SHIFT 9 | ||
300 | static inline uint32_t A4XX_RB_MRT_BUF_INFO_DITHER_MODE(enum adreno_rb_dither_mode val) | ||
301 | { | ||
302 | return ((val) << A4XX_RB_MRT_BUF_INFO_DITHER_MODE__SHIFT) & A4XX_RB_MRT_BUF_INFO_DITHER_MODE__MASK; | ||
303 | } | ||
304 | #define A4XX_RB_MRT_BUF_INFO_COLOR_SWAP__MASK 0x00001800 | ||
305 | #define A4XX_RB_MRT_BUF_INFO_COLOR_SWAP__SHIFT 11 | ||
306 | static inline uint32_t A4XX_RB_MRT_BUF_INFO_COLOR_SWAP(enum a3xx_color_swap val) | ||
307 | { | ||
308 | return ((val) << A4XX_RB_MRT_BUF_INFO_COLOR_SWAP__SHIFT) & A4XX_RB_MRT_BUF_INFO_COLOR_SWAP__MASK; | ||
309 | } | ||
310 | #define A4XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__MASK 0x007fc000 | ||
311 | #define A4XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__SHIFT 14 | ||
312 | static inline uint32_t A4XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH(uint32_t val) | ||
313 | { | ||
314 | return ((val >> 4) << A4XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__SHIFT) & A4XX_RB_MRT_BUF_INFO_COLOR_BUF_PITCH__MASK; | ||
315 | } | ||
316 | |||
317 | static inline uint32_t REG_A4XX_RB_MRT_BASE(uint32_t i0) { return 0x000020a6 + 0x5*i0; } | ||
318 | |||
319 | static inline uint32_t REG_A4XX_RB_MRT_CONTROL3(uint32_t i0) { return 0x000020a7 + 0x5*i0; } | ||
320 | #define A4XX_RB_MRT_CONTROL3_STRIDE__MASK 0x0001fff8 | ||
321 | #define A4XX_RB_MRT_CONTROL3_STRIDE__SHIFT 3 | ||
322 | static inline uint32_t A4XX_RB_MRT_CONTROL3_STRIDE(uint32_t val) | ||
323 | { | ||
324 | return ((val) << A4XX_RB_MRT_CONTROL3_STRIDE__SHIFT) & A4XX_RB_MRT_CONTROL3_STRIDE__MASK; | ||
325 | } | ||
326 | |||
327 | static inline uint32_t REG_A4XX_RB_MRT_BLEND_CONTROL(uint32_t i0) { return 0x000020a8 + 0x5*i0; } | ||
328 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_SRC_FACTOR__MASK 0x0000001f | ||
329 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_SRC_FACTOR__SHIFT 0 | ||
330 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_RGB_SRC_FACTOR(enum adreno_rb_blend_factor val) | ||
331 | { | ||
332 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_RGB_SRC_FACTOR__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_RGB_SRC_FACTOR__MASK; | ||
333 | } | ||
334 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_BLEND_OPCODE__MASK 0x000000e0 | ||
335 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_BLEND_OPCODE__SHIFT 5 | ||
336 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_RGB_BLEND_OPCODE(enum a4xx_rb_blend_opcode val) | ||
337 | { | ||
338 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_RGB_BLEND_OPCODE__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_RGB_BLEND_OPCODE__MASK; | ||
339 | } | ||
340 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_DEST_FACTOR__MASK 0x00001f00 | ||
341 | #define A4XX_RB_MRT_BLEND_CONTROL_RGB_DEST_FACTOR__SHIFT 8 | ||
342 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_RGB_DEST_FACTOR(enum adreno_rb_blend_factor val) | ||
343 | { | ||
344 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_RGB_DEST_FACTOR__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_RGB_DEST_FACTOR__MASK; | ||
345 | } | ||
346 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_SRC_FACTOR__MASK 0x001f0000 | ||
347 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_SRC_FACTOR__SHIFT 16 | ||
348 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_ALPHA_SRC_FACTOR(enum adreno_rb_blend_factor val) | ||
349 | { | ||
350 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_ALPHA_SRC_FACTOR__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_ALPHA_SRC_FACTOR__MASK; | ||
351 | } | ||
352 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_BLEND_OPCODE__MASK 0x00e00000 | ||
353 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_BLEND_OPCODE__SHIFT 21 | ||
354 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_ALPHA_BLEND_OPCODE(enum a4xx_rb_blend_opcode val) | ||
355 | { | ||
356 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_ALPHA_BLEND_OPCODE__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_ALPHA_BLEND_OPCODE__MASK; | ||
357 | } | ||
358 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_DEST_FACTOR__MASK 0x1f000000 | ||
359 | #define A4XX_RB_MRT_BLEND_CONTROL_ALPHA_DEST_FACTOR__SHIFT 24 | ||
360 | static inline uint32_t A4XX_RB_MRT_BLEND_CONTROL_ALPHA_DEST_FACTOR(enum adreno_rb_blend_factor val) | ||
361 | { | ||
362 | return ((val) << A4XX_RB_MRT_BLEND_CONTROL_ALPHA_DEST_FACTOR__SHIFT) & A4XX_RB_MRT_BLEND_CONTROL_ALPHA_DEST_FACTOR__MASK; | ||
363 | } | ||
364 | |||
365 | #define REG_A4XX_RB_ALPHA_CONTROL 0x000020f8 | ||
366 | #define A4XX_RB_ALPHA_CONTROL_ALPHA_TEST 0x00000100 | ||
367 | #define A4XX_RB_ALPHA_CONTROL_ALPHA_TEST_FUNC__MASK 0x00000e00 | ||
368 | #define A4XX_RB_ALPHA_CONTROL_ALPHA_TEST_FUNC__SHIFT 9 | ||
369 | static inline uint32_t A4XX_RB_ALPHA_CONTROL_ALPHA_TEST_FUNC(enum adreno_compare_func val) | ||
370 | { | ||
371 | return ((val) << A4XX_RB_ALPHA_CONTROL_ALPHA_TEST_FUNC__SHIFT) & A4XX_RB_ALPHA_CONTROL_ALPHA_TEST_FUNC__MASK; | ||
372 | } | ||
373 | |||
374 | #define REG_A4XX_RB_FS_OUTPUT 0x000020f9 | ||
375 | #define A4XX_RB_FS_OUTPUT_ENABLE_COLOR_PIPE 0x00000001 | ||
376 | #define A4XX_RB_FS_OUTPUT_FAST_CLEAR 0x00000100 | ||
377 | #define A4XX_RB_FS_OUTPUT_SAMPLE_MASK__MASK 0xffff0000 | ||
378 | #define A4XX_RB_FS_OUTPUT_SAMPLE_MASK__SHIFT 16 | ||
379 | static inline uint32_t A4XX_RB_FS_OUTPUT_SAMPLE_MASK(uint32_t val) | ||
380 | { | ||
381 | return ((val) << A4XX_RB_FS_OUTPUT_SAMPLE_MASK__SHIFT) & A4XX_RB_FS_OUTPUT_SAMPLE_MASK__MASK; | ||
382 | } | ||
383 | |||
384 | #define REG_A4XX_RB_RENDER_CONTROL3 0x000020fb | ||
385 | #define A4XX_RB_RENDER_CONTROL3_COMPONENT_ENABLE__MASK 0x0000001f | ||
386 | #define A4XX_RB_RENDER_CONTROL3_COMPONENT_ENABLE__SHIFT 0 | ||
387 | static inline uint32_t A4XX_RB_RENDER_CONTROL3_COMPONENT_ENABLE(uint32_t val) | ||
388 | { | ||
389 | return ((val) << A4XX_RB_RENDER_CONTROL3_COMPONENT_ENABLE__SHIFT) & A4XX_RB_RENDER_CONTROL3_COMPONENT_ENABLE__MASK; | ||
390 | } | ||
391 | |||
392 | #define REG_A4XX_RB_COPY_CONTROL 0x000020fc | ||
393 | #define A4XX_RB_COPY_CONTROL_MSAA_RESOLVE__MASK 0x00000003 | ||
394 | #define A4XX_RB_COPY_CONTROL_MSAA_RESOLVE__SHIFT 0 | ||
395 | static inline uint32_t A4XX_RB_COPY_CONTROL_MSAA_RESOLVE(enum a3xx_msaa_samples val) | ||
396 | { | ||
397 | return ((val) << A4XX_RB_COPY_CONTROL_MSAA_RESOLVE__SHIFT) & A4XX_RB_COPY_CONTROL_MSAA_RESOLVE__MASK; | ||
398 | } | ||
399 | #define A4XX_RB_COPY_CONTROL_MODE__MASK 0x00000070 | ||
400 | #define A4XX_RB_COPY_CONTROL_MODE__SHIFT 4 | ||
401 | static inline uint32_t A4XX_RB_COPY_CONTROL_MODE(enum adreno_rb_copy_control_mode val) | ||
402 | { | ||
403 | return ((val) << A4XX_RB_COPY_CONTROL_MODE__SHIFT) & A4XX_RB_COPY_CONTROL_MODE__MASK; | ||
404 | } | ||
405 | #define A4XX_RB_COPY_CONTROL_FASTCLEAR__MASK 0x00000f00 | ||
406 | #define A4XX_RB_COPY_CONTROL_FASTCLEAR__SHIFT 8 | ||
407 | static inline uint32_t A4XX_RB_COPY_CONTROL_FASTCLEAR(uint32_t val) | ||
408 | { | ||
409 | return ((val) << A4XX_RB_COPY_CONTROL_FASTCLEAR__SHIFT) & A4XX_RB_COPY_CONTROL_FASTCLEAR__MASK; | ||
410 | } | ||
411 | #define A4XX_RB_COPY_CONTROL_GMEM_BASE__MASK 0xffffc000 | ||
412 | #define A4XX_RB_COPY_CONTROL_GMEM_BASE__SHIFT 14 | ||
413 | static inline uint32_t A4XX_RB_COPY_CONTROL_GMEM_BASE(uint32_t val) | ||
414 | { | ||
415 | return ((val >> 14) << A4XX_RB_COPY_CONTROL_GMEM_BASE__SHIFT) & A4XX_RB_COPY_CONTROL_GMEM_BASE__MASK; | ||
416 | } | ||
417 | |||
418 | #define REG_A4XX_RB_COPY_DEST_BASE 0x000020fd | ||
419 | #define A4XX_RB_COPY_DEST_BASE_BASE__MASK 0xfffffff0 | ||
420 | #define A4XX_RB_COPY_DEST_BASE_BASE__SHIFT 4 | ||
421 | static inline uint32_t A4XX_RB_COPY_DEST_BASE_BASE(uint32_t val) | ||
422 | { | ||
423 | return ((val >> 4) << A4XX_RB_COPY_DEST_BASE_BASE__SHIFT) & A4XX_RB_COPY_DEST_BASE_BASE__MASK; | ||
424 | } | ||
425 | |||
426 | #define REG_A4XX_RB_COPY_DEST_PITCH 0x000020fe | ||
427 | #define A4XX_RB_COPY_DEST_PITCH_PITCH__MASK 0xffffffff | ||
428 | #define A4XX_RB_COPY_DEST_PITCH_PITCH__SHIFT 0 | ||
429 | static inline uint32_t A4XX_RB_COPY_DEST_PITCH_PITCH(uint32_t val) | ||
430 | { | ||
431 | return ((val >> 5) << A4XX_RB_COPY_DEST_PITCH_PITCH__SHIFT) & A4XX_RB_COPY_DEST_PITCH_PITCH__MASK; | ||
432 | } | ||
433 | |||
434 | #define REG_A4XX_RB_COPY_DEST_INFO 0x000020ff | ||
435 | #define A4XX_RB_COPY_DEST_INFO_FORMAT__MASK 0x000000fc | ||
436 | #define A4XX_RB_COPY_DEST_INFO_FORMAT__SHIFT 2 | ||
437 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_FORMAT(enum a4xx_color_fmt val) | ||
438 | { | ||
439 | return ((val) << A4XX_RB_COPY_DEST_INFO_FORMAT__SHIFT) & A4XX_RB_COPY_DEST_INFO_FORMAT__MASK; | ||
440 | } | ||
441 | #define A4XX_RB_COPY_DEST_INFO_SWAP__MASK 0x00000300 | ||
442 | #define A4XX_RB_COPY_DEST_INFO_SWAP__SHIFT 8 | ||
443 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_SWAP(enum a3xx_color_swap val) | ||
444 | { | ||
445 | return ((val) << A4XX_RB_COPY_DEST_INFO_SWAP__SHIFT) & A4XX_RB_COPY_DEST_INFO_SWAP__MASK; | ||
446 | } | ||
447 | #define A4XX_RB_COPY_DEST_INFO_DITHER_MODE__MASK 0x00000c00 | ||
448 | #define A4XX_RB_COPY_DEST_INFO_DITHER_MODE__SHIFT 10 | ||
449 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_DITHER_MODE(enum adreno_rb_dither_mode val) | ||
450 | { | ||
451 | return ((val) << A4XX_RB_COPY_DEST_INFO_DITHER_MODE__SHIFT) & A4XX_RB_COPY_DEST_INFO_DITHER_MODE__MASK; | ||
452 | } | ||
453 | #define A4XX_RB_COPY_DEST_INFO_COMPONENT_ENABLE__MASK 0x0003c000 | ||
454 | #define A4XX_RB_COPY_DEST_INFO_COMPONENT_ENABLE__SHIFT 14 | ||
455 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_COMPONENT_ENABLE(uint32_t val) | ||
456 | { | ||
457 | return ((val) << A4XX_RB_COPY_DEST_INFO_COMPONENT_ENABLE__SHIFT) & A4XX_RB_COPY_DEST_INFO_COMPONENT_ENABLE__MASK; | ||
458 | } | ||
459 | #define A4XX_RB_COPY_DEST_INFO_ENDIAN__MASK 0x001c0000 | ||
460 | #define A4XX_RB_COPY_DEST_INFO_ENDIAN__SHIFT 18 | ||
461 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_ENDIAN(enum adreno_rb_surface_endian val) | ||
462 | { | ||
463 | return ((val) << A4XX_RB_COPY_DEST_INFO_ENDIAN__SHIFT) & A4XX_RB_COPY_DEST_INFO_ENDIAN__MASK; | ||
464 | } | ||
465 | #define A4XX_RB_COPY_DEST_INFO_TILE__MASK 0x03000000 | ||
466 | #define A4XX_RB_COPY_DEST_INFO_TILE__SHIFT 24 | ||
467 | static inline uint32_t A4XX_RB_COPY_DEST_INFO_TILE(enum a4xx_tile_mode val) | ||
468 | { | ||
469 | return ((val) << A4XX_RB_COPY_DEST_INFO_TILE__SHIFT) & A4XX_RB_COPY_DEST_INFO_TILE__MASK; | ||
470 | } | ||
471 | |||
472 | #define REG_A4XX_RB_FS_OUTPUT_REG 0x00002100 | ||
473 | #define A4XX_RB_FS_OUTPUT_REG_COLOR_PIPE_ENABLE 0x00000001 | ||
474 | #define A4XX_RB_FS_OUTPUT_REG_FRAG_WRITES_Z 0x00000020 | ||
475 | |||
476 | #define REG_A4XX_RB_DEPTH_CONTROL 0x00002101 | ||
477 | #define A4XX_RB_DEPTH_CONTROL_FRAG_WRITES_Z 0x00000001 | ||
478 | #define A4XX_RB_DEPTH_CONTROL_Z_ENABLE 0x00000002 | ||
479 | #define A4XX_RB_DEPTH_CONTROL_Z_WRITE_ENABLE 0x00000004 | ||
480 | #define A4XX_RB_DEPTH_CONTROL_ZFUNC__MASK 0x00000070 | ||
481 | #define A4XX_RB_DEPTH_CONTROL_ZFUNC__SHIFT 4 | ||
482 | static inline uint32_t A4XX_RB_DEPTH_CONTROL_ZFUNC(enum adreno_compare_func val) | ||
483 | { | ||
484 | return ((val) << A4XX_RB_DEPTH_CONTROL_ZFUNC__SHIFT) & A4XX_RB_DEPTH_CONTROL_ZFUNC__MASK; | ||
485 | } | ||
486 | #define A4XX_RB_DEPTH_CONTROL_BF_ENABLE 0x00000080 | ||
487 | #define A4XX_RB_DEPTH_CONTROL_EARLY_Z_DISABLE 0x00010000 | ||
488 | #define A4XX_RB_DEPTH_CONTROL_Z_TEST_ENABLE 0x80000000 | ||
489 | |||
490 | #define REG_A4XX_RB_DEPTH_CLEAR 0x00002102 | ||
491 | |||
492 | #define REG_A4XX_RB_DEPTH_INFO 0x00002103 | ||
493 | #define A4XX_RB_DEPTH_INFO_DEPTH_FORMAT__MASK 0x00000003 | ||
494 | #define A4XX_RB_DEPTH_INFO_DEPTH_FORMAT__SHIFT 0 | ||
495 | static inline uint32_t A4XX_RB_DEPTH_INFO_DEPTH_FORMAT(enum a4xx_depth_format val) | ||
496 | { | ||
497 | return ((val) << A4XX_RB_DEPTH_INFO_DEPTH_FORMAT__SHIFT) & A4XX_RB_DEPTH_INFO_DEPTH_FORMAT__MASK; | ||
498 | } | ||
499 | #define A4XX_RB_DEPTH_INFO_DEPTH_BASE__MASK 0xfffff000 | ||
500 | #define A4XX_RB_DEPTH_INFO_DEPTH_BASE__SHIFT 12 | ||
501 | static inline uint32_t A4XX_RB_DEPTH_INFO_DEPTH_BASE(uint32_t val) | ||
502 | { | ||
503 | return ((val >> 12) << A4XX_RB_DEPTH_INFO_DEPTH_BASE__SHIFT) & A4XX_RB_DEPTH_INFO_DEPTH_BASE__MASK; | ||
504 | } | ||
505 | |||
506 | #define REG_A4XX_RB_DEPTH_PITCH 0x00002104 | ||
507 | #define A4XX_RB_DEPTH_PITCH__MASK 0xffffffff | ||
508 | #define A4XX_RB_DEPTH_PITCH__SHIFT 0 | ||
509 | static inline uint32_t A4XX_RB_DEPTH_PITCH(uint32_t val) | ||
510 | { | ||
511 | return ((val >> 4) << A4XX_RB_DEPTH_PITCH__SHIFT) & A4XX_RB_DEPTH_PITCH__MASK; | ||
512 | } | ||
513 | |||
514 | #define REG_A4XX_RB_DEPTH_PITCH2 0x00002105 | ||
515 | #define A4XX_RB_DEPTH_PITCH2__MASK 0xffffffff | ||
516 | #define A4XX_RB_DEPTH_PITCH2__SHIFT 0 | ||
517 | static inline uint32_t A4XX_RB_DEPTH_PITCH2(uint32_t val) | ||
518 | { | ||
519 | return ((val >> 4) << A4XX_RB_DEPTH_PITCH2__SHIFT) & A4XX_RB_DEPTH_PITCH2__MASK; | ||
520 | } | ||
521 | |||
522 | #define REG_A4XX_RB_STENCIL_CONTROL 0x00002106 | ||
523 | #define A4XX_RB_STENCIL_CONTROL_STENCIL_ENABLE 0x00000001 | ||
524 | #define A4XX_RB_STENCIL_CONTROL_STENCIL_ENABLE_BF 0x00000002 | ||
525 | #define A4XX_RB_STENCIL_CONTROL_STENCIL_READ 0x00000004 | ||
526 | #define A4XX_RB_STENCIL_CONTROL_FUNC__MASK 0x00000700 | ||
527 | #define A4XX_RB_STENCIL_CONTROL_FUNC__SHIFT 8 | ||
528 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_FUNC(enum adreno_compare_func val) | ||
529 | { | ||
530 | return ((val) << A4XX_RB_STENCIL_CONTROL_FUNC__SHIFT) & A4XX_RB_STENCIL_CONTROL_FUNC__MASK; | ||
531 | } | ||
532 | #define A4XX_RB_STENCIL_CONTROL_FAIL__MASK 0x00003800 | ||
533 | #define A4XX_RB_STENCIL_CONTROL_FAIL__SHIFT 11 | ||
534 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_FAIL(enum adreno_stencil_op val) | ||
535 | { | ||
536 | return ((val) << A4XX_RB_STENCIL_CONTROL_FAIL__SHIFT) & A4XX_RB_STENCIL_CONTROL_FAIL__MASK; | ||
537 | } | ||
538 | #define A4XX_RB_STENCIL_CONTROL_ZPASS__MASK 0x0001c000 | ||
539 | #define A4XX_RB_STENCIL_CONTROL_ZPASS__SHIFT 14 | ||
540 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_ZPASS(enum adreno_stencil_op val) | ||
541 | { | ||
542 | return ((val) << A4XX_RB_STENCIL_CONTROL_ZPASS__SHIFT) & A4XX_RB_STENCIL_CONTROL_ZPASS__MASK; | ||
543 | } | ||
544 | #define A4XX_RB_STENCIL_CONTROL_ZFAIL__MASK 0x000e0000 | ||
545 | #define A4XX_RB_STENCIL_CONTROL_ZFAIL__SHIFT 17 | ||
546 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_ZFAIL(enum adreno_stencil_op val) | ||
547 | { | ||
548 | return ((val) << A4XX_RB_STENCIL_CONTROL_ZFAIL__SHIFT) & A4XX_RB_STENCIL_CONTROL_ZFAIL__MASK; | ||
549 | } | ||
550 | #define A4XX_RB_STENCIL_CONTROL_FUNC_BF__MASK 0x00700000 | ||
551 | #define A4XX_RB_STENCIL_CONTROL_FUNC_BF__SHIFT 20 | ||
552 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_FUNC_BF(enum adreno_compare_func val) | ||
553 | { | ||
554 | return ((val) << A4XX_RB_STENCIL_CONTROL_FUNC_BF__SHIFT) & A4XX_RB_STENCIL_CONTROL_FUNC_BF__MASK; | ||
555 | } | ||
556 | #define A4XX_RB_STENCIL_CONTROL_FAIL_BF__MASK 0x03800000 | ||
557 | #define A4XX_RB_STENCIL_CONTROL_FAIL_BF__SHIFT 23 | ||
558 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_FAIL_BF(enum adreno_stencil_op val) | ||
559 | { | ||
560 | return ((val) << A4XX_RB_STENCIL_CONTROL_FAIL_BF__SHIFT) & A4XX_RB_STENCIL_CONTROL_FAIL_BF__MASK; | ||
561 | } | ||
562 | #define A4XX_RB_STENCIL_CONTROL_ZPASS_BF__MASK 0x1c000000 | ||
563 | #define A4XX_RB_STENCIL_CONTROL_ZPASS_BF__SHIFT 26 | ||
564 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_ZPASS_BF(enum adreno_stencil_op val) | ||
565 | { | ||
566 | return ((val) << A4XX_RB_STENCIL_CONTROL_ZPASS_BF__SHIFT) & A4XX_RB_STENCIL_CONTROL_ZPASS_BF__MASK; | ||
567 | } | ||
568 | #define A4XX_RB_STENCIL_CONTROL_ZFAIL_BF__MASK 0xe0000000 | ||
569 | #define A4XX_RB_STENCIL_CONTROL_ZFAIL_BF__SHIFT 29 | ||
570 | static inline uint32_t A4XX_RB_STENCIL_CONTROL_ZFAIL_BF(enum adreno_stencil_op val) | ||
571 | { | ||
572 | return ((val) << A4XX_RB_STENCIL_CONTROL_ZFAIL_BF__SHIFT) & A4XX_RB_STENCIL_CONTROL_ZFAIL_BF__MASK; | ||
573 | } | ||
574 | |||
575 | #define REG_A4XX_RB_STENCIL_CONTROL2 0x00002107 | ||
576 | #define A4XX_RB_STENCIL_CONTROL2_STENCIL_BUFFER 0x00000001 | ||
577 | |||
578 | #define REG_A4XX_RB_STENCILREFMASK 0x0000210b | ||
579 | #define A4XX_RB_STENCILREFMASK_STENCILREF__MASK 0x000000ff | ||
580 | #define A4XX_RB_STENCILREFMASK_STENCILREF__SHIFT 0 | ||
581 | static inline uint32_t A4XX_RB_STENCILREFMASK_STENCILREF(uint32_t val) | ||
582 | { | ||
583 | return ((val) << A4XX_RB_STENCILREFMASK_STENCILREF__SHIFT) & A4XX_RB_STENCILREFMASK_STENCILREF__MASK; | ||
584 | } | ||
585 | #define A4XX_RB_STENCILREFMASK_STENCILMASK__MASK 0x0000ff00 | ||
586 | #define A4XX_RB_STENCILREFMASK_STENCILMASK__SHIFT 8 | ||
587 | static inline uint32_t A4XX_RB_STENCILREFMASK_STENCILMASK(uint32_t val) | ||
588 | { | ||
589 | return ((val) << A4XX_RB_STENCILREFMASK_STENCILMASK__SHIFT) & A4XX_RB_STENCILREFMASK_STENCILMASK__MASK; | ||
590 | } | ||
591 | #define A4XX_RB_STENCILREFMASK_STENCILWRITEMASK__MASK 0x00ff0000 | ||
592 | #define A4XX_RB_STENCILREFMASK_STENCILWRITEMASK__SHIFT 16 | ||
593 | static inline uint32_t A4XX_RB_STENCILREFMASK_STENCILWRITEMASK(uint32_t val) | ||
594 | { | ||
595 | return ((val) << A4XX_RB_STENCILREFMASK_STENCILWRITEMASK__SHIFT) & A4XX_RB_STENCILREFMASK_STENCILWRITEMASK__MASK; | ||
596 | } | ||
597 | |||
598 | #define REG_A4XX_RB_STENCILREFMASK_BF 0x0000210c | ||
599 | #define A4XX_RB_STENCILREFMASK_BF_STENCILREF__MASK 0x000000ff | ||
600 | #define A4XX_RB_STENCILREFMASK_BF_STENCILREF__SHIFT 0 | ||
601 | static inline uint32_t A4XX_RB_STENCILREFMASK_BF_STENCILREF(uint32_t val) | ||
602 | { | ||
603 | return ((val) << A4XX_RB_STENCILREFMASK_BF_STENCILREF__SHIFT) & A4XX_RB_STENCILREFMASK_BF_STENCILREF__MASK; | ||
604 | } | ||
605 | #define A4XX_RB_STENCILREFMASK_BF_STENCILMASK__MASK 0x0000ff00 | ||
606 | #define A4XX_RB_STENCILREFMASK_BF_STENCILMASK__SHIFT 8 | ||
607 | static inline uint32_t A4XX_RB_STENCILREFMASK_BF_STENCILMASK(uint32_t val) | ||
608 | { | ||
609 | return ((val) << A4XX_RB_STENCILREFMASK_BF_STENCILMASK__SHIFT) & A4XX_RB_STENCILREFMASK_BF_STENCILMASK__MASK; | ||
610 | } | ||
611 | #define A4XX_RB_STENCILREFMASK_BF_STENCILWRITEMASK__MASK 0x00ff0000 | ||
612 | #define A4XX_RB_STENCILREFMASK_BF_STENCILWRITEMASK__SHIFT 16 | ||
613 | static inline uint32_t A4XX_RB_STENCILREFMASK_BF_STENCILWRITEMASK(uint32_t val) | ||
614 | { | ||
615 | return ((val) << A4XX_RB_STENCILREFMASK_BF_STENCILWRITEMASK__SHIFT) & A4XX_RB_STENCILREFMASK_BF_STENCILWRITEMASK__MASK; | ||
616 | } | ||
617 | |||
618 | #define REG_A4XX_RB_BIN_OFFSET 0x0000210d | ||
619 | #define A4XX_RB_BIN_OFFSET_WINDOW_OFFSET_DISABLE 0x80000000 | ||
620 | #define A4XX_RB_BIN_OFFSET_X__MASK 0x00007fff | ||
621 | #define A4XX_RB_BIN_OFFSET_X__SHIFT 0 | ||
622 | static inline uint32_t A4XX_RB_BIN_OFFSET_X(uint32_t val) | ||
623 | { | ||
624 | return ((val) << A4XX_RB_BIN_OFFSET_X__SHIFT) & A4XX_RB_BIN_OFFSET_X__MASK; | ||
625 | } | ||
626 | #define A4XX_RB_BIN_OFFSET_Y__MASK 0x7fff0000 | ||
627 | #define A4XX_RB_BIN_OFFSET_Y__SHIFT 16 | ||
628 | static inline uint32_t A4XX_RB_BIN_OFFSET_Y(uint32_t val) | ||
629 | { | ||
630 | return ((val) << A4XX_RB_BIN_OFFSET_Y__SHIFT) & A4XX_RB_BIN_OFFSET_Y__MASK; | ||
631 | } | ||
632 | |||
633 | #define REG_A4XX_RB_VPORT_Z_CLAMP_MAX_15 0x0000213f | ||
634 | |||
635 | #define REG_A4XX_RBBM_HW_VERSION 0x00000000 | ||
636 | |||
637 | #define REG_A4XX_RBBM_HW_CONFIGURATION 0x00000002 | ||
638 | |||
639 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_TP(uint32_t i0) { return 0x00000004 + 0x1*i0; } | ||
640 | |||
641 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_TP_REG(uint32_t i0) { return 0x00000004 + 0x1*i0; } | ||
642 | |||
643 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_TP(uint32_t i0) { return 0x00000008 + 0x1*i0; } | ||
644 | |||
645 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_TP_REG(uint32_t i0) { return 0x00000008 + 0x1*i0; } | ||
646 | |||
647 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_TP(uint32_t i0) { return 0x0000000c + 0x1*i0; } | ||
648 | |||
649 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_TP_REG(uint32_t i0) { return 0x0000000c + 0x1*i0; } | ||
650 | |||
651 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_TP(uint32_t i0) { return 0x00000010 + 0x1*i0; } | ||
652 | |||
653 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_TP_REG(uint32_t i0) { return 0x00000010 + 0x1*i0; } | ||
654 | |||
655 | #define REG_A4XX_RBBM_CLOCK_CTL_UCHE 0x00000014 | ||
656 | |||
657 | #define REG_A4XX_RBBM_CLOCK_CTL2_UCHE 0x00000015 | ||
658 | |||
659 | #define REG_A4XX_RBBM_CLOCK_CTL3_UCHE 0x00000016 | ||
660 | |||
661 | #define REG_A4XX_RBBM_CLOCK_CTL4_UCHE 0x00000017 | ||
662 | |||
663 | #define REG_A4XX_RBBM_CLOCK_HYST_UCHE 0x00000018 | ||
664 | |||
665 | #define REG_A4XX_RBBM_CLOCK_DELAY_UCHE 0x00000019 | ||
666 | |||
667 | #define REG_A4XX_RBBM_CLOCK_MODE_GPC 0x0000001a | ||
668 | |||
669 | #define REG_A4XX_RBBM_CLOCK_DELAY_GPC 0x0000001b | ||
670 | |||
671 | #define REG_A4XX_RBBM_CLOCK_HYST_GPC 0x0000001c | ||
672 | |||
673 | #define REG_A4XX_RBBM_CLOCK_CTL_TSE_RAS_RBBM 0x0000001d | ||
674 | |||
675 | #define REG_A4XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM 0x0000001e | ||
676 | |||
677 | #define REG_A4XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM 0x0000001f | ||
678 | |||
679 | #define REG_A4XX_RBBM_CLOCK_CTL 0x00000020 | ||
680 | |||
681 | #define REG_A4XX_RBBM_SP_HYST_CNT 0x00000021 | ||
682 | |||
683 | #define REG_A4XX_RBBM_SW_RESET_CMD 0x00000022 | ||
684 | |||
685 | #define REG_A4XX_RBBM_AHB_CTL0 0x00000023 | ||
686 | |||
687 | #define REG_A4XX_RBBM_AHB_CTL1 0x00000024 | ||
688 | |||
689 | #define REG_A4XX_RBBM_AHB_CMD 0x00000025 | ||
690 | |||
691 | #define REG_A4XX_RBBM_RB_SUB_BLOCK_SEL_CTL 0x00000026 | ||
692 | |||
693 | #define REG_A4XX_RBBM_RAM_ACC_63_32 0x00000028 | ||
694 | |||
695 | #define REG_A4XX_RBBM_WAIT_IDLE_CLOCKS_CTL 0x0000002b | ||
696 | |||
697 | #define REG_A4XX_RBBM_INTERFACE_HANG_INT_CTL 0x0000002f | ||
698 | |||
699 | #define REG_A4XX_RBBM_INTERFACE_HANG_MASK_CTL4 0x00000034 | ||
700 | |||
701 | #define REG_A4XX_RBBM_INT_CLEAR_CMD 0x00000036 | ||
702 | |||
703 | #define REG_A4XX_RBBM_INT_0_MASK 0x00000037 | ||
704 | |||
705 | #define REG_A4XX_RBBM_RBBM_CTL 0x0000003e | ||
706 | |||
707 | #define REG_A4XX_RBBM_AHB_DEBUG_CTL 0x0000003f | ||
708 | |||
709 | #define REG_A4XX_RBBM_VBIF_DEBUG_CTL 0x00000041 | ||
710 | |||
711 | #define REG_A4XX_RBBM_CLOCK_CTL2 0x00000042 | ||
712 | |||
713 | #define REG_A4XX_RBBM_BLOCK_SW_RESET_CMD 0x00000045 | ||
714 | |||
715 | #define REG_A4XX_RBBM_RESET_CYCLES 0x00000047 | ||
716 | |||
717 | #define REG_A4XX_RBBM_EXT_TRACE_BUS_CTL 0x00000049 | ||
718 | |||
719 | #define REG_A4XX_RBBM_CFG_DEBBUS_SEL_A 0x0000004a | ||
720 | |||
721 | #define REG_A4XX_RBBM_CFG_DEBBUS_SEL_B 0x0000004b | ||
722 | |||
723 | #define REG_A4XX_RBBM_CFG_DEBBUS_SEL_C 0x0000004c | ||
724 | |||
725 | #define REG_A4XX_RBBM_CFG_DEBBUS_SEL_D 0x0000004d | ||
726 | |||
727 | #define REG_A4XX_RBBM_PERFCTR_CP_0_LO 0x0000009c | ||
728 | |||
729 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_SP(uint32_t i0) { return 0x00000068 + 0x1*i0; } | ||
730 | |||
731 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_SP_REG(uint32_t i0) { return 0x00000068 + 0x1*i0; } | ||
732 | |||
733 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_SP(uint32_t i0) { return 0x0000006c + 0x1*i0; } | ||
734 | |||
735 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_SP_REG(uint32_t i0) { return 0x0000006c + 0x1*i0; } | ||
736 | |||
737 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_SP(uint32_t i0) { return 0x00000070 + 0x1*i0; } | ||
738 | |||
739 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_SP_REG(uint32_t i0) { return 0x00000070 + 0x1*i0; } | ||
740 | |||
741 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_SP(uint32_t i0) { return 0x00000074 + 0x1*i0; } | ||
742 | |||
743 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_SP_REG(uint32_t i0) { return 0x00000074 + 0x1*i0; } | ||
744 | |||
745 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_RB(uint32_t i0) { return 0x00000078 + 0x1*i0; } | ||
746 | |||
747 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_RB_REG(uint32_t i0) { return 0x00000078 + 0x1*i0; } | ||
748 | |||
749 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_RB(uint32_t i0) { return 0x0000007c + 0x1*i0; } | ||
750 | |||
751 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL2_RB_REG(uint32_t i0) { return 0x0000007c + 0x1*i0; } | ||
752 | |||
753 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_MARB_CCU(uint32_t i0) { return 0x00000082 + 0x1*i0; } | ||
754 | |||
755 | static inline uint32_t REG_A4XX_RBBM_CLOCK_CTL_MARB_CCU_REG(uint32_t i0) { return 0x00000082 + 0x1*i0; } | ||
756 | |||
757 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_RB_MARB_CCU(uint32_t i0) { return 0x00000086 + 0x1*i0; } | ||
758 | |||
759 | static inline uint32_t REG_A4XX_RBBM_CLOCK_HYST_RB_MARB_CCU_REG(uint32_t i0) { return 0x00000086 + 0x1*i0; } | ||
760 | |||
761 | #define REG_A4XX_RBBM_CLOCK_HYST_COM_DCOM 0x00000080 | ||
762 | |||
763 | #define REG_A4XX_RBBM_CLOCK_CTL_COM_DCOM 0x00000081 | ||
764 | |||
765 | #define REG_A4XX_RBBM_CLOCK_CTL_HLSQ 0x0000008a | ||
766 | |||
767 | #define REG_A4XX_RBBM_CLOCK_HYST_HLSQ 0x0000008b | ||
768 | |||
769 | #define REG_A4XX_RBBM_CLOCK_DELAY_HLSQ 0x0000008c | ||
770 | |||
771 | #define REG_A4XX_RBBM_CLOCK_DELAY_COM_DCOM 0x0000008d | ||
772 | |||
773 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_RB_MARB_CCU_L1(uint32_t i0) { return 0x0000008e + 0x1*i0; } | ||
774 | |||
775 | static inline uint32_t REG_A4XX_RBBM_CLOCK_DELAY_RB_MARB_CCU_L1_REG(uint32_t i0) { return 0x0000008e + 0x1*i0; } | ||
776 | |||
777 | #define REG_A4XX_RBBM_PERFCTR_PWR_1_LO 0x00000168 | ||
778 | |||
779 | #define REG_A4XX_RBBM_PERFCTR_CTL 0x00000170 | ||
780 | |||
781 | #define REG_A4XX_RBBM_PERFCTR_LOAD_CMD0 0x00000171 | ||
782 | |||
783 | #define REG_A4XX_RBBM_PERFCTR_LOAD_CMD1 0x00000172 | ||
784 | |||
785 | #define REG_A4XX_RBBM_PERFCTR_LOAD_CMD2 0x00000173 | ||
786 | |||
787 | #define REG_A4XX_RBBM_PERFCTR_LOAD_VALUE_LO 0x00000174 | ||
788 | |||
789 | #define REG_A4XX_RBBM_PERFCTR_LOAD_VALUE_HI 0x00000175 | ||
790 | |||
791 | #define REG_A4XX_RBBM_GPU_BUSY_MASKED 0x0000017a | ||
792 | |||
793 | #define REG_A4XX_RBBM_INT_0_STATUS 0x0000017d | ||
794 | |||
795 | #define REG_A4XX_RBBM_CLOCK_STATUS 0x00000182 | ||
796 | |||
797 | #define REG_A4XX_RBBM_AHB_STATUS 0x00000189 | ||
798 | |||
799 | #define REG_A4XX_RBBM_AHB_ME_SPLIT_STATUS 0x0000018c | ||
800 | |||
801 | #define REG_A4XX_RBBM_AHB_PFP_SPLIT_STATUS 0x0000018d | ||
802 | |||
803 | #define REG_A4XX_RBBM_AHB_ERROR_STATUS 0x0000018f | ||
804 | |||
805 | #define REG_A4XX_RBBM_STATUS 0x00000191 | ||
806 | #define A4XX_RBBM_STATUS_HI_BUSY 0x00000001 | ||
807 | #define A4XX_RBBM_STATUS_CP_ME_BUSY 0x00000002 | ||
808 | #define A4XX_RBBM_STATUS_CP_PFP_BUSY 0x00000004 | ||
809 | #define A4XX_RBBM_STATUS_CP_NRT_BUSY 0x00004000 | ||
810 | #define A4XX_RBBM_STATUS_VBIF_BUSY 0x00008000 | ||
811 | #define A4XX_RBBM_STATUS_TSE_BUSY 0x00010000 | ||
812 | #define A4XX_RBBM_STATUS_RAS_BUSY 0x00020000 | ||
813 | #define A4XX_RBBM_STATUS_RB_BUSY 0x00040000 | ||
814 | #define A4XX_RBBM_STATUS_PC_DCALL_BUSY 0x00080000 | ||
815 | #define A4XX_RBBM_STATUS_PC_VSD_BUSY 0x00100000 | ||
816 | #define A4XX_RBBM_STATUS_VFD_BUSY 0x00200000 | ||
817 | #define A4XX_RBBM_STATUS_VPC_BUSY 0x00400000 | ||
818 | #define A4XX_RBBM_STATUS_UCHE_BUSY 0x00800000 | ||
819 | #define A4XX_RBBM_STATUS_SP_BUSY 0x01000000 | ||
820 | #define A4XX_RBBM_STATUS_TPL1_BUSY 0x02000000 | ||
821 | #define A4XX_RBBM_STATUS_MARB_BUSY 0x04000000 | ||
822 | #define A4XX_RBBM_STATUS_VSC_BUSY 0x08000000 | ||
823 | #define A4XX_RBBM_STATUS_ARB_BUSY 0x10000000 | ||
824 | #define A4XX_RBBM_STATUS_HLSQ_BUSY 0x20000000 | ||
825 | #define A4XX_RBBM_STATUS_GPU_BUSY_NOHC 0x40000000 | ||
826 | #define A4XX_RBBM_STATUS_GPU_BUSY 0x80000000 | ||
827 | |||
828 | #define REG_A4XX_RBBM_INTERFACE_RRDY_STATUS5 0x0000019f | ||
829 | |||
830 | #define REG_A4XX_CP_SCRATCH_UMASK 0x00000228 | ||
831 | |||
832 | #define REG_A4XX_CP_SCRATCH_ADDR 0x00000229 | ||
833 | |||
834 | #define REG_A4XX_CP_RB_BASE 0x00000200 | ||
835 | |||
836 | #define REG_A4XX_CP_RB_CNTL 0x00000201 | ||
837 | |||
838 | #define REG_A4XX_CP_RB_WPTR 0x00000205 | ||
839 | |||
840 | #define REG_A4XX_CP_RB_RPTR_ADDR 0x00000203 | ||
841 | |||
842 | #define REG_A4XX_CP_RB_RPTR 0x00000204 | ||
843 | |||
844 | #define REG_A4XX_CP_IB1_BASE 0x00000206 | ||
845 | |||
846 | #define REG_A4XX_CP_IB1_BUFSZ 0x00000207 | ||
847 | |||
848 | #define REG_A4XX_CP_IB2_BASE 0x00000208 | ||
849 | |||
850 | #define REG_A4XX_CP_IB2_BUFSZ 0x00000209 | ||
851 | |||
852 | #define REG_A4XX_CP_ME_RB_DONE_DATA 0x00000217 | ||
853 | |||
854 | #define REG_A4XX_CP_QUEUE_THRESH2 0x00000219 | ||
855 | |||
856 | #define REG_A4XX_CP_MERCIU_SIZE 0x0000021b | ||
857 | |||
858 | #define REG_A4XX_CP_ROQ_ADDR 0x0000021c | ||
859 | |||
860 | #define REG_A4XX_CP_ROQ_DATA 0x0000021d | ||
861 | |||
862 | #define REG_A4XX_CP_MEQ_ADDR 0x0000021e | ||
863 | |||
864 | #define REG_A4XX_CP_MEQ_DATA 0x0000021f | ||
865 | |||
866 | #define REG_A4XX_CP_MERCIU_ADDR 0x00000220 | ||
867 | |||
868 | #define REG_A4XX_CP_MERCIU_DATA 0x00000221 | ||
869 | |||
870 | #define REG_A4XX_CP_MERCIU_DATA2 0x00000222 | ||
871 | |||
872 | #define REG_A4XX_CP_PFP_UCODE_ADDR 0x00000223 | ||
873 | |||
874 | #define REG_A4XX_CP_PFP_UCODE_DATA 0x00000224 | ||
875 | |||
876 | #define REG_A4XX_CP_ME_RAM_WADDR 0x00000225 | ||
877 | |||
878 | #define REG_A4XX_CP_ME_RAM_RADDR 0x00000226 | ||
879 | |||
880 | #define REG_A4XX_CP_ME_RAM_DATA 0x00000227 | ||
881 | |||
882 | #define REG_A4XX_CP_PREEMPT 0x0000022a | ||
883 | |||
884 | #define REG_A4XX_CP_CNTL 0x0000022c | ||
885 | |||
886 | #define REG_A4XX_CP_ME_CNTL 0x0000022d | ||
887 | |||
888 | #define REG_A4XX_CP_DEBUG 0x0000022e | ||
889 | |||
890 | #define REG_A4XX_CP_DEBUG_ECO_CONTROL 0x00000231 | ||
891 | |||
892 | #define REG_A4XX_CP_DRAW_STATE_ADDR 0x00000232 | ||
893 | |||
894 | #define REG_A4XX_CP_PROTECT_REG_0 0x00000240 | ||
895 | |||
896 | static inline uint32_t REG_A4XX_CP_PROTECT(uint32_t i0) { return 0x00000240 + 0x1*i0; } | ||
897 | |||
898 | static inline uint32_t REG_A4XX_CP_PROTECT_REG(uint32_t i0) { return 0x00000240 + 0x1*i0; } | ||
899 | |||
900 | #define REG_A4XX_CP_PROTECT_CTRL 0x00000250 | ||
901 | |||
902 | #define REG_A4XX_CP_ST_BASE 0x000004c0 | ||
903 | |||
904 | #define REG_A4XX_CP_STQ_AVAIL 0x000004ce | ||
905 | |||
906 | #define REG_A4XX_CP_MERCIU_STAT 0x000004d0 | ||
907 | |||
908 | #define REG_A4XX_CP_WFI_PEND_CTR 0x000004d2 | ||
909 | |||
910 | #define REG_A4XX_CP_HW_FAULT 0x000004d8 | ||
911 | |||
912 | #define REG_A4XX_CP_PROTECT_STATUS 0x000004da | ||
913 | |||
914 | #define REG_A4XX_CP_EVENTS_IN_FLIGHT 0x000004dd | ||
915 | |||
916 | #define REG_A4XX_CP_PERFCTR_CP_SEL_0 0x00000500 | ||
917 | |||
918 | #define REG_A4XX_CP_PERFCOMBINER_SELECT 0x0000050b | ||
919 | |||
920 | static inline uint32_t REG_A4XX_CP_SCRATCH(uint32_t i0) { return 0x00000578 + 0x1*i0; } | ||
921 | |||
922 | static inline uint32_t REG_A4XX_CP_SCRATCH_REG(uint32_t i0) { return 0x00000578 + 0x1*i0; } | ||
923 | |||
924 | #define REG_A4XX_SP_VS_STATUS 0x00000ec0 | ||
925 | |||
926 | #define REG_A4XX_SP_PERFCTR_SP_SEL_11 0x00000ecf | ||
927 | |||
928 | #define REG_A4XX_SP_SP_CTRL_REG 0x000022c0 | ||
929 | #define A4XX_SP_SP_CTRL_REG_BINNING_PASS 0x00080000 | ||
930 | |||
931 | #define REG_A4XX_SP_INSTR_CACHE_CTRL 0x000022c1 | ||
932 | |||
933 | #define REG_A4XX_SP_VS_CTRL_REG0 0x000022c4 | ||
934 | #define A4XX_SP_VS_CTRL_REG0_THREADMODE__MASK 0x00000001 | ||
935 | #define A4XX_SP_VS_CTRL_REG0_THREADMODE__SHIFT 0 | ||
936 | static inline uint32_t A4XX_SP_VS_CTRL_REG0_THREADMODE(enum a3xx_threadmode val) | ||
937 | { | ||
938 | return ((val) << A4XX_SP_VS_CTRL_REG0_THREADMODE__SHIFT) & A4XX_SP_VS_CTRL_REG0_THREADMODE__MASK; | ||
939 | } | ||
940 | #define A4XX_SP_VS_CTRL_REG0_VARYING 0x00000002 | ||
941 | #define A4XX_SP_VS_CTRL_REG0_CACHEINVALID 0x00000004 | ||
942 | #define A4XX_SP_VS_CTRL_REG0_HALFREGFOOTPRINT__MASK 0x000003f0 | ||
943 | #define A4XX_SP_VS_CTRL_REG0_HALFREGFOOTPRINT__SHIFT 4 | ||
944 | static inline uint32_t A4XX_SP_VS_CTRL_REG0_HALFREGFOOTPRINT(uint32_t val) | ||
945 | { | ||
946 | return ((val) << A4XX_SP_VS_CTRL_REG0_HALFREGFOOTPRINT__SHIFT) & A4XX_SP_VS_CTRL_REG0_HALFREGFOOTPRINT__MASK; | ||
947 | } | ||
948 | #define A4XX_SP_VS_CTRL_REG0_FULLREGFOOTPRINT__MASK 0x0003fc00 | ||
949 | #define A4XX_SP_VS_CTRL_REG0_FULLREGFOOTPRINT__SHIFT 10 | ||
950 | static inline uint32_t A4XX_SP_VS_CTRL_REG0_FULLREGFOOTPRINT(uint32_t val) | ||
951 | { | ||
952 | return ((val) << A4XX_SP_VS_CTRL_REG0_FULLREGFOOTPRINT__SHIFT) & A4XX_SP_VS_CTRL_REG0_FULLREGFOOTPRINT__MASK; | ||
953 | } | ||
954 | #define A4XX_SP_VS_CTRL_REG0_INOUTREGOVERLAP__MASK 0x000c0000 | ||
955 | #define A4XX_SP_VS_CTRL_REG0_INOUTREGOVERLAP__SHIFT 18 | ||
956 | static inline uint32_t A4XX_SP_VS_CTRL_REG0_INOUTREGOVERLAP(uint32_t val) | ||
957 | { | ||
958 | return ((val) << A4XX_SP_VS_CTRL_REG0_INOUTREGOVERLAP__SHIFT) & A4XX_SP_VS_CTRL_REG0_INOUTREGOVERLAP__MASK; | ||
959 | } | ||
960 | #define A4XX_SP_VS_CTRL_REG0_THREADSIZE__MASK 0x00100000 | ||
961 | #define A4XX_SP_VS_CTRL_REG0_THREADSIZE__SHIFT 20 | ||
962 | static inline uint32_t A4XX_SP_VS_CTRL_REG0_THREADSIZE(enum a3xx_threadsize val) | ||
963 | { | ||
964 | return ((val) << A4XX_SP_VS_CTRL_REG0_THREADSIZE__SHIFT) & A4XX_SP_VS_CTRL_REG0_THREADSIZE__MASK; | ||
965 | } | ||
966 | #define A4XX_SP_VS_CTRL_REG0_SUPERTHREADMODE 0x00200000 | ||
967 | #define A4XX_SP_VS_CTRL_REG0_PIXLODENABLE 0x00400000 | ||
968 | |||
969 | #define REG_A4XX_SP_VS_CTRL_REG1 0x000022c5 | ||
970 | #define A4XX_SP_VS_CTRL_REG1_CONSTLENGTH__MASK 0x000000ff | ||
971 | #define A4XX_SP_VS_CTRL_REG1_CONSTLENGTH__SHIFT 0 | ||
972 | static inline uint32_t A4XX_SP_VS_CTRL_REG1_CONSTLENGTH(uint32_t val) | ||
973 | { | ||
974 | return ((val) << A4XX_SP_VS_CTRL_REG1_CONSTLENGTH__SHIFT) & A4XX_SP_VS_CTRL_REG1_CONSTLENGTH__MASK; | ||
975 | } | ||
976 | #define A4XX_SP_VS_CTRL_REG1_INITIALOUTSTANDING__MASK 0x7f000000 | ||
977 | #define A4XX_SP_VS_CTRL_REG1_INITIALOUTSTANDING__SHIFT 24 | ||
978 | static inline uint32_t A4XX_SP_VS_CTRL_REG1_INITIALOUTSTANDING(uint32_t val) | ||
979 | { | ||
980 | return ((val) << A4XX_SP_VS_CTRL_REG1_INITIALOUTSTANDING__SHIFT) & A4XX_SP_VS_CTRL_REG1_INITIALOUTSTANDING__MASK; | ||
981 | } | ||
982 | |||
983 | #define REG_A4XX_SP_VS_PARAM_REG 0x000022c6 | ||
984 | #define A4XX_SP_VS_PARAM_REG_POSREGID__MASK 0x000000ff | ||
985 | #define A4XX_SP_VS_PARAM_REG_POSREGID__SHIFT 0 | ||
986 | static inline uint32_t A4XX_SP_VS_PARAM_REG_POSREGID(uint32_t val) | ||
987 | { | ||
988 | return ((val) << A4XX_SP_VS_PARAM_REG_POSREGID__SHIFT) & A4XX_SP_VS_PARAM_REG_POSREGID__MASK; | ||
989 | } | ||
990 | #define A4XX_SP_VS_PARAM_REG_PSIZEREGID__MASK 0x0000ff00 | ||
991 | #define A4XX_SP_VS_PARAM_REG_PSIZEREGID__SHIFT 8 | ||
992 | static inline uint32_t A4XX_SP_VS_PARAM_REG_PSIZEREGID(uint32_t val) | ||
993 | { | ||
994 | return ((val) << A4XX_SP_VS_PARAM_REG_PSIZEREGID__SHIFT) & A4XX_SP_VS_PARAM_REG_PSIZEREGID__MASK; | ||
995 | } | ||
996 | #define A4XX_SP_VS_PARAM_REG_TOTALVSOUTVAR__MASK 0xfff00000 | ||
997 | #define A4XX_SP_VS_PARAM_REG_TOTALVSOUTVAR__SHIFT 20 | ||
998 | static inline uint32_t A4XX_SP_VS_PARAM_REG_TOTALVSOUTVAR(uint32_t val) | ||
999 | { | ||
1000 | return ((val) << A4XX_SP_VS_PARAM_REG_TOTALVSOUTVAR__SHIFT) & A4XX_SP_VS_PARAM_REG_TOTALVSOUTVAR__MASK; | ||
1001 | } | ||
1002 | |||
1003 | static inline uint32_t REG_A4XX_SP_VS_OUT(uint32_t i0) { return 0x000022c7 + 0x1*i0; } | ||
1004 | |||
1005 | static inline uint32_t REG_A4XX_SP_VS_OUT_REG(uint32_t i0) { return 0x000022c7 + 0x1*i0; } | ||
1006 | #define A4XX_SP_VS_OUT_REG_A_REGID__MASK 0x000001ff | ||
1007 | #define A4XX_SP_VS_OUT_REG_A_REGID__SHIFT 0 | ||
1008 | static inline uint32_t A4XX_SP_VS_OUT_REG_A_REGID(uint32_t val) | ||
1009 | { | ||
1010 | return ((val) << A4XX_SP_VS_OUT_REG_A_REGID__SHIFT) & A4XX_SP_VS_OUT_REG_A_REGID__MASK; | ||
1011 | } | ||
1012 | #define A4XX_SP_VS_OUT_REG_A_COMPMASK__MASK 0x00001e00 | ||
1013 | #define A4XX_SP_VS_OUT_REG_A_COMPMASK__SHIFT 9 | ||
1014 | static inline uint32_t A4XX_SP_VS_OUT_REG_A_COMPMASK(uint32_t val) | ||
1015 | { | ||
1016 | return ((val) << A4XX_SP_VS_OUT_REG_A_COMPMASK__SHIFT) & A4XX_SP_VS_OUT_REG_A_COMPMASK__MASK; | ||
1017 | } | ||
1018 | #define A4XX_SP_VS_OUT_REG_B_REGID__MASK 0x01ff0000 | ||
1019 | #define A4XX_SP_VS_OUT_REG_B_REGID__SHIFT 16 | ||
1020 | static inline uint32_t A4XX_SP_VS_OUT_REG_B_REGID(uint32_t val) | ||
1021 | { | ||
1022 | return ((val) << A4XX_SP_VS_OUT_REG_B_REGID__SHIFT) & A4XX_SP_VS_OUT_REG_B_REGID__MASK; | ||
1023 | } | ||
1024 | #define A4XX_SP_VS_OUT_REG_B_COMPMASK__MASK 0x1e000000 | ||
1025 | #define A4XX_SP_VS_OUT_REG_B_COMPMASK__SHIFT 25 | ||
1026 | static inline uint32_t A4XX_SP_VS_OUT_REG_B_COMPMASK(uint32_t val) | ||
1027 | { | ||
1028 | return ((val) << A4XX_SP_VS_OUT_REG_B_COMPMASK__SHIFT) & A4XX_SP_VS_OUT_REG_B_COMPMASK__MASK; | ||
1029 | } | ||
1030 | |||
1031 | static inline uint32_t REG_A4XX_SP_VS_VPC_DST(uint32_t i0) { return 0x000022d8 + 0x1*i0; } | ||
1032 | |||
1033 | static inline uint32_t REG_A4XX_SP_VS_VPC_DST_REG(uint32_t i0) { return 0x000022d8 + 0x1*i0; } | ||
1034 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC0__MASK 0x000000ff | ||
1035 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC0__SHIFT 0 | ||
1036 | static inline uint32_t A4XX_SP_VS_VPC_DST_REG_OUTLOC0(uint32_t val) | ||
1037 | { | ||
1038 | return ((val) << A4XX_SP_VS_VPC_DST_REG_OUTLOC0__SHIFT) & A4XX_SP_VS_VPC_DST_REG_OUTLOC0__MASK; | ||
1039 | } | ||
1040 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC1__MASK 0x0000ff00 | ||
1041 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC1__SHIFT 8 | ||
1042 | static inline uint32_t A4XX_SP_VS_VPC_DST_REG_OUTLOC1(uint32_t val) | ||
1043 | { | ||
1044 | return ((val) << A4XX_SP_VS_VPC_DST_REG_OUTLOC1__SHIFT) & A4XX_SP_VS_VPC_DST_REG_OUTLOC1__MASK; | ||
1045 | } | ||
1046 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC2__MASK 0x00ff0000 | ||
1047 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC2__SHIFT 16 | ||
1048 | static inline uint32_t A4XX_SP_VS_VPC_DST_REG_OUTLOC2(uint32_t val) | ||
1049 | { | ||
1050 | return ((val) << A4XX_SP_VS_VPC_DST_REG_OUTLOC2__SHIFT) & A4XX_SP_VS_VPC_DST_REG_OUTLOC2__MASK; | ||
1051 | } | ||
1052 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC3__MASK 0xff000000 | ||
1053 | #define A4XX_SP_VS_VPC_DST_REG_OUTLOC3__SHIFT 24 | ||
1054 | static inline uint32_t A4XX_SP_VS_VPC_DST_REG_OUTLOC3(uint32_t val) | ||
1055 | { | ||
1056 | return ((val) << A4XX_SP_VS_VPC_DST_REG_OUTLOC3__SHIFT) & A4XX_SP_VS_VPC_DST_REG_OUTLOC3__MASK; | ||
1057 | } | ||
1058 | |||
1059 | #define REG_A4XX_SP_VS_OBJ_OFFSET_REG 0x000022e0 | ||
1060 | #define A4XX_SP_VS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK 0x01ff0000 | ||
1061 | #define A4XX_SP_VS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT 16 | ||
1062 | static inline uint32_t A4XX_SP_VS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1063 | { | ||
1064 | return ((val) << A4XX_SP_VS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_SP_VS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK; | ||
1065 | } | ||
1066 | #define A4XX_SP_VS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK 0xfe000000 | ||
1067 | #define A4XX_SP_VS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT 25 | ||
1068 | static inline uint32_t A4XX_SP_VS_OBJ_OFFSET_REG_SHADEROBJOFFSET(uint32_t val) | ||
1069 | { | ||
1070 | return ((val) << A4XX_SP_VS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT) & A4XX_SP_VS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK; | ||
1071 | } | ||
1072 | |||
1073 | #define REG_A4XX_SP_VS_OBJ_START 0x000022e1 | ||
1074 | |||
1075 | #define REG_A4XX_SP_VS_PVT_MEM_PARAM 0x000022e2 | ||
1076 | |||
1077 | #define REG_A4XX_SP_VS_PVT_MEM_ADDR 0x000022e3 | ||
1078 | |||
1079 | #define REG_A4XX_SP_VS_LENGTH_REG 0x000022e5 | ||
1080 | |||
1081 | #define REG_A4XX_SP_FS_CTRL_REG0 0x000022e8 | ||
1082 | #define A4XX_SP_FS_CTRL_REG0_THREADMODE__MASK 0x00000001 | ||
1083 | #define A4XX_SP_FS_CTRL_REG0_THREADMODE__SHIFT 0 | ||
1084 | static inline uint32_t A4XX_SP_FS_CTRL_REG0_THREADMODE(enum a3xx_threadmode val) | ||
1085 | { | ||
1086 | return ((val) << A4XX_SP_FS_CTRL_REG0_THREADMODE__SHIFT) & A4XX_SP_FS_CTRL_REG0_THREADMODE__MASK; | ||
1087 | } | ||
1088 | #define A4XX_SP_FS_CTRL_REG0_VARYING 0x00000002 | ||
1089 | #define A4XX_SP_FS_CTRL_REG0_CACHEINVALID 0x00000004 | ||
1090 | #define A4XX_SP_FS_CTRL_REG0_HALFREGFOOTPRINT__MASK 0x000003f0 | ||
1091 | #define A4XX_SP_FS_CTRL_REG0_HALFREGFOOTPRINT__SHIFT 4 | ||
1092 | static inline uint32_t A4XX_SP_FS_CTRL_REG0_HALFREGFOOTPRINT(uint32_t val) | ||
1093 | { | ||
1094 | return ((val) << A4XX_SP_FS_CTRL_REG0_HALFREGFOOTPRINT__SHIFT) & A4XX_SP_FS_CTRL_REG0_HALFREGFOOTPRINT__MASK; | ||
1095 | } | ||
1096 | #define A4XX_SP_FS_CTRL_REG0_FULLREGFOOTPRINT__MASK 0x0003fc00 | ||
1097 | #define A4XX_SP_FS_CTRL_REG0_FULLREGFOOTPRINT__SHIFT 10 | ||
1098 | static inline uint32_t A4XX_SP_FS_CTRL_REG0_FULLREGFOOTPRINT(uint32_t val) | ||
1099 | { | ||
1100 | return ((val) << A4XX_SP_FS_CTRL_REG0_FULLREGFOOTPRINT__SHIFT) & A4XX_SP_FS_CTRL_REG0_FULLREGFOOTPRINT__MASK; | ||
1101 | } | ||
1102 | #define A4XX_SP_FS_CTRL_REG0_INOUTREGOVERLAP__MASK 0x000c0000 | ||
1103 | #define A4XX_SP_FS_CTRL_REG0_INOUTREGOVERLAP__SHIFT 18 | ||
1104 | static inline uint32_t A4XX_SP_FS_CTRL_REG0_INOUTREGOVERLAP(uint32_t val) | ||
1105 | { | ||
1106 | return ((val) << A4XX_SP_FS_CTRL_REG0_INOUTREGOVERLAP__SHIFT) & A4XX_SP_FS_CTRL_REG0_INOUTREGOVERLAP__MASK; | ||
1107 | } | ||
1108 | #define A4XX_SP_FS_CTRL_REG0_THREADSIZE__MASK 0x00100000 | ||
1109 | #define A4XX_SP_FS_CTRL_REG0_THREADSIZE__SHIFT 20 | ||
1110 | static inline uint32_t A4XX_SP_FS_CTRL_REG0_THREADSIZE(enum a3xx_threadsize val) | ||
1111 | { | ||
1112 | return ((val) << A4XX_SP_FS_CTRL_REG0_THREADSIZE__SHIFT) & A4XX_SP_FS_CTRL_REG0_THREADSIZE__MASK; | ||
1113 | } | ||
1114 | #define A4XX_SP_FS_CTRL_REG0_SUPERTHREADMODE 0x00200000 | ||
1115 | #define A4XX_SP_FS_CTRL_REG0_PIXLODENABLE 0x00400000 | ||
1116 | |||
1117 | #define REG_A4XX_SP_FS_CTRL_REG1 0x000022e9 | ||
1118 | #define A4XX_SP_FS_CTRL_REG1_CONSTLENGTH__MASK 0x000000ff | ||
1119 | #define A4XX_SP_FS_CTRL_REG1_CONSTLENGTH__SHIFT 0 | ||
1120 | static inline uint32_t A4XX_SP_FS_CTRL_REG1_CONSTLENGTH(uint32_t val) | ||
1121 | { | ||
1122 | return ((val) << A4XX_SP_FS_CTRL_REG1_CONSTLENGTH__SHIFT) & A4XX_SP_FS_CTRL_REG1_CONSTLENGTH__MASK; | ||
1123 | } | ||
1124 | #define A4XX_SP_FS_CTRL_REG1_VARYING 0x00100000 | ||
1125 | |||
1126 | #define REG_A4XX_SP_FS_OBJ_OFFSET_REG 0x000022ea | ||
1127 | #define A4XX_SP_FS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK 0x01ff0000 | ||
1128 | #define A4XX_SP_FS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT 16 | ||
1129 | static inline uint32_t A4XX_SP_FS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1130 | { | ||
1131 | return ((val) << A4XX_SP_FS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_SP_FS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK; | ||
1132 | } | ||
1133 | #define A4XX_SP_FS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK 0xfe000000 | ||
1134 | #define A4XX_SP_FS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT 25 | ||
1135 | static inline uint32_t A4XX_SP_FS_OBJ_OFFSET_REG_SHADEROBJOFFSET(uint32_t val) | ||
1136 | { | ||
1137 | return ((val) << A4XX_SP_FS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT) & A4XX_SP_FS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK; | ||
1138 | } | ||
1139 | |||
1140 | #define REG_A4XX_SP_FS_OBJ_START 0x000022eb | ||
1141 | |||
1142 | #define REG_A4XX_SP_FS_PVT_MEM_PARAM 0x000022ec | ||
1143 | |||
1144 | #define REG_A4XX_SP_FS_PVT_MEM_ADDR 0x000022ed | ||
1145 | |||
1146 | #define REG_A4XX_SP_FS_LENGTH_REG 0x000022ef | ||
1147 | |||
1148 | #define REG_A4XX_SP_FS_OUTPUT_REG 0x000022f0 | ||
1149 | #define A4XX_SP_FS_OUTPUT_REG_DEPTH_ENABLE 0x00000080 | ||
1150 | #define A4XX_SP_FS_OUTPUT_REG_DEPTH_REGID__MASK 0x0000ff00 | ||
1151 | #define A4XX_SP_FS_OUTPUT_REG_DEPTH_REGID__SHIFT 8 | ||
1152 | static inline uint32_t A4XX_SP_FS_OUTPUT_REG_DEPTH_REGID(uint32_t val) | ||
1153 | { | ||
1154 | return ((val) << A4XX_SP_FS_OUTPUT_REG_DEPTH_REGID__SHIFT) & A4XX_SP_FS_OUTPUT_REG_DEPTH_REGID__MASK; | ||
1155 | } | ||
1156 | |||
1157 | static inline uint32_t REG_A4XX_SP_FS_MRT(uint32_t i0) { return 0x000022f1 + 0x1*i0; } | ||
1158 | |||
1159 | static inline uint32_t REG_A4XX_SP_FS_MRT_REG(uint32_t i0) { return 0x000022f1 + 0x1*i0; } | ||
1160 | #define A4XX_SP_FS_MRT_REG_REGID__MASK 0x000000ff | ||
1161 | #define A4XX_SP_FS_MRT_REG_REGID__SHIFT 0 | ||
1162 | static inline uint32_t A4XX_SP_FS_MRT_REG_REGID(uint32_t val) | ||
1163 | { | ||
1164 | return ((val) << A4XX_SP_FS_MRT_REG_REGID__SHIFT) & A4XX_SP_FS_MRT_REG_REGID__MASK; | ||
1165 | } | ||
1166 | #define A4XX_SP_FS_MRT_REG_HALF_PRECISION 0x00000100 | ||
1167 | #define A4XX_SP_FS_MRT_REG_MRTFORMAT__MASK 0x0003f000 | ||
1168 | #define A4XX_SP_FS_MRT_REG_MRTFORMAT__SHIFT 12 | ||
1169 | static inline uint32_t A4XX_SP_FS_MRT_REG_MRTFORMAT(enum a4xx_color_fmt val) | ||
1170 | { | ||
1171 | return ((val) << A4XX_SP_FS_MRT_REG_MRTFORMAT__SHIFT) & A4XX_SP_FS_MRT_REG_MRTFORMAT__MASK; | ||
1172 | } | ||
1173 | |||
1174 | #define REG_A4XX_SP_HS_OBJ_OFFSET_REG 0x0000230d | ||
1175 | #define A4XX_SP_HS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK 0x01ff0000 | ||
1176 | #define A4XX_SP_HS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT 16 | ||
1177 | static inline uint32_t A4XX_SP_HS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1178 | { | ||
1179 | return ((val) << A4XX_SP_HS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_SP_HS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK; | ||
1180 | } | ||
1181 | #define A4XX_SP_HS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK 0xfe000000 | ||
1182 | #define A4XX_SP_HS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT 25 | ||
1183 | static inline uint32_t A4XX_SP_HS_OBJ_OFFSET_REG_SHADEROBJOFFSET(uint32_t val) | ||
1184 | { | ||
1185 | return ((val) << A4XX_SP_HS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT) & A4XX_SP_HS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK; | ||
1186 | } | ||
1187 | |||
1188 | #define REG_A4XX_SP_DS_OBJ_OFFSET_REG 0x00002334 | ||
1189 | #define A4XX_SP_DS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK 0x01ff0000 | ||
1190 | #define A4XX_SP_DS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT 16 | ||
1191 | static inline uint32_t A4XX_SP_DS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1192 | { | ||
1193 | return ((val) << A4XX_SP_DS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_SP_DS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK; | ||
1194 | } | ||
1195 | #define A4XX_SP_DS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK 0xfe000000 | ||
1196 | #define A4XX_SP_DS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT 25 | ||
1197 | static inline uint32_t A4XX_SP_DS_OBJ_OFFSET_REG_SHADEROBJOFFSET(uint32_t val) | ||
1198 | { | ||
1199 | return ((val) << A4XX_SP_DS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT) & A4XX_SP_DS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK; | ||
1200 | } | ||
1201 | |||
1202 | #define REG_A4XX_SP_GS_OBJ_OFFSET_REG 0x0000235b | ||
1203 | #define A4XX_SP_GS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK 0x01ff0000 | ||
1204 | #define A4XX_SP_GS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT 16 | ||
1205 | static inline uint32_t A4XX_SP_GS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1206 | { | ||
1207 | return ((val) << A4XX_SP_GS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_SP_GS_OBJ_OFFSET_REG_CONSTOBJECTOFFSET__MASK; | ||
1208 | } | ||
1209 | #define A4XX_SP_GS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK 0xfe000000 | ||
1210 | #define A4XX_SP_GS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT 25 | ||
1211 | static inline uint32_t A4XX_SP_GS_OBJ_OFFSET_REG_SHADEROBJOFFSET(uint32_t val) | ||
1212 | { | ||
1213 | return ((val) << A4XX_SP_GS_OBJ_OFFSET_REG_SHADEROBJOFFSET__SHIFT) & A4XX_SP_GS_OBJ_OFFSET_REG_SHADEROBJOFFSET__MASK; | ||
1214 | } | ||
1215 | |||
1216 | #define REG_A4XX_SP_GS_LENGTH_REG 0x00002360 | ||
1217 | |||
1218 | #define REG_A4XX_VPC_DEBUG_RAM_SEL 0x00000e60 | ||
1219 | |||
1220 | #define REG_A4XX_VPC_DEBUG_RAM_READ 0x00000e61 | ||
1221 | |||
1222 | #define REG_A4XX_VPC_DEBUG_ECO_CONTROL 0x00000e64 | ||
1223 | |||
1224 | #define REG_A4XX_VPC_PERFCTR_VPC_SEL_3 0x00000e68 | ||
1225 | |||
1226 | #define REG_A4XX_VPC_ATTR 0x00002140 | ||
1227 | #define A4XX_VPC_ATTR_TOTALATTR__MASK 0x000001ff | ||
1228 | #define A4XX_VPC_ATTR_TOTALATTR__SHIFT 0 | ||
1229 | static inline uint32_t A4XX_VPC_ATTR_TOTALATTR(uint32_t val) | ||
1230 | { | ||
1231 | return ((val) << A4XX_VPC_ATTR_TOTALATTR__SHIFT) & A4XX_VPC_ATTR_TOTALATTR__MASK; | ||
1232 | } | ||
1233 | #define A4XX_VPC_ATTR_PSIZE 0x00000200 | ||
1234 | #define A4XX_VPC_ATTR_THRDASSIGN__MASK 0x00003000 | ||
1235 | #define A4XX_VPC_ATTR_THRDASSIGN__SHIFT 12 | ||
1236 | static inline uint32_t A4XX_VPC_ATTR_THRDASSIGN(uint32_t val) | ||
1237 | { | ||
1238 | return ((val) << A4XX_VPC_ATTR_THRDASSIGN__SHIFT) & A4XX_VPC_ATTR_THRDASSIGN__MASK; | ||
1239 | } | ||
1240 | #define A4XX_VPC_ATTR_ENABLE 0x02000000 | ||
1241 | |||
1242 | #define REG_A4XX_VPC_PACK 0x00002141 | ||
1243 | #define A4XX_VPC_PACK_NUMBYPASSVAR__MASK 0x000000ff | ||
1244 | #define A4XX_VPC_PACK_NUMBYPASSVAR__SHIFT 0 | ||
1245 | static inline uint32_t A4XX_VPC_PACK_NUMBYPASSVAR(uint32_t val) | ||
1246 | { | ||
1247 | return ((val) << A4XX_VPC_PACK_NUMBYPASSVAR__SHIFT) & A4XX_VPC_PACK_NUMBYPASSVAR__MASK; | ||
1248 | } | ||
1249 | #define A4XX_VPC_PACK_NUMFPNONPOSVAR__MASK 0x0000ff00 | ||
1250 | #define A4XX_VPC_PACK_NUMFPNONPOSVAR__SHIFT 8 | ||
1251 | static inline uint32_t A4XX_VPC_PACK_NUMFPNONPOSVAR(uint32_t val) | ||
1252 | { | ||
1253 | return ((val) << A4XX_VPC_PACK_NUMFPNONPOSVAR__SHIFT) & A4XX_VPC_PACK_NUMFPNONPOSVAR__MASK; | ||
1254 | } | ||
1255 | #define A4XX_VPC_PACK_NUMNONPOSVSVAR__MASK 0x00ff0000 | ||
1256 | #define A4XX_VPC_PACK_NUMNONPOSVSVAR__SHIFT 16 | ||
1257 | static inline uint32_t A4XX_VPC_PACK_NUMNONPOSVSVAR(uint32_t val) | ||
1258 | { | ||
1259 | return ((val) << A4XX_VPC_PACK_NUMNONPOSVSVAR__SHIFT) & A4XX_VPC_PACK_NUMNONPOSVSVAR__MASK; | ||
1260 | } | ||
1261 | |||
1262 | static inline uint32_t REG_A4XX_VPC_VARYING_INTERP(uint32_t i0) { return 0x00002142 + 0x1*i0; } | ||
1263 | |||
1264 | static inline uint32_t REG_A4XX_VPC_VARYING_INTERP_MODE(uint32_t i0) { return 0x00002142 + 0x1*i0; } | ||
1265 | |||
1266 | static inline uint32_t REG_A4XX_VPC_VARYING_PS_REPL(uint32_t i0) { return 0x0000214a + 0x1*i0; } | ||
1267 | |||
1268 | static inline uint32_t REG_A4XX_VPC_VARYING_PS_REPL_MODE(uint32_t i0) { return 0x0000214a + 0x1*i0; } | ||
1269 | |||
1270 | #define REG_A4XX_VPC_SO_FLUSH_WADDR_3 0x0000216e | ||
1271 | |||
1272 | #define REG_A4XX_VSC_BIN_SIZE 0x00000c00 | ||
1273 | #define A4XX_VSC_BIN_SIZE_WIDTH__MASK 0x0000001f | ||
1274 | #define A4XX_VSC_BIN_SIZE_WIDTH__SHIFT 0 | ||
1275 | static inline uint32_t A4XX_VSC_BIN_SIZE_WIDTH(uint32_t val) | ||
1276 | { | ||
1277 | return ((val >> 5) << A4XX_VSC_BIN_SIZE_WIDTH__SHIFT) & A4XX_VSC_BIN_SIZE_WIDTH__MASK; | ||
1278 | } | ||
1279 | #define A4XX_VSC_BIN_SIZE_HEIGHT__MASK 0x000003e0 | ||
1280 | #define A4XX_VSC_BIN_SIZE_HEIGHT__SHIFT 5 | ||
1281 | static inline uint32_t A4XX_VSC_BIN_SIZE_HEIGHT(uint32_t val) | ||
1282 | { | ||
1283 | return ((val >> 5) << A4XX_VSC_BIN_SIZE_HEIGHT__SHIFT) & A4XX_VSC_BIN_SIZE_HEIGHT__MASK; | ||
1284 | } | ||
1285 | |||
1286 | #define REG_A4XX_VSC_SIZE_ADDRESS 0x00000c01 | ||
1287 | |||
1288 | #define REG_A4XX_VSC_SIZE_ADDRESS2 0x00000c02 | ||
1289 | |||
1290 | #define REG_A4XX_VSC_DEBUG_ECO_CONTROL 0x00000c03 | ||
1291 | |||
1292 | static inline uint32_t REG_A4XX_VSC_PIPE_CONFIG(uint32_t i0) { return 0x00000c08 + 0x1*i0; } | ||
1293 | |||
1294 | static inline uint32_t REG_A4XX_VSC_PIPE_CONFIG_REG(uint32_t i0) { return 0x00000c08 + 0x1*i0; } | ||
1295 | #define A4XX_VSC_PIPE_CONFIG_REG_X__MASK 0x000003ff | ||
1296 | #define A4XX_VSC_PIPE_CONFIG_REG_X__SHIFT 0 | ||
1297 | static inline uint32_t A4XX_VSC_PIPE_CONFIG_REG_X(uint32_t val) | ||
1298 | { | ||
1299 | return ((val) << A4XX_VSC_PIPE_CONFIG_REG_X__SHIFT) & A4XX_VSC_PIPE_CONFIG_REG_X__MASK; | ||
1300 | } | ||
1301 | #define A4XX_VSC_PIPE_CONFIG_REG_Y__MASK 0x000ffc00 | ||
1302 | #define A4XX_VSC_PIPE_CONFIG_REG_Y__SHIFT 10 | ||
1303 | static inline uint32_t A4XX_VSC_PIPE_CONFIG_REG_Y(uint32_t val) | ||
1304 | { | ||
1305 | return ((val) << A4XX_VSC_PIPE_CONFIG_REG_Y__SHIFT) & A4XX_VSC_PIPE_CONFIG_REG_Y__MASK; | ||
1306 | } | ||
1307 | #define A4XX_VSC_PIPE_CONFIG_REG_W__MASK 0x00f00000 | ||
1308 | #define A4XX_VSC_PIPE_CONFIG_REG_W__SHIFT 20 | ||
1309 | static inline uint32_t A4XX_VSC_PIPE_CONFIG_REG_W(uint32_t val) | ||
1310 | { | ||
1311 | return ((val) << A4XX_VSC_PIPE_CONFIG_REG_W__SHIFT) & A4XX_VSC_PIPE_CONFIG_REG_W__MASK; | ||
1312 | } | ||
1313 | #define A4XX_VSC_PIPE_CONFIG_REG_H__MASK 0x0f000000 | ||
1314 | #define A4XX_VSC_PIPE_CONFIG_REG_H__SHIFT 24 | ||
1315 | static inline uint32_t A4XX_VSC_PIPE_CONFIG_REG_H(uint32_t val) | ||
1316 | { | ||
1317 | return ((val) << A4XX_VSC_PIPE_CONFIG_REG_H__SHIFT) & A4XX_VSC_PIPE_CONFIG_REG_H__MASK; | ||
1318 | } | ||
1319 | |||
1320 | static inline uint32_t REG_A4XX_VSC_PIPE_DATA_ADDRESS(uint32_t i0) { return 0x00000c10 + 0x1*i0; } | ||
1321 | |||
1322 | static inline uint32_t REG_A4XX_VSC_PIPE_DATA_ADDRESS_REG(uint32_t i0) { return 0x00000c10 + 0x1*i0; } | ||
1323 | |||
1324 | static inline uint32_t REG_A4XX_VSC_PIPE_DATA_LENGTH(uint32_t i0) { return 0x00000c18 + 0x1*i0; } | ||
1325 | |||
1326 | static inline uint32_t REG_A4XX_VSC_PIPE_DATA_LENGTH_REG(uint32_t i0) { return 0x00000c18 + 0x1*i0; } | ||
1327 | |||
1328 | #define REG_A4XX_VSC_PIPE_PARTIAL_POSN_1 0x00000c41 | ||
1329 | |||
1330 | #define REG_A4XX_VSC_PERFCTR_VSC_SEL_0 0x00000c50 | ||
1331 | |||
1332 | #define REG_A4XX_VSC_PERFCTR_VSC_SEL_1 0x00000c51 | ||
1333 | |||
1334 | #define REG_A4XX_VFD_DEBUG_CONTROL 0x00000e40 | ||
1335 | |||
1336 | #define REG_A4XX_VFD_PERFCTR_VFD_SEL_7 0x00000e4a | ||
1337 | |||
1338 | #define REG_A4XX_VFD_CONTROL_0 0x00002200 | ||
1339 | #define A4XX_VFD_CONTROL_0_TOTALATTRTOVS__MASK 0x000000ff | ||
1340 | #define A4XX_VFD_CONTROL_0_TOTALATTRTOVS__SHIFT 0 | ||
1341 | static inline uint32_t A4XX_VFD_CONTROL_0_TOTALATTRTOVS(uint32_t val) | ||
1342 | { | ||
1343 | return ((val) << A4XX_VFD_CONTROL_0_TOTALATTRTOVS__SHIFT) & A4XX_VFD_CONTROL_0_TOTALATTRTOVS__MASK; | ||
1344 | } | ||
1345 | #define A4XX_VFD_CONTROL_0_BYPASSATTROVS__MASK 0x0001fe00 | ||
1346 | #define A4XX_VFD_CONTROL_0_BYPASSATTROVS__SHIFT 9 | ||
1347 | static inline uint32_t A4XX_VFD_CONTROL_0_BYPASSATTROVS(uint32_t val) | ||
1348 | { | ||
1349 | return ((val) << A4XX_VFD_CONTROL_0_BYPASSATTROVS__SHIFT) & A4XX_VFD_CONTROL_0_BYPASSATTROVS__MASK; | ||
1350 | } | ||
1351 | #define A4XX_VFD_CONTROL_0_STRMDECINSTRCNT__MASK 0x03f00000 | ||
1352 | #define A4XX_VFD_CONTROL_0_STRMDECINSTRCNT__SHIFT 20 | ||
1353 | static inline uint32_t A4XX_VFD_CONTROL_0_STRMDECINSTRCNT(uint32_t val) | ||
1354 | { | ||
1355 | return ((val) << A4XX_VFD_CONTROL_0_STRMDECINSTRCNT__SHIFT) & A4XX_VFD_CONTROL_0_STRMDECINSTRCNT__MASK; | ||
1356 | } | ||
1357 | #define A4XX_VFD_CONTROL_0_STRMFETCHINSTRCNT__MASK 0xfc000000 | ||
1358 | #define A4XX_VFD_CONTROL_0_STRMFETCHINSTRCNT__SHIFT 26 | ||
1359 | static inline uint32_t A4XX_VFD_CONTROL_0_STRMFETCHINSTRCNT(uint32_t val) | ||
1360 | { | ||
1361 | return ((val) << A4XX_VFD_CONTROL_0_STRMFETCHINSTRCNT__SHIFT) & A4XX_VFD_CONTROL_0_STRMFETCHINSTRCNT__MASK; | ||
1362 | } | ||
1363 | |||
1364 | #define REG_A4XX_VFD_CONTROL_1 0x00002201 | ||
1365 | #define A4XX_VFD_CONTROL_1_MAXSTORAGE__MASK 0x0000ffff | ||
1366 | #define A4XX_VFD_CONTROL_1_MAXSTORAGE__SHIFT 0 | ||
1367 | static inline uint32_t A4XX_VFD_CONTROL_1_MAXSTORAGE(uint32_t val) | ||
1368 | { | ||
1369 | return ((val) << A4XX_VFD_CONTROL_1_MAXSTORAGE__SHIFT) & A4XX_VFD_CONTROL_1_MAXSTORAGE__MASK; | ||
1370 | } | ||
1371 | #define A4XX_VFD_CONTROL_1_REGID4VTX__MASK 0x00ff0000 | ||
1372 | #define A4XX_VFD_CONTROL_1_REGID4VTX__SHIFT 16 | ||
1373 | static inline uint32_t A4XX_VFD_CONTROL_1_REGID4VTX(uint32_t val) | ||
1374 | { | ||
1375 | return ((val) << A4XX_VFD_CONTROL_1_REGID4VTX__SHIFT) & A4XX_VFD_CONTROL_1_REGID4VTX__MASK; | ||
1376 | } | ||
1377 | #define A4XX_VFD_CONTROL_1_REGID4INST__MASK 0xff000000 | ||
1378 | #define A4XX_VFD_CONTROL_1_REGID4INST__SHIFT 24 | ||
1379 | static inline uint32_t A4XX_VFD_CONTROL_1_REGID4INST(uint32_t val) | ||
1380 | { | ||
1381 | return ((val) << A4XX_VFD_CONTROL_1_REGID4INST__SHIFT) & A4XX_VFD_CONTROL_1_REGID4INST__MASK; | ||
1382 | } | ||
1383 | |||
1384 | #define REG_A4XX_VFD_CONTROL_2 0x00002202 | ||
1385 | |||
1386 | #define REG_A4XX_VFD_CONTROL_3 0x00002203 | ||
1387 | |||
1388 | #define REG_A4XX_VFD_CONTROL_4 0x00002204 | ||
1389 | |||
1390 | #define REG_A4XX_VFD_INDEX_OFFSET 0x00002208 | ||
1391 | |||
1392 | static inline uint32_t REG_A4XX_VFD_FETCH(uint32_t i0) { return 0x0000220a + 0x4*i0; } | ||
1393 | |||
1394 | static inline uint32_t REG_A4XX_VFD_FETCH_INSTR_0(uint32_t i0) { return 0x0000220a + 0x4*i0; } | ||
1395 | #define A4XX_VFD_FETCH_INSTR_0_FETCHSIZE__MASK 0x0000007f | ||
1396 | #define A4XX_VFD_FETCH_INSTR_0_FETCHSIZE__SHIFT 0 | ||
1397 | static inline uint32_t A4XX_VFD_FETCH_INSTR_0_FETCHSIZE(uint32_t val) | ||
1398 | { | ||
1399 | return ((val) << A4XX_VFD_FETCH_INSTR_0_FETCHSIZE__SHIFT) & A4XX_VFD_FETCH_INSTR_0_FETCHSIZE__MASK; | ||
1400 | } | ||
1401 | #define A4XX_VFD_FETCH_INSTR_0_BUFSTRIDE__MASK 0x0001ff80 | ||
1402 | #define A4XX_VFD_FETCH_INSTR_0_BUFSTRIDE__SHIFT 7 | ||
1403 | static inline uint32_t A4XX_VFD_FETCH_INSTR_0_BUFSTRIDE(uint32_t val) | ||
1404 | { | ||
1405 | return ((val) << A4XX_VFD_FETCH_INSTR_0_BUFSTRIDE__SHIFT) & A4XX_VFD_FETCH_INSTR_0_BUFSTRIDE__MASK; | ||
1406 | } | ||
1407 | #define A4XX_VFD_FETCH_INSTR_0_SWITCHNEXT 0x00080000 | ||
1408 | #define A4XX_VFD_FETCH_INSTR_0_STEPRATE__MASK 0xff000000 | ||
1409 | #define A4XX_VFD_FETCH_INSTR_0_STEPRATE__SHIFT 24 | ||
1410 | static inline uint32_t A4XX_VFD_FETCH_INSTR_0_STEPRATE(uint32_t val) | ||
1411 | { | ||
1412 | return ((val) << A4XX_VFD_FETCH_INSTR_0_STEPRATE__SHIFT) & A4XX_VFD_FETCH_INSTR_0_STEPRATE__MASK; | ||
1413 | } | ||
1414 | |||
1415 | static inline uint32_t REG_A4XX_VFD_FETCH_INSTR_1(uint32_t i0) { return 0x0000220b + 0x4*i0; } | ||
1416 | |||
1417 | static inline uint32_t REG_A4XX_VFD_FETCH_INSTR_2(uint32_t i0) { return 0x0000220c + 0x4*i0; } | ||
1418 | #define A4XX_VFD_FETCH_INSTR_2_SIZE__MASK 0xfffffff0 | ||
1419 | #define A4XX_VFD_FETCH_INSTR_2_SIZE__SHIFT 4 | ||
1420 | static inline uint32_t A4XX_VFD_FETCH_INSTR_2_SIZE(uint32_t val) | ||
1421 | { | ||
1422 | return ((val >> 4) << A4XX_VFD_FETCH_INSTR_2_SIZE__SHIFT) & A4XX_VFD_FETCH_INSTR_2_SIZE__MASK; | ||
1423 | } | ||
1424 | |||
1425 | static inline uint32_t REG_A4XX_VFD_FETCH_INSTR_3(uint32_t i0) { return 0x0000220d + 0x4*i0; } | ||
1426 | |||
1427 | static inline uint32_t REG_A4XX_VFD_DECODE(uint32_t i0) { return 0x0000228a + 0x1*i0; } | ||
1428 | |||
1429 | static inline uint32_t REG_A4XX_VFD_DECODE_INSTR(uint32_t i0) { return 0x0000228a + 0x1*i0; } | ||
1430 | #define A4XX_VFD_DECODE_INSTR_WRITEMASK__MASK 0x0000000f | ||
1431 | #define A4XX_VFD_DECODE_INSTR_WRITEMASK__SHIFT 0 | ||
1432 | static inline uint32_t A4XX_VFD_DECODE_INSTR_WRITEMASK(uint32_t val) | ||
1433 | { | ||
1434 | return ((val) << A4XX_VFD_DECODE_INSTR_WRITEMASK__SHIFT) & A4XX_VFD_DECODE_INSTR_WRITEMASK__MASK; | ||
1435 | } | ||
1436 | #define A4XX_VFD_DECODE_INSTR_CONSTFILL 0x00000010 | ||
1437 | #define A4XX_VFD_DECODE_INSTR_FORMAT__MASK 0x00000fc0 | ||
1438 | #define A4XX_VFD_DECODE_INSTR_FORMAT__SHIFT 6 | ||
1439 | static inline uint32_t A4XX_VFD_DECODE_INSTR_FORMAT(enum a4xx_vtx_fmt val) | ||
1440 | { | ||
1441 | return ((val) << A4XX_VFD_DECODE_INSTR_FORMAT__SHIFT) & A4XX_VFD_DECODE_INSTR_FORMAT__MASK; | ||
1442 | } | ||
1443 | #define A4XX_VFD_DECODE_INSTR_REGID__MASK 0x000ff000 | ||
1444 | #define A4XX_VFD_DECODE_INSTR_REGID__SHIFT 12 | ||
1445 | static inline uint32_t A4XX_VFD_DECODE_INSTR_REGID(uint32_t val) | ||
1446 | { | ||
1447 | return ((val) << A4XX_VFD_DECODE_INSTR_REGID__SHIFT) & A4XX_VFD_DECODE_INSTR_REGID__MASK; | ||
1448 | } | ||
1449 | #define A4XX_VFD_DECODE_INSTR_SWAP__MASK 0x00c00000 | ||
1450 | #define A4XX_VFD_DECODE_INSTR_SWAP__SHIFT 22 | ||
1451 | static inline uint32_t A4XX_VFD_DECODE_INSTR_SWAP(enum a3xx_color_swap val) | ||
1452 | { | ||
1453 | return ((val) << A4XX_VFD_DECODE_INSTR_SWAP__SHIFT) & A4XX_VFD_DECODE_INSTR_SWAP__MASK; | ||
1454 | } | ||
1455 | #define A4XX_VFD_DECODE_INSTR_SHIFTCNT__MASK 0x1f000000 | ||
1456 | #define A4XX_VFD_DECODE_INSTR_SHIFTCNT__SHIFT 24 | ||
1457 | static inline uint32_t A4XX_VFD_DECODE_INSTR_SHIFTCNT(uint32_t val) | ||
1458 | { | ||
1459 | return ((val) << A4XX_VFD_DECODE_INSTR_SHIFTCNT__SHIFT) & A4XX_VFD_DECODE_INSTR_SHIFTCNT__MASK; | ||
1460 | } | ||
1461 | #define A4XX_VFD_DECODE_INSTR_LASTCOMPVALID 0x20000000 | ||
1462 | #define A4XX_VFD_DECODE_INSTR_SWITCHNEXT 0x40000000 | ||
1463 | |||
1464 | #define REG_A4XX_TPL1_DEBUG_ECO_CONTROL 0x00000f00 | ||
1465 | |||
1466 | #define REG_A4XX_TPL1_PERFCTR_TP_SEL_7 0x00000f0b | ||
1467 | |||
1468 | #define REG_A4XX_TPL1_TP_TEX_OFFSET 0x00002380 | ||
1469 | |||
1470 | #define REG_A4XX_TPL1_TP_CS_TEXMEMOBJ_BASE_ADDR 0x000023a6 | ||
1471 | |||
1472 | #define REG_A4XX_GRAS_TSE_STATUS 0x00000c80 | ||
1473 | |||
1474 | #define REG_A4XX_GRAS_DEBUG_ECO_CONTROL 0x00000c81 | ||
1475 | |||
1476 | #define REG_A4XX_GRAS_PERFCTR_TSE_SEL_0 0x00000c88 | ||
1477 | |||
1478 | #define REG_A4XX_GRAS_PERFCTR_TSE_SEL_3 0x00000c8b | ||
1479 | |||
1480 | #define REG_A4XX_GRAS_CL_CLIP_CNTL 0x00002000 | ||
1481 | |||
1482 | #define REG_A4XX_GRAS_CLEAR_CNTL 0x00002003 | ||
1483 | #define A4XX_GRAS_CLEAR_CNTL_NOT_FASTCLEAR 0x00000001 | ||
1484 | |||
1485 | #define REG_A4XX_GRAS_CL_GB_CLIP_ADJ 0x00002004 | ||
1486 | #define A4XX_GRAS_CL_GB_CLIP_ADJ_HORZ__MASK 0x000003ff | ||
1487 | #define A4XX_GRAS_CL_GB_CLIP_ADJ_HORZ__SHIFT 0 | ||
1488 | static inline uint32_t A4XX_GRAS_CL_GB_CLIP_ADJ_HORZ(uint32_t val) | ||
1489 | { | ||
1490 | return ((val) << A4XX_GRAS_CL_GB_CLIP_ADJ_HORZ__SHIFT) & A4XX_GRAS_CL_GB_CLIP_ADJ_HORZ__MASK; | ||
1491 | } | ||
1492 | #define A4XX_GRAS_CL_GB_CLIP_ADJ_VERT__MASK 0x000ffc00 | ||
1493 | #define A4XX_GRAS_CL_GB_CLIP_ADJ_VERT__SHIFT 10 | ||
1494 | static inline uint32_t A4XX_GRAS_CL_GB_CLIP_ADJ_VERT(uint32_t val) | ||
1495 | { | ||
1496 | return ((val) << A4XX_GRAS_CL_GB_CLIP_ADJ_VERT__SHIFT) & A4XX_GRAS_CL_GB_CLIP_ADJ_VERT__MASK; | ||
1497 | } | ||
1498 | |||
1499 | #define REG_A4XX_GRAS_CL_VPORT_XOFFSET_0 0x00002008 | ||
1500 | #define A4XX_GRAS_CL_VPORT_XOFFSET_0__MASK 0xffffffff | ||
1501 | #define A4XX_GRAS_CL_VPORT_XOFFSET_0__SHIFT 0 | ||
1502 | static inline uint32_t A4XX_GRAS_CL_VPORT_XOFFSET_0(float val) | ||
1503 | { | ||
1504 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_XOFFSET_0__SHIFT) & A4XX_GRAS_CL_VPORT_XOFFSET_0__MASK; | ||
1505 | } | ||
1506 | |||
1507 | #define REG_A4XX_GRAS_CL_VPORT_XSCALE_0 0x00002009 | ||
1508 | #define A4XX_GRAS_CL_VPORT_XSCALE_0__MASK 0xffffffff | ||
1509 | #define A4XX_GRAS_CL_VPORT_XSCALE_0__SHIFT 0 | ||
1510 | static inline uint32_t A4XX_GRAS_CL_VPORT_XSCALE_0(float val) | ||
1511 | { | ||
1512 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_XSCALE_0__SHIFT) & A4XX_GRAS_CL_VPORT_XSCALE_0__MASK; | ||
1513 | } | ||
1514 | |||
1515 | #define REG_A4XX_GRAS_CL_VPORT_YOFFSET_0 0x0000200a | ||
1516 | #define A4XX_GRAS_CL_VPORT_YOFFSET_0__MASK 0xffffffff | ||
1517 | #define A4XX_GRAS_CL_VPORT_YOFFSET_0__SHIFT 0 | ||
1518 | static inline uint32_t A4XX_GRAS_CL_VPORT_YOFFSET_0(float val) | ||
1519 | { | ||
1520 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_YOFFSET_0__SHIFT) & A4XX_GRAS_CL_VPORT_YOFFSET_0__MASK; | ||
1521 | } | ||
1522 | |||
1523 | #define REG_A4XX_GRAS_CL_VPORT_YSCALE_0 0x0000200b | ||
1524 | #define A4XX_GRAS_CL_VPORT_YSCALE_0__MASK 0xffffffff | ||
1525 | #define A4XX_GRAS_CL_VPORT_YSCALE_0__SHIFT 0 | ||
1526 | static inline uint32_t A4XX_GRAS_CL_VPORT_YSCALE_0(float val) | ||
1527 | { | ||
1528 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_YSCALE_0__SHIFT) & A4XX_GRAS_CL_VPORT_YSCALE_0__MASK; | ||
1529 | } | ||
1530 | |||
1531 | #define REG_A4XX_GRAS_CL_VPORT_ZOFFSET_0 0x0000200c | ||
1532 | #define A4XX_GRAS_CL_VPORT_ZOFFSET_0__MASK 0xffffffff | ||
1533 | #define A4XX_GRAS_CL_VPORT_ZOFFSET_0__SHIFT 0 | ||
1534 | static inline uint32_t A4XX_GRAS_CL_VPORT_ZOFFSET_0(float val) | ||
1535 | { | ||
1536 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_ZOFFSET_0__SHIFT) & A4XX_GRAS_CL_VPORT_ZOFFSET_0__MASK; | ||
1537 | } | ||
1538 | |||
1539 | #define REG_A4XX_GRAS_CL_VPORT_ZSCALE_0 0x0000200d | ||
1540 | #define A4XX_GRAS_CL_VPORT_ZSCALE_0__MASK 0xffffffff | ||
1541 | #define A4XX_GRAS_CL_VPORT_ZSCALE_0__SHIFT 0 | ||
1542 | static inline uint32_t A4XX_GRAS_CL_VPORT_ZSCALE_0(float val) | ||
1543 | { | ||
1544 | return ((fui(val)) << A4XX_GRAS_CL_VPORT_ZSCALE_0__SHIFT) & A4XX_GRAS_CL_VPORT_ZSCALE_0__MASK; | ||
1545 | } | ||
1546 | |||
1547 | #define REG_A4XX_GRAS_SU_POINT_MINMAX 0x00002070 | ||
1548 | #define A4XX_GRAS_SU_POINT_MINMAX_MIN__MASK 0x0000ffff | ||
1549 | #define A4XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT 0 | ||
1550 | static inline uint32_t A4XX_GRAS_SU_POINT_MINMAX_MIN(float val) | ||
1551 | { | ||
1552 | return ((((uint32_t)(val * 16.0))) << A4XX_GRAS_SU_POINT_MINMAX_MIN__SHIFT) & A4XX_GRAS_SU_POINT_MINMAX_MIN__MASK; | ||
1553 | } | ||
1554 | #define A4XX_GRAS_SU_POINT_MINMAX_MAX__MASK 0xffff0000 | ||
1555 | #define A4XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT 16 | ||
1556 | static inline uint32_t A4XX_GRAS_SU_POINT_MINMAX_MAX(float val) | ||
1557 | { | ||
1558 | return ((((uint32_t)(val * 16.0))) << A4XX_GRAS_SU_POINT_MINMAX_MAX__SHIFT) & A4XX_GRAS_SU_POINT_MINMAX_MAX__MASK; | ||
1559 | } | ||
1560 | |||
1561 | #define REG_A4XX_GRAS_SU_POINT_SIZE 0x00002071 | ||
1562 | #define A4XX_GRAS_SU_POINT_SIZE__MASK 0xffffffff | ||
1563 | #define A4XX_GRAS_SU_POINT_SIZE__SHIFT 0 | ||
1564 | static inline uint32_t A4XX_GRAS_SU_POINT_SIZE(float val) | ||
1565 | { | ||
1566 | return ((((int32_t)(val * 16.0))) << A4XX_GRAS_SU_POINT_SIZE__SHIFT) & A4XX_GRAS_SU_POINT_SIZE__MASK; | ||
1567 | } | ||
1568 | |||
1569 | #define REG_A4XX_GRAS_ALPHA_CONTROL 0x00002073 | ||
1570 | #define A4XX_GRAS_ALPHA_CONTROL_ALPHA_TEST_ENABLE 0x00000004 | ||
1571 | |||
1572 | #define REG_A4XX_GRAS_SU_POLY_OFFSET_SCALE 0x00002074 | ||
1573 | #define A4XX_GRAS_SU_POLY_OFFSET_SCALE__MASK 0xffffffff | ||
1574 | #define A4XX_GRAS_SU_POLY_OFFSET_SCALE__SHIFT 0 | ||
1575 | static inline uint32_t A4XX_GRAS_SU_POLY_OFFSET_SCALE(float val) | ||
1576 | { | ||
1577 | return ((fui(val)) << A4XX_GRAS_SU_POLY_OFFSET_SCALE__SHIFT) & A4XX_GRAS_SU_POLY_OFFSET_SCALE__MASK; | ||
1578 | } | ||
1579 | |||
1580 | #define REG_A4XX_GRAS_SU_POLY_OFFSET_OFFSET 0x00002075 | ||
1581 | #define A4XX_GRAS_SU_POLY_OFFSET_OFFSET__MASK 0xffffffff | ||
1582 | #define A4XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT 0 | ||
1583 | static inline uint32_t A4XX_GRAS_SU_POLY_OFFSET_OFFSET(float val) | ||
1584 | { | ||
1585 | return ((fui(val)) << A4XX_GRAS_SU_POLY_OFFSET_OFFSET__SHIFT) & A4XX_GRAS_SU_POLY_OFFSET_OFFSET__MASK; | ||
1586 | } | ||
1587 | |||
1588 | #define REG_A4XX_GRAS_SC_EXTENT_WINDOW_TL 0x0000209f | ||
1589 | |||
1590 | #define REG_A4XX_GRAS_SC_SCREEN_SCISSOR_TL 0x0000207c | ||
1591 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_TL_WINDOW_OFFSET_DISABLE 0x80000000 | ||
1592 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_TL_X__MASK 0x00007fff | ||
1593 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_TL_X__SHIFT 0 | ||
1594 | static inline uint32_t A4XX_GRAS_SC_SCREEN_SCISSOR_TL_X(uint32_t val) | ||
1595 | { | ||
1596 | return ((val) << A4XX_GRAS_SC_SCREEN_SCISSOR_TL_X__SHIFT) & A4XX_GRAS_SC_SCREEN_SCISSOR_TL_X__MASK; | ||
1597 | } | ||
1598 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_TL_Y__MASK 0x7fff0000 | ||
1599 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_TL_Y__SHIFT 16 | ||
1600 | static inline uint32_t A4XX_GRAS_SC_SCREEN_SCISSOR_TL_Y(uint32_t val) | ||
1601 | { | ||
1602 | return ((val) << A4XX_GRAS_SC_SCREEN_SCISSOR_TL_Y__SHIFT) & A4XX_GRAS_SC_SCREEN_SCISSOR_TL_Y__MASK; | ||
1603 | } | ||
1604 | |||
1605 | #define REG_A4XX_GRAS_SC_SCREEN_SCISSOR_BR 0x0000207d | ||
1606 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_BR_WINDOW_OFFSET_DISABLE 0x80000000 | ||
1607 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_BR_X__MASK 0x00007fff | ||
1608 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_BR_X__SHIFT 0 | ||
1609 | static inline uint32_t A4XX_GRAS_SC_SCREEN_SCISSOR_BR_X(uint32_t val) | ||
1610 | { | ||
1611 | return ((val) << A4XX_GRAS_SC_SCREEN_SCISSOR_BR_X__SHIFT) & A4XX_GRAS_SC_SCREEN_SCISSOR_BR_X__MASK; | ||
1612 | } | ||
1613 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_BR_Y__MASK 0x7fff0000 | ||
1614 | #define A4XX_GRAS_SC_SCREEN_SCISSOR_BR_Y__SHIFT 16 | ||
1615 | static inline uint32_t A4XX_GRAS_SC_SCREEN_SCISSOR_BR_Y(uint32_t val) | ||
1616 | { | ||
1617 | return ((val) << A4XX_GRAS_SC_SCREEN_SCISSOR_BR_Y__SHIFT) & A4XX_GRAS_SC_SCREEN_SCISSOR_BR_Y__MASK; | ||
1618 | } | ||
1619 | |||
1620 | #define REG_A4XX_GRAS_SC_WINDOW_SCISSOR_BR 0x0000209c | ||
1621 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_BR_WINDOW_OFFSET_DISABLE 0x80000000 | ||
1622 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_BR_X__MASK 0x00007fff | ||
1623 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_BR_X__SHIFT 0 | ||
1624 | static inline uint32_t A4XX_GRAS_SC_WINDOW_SCISSOR_BR_X(uint32_t val) | ||
1625 | { | ||
1626 | return ((val) << A4XX_GRAS_SC_WINDOW_SCISSOR_BR_X__SHIFT) & A4XX_GRAS_SC_WINDOW_SCISSOR_BR_X__MASK; | ||
1627 | } | ||
1628 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_BR_Y__MASK 0x7fff0000 | ||
1629 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_BR_Y__SHIFT 16 | ||
1630 | static inline uint32_t A4XX_GRAS_SC_WINDOW_SCISSOR_BR_Y(uint32_t val) | ||
1631 | { | ||
1632 | return ((val) << A4XX_GRAS_SC_WINDOW_SCISSOR_BR_Y__SHIFT) & A4XX_GRAS_SC_WINDOW_SCISSOR_BR_Y__MASK; | ||
1633 | } | ||
1634 | |||
1635 | #define REG_A4XX_GRAS_SC_WINDOW_SCISSOR_TL 0x0000209d | ||
1636 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_TL_WINDOW_OFFSET_DISABLE 0x80000000 | ||
1637 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_TL_X__MASK 0x00007fff | ||
1638 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_TL_X__SHIFT 0 | ||
1639 | static inline uint32_t A4XX_GRAS_SC_WINDOW_SCISSOR_TL_X(uint32_t val) | ||
1640 | { | ||
1641 | return ((val) << A4XX_GRAS_SC_WINDOW_SCISSOR_TL_X__SHIFT) & A4XX_GRAS_SC_WINDOW_SCISSOR_TL_X__MASK; | ||
1642 | } | ||
1643 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_TL_Y__MASK 0x7fff0000 | ||
1644 | #define A4XX_GRAS_SC_WINDOW_SCISSOR_TL_Y__SHIFT 16 | ||
1645 | static inline uint32_t A4XX_GRAS_SC_WINDOW_SCISSOR_TL_Y(uint32_t val) | ||
1646 | { | ||
1647 | return ((val) << A4XX_GRAS_SC_WINDOW_SCISSOR_TL_Y__SHIFT) & A4XX_GRAS_SC_WINDOW_SCISSOR_TL_Y__MASK; | ||
1648 | } | ||
1649 | |||
1650 | #define REG_A4XX_GRAS_DEPTH_CONTROL 0x00002077 | ||
1651 | #define A4XX_GRAS_DEPTH_CONTROL_FORMAT__MASK 0x00000003 | ||
1652 | #define A4XX_GRAS_DEPTH_CONTROL_FORMAT__SHIFT 0 | ||
1653 | static inline uint32_t A4XX_GRAS_DEPTH_CONTROL_FORMAT(enum a4xx_depth_format val) | ||
1654 | { | ||
1655 | return ((val) << A4XX_GRAS_DEPTH_CONTROL_FORMAT__SHIFT) & A4XX_GRAS_DEPTH_CONTROL_FORMAT__MASK; | ||
1656 | } | ||
1657 | |||
1658 | #define REG_A4XX_GRAS_SU_MODE_CONTROL 0x00002078 | ||
1659 | #define A4XX_GRAS_SU_MODE_CONTROL_CULL_FRONT 0x00000001 | ||
1660 | #define A4XX_GRAS_SU_MODE_CONTROL_CULL_BACK 0x00000002 | ||
1661 | #define A4XX_GRAS_SU_MODE_CONTROL_FRONT_CW 0x00000004 | ||
1662 | #define A4XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__MASK 0x000007f8 | ||
1663 | #define A4XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT 3 | ||
1664 | static inline uint32_t A4XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH(float val) | ||
1665 | { | ||
1666 | return ((((int32_t)(val * 4.0))) << A4XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__SHIFT) & A4XX_GRAS_SU_MODE_CONTROL_LINEHALFWIDTH__MASK; | ||
1667 | } | ||
1668 | #define A4XX_GRAS_SU_MODE_CONTROL_POLY_OFFSET 0x00000800 | ||
1669 | #define A4XX_GRAS_SU_MODE_CONTROL_RENDERING_PASS 0x00100000 | ||
1670 | |||
1671 | #define REG_A4XX_GRAS_SC_CONTROL 0x0000207b | ||
1672 | #define A4XX_GRAS_SC_CONTROL_RENDER_MODE__MASK 0x0000000c | ||
1673 | #define A4XX_GRAS_SC_CONTROL_RENDER_MODE__SHIFT 2 | ||
1674 | static inline uint32_t A4XX_GRAS_SC_CONTROL_RENDER_MODE(enum a3xx_render_mode val) | ||
1675 | { | ||
1676 | return ((val) << A4XX_GRAS_SC_CONTROL_RENDER_MODE__SHIFT) & A4XX_GRAS_SC_CONTROL_RENDER_MODE__MASK; | ||
1677 | } | ||
1678 | #define A4XX_GRAS_SC_CONTROL_MSAA_SAMPLES__MASK 0x00000380 | ||
1679 | #define A4XX_GRAS_SC_CONTROL_MSAA_SAMPLES__SHIFT 7 | ||
1680 | static inline uint32_t A4XX_GRAS_SC_CONTROL_MSAA_SAMPLES(uint32_t val) | ||
1681 | { | ||
1682 | return ((val) << A4XX_GRAS_SC_CONTROL_MSAA_SAMPLES__SHIFT) & A4XX_GRAS_SC_CONTROL_MSAA_SAMPLES__MASK; | ||
1683 | } | ||
1684 | #define A4XX_GRAS_SC_CONTROL_MSAA_DISABLE 0x00000800 | ||
1685 | #define A4XX_GRAS_SC_CONTROL_RASTER_MODE__MASK 0x0000f000 | ||
1686 | #define A4XX_GRAS_SC_CONTROL_RASTER_MODE__SHIFT 12 | ||
1687 | static inline uint32_t A4XX_GRAS_SC_CONTROL_RASTER_MODE(uint32_t val) | ||
1688 | { | ||
1689 | return ((val) << A4XX_GRAS_SC_CONTROL_RASTER_MODE__SHIFT) & A4XX_GRAS_SC_CONTROL_RASTER_MODE__MASK; | ||
1690 | } | ||
1691 | |||
1692 | #define REG_A4XX_UCHE_CACHE_MODE_CONTROL 0x00000e80 | ||
1693 | |||
1694 | #define REG_A4XX_UCHE_TRAP_BASE_LO 0x00000e83 | ||
1695 | |||
1696 | #define REG_A4XX_UCHE_TRAP_BASE_HI 0x00000e84 | ||
1697 | |||
1698 | #define REG_A4XX_UCHE_CACHE_STATUS 0x00000e88 | ||
1699 | |||
1700 | #define REG_A4XX_UCHE_INVALIDATE0 0x00000e8a | ||
1701 | |||
1702 | #define REG_A4XX_UCHE_INVALIDATE1 0x00000e8b | ||
1703 | |||
1704 | #define REG_A4XX_UCHE_CACHE_WAYS_VFD 0x00000e8c | ||
1705 | |||
1706 | #define REG_A4XX_UCHE_PERFCTR_UCHE_SEL_7 0x00000e95 | ||
1707 | |||
1708 | #define REG_A4XX_HLSQ_TIMEOUT_THRESHOLD 0x00000e00 | ||
1709 | |||
1710 | #define REG_A4XX_HLSQ_DEBUG_ECO_CONTROL 0x00000e04 | ||
1711 | |||
1712 | #define REG_A4XX_HLSQ_PERF_PIPE_MASK 0x00000e0e | ||
1713 | |||
1714 | #define REG_A4XX_HLSQ_CONTROL_0_REG 0x000023c0 | ||
1715 | #define A4XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE__MASK 0x00000010 | ||
1716 | #define A4XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE__SHIFT 4 | ||
1717 | static inline uint32_t A4XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE(enum a3xx_threadsize val) | ||
1718 | { | ||
1719 | return ((val) << A4XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE__SHIFT) & A4XX_HLSQ_CONTROL_0_REG_FSTHREADSIZE__MASK; | ||
1720 | } | ||
1721 | #define A4XX_HLSQ_CONTROL_0_REG_FSSUPERTHREADENABLE 0x00000040 | ||
1722 | #define A4XX_HLSQ_CONTROL_0_REG_SPSHADERRESTART 0x00000200 | ||
1723 | #define A4XX_HLSQ_CONTROL_0_REG_RESERVED2 0x00000400 | ||
1724 | #define A4XX_HLSQ_CONTROL_0_REG_CHUNKDISABLE 0x04000000 | ||
1725 | #define A4XX_HLSQ_CONTROL_0_REG_CONSTMODE__MASK 0x08000000 | ||
1726 | #define A4XX_HLSQ_CONTROL_0_REG_CONSTMODE__SHIFT 27 | ||
1727 | static inline uint32_t A4XX_HLSQ_CONTROL_0_REG_CONSTMODE(uint32_t val) | ||
1728 | { | ||
1729 | return ((val) << A4XX_HLSQ_CONTROL_0_REG_CONSTMODE__SHIFT) & A4XX_HLSQ_CONTROL_0_REG_CONSTMODE__MASK; | ||
1730 | } | ||
1731 | #define A4XX_HLSQ_CONTROL_0_REG_LAZYUPDATEDISABLE 0x10000000 | ||
1732 | #define A4XX_HLSQ_CONTROL_0_REG_SPCONSTFULLUPDATE 0x20000000 | ||
1733 | #define A4XX_HLSQ_CONTROL_0_REG_TPFULLUPDATE 0x40000000 | ||
1734 | #define A4XX_HLSQ_CONTROL_0_REG_SINGLECONTEXT 0x80000000 | ||
1735 | |||
1736 | #define REG_A4XX_HLSQ_CONTROL_1_REG 0x000023c1 | ||
1737 | #define A4XX_HLSQ_CONTROL_1_REG_VSTHREADSIZE__MASK 0x00000040 | ||
1738 | #define A4XX_HLSQ_CONTROL_1_REG_VSTHREADSIZE__SHIFT 6 | ||
1739 | static inline uint32_t A4XX_HLSQ_CONTROL_1_REG_VSTHREADSIZE(enum a3xx_threadsize val) | ||
1740 | { | ||
1741 | return ((val) << A4XX_HLSQ_CONTROL_1_REG_VSTHREADSIZE__SHIFT) & A4XX_HLSQ_CONTROL_1_REG_VSTHREADSIZE__MASK; | ||
1742 | } | ||
1743 | #define A4XX_HLSQ_CONTROL_1_REG_VSSUPERTHREADENABLE 0x00000100 | ||
1744 | #define A4XX_HLSQ_CONTROL_1_REG_RESERVED1 0x00000200 | ||
1745 | #define A4XX_HLSQ_CONTROL_1_REG_ZWCOORD 0x02000000 | ||
1746 | |||
1747 | #define REG_A4XX_HLSQ_CONTROL_2_REG 0x000023c2 | ||
1748 | #define A4XX_HLSQ_CONTROL_2_REG_PRIMALLOCTHRESHOLD__MASK 0xfc000000 | ||
1749 | #define A4XX_HLSQ_CONTROL_2_REG_PRIMALLOCTHRESHOLD__SHIFT 26 | ||
1750 | static inline uint32_t A4XX_HLSQ_CONTROL_2_REG_PRIMALLOCTHRESHOLD(uint32_t val) | ||
1751 | { | ||
1752 | return ((val) << A4XX_HLSQ_CONTROL_2_REG_PRIMALLOCTHRESHOLD__SHIFT) & A4XX_HLSQ_CONTROL_2_REG_PRIMALLOCTHRESHOLD__MASK; | ||
1753 | } | ||
1754 | |||
1755 | #define REG_A4XX_HLSQ_CONTROL_3_REG 0x000023c3 | ||
1756 | #define A4XX_HLSQ_CONTROL_3_REG_REGID__MASK 0x000000ff | ||
1757 | #define A4XX_HLSQ_CONTROL_3_REG_REGID__SHIFT 0 | ||
1758 | static inline uint32_t A4XX_HLSQ_CONTROL_3_REG_REGID(uint32_t val) | ||
1759 | { | ||
1760 | return ((val) << A4XX_HLSQ_CONTROL_3_REG_REGID__SHIFT) & A4XX_HLSQ_CONTROL_3_REG_REGID__MASK; | ||
1761 | } | ||
1762 | |||
1763 | #define REG_A4XX_HLSQ_VS_CONTROL_REG 0x000023c5 | ||
1764 | #define A4XX_HLSQ_VS_CONTROL_REG_CONSTLENGTH__MASK 0x000000ff | ||
1765 | #define A4XX_HLSQ_VS_CONTROL_REG_CONSTLENGTH__SHIFT 0 | ||
1766 | static inline uint32_t A4XX_HLSQ_VS_CONTROL_REG_CONSTLENGTH(uint32_t val) | ||
1767 | { | ||
1768 | return ((val) << A4XX_HLSQ_VS_CONTROL_REG_CONSTLENGTH__SHIFT) & A4XX_HLSQ_VS_CONTROL_REG_CONSTLENGTH__MASK; | ||
1769 | } | ||
1770 | #define A4XX_HLSQ_VS_CONTROL_REG_CONSTOBJECTOFFSET__MASK 0x0000ff00 | ||
1771 | #define A4XX_HLSQ_VS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT 8 | ||
1772 | static inline uint32_t A4XX_HLSQ_VS_CONTROL_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1773 | { | ||
1774 | return ((val) << A4XX_HLSQ_VS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_HLSQ_VS_CONTROL_REG_CONSTOBJECTOFFSET__MASK; | ||
1775 | } | ||
1776 | #define A4XX_HLSQ_VS_CONTROL_REG_SHADEROBJOFFSET__MASK 0x00fe0000 | ||
1777 | #define A4XX_HLSQ_VS_CONTROL_REG_SHADEROBJOFFSET__SHIFT 17 | ||
1778 | static inline uint32_t A4XX_HLSQ_VS_CONTROL_REG_SHADEROBJOFFSET(uint32_t val) | ||
1779 | { | ||
1780 | return ((val) << A4XX_HLSQ_VS_CONTROL_REG_SHADEROBJOFFSET__SHIFT) & A4XX_HLSQ_VS_CONTROL_REG_SHADEROBJOFFSET__MASK; | ||
1781 | } | ||
1782 | #define A4XX_HLSQ_VS_CONTROL_REG_INSTRLENGTH__MASK 0xff000000 | ||
1783 | #define A4XX_HLSQ_VS_CONTROL_REG_INSTRLENGTH__SHIFT 24 | ||
1784 | static inline uint32_t A4XX_HLSQ_VS_CONTROL_REG_INSTRLENGTH(uint32_t val) | ||
1785 | { | ||
1786 | return ((val) << A4XX_HLSQ_VS_CONTROL_REG_INSTRLENGTH__SHIFT) & A4XX_HLSQ_VS_CONTROL_REG_INSTRLENGTH__MASK; | ||
1787 | } | ||
1788 | |||
1789 | #define REG_A4XX_HLSQ_FS_CONTROL_REG 0x000023c6 | ||
1790 | #define A4XX_HLSQ_FS_CONTROL_REG_CONSTLENGTH__MASK 0x000000ff | ||
1791 | #define A4XX_HLSQ_FS_CONTROL_REG_CONSTLENGTH__SHIFT 0 | ||
1792 | static inline uint32_t A4XX_HLSQ_FS_CONTROL_REG_CONSTLENGTH(uint32_t val) | ||
1793 | { | ||
1794 | return ((val) << A4XX_HLSQ_FS_CONTROL_REG_CONSTLENGTH__SHIFT) & A4XX_HLSQ_FS_CONTROL_REG_CONSTLENGTH__MASK; | ||
1795 | } | ||
1796 | #define A4XX_HLSQ_FS_CONTROL_REG_CONSTOBJECTOFFSET__MASK 0x0000ff00 | ||
1797 | #define A4XX_HLSQ_FS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT 8 | ||
1798 | static inline uint32_t A4XX_HLSQ_FS_CONTROL_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1799 | { | ||
1800 | return ((val) << A4XX_HLSQ_FS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_HLSQ_FS_CONTROL_REG_CONSTOBJECTOFFSET__MASK; | ||
1801 | } | ||
1802 | #define A4XX_HLSQ_FS_CONTROL_REG_SHADEROBJOFFSET__MASK 0x00fe0000 | ||
1803 | #define A4XX_HLSQ_FS_CONTROL_REG_SHADEROBJOFFSET__SHIFT 17 | ||
1804 | static inline uint32_t A4XX_HLSQ_FS_CONTROL_REG_SHADEROBJOFFSET(uint32_t val) | ||
1805 | { | ||
1806 | return ((val) << A4XX_HLSQ_FS_CONTROL_REG_SHADEROBJOFFSET__SHIFT) & A4XX_HLSQ_FS_CONTROL_REG_SHADEROBJOFFSET__MASK; | ||
1807 | } | ||
1808 | #define A4XX_HLSQ_FS_CONTROL_REG_INSTRLENGTH__MASK 0xff000000 | ||
1809 | #define A4XX_HLSQ_FS_CONTROL_REG_INSTRLENGTH__SHIFT 24 | ||
1810 | static inline uint32_t A4XX_HLSQ_FS_CONTROL_REG_INSTRLENGTH(uint32_t val) | ||
1811 | { | ||
1812 | return ((val) << A4XX_HLSQ_FS_CONTROL_REG_INSTRLENGTH__SHIFT) & A4XX_HLSQ_FS_CONTROL_REG_INSTRLENGTH__MASK; | ||
1813 | } | ||
1814 | |||
1815 | #define REG_A4XX_HLSQ_HS_CONTROL_REG 0x000023c7 | ||
1816 | #define A4XX_HLSQ_HS_CONTROL_REG_CONSTLENGTH__MASK 0x000000ff | ||
1817 | #define A4XX_HLSQ_HS_CONTROL_REG_CONSTLENGTH__SHIFT 0 | ||
1818 | static inline uint32_t A4XX_HLSQ_HS_CONTROL_REG_CONSTLENGTH(uint32_t val) | ||
1819 | { | ||
1820 | return ((val) << A4XX_HLSQ_HS_CONTROL_REG_CONSTLENGTH__SHIFT) & A4XX_HLSQ_HS_CONTROL_REG_CONSTLENGTH__MASK; | ||
1821 | } | ||
1822 | #define A4XX_HLSQ_HS_CONTROL_REG_CONSTOBJECTOFFSET__MASK 0x0000ff00 | ||
1823 | #define A4XX_HLSQ_HS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT 8 | ||
1824 | static inline uint32_t A4XX_HLSQ_HS_CONTROL_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1825 | { | ||
1826 | return ((val) << A4XX_HLSQ_HS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_HLSQ_HS_CONTROL_REG_CONSTOBJECTOFFSET__MASK; | ||
1827 | } | ||
1828 | #define A4XX_HLSQ_HS_CONTROL_REG_SHADEROBJOFFSET__MASK 0x00fe0000 | ||
1829 | #define A4XX_HLSQ_HS_CONTROL_REG_SHADEROBJOFFSET__SHIFT 17 | ||
1830 | static inline uint32_t A4XX_HLSQ_HS_CONTROL_REG_SHADEROBJOFFSET(uint32_t val) | ||
1831 | { | ||
1832 | return ((val) << A4XX_HLSQ_HS_CONTROL_REG_SHADEROBJOFFSET__SHIFT) & A4XX_HLSQ_HS_CONTROL_REG_SHADEROBJOFFSET__MASK; | ||
1833 | } | ||
1834 | #define A4XX_HLSQ_HS_CONTROL_REG_INSTRLENGTH__MASK 0xff000000 | ||
1835 | #define A4XX_HLSQ_HS_CONTROL_REG_INSTRLENGTH__SHIFT 24 | ||
1836 | static inline uint32_t A4XX_HLSQ_HS_CONTROL_REG_INSTRLENGTH(uint32_t val) | ||
1837 | { | ||
1838 | return ((val) << A4XX_HLSQ_HS_CONTROL_REG_INSTRLENGTH__SHIFT) & A4XX_HLSQ_HS_CONTROL_REG_INSTRLENGTH__MASK; | ||
1839 | } | ||
1840 | |||
1841 | #define REG_A4XX_HLSQ_DS_CONTROL_REG 0x000023c8 | ||
1842 | #define A4XX_HLSQ_DS_CONTROL_REG_CONSTLENGTH__MASK 0x000000ff | ||
1843 | #define A4XX_HLSQ_DS_CONTROL_REG_CONSTLENGTH__SHIFT 0 | ||
1844 | static inline uint32_t A4XX_HLSQ_DS_CONTROL_REG_CONSTLENGTH(uint32_t val) | ||
1845 | { | ||
1846 | return ((val) << A4XX_HLSQ_DS_CONTROL_REG_CONSTLENGTH__SHIFT) & A4XX_HLSQ_DS_CONTROL_REG_CONSTLENGTH__MASK; | ||
1847 | } | ||
1848 | #define A4XX_HLSQ_DS_CONTROL_REG_CONSTOBJECTOFFSET__MASK 0x0000ff00 | ||
1849 | #define A4XX_HLSQ_DS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT 8 | ||
1850 | static inline uint32_t A4XX_HLSQ_DS_CONTROL_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1851 | { | ||
1852 | return ((val) << A4XX_HLSQ_DS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_HLSQ_DS_CONTROL_REG_CONSTOBJECTOFFSET__MASK; | ||
1853 | } | ||
1854 | #define A4XX_HLSQ_DS_CONTROL_REG_SHADEROBJOFFSET__MASK 0x00fe0000 | ||
1855 | #define A4XX_HLSQ_DS_CONTROL_REG_SHADEROBJOFFSET__SHIFT 17 | ||
1856 | static inline uint32_t A4XX_HLSQ_DS_CONTROL_REG_SHADEROBJOFFSET(uint32_t val) | ||
1857 | { | ||
1858 | return ((val) << A4XX_HLSQ_DS_CONTROL_REG_SHADEROBJOFFSET__SHIFT) & A4XX_HLSQ_DS_CONTROL_REG_SHADEROBJOFFSET__MASK; | ||
1859 | } | ||
1860 | #define A4XX_HLSQ_DS_CONTROL_REG_INSTRLENGTH__MASK 0xff000000 | ||
1861 | #define A4XX_HLSQ_DS_CONTROL_REG_INSTRLENGTH__SHIFT 24 | ||
1862 | static inline uint32_t A4XX_HLSQ_DS_CONTROL_REG_INSTRLENGTH(uint32_t val) | ||
1863 | { | ||
1864 | return ((val) << A4XX_HLSQ_DS_CONTROL_REG_INSTRLENGTH__SHIFT) & A4XX_HLSQ_DS_CONTROL_REG_INSTRLENGTH__MASK; | ||
1865 | } | ||
1866 | |||
1867 | #define REG_A4XX_HLSQ_GS_CONTROL_REG 0x000023c9 | ||
1868 | #define A4XX_HLSQ_GS_CONTROL_REG_CONSTLENGTH__MASK 0x000000ff | ||
1869 | #define A4XX_HLSQ_GS_CONTROL_REG_CONSTLENGTH__SHIFT 0 | ||
1870 | static inline uint32_t A4XX_HLSQ_GS_CONTROL_REG_CONSTLENGTH(uint32_t val) | ||
1871 | { | ||
1872 | return ((val) << A4XX_HLSQ_GS_CONTROL_REG_CONSTLENGTH__SHIFT) & A4XX_HLSQ_GS_CONTROL_REG_CONSTLENGTH__MASK; | ||
1873 | } | ||
1874 | #define A4XX_HLSQ_GS_CONTROL_REG_CONSTOBJECTOFFSET__MASK 0x0000ff00 | ||
1875 | #define A4XX_HLSQ_GS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT 8 | ||
1876 | static inline uint32_t A4XX_HLSQ_GS_CONTROL_REG_CONSTOBJECTOFFSET(uint32_t val) | ||
1877 | { | ||
1878 | return ((val) << A4XX_HLSQ_GS_CONTROL_REG_CONSTOBJECTOFFSET__SHIFT) & A4XX_HLSQ_GS_CONTROL_REG_CONSTOBJECTOFFSET__MASK; | ||
1879 | } | ||
1880 | #define A4XX_HLSQ_GS_CONTROL_REG_SHADEROBJOFFSET__MASK 0x00fe0000 | ||
1881 | #define A4XX_HLSQ_GS_CONTROL_REG_SHADEROBJOFFSET__SHIFT 17 | ||
1882 | static inline uint32_t A4XX_HLSQ_GS_CONTROL_REG_SHADEROBJOFFSET(uint32_t val) | ||
1883 | { | ||
1884 | return ((val) << A4XX_HLSQ_GS_CONTROL_REG_SHADEROBJOFFSET__SHIFT) & A4XX_HLSQ_GS_CONTROL_REG_SHADEROBJOFFSET__MASK; | ||
1885 | } | ||
1886 | #define A4XX_HLSQ_GS_CONTROL_REG_INSTRLENGTH__MASK 0xff000000 | ||
1887 | #define A4XX_HLSQ_GS_CONTROL_REG_INSTRLENGTH__SHIFT 24 | ||
1888 | static inline uint32_t A4XX_HLSQ_GS_CONTROL_REG_INSTRLENGTH(uint32_t val) | ||
1889 | { | ||
1890 | return ((val) << A4XX_HLSQ_GS_CONTROL_REG_INSTRLENGTH__SHIFT) & A4XX_HLSQ_GS_CONTROL_REG_INSTRLENGTH__MASK; | ||
1891 | } | ||
1892 | |||
1893 | #define REG_A4XX_HLSQ_UPDATE_CONTROL 0x000023db | ||
1894 | |||
1895 | #define REG_A4XX_PC_BINNING_COMMAND 0x00000d00 | ||
1896 | #define A4XX_PC_BINNING_COMMAND_BINNING_ENABLE 0x00000001 | ||
1897 | |||
1898 | #define REG_A4XX_PC_DRAWCALL_SETUP_OVERRIDE 0x00000d0c | ||
1899 | |||
1900 | #define REG_A4XX_PC_PERFCTR_PC_SEL_0 0x00000d10 | ||
1901 | |||
1902 | #define REG_A4XX_PC_PERFCTR_PC_SEL_7 0x00000d17 | ||
1903 | |||
1904 | #define REG_A4XX_PC_BIN_BASE 0x000021c0 | ||
1905 | |||
1906 | #define REG_A4XX_PC_PRIM_VTX_CNTL 0x000021c4 | ||
1907 | #define A4XX_PC_PRIM_VTX_CNTL_VAROUT 0x00000001 | ||
1908 | #define A4XX_PC_PRIM_VTX_CNTL_PROVOKING_VTX_LAST 0x02000000 | ||
1909 | #define A4XX_PC_PRIM_VTX_CNTL_PSIZE 0x04000000 | ||
1910 | |||
1911 | #define REG_A4XX_UNKNOWN_21C5 0x000021c5 | ||
1912 | |||
1913 | #define REG_A4XX_PC_RESTART_INDEX 0x000021c6 | ||
1914 | |||
1915 | #define REG_A4XX_PC_GS_PARAM 0x000021e5 | ||
1916 | |||
1917 | #define REG_A4XX_PC_HS_PARAM 0x000021e7 | ||
1918 | |||
1919 | #define REG_A4XX_VBIF_VERSION 0x00003000 | ||
1920 | |||
1921 | #define REG_A4XX_VBIF_CLKON 0x00003001 | ||
1922 | #define A4XX_VBIF_CLKON_FORCE_ON_TESTBUS 0x00000001 | ||
1923 | |||
1924 | #define REG_A4XX_VBIF_ABIT_SORT 0x0000301c | ||
1925 | |||
1926 | #define REG_A4XX_VBIF_ABIT_SORT_CONF 0x0000301d | ||
1927 | |||
1928 | #define REG_A4XX_VBIF_GATE_OFF_WRREQ_EN 0x0000302a | ||
1929 | |||
1930 | #define REG_A4XX_VBIF_IN_RD_LIM_CONF0 0x0000302c | ||
1931 | |||
1932 | #define REG_A4XX_VBIF_IN_RD_LIM_CONF1 0x0000302d | ||
1933 | |||
1934 | #define REG_A4XX_VBIF_IN_WR_LIM_CONF0 0x00003030 | ||
1935 | |||
1936 | #define REG_A4XX_VBIF_IN_WR_LIM_CONF1 0x00003031 | ||
1937 | |||
1938 | #define REG_A4XX_VBIF_ROUND_ROBIN_QOS_ARB 0x00003049 | ||
1939 | |||
1940 | #define REG_A4XX_UNKNOWN_0CC5 0x00000cc5 | ||
1941 | |||
1942 | #define REG_A4XX_UNKNOWN_0CC6 0x00000cc6 | ||
1943 | |||
1944 | #define REG_A4XX_UNKNOWN_0D01 0x00000d01 | ||
1945 | |||
1946 | #define REG_A4XX_UNKNOWN_0E05 0x00000e05 | ||
1947 | |||
1948 | #define REG_A4XX_UNKNOWN_0E42 0x00000e42 | ||
1949 | |||
1950 | #define REG_A4XX_UNKNOWN_0EC2 0x00000ec2 | ||
1951 | |||
1952 | #define REG_A4XX_UNKNOWN_0EC3 0x00000ec3 | ||
1953 | |||
1954 | #define REG_A4XX_UNKNOWN_0F03 0x00000f03 | ||
1955 | |||
1956 | #define REG_A4XX_UNKNOWN_2001 0x00002001 | ||
1957 | |||
1958 | #define REG_A4XX_UNKNOWN_209B 0x0000209b | ||
1959 | |||
1960 | #define REG_A4XX_UNKNOWN_20EF 0x000020ef | ||
1961 | |||
1962 | #define REG_A4XX_UNKNOWN_20F0 0x000020f0 | ||
1963 | |||
1964 | #define REG_A4XX_UNKNOWN_20F1 0x000020f1 | ||
1965 | |||
1966 | #define REG_A4XX_UNKNOWN_20F2 0x000020f2 | ||
1967 | |||
1968 | #define REG_A4XX_UNKNOWN_20F3 0x000020f3 | ||
1969 | |||
1970 | #define REG_A4XX_UNKNOWN_20F4 0x000020f4 | ||
1971 | |||
1972 | #define REG_A4XX_UNKNOWN_20F5 0x000020f5 | ||
1973 | |||
1974 | #define REG_A4XX_UNKNOWN_20F6 0x000020f6 | ||
1975 | |||
1976 | #define REG_A4XX_UNKNOWN_20F7 0x000020f7 | ||
1977 | |||
1978 | #define REG_A4XX_UNKNOWN_2152 0x00002152 | ||
1979 | |||
1980 | #define REG_A4XX_UNKNOWN_2153 0x00002153 | ||
1981 | |||
1982 | #define REG_A4XX_UNKNOWN_2154 0x00002154 | ||
1983 | |||
1984 | #define REG_A4XX_UNKNOWN_2155 0x00002155 | ||
1985 | |||
1986 | #define REG_A4XX_UNKNOWN_2156 0x00002156 | ||
1987 | |||
1988 | #define REG_A4XX_UNKNOWN_2157 0x00002157 | ||
1989 | |||
1990 | #define REG_A4XX_UNKNOWN_21C3 0x000021c3 | ||
1991 | |||
1992 | #define REG_A4XX_UNKNOWN_21E6 0x000021e6 | ||
1993 | |||
1994 | #define REG_A4XX_UNKNOWN_2209 0x00002209 | ||
1995 | |||
1996 | #define REG_A4XX_UNKNOWN_22D7 0x000022d7 | ||
1997 | |||
1998 | #define REG_A4XX_UNKNOWN_2381 0x00002381 | ||
1999 | |||
2000 | #define REG_A4XX_UNKNOWN_23A0 0x000023a0 | ||
2001 | |||
2002 | #define REG_A4XX_TEX_SAMP_0 0x00000000 | ||
2003 | #define A4XX_TEX_SAMP_0_XY_MAG__MASK 0x00000006 | ||
2004 | #define A4XX_TEX_SAMP_0_XY_MAG__SHIFT 1 | ||
2005 | static inline uint32_t A4XX_TEX_SAMP_0_XY_MAG(enum a4xx_tex_filter val) | ||
2006 | { | ||
2007 | return ((val) << A4XX_TEX_SAMP_0_XY_MAG__SHIFT) & A4XX_TEX_SAMP_0_XY_MAG__MASK; | ||
2008 | } | ||
2009 | #define A4XX_TEX_SAMP_0_XY_MIN__MASK 0x00000018 | ||
2010 | #define A4XX_TEX_SAMP_0_XY_MIN__SHIFT 3 | ||
2011 | static inline uint32_t A4XX_TEX_SAMP_0_XY_MIN(enum a4xx_tex_filter val) | ||
2012 | { | ||
2013 | return ((val) << A4XX_TEX_SAMP_0_XY_MIN__SHIFT) & A4XX_TEX_SAMP_0_XY_MIN__MASK; | ||
2014 | } | ||
2015 | #define A4XX_TEX_SAMP_0_WRAP_S__MASK 0x000000e0 | ||
2016 | #define A4XX_TEX_SAMP_0_WRAP_S__SHIFT 5 | ||
2017 | static inline uint32_t A4XX_TEX_SAMP_0_WRAP_S(enum a4xx_tex_clamp val) | ||
2018 | { | ||
2019 | return ((val) << A4XX_TEX_SAMP_0_WRAP_S__SHIFT) & A4XX_TEX_SAMP_0_WRAP_S__MASK; | ||
2020 | } | ||
2021 | #define A4XX_TEX_SAMP_0_WRAP_T__MASK 0x00000700 | ||
2022 | #define A4XX_TEX_SAMP_0_WRAP_T__SHIFT 8 | ||
2023 | static inline uint32_t A4XX_TEX_SAMP_0_WRAP_T(enum a4xx_tex_clamp val) | ||
2024 | { | ||
2025 | return ((val) << A4XX_TEX_SAMP_0_WRAP_T__SHIFT) & A4XX_TEX_SAMP_0_WRAP_T__MASK; | ||
2026 | } | ||
2027 | #define A4XX_TEX_SAMP_0_WRAP_R__MASK 0x00003800 | ||
2028 | #define A4XX_TEX_SAMP_0_WRAP_R__SHIFT 11 | ||
2029 | static inline uint32_t A4XX_TEX_SAMP_0_WRAP_R(enum a4xx_tex_clamp val) | ||
2030 | { | ||
2031 | return ((val) << A4XX_TEX_SAMP_0_WRAP_R__SHIFT) & A4XX_TEX_SAMP_0_WRAP_R__MASK; | ||
2032 | } | ||
2033 | |||
2034 | #define REG_A4XX_TEX_SAMP_1 0x00000001 | ||
2035 | #define A4XX_TEX_SAMP_1_COMPARE_FUNC__MASK 0x0000000e | ||
2036 | #define A4XX_TEX_SAMP_1_COMPARE_FUNC__SHIFT 1 | ||
2037 | static inline uint32_t A4XX_TEX_SAMP_1_COMPARE_FUNC(enum adreno_compare_func val) | ||
2038 | { | ||
2039 | return ((val) << A4XX_TEX_SAMP_1_COMPARE_FUNC__SHIFT) & A4XX_TEX_SAMP_1_COMPARE_FUNC__MASK; | ||
2040 | } | ||
2041 | #define A4XX_TEX_SAMP_1_MAX_LOD__MASK 0x000fff00 | ||
2042 | #define A4XX_TEX_SAMP_1_MAX_LOD__SHIFT 8 | ||
2043 | static inline uint32_t A4XX_TEX_SAMP_1_MAX_LOD(float val) | ||
2044 | { | ||
2045 | return ((((uint32_t)(val * 64.0))) << A4XX_TEX_SAMP_1_MAX_LOD__SHIFT) & A4XX_TEX_SAMP_1_MAX_LOD__MASK; | ||
2046 | } | ||
2047 | #define A4XX_TEX_SAMP_1_MIN_LOD__MASK 0xfff00000 | ||
2048 | #define A4XX_TEX_SAMP_1_MIN_LOD__SHIFT 20 | ||
2049 | static inline uint32_t A4XX_TEX_SAMP_1_MIN_LOD(float val) | ||
2050 | { | ||
2051 | return ((((uint32_t)(val * 64.0))) << A4XX_TEX_SAMP_1_MIN_LOD__SHIFT) & A4XX_TEX_SAMP_1_MIN_LOD__MASK; | ||
2052 | } | ||
2053 | |||
2054 | #define REG_A4XX_TEX_CONST_0 0x00000000 | ||
2055 | #define A4XX_TEX_CONST_0_TILED 0x00000001 | ||
2056 | #define A4XX_TEX_CONST_0_SWIZ_X__MASK 0x00000070 | ||
2057 | #define A4XX_TEX_CONST_0_SWIZ_X__SHIFT 4 | ||
2058 | static inline uint32_t A4XX_TEX_CONST_0_SWIZ_X(enum a4xx_tex_swiz val) | ||
2059 | { | ||
2060 | return ((val) << A4XX_TEX_CONST_0_SWIZ_X__SHIFT) & A4XX_TEX_CONST_0_SWIZ_X__MASK; | ||
2061 | } | ||
2062 | #define A4XX_TEX_CONST_0_SWIZ_Y__MASK 0x00000380 | ||
2063 | #define A4XX_TEX_CONST_0_SWIZ_Y__SHIFT 7 | ||
2064 | static inline uint32_t A4XX_TEX_CONST_0_SWIZ_Y(enum a4xx_tex_swiz val) | ||
2065 | { | ||
2066 | return ((val) << A4XX_TEX_CONST_0_SWIZ_Y__SHIFT) & A4XX_TEX_CONST_0_SWIZ_Y__MASK; | ||
2067 | } | ||
2068 | #define A4XX_TEX_CONST_0_SWIZ_Z__MASK 0x00001c00 | ||
2069 | #define A4XX_TEX_CONST_0_SWIZ_Z__SHIFT 10 | ||
2070 | static inline uint32_t A4XX_TEX_CONST_0_SWIZ_Z(enum a4xx_tex_swiz val) | ||
2071 | { | ||
2072 | return ((val) << A4XX_TEX_CONST_0_SWIZ_Z__SHIFT) & A4XX_TEX_CONST_0_SWIZ_Z__MASK; | ||
2073 | } | ||
2074 | #define A4XX_TEX_CONST_0_SWIZ_W__MASK 0x0000e000 | ||
2075 | #define A4XX_TEX_CONST_0_SWIZ_W__SHIFT 13 | ||
2076 | static inline uint32_t A4XX_TEX_CONST_0_SWIZ_W(enum a4xx_tex_swiz val) | ||
2077 | { | ||
2078 | return ((val) << A4XX_TEX_CONST_0_SWIZ_W__SHIFT) & A4XX_TEX_CONST_0_SWIZ_W__MASK; | ||
2079 | } | ||
2080 | #define A4XX_TEX_CONST_0_FMT__MASK 0x1fc00000 | ||
2081 | #define A4XX_TEX_CONST_0_FMT__SHIFT 22 | ||
2082 | static inline uint32_t A4XX_TEX_CONST_0_FMT(enum a4xx_tex_fmt val) | ||
2083 | { | ||
2084 | return ((val) << A4XX_TEX_CONST_0_FMT__SHIFT) & A4XX_TEX_CONST_0_FMT__MASK; | ||
2085 | } | ||
2086 | #define A4XX_TEX_CONST_0_TYPE__MASK 0x60000000 | ||
2087 | #define A4XX_TEX_CONST_0_TYPE__SHIFT 29 | ||
2088 | static inline uint32_t A4XX_TEX_CONST_0_TYPE(enum a4xx_tex_type val) | ||
2089 | { | ||
2090 | return ((val) << A4XX_TEX_CONST_0_TYPE__SHIFT) & A4XX_TEX_CONST_0_TYPE__MASK; | ||
2091 | } | ||
2092 | |||
2093 | #define REG_A4XX_TEX_CONST_1 0x00000001 | ||
2094 | #define A4XX_TEX_CONST_1_HEIGHT__MASK 0x00007fff | ||
2095 | #define A4XX_TEX_CONST_1_HEIGHT__SHIFT 0 | ||
2096 | static inline uint32_t A4XX_TEX_CONST_1_HEIGHT(uint32_t val) | ||
2097 | { | ||
2098 | return ((val) << A4XX_TEX_CONST_1_HEIGHT__SHIFT) & A4XX_TEX_CONST_1_HEIGHT__MASK; | ||
2099 | } | ||
2100 | #define A4XX_TEX_CONST_1_WIDTH__MASK 0x1fff8000 | ||
2101 | #define A4XX_TEX_CONST_1_WIDTH__SHIFT 15 | ||
2102 | static inline uint32_t A4XX_TEX_CONST_1_WIDTH(uint32_t val) | ||
2103 | { | ||
2104 | return ((val) << A4XX_TEX_CONST_1_WIDTH__SHIFT) & A4XX_TEX_CONST_1_WIDTH__MASK; | ||
2105 | } | ||
2106 | |||
2107 | #define REG_A4XX_TEX_CONST_2 0x00000002 | ||
2108 | #define A4XX_TEX_CONST_2_PITCH__MASK 0x3ffffe00 | ||
2109 | #define A4XX_TEX_CONST_2_PITCH__SHIFT 9 | ||
2110 | static inline uint32_t A4XX_TEX_CONST_2_PITCH(uint32_t val) | ||
2111 | { | ||
2112 | return ((val) << A4XX_TEX_CONST_2_PITCH__SHIFT) & A4XX_TEX_CONST_2_PITCH__MASK; | ||
2113 | } | ||
2114 | #define A4XX_TEX_CONST_2_SWAP__MASK 0xc0000000 | ||
2115 | #define A4XX_TEX_CONST_2_SWAP__SHIFT 30 | ||
2116 | static inline uint32_t A4XX_TEX_CONST_2_SWAP(enum a3xx_color_swap val) | ||
2117 | { | ||
2118 | return ((val) << A4XX_TEX_CONST_2_SWAP__SHIFT) & A4XX_TEX_CONST_2_SWAP__MASK; | ||
2119 | } | ||
2120 | |||
2121 | #define REG_A4XX_TEX_CONST_3 0x00000003 | ||
2122 | #define A4XX_TEX_CONST_3_LAYERSZ__MASK 0x0000000f | ||
2123 | #define A4XX_TEX_CONST_3_LAYERSZ__SHIFT 0 | ||
2124 | static inline uint32_t A4XX_TEX_CONST_3_LAYERSZ(uint32_t val) | ||
2125 | { | ||
2126 | return ((val >> 12) << A4XX_TEX_CONST_3_LAYERSZ__SHIFT) & A4XX_TEX_CONST_3_LAYERSZ__MASK; | ||
2127 | } | ||
2128 | |||
2129 | #define REG_A4XX_TEX_CONST_4 0x00000004 | ||
2130 | #define A4XX_TEX_CONST_4_BASE__MASK 0xffffffff | ||
2131 | #define A4XX_TEX_CONST_4_BASE__SHIFT 0 | ||
2132 | static inline uint32_t A4XX_TEX_CONST_4_BASE(uint32_t val) | ||
2133 | { | ||
2134 | return ((val) << A4XX_TEX_CONST_4_BASE__SHIFT) & A4XX_TEX_CONST_4_BASE__MASK; | ||
2135 | } | ||
2136 | |||
2137 | #define REG_A4XX_TEX_CONST_5 0x00000005 | ||
2138 | |||
2139 | #define REG_A4XX_TEX_CONST_6 0x00000006 | ||
2140 | |||
2141 | #define REG_A4XX_TEX_CONST_7 0x00000007 | ||
2142 | |||
2143 | |||
2144 | #endif /* A4XX_XML */ | ||
diff --git a/drivers/gpu/drm/msm/adreno/a4xx_gpu.c b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c new file mode 100644 index 000000000000..91221836c5ad --- /dev/null +++ b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c | |||
@@ -0,0 +1,604 @@ | |||
1 | /* Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
2 | * | ||
3 | * This program is free software; you can redistribute it and/or modify | ||
4 | * it under the terms of the GNU General Public License version 2 and | ||
5 | * only version 2 as published by the Free Software Foundation. | ||
6 | * | ||
7 | * This program is distributed in the hope that it will be useful, | ||
8 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
9 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
10 | * GNU General Public License for more details. | ||
11 | * | ||
12 | */ | ||
13 | #include "a4xx_gpu.h" | ||
14 | #ifdef CONFIG_MSM_OCMEM | ||
15 | # include <soc/qcom/ocmem.h> | ||
16 | #endif | ||
17 | |||
18 | #define A4XX_INT0_MASK \ | ||
19 | (A4XX_INT0_RBBM_AHB_ERROR | \ | ||
20 | A4XX_INT0_RBBM_ATB_BUS_OVERFLOW | \ | ||
21 | A4XX_INT0_CP_T0_PACKET_IN_IB | \ | ||
22 | A4XX_INT0_CP_OPCODE_ERROR | \ | ||
23 | A4XX_INT0_CP_RESERVED_BIT_ERROR | \ | ||
24 | A4XX_INT0_CP_HW_FAULT | \ | ||
25 | A4XX_INT0_CP_IB1_INT | \ | ||
26 | A4XX_INT0_CP_IB2_INT | \ | ||
27 | A4XX_INT0_CP_RB_INT | \ | ||
28 | A4XX_INT0_CP_REG_PROTECT_FAULT | \ | ||
29 | A4XX_INT0_CP_AHB_ERROR_HALT | \ | ||
30 | A4XX_INT0_UCHE_OOB_ACCESS) | ||
31 | |||
32 | extern bool hang_debug; | ||
33 | static void a4xx_dump(struct msm_gpu *gpu); | ||
34 | |||
35 | /* | ||
36 | * a4xx_enable_hwcg() - Program the clock control registers | ||
37 | * @device: The adreno device pointer | ||
38 | */ | ||
39 | static void a4xx_enable_hwcg(struct msm_gpu *gpu) | ||
40 | { | ||
41 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | ||
42 | unsigned int i; | ||
43 | for (i = 0; i < 4; i++) | ||
44 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_TP(i), 0x02222202); | ||
45 | for (i = 0; i < 4; i++) | ||
46 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2_TP(i), 0x00002222); | ||
47 | for (i = 0; i < 4; i++) | ||
48 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_TP(i), 0x0E739CE7); | ||
49 | for (i = 0; i < 4; i++) | ||
50 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_TP(i), 0x00111111); | ||
51 | for (i = 0; i < 4; i++) | ||
52 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_SP(i), 0x22222222); | ||
53 | for (i = 0; i < 4; i++) | ||
54 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2_SP(i), 0x00222222); | ||
55 | for (i = 0; i < 4; i++) | ||
56 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_SP(i), 0x00000104); | ||
57 | for (i = 0; i < 4; i++) | ||
58 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_SP(i), 0x00000081); | ||
59 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_UCHE, 0x22222222); | ||
60 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2_UCHE, 0x02222222); | ||
61 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL3_UCHE, 0x00000000); | ||
62 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL4_UCHE, 0x00000000); | ||
63 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_UCHE, 0x00004444); | ||
64 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_UCHE, 0x00001112); | ||
65 | for (i = 0; i < 4; i++) | ||
66 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_RB(i), 0x22222222); | ||
67 | |||
68 | /* Disable L1 clocking in A420 due to CCU issues with it */ | ||
69 | for (i = 0; i < 4; i++) { | ||
70 | if (adreno_is_a420(adreno_gpu)) { | ||
71 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2_RB(i), | ||
72 | 0x00002020); | ||
73 | } else { | ||
74 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2_RB(i), | ||
75 | 0x00022020); | ||
76 | } | ||
77 | } | ||
78 | |||
79 | for (i = 0; i < 4; i++) { | ||
80 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_MARB_CCU(i), | ||
81 | 0x00000922); | ||
82 | } | ||
83 | |||
84 | for (i = 0; i < 4; i++) { | ||
85 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_RB_MARB_CCU(i), | ||
86 | 0x00000000); | ||
87 | } | ||
88 | |||
89 | for (i = 0; i < 4; i++) { | ||
90 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_RB_MARB_CCU_L1(i), | ||
91 | 0x00000001); | ||
92 | } | ||
93 | |||
94 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_MODE_GPC, 0x02222222); | ||
95 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_GPC, 0x04100104); | ||
96 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_GPC, 0x00022222); | ||
97 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_COM_DCOM, 0x00000022); | ||
98 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_COM_DCOM, 0x0000010F); | ||
99 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_COM_DCOM, 0x00000022); | ||
100 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_TSE_RAS_RBBM, 0x00222222); | ||
101 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00004104); | ||
102 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00000222); | ||
103 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL_HLSQ , 0x00000000); | ||
104 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_HYST_HLSQ, 0x00000000); | ||
105 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_HLSQ, 0x00020000); | ||
106 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL, 0xAAAAAAAA); | ||
107 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_CTL2, 0); | ||
108 | } | ||
109 | |||
110 | static void a4xx_me_init(struct msm_gpu *gpu) | ||
111 | { | ||
112 | struct msm_ringbuffer *ring = gpu->rb; | ||
113 | |||
114 | OUT_PKT3(ring, CP_ME_INIT, 17); | ||
115 | OUT_RING(ring, 0x000003f7); | ||
116 | OUT_RING(ring, 0x00000000); | ||
117 | OUT_RING(ring, 0x00000000); | ||
118 | OUT_RING(ring, 0x00000000); | ||
119 | OUT_RING(ring, 0x00000080); | ||
120 | OUT_RING(ring, 0x00000100); | ||
121 | OUT_RING(ring, 0x00000180); | ||
122 | OUT_RING(ring, 0x00006600); | ||
123 | OUT_RING(ring, 0x00000150); | ||
124 | OUT_RING(ring, 0x0000014e); | ||
125 | OUT_RING(ring, 0x00000154); | ||
126 | OUT_RING(ring, 0x00000001); | ||
127 | OUT_RING(ring, 0x00000000); | ||
128 | OUT_RING(ring, 0x00000000); | ||
129 | OUT_RING(ring, 0x00000000); | ||
130 | OUT_RING(ring, 0x00000000); | ||
131 | OUT_RING(ring, 0x00000000); | ||
132 | |||
133 | gpu->funcs->flush(gpu); | ||
134 | gpu->funcs->idle(gpu); | ||
135 | } | ||
136 | |||
137 | static int a4xx_hw_init(struct msm_gpu *gpu) | ||
138 | { | ||
139 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | ||
140 | struct a4xx_gpu *a4xx_gpu = to_a4xx_gpu(adreno_gpu); | ||
141 | uint32_t *ptr, len; | ||
142 | int i, ret; | ||
143 | |||
144 | if (adreno_is_a4xx(adreno_gpu)) { | ||
145 | gpu_write(gpu, REG_A4XX_VBIF_ABIT_SORT, 0x0001001F); | ||
146 | gpu_write(gpu, REG_A4XX_VBIF_ABIT_SORT_CONF, 0x000000A4); | ||
147 | gpu_write(gpu, REG_A4XX_VBIF_GATE_OFF_WRREQ_EN, 0x00000001); | ||
148 | gpu_write(gpu, REG_A4XX_VBIF_IN_RD_LIM_CONF0, 0x18181818); | ||
149 | gpu_write(gpu, REG_A4XX_VBIF_IN_RD_LIM_CONF1, 0x00000018); | ||
150 | gpu_write(gpu, REG_A4XX_VBIF_IN_WR_LIM_CONF0, 0x18181818); | ||
151 | gpu_write(gpu, REG_A4XX_VBIF_IN_WR_LIM_CONF1, 0x00000018); | ||
152 | gpu_write(gpu, REG_A4XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x00000003); | ||
153 | } else { | ||
154 | BUG(); | ||
155 | } | ||
156 | |||
157 | /* Make all blocks contribute to the GPU BUSY perf counter */ | ||
158 | gpu_write(gpu, REG_A4XX_RBBM_GPU_BUSY_MASKED, 0xffffffff); | ||
159 | |||
160 | /* Tune the hystersis counters for SP and CP idle detection */ | ||
161 | gpu_write(gpu, REG_A4XX_RBBM_SP_HYST_CNT, 0x10); | ||
162 | gpu_write(gpu, REG_A4XX_RBBM_WAIT_IDLE_CLOCKS_CTL, 0x10); | ||
163 | |||
164 | /* Enable the RBBM error reporting bits */ | ||
165 | gpu_write(gpu, REG_A4XX_RBBM_AHB_CTL0, 0x00000001); | ||
166 | |||
167 | /* Enable AHB error reporting*/ | ||
168 | gpu_write(gpu, REG_A4XX_RBBM_AHB_CTL1, 0xa6ffffff); | ||
169 | |||
170 | /* Enable power counters*/ | ||
171 | gpu_write(gpu, REG_A4XX_RBBM_RBBM_CTL, 0x00000030); | ||
172 | |||
173 | /* | ||
174 | * Turn on hang detection - this spews a lot of useful information | ||
175 | * into the RBBM registers on a hang: | ||
176 | */ | ||
177 | gpu_write(gpu, REG_A4XX_RBBM_INTERFACE_HANG_INT_CTL, | ||
178 | (1 << 30) | 0xFFFF); | ||
179 | |||
180 | gpu_write(gpu, REG_A4XX_RB_GMEM_BASE_ADDR, | ||
181 | (unsigned int)(a4xx_gpu->ocmem_base >> 14)); | ||
182 | |||
183 | /* Turn on performance counters: */ | ||
184 | gpu_write(gpu, REG_A4XX_RBBM_PERFCTR_CTL, 0x01); | ||
185 | |||
186 | /* Disable L2 bypass to avoid UCHE out of bounds errors */ | ||
187 | gpu_write(gpu, REG_A4XX_UCHE_TRAP_BASE_LO, 0xffff0000); | ||
188 | gpu_write(gpu, REG_A4XX_UCHE_TRAP_BASE_HI, 0xffff0000); | ||
189 | |||
190 | gpu_write(gpu, REG_A4XX_CP_DEBUG, (1 << 25) | | ||
191 | (adreno_is_a420(adreno_gpu) ? (1 << 29) : 0)); | ||
192 | |||
193 | a4xx_enable_hwcg(gpu); | ||
194 | |||
195 | /* | ||
196 | * For A420 set RBBM_CLOCK_DELAY_HLSQ.CGC_HLSQ_TP_EARLY_CYC >= 2 | ||
197 | * due to timing issue with HLSQ_TP_CLK_EN | ||
198 | */ | ||
199 | if (adreno_is_a420(adreno_gpu)) { | ||
200 | unsigned int val; | ||
201 | val = gpu_read(gpu, REG_A4XX_RBBM_CLOCK_DELAY_HLSQ); | ||
202 | val &= ~A4XX_CGC_HLSQ_EARLY_CYC__MASK; | ||
203 | val |= 2 << A4XX_CGC_HLSQ_EARLY_CYC__SHIFT; | ||
204 | gpu_write(gpu, REG_A4XX_RBBM_CLOCK_DELAY_HLSQ, val); | ||
205 | } | ||
206 | |||
207 | ret = adreno_hw_init(gpu); | ||
208 | if (ret) | ||
209 | return ret; | ||
210 | |||
211 | /* setup access protection: */ | ||
212 | gpu_write(gpu, REG_A4XX_CP_PROTECT_CTRL, 0x00000007); | ||
213 | |||
214 | /* RBBM registers */ | ||
215 | gpu_write(gpu, REG_A4XX_CP_PROTECT(0), 0x62000010); | ||
216 | gpu_write(gpu, REG_A4XX_CP_PROTECT(1), 0x63000020); | ||
217 | gpu_write(gpu, REG_A4XX_CP_PROTECT(2), 0x64000040); | ||
218 | gpu_write(gpu, REG_A4XX_CP_PROTECT(3), 0x65000080); | ||
219 | gpu_write(gpu, REG_A4XX_CP_PROTECT(4), 0x66000100); | ||
220 | gpu_write(gpu, REG_A4XX_CP_PROTECT(5), 0x64000200); | ||
221 | |||
222 | /* CP registers */ | ||
223 | gpu_write(gpu, REG_A4XX_CP_PROTECT(6), 0x67000800); | ||
224 | gpu_write(gpu, REG_A4XX_CP_PROTECT(7), 0x64001600); | ||
225 | |||
226 | |||
227 | /* RB registers */ | ||
228 | gpu_write(gpu, REG_A4XX_CP_PROTECT(8), 0x60003300); | ||
229 | |||
230 | /* HLSQ registers */ | ||
231 | gpu_write(gpu, REG_A4XX_CP_PROTECT(9), 0x60003800); | ||
232 | |||
233 | /* VPC registers */ | ||
234 | gpu_write(gpu, REG_A4XX_CP_PROTECT(10), 0x61003980); | ||
235 | |||
236 | /* SMMU registers */ | ||
237 | gpu_write(gpu, REG_A4XX_CP_PROTECT(11), 0x6e010000); | ||
238 | |||
239 | gpu_write(gpu, REG_A4XX_RBBM_INT_0_MASK, A4XX_INT0_MASK); | ||
240 | |||
241 | ret = adreno_hw_init(gpu); | ||
242 | if (ret) | ||
243 | return ret; | ||
244 | |||
245 | /* Load PM4: */ | ||
246 | ptr = (uint32_t *)(adreno_gpu->pm4->data); | ||
247 | len = adreno_gpu->pm4->size / 4; | ||
248 | DBG("loading PM4 ucode version: %u", ptr[0]); | ||
249 | gpu_write(gpu, REG_A4XX_CP_ME_RAM_WADDR, 0); | ||
250 | for (i = 1; i < len; i++) | ||
251 | gpu_write(gpu, REG_A4XX_CP_ME_RAM_DATA, ptr[i]); | ||
252 | |||
253 | /* Load PFP: */ | ||
254 | ptr = (uint32_t *)(adreno_gpu->pfp->data); | ||
255 | len = adreno_gpu->pfp->size / 4; | ||
256 | DBG("loading PFP ucode version: %u", ptr[0]); | ||
257 | |||
258 | gpu_write(gpu, REG_A4XX_CP_PFP_UCODE_ADDR, 0); | ||
259 | for (i = 1; i < len; i++) | ||
260 | gpu_write(gpu, REG_A4XX_CP_PFP_UCODE_DATA, ptr[i]); | ||
261 | |||
262 | /* clear ME_HALT to start micro engine */ | ||
263 | gpu_write(gpu, REG_A4XX_CP_ME_CNTL, 0); | ||
264 | |||
265 | a4xx_me_init(gpu); | ||
266 | return 0; | ||
267 | } | ||
268 | |||
269 | static void a4xx_recover(struct msm_gpu *gpu) | ||
270 | { | ||
271 | /* dump registers before resetting gpu, if enabled: */ | ||
272 | if (hang_debug) | ||
273 | a4xx_dump(gpu); | ||
274 | |||
275 | gpu_write(gpu, REG_A4XX_RBBM_SW_RESET_CMD, 1); | ||
276 | gpu_read(gpu, REG_A4XX_RBBM_SW_RESET_CMD); | ||
277 | gpu_write(gpu, REG_A4XX_RBBM_SW_RESET_CMD, 0); | ||
278 | adreno_recover(gpu); | ||
279 | } | ||
280 | |||
281 | static void a4xx_destroy(struct msm_gpu *gpu) | ||
282 | { | ||
283 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | ||
284 | struct a4xx_gpu *a4xx_gpu = to_a4xx_gpu(adreno_gpu); | ||
285 | |||
286 | DBG("%s", gpu->name); | ||
287 | |||
288 | adreno_gpu_cleanup(adreno_gpu); | ||
289 | |||
290 | #ifdef CONFIG_MSM_OCMEM | ||
291 | if (a4xx_gpu->ocmem_base) | ||
292 | ocmem_free(OCMEM_GRAPHICS, a4xx_gpu->ocmem_hdl); | ||
293 | #endif | ||
294 | |||
295 | kfree(a4xx_gpu); | ||
296 | } | ||
297 | |||
298 | static void a4xx_idle(struct msm_gpu *gpu) | ||
299 | { | ||
300 | /* wait for ringbuffer to drain: */ | ||
301 | adreno_idle(gpu); | ||
302 | |||
303 | /* then wait for GPU to finish: */ | ||
304 | if (spin_until(!(gpu_read(gpu, REG_A4XX_RBBM_STATUS) & | ||
305 | A4XX_RBBM_STATUS_GPU_BUSY))) | ||
306 | DRM_ERROR("%s: timeout waiting for GPU to idle!\n", gpu->name); | ||
307 | |||
308 | /* TODO maybe we need to reset GPU here to recover from hang? */ | ||
309 | } | ||
310 | |||
311 | static irqreturn_t a4xx_irq(struct msm_gpu *gpu) | ||
312 | { | ||
313 | uint32_t status; | ||
314 | |||
315 | status = gpu_read(gpu, REG_A4XX_RBBM_INT_0_STATUS); | ||
316 | DBG("%s: Int status %08x", gpu->name, status); | ||
317 | |||
318 | gpu_write(gpu, REG_A4XX_RBBM_INT_CLEAR_CMD, status); | ||
319 | |||
320 | msm_gpu_retire(gpu); | ||
321 | |||
322 | return IRQ_HANDLED; | ||
323 | } | ||
324 | |||
325 | static const unsigned int a4xx_registers[] = { | ||
326 | /* RBBM */ | ||
327 | 0x0000, 0x0002, 0x0004, 0x0021, 0x0023, 0x0024, 0x0026, 0x0026, | ||
328 | 0x0028, 0x002B, 0x002E, 0x0034, 0x0037, 0x0044, 0x0047, 0x0066, | ||
329 | 0x0068, 0x0095, 0x009C, 0x0170, 0x0174, 0x01AF, | ||
330 | /* CP */ | ||
331 | 0x0200, 0x0233, 0x0240, 0x0250, 0x04C0, 0x04DD, 0x0500, 0x050B, | ||
332 | 0x0578, 0x058F, | ||
333 | /* VSC */ | ||
334 | 0x0C00, 0x0C03, 0x0C08, 0x0C41, 0x0C50, 0x0C51, | ||
335 | /* GRAS */ | ||
336 | 0x0C80, 0x0C81, 0x0C88, 0x0C8F, | ||
337 | /* RB */ | ||
338 | 0x0CC0, 0x0CC0, 0x0CC4, 0x0CD2, | ||
339 | /* PC */ | ||
340 | 0x0D00, 0x0D0C, 0x0D10, 0x0D17, 0x0D20, 0x0D23, | ||
341 | /* VFD */ | ||
342 | 0x0E40, 0x0E4A, | ||
343 | /* VPC */ | ||
344 | 0x0E60, 0x0E61, 0x0E63, 0x0E68, | ||
345 | /* UCHE */ | ||
346 | 0x0E80, 0x0E84, 0x0E88, 0x0E95, | ||
347 | /* VMIDMT */ | ||
348 | 0x1000, 0x1000, 0x1002, 0x1002, 0x1004, 0x1004, 0x1008, 0x100A, | ||
349 | 0x100C, 0x100D, 0x100F, 0x1010, 0x1012, 0x1016, 0x1024, 0x1024, | ||
350 | 0x1027, 0x1027, 0x1100, 0x1100, 0x1102, 0x1102, 0x1104, 0x1104, | ||
351 | 0x1110, 0x1110, 0x1112, 0x1116, 0x1124, 0x1124, 0x1300, 0x1300, | ||
352 | 0x1380, 0x1380, | ||
353 | /* GRAS CTX 0 */ | ||
354 | 0x2000, 0x2004, 0x2008, 0x2067, 0x2070, 0x2078, 0x207B, 0x216E, | ||
355 | /* PC CTX 0 */ | ||
356 | 0x21C0, 0x21C6, 0x21D0, 0x21D0, 0x21D9, 0x21D9, 0x21E5, 0x21E7, | ||
357 | /* VFD CTX 0 */ | ||
358 | 0x2200, 0x2204, 0x2208, 0x22A9, | ||
359 | /* GRAS CTX 1 */ | ||
360 | 0x2400, 0x2404, 0x2408, 0x2467, 0x2470, 0x2478, 0x247B, 0x256E, | ||
361 | /* PC CTX 1 */ | ||
362 | 0x25C0, 0x25C6, 0x25D0, 0x25D0, 0x25D9, 0x25D9, 0x25E5, 0x25E7, | ||
363 | /* VFD CTX 1 */ | ||
364 | 0x2600, 0x2604, 0x2608, 0x26A9, | ||
365 | /* XPU */ | ||
366 | 0x2C00, 0x2C01, 0x2C10, 0x2C10, 0x2C12, 0x2C16, 0x2C1D, 0x2C20, | ||
367 | 0x2C28, 0x2C28, 0x2C30, 0x2C30, 0x2C32, 0x2C36, 0x2C40, 0x2C40, | ||
368 | 0x2C50, 0x2C50, 0x2C52, 0x2C56, 0x2C80, 0x2C80, 0x2C94, 0x2C95, | ||
369 | /* VBIF */ | ||
370 | 0x3000, 0x3007, 0x300C, 0x3014, 0x3018, 0x301D, 0x3020, 0x3022, | ||
371 | 0x3024, 0x3026, 0x3028, 0x302A, 0x302C, 0x302D, 0x3030, 0x3031, | ||
372 | 0x3034, 0x3036, 0x3038, 0x3038, 0x303C, 0x303D, 0x3040, 0x3040, | ||
373 | 0x3049, 0x3049, 0x3058, 0x3058, 0x305B, 0x3061, 0x3064, 0x3068, | ||
374 | 0x306C, 0x306D, 0x3080, 0x3088, 0x308B, 0x308C, 0x3090, 0x3094, | ||
375 | 0x3098, 0x3098, 0x309C, 0x309C, 0x30C0, 0x30C0, 0x30C8, 0x30C8, | ||
376 | 0x30D0, 0x30D0, 0x30D8, 0x30D8, 0x30E0, 0x30E0, 0x3100, 0x3100, | ||
377 | 0x3108, 0x3108, 0x3110, 0x3110, 0x3118, 0x3118, 0x3120, 0x3120, | ||
378 | 0x3124, 0x3125, 0x3129, 0x3129, 0x3131, 0x3131, 0x330C, 0x330C, | ||
379 | 0x3310, 0x3310, 0x3400, 0x3401, 0x3410, 0x3410, 0x3412, 0x3416, | ||
380 | 0x341D, 0x3420, 0x3428, 0x3428, 0x3430, 0x3430, 0x3432, 0x3436, | ||
381 | 0x3440, 0x3440, 0x3450, 0x3450, 0x3452, 0x3456, 0x3480, 0x3480, | ||
382 | 0x3494, 0x3495, 0x4000, 0x4000, 0x4002, 0x4002, 0x4004, 0x4004, | ||
383 | 0x4008, 0x400A, 0x400C, 0x400D, 0x400F, 0x4012, 0x4014, 0x4016, | ||
384 | 0x401D, 0x401D, 0x4020, 0x4027, 0x4060, 0x4062, 0x4200, 0x4200, | ||
385 | 0x4300, 0x4300, 0x4400, 0x4400, 0x4500, 0x4500, 0x4800, 0x4802, | ||
386 | 0x480F, 0x480F, 0x4811, 0x4811, 0x4813, 0x4813, 0x4815, 0x4816, | ||
387 | 0x482B, 0x482B, 0x4857, 0x4857, 0x4883, 0x4883, 0x48AF, 0x48AF, | ||
388 | 0x48C5, 0x48C5, 0x48E5, 0x48E5, 0x4905, 0x4905, 0x4925, 0x4925, | ||
389 | 0x4945, 0x4945, 0x4950, 0x4950, 0x495B, 0x495B, 0x4980, 0x498E, | ||
390 | 0x4B00, 0x4B00, 0x4C00, 0x4C00, 0x4D00, 0x4D00, 0x4E00, 0x4E00, | ||
391 | 0x4E80, 0x4E80, 0x4F00, 0x4F00, 0x4F08, 0x4F08, 0x4F10, 0x4F10, | ||
392 | 0x4F18, 0x4F18, 0x4F20, 0x4F20, 0x4F30, 0x4F30, 0x4F60, 0x4F60, | ||
393 | 0x4F80, 0x4F81, 0x4F88, 0x4F89, 0x4FEE, 0x4FEE, 0x4FF3, 0x4FF3, | ||
394 | 0x6000, 0x6001, 0x6008, 0x600F, 0x6014, 0x6016, 0x6018, 0x601B, | ||
395 | 0x61FD, 0x61FD, 0x623C, 0x623C, 0x6380, 0x6380, 0x63A0, 0x63A0, | ||
396 | 0x63C0, 0x63C1, 0x63C8, 0x63C9, 0x63D0, 0x63D4, 0x63D6, 0x63D6, | ||
397 | 0x63EE, 0x63EE, 0x6400, 0x6401, 0x6408, 0x640F, 0x6414, 0x6416, | ||
398 | 0x6418, 0x641B, 0x65FD, 0x65FD, 0x663C, 0x663C, 0x6780, 0x6780, | ||
399 | 0x67A0, 0x67A0, 0x67C0, 0x67C1, 0x67C8, 0x67C9, 0x67D0, 0x67D4, | ||
400 | 0x67D6, 0x67D6, 0x67EE, 0x67EE, 0x6800, 0x6801, 0x6808, 0x680F, | ||
401 | 0x6814, 0x6816, 0x6818, 0x681B, 0x69FD, 0x69FD, 0x6A3C, 0x6A3C, | ||
402 | 0x6B80, 0x6B80, 0x6BA0, 0x6BA0, 0x6BC0, 0x6BC1, 0x6BC8, 0x6BC9, | ||
403 | 0x6BD0, 0x6BD4, 0x6BD6, 0x6BD6, 0x6BEE, 0x6BEE, | ||
404 | ~0 /* sentinel */ | ||
405 | }; | ||
406 | |||
407 | #ifdef CONFIG_DEBUG_FS | ||
408 | static void a4xx_show(struct msm_gpu *gpu, struct seq_file *m) | ||
409 | { | ||
410 | gpu->funcs->pm_resume(gpu); | ||
411 | |||
412 | seq_printf(m, "status: %08x\n", | ||
413 | gpu_read(gpu, REG_A4XX_RBBM_STATUS)); | ||
414 | gpu->funcs->pm_suspend(gpu); | ||
415 | |||
416 | adreno_show(gpu, m); | ||
417 | |||
418 | } | ||
419 | #endif | ||
420 | |||
421 | /* Register offset defines for A4XX, in order of enum adreno_regs */ | ||
422 | static const unsigned int a4xx_register_offsets[REG_ADRENO_REGISTER_MAX] = { | ||
423 | REG_ADRENO_DEFINE(REG_ADRENO_CP_DEBUG, REG_A4XX_CP_DEBUG), | ||
424 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_WADDR, REG_A4XX_CP_ME_RAM_WADDR), | ||
425 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_DATA, REG_A4XX_CP_ME_RAM_DATA), | ||
426 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PFP_UCODE_DATA, | ||
427 | REG_A4XX_CP_PFP_UCODE_DATA), | ||
428 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PFP_UCODE_ADDR, | ||
429 | REG_A4XX_CP_PFP_UCODE_ADDR), | ||
430 | REG_ADRENO_DEFINE(REG_ADRENO_CP_WFI_PEND_CTR, REG_A4XX_CP_WFI_PEND_CTR), | ||
431 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_BASE, REG_A4XX_CP_RB_BASE), | ||
432 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_RPTR_ADDR, REG_A4XX_CP_RB_RPTR_ADDR), | ||
433 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_RPTR, REG_A4XX_CP_RB_RPTR), | ||
434 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_WPTR, REG_A4XX_CP_RB_WPTR), | ||
435 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PROTECT_CTRL, REG_A4XX_CP_PROTECT_CTRL), | ||
436 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_CNTL, REG_A4XX_CP_ME_CNTL), | ||
437 | REG_ADRENO_DEFINE(REG_ADRENO_CP_RB_CNTL, REG_A4XX_CP_RB_CNTL), | ||
438 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB1_BASE, REG_A4XX_CP_IB1_BASE), | ||
439 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB1_BUFSZ, REG_A4XX_CP_IB1_BUFSZ), | ||
440 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB2_BASE, REG_A4XX_CP_IB2_BASE), | ||
441 | REG_ADRENO_DEFINE(REG_ADRENO_CP_IB2_BUFSZ, REG_A4XX_CP_IB2_BUFSZ), | ||
442 | REG_ADRENO_DEFINE(REG_ADRENO_CP_TIMESTAMP, REG_AXXX_CP_SCRATCH_REG0), | ||
443 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ME_RAM_RADDR, REG_A4XX_CP_ME_RAM_RADDR), | ||
444 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ROQ_ADDR, REG_A4XX_CP_ROQ_ADDR), | ||
445 | REG_ADRENO_DEFINE(REG_ADRENO_CP_ROQ_DATA, REG_A4XX_CP_ROQ_DATA), | ||
446 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_ADDR, REG_A4XX_CP_MERCIU_ADDR), | ||
447 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_DATA, REG_A4XX_CP_MERCIU_DATA), | ||
448 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MERCIU_DATA2, REG_A4XX_CP_MERCIU_DATA2), | ||
449 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MEQ_ADDR, REG_A4XX_CP_MEQ_ADDR), | ||
450 | REG_ADRENO_DEFINE(REG_ADRENO_CP_MEQ_DATA, REG_A4XX_CP_MEQ_DATA), | ||
451 | REG_ADRENO_DEFINE(REG_ADRENO_CP_HW_FAULT, REG_A4XX_CP_HW_FAULT), | ||
452 | REG_ADRENO_DEFINE(REG_ADRENO_CP_PROTECT_STATUS, | ||
453 | REG_A4XX_CP_PROTECT_STATUS), | ||
454 | REG_ADRENO_DEFINE(REG_ADRENO_SCRATCH_ADDR, REG_A4XX_CP_SCRATCH_ADDR), | ||
455 | REG_ADRENO_DEFINE(REG_ADRENO_SCRATCH_UMSK, REG_A4XX_CP_SCRATCH_UMASK), | ||
456 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_STATUS, REG_A4XX_RBBM_STATUS), | ||
457 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_CTL, | ||
458 | REG_A4XX_RBBM_PERFCTR_CTL), | ||
459 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_CMD0, | ||
460 | REG_A4XX_RBBM_PERFCTR_LOAD_CMD0), | ||
461 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_CMD1, | ||
462 | REG_A4XX_RBBM_PERFCTR_LOAD_CMD1), | ||
463 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_CMD2, | ||
464 | REG_A4XX_RBBM_PERFCTR_LOAD_CMD2), | ||
465 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_PWR_1_LO, | ||
466 | REG_A4XX_RBBM_PERFCTR_PWR_1_LO), | ||
467 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_0_MASK, REG_A4XX_RBBM_INT_0_MASK), | ||
468 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_0_STATUS, | ||
469 | REG_A4XX_RBBM_INT_0_STATUS), | ||
470 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_ERROR_STATUS, | ||
471 | REG_A4XX_RBBM_AHB_ERROR_STATUS), | ||
472 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_CMD, REG_A4XX_RBBM_AHB_CMD), | ||
473 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_CLOCK_CTL, REG_A4XX_RBBM_CLOCK_CTL), | ||
474 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_ME_SPLIT_STATUS, | ||
475 | REG_A4XX_RBBM_AHB_ME_SPLIT_STATUS), | ||
476 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_AHB_PFP_SPLIT_STATUS, | ||
477 | REG_A4XX_RBBM_AHB_PFP_SPLIT_STATUS), | ||
478 | REG_ADRENO_DEFINE(REG_ADRENO_VPC_DEBUG_RAM_SEL, | ||
479 | REG_A4XX_VPC_DEBUG_RAM_SEL), | ||
480 | REG_ADRENO_DEFINE(REG_ADRENO_VPC_DEBUG_RAM_READ, | ||
481 | REG_A4XX_VPC_DEBUG_RAM_READ), | ||
482 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_INT_CLEAR_CMD, | ||
483 | REG_A4XX_RBBM_INT_CLEAR_CMD), | ||
484 | REG_ADRENO_DEFINE(REG_ADRENO_VSC_SIZE_ADDRESS, | ||
485 | REG_A4XX_VSC_SIZE_ADDRESS), | ||
486 | REG_ADRENO_DEFINE(REG_ADRENO_VFD_CONTROL_0, REG_A4XX_VFD_CONTROL_0), | ||
487 | REG_ADRENO_DEFINE(REG_ADRENO_SP_VS_PVT_MEM_ADDR_REG, | ||
488 | REG_A4XX_SP_VS_PVT_MEM_ADDR), | ||
489 | REG_ADRENO_DEFINE(REG_ADRENO_SP_FS_PVT_MEM_ADDR_REG, | ||
490 | REG_A4XX_SP_FS_PVT_MEM_ADDR), | ||
491 | REG_ADRENO_DEFINE(REG_ADRENO_SP_VS_OBJ_START_REG, | ||
492 | REG_A4XX_SP_VS_OBJ_START), | ||
493 | REG_ADRENO_DEFINE(REG_ADRENO_SP_FS_OBJ_START_REG, | ||
494 | REG_A4XX_SP_FS_OBJ_START), | ||
495 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_RBBM_CTL, REG_A4XX_RBBM_RBBM_CTL), | ||
496 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_SW_RESET_CMD, | ||
497 | REG_A4XX_RBBM_SW_RESET_CMD), | ||
498 | REG_ADRENO_DEFINE(REG_ADRENO_UCHE_INVALIDATE0, | ||
499 | REG_A4XX_UCHE_INVALIDATE0), | ||
500 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_LO, | ||
501 | REG_A4XX_RBBM_PERFCTR_LOAD_VALUE_LO), | ||
502 | REG_ADRENO_DEFINE(REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_HI, | ||
503 | REG_A4XX_RBBM_PERFCTR_LOAD_VALUE_HI), | ||
504 | }; | ||
505 | |||
506 | static void a4xx_dump(struct msm_gpu *gpu) | ||
507 | { | ||
508 | adreno_dump(gpu); | ||
509 | printk("status: %08x\n", | ||
510 | gpu_read(gpu, REG_A4XX_RBBM_STATUS)); | ||
511 | adreno_dump(gpu); | ||
512 | } | ||
513 | |||
514 | static const struct adreno_gpu_funcs funcs = { | ||
515 | .base = { | ||
516 | .get_param = adreno_get_param, | ||
517 | .hw_init = a4xx_hw_init, | ||
518 | .pm_suspend = msm_gpu_pm_suspend, | ||
519 | .pm_resume = msm_gpu_pm_resume, | ||
520 | .recover = a4xx_recover, | ||
521 | .last_fence = adreno_last_fence, | ||
522 | .submit = adreno_submit, | ||
523 | .flush = adreno_flush, | ||
524 | .idle = a4xx_idle, | ||
525 | .irq = a4xx_irq, | ||
526 | .destroy = a4xx_destroy, | ||
527 | #ifdef CONFIG_DEBUG_FS | ||
528 | .show = a4xx_show, | ||
529 | #endif | ||
530 | }, | ||
531 | }; | ||
532 | |||
533 | struct msm_gpu *a4xx_gpu_init(struct drm_device *dev) | ||
534 | { | ||
535 | struct a4xx_gpu *a4xx_gpu = NULL; | ||
536 | struct adreno_gpu *adreno_gpu; | ||
537 | struct msm_gpu *gpu; | ||
538 | struct msm_drm_private *priv = dev->dev_private; | ||
539 | struct platform_device *pdev = priv->gpu_pdev; | ||
540 | int ret; | ||
541 | |||
542 | if (!pdev) { | ||
543 | dev_err(dev->dev, "no a4xx device\n"); | ||
544 | ret = -ENXIO; | ||
545 | goto fail; | ||
546 | } | ||
547 | |||
548 | a4xx_gpu = kzalloc(sizeof(*a4xx_gpu), GFP_KERNEL); | ||
549 | if (!a4xx_gpu) { | ||
550 | ret = -ENOMEM; | ||
551 | goto fail; | ||
552 | } | ||
553 | |||
554 | adreno_gpu = &a4xx_gpu->base; | ||
555 | gpu = &adreno_gpu->base; | ||
556 | |||
557 | a4xx_gpu->pdev = pdev; | ||
558 | |||
559 | gpu->perfcntrs = NULL; | ||
560 | gpu->num_perfcntrs = 0; | ||
561 | |||
562 | adreno_gpu->registers = a4xx_registers; | ||
563 | adreno_gpu->reg_offsets = a4xx_register_offsets; | ||
564 | |||
565 | ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs); | ||
566 | if (ret) | ||
567 | goto fail; | ||
568 | |||
569 | /* if needed, allocate gmem: */ | ||
570 | if (adreno_is_a4xx(adreno_gpu)) { | ||
571 | #ifdef CONFIG_MSM_OCMEM | ||
572 | /* TODO this is different/missing upstream: */ | ||
573 | struct ocmem_buf *ocmem_hdl = | ||
574 | ocmem_allocate(OCMEM_GRAPHICS, adreno_gpu->gmem); | ||
575 | |||
576 | a4xx_gpu->ocmem_hdl = ocmem_hdl; | ||
577 | a4xx_gpu->ocmem_base = ocmem_hdl->addr; | ||
578 | adreno_gpu->gmem = ocmem_hdl->len; | ||
579 | DBG("using %dK of OCMEM at 0x%08x", adreno_gpu->gmem / 1024, | ||
580 | a4xx_gpu->ocmem_base); | ||
581 | #endif | ||
582 | } | ||
583 | |||
584 | if (!gpu->mmu) { | ||
585 | /* TODO we think it is possible to configure the GPU to | ||
586 | * restrict access to VRAM carveout. But the required | ||
587 | * registers are unknown. For now just bail out and | ||
588 | * limp along with just modesetting. If it turns out | ||
589 | * to not be possible to restrict access, then we must | ||
590 | * implement a cmdstream validator. | ||
591 | */ | ||
592 | dev_err(dev->dev, "No memory protection without IOMMU\n"); | ||
593 | ret = -ENXIO; | ||
594 | goto fail; | ||
595 | } | ||
596 | |||
597 | return gpu; | ||
598 | |||
599 | fail: | ||
600 | if (a4xx_gpu) | ||
601 | a4xx_destroy(&a4xx_gpu->base.base); | ||
602 | |||
603 | return ERR_PTR(ret); | ||
604 | } | ||
diff --git a/drivers/gpu/drm/msm/adreno/a4xx_gpu.h b/drivers/gpu/drm/msm/adreno/a4xx_gpu.h new file mode 100644 index 000000000000..01247204ac92 --- /dev/null +++ b/drivers/gpu/drm/msm/adreno/a4xx_gpu.h | |||
@@ -0,0 +1,34 @@ | |||
1 | /* Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
2 | * | ||
3 | * This program is free software; you can redistribute it and/or modify | ||
4 | * it under the terms of the GNU General Public License version 2 and | ||
5 | * only version 2 as published by the Free Software Foundation. | ||
6 | * | ||
7 | * This program is distributed in the hope that it will be useful, | ||
8 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
9 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
10 | * GNU General Public License for more details. | ||
11 | * | ||
12 | */ | ||
13 | #ifndef __A4XX_GPU_H__ | ||
14 | #define __A4XX_GPU_H__ | ||
15 | |||
16 | #include "adreno_gpu.h" | ||
17 | |||
18 | /* arrg, somehow fb.h is getting pulled in: */ | ||
19 | #undef ROP_COPY | ||
20 | #undef ROP_XOR | ||
21 | |||
22 | #include "a4xx.xml.h" | ||
23 | |||
24 | struct a4xx_gpu { | ||
25 | struct adreno_gpu base; | ||
26 | struct platform_device *pdev; | ||
27 | |||
28 | /* if OCMEM is used for GMEM: */ | ||
29 | uint32_t ocmem_base; | ||
30 | void *ocmem_hdl; | ||
31 | }; | ||
32 | #define to_a4xx_gpu(x) container_of(x, struct a4xx_gpu, base) | ||
33 | |||
34 | #endif /* __A4XX_GPU_H__ */ | ||
diff --git a/drivers/gpu/drm/msm/adreno/adreno_common.xml.h b/drivers/gpu/drm/msm/adreno/adreno_common.xml.h index cc341bc62b51..a4b33af9338d 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_common.xml.h +++ b/drivers/gpu/drm/msm/adreno/adreno_common.xml.h | |||
@@ -11,10 +11,10 @@ The rules-ng-ng source files this header was generated from are: | |||
11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 9859 bytes, from 2014-06-02 15:21:30) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 10551 bytes, from 2014-11-13 22:44:30) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 14960 bytes, from 2014-07-27 17:22:13) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 15053 bytes, from 2014-11-09 15:45:47) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 58020 bytes, from 2014-08-01 12:22:48) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 63169 bytes, from 2014-11-13 22:44:18) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 41068 bytes, from 2014-08-01 12:22:48) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 49097 bytes, from 2014-11-14 15:38:00) |
18 | 18 | ||
19 | Copyright (C) 2013-2014 by the following authors: | 19 | Copyright (C) 2013-2014 by the following authors: |
20 | - Rob Clark <robdclark@gmail.com> (robclark) | 20 | - Rob Clark <robdclark@gmail.com> (robclark) |
@@ -105,6 +105,7 @@ enum adreno_rb_dither_mode { | |||
105 | enum adreno_rb_depth_format { | 105 | enum adreno_rb_depth_format { |
106 | DEPTHX_16 = 0, | 106 | DEPTHX_16 = 0, |
107 | DEPTHX_24_8 = 1, | 107 | DEPTHX_24_8 = 1, |
108 | DEPTHX_32 = 2, | ||
108 | }; | 109 | }; |
109 | 110 | ||
110 | enum adreno_rb_copy_control_mode { | 111 | enum adreno_rb_copy_control_mode { |
@@ -132,6 +133,7 @@ enum a3xx_threadmode { | |||
132 | }; | 133 | }; |
133 | 134 | ||
134 | enum a3xx_instrbuffermode { | 135 | enum a3xx_instrbuffermode { |
136 | CACHE = 0, | ||
135 | BUFFER = 1, | 137 | BUFFER = 1, |
136 | }; | 138 | }; |
137 | 139 | ||
@@ -140,6 +142,13 @@ enum a3xx_threadsize { | |||
140 | FOUR_QUADS = 1, | 142 | FOUR_QUADS = 1, |
141 | }; | 143 | }; |
142 | 144 | ||
145 | enum a3xx_color_swap { | ||
146 | WZYX = 0, | ||
147 | WXYZ = 1, | ||
148 | ZYXW = 2, | ||
149 | XYZW = 3, | ||
150 | }; | ||
151 | |||
143 | #define REG_AXXX_CP_RB_BASE 0x000001c0 | 152 | #define REG_AXXX_CP_RB_BASE 0x000001c0 |
144 | 153 | ||
145 | #define REG_AXXX_CP_RB_CNTL 0x000001c1 | 154 | #define REG_AXXX_CP_RB_CNTL 0x000001c1 |
diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c index 7ab85af3a7db..be83dee83d08 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_device.c +++ b/drivers/gpu/drm/msm/adreno/adreno_device.c | |||
@@ -2,6 +2,8 @@ | |||
2 | * Copyright (C) 2013-2014 Red Hat | 2 | * Copyright (C) 2013-2014 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 3 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 4 | * |
5 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
6 | * | ||
5 | * This program is free software; you can redistribute it and/or modify it | 7 | * This program is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 as published by | 8 | * under the terms of the GNU General Public License version 2 as published by |
7 | * the Free Software Foundation. | 9 | * the Free Software Foundation. |
@@ -28,6 +30,7 @@ MODULE_PARM_DESC(hang_debug, "Dump registers when hang is detected (can be slow! | |||
28 | module_param_named(hang_debug, hang_debug, bool, 0600); | 30 | module_param_named(hang_debug, hang_debug, bool, 0600); |
29 | 31 | ||
30 | struct msm_gpu *a3xx_gpu_init(struct drm_device *dev); | 32 | struct msm_gpu *a3xx_gpu_init(struct drm_device *dev); |
33 | struct msm_gpu *a4xx_gpu_init(struct drm_device *dev); | ||
31 | 34 | ||
32 | static const struct adreno_info gpulist[] = { | 35 | static const struct adreno_info gpulist[] = { |
33 | { | 36 | { |
@@ -54,6 +57,14 @@ static const struct adreno_info gpulist[] = { | |||
54 | .pfpfw = "a330_pfp.fw", | 57 | .pfpfw = "a330_pfp.fw", |
55 | .gmem = SZ_1M, | 58 | .gmem = SZ_1M, |
56 | .init = a3xx_gpu_init, | 59 | .init = a3xx_gpu_init, |
60 | }, { | ||
61 | .rev = ADRENO_REV(4, 2, 0, ANY_ID), | ||
62 | .revn = 420, | ||
63 | .name = "A420", | ||
64 | .pm4fw = "a420_pm4.fw", | ||
65 | .pfpfw = "a420_pfp.fw", | ||
66 | .gmem = (SZ_1M + SZ_512K), | ||
67 | .init = a4xx_gpu_init, | ||
57 | }, | 68 | }, |
58 | }; | 69 | }; |
59 | 70 | ||
@@ -61,6 +72,8 @@ MODULE_FIRMWARE("a300_pm4.fw"); | |||
61 | MODULE_FIRMWARE("a300_pfp.fw"); | 72 | MODULE_FIRMWARE("a300_pfp.fw"); |
62 | MODULE_FIRMWARE("a330_pm4.fw"); | 73 | MODULE_FIRMWARE("a330_pm4.fw"); |
63 | MODULE_FIRMWARE("a330_pfp.fw"); | 74 | MODULE_FIRMWARE("a330_pfp.fw"); |
75 | MODULE_FIRMWARE("a420_pm4.fw"); | ||
76 | MODULE_FIRMWARE("a420_pfp.fw"); | ||
64 | 77 | ||
65 | static inline bool _rev_match(uint8_t entry, uint8_t id) | 78 | static inline bool _rev_match(uint8_t entry, uint8_t id) |
66 | { | 79 | { |
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c index 6afa29167fee..aa873048308b 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c | |||
@@ -2,6 +2,8 @@ | |||
2 | * Copyright (C) 2013 Red Hat | 2 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 3 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 4 | * |
5 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
6 | * | ||
5 | * This program is free software; you can redistribute it and/or modify it | 7 | * This program is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 as published by | 8 | * under the terms of the GNU General Public License version 2 as published by |
7 | * the Free Software Foundation. | 9 | * the Free Software Foundation. |
@@ -63,19 +65,21 @@ int adreno_hw_init(struct msm_gpu *gpu) | |||
63 | } | 65 | } |
64 | 66 | ||
65 | /* Setup REG_CP_RB_CNTL: */ | 67 | /* Setup REG_CP_RB_CNTL: */ |
66 | gpu_write(gpu, REG_AXXX_CP_RB_CNTL, | 68 | adreno_gpu_write(adreno_gpu, REG_ADRENO_CP_RB_CNTL, |
67 | /* size is log2(quad-words): */ | 69 | /* size is log2(quad-words): */ |
68 | AXXX_CP_RB_CNTL_BUFSZ(ilog2(gpu->rb->size / 8)) | | 70 | AXXX_CP_RB_CNTL_BUFSZ(ilog2(gpu->rb->size / 8)) | |
69 | AXXX_CP_RB_CNTL_BLKSZ(ilog2(RB_BLKSIZE / 8))); | 71 | AXXX_CP_RB_CNTL_BLKSZ(ilog2(RB_BLKSIZE / 8))); |
70 | 72 | ||
71 | /* Setup ringbuffer address: */ | 73 | /* Setup ringbuffer address: */ |
72 | gpu_write(gpu, REG_AXXX_CP_RB_BASE, gpu->rb_iova); | 74 | adreno_gpu_write(adreno_gpu, REG_ADRENO_CP_RB_BASE, gpu->rb_iova); |
73 | gpu_write(gpu, REG_AXXX_CP_RB_RPTR_ADDR, rbmemptr(adreno_gpu, rptr)); | 75 | adreno_gpu_write(adreno_gpu, REG_ADRENO_CP_RB_RPTR_ADDR, |
76 | rbmemptr(adreno_gpu, rptr)); | ||
74 | 77 | ||
75 | /* Setup scratch/timestamp: */ | 78 | /* Setup scratch/timestamp: */ |
76 | gpu_write(gpu, REG_AXXX_SCRATCH_ADDR, rbmemptr(adreno_gpu, fence)); | 79 | adreno_gpu_write(adreno_gpu, REG_ADRENO_SCRATCH_ADDR, |
80 | rbmemptr(adreno_gpu, fence)); | ||
77 | 81 | ||
78 | gpu_write(gpu, REG_AXXX_SCRATCH_UMSK, 0x1); | 82 | adreno_gpu_write(adreno_gpu, REG_ADRENO_SCRATCH_UMSK, 0x1); |
79 | 83 | ||
80 | return 0; | 84 | return 0; |
81 | } | 85 | } |
@@ -151,7 +155,7 @@ int adreno_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, | |||
151 | OUT_PKT0(ring, REG_AXXX_CP_SCRATCH_REG2, 1); | 155 | OUT_PKT0(ring, REG_AXXX_CP_SCRATCH_REG2, 1); |
152 | OUT_RING(ring, submit->fence); | 156 | OUT_RING(ring, submit->fence); |
153 | 157 | ||
154 | if (adreno_is_a3xx(adreno_gpu)) { | 158 | if (adreno_is_a3xx(adreno_gpu) || adreno_is_a4xx(adreno_gpu)) { |
155 | /* Flush HLSQ lazy updates to make sure there is nothing | 159 | /* Flush HLSQ lazy updates to make sure there is nothing |
156 | * pending for indirect loads after the timestamp has | 160 | * pending for indirect loads after the timestamp has |
157 | * passed: | 161 | * passed: |
@@ -188,12 +192,13 @@ int adreno_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, | |||
188 | 192 | ||
189 | void adreno_flush(struct msm_gpu *gpu) | 193 | void adreno_flush(struct msm_gpu *gpu) |
190 | { | 194 | { |
195 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | ||
191 | uint32_t wptr = get_wptr(gpu->rb); | 196 | uint32_t wptr = get_wptr(gpu->rb); |
192 | 197 | ||
193 | /* ensure writes to ringbuffer have hit system memory: */ | 198 | /* ensure writes to ringbuffer have hit system memory: */ |
194 | mb(); | 199 | mb(); |
195 | 200 | ||
196 | gpu_write(gpu, REG_AXXX_CP_RB_WPTR, wptr); | 201 | adreno_gpu_write(adreno_gpu, REG_ADRENO_CP_RB_WPTR, wptr); |
197 | } | 202 | } |
198 | 203 | ||
199 | void adreno_idle(struct msm_gpu *gpu) | 204 | void adreno_idle(struct msm_gpu *gpu) |
@@ -319,6 +324,12 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev, | |||
319 | DBG("fast_rate=%u, slow_rate=%u, bus_freq=%u", | 324 | DBG("fast_rate=%u, slow_rate=%u, bus_freq=%u", |
320 | gpu->fast_rate, gpu->slow_rate, gpu->bus_freq); | 325 | gpu->fast_rate, gpu->slow_rate, gpu->bus_freq); |
321 | 326 | ||
327 | ret = msm_gpu_init(drm, pdev, &adreno_gpu->base, &funcs->base, | ||
328 | adreno_gpu->info->name, "kgsl_3d0_reg_memory", "kgsl_3d0_irq", | ||
329 | RB_SIZE); | ||
330 | if (ret) | ||
331 | return ret; | ||
332 | |||
322 | ret = request_firmware(&adreno_gpu->pm4, adreno_gpu->info->pm4fw, drm->dev); | 333 | ret = request_firmware(&adreno_gpu->pm4, adreno_gpu->info->pm4fw, drm->dev); |
323 | if (ret) { | 334 | if (ret) { |
324 | dev_err(drm->dev, "failed to load %s PM4 firmware: %d\n", | 335 | dev_err(drm->dev, "failed to load %s PM4 firmware: %d\n", |
@@ -333,12 +344,6 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev, | |||
333 | return ret; | 344 | return ret; |
334 | } | 345 | } |
335 | 346 | ||
336 | ret = msm_gpu_init(drm, pdev, &adreno_gpu->base, &funcs->base, | ||
337 | adreno_gpu->info->name, "kgsl_3d0_reg_memory", "kgsl_3d0_irq", | ||
338 | RB_SIZE); | ||
339 | if (ret) | ||
340 | return ret; | ||
341 | |||
342 | mmu = gpu->mmu; | 347 | mmu = gpu->mmu; |
343 | if (mmu) { | 348 | if (mmu) { |
344 | ret = mmu->funcs->attach(mmu, iommu_ports, | 349 | ret = mmu->funcs->attach(mmu, iommu_ports, |
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h index 52f051579753..a0cc30977e67 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h | |||
@@ -2,6 +2,8 @@ | |||
2 | * Copyright (C) 2013 Red Hat | 2 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 3 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 4 | * |
5 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
6 | * | ||
5 | * This program is free software; you can redistribute it and/or modify it | 7 | * This program is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 as published by | 8 | * under the terms of the GNU General Public License version 2 as published by |
7 | * the Free Software Foundation. | 9 | * the Free Software Foundation. |
@@ -25,6 +27,81 @@ | |||
25 | #include "adreno_common.xml.h" | 27 | #include "adreno_common.xml.h" |
26 | #include "adreno_pm4.xml.h" | 28 | #include "adreno_pm4.xml.h" |
27 | 29 | ||
30 | #define REG_ADRENO_DEFINE(_offset, _reg) [_offset] = (_reg) + 1 | ||
31 | /** | ||
32 | * adreno_regs: List of registers that are used in across all | ||
33 | * 3D devices. Each device type has different offset value for the same | ||
34 | * register, so an array of register offsets are declared for every device | ||
35 | * and are indexed by the enumeration values defined in this enum | ||
36 | */ | ||
37 | enum adreno_regs { | ||
38 | REG_ADRENO_CP_DEBUG, | ||
39 | REG_ADRENO_CP_ME_RAM_WADDR, | ||
40 | REG_ADRENO_CP_ME_RAM_DATA, | ||
41 | REG_ADRENO_CP_PFP_UCODE_DATA, | ||
42 | REG_ADRENO_CP_PFP_UCODE_ADDR, | ||
43 | REG_ADRENO_CP_WFI_PEND_CTR, | ||
44 | REG_ADRENO_CP_RB_BASE, | ||
45 | REG_ADRENO_CP_RB_RPTR_ADDR, | ||
46 | REG_ADRENO_CP_RB_RPTR, | ||
47 | REG_ADRENO_CP_RB_WPTR, | ||
48 | REG_ADRENO_CP_PROTECT_CTRL, | ||
49 | REG_ADRENO_CP_ME_CNTL, | ||
50 | REG_ADRENO_CP_RB_CNTL, | ||
51 | REG_ADRENO_CP_IB1_BASE, | ||
52 | REG_ADRENO_CP_IB1_BUFSZ, | ||
53 | REG_ADRENO_CP_IB2_BASE, | ||
54 | REG_ADRENO_CP_IB2_BUFSZ, | ||
55 | REG_ADRENO_CP_TIMESTAMP, | ||
56 | REG_ADRENO_CP_ME_RAM_RADDR, | ||
57 | REG_ADRENO_CP_ROQ_ADDR, | ||
58 | REG_ADRENO_CP_ROQ_DATA, | ||
59 | REG_ADRENO_CP_MERCIU_ADDR, | ||
60 | REG_ADRENO_CP_MERCIU_DATA, | ||
61 | REG_ADRENO_CP_MERCIU_DATA2, | ||
62 | REG_ADRENO_CP_MEQ_ADDR, | ||
63 | REG_ADRENO_CP_MEQ_DATA, | ||
64 | REG_ADRENO_CP_HW_FAULT, | ||
65 | REG_ADRENO_CP_PROTECT_STATUS, | ||
66 | REG_ADRENO_SCRATCH_ADDR, | ||
67 | REG_ADRENO_SCRATCH_UMSK, | ||
68 | REG_ADRENO_SCRATCH_REG2, | ||
69 | REG_ADRENO_RBBM_STATUS, | ||
70 | REG_ADRENO_RBBM_PERFCTR_CTL, | ||
71 | REG_ADRENO_RBBM_PERFCTR_LOAD_CMD0, | ||
72 | REG_ADRENO_RBBM_PERFCTR_LOAD_CMD1, | ||
73 | REG_ADRENO_RBBM_PERFCTR_LOAD_CMD2, | ||
74 | REG_ADRENO_RBBM_PERFCTR_PWR_1_LO, | ||
75 | REG_ADRENO_RBBM_INT_0_MASK, | ||
76 | REG_ADRENO_RBBM_INT_0_STATUS, | ||
77 | REG_ADRENO_RBBM_AHB_ERROR_STATUS, | ||
78 | REG_ADRENO_RBBM_PM_OVERRIDE2, | ||
79 | REG_ADRENO_RBBM_AHB_CMD, | ||
80 | REG_ADRENO_RBBM_INT_CLEAR_CMD, | ||
81 | REG_ADRENO_RBBM_SW_RESET_CMD, | ||
82 | REG_ADRENO_RBBM_CLOCK_CTL, | ||
83 | REG_ADRENO_RBBM_AHB_ME_SPLIT_STATUS, | ||
84 | REG_ADRENO_RBBM_AHB_PFP_SPLIT_STATUS, | ||
85 | REG_ADRENO_VPC_DEBUG_RAM_SEL, | ||
86 | REG_ADRENO_VPC_DEBUG_RAM_READ, | ||
87 | REG_ADRENO_VSC_SIZE_ADDRESS, | ||
88 | REG_ADRENO_VFD_CONTROL_0, | ||
89 | REG_ADRENO_VFD_INDEX_MAX, | ||
90 | REG_ADRENO_SP_VS_PVT_MEM_ADDR_REG, | ||
91 | REG_ADRENO_SP_FS_PVT_MEM_ADDR_REG, | ||
92 | REG_ADRENO_SP_VS_OBJ_START_REG, | ||
93 | REG_ADRENO_SP_FS_OBJ_START_REG, | ||
94 | REG_ADRENO_PA_SC_AA_CONFIG, | ||
95 | REG_ADRENO_SQ_GPR_MANAGEMENT, | ||
96 | REG_ADRENO_SQ_INST_STORE_MANAGMENT, | ||
97 | REG_ADRENO_TP0_CHICKEN, | ||
98 | REG_ADRENO_RBBM_RBBM_CTL, | ||
99 | REG_ADRENO_UCHE_INVALIDATE0, | ||
100 | REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_LO, | ||
101 | REG_ADRENO_RBBM_PERFCTR_LOAD_VALUE_HI, | ||
102 | REG_ADRENO_REGISTER_MAX, | ||
103 | }; | ||
104 | |||
28 | struct adreno_rev { | 105 | struct adreno_rev { |
29 | uint8_t core; | 106 | uint8_t core; |
30 | uint8_t major; | 107 | uint8_t major; |
@@ -76,6 +153,13 @@ struct adreno_gpu { | |||
76 | struct adreno_rbmemptrs *memptrs; | 153 | struct adreno_rbmemptrs *memptrs; |
77 | struct drm_gem_object *memptrs_bo; | 154 | struct drm_gem_object *memptrs_bo; |
78 | uint32_t memptrs_iova; | 155 | uint32_t memptrs_iova; |
156 | |||
157 | /* | ||
158 | * Register offsets are different between some GPUs. | ||
159 | * GPU specific offsets will be exported by GPU specific | ||
160 | * code (a3xx_gpu.c) and stored in this common location. | ||
161 | */ | ||
162 | const unsigned int *reg_offsets; | ||
79 | }; | 163 | }; |
80 | #define to_adreno_gpu(x) container_of(x, struct adreno_gpu, base) | 164 | #define to_adreno_gpu(x) container_of(x, struct adreno_gpu, base) |
81 | 165 | ||
@@ -128,6 +212,16 @@ static inline bool adreno_is_a330v2(struct adreno_gpu *gpu) | |||
128 | return adreno_is_a330(gpu) && (gpu->rev.patchid > 0); | 212 | return adreno_is_a330(gpu) && (gpu->rev.patchid > 0); |
129 | } | 213 | } |
130 | 214 | ||
215 | static inline bool adreno_is_a4xx(struct adreno_gpu *gpu) | ||
216 | { | ||
217 | return (gpu->revn >= 400) && (gpu->revn < 500); | ||
218 | } | ||
219 | |||
220 | static inline int adreno_is_a420(struct adreno_gpu *gpu) | ||
221 | { | ||
222 | return gpu->revn == 420; | ||
223 | } | ||
224 | |||
131 | int adreno_get_param(struct msm_gpu *gpu, uint32_t param, uint64_t *value); | 225 | int adreno_get_param(struct msm_gpu *gpu, uint32_t param, uint64_t *value); |
132 | int adreno_hw_init(struct msm_gpu *gpu); | 226 | int adreno_hw_init(struct msm_gpu *gpu); |
133 | uint32_t adreno_last_fence(struct msm_gpu *gpu); | 227 | uint32_t adreno_last_fence(struct msm_gpu *gpu); |
@@ -171,5 +265,37 @@ OUT_PKT3(struct msm_ringbuffer *ring, uint8_t opcode, uint16_t cnt) | |||
171 | OUT_RING(ring, CP_TYPE3_PKT | ((cnt-1) << 16) | ((opcode & 0xFF) << 8)); | 265 | OUT_RING(ring, CP_TYPE3_PKT | ((cnt-1) << 16) | ((opcode & 0xFF) << 8)); |
172 | } | 266 | } |
173 | 267 | ||
268 | /* | ||
269 | * adreno_checkreg_off() - Checks the validity of a register enum | ||
270 | * @gpu: Pointer to struct adreno_gpu | ||
271 | * @offset_name: The register enum that is checked | ||
272 | */ | ||
273 | static inline bool adreno_reg_check(struct adreno_gpu *gpu, | ||
274 | enum adreno_regs offset_name) | ||
275 | { | ||
276 | if (offset_name >= REG_ADRENO_REGISTER_MAX || | ||
277 | !gpu->reg_offsets[offset_name]) { | ||
278 | BUG(); | ||
279 | } | ||
280 | return true; | ||
281 | } | ||
282 | |||
283 | static inline u32 adreno_gpu_read(struct adreno_gpu *gpu, | ||
284 | enum adreno_regs offset_name) | ||
285 | { | ||
286 | u32 reg = gpu->reg_offsets[offset_name]; | ||
287 | u32 val = 0; | ||
288 | if(adreno_reg_check(gpu,offset_name)) | ||
289 | val = gpu_read(&gpu->base, reg - 1); | ||
290 | return val; | ||
291 | } | ||
292 | |||
293 | static inline void adreno_gpu_write(struct adreno_gpu *gpu, | ||
294 | enum adreno_regs offset_name, u32 data) | ||
295 | { | ||
296 | u32 reg = gpu->reg_offsets[offset_name]; | ||
297 | if(adreno_reg_check(gpu, offset_name)) | ||
298 | gpu_write(&gpu->base, reg - 1, data); | ||
299 | } | ||
174 | 300 | ||
175 | #endif /* __ADRENO_GPU_H__ */ | 301 | #endif /* __ADRENO_GPU_H__ */ |
diff --git a/drivers/gpu/drm/msm/adreno/adreno_pm4.xml.h b/drivers/gpu/drm/msm/adreno/adreno_pm4.xml.h index 6ef43f66c30a..6a75cee94d81 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_pm4.xml.h +++ b/drivers/gpu/drm/msm/adreno/adreno_pm4.xml.h | |||
@@ -11,10 +11,10 @@ The rules-ng-ng source files this header was generated from are: | |||
11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/adreno.xml ( 364 bytes, from 2013-11-30 14:47:15) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a2xx.xml ( 32901 bytes, from 2014-06-02 15:21:30) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 9859 bytes, from 2014-06-02 15:21:30) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_common.xml ( 10551 bytes, from 2014-11-13 22:44:30) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 14960 bytes, from 2014-07-27 17:22:13) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/adreno_pm4.xml ( 15053 bytes, from 2014-11-09 15:45:47) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 58020 bytes, from 2014-08-01 12:22:48) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a3xx.xml ( 63169 bytes, from 2014-11-13 22:44:18) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 41068 bytes, from 2014-08-01 12:22:48) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/adreno/a4xx.xml ( 49097 bytes, from 2014-11-14 15:38:00) |
18 | 18 | ||
19 | Copyright (C) 2013-2014 by the following authors: | 19 | Copyright (C) 2013-2014 by the following authors: |
20 | - Rob Clark <robdclark@gmail.com> (robclark) | 20 | - Rob Clark <robdclark@gmail.com> (robclark) |
@@ -157,6 +157,7 @@ enum adreno_pm4_type3_packets { | |||
157 | CP_IM_STORE = 44, | 157 | CP_IM_STORE = 44, |
158 | CP_SET_DRAW_INIT_FLAGS = 75, | 158 | CP_SET_DRAW_INIT_FLAGS = 75, |
159 | CP_SET_PROTECTED_MODE = 95, | 159 | CP_SET_PROTECTED_MODE = 95, |
160 | CP_BOOTSTRAP_UCODE = 111, | ||
160 | CP_LOAD_STATE = 48, | 161 | CP_LOAD_STATE = 48, |
161 | CP_COND_INDIRECT_BUFFER_PFE = 58, | 162 | CP_COND_INDIRECT_BUFFER_PFE = 58, |
162 | CP_COND_INDIRECT_BUFFER_PFD = 50, | 163 | CP_COND_INDIRECT_BUFFER_PFD = 50, |
@@ -278,11 +279,11 @@ static inline uint32_t CP_DRAW_INDX_1_INDEX_SIZE(enum pc_di_index_size val) | |||
278 | #define CP_DRAW_INDX_1_NOT_EOP 0x00001000 | 279 | #define CP_DRAW_INDX_1_NOT_EOP 0x00001000 |
279 | #define CP_DRAW_INDX_1_SMALL_INDEX 0x00002000 | 280 | #define CP_DRAW_INDX_1_SMALL_INDEX 0x00002000 |
280 | #define CP_DRAW_INDX_1_PRE_DRAW_INITIATOR_ENABLE 0x00004000 | 281 | #define CP_DRAW_INDX_1_PRE_DRAW_INITIATOR_ENABLE 0x00004000 |
281 | #define CP_DRAW_INDX_1_NUM_INDICES__MASK 0xffff0000 | 282 | #define CP_DRAW_INDX_1_NUM_INSTANCES__MASK 0xff000000 |
282 | #define CP_DRAW_INDX_1_NUM_INDICES__SHIFT 16 | 283 | #define CP_DRAW_INDX_1_NUM_INSTANCES__SHIFT 24 |
283 | static inline uint32_t CP_DRAW_INDX_1_NUM_INDICES(uint32_t val) | 284 | static inline uint32_t CP_DRAW_INDX_1_NUM_INSTANCES(uint32_t val) |
284 | { | 285 | { |
285 | return ((val) << CP_DRAW_INDX_1_NUM_INDICES__SHIFT) & CP_DRAW_INDX_1_NUM_INDICES__MASK; | 286 | return ((val) << CP_DRAW_INDX_1_NUM_INSTANCES__SHIFT) & CP_DRAW_INDX_1_NUM_INSTANCES__MASK; |
286 | } | 287 | } |
287 | 288 | ||
288 | #define REG_CP_DRAW_INDX_2 0x00000002 | 289 | #define REG_CP_DRAW_INDX_2 0x00000002 |
@@ -293,20 +294,20 @@ static inline uint32_t CP_DRAW_INDX_2_NUM_INDICES(uint32_t val) | |||
293 | return ((val) << CP_DRAW_INDX_2_NUM_INDICES__SHIFT) & CP_DRAW_INDX_2_NUM_INDICES__MASK; | 294 | return ((val) << CP_DRAW_INDX_2_NUM_INDICES__SHIFT) & CP_DRAW_INDX_2_NUM_INDICES__MASK; |
294 | } | 295 | } |
295 | 296 | ||
296 | #define REG_CP_DRAW_INDX_2 0x00000002 | 297 | #define REG_CP_DRAW_INDX_3 0x00000003 |
297 | #define CP_DRAW_INDX_2_INDX_BASE__MASK 0xffffffff | 298 | #define CP_DRAW_INDX_3_INDX_BASE__MASK 0xffffffff |
298 | #define CP_DRAW_INDX_2_INDX_BASE__SHIFT 0 | 299 | #define CP_DRAW_INDX_3_INDX_BASE__SHIFT 0 |
299 | static inline uint32_t CP_DRAW_INDX_2_INDX_BASE(uint32_t val) | 300 | static inline uint32_t CP_DRAW_INDX_3_INDX_BASE(uint32_t val) |
300 | { | 301 | { |
301 | return ((val) << CP_DRAW_INDX_2_INDX_BASE__SHIFT) & CP_DRAW_INDX_2_INDX_BASE__MASK; | 302 | return ((val) << CP_DRAW_INDX_3_INDX_BASE__SHIFT) & CP_DRAW_INDX_3_INDX_BASE__MASK; |
302 | } | 303 | } |
303 | 304 | ||
304 | #define REG_CP_DRAW_INDX_2 0x00000002 | 305 | #define REG_CP_DRAW_INDX_4 0x00000004 |
305 | #define CP_DRAW_INDX_2_INDX_SIZE__MASK 0xffffffff | 306 | #define CP_DRAW_INDX_4_INDX_SIZE__MASK 0xffffffff |
306 | #define CP_DRAW_INDX_2_INDX_SIZE__SHIFT 0 | 307 | #define CP_DRAW_INDX_4_INDX_SIZE__SHIFT 0 |
307 | static inline uint32_t CP_DRAW_INDX_2_INDX_SIZE(uint32_t val) | 308 | static inline uint32_t CP_DRAW_INDX_4_INDX_SIZE(uint32_t val) |
308 | { | 309 | { |
309 | return ((val) << CP_DRAW_INDX_2_INDX_SIZE__SHIFT) & CP_DRAW_INDX_2_INDX_SIZE__MASK; | 310 | return ((val) << CP_DRAW_INDX_4_INDX_SIZE__SHIFT) & CP_DRAW_INDX_4_INDX_SIZE__MASK; |
310 | } | 311 | } |
311 | 312 | ||
312 | #define REG_CP_DRAW_INDX_2_0 0x00000000 | 313 | #define REG_CP_DRAW_INDX_2_0 0x00000000 |
@@ -345,11 +346,11 @@ static inline uint32_t CP_DRAW_INDX_2_1_INDEX_SIZE(enum pc_di_index_size val) | |||
345 | #define CP_DRAW_INDX_2_1_NOT_EOP 0x00001000 | 346 | #define CP_DRAW_INDX_2_1_NOT_EOP 0x00001000 |
346 | #define CP_DRAW_INDX_2_1_SMALL_INDEX 0x00002000 | 347 | #define CP_DRAW_INDX_2_1_SMALL_INDEX 0x00002000 |
347 | #define CP_DRAW_INDX_2_1_PRE_DRAW_INITIATOR_ENABLE 0x00004000 | 348 | #define CP_DRAW_INDX_2_1_PRE_DRAW_INITIATOR_ENABLE 0x00004000 |
348 | #define CP_DRAW_INDX_2_1_NUM_INDICES__MASK 0xffff0000 | 349 | #define CP_DRAW_INDX_2_1_NUM_INSTANCES__MASK 0xff000000 |
349 | #define CP_DRAW_INDX_2_1_NUM_INDICES__SHIFT 16 | 350 | #define CP_DRAW_INDX_2_1_NUM_INSTANCES__SHIFT 24 |
350 | static inline uint32_t CP_DRAW_INDX_2_1_NUM_INDICES(uint32_t val) | 351 | static inline uint32_t CP_DRAW_INDX_2_1_NUM_INSTANCES(uint32_t val) |
351 | { | 352 | { |
352 | return ((val) << CP_DRAW_INDX_2_1_NUM_INDICES__SHIFT) & CP_DRAW_INDX_2_1_NUM_INDICES__MASK; | 353 | return ((val) << CP_DRAW_INDX_2_1_NUM_INSTANCES__SHIFT) & CP_DRAW_INDX_2_1_NUM_INSTANCES__MASK; |
353 | } | 354 | } |
354 | 355 | ||
355 | #define REG_CP_DRAW_INDX_2_2 0x00000002 | 356 | #define REG_CP_DRAW_INDX_2_2 0x00000002 |
@@ -388,11 +389,11 @@ static inline uint32_t CP_DRAW_INDX_OFFSET_0_INDEX_SIZE(enum pc_di_index_size va | |||
388 | #define CP_DRAW_INDX_OFFSET_0_NOT_EOP 0x00001000 | 389 | #define CP_DRAW_INDX_OFFSET_0_NOT_EOP 0x00001000 |
389 | #define CP_DRAW_INDX_OFFSET_0_SMALL_INDEX 0x00002000 | 390 | #define CP_DRAW_INDX_OFFSET_0_SMALL_INDEX 0x00002000 |
390 | #define CP_DRAW_INDX_OFFSET_0_PRE_DRAW_INITIATOR_ENABLE 0x00004000 | 391 | #define CP_DRAW_INDX_OFFSET_0_PRE_DRAW_INITIATOR_ENABLE 0x00004000 |
391 | #define CP_DRAW_INDX_OFFSET_0_NUM_INDICES__MASK 0xffff0000 | 392 | #define CP_DRAW_INDX_OFFSET_0_NUM_INSTANCES__MASK 0xffff0000 |
392 | #define CP_DRAW_INDX_OFFSET_0_NUM_INDICES__SHIFT 16 | 393 | #define CP_DRAW_INDX_OFFSET_0_NUM_INSTANCES__SHIFT 16 |
393 | static inline uint32_t CP_DRAW_INDX_OFFSET_0_NUM_INDICES(uint32_t val) | 394 | static inline uint32_t CP_DRAW_INDX_OFFSET_0_NUM_INSTANCES(uint32_t val) |
394 | { | 395 | { |
395 | return ((val) << CP_DRAW_INDX_OFFSET_0_NUM_INDICES__SHIFT) & CP_DRAW_INDX_OFFSET_0_NUM_INDICES__MASK; | 396 | return ((val) << CP_DRAW_INDX_OFFSET_0_NUM_INSTANCES__SHIFT) & CP_DRAW_INDX_OFFSET_0_NUM_INSTANCES__MASK; |
396 | } | 397 | } |
397 | 398 | ||
398 | #define REG_CP_DRAW_INDX_OFFSET_1 0x00000001 | 399 | #define REG_CP_DRAW_INDX_OFFSET_1 0x00000001 |
@@ -405,20 +406,22 @@ static inline uint32_t CP_DRAW_INDX_OFFSET_2_NUM_INDICES(uint32_t val) | |||
405 | return ((val) << CP_DRAW_INDX_OFFSET_2_NUM_INDICES__SHIFT) & CP_DRAW_INDX_OFFSET_2_NUM_INDICES__MASK; | 406 | return ((val) << CP_DRAW_INDX_OFFSET_2_NUM_INDICES__SHIFT) & CP_DRAW_INDX_OFFSET_2_NUM_INDICES__MASK; |
406 | } | 407 | } |
407 | 408 | ||
408 | #define REG_CP_DRAW_INDX_OFFSET_2 0x00000002 | 409 | #define REG_CP_DRAW_INDX_OFFSET_3 0x00000003 |
409 | #define CP_DRAW_INDX_OFFSET_2_INDX_BASE__MASK 0xffffffff | 410 | |
410 | #define CP_DRAW_INDX_OFFSET_2_INDX_BASE__SHIFT 0 | 411 | #define REG_CP_DRAW_INDX_OFFSET_4 0x00000004 |
411 | static inline uint32_t CP_DRAW_INDX_OFFSET_2_INDX_BASE(uint32_t val) | 412 | #define CP_DRAW_INDX_OFFSET_4_INDX_BASE__MASK 0xffffffff |
413 | #define CP_DRAW_INDX_OFFSET_4_INDX_BASE__SHIFT 0 | ||
414 | static inline uint32_t CP_DRAW_INDX_OFFSET_4_INDX_BASE(uint32_t val) | ||
412 | { | 415 | { |
413 | return ((val) << CP_DRAW_INDX_OFFSET_2_INDX_BASE__SHIFT) & CP_DRAW_INDX_OFFSET_2_INDX_BASE__MASK; | 416 | return ((val) << CP_DRAW_INDX_OFFSET_4_INDX_BASE__SHIFT) & CP_DRAW_INDX_OFFSET_4_INDX_BASE__MASK; |
414 | } | 417 | } |
415 | 418 | ||
416 | #define REG_CP_DRAW_INDX_OFFSET_2 0x00000002 | 419 | #define REG_CP_DRAW_INDX_OFFSET_5 0x00000005 |
417 | #define CP_DRAW_INDX_OFFSET_2_INDX_SIZE__MASK 0xffffffff | 420 | #define CP_DRAW_INDX_OFFSET_5_INDX_SIZE__MASK 0xffffffff |
418 | #define CP_DRAW_INDX_OFFSET_2_INDX_SIZE__SHIFT 0 | 421 | #define CP_DRAW_INDX_OFFSET_5_INDX_SIZE__SHIFT 0 |
419 | static inline uint32_t CP_DRAW_INDX_OFFSET_2_INDX_SIZE(uint32_t val) | 422 | static inline uint32_t CP_DRAW_INDX_OFFSET_5_INDX_SIZE(uint32_t val) |
420 | { | 423 | { |
421 | return ((val) << CP_DRAW_INDX_OFFSET_2_INDX_SIZE__SHIFT) & CP_DRAW_INDX_OFFSET_2_INDX_SIZE__MASK; | 424 | return ((val) << CP_DRAW_INDX_OFFSET_5_INDX_SIZE__SHIFT) & CP_DRAW_INDX_OFFSET_5_INDX_SIZE__MASK; |
422 | } | 425 | } |
423 | 426 | ||
424 | #define REG_CP_SET_DRAW_STATE_0 0x00000000 | 427 | #define REG_CP_SET_DRAW_STATE_0 0x00000000 |
diff --git a/drivers/gpu/drm/msm/dsi/dsi.xml.h b/drivers/gpu/drm/msm/dsi/dsi.xml.h index e965898dfda6..448438b759b4 100644 --- a/drivers/gpu/drm/msm/dsi/dsi.xml.h +++ b/drivers/gpu/drm/msm/dsi/dsi.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/dsi/mmss_cc.xml.h b/drivers/gpu/drm/msm/dsi/mmss_cc.xml.h index f2bdda957205..c102a7f074ac 100644 --- a/drivers/gpu/drm/msm/dsi/mmss_cc.xml.h +++ b/drivers/gpu/drm/msm/dsi/mmss_cc.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/dsi/sfpb.xml.h b/drivers/gpu/drm/msm/dsi/sfpb.xml.h index e5b071ffd865..a900134bdf33 100644 --- a/drivers/gpu/drm/msm/dsi/sfpb.xml.h +++ b/drivers/gpu/drm/msm/dsi/sfpb.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi.c b/drivers/gpu/drm/msm/hdmi/hdmi.c index 9d00dcba6959..062c68725376 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi.c +++ b/drivers/gpu/drm/msm/hdmi/hdmi.c | |||
@@ -15,6 +15,7 @@ | |||
15 | * this program. If not, see <http://www.gnu.org/licenses/>. | 15 | * this program. If not, see <http://www.gnu.org/licenses/>. |
16 | */ | 16 | */ |
17 | 17 | ||
18 | #include <linux/of_irq.h> | ||
18 | #include "hdmi.h" | 19 | #include "hdmi.h" |
19 | 20 | ||
20 | void hdmi_set_mode(struct hdmi *hdmi, bool power_on) | 21 | void hdmi_set_mode(struct hdmi *hdmi, bool power_on) |
@@ -39,7 +40,7 @@ void hdmi_set_mode(struct hdmi *hdmi, bool power_on) | |||
39 | power_on ? "Enable" : "Disable", ctrl); | 40 | power_on ? "Enable" : "Disable", ctrl); |
40 | } | 41 | } |
41 | 42 | ||
42 | irqreturn_t hdmi_irq(int irq, void *dev_id) | 43 | static irqreturn_t hdmi_irq(int irq, void *dev_id) |
43 | { | 44 | { |
44 | struct hdmi *hdmi = dev_id; | 45 | struct hdmi *hdmi = dev_id; |
45 | 46 | ||
@@ -54,9 +55,8 @@ irqreturn_t hdmi_irq(int irq, void *dev_id) | |||
54 | return IRQ_HANDLED; | 55 | return IRQ_HANDLED; |
55 | } | 56 | } |
56 | 57 | ||
57 | void hdmi_destroy(struct kref *kref) | 58 | static void hdmi_destroy(struct hdmi *hdmi) |
58 | { | 59 | { |
59 | struct hdmi *hdmi = container_of(kref, struct hdmi, refcount); | ||
60 | struct hdmi_phy *phy = hdmi->phy; | 60 | struct hdmi_phy *phy = hdmi->phy; |
61 | 61 | ||
62 | if (phy) | 62 | if (phy) |
@@ -68,37 +68,24 @@ void hdmi_destroy(struct kref *kref) | |||
68 | platform_set_drvdata(hdmi->pdev, NULL); | 68 | platform_set_drvdata(hdmi->pdev, NULL); |
69 | } | 69 | } |
70 | 70 | ||
71 | /* initialize connector */ | 71 | /* construct hdmi at bind/probe time, grab all the resources. If |
72 | struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | 72 | * we are to EPROBE_DEFER we want to do it here, rather than later |
73 | * at modeset_init() time | ||
74 | */ | ||
75 | static struct hdmi *hdmi_init(struct platform_device *pdev) | ||
73 | { | 76 | { |
77 | struct hdmi_platform_config *config = pdev->dev.platform_data; | ||
74 | struct hdmi *hdmi = NULL; | 78 | struct hdmi *hdmi = NULL; |
75 | struct msm_drm_private *priv = dev->dev_private; | ||
76 | struct platform_device *pdev = priv->hdmi_pdev; | ||
77 | struct hdmi_platform_config *config; | ||
78 | int i, ret; | 79 | int i, ret; |
79 | 80 | ||
80 | if (!pdev) { | 81 | hdmi = devm_kzalloc(&pdev->dev, sizeof(*hdmi), GFP_KERNEL); |
81 | dev_err(dev->dev, "no hdmi device\n"); | ||
82 | ret = -ENXIO; | ||
83 | goto fail; | ||
84 | } | ||
85 | |||
86 | config = pdev->dev.platform_data; | ||
87 | |||
88 | hdmi = kzalloc(sizeof(*hdmi), GFP_KERNEL); | ||
89 | if (!hdmi) { | 82 | if (!hdmi) { |
90 | ret = -ENOMEM; | 83 | ret = -ENOMEM; |
91 | goto fail; | 84 | goto fail; |
92 | } | 85 | } |
93 | 86 | ||
94 | kref_init(&hdmi->refcount); | ||
95 | |||
96 | hdmi->dev = dev; | ||
97 | hdmi->pdev = pdev; | 87 | hdmi->pdev = pdev; |
98 | hdmi->config = config; | 88 | hdmi->config = config; |
99 | hdmi->encoder = encoder; | ||
100 | |||
101 | hdmi_audio_infoframe_init(&hdmi->audio.infoframe); | ||
102 | 89 | ||
103 | /* not sure about which phy maps to which msm.. probably I miss some */ | 90 | /* not sure about which phy maps to which msm.. probably I miss some */ |
104 | if (config->phy_init) | 91 | if (config->phy_init) |
@@ -108,7 +95,7 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
108 | 95 | ||
109 | if (IS_ERR(hdmi->phy)) { | 96 | if (IS_ERR(hdmi->phy)) { |
110 | ret = PTR_ERR(hdmi->phy); | 97 | ret = PTR_ERR(hdmi->phy); |
111 | dev_err(dev->dev, "failed to load phy: %d\n", ret); | 98 | dev_err(&pdev->dev, "failed to load phy: %d\n", ret); |
112 | hdmi->phy = NULL; | 99 | hdmi->phy = NULL; |
113 | goto fail; | 100 | goto fail; |
114 | } | 101 | } |
@@ -127,7 +114,7 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
127 | config->hpd_reg_names[i]); | 114 | config->hpd_reg_names[i]); |
128 | if (IS_ERR(reg)) { | 115 | if (IS_ERR(reg)) { |
129 | ret = PTR_ERR(reg); | 116 | ret = PTR_ERR(reg); |
130 | dev_err(dev->dev, "failed to get hpd regulator: %s (%d)\n", | 117 | dev_err(&pdev->dev, "failed to get hpd regulator: %s (%d)\n", |
131 | config->hpd_reg_names[i], ret); | 118 | config->hpd_reg_names[i], ret); |
132 | goto fail; | 119 | goto fail; |
133 | } | 120 | } |
@@ -143,7 +130,7 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
143 | config->pwr_reg_names[i]); | 130 | config->pwr_reg_names[i]); |
144 | if (IS_ERR(reg)) { | 131 | if (IS_ERR(reg)) { |
145 | ret = PTR_ERR(reg); | 132 | ret = PTR_ERR(reg); |
146 | dev_err(dev->dev, "failed to get pwr regulator: %s (%d)\n", | 133 | dev_err(&pdev->dev, "failed to get pwr regulator: %s (%d)\n", |
147 | config->pwr_reg_names[i], ret); | 134 | config->pwr_reg_names[i], ret); |
148 | goto fail; | 135 | goto fail; |
149 | } | 136 | } |
@@ -158,7 +145,7 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
158 | clk = devm_clk_get(&pdev->dev, config->hpd_clk_names[i]); | 145 | clk = devm_clk_get(&pdev->dev, config->hpd_clk_names[i]); |
159 | if (IS_ERR(clk)) { | 146 | if (IS_ERR(clk)) { |
160 | ret = PTR_ERR(clk); | 147 | ret = PTR_ERR(clk); |
161 | dev_err(dev->dev, "failed to get hpd clk: %s (%d)\n", | 148 | dev_err(&pdev->dev, "failed to get hpd clk: %s (%d)\n", |
162 | config->hpd_clk_names[i], ret); | 149 | config->hpd_clk_names[i], ret); |
163 | goto fail; | 150 | goto fail; |
164 | } | 151 | } |
@@ -173,7 +160,7 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
173 | clk = devm_clk_get(&pdev->dev, config->pwr_clk_names[i]); | 160 | clk = devm_clk_get(&pdev->dev, config->pwr_clk_names[i]); |
174 | if (IS_ERR(clk)) { | 161 | if (IS_ERR(clk)) { |
175 | ret = PTR_ERR(clk); | 162 | ret = PTR_ERR(clk); |
176 | dev_err(dev->dev, "failed to get pwr clk: %s (%d)\n", | 163 | dev_err(&pdev->dev, "failed to get pwr clk: %s (%d)\n", |
177 | config->pwr_clk_names[i], ret); | 164 | config->pwr_clk_names[i], ret); |
178 | goto fail; | 165 | goto fail; |
179 | } | 166 | } |
@@ -184,11 +171,40 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
184 | hdmi->i2c = hdmi_i2c_init(hdmi); | 171 | hdmi->i2c = hdmi_i2c_init(hdmi); |
185 | if (IS_ERR(hdmi->i2c)) { | 172 | if (IS_ERR(hdmi->i2c)) { |
186 | ret = PTR_ERR(hdmi->i2c); | 173 | ret = PTR_ERR(hdmi->i2c); |
187 | dev_err(dev->dev, "failed to get i2c: %d\n", ret); | 174 | dev_err(&pdev->dev, "failed to get i2c: %d\n", ret); |
188 | hdmi->i2c = NULL; | 175 | hdmi->i2c = NULL; |
189 | goto fail; | 176 | goto fail; |
190 | } | 177 | } |
191 | 178 | ||
179 | return hdmi; | ||
180 | |||
181 | fail: | ||
182 | if (hdmi) | ||
183 | hdmi_destroy(hdmi); | ||
184 | |||
185 | return ERR_PTR(ret); | ||
186 | } | ||
187 | |||
188 | /* Second part of initialization, the drm/kms level modeset_init, | ||
189 | * constructs/initializes mode objects, etc, is called from master | ||
190 | * driver (not hdmi sub-device's probe/bind!) | ||
191 | * | ||
192 | * Any resource (regulator/clk/etc) which could be missing at boot | ||
193 | * should be handled in hdmi_init() so that failure happens from | ||
194 | * hdmi sub-device's probe. | ||
195 | */ | ||
196 | int hdmi_modeset_init(struct hdmi *hdmi, | ||
197 | struct drm_device *dev, struct drm_encoder *encoder) | ||
198 | { | ||
199 | struct msm_drm_private *priv = dev->dev_private; | ||
200 | struct platform_device *pdev = hdmi->pdev; | ||
201 | int ret; | ||
202 | |||
203 | hdmi->dev = dev; | ||
204 | hdmi->encoder = encoder; | ||
205 | |||
206 | hdmi_audio_infoframe_init(&hdmi->audio.infoframe); | ||
207 | |||
192 | hdmi->bridge = hdmi_bridge_init(hdmi); | 208 | hdmi->bridge = hdmi_bridge_init(hdmi); |
193 | if (IS_ERR(hdmi->bridge)) { | 209 | if (IS_ERR(hdmi->bridge)) { |
194 | ret = PTR_ERR(hdmi->bridge); | 210 | ret = PTR_ERR(hdmi->bridge); |
@@ -205,22 +221,20 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
205 | goto fail; | 221 | goto fail; |
206 | } | 222 | } |
207 | 223 | ||
208 | if (!config->shared_irq) { | 224 | hdmi->irq = irq_of_parse_and_map(pdev->dev.of_node, 0); |
209 | hdmi->irq = platform_get_irq(pdev, 0); | 225 | if (hdmi->irq < 0) { |
210 | if (hdmi->irq < 0) { | 226 | ret = hdmi->irq; |
211 | ret = hdmi->irq; | 227 | dev_err(dev->dev, "failed to get irq: %d\n", ret); |
212 | dev_err(dev->dev, "failed to get irq: %d\n", ret); | 228 | goto fail; |
213 | goto fail; | 229 | } |
214 | } | ||
215 | 230 | ||
216 | ret = devm_request_threaded_irq(&pdev->dev, hdmi->irq, | 231 | ret = devm_request_irq(&pdev->dev, hdmi->irq, |
217 | NULL, hdmi_irq, IRQF_TRIGGER_HIGH | IRQF_ONESHOT, | 232 | hdmi_irq, IRQF_TRIGGER_HIGH | IRQF_ONESHOT, |
218 | "hdmi_isr", hdmi); | 233 | "hdmi_isr", hdmi); |
219 | if (ret < 0) { | 234 | if (ret < 0) { |
220 | dev_err(dev->dev, "failed to request IRQ%u: %d\n", | 235 | dev_err(dev->dev, "failed to request IRQ%u: %d\n", |
221 | hdmi->irq, ret); | 236 | hdmi->irq, ret); |
222 | goto fail; | 237 | goto fail; |
223 | } | ||
224 | } | 238 | } |
225 | 239 | ||
226 | encoder->bridge = hdmi->bridge; | 240 | encoder->bridge = hdmi->bridge; |
@@ -230,19 +244,20 @@ struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder) | |||
230 | 244 | ||
231 | platform_set_drvdata(pdev, hdmi); | 245 | platform_set_drvdata(pdev, hdmi); |
232 | 246 | ||
233 | return hdmi; | 247 | return 0; |
234 | 248 | ||
235 | fail: | 249 | fail: |
236 | if (hdmi) { | 250 | /* bridge/connector are normally destroyed by drm: */ |
237 | /* bridge/connector are normally destroyed by drm: */ | 251 | if (hdmi->bridge) { |
238 | if (hdmi->bridge) | 252 | hdmi->bridge->funcs->destroy(hdmi->bridge); |
239 | hdmi->bridge->funcs->destroy(hdmi->bridge); | 253 | hdmi->bridge = NULL; |
240 | if (hdmi->connector) | 254 | } |
241 | hdmi->connector->funcs->destroy(hdmi->connector); | 255 | if (hdmi->connector) { |
242 | hdmi_destroy(&hdmi->refcount); | 256 | hdmi->connector->funcs->destroy(hdmi->connector); |
257 | hdmi->connector = NULL; | ||
243 | } | 258 | } |
244 | 259 | ||
245 | return ERR_PTR(ret); | 260 | return ret; |
246 | } | 261 | } |
247 | 262 | ||
248 | /* | 263 | /* |
@@ -251,13 +266,6 @@ fail: | |||
251 | 266 | ||
252 | #include <linux/of_gpio.h> | 267 | #include <linux/of_gpio.h> |
253 | 268 | ||
254 | static void set_hdmi_pdev(struct drm_device *dev, | ||
255 | struct platform_device *pdev) | ||
256 | { | ||
257 | struct msm_drm_private *priv = dev->dev_private; | ||
258 | priv->hdmi_pdev = pdev; | ||
259 | } | ||
260 | |||
261 | #ifdef CONFIG_OF | 269 | #ifdef CONFIG_OF |
262 | static int get_gpio(struct device *dev, struct device_node *of_node, const char *name) | 270 | static int get_gpio(struct device *dev, struct device_node *of_node, const char *name) |
263 | { | 271 | { |
@@ -278,7 +286,10 @@ static int get_gpio(struct device *dev, struct device_node *of_node, const char | |||
278 | 286 | ||
279 | static int hdmi_bind(struct device *dev, struct device *master, void *data) | 287 | static int hdmi_bind(struct device *dev, struct device *master, void *data) |
280 | { | 288 | { |
289 | struct drm_device *drm = dev_get_drvdata(master); | ||
290 | struct msm_drm_private *priv = drm->dev_private; | ||
281 | static struct hdmi_platform_config config = {}; | 291 | static struct hdmi_platform_config config = {}; |
292 | struct hdmi *hdmi; | ||
282 | #ifdef CONFIG_OF | 293 | #ifdef CONFIG_OF |
283 | struct device_node *of_node = dev->of_node; | 294 | struct device_node *of_node = dev->of_node; |
284 | 295 | ||
@@ -298,7 +309,6 @@ static int hdmi_bind(struct device *dev, struct device *master, void *data) | |||
298 | config.hpd_clk_cnt = ARRAY_SIZE(hpd_clk_names); | 309 | config.hpd_clk_cnt = ARRAY_SIZE(hpd_clk_names); |
299 | config.pwr_clk_names = pwr_clk_names; | 310 | config.pwr_clk_names = pwr_clk_names; |
300 | config.pwr_clk_cnt = ARRAY_SIZE(pwr_clk_names); | 311 | config.pwr_clk_cnt = ARRAY_SIZE(pwr_clk_names); |
301 | config.shared_irq = true; | ||
302 | } else if (of_device_is_compatible(of_node, "qcom,hdmi-tx-8960")) { | 312 | } else if (of_device_is_compatible(of_node, "qcom,hdmi-tx-8960")) { |
303 | static const char *hpd_clk_names[] = {"core_clk", "master_iface_clk", "slave_iface_clk"}; | 313 | static const char *hpd_clk_names[] = {"core_clk", "master_iface_clk", "slave_iface_clk"}; |
304 | static const char *hpd_reg_names[] = {"core-vdda", "hdmi-mux"}; | 314 | static const char *hpd_reg_names[] = {"core-vdda", "hdmi-mux"}; |
@@ -369,14 +379,22 @@ static int hdmi_bind(struct device *dev, struct device *master, void *data) | |||
369 | } | 379 | } |
370 | #endif | 380 | #endif |
371 | dev->platform_data = &config; | 381 | dev->platform_data = &config; |
372 | set_hdmi_pdev(dev_get_drvdata(master), to_platform_device(dev)); | 382 | hdmi = hdmi_init(to_platform_device(dev)); |
383 | if (IS_ERR(hdmi)) | ||
384 | return PTR_ERR(hdmi); | ||
385 | priv->hdmi = hdmi; | ||
373 | return 0; | 386 | return 0; |
374 | } | 387 | } |
375 | 388 | ||
376 | static void hdmi_unbind(struct device *dev, struct device *master, | 389 | static void hdmi_unbind(struct device *dev, struct device *master, |
377 | void *data) | 390 | void *data) |
378 | { | 391 | { |
379 | set_hdmi_pdev(dev_get_drvdata(master), NULL); | 392 | struct drm_device *drm = dev_get_drvdata(master); |
393 | struct msm_drm_private *priv = drm->dev_private; | ||
394 | if (priv->hdmi) { | ||
395 | hdmi_destroy(priv->hdmi); | ||
396 | priv->hdmi = NULL; | ||
397 | } | ||
380 | } | 398 | } |
381 | 399 | ||
382 | static const struct component_ops hdmi_ops = { | 400 | static const struct component_ops hdmi_ops = { |
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi.h b/drivers/gpu/drm/msm/hdmi/hdmi.h index b981995410b5..43e654f751b7 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi.h +++ b/drivers/gpu/drm/msm/hdmi/hdmi.h | |||
@@ -38,8 +38,6 @@ struct hdmi_audio { | |||
38 | }; | 38 | }; |
39 | 39 | ||
40 | struct hdmi { | 40 | struct hdmi { |
41 | struct kref refcount; | ||
42 | |||
43 | struct drm_device *dev; | 41 | struct drm_device *dev; |
44 | struct platform_device *pdev; | 42 | struct platform_device *pdev; |
45 | 43 | ||
@@ -97,13 +95,9 @@ struct hdmi_platform_config { | |||
97 | /* gpio's: */ | 95 | /* gpio's: */ |
98 | int ddc_clk_gpio, ddc_data_gpio, hpd_gpio, mux_en_gpio, mux_sel_gpio; | 96 | int ddc_clk_gpio, ddc_data_gpio, hpd_gpio, mux_en_gpio, mux_sel_gpio; |
99 | int mux_lpm_gpio; | 97 | int mux_lpm_gpio; |
100 | |||
101 | /* older devices had their own irq, mdp5+ it is shared w/ mdp: */ | ||
102 | bool shared_irq; | ||
103 | }; | 98 | }; |
104 | 99 | ||
105 | void hdmi_set_mode(struct hdmi *hdmi, bool power_on); | 100 | void hdmi_set_mode(struct hdmi *hdmi, bool power_on); |
106 | void hdmi_destroy(struct kref *kref); | ||
107 | 101 | ||
108 | static inline void hdmi_write(struct hdmi *hdmi, u32 reg, u32 data) | 102 | static inline void hdmi_write(struct hdmi *hdmi, u32 reg, u32 data) |
109 | { | 103 | { |
@@ -115,17 +109,6 @@ static inline u32 hdmi_read(struct hdmi *hdmi, u32 reg) | |||
115 | return msm_readl(hdmi->mmio + reg); | 109 | return msm_readl(hdmi->mmio + reg); |
116 | } | 110 | } |
117 | 111 | ||
118 | static inline struct hdmi * hdmi_reference(struct hdmi *hdmi) | ||
119 | { | ||
120 | kref_get(&hdmi->refcount); | ||
121 | return hdmi; | ||
122 | } | ||
123 | |||
124 | static inline void hdmi_unreference(struct hdmi *hdmi) | ||
125 | { | ||
126 | kref_put(&hdmi->refcount, hdmi_destroy); | ||
127 | } | ||
128 | |||
129 | /* | 112 | /* |
130 | * The phy appears to be different, for example between 8960 and 8x60, | 113 | * The phy appears to be different, for example between 8960 and 8x60, |
131 | * so split the phy related functions out and load the correct one at | 114 | * so split the phy related functions out and load the correct one at |
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi.xml.h b/drivers/gpu/drm/msm/hdmi/hdmi.xml.h index 76fd0cfc6558..5b0844befbab 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi.xml.h +++ b/drivers/gpu/drm/msm/hdmi/hdmi.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi_bridge.c b/drivers/gpu/drm/msm/hdmi/hdmi_bridge.c index f6cf745c249e..6902ad6da710 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi_bridge.c +++ b/drivers/gpu/drm/msm/hdmi/hdmi_bridge.c | |||
@@ -26,7 +26,6 @@ struct hdmi_bridge { | |||
26 | static void hdmi_bridge_destroy(struct drm_bridge *bridge) | 26 | static void hdmi_bridge_destroy(struct drm_bridge *bridge) |
27 | { | 27 | { |
28 | struct hdmi_bridge *hdmi_bridge = to_hdmi_bridge(bridge); | 28 | struct hdmi_bridge *hdmi_bridge = to_hdmi_bridge(bridge); |
29 | hdmi_unreference(hdmi_bridge->hdmi); | ||
30 | drm_bridge_cleanup(bridge); | 29 | drm_bridge_cleanup(bridge); |
31 | kfree(hdmi_bridge); | 30 | kfree(hdmi_bridge); |
32 | } | 31 | } |
@@ -218,7 +217,7 @@ struct drm_bridge *hdmi_bridge_init(struct hdmi *hdmi) | |||
218 | goto fail; | 217 | goto fail; |
219 | } | 218 | } |
220 | 219 | ||
221 | hdmi_bridge->hdmi = hdmi_reference(hdmi); | 220 | hdmi_bridge->hdmi = hdmi; |
222 | 221 | ||
223 | bridge = &hdmi_bridge->base; | 222 | bridge = &hdmi_bridge->base; |
224 | 223 | ||
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi_connector.c b/drivers/gpu/drm/msm/hdmi/hdmi_connector.c index 4aca2a3c667c..fbebb0405d76 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi_connector.c +++ b/drivers/gpu/drm/msm/hdmi/hdmi_connector.c | |||
@@ -330,8 +330,6 @@ static void hdmi_connector_destroy(struct drm_connector *connector) | |||
330 | drm_connector_unregister(connector); | 330 | drm_connector_unregister(connector); |
331 | drm_connector_cleanup(connector); | 331 | drm_connector_cleanup(connector); |
332 | 332 | ||
333 | hdmi_unreference(hdmi_connector->hdmi); | ||
334 | |||
335 | kfree(hdmi_connector); | 333 | kfree(hdmi_connector); |
336 | } | 334 | } |
337 | 335 | ||
@@ -401,6 +399,9 @@ static const struct drm_connector_funcs hdmi_connector_funcs = { | |||
401 | .detect = hdmi_connector_detect, | 399 | .detect = hdmi_connector_detect, |
402 | .fill_modes = drm_helper_probe_single_connector_modes, | 400 | .fill_modes = drm_helper_probe_single_connector_modes, |
403 | .destroy = hdmi_connector_destroy, | 401 | .destroy = hdmi_connector_destroy, |
402 | .reset = drm_atomic_helper_connector_reset, | ||
403 | .atomic_duplicate_state = drm_atomic_helper_connector_duplicate_state, | ||
404 | .atomic_destroy_state = drm_atomic_helper_connector_destroy_state, | ||
404 | }; | 405 | }; |
405 | 406 | ||
406 | static const struct drm_connector_helper_funcs hdmi_connector_helper_funcs = { | 407 | static const struct drm_connector_helper_funcs hdmi_connector_helper_funcs = { |
@@ -422,7 +423,7 @@ struct drm_connector *hdmi_connector_init(struct hdmi *hdmi) | |||
422 | goto fail; | 423 | goto fail; |
423 | } | 424 | } |
424 | 425 | ||
425 | hdmi_connector->hdmi = hdmi_reference(hdmi); | 426 | hdmi_connector->hdmi = hdmi; |
426 | INIT_WORK(&hdmi_connector->hpd_work, hotplug_work); | 427 | INIT_WORK(&hdmi_connector->hpd_work, hotplug_work); |
427 | 428 | ||
428 | connector = &hdmi_connector->base; | 429 | connector = &hdmi_connector->base; |
diff --git a/drivers/gpu/drm/msm/hdmi/hdmi_phy_8960.c b/drivers/gpu/drm/msm/hdmi/hdmi_phy_8960.c index f408b69486a8..eeed006eed13 100644 --- a/drivers/gpu/drm/msm/hdmi/hdmi_phy_8960.c +++ b/drivers/gpu/drm/msm/hdmi/hdmi_phy_8960.c | |||
@@ -510,7 +510,7 @@ struct hdmi_phy *hdmi_phy_8960_init(struct hdmi *hdmi) | |||
510 | 510 | ||
511 | #ifdef CONFIG_COMMON_CLK | 511 | #ifdef CONFIG_COMMON_CLK |
512 | phy_8960->pll_hw.init = &pll_init; | 512 | phy_8960->pll_hw.init = &pll_init; |
513 | phy_8960->pll = devm_clk_register(hdmi->dev->dev, &phy_8960->pll_hw); | 513 | phy_8960->pll = devm_clk_register(&hdmi->pdev->dev, &phy_8960->pll_hw); |
514 | if (IS_ERR(phy_8960->pll)) { | 514 | if (IS_ERR(phy_8960->pll)) { |
515 | ret = PTR_ERR(phy_8960->pll); | 515 | ret = PTR_ERR(phy_8960->pll); |
516 | phy_8960->pll = NULL; | 516 | phy_8960->pll = NULL; |
diff --git a/drivers/gpu/drm/msm/hdmi/qfprom.xml.h b/drivers/gpu/drm/msm/hdmi/qfprom.xml.h index d53c29327df9..29bd796797de 100644 --- a/drivers/gpu/drm/msm/hdmi/qfprom.xml.h +++ b/drivers/gpu/drm/msm/hdmi/qfprom.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4.xml.h b/drivers/gpu/drm/msm/mdp/mdp4/mdp4.xml.h index 03c0bd9cd5b9..a4a7f8c7122a 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4.xml.h +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4.xml.h | |||
@@ -10,12 +10,12 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20457 bytes, from 2014-08-01 12:22:48) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2014-07-17 15:34:33) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-07-17 15:34:33) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-08-01 12:23:53) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_crtc.c b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_crtc.c index 7d00f7fb5773..a7672e100d8b 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_crtc.c +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_crtc.c | |||
@@ -25,8 +25,6 @@ | |||
25 | struct mdp4_crtc { | 25 | struct mdp4_crtc { |
26 | struct drm_crtc base; | 26 | struct drm_crtc base; |
27 | char name[8]; | 27 | char name[8]; |
28 | struct drm_plane *plane; | ||
29 | struct drm_plane *planes[8]; | ||
30 | int id; | 28 | int id; |
31 | int ovlp; | 29 | int ovlp; |
32 | enum mdp4_dma dma; | 30 | enum mdp4_dma dma; |
@@ -52,25 +50,11 @@ struct mdp4_crtc { | |||
52 | 50 | ||
53 | /* if there is a pending flip, these will be non-null: */ | 51 | /* if there is a pending flip, these will be non-null: */ |
54 | struct drm_pending_vblank_event *event; | 52 | struct drm_pending_vblank_event *event; |
55 | struct msm_fence_cb pageflip_cb; | ||
56 | 53 | ||
57 | #define PENDING_CURSOR 0x1 | 54 | #define PENDING_CURSOR 0x1 |
58 | #define PENDING_FLIP 0x2 | 55 | #define PENDING_FLIP 0x2 |
59 | atomic_t pending; | 56 | atomic_t pending; |
60 | 57 | ||
61 | /* the fb that we logically (from PoV of KMS API) hold a ref | ||
62 | * to. Which we may not yet be scanning out (we may still | ||
63 | * be scanning out previous in case of page_flip while waiting | ||
64 | * for gpu rendering to complete: | ||
65 | */ | ||
66 | struct drm_framebuffer *fb; | ||
67 | |||
68 | /* the fb that we currently hold a scanout ref to: */ | ||
69 | struct drm_framebuffer *scanout_fb; | ||
70 | |||
71 | /* for unref'ing framebuffers after scanout completes: */ | ||
72 | struct drm_flip_work unref_fb_work; | ||
73 | |||
74 | /* for unref'ing cursor bo's after scanout completes: */ | 58 | /* for unref'ing cursor bo's after scanout completes: */ |
75 | struct drm_flip_work unref_cursor_work; | 59 | struct drm_flip_work unref_cursor_work; |
76 | 60 | ||
@@ -97,15 +81,14 @@ static void crtc_flush(struct drm_crtc *crtc) | |||
97 | { | 81 | { |
98 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 82 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
99 | struct mdp4_kms *mdp4_kms = get_kms(crtc); | 83 | struct mdp4_kms *mdp4_kms = get_kms(crtc); |
100 | uint32_t i, flush = 0; | 84 | struct drm_plane *plane; |
85 | uint32_t flush = 0; | ||
101 | 86 | ||
102 | for (i = 0; i < ARRAY_SIZE(mdp4_crtc->planes); i++) { | 87 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
103 | struct drm_plane *plane = mdp4_crtc->planes[i]; | 88 | enum mdp4_pipe pipe_id = mdp4_plane_pipe(plane); |
104 | if (plane) { | 89 | flush |= pipe2flush(pipe_id); |
105 | enum mdp4_pipe pipe_id = mdp4_plane_pipe(plane); | ||
106 | flush |= pipe2flush(pipe_id); | ||
107 | } | ||
108 | } | 90 | } |
91 | |||
109 | flush |= ovlp2flush(mdp4_crtc->ovlp); | 92 | flush |= ovlp2flush(mdp4_crtc->ovlp); |
110 | 93 | ||
111 | DBG("%s: flush=%08x", mdp4_crtc->name, flush); | 94 | DBG("%s: flush=%08x", mdp4_crtc->name, flush); |
@@ -113,47 +96,6 @@ static void crtc_flush(struct drm_crtc *crtc) | |||
113 | mdp4_write(mdp4_kms, REG_MDP4_OVERLAY_FLUSH, flush); | 96 | mdp4_write(mdp4_kms, REG_MDP4_OVERLAY_FLUSH, flush); |
114 | } | 97 | } |
115 | 98 | ||
116 | static void update_fb(struct drm_crtc *crtc, struct drm_framebuffer *new_fb) | ||
117 | { | ||
118 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | ||
119 | struct drm_framebuffer *old_fb = mdp4_crtc->fb; | ||
120 | |||
121 | /* grab reference to incoming scanout fb: */ | ||
122 | drm_framebuffer_reference(new_fb); | ||
123 | mdp4_crtc->base.primary->fb = new_fb; | ||
124 | mdp4_crtc->fb = new_fb; | ||
125 | |||
126 | if (old_fb) | ||
127 | drm_flip_work_queue(&mdp4_crtc->unref_fb_work, old_fb); | ||
128 | } | ||
129 | |||
130 | /* unlike update_fb(), take a ref to the new scanout fb *before* updating | ||
131 | * plane, then call this. Needed to ensure we don't unref the buffer that | ||
132 | * is actually still being scanned out. | ||
133 | * | ||
134 | * Note that this whole thing goes away with atomic.. since we can defer | ||
135 | * calling into driver until rendering is done. | ||
136 | */ | ||
137 | static void update_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb) | ||
138 | { | ||
139 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | ||
140 | |||
141 | /* flush updates, to make sure hw is updated to new scanout fb, | ||
142 | * so that we can safely queue unref to current fb (ie. next | ||
143 | * vblank we know hw is done w/ previous scanout_fb). | ||
144 | */ | ||
145 | crtc_flush(crtc); | ||
146 | |||
147 | if (mdp4_crtc->scanout_fb) | ||
148 | drm_flip_work_queue(&mdp4_crtc->unref_fb_work, | ||
149 | mdp4_crtc->scanout_fb); | ||
150 | |||
151 | mdp4_crtc->scanout_fb = fb; | ||
152 | |||
153 | /* enable vblank to complete flip: */ | ||
154 | request_pending(crtc, PENDING_FLIP); | ||
155 | } | ||
156 | |||
157 | /* if file!=NULL, this is preclose potential cancel-flip path */ | 99 | /* if file!=NULL, this is preclose potential cancel-flip path */ |
158 | static void complete_flip(struct drm_crtc *crtc, struct drm_file *file) | 100 | static void complete_flip(struct drm_crtc *crtc, struct drm_file *file) |
159 | { | 101 | { |
@@ -171,38 +113,13 @@ static void complete_flip(struct drm_crtc *crtc, struct drm_file *file) | |||
171 | */ | 113 | */ |
172 | if (!file || (event->base.file_priv == file)) { | 114 | if (!file || (event->base.file_priv == file)) { |
173 | mdp4_crtc->event = NULL; | 115 | mdp4_crtc->event = NULL; |
116 | DBG("%s: send event: %p", mdp4_crtc->name, event); | ||
174 | drm_send_vblank_event(dev, mdp4_crtc->id, event); | 117 | drm_send_vblank_event(dev, mdp4_crtc->id, event); |
175 | } | 118 | } |
176 | } | 119 | } |
177 | spin_unlock_irqrestore(&dev->event_lock, flags); | 120 | spin_unlock_irqrestore(&dev->event_lock, flags); |
178 | } | 121 | } |
179 | 122 | ||
180 | static void pageflip_cb(struct msm_fence_cb *cb) | ||
181 | { | ||
182 | struct mdp4_crtc *mdp4_crtc = | ||
183 | container_of(cb, struct mdp4_crtc, pageflip_cb); | ||
184 | struct drm_crtc *crtc = &mdp4_crtc->base; | ||
185 | struct drm_framebuffer *fb = crtc->primary->fb; | ||
186 | |||
187 | if (!fb) | ||
188 | return; | ||
189 | |||
190 | drm_framebuffer_reference(fb); | ||
191 | mdp4_plane_set_scanout(mdp4_crtc->plane, fb); | ||
192 | update_scanout(crtc, fb); | ||
193 | } | ||
194 | |||
195 | static void unref_fb_worker(struct drm_flip_work *work, void *val) | ||
196 | { | ||
197 | struct mdp4_crtc *mdp4_crtc = | ||
198 | container_of(work, struct mdp4_crtc, unref_fb_work); | ||
199 | struct drm_device *dev = mdp4_crtc->base.dev; | ||
200 | |||
201 | mutex_lock(&dev->mode_config.mutex); | ||
202 | drm_framebuffer_unreference(val); | ||
203 | mutex_unlock(&dev->mode_config.mutex); | ||
204 | } | ||
205 | |||
206 | static void unref_cursor_worker(struct drm_flip_work *work, void *val) | 123 | static void unref_cursor_worker(struct drm_flip_work *work, void *val) |
207 | { | 124 | { |
208 | struct mdp4_crtc *mdp4_crtc = | 125 | struct mdp4_crtc *mdp4_crtc = |
@@ -218,7 +135,6 @@ static void mdp4_crtc_destroy(struct drm_crtc *crtc) | |||
218 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 135 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
219 | 136 | ||
220 | drm_crtc_cleanup(crtc); | 137 | drm_crtc_cleanup(crtc); |
221 | drm_flip_work_cleanup(&mdp4_crtc->unref_fb_work); | ||
222 | drm_flip_work_cleanup(&mdp4_crtc->unref_cursor_work); | 138 | drm_flip_work_cleanup(&mdp4_crtc->unref_cursor_work); |
223 | 139 | ||
224 | kfree(mdp4_crtc); | 140 | kfree(mdp4_crtc); |
@@ -251,57 +167,70 @@ static bool mdp4_crtc_mode_fixup(struct drm_crtc *crtc, | |||
251 | return true; | 167 | return true; |
252 | } | 168 | } |
253 | 169 | ||
254 | static void blend_setup(struct drm_crtc *crtc) | 170 | /* statically (for now) map planes to mixer stage (z-order): */ |
171 | static const int idxs[] = { | ||
172 | [VG1] = 1, | ||
173 | [VG2] = 2, | ||
174 | [RGB1] = 0, | ||
175 | [RGB2] = 0, | ||
176 | [RGB3] = 0, | ||
177 | [VG3] = 3, | ||
178 | [VG4] = 4, | ||
179 | |||
180 | }; | ||
181 | |||
182 | /* setup mixer config, for which we need to consider all crtc's and | ||
183 | * the planes attached to them | ||
184 | * | ||
185 | * TODO may possibly need some extra locking here | ||
186 | */ | ||
187 | static void setup_mixer(struct mdp4_kms *mdp4_kms) | ||
255 | { | 188 | { |
256 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 189 | struct drm_mode_config *config = &mdp4_kms->dev->mode_config; |
257 | struct mdp4_kms *mdp4_kms = get_kms(crtc); | 190 | struct drm_crtc *crtc; |
258 | int i, ovlp = mdp4_crtc->ovlp; | ||
259 | uint32_t mixer_cfg = 0; | 191 | uint32_t mixer_cfg = 0; |
260 | static const enum mdp_mixer_stage_id stages[] = { | 192 | static const enum mdp_mixer_stage_id stages[] = { |
261 | STAGE_BASE, STAGE0, STAGE1, STAGE2, STAGE3, | 193 | STAGE_BASE, STAGE0, STAGE1, STAGE2, STAGE3, |
262 | }; | 194 | }; |
263 | /* statically (for now) map planes to mixer stage (z-order): */ | ||
264 | static const int idxs[] = { | ||
265 | [VG1] = 1, | ||
266 | [VG2] = 2, | ||
267 | [RGB1] = 0, | ||
268 | [RGB2] = 0, | ||
269 | [RGB3] = 0, | ||
270 | [VG3] = 3, | ||
271 | [VG4] = 4, | ||
272 | 195 | ||
273 | }; | 196 | list_for_each_entry(crtc, &config->crtc_list, head) { |
274 | bool alpha[4]= { false, false, false, false }; | 197 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
198 | struct drm_plane *plane; | ||
275 | 199 | ||
276 | /* Don't rely on value read back from hw, but instead use our | 200 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
277 | * own shadowed value. Possibly disable/reenable looses the | 201 | enum mdp4_pipe pipe_id = mdp4_plane_pipe(plane); |
278 | * previous value and goes back to power-on default? | 202 | int idx = idxs[pipe_id]; |
279 | */ | 203 | mixer_cfg = mixercfg(mixer_cfg, mdp4_crtc->mixer, |
280 | mixer_cfg = mdp4_kms->mixer_cfg; | 204 | pipe_id, stages[idx]); |
205 | } | ||
206 | } | ||
207 | |||
208 | mdp4_write(mdp4_kms, REG_MDP4_LAYERMIXER_IN_CFG, mixer_cfg); | ||
209 | } | ||
210 | |||
211 | static void blend_setup(struct drm_crtc *crtc) | ||
212 | { | ||
213 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | ||
214 | struct mdp4_kms *mdp4_kms = get_kms(crtc); | ||
215 | struct drm_plane *plane; | ||
216 | int i, ovlp = mdp4_crtc->ovlp; | ||
217 | bool alpha[4]= { false, false, false, false }; | ||
281 | 218 | ||
282 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_LOW0(ovlp), 0); | 219 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_LOW0(ovlp), 0); |
283 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_LOW1(ovlp), 0); | 220 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_LOW1(ovlp), 0); |
284 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_HIGH0(ovlp), 0); | 221 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_HIGH0(ovlp), 0); |
285 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_HIGH1(ovlp), 0); | 222 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_TRANSP_HIGH1(ovlp), 0); |
286 | 223 | ||
287 | for (i = 0; i < ARRAY_SIZE(mdp4_crtc->planes); i++) { | 224 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
288 | struct drm_plane *plane = mdp4_crtc->planes[i]; | 225 | enum mdp4_pipe pipe_id = mdp4_plane_pipe(plane); |
289 | if (plane) { | 226 | int idx = idxs[pipe_id]; |
290 | enum mdp4_pipe pipe_id = mdp4_plane_pipe(plane); | 227 | if (idx > 0) { |
291 | int idx = idxs[pipe_id]; | 228 | const struct mdp_format *format = |
292 | if (idx > 0) { | ||
293 | const struct mdp_format *format = | ||
294 | to_mdp_format(msm_framebuffer_format(plane->fb)); | 229 | to_mdp_format(msm_framebuffer_format(plane->fb)); |
295 | alpha[idx-1] = format->alpha_enable; | 230 | alpha[idx-1] = format->alpha_enable; |
296 | } | ||
297 | mixer_cfg = mixercfg(mixer_cfg, mdp4_crtc->mixer, | ||
298 | pipe_id, stages[idx]); | ||
299 | } | 231 | } |
300 | } | 232 | } |
301 | 233 | ||
302 | /* this shouldn't happen.. and seems to cause underflow: */ | ||
303 | WARN_ON(!mixer_cfg); | ||
304 | |||
305 | for (i = 0; i < 4; i++) { | 234 | for (i = 0; i < 4; i++) { |
306 | uint32_t op; | 235 | uint32_t op; |
307 | 236 | ||
@@ -324,22 +253,21 @@ static void blend_setup(struct drm_crtc *crtc) | |||
324 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_STAGE_TRANSP_HIGH1(ovlp, i), 0); | 253 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_STAGE_TRANSP_HIGH1(ovlp, i), 0); |
325 | } | 254 | } |
326 | 255 | ||
327 | mdp4_kms->mixer_cfg = mixer_cfg; | 256 | setup_mixer(mdp4_kms); |
328 | mdp4_write(mdp4_kms, REG_MDP4_LAYERMIXER_IN_CFG, mixer_cfg); | ||
329 | } | 257 | } |
330 | 258 | ||
331 | static int mdp4_crtc_mode_set(struct drm_crtc *crtc, | 259 | static void mdp4_crtc_mode_set_nofb(struct drm_crtc *crtc) |
332 | struct drm_display_mode *mode, | ||
333 | struct drm_display_mode *adjusted_mode, | ||
334 | int x, int y, | ||
335 | struct drm_framebuffer *old_fb) | ||
336 | { | 260 | { |
337 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 261 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
338 | struct mdp4_kms *mdp4_kms = get_kms(crtc); | 262 | struct mdp4_kms *mdp4_kms = get_kms(crtc); |
339 | enum mdp4_dma dma = mdp4_crtc->dma; | 263 | enum mdp4_dma dma = mdp4_crtc->dma; |
340 | int ret, ovlp = mdp4_crtc->ovlp; | 264 | int ovlp = mdp4_crtc->ovlp; |
265 | struct drm_display_mode *mode; | ||
266 | |||
267 | if (WARN_ON(!crtc->state)) | ||
268 | return; | ||
341 | 269 | ||
342 | mode = adjusted_mode; | 270 | mode = &crtc->state->adjusted_mode; |
343 | 271 | ||
344 | DBG("%s: set mode: %d:\"%s\" %d %d %d %d %d %d %d %d %d %d 0x%x 0x%x", | 272 | DBG("%s: set mode: %d:\"%s\" %d %d %d %d %d %d %d %d %d %d 0x%x 0x%x", |
345 | mdp4_crtc->name, mode->base.id, mode->name, | 273 | mdp4_crtc->name, mode->base.id, mode->name, |
@@ -350,28 +278,13 @@ static int mdp4_crtc_mode_set(struct drm_crtc *crtc, | |||
350 | mode->vsync_end, mode->vtotal, | 278 | mode->vsync_end, mode->vtotal, |
351 | mode->type, mode->flags); | 279 | mode->type, mode->flags); |
352 | 280 | ||
353 | /* grab extra ref for update_scanout() */ | ||
354 | drm_framebuffer_reference(crtc->primary->fb); | ||
355 | |||
356 | ret = mdp4_plane_mode_set(mdp4_crtc->plane, crtc, crtc->primary->fb, | ||
357 | 0, 0, mode->hdisplay, mode->vdisplay, | ||
358 | x << 16, y << 16, | ||
359 | mode->hdisplay << 16, mode->vdisplay << 16); | ||
360 | if (ret) { | ||
361 | drm_framebuffer_unreference(crtc->primary->fb); | ||
362 | dev_err(crtc->dev->dev, "%s: failed to set mode on plane: %d\n", | ||
363 | mdp4_crtc->name, ret); | ||
364 | return ret; | ||
365 | } | ||
366 | |||
367 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_SIZE(dma), | 281 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_SIZE(dma), |
368 | MDP4_DMA_SRC_SIZE_WIDTH(mode->hdisplay) | | 282 | MDP4_DMA_SRC_SIZE_WIDTH(mode->hdisplay) | |
369 | MDP4_DMA_SRC_SIZE_HEIGHT(mode->vdisplay)); | 283 | MDP4_DMA_SRC_SIZE_HEIGHT(mode->vdisplay)); |
370 | 284 | ||
371 | /* take data from pipe: */ | 285 | /* take data from pipe: */ |
372 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_BASE(dma), 0); | 286 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_BASE(dma), 0); |
373 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_STRIDE(dma), | 287 | mdp4_write(mdp4_kms, REG_MDP4_DMA_SRC_STRIDE(dma), 0); |
374 | crtc->primary->fb->pitches[0]); | ||
375 | mdp4_write(mdp4_kms, REG_MDP4_DMA_DST_SIZE(dma), | 288 | mdp4_write(mdp4_kms, REG_MDP4_DMA_DST_SIZE(dma), |
376 | MDP4_DMA_DST_SIZE_WIDTH(0) | | 289 | MDP4_DMA_DST_SIZE_WIDTH(0) | |
377 | MDP4_DMA_DST_SIZE_HEIGHT(0)); | 290 | MDP4_DMA_DST_SIZE_HEIGHT(0)); |
@@ -380,8 +293,7 @@ static int mdp4_crtc_mode_set(struct drm_crtc *crtc, | |||
380 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_SIZE(ovlp), | 293 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_SIZE(ovlp), |
381 | MDP4_OVLP_SIZE_WIDTH(mode->hdisplay) | | 294 | MDP4_OVLP_SIZE_WIDTH(mode->hdisplay) | |
382 | MDP4_OVLP_SIZE_HEIGHT(mode->vdisplay)); | 295 | MDP4_OVLP_SIZE_HEIGHT(mode->vdisplay)); |
383 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_STRIDE(ovlp), | 296 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_STRIDE(ovlp), 0); |
384 | crtc->primary->fb->pitches[0]); | ||
385 | 297 | ||
386 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_CFG(ovlp), 1); | 298 | mdp4_write(mdp4_kms, REG_MDP4_OVLP_CFG(ovlp), 1); |
387 | 299 | ||
@@ -390,11 +302,6 @@ static int mdp4_crtc_mode_set(struct drm_crtc *crtc, | |||
390 | mdp4_write(mdp4_kms, REG_MDP4_DMA_E_QUANT(1), 0x00ff0000); | 302 | mdp4_write(mdp4_kms, REG_MDP4_DMA_E_QUANT(1), 0x00ff0000); |
391 | mdp4_write(mdp4_kms, REG_MDP4_DMA_E_QUANT(2), 0x00ff0000); | 303 | mdp4_write(mdp4_kms, REG_MDP4_DMA_E_QUANT(2), 0x00ff0000); |
392 | } | 304 | } |
393 | |||
394 | update_fb(crtc, crtc->primary->fb); | ||
395 | update_scanout(crtc, crtc->primary->fb); | ||
396 | |||
397 | return 0; | ||
398 | } | 305 | } |
399 | 306 | ||
400 | static void mdp4_crtc_prepare(struct drm_crtc *crtc) | 307 | static void mdp4_crtc_prepare(struct drm_crtc *crtc) |
@@ -416,60 +323,51 @@ static void mdp4_crtc_commit(struct drm_crtc *crtc) | |||
416 | drm_crtc_vblank_put(crtc); | 323 | drm_crtc_vblank_put(crtc); |
417 | } | 324 | } |
418 | 325 | ||
419 | static int mdp4_crtc_mode_set_base(struct drm_crtc *crtc, int x, int y, | 326 | static void mdp4_crtc_load_lut(struct drm_crtc *crtc) |
420 | struct drm_framebuffer *old_fb) | 327 | { |
328 | } | ||
329 | |||
330 | static int mdp4_crtc_atomic_check(struct drm_crtc *crtc, | ||
331 | struct drm_crtc_state *state) | ||
421 | { | 332 | { |
422 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 333 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
423 | struct drm_plane *plane = mdp4_crtc->plane; | 334 | struct drm_device *dev = crtc->dev; |
424 | struct drm_display_mode *mode = &crtc->mode; | ||
425 | int ret; | ||
426 | 335 | ||
427 | /* grab extra ref for update_scanout() */ | 336 | DBG("%s: check", mdp4_crtc->name); |
428 | drm_framebuffer_reference(crtc->primary->fb); | ||
429 | 337 | ||
430 | ret = mdp4_plane_mode_set(plane, crtc, crtc->primary->fb, | 338 | if (mdp4_crtc->event) { |
431 | 0, 0, mode->hdisplay, mode->vdisplay, | 339 | dev_err(dev->dev, "already pending flip!\n"); |
432 | x << 16, y << 16, | 340 | return -EBUSY; |
433 | mode->hdisplay << 16, mode->vdisplay << 16); | ||
434 | if (ret) { | ||
435 | drm_framebuffer_unreference(crtc->primary->fb); | ||
436 | return ret; | ||
437 | } | 341 | } |
438 | 342 | ||
439 | update_fb(crtc, crtc->primary->fb); | 343 | // TODO anything else to check? |
440 | update_scanout(crtc, crtc->primary->fb); | ||
441 | 344 | ||
442 | return 0; | 345 | return 0; |
443 | } | 346 | } |
444 | 347 | ||
445 | static void mdp4_crtc_load_lut(struct drm_crtc *crtc) | 348 | static void mdp4_crtc_atomic_begin(struct drm_crtc *crtc) |
446 | { | 349 | { |
350 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | ||
351 | DBG("%s: begin", mdp4_crtc->name); | ||
447 | } | 352 | } |
448 | 353 | ||
449 | static int mdp4_crtc_page_flip(struct drm_crtc *crtc, | 354 | static void mdp4_crtc_atomic_flush(struct drm_crtc *crtc) |
450 | struct drm_framebuffer *new_fb, | ||
451 | struct drm_pending_vblank_event *event, | ||
452 | uint32_t page_flip_flags) | ||
453 | { | 355 | { |
454 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | 356 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
455 | struct drm_device *dev = crtc->dev; | 357 | struct drm_device *dev = crtc->dev; |
456 | struct drm_gem_object *obj; | ||
457 | unsigned long flags; | 358 | unsigned long flags; |
458 | 359 | ||
459 | if (mdp4_crtc->event) { | 360 | DBG("%s: flush", mdp4_crtc->name); |
460 | dev_err(dev->dev, "already pending flip!\n"); | ||
461 | return -EBUSY; | ||
462 | } | ||
463 | 361 | ||
464 | obj = msm_framebuffer_bo(new_fb, 0); | 362 | WARN_ON(mdp4_crtc->event); |
465 | 363 | ||
466 | spin_lock_irqsave(&dev->event_lock, flags); | 364 | spin_lock_irqsave(&dev->event_lock, flags); |
467 | mdp4_crtc->event = event; | 365 | mdp4_crtc->event = crtc->state->event; |
468 | spin_unlock_irqrestore(&dev->event_lock, flags); | 366 | spin_unlock_irqrestore(&dev->event_lock, flags); |
469 | 367 | ||
470 | update_fb(crtc, new_fb); | 368 | blend_setup(crtc); |
471 | 369 | crtc_flush(crtc); | |
472 | return msm_gem_queue_inactive_cb(obj, &mdp4_crtc->pageflip_cb); | 370 | request_pending(crtc, PENDING_FLIP); |
473 | } | 371 | } |
474 | 372 | ||
475 | static int mdp4_crtc_set_property(struct drm_crtc *crtc, | 373 | static int mdp4_crtc_set_property(struct drm_crtc *crtc, |
@@ -607,22 +505,29 @@ static int mdp4_crtc_cursor_move(struct drm_crtc *crtc, int x, int y) | |||
607 | } | 505 | } |
608 | 506 | ||
609 | static const struct drm_crtc_funcs mdp4_crtc_funcs = { | 507 | static const struct drm_crtc_funcs mdp4_crtc_funcs = { |
610 | .set_config = drm_crtc_helper_set_config, | 508 | .set_config = drm_atomic_helper_set_config, |
611 | .destroy = mdp4_crtc_destroy, | 509 | .destroy = mdp4_crtc_destroy, |
612 | .page_flip = mdp4_crtc_page_flip, | 510 | .page_flip = drm_atomic_helper_page_flip, |
613 | .set_property = mdp4_crtc_set_property, | 511 | .set_property = mdp4_crtc_set_property, |
614 | .cursor_set = mdp4_crtc_cursor_set, | 512 | .cursor_set = mdp4_crtc_cursor_set, |
615 | .cursor_move = mdp4_crtc_cursor_move, | 513 | .cursor_move = mdp4_crtc_cursor_move, |
514 | .reset = drm_atomic_helper_crtc_reset, | ||
515 | .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state, | ||
516 | .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state, | ||
616 | }; | 517 | }; |
617 | 518 | ||
618 | static const struct drm_crtc_helper_funcs mdp4_crtc_helper_funcs = { | 519 | static const struct drm_crtc_helper_funcs mdp4_crtc_helper_funcs = { |
619 | .dpms = mdp4_crtc_dpms, | 520 | .dpms = mdp4_crtc_dpms, |
620 | .mode_fixup = mdp4_crtc_mode_fixup, | 521 | .mode_fixup = mdp4_crtc_mode_fixup, |
621 | .mode_set = mdp4_crtc_mode_set, | 522 | .mode_set_nofb = mdp4_crtc_mode_set_nofb, |
523 | .mode_set = drm_helper_crtc_mode_set, | ||
524 | .mode_set_base = drm_helper_crtc_mode_set_base, | ||
622 | .prepare = mdp4_crtc_prepare, | 525 | .prepare = mdp4_crtc_prepare, |
623 | .commit = mdp4_crtc_commit, | 526 | .commit = mdp4_crtc_commit, |
624 | .mode_set_base = mdp4_crtc_mode_set_base, | ||
625 | .load_lut = mdp4_crtc_load_lut, | 527 | .load_lut = mdp4_crtc_load_lut, |
528 | .atomic_check = mdp4_crtc_atomic_check, | ||
529 | .atomic_begin = mdp4_crtc_atomic_begin, | ||
530 | .atomic_flush = mdp4_crtc_atomic_flush, | ||
626 | }; | 531 | }; |
627 | 532 | ||
628 | static void mdp4_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) | 533 | static void mdp4_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) |
@@ -638,7 +543,6 @@ static void mdp4_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) | |||
638 | 543 | ||
639 | if (pending & PENDING_FLIP) { | 544 | if (pending & PENDING_FLIP) { |
640 | complete_flip(crtc, NULL); | 545 | complete_flip(crtc, NULL); |
641 | drm_flip_work_commit(&mdp4_crtc->unref_fb_work, priv->wq); | ||
642 | } | 546 | } |
643 | 547 | ||
644 | if (pending & PENDING_CURSOR) { | 548 | if (pending & PENDING_CURSOR) { |
@@ -663,7 +567,8 @@ uint32_t mdp4_crtc_vblank(struct drm_crtc *crtc) | |||
663 | 567 | ||
664 | void mdp4_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file) | 568 | void mdp4_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file) |
665 | { | 569 | { |
666 | DBG("cancel: %p", file); | 570 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); |
571 | DBG("%s: cancel: %p", mdp4_crtc->name, file); | ||
667 | complete_flip(crtc, file); | 572 | complete_flip(crtc, file); |
668 | } | 573 | } |
669 | 574 | ||
@@ -717,35 +622,6 @@ void mdp4_crtc_set_intf(struct drm_crtc *crtc, enum mdp4_intf intf, int mixer) | |||
717 | mdp4_write(mdp4_kms, REG_MDP4_DISP_INTF_SEL, intf_sel); | 622 | mdp4_write(mdp4_kms, REG_MDP4_DISP_INTF_SEL, intf_sel); |
718 | } | 623 | } |
719 | 624 | ||
720 | static void set_attach(struct drm_crtc *crtc, enum mdp4_pipe pipe_id, | ||
721 | struct drm_plane *plane) | ||
722 | { | ||
723 | struct mdp4_crtc *mdp4_crtc = to_mdp4_crtc(crtc); | ||
724 | |||
725 | BUG_ON(pipe_id >= ARRAY_SIZE(mdp4_crtc->planes)); | ||
726 | |||
727 | if (mdp4_crtc->planes[pipe_id] == plane) | ||
728 | return; | ||
729 | |||
730 | mdp4_crtc->planes[pipe_id] = plane; | ||
731 | blend_setup(crtc); | ||
732 | if (mdp4_crtc->enabled && (plane != mdp4_crtc->plane)) | ||
733 | crtc_flush(crtc); | ||
734 | } | ||
735 | |||
736 | void mdp4_crtc_attach(struct drm_crtc *crtc, struct drm_plane *plane) | ||
737 | { | ||
738 | set_attach(crtc, mdp4_plane_pipe(plane), plane); | ||
739 | } | ||
740 | |||
741 | void mdp4_crtc_detach(struct drm_crtc *crtc, struct drm_plane *plane) | ||
742 | { | ||
743 | /* don't actually detatch our primary plane: */ | ||
744 | if (to_mdp4_crtc(crtc)->plane == plane) | ||
745 | return; | ||
746 | set_attach(crtc, mdp4_plane_pipe(plane), NULL); | ||
747 | } | ||
748 | |||
749 | static const char *dma_names[] = { | 625 | static const char *dma_names[] = { |
750 | "DMA_P", "DMA_S", "DMA_E", | 626 | "DMA_P", "DMA_S", "DMA_E", |
751 | }; | 627 | }; |
@@ -757,17 +633,13 @@ struct drm_crtc *mdp4_crtc_init(struct drm_device *dev, | |||
757 | { | 633 | { |
758 | struct drm_crtc *crtc = NULL; | 634 | struct drm_crtc *crtc = NULL; |
759 | struct mdp4_crtc *mdp4_crtc; | 635 | struct mdp4_crtc *mdp4_crtc; |
760 | int ret; | ||
761 | 636 | ||
762 | mdp4_crtc = kzalloc(sizeof(*mdp4_crtc), GFP_KERNEL); | 637 | mdp4_crtc = kzalloc(sizeof(*mdp4_crtc), GFP_KERNEL); |
763 | if (!mdp4_crtc) { | 638 | if (!mdp4_crtc) |
764 | ret = -ENOMEM; | 639 | return ERR_PTR(-ENOMEM); |
765 | goto fail; | ||
766 | } | ||
767 | 640 | ||
768 | crtc = &mdp4_crtc->base; | 641 | crtc = &mdp4_crtc->base; |
769 | 642 | ||
770 | mdp4_crtc->plane = plane; | ||
771 | mdp4_crtc->id = id; | 643 | mdp4_crtc->id = id; |
772 | 644 | ||
773 | mdp4_crtc->ovlp = ovlp_id; | 645 | mdp4_crtc->ovlp = ovlp_id; |
@@ -784,26 +656,14 @@ struct drm_crtc *mdp4_crtc_init(struct drm_device *dev, | |||
784 | 656 | ||
785 | spin_lock_init(&mdp4_crtc->cursor.lock); | 657 | spin_lock_init(&mdp4_crtc->cursor.lock); |
786 | 658 | ||
787 | ret = drm_flip_work_init(&mdp4_crtc->unref_fb_work, 16, | 659 | drm_flip_work_init(&mdp4_crtc->unref_cursor_work, |
788 | "unref fb", unref_fb_worker); | ||
789 | if (ret) | ||
790 | goto fail; | ||
791 | |||
792 | ret = drm_flip_work_init(&mdp4_crtc->unref_cursor_work, 64, | ||
793 | "unref cursor", unref_cursor_worker); | 660 | "unref cursor", unref_cursor_worker); |
794 | 661 | ||
795 | INIT_FENCE_CB(&mdp4_crtc->pageflip_cb, pageflip_cb); | ||
796 | |||
797 | drm_crtc_init_with_planes(dev, crtc, plane, NULL, &mdp4_crtc_funcs); | 662 | drm_crtc_init_with_planes(dev, crtc, plane, NULL, &mdp4_crtc_funcs); |
798 | drm_crtc_helper_add(crtc, &mdp4_crtc_helper_funcs); | 663 | drm_crtc_helper_add(crtc, &mdp4_crtc_helper_funcs); |
664 | plane->crtc = crtc; | ||
799 | 665 | ||
800 | mdp4_plane_install_properties(mdp4_crtc->plane, &crtc->base); | 666 | mdp4_plane_install_properties(plane, &crtc->base); |
801 | 667 | ||
802 | return crtc; | 668 | return crtc; |
803 | |||
804 | fail: | ||
805 | if (crtc) | ||
806 | mdp4_crtc_destroy(crtc); | ||
807 | |||
808 | return ERR_PTR(ret); | ||
809 | } | 669 | } |
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.c b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.c index 79d804e61cc4..a62109e4ae0d 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.c +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.c | |||
@@ -228,7 +228,6 @@ static int modeset_init(struct mdp4_kms *mdp4_kms) | |||
228 | struct drm_encoder *encoder; | 228 | struct drm_encoder *encoder; |
229 | struct drm_connector *connector; | 229 | struct drm_connector *connector; |
230 | struct drm_panel *panel; | 230 | struct drm_panel *panel; |
231 | struct hdmi *hdmi; | ||
232 | int ret; | 231 | int ret; |
233 | 232 | ||
234 | /* construct non-private planes: */ | 233 | /* construct non-private planes: */ |
@@ -326,11 +325,13 @@ static int modeset_init(struct mdp4_kms *mdp4_kms) | |||
326 | priv->crtcs[priv->num_crtcs++] = crtc; | 325 | priv->crtcs[priv->num_crtcs++] = crtc; |
327 | priv->encoders[priv->num_encoders++] = encoder; | 326 | priv->encoders[priv->num_encoders++] = encoder; |
328 | 327 | ||
329 | hdmi = hdmi_init(dev, encoder); | 328 | if (priv->hdmi) { |
330 | if (IS_ERR(hdmi)) { | 329 | /* Construct bridge/connector for HDMI: */ |
331 | ret = PTR_ERR(hdmi); | 330 | ret = hdmi_modeset_init(priv->hdmi, dev, encoder); |
332 | dev_err(dev->dev, "failed to initialize HDMI: %d\n", ret); | 331 | if (ret) { |
333 | goto fail; | 332 | dev_err(dev->dev, "failed to initialize HDMI: %d\n", ret); |
333 | goto fail; | ||
334 | } | ||
334 | } | 335 | } |
335 | 336 | ||
336 | return 0; | 337 | return 0; |
@@ -381,6 +382,10 @@ struct msm_kms *mdp4_kms_init(struct drm_device *dev) | |||
381 | if (IS_ERR(mdp4_kms->dsi_pll_vddio)) | 382 | if (IS_ERR(mdp4_kms->dsi_pll_vddio)) |
382 | mdp4_kms->dsi_pll_vddio = NULL; | 383 | mdp4_kms->dsi_pll_vddio = NULL; |
383 | 384 | ||
385 | /* NOTE: driver for this regulator still missing upstream.. use | ||
386 | * _get_exclusive() and ignore the error if it does not exist | ||
387 | * (and hope that the bootloader left it on for us) | ||
388 | */ | ||
384 | mdp4_kms->vdd = devm_regulator_get_exclusive(&pdev->dev, "vdd"); | 389 | mdp4_kms->vdd = devm_regulator_get_exclusive(&pdev->dev, "vdd"); |
385 | if (IS_ERR(mdp4_kms->vdd)) | 390 | if (IS_ERR(mdp4_kms->vdd)) |
386 | mdp4_kms->vdd = NULL; | 391 | mdp4_kms->vdd = NULL; |
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.h b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.h index 9ff6e7ccfe90..cbd77bc626d5 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.h +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_kms.h | |||
@@ -32,13 +32,6 @@ struct mdp4_kms { | |||
32 | 32 | ||
33 | int rev; | 33 | int rev; |
34 | 34 | ||
35 | /* Shadow value for MDP4_LAYERMIXER_IN_CFG.. since setup for all | ||
36 | * crtcs/encoders is in one shared register, we need to update it | ||
37 | * via read/modify/write. But to avoid getting confused by power- | ||
38 | * on-default values after resume, use this shadow value instead: | ||
39 | */ | ||
40 | uint32_t mixer_cfg; | ||
41 | |||
42 | /* mapper-id used to request GEM buffer mapped for scanout: */ | 35 | /* mapper-id used to request GEM buffer mapped for scanout: */ |
43 | int id; | 36 | int id; |
44 | 37 | ||
@@ -194,14 +187,6 @@ uint32_t mdp4_get_formats(enum mdp4_pipe pipe_id, uint32_t *pixel_formats, | |||
194 | 187 | ||
195 | void mdp4_plane_install_properties(struct drm_plane *plane, | 188 | void mdp4_plane_install_properties(struct drm_plane *plane, |
196 | struct drm_mode_object *obj); | 189 | struct drm_mode_object *obj); |
197 | void mdp4_plane_set_scanout(struct drm_plane *plane, | ||
198 | struct drm_framebuffer *fb); | ||
199 | int mdp4_plane_mode_set(struct drm_plane *plane, | ||
200 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | ||
201 | int crtc_x, int crtc_y, | ||
202 | unsigned int crtc_w, unsigned int crtc_h, | ||
203 | uint32_t src_x, uint32_t src_y, | ||
204 | uint32_t src_w, uint32_t src_h); | ||
205 | enum mdp4_pipe mdp4_plane_pipe(struct drm_plane *plane); | 190 | enum mdp4_pipe mdp4_plane_pipe(struct drm_plane *plane); |
206 | struct drm_plane *mdp4_plane_init(struct drm_device *dev, | 191 | struct drm_plane *mdp4_plane_init(struct drm_device *dev, |
207 | enum mdp4_pipe pipe_id, bool private_plane); | 192 | enum mdp4_pipe pipe_id, bool private_plane); |
@@ -210,8 +195,6 @@ uint32_t mdp4_crtc_vblank(struct drm_crtc *crtc); | |||
210 | void mdp4_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file); | 195 | void mdp4_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file); |
211 | void mdp4_crtc_set_config(struct drm_crtc *crtc, uint32_t config); | 196 | void mdp4_crtc_set_config(struct drm_crtc *crtc, uint32_t config); |
212 | void mdp4_crtc_set_intf(struct drm_crtc *crtc, enum mdp4_intf intf, int mixer); | 197 | void mdp4_crtc_set_intf(struct drm_crtc *crtc, enum mdp4_intf intf, int mixer); |
213 | void mdp4_crtc_attach(struct drm_crtc *crtc, struct drm_plane *plane); | ||
214 | void mdp4_crtc_detach(struct drm_crtc *crtc, struct drm_plane *plane); | ||
215 | struct drm_crtc *mdp4_crtc_init(struct drm_device *dev, | 198 | struct drm_crtc *mdp4_crtc_init(struct drm_device *dev, |
216 | struct drm_plane *plane, int id, int ovlp_id, | 199 | struct drm_plane *plane, int id, int ovlp_id, |
217 | enum mdp4_dma dma_id); | 200 | enum mdp4_dma dma_id); |
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_lvds_connector.c b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_lvds_connector.c index 310034688c15..4ddc28e1275b 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_lvds_connector.c +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_lvds_connector.c | |||
@@ -98,6 +98,9 @@ static const struct drm_connector_funcs mdp4_lvds_connector_funcs = { | |||
98 | .detect = mdp4_lvds_connector_detect, | 98 | .detect = mdp4_lvds_connector_detect, |
99 | .fill_modes = drm_helper_probe_single_connector_modes, | 99 | .fill_modes = drm_helper_probe_single_connector_modes, |
100 | .destroy = mdp4_lvds_connector_destroy, | 100 | .destroy = mdp4_lvds_connector_destroy, |
101 | .reset = drm_atomic_helper_connector_reset, | ||
102 | .atomic_duplicate_state = drm_atomic_helper_connector_duplicate_state, | ||
103 | .atomic_destroy_state = drm_atomic_helper_connector_destroy_state, | ||
101 | }; | 104 | }; |
102 | 105 | ||
103 | static const struct drm_connector_helper_funcs mdp4_lvds_connector_helper_funcs = { | 106 | static const struct drm_connector_helper_funcs mdp4_lvds_connector_helper_funcs = { |
diff --git a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_plane.c b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_plane.c index 66f33dba1ebb..1e5ebe83647d 100644 --- a/drivers/gpu/drm/msm/mdp/mdp4/mdp4_plane.c +++ b/drivers/gpu/drm/msm/mdp/mdp4/mdp4_plane.c | |||
@@ -31,47 +31,26 @@ struct mdp4_plane { | |||
31 | }; | 31 | }; |
32 | #define to_mdp4_plane(x) container_of(x, struct mdp4_plane, base) | 32 | #define to_mdp4_plane(x) container_of(x, struct mdp4_plane, base) |
33 | 33 | ||
34 | static struct mdp4_kms *get_kms(struct drm_plane *plane) | 34 | static void mdp4_plane_set_scanout(struct drm_plane *plane, |
35 | { | 35 | struct drm_framebuffer *fb); |
36 | struct msm_drm_private *priv = plane->dev->dev_private; | 36 | static int mdp4_plane_mode_set(struct drm_plane *plane, |
37 | return to_mdp4_kms(to_mdp_kms(priv->kms)); | ||
38 | } | ||
39 | |||
40 | static int mdp4_plane_update(struct drm_plane *plane, | ||
41 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | 37 | struct drm_crtc *crtc, struct drm_framebuffer *fb, |
42 | int crtc_x, int crtc_y, | 38 | int crtc_x, int crtc_y, |
43 | unsigned int crtc_w, unsigned int crtc_h, | 39 | unsigned int crtc_w, unsigned int crtc_h, |
44 | uint32_t src_x, uint32_t src_y, | 40 | uint32_t src_x, uint32_t src_y, |
45 | uint32_t src_w, uint32_t src_h) | 41 | uint32_t src_w, uint32_t src_h); |
46 | { | ||
47 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | ||
48 | |||
49 | mdp4_plane->enabled = true; | ||
50 | |||
51 | if (plane->fb) | ||
52 | drm_framebuffer_unreference(plane->fb); | ||
53 | |||
54 | drm_framebuffer_reference(fb); | ||
55 | |||
56 | return mdp4_plane_mode_set(plane, crtc, fb, | ||
57 | crtc_x, crtc_y, crtc_w, crtc_h, | ||
58 | src_x, src_y, src_w, src_h); | ||
59 | } | ||
60 | 42 | ||
61 | static int mdp4_plane_disable(struct drm_plane *plane) | 43 | static struct mdp4_kms *get_kms(struct drm_plane *plane) |
62 | { | 44 | { |
63 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | 45 | struct msm_drm_private *priv = plane->dev->dev_private; |
64 | DBG("%s: disable", mdp4_plane->name); | 46 | return to_mdp4_kms(to_mdp_kms(priv->kms)); |
65 | if (plane->crtc) | ||
66 | mdp4_crtc_detach(plane->crtc, plane); | ||
67 | return 0; | ||
68 | } | 47 | } |
69 | 48 | ||
70 | static void mdp4_plane_destroy(struct drm_plane *plane) | 49 | static void mdp4_plane_destroy(struct drm_plane *plane) |
71 | { | 50 | { |
72 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | 51 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); |
73 | 52 | ||
74 | mdp4_plane_disable(plane); | 53 | drm_plane_helper_disable(plane); |
75 | drm_plane_cleanup(plane); | 54 | drm_plane_cleanup(plane); |
76 | 55 | ||
77 | kfree(mdp4_plane); | 56 | kfree(mdp4_plane); |
@@ -92,19 +71,75 @@ int mdp4_plane_set_property(struct drm_plane *plane, | |||
92 | } | 71 | } |
93 | 72 | ||
94 | static const struct drm_plane_funcs mdp4_plane_funcs = { | 73 | static const struct drm_plane_funcs mdp4_plane_funcs = { |
95 | .update_plane = mdp4_plane_update, | 74 | .update_plane = drm_atomic_helper_update_plane, |
96 | .disable_plane = mdp4_plane_disable, | 75 | .disable_plane = drm_atomic_helper_disable_plane, |
97 | .destroy = mdp4_plane_destroy, | 76 | .destroy = mdp4_plane_destroy, |
98 | .set_property = mdp4_plane_set_property, | 77 | .set_property = mdp4_plane_set_property, |
78 | .reset = drm_atomic_helper_plane_reset, | ||
79 | .atomic_duplicate_state = drm_atomic_helper_plane_duplicate_state, | ||
80 | .atomic_destroy_state = drm_atomic_helper_plane_destroy_state, | ||
99 | }; | 81 | }; |
100 | 82 | ||
101 | void mdp4_plane_set_scanout(struct drm_plane *plane, | 83 | static int mdp4_plane_prepare_fb(struct drm_plane *plane, |
84 | struct drm_framebuffer *fb) | ||
85 | { | ||
86 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | ||
87 | struct mdp4_kms *mdp4_kms = get_kms(plane); | ||
88 | |||
89 | DBG("%s: prepare: FB[%u]", mdp4_plane->name, fb->base.id); | ||
90 | return msm_framebuffer_prepare(fb, mdp4_kms->id); | ||
91 | } | ||
92 | |||
93 | static void mdp4_plane_cleanup_fb(struct drm_plane *plane, | ||
94 | struct drm_framebuffer *fb) | ||
95 | { | ||
96 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | ||
97 | struct mdp4_kms *mdp4_kms = get_kms(plane); | ||
98 | |||
99 | DBG("%s: cleanup: FB[%u]", mdp4_plane->name, fb->base.id); | ||
100 | msm_framebuffer_cleanup(fb, mdp4_kms->id); | ||
101 | } | ||
102 | |||
103 | |||
104 | static int mdp4_plane_atomic_check(struct drm_plane *plane, | ||
105 | struct drm_plane_state *state) | ||
106 | { | ||
107 | return 0; | ||
108 | } | ||
109 | |||
110 | static void mdp4_plane_atomic_update(struct drm_plane *plane, | ||
111 | struct drm_plane_state *old_state) | ||
112 | { | ||
113 | struct drm_plane_state *state = plane->state; | ||
114 | int ret; | ||
115 | |||
116 | ret = mdp4_plane_mode_set(plane, | ||
117 | state->crtc, state->fb, | ||
118 | state->crtc_x, state->crtc_y, | ||
119 | state->crtc_w, state->crtc_h, | ||
120 | state->src_x, state->src_y, | ||
121 | state->src_w, state->src_h); | ||
122 | /* atomic_check should have ensured that this doesn't fail */ | ||
123 | WARN_ON(ret < 0); | ||
124 | } | ||
125 | |||
126 | static const struct drm_plane_helper_funcs mdp4_plane_helper_funcs = { | ||
127 | .prepare_fb = mdp4_plane_prepare_fb, | ||
128 | .cleanup_fb = mdp4_plane_cleanup_fb, | ||
129 | .atomic_check = mdp4_plane_atomic_check, | ||
130 | .atomic_update = mdp4_plane_atomic_update, | ||
131 | }; | ||
132 | |||
133 | static void mdp4_plane_set_scanout(struct drm_plane *plane, | ||
102 | struct drm_framebuffer *fb) | 134 | struct drm_framebuffer *fb) |
103 | { | 135 | { |
104 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); | 136 | struct mdp4_plane *mdp4_plane = to_mdp4_plane(plane); |
105 | struct mdp4_kms *mdp4_kms = get_kms(plane); | 137 | struct mdp4_kms *mdp4_kms = get_kms(plane); |
106 | enum mdp4_pipe pipe = mdp4_plane->pipe; | 138 | enum mdp4_pipe pipe = mdp4_plane->pipe; |
107 | uint32_t iova; | 139 | uint32_t iova = msm_framebuffer_iova(fb, mdp4_kms->id, 0); |
140 | |||
141 | DBG("%s: set_scanout: %08x (%u)", mdp4_plane->name, | ||
142 | iova, fb->pitches[0]); | ||
108 | 143 | ||
109 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_SRC_STRIDE_A(pipe), | 144 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_SRC_STRIDE_A(pipe), |
110 | MDP4_PIPE_SRC_STRIDE_A_P0(fb->pitches[0]) | | 145 | MDP4_PIPE_SRC_STRIDE_A_P0(fb->pitches[0]) | |
@@ -114,7 +149,6 @@ void mdp4_plane_set_scanout(struct drm_plane *plane, | |||
114 | MDP4_PIPE_SRC_STRIDE_B_P2(fb->pitches[2]) | | 149 | MDP4_PIPE_SRC_STRIDE_B_P2(fb->pitches[2]) | |
115 | MDP4_PIPE_SRC_STRIDE_B_P3(fb->pitches[3])); | 150 | MDP4_PIPE_SRC_STRIDE_B_P3(fb->pitches[3])); |
116 | 151 | ||
117 | msm_gem_get_iova(msm_framebuffer_bo(fb, 0), mdp4_kms->id, &iova); | ||
118 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_SRCP0_BASE(pipe), iova); | 152 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_SRCP0_BASE(pipe), iova); |
119 | 153 | ||
120 | plane->fb = fb; | 154 | plane->fb = fb; |
@@ -122,7 +156,7 @@ void mdp4_plane_set_scanout(struct drm_plane *plane, | |||
122 | 156 | ||
123 | #define MDP4_VG_PHASE_STEP_DEFAULT 0x20000000 | 157 | #define MDP4_VG_PHASE_STEP_DEFAULT 0x20000000 |
124 | 158 | ||
125 | int mdp4_plane_mode_set(struct drm_plane *plane, | 159 | static int mdp4_plane_mode_set(struct drm_plane *plane, |
126 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | 160 | struct drm_crtc *crtc, struct drm_framebuffer *fb, |
127 | int crtc_x, int crtc_y, | 161 | int crtc_x, int crtc_y, |
128 | unsigned int crtc_w, unsigned int crtc_h, | 162 | unsigned int crtc_w, unsigned int crtc_h, |
@@ -137,6 +171,11 @@ int mdp4_plane_mode_set(struct drm_plane *plane, | |||
137 | uint32_t phasex_step = MDP4_VG_PHASE_STEP_DEFAULT; | 171 | uint32_t phasex_step = MDP4_VG_PHASE_STEP_DEFAULT; |
138 | uint32_t phasey_step = MDP4_VG_PHASE_STEP_DEFAULT; | 172 | uint32_t phasey_step = MDP4_VG_PHASE_STEP_DEFAULT; |
139 | 173 | ||
174 | if (!(crtc && fb)) { | ||
175 | DBG("%s: disabled!", mdp4_plane->name); | ||
176 | return 0; | ||
177 | } | ||
178 | |||
140 | /* src values are in Q16 fixed point, convert to integer: */ | 179 | /* src values are in Q16 fixed point, convert to integer: */ |
141 | src_x = src_x >> 16; | 180 | src_x = src_x >> 16; |
142 | src_y = src_y >> 16; | 181 | src_y = src_y >> 16; |
@@ -197,9 +236,6 @@ int mdp4_plane_mode_set(struct drm_plane *plane, | |||
197 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_PHASEX_STEP(pipe), phasex_step); | 236 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_PHASEX_STEP(pipe), phasex_step); |
198 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_PHASEY_STEP(pipe), phasey_step); | 237 | mdp4_write(mdp4_kms, REG_MDP4_PIPE_PHASEY_STEP(pipe), phasey_step); |
199 | 238 | ||
200 | /* TODO detach from old crtc (if we had more than one) */ | ||
201 | mdp4_crtc_attach(crtc, plane); | ||
202 | |||
203 | return 0; | 239 | return 0; |
204 | } | 240 | } |
205 | 241 | ||
@@ -239,9 +275,12 @@ struct drm_plane *mdp4_plane_init(struct drm_device *dev, | |||
239 | ARRAY_SIZE(mdp4_plane->formats)); | 275 | ARRAY_SIZE(mdp4_plane->formats)); |
240 | 276 | ||
241 | type = private_plane ? DRM_PLANE_TYPE_PRIMARY : DRM_PLANE_TYPE_OVERLAY; | 277 | type = private_plane ? DRM_PLANE_TYPE_PRIMARY : DRM_PLANE_TYPE_OVERLAY; |
242 | drm_universal_plane_init(dev, plane, 0xff, &mdp4_plane_funcs, | 278 | ret = drm_universal_plane_init(dev, plane, 0xff, &mdp4_plane_funcs, |
243 | mdp4_plane->formats, mdp4_plane->nformats, | 279 | mdp4_plane->formats, mdp4_plane->nformats, type); |
244 | type); | 280 | if (ret) |
281 | goto fail; | ||
282 | |||
283 | drm_plane_helper_add(plane, &mdp4_plane_helper_funcs); | ||
245 | 284 | ||
246 | mdp4_plane_install_properties(plane, &plane->base); | 285 | mdp4_plane_install_properties(plane, &plane->base); |
247 | 286 | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5.xml.h b/drivers/gpu/drm/msm/mdp/mdp5/mdp5.xml.h index 67f4f896ba8c..e87ef5512cb0 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5.xml.h +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5.xml.h | |||
@@ -10,14 +10,14 @@ git clone https://github.com/freedreno/envytools.git | |||
10 | The rules-ng-ng source files this header was generated from are: | 10 | The rules-ng-ng source files this header was generated from are: |
11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) | 11 | - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 647 bytes, from 2013-11-30 14:45:35) |
12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) | 12 | - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) |
13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 17996 bytes, from 2013-12-01 19:10:31) | 13 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20136 bytes, from 2014-10-31 16:51:39) |
14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1615 bytes, from 2013-11-30 15:00:52) | 14 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 1940 bytes, from 2014-10-31 16:51:39) |
15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 22517 bytes, from 2014-06-25 12:55:02) | 15 | - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 23963 bytes, from 2014-10-31 16:51:46) |
16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) | 16 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) |
17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) | 17 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) |
18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1544 bytes, from 2013-08-16 19:17:05) | 18 | - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) |
19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) | 19 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) |
20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-06-25 12:53:44) | 20 | - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 23613 bytes, from 2014-07-17 15:33:30) |
21 | 21 | ||
22 | Copyright (C) 2013-2014 by the following authors: | 22 | Copyright (C) 2013-2014 by the following authors: |
23 | - Rob Clark <robdclark@gmail.com> (robclark) | 23 | - Rob Clark <robdclark@gmail.com> (robclark) |
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.c new file mode 100644 index 000000000000..b0a44310cf2a --- /dev/null +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.c | |||
@@ -0,0 +1,207 @@ | |||
1 | /* | ||
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
3 | * | ||
4 | * This program is free software; you can redistribute it and/or modify | ||
5 | * it under the terms of the GNU General Public License version 2 and | ||
6 | * only version 2 as published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope that it will be useful, | ||
9 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
11 | * GNU General Public License for more details. | ||
12 | */ | ||
13 | |||
14 | #include "mdp5_kms.h" | ||
15 | #include "mdp5_cfg.h" | ||
16 | |||
17 | struct mdp5_cfg_handler { | ||
18 | int revision; | ||
19 | struct mdp5_cfg config; | ||
20 | }; | ||
21 | |||
22 | /* mdp5_cfg must be exposed (used in mdp5.xml.h) */ | ||
23 | const struct mdp5_cfg_hw *mdp5_cfg = NULL; | ||
24 | |||
25 | const struct mdp5_cfg_hw msm8x74_config = { | ||
26 | .name = "msm8x74", | ||
27 | .smp = { | ||
28 | .mmb_count = 22, | ||
29 | .mmb_size = 4096, | ||
30 | }, | ||
31 | .ctl = { | ||
32 | .count = 5, | ||
33 | .base = { 0x00600, 0x00700, 0x00800, 0x00900, 0x00a00 }, | ||
34 | }, | ||
35 | .pipe_vig = { | ||
36 | .count = 3, | ||
37 | .base = { 0x01200, 0x01600, 0x01a00 }, | ||
38 | }, | ||
39 | .pipe_rgb = { | ||
40 | .count = 3, | ||
41 | .base = { 0x01e00, 0x02200, 0x02600 }, | ||
42 | }, | ||
43 | .pipe_dma = { | ||
44 | .count = 2, | ||
45 | .base = { 0x02a00, 0x02e00 }, | ||
46 | }, | ||
47 | .lm = { | ||
48 | .count = 5, | ||
49 | .base = { 0x03200, 0x03600, 0x03a00, 0x03e00, 0x04200 }, | ||
50 | .nb_stages = 5, | ||
51 | }, | ||
52 | .dspp = { | ||
53 | .count = 3, | ||
54 | .base = { 0x04600, 0x04a00, 0x04e00 }, | ||
55 | }, | ||
56 | .ad = { | ||
57 | .count = 2, | ||
58 | .base = { 0x13100, 0x13300 }, /* NOTE: no ad in v1.0 */ | ||
59 | }, | ||
60 | .intf = { | ||
61 | .count = 4, | ||
62 | .base = { 0x12500, 0x12700, 0x12900, 0x12b00 }, | ||
63 | }, | ||
64 | .max_clk = 200000000, | ||
65 | }; | ||
66 | |||
67 | const struct mdp5_cfg_hw apq8084_config = { | ||
68 | .name = "apq8084", | ||
69 | .smp = { | ||
70 | .mmb_count = 44, | ||
71 | .mmb_size = 8192, | ||
72 | .reserved_state[0] = GENMASK(7, 0), /* first 8 MMBs */ | ||
73 | .reserved[CID_RGB0] = 2, | ||
74 | .reserved[CID_RGB1] = 2, | ||
75 | .reserved[CID_RGB2] = 2, | ||
76 | .reserved[CID_RGB3] = 2, | ||
77 | }, | ||
78 | .ctl = { | ||
79 | .count = 5, | ||
80 | .base = { 0x00600, 0x00700, 0x00800, 0x00900, 0x00a00 }, | ||
81 | }, | ||
82 | .pipe_vig = { | ||
83 | .count = 4, | ||
84 | .base = { 0x01200, 0x01600, 0x01a00, 0x01e00 }, | ||
85 | }, | ||
86 | .pipe_rgb = { | ||
87 | .count = 4, | ||
88 | .base = { 0x02200, 0x02600, 0x02a00, 0x02e00 }, | ||
89 | }, | ||
90 | .pipe_dma = { | ||
91 | .count = 2, | ||
92 | .base = { 0x03200, 0x03600 }, | ||
93 | }, | ||
94 | .lm = { | ||
95 | .count = 6, | ||
96 | .base = { 0x03a00, 0x03e00, 0x04200, 0x04600, 0x04a00, 0x04e00 }, | ||
97 | .nb_stages = 5, | ||
98 | }, | ||
99 | .dspp = { | ||
100 | .count = 4, | ||
101 | .base = { 0x05200, 0x05600, 0x05a00, 0x05e00 }, | ||
102 | |||
103 | }, | ||
104 | .ad = { | ||
105 | .count = 3, | ||
106 | .base = { 0x13500, 0x13700, 0x13900 }, | ||
107 | }, | ||
108 | .intf = { | ||
109 | .count = 5, | ||
110 | .base = { 0x12500, 0x12700, 0x12900, 0x12b00, 0x12d00 }, | ||
111 | }, | ||
112 | .max_clk = 320000000, | ||
113 | }; | ||
114 | |||
115 | static const struct mdp5_cfg_handler cfg_handlers[] = { | ||
116 | { .revision = 0, .config = { .hw = &msm8x74_config } }, | ||
117 | { .revision = 2, .config = { .hw = &msm8x74_config } }, | ||
118 | { .revision = 3, .config = { .hw = &apq8084_config } }, | ||
119 | }; | ||
120 | |||
121 | |||
122 | static struct mdp5_cfg_platform *mdp5_get_config(struct platform_device *dev); | ||
123 | |||
124 | const struct mdp5_cfg_hw *mdp5_cfg_get_hw_config(struct mdp5_cfg_handler *cfg_handler) | ||
125 | { | ||
126 | return cfg_handler->config.hw; | ||
127 | } | ||
128 | |||
129 | struct mdp5_cfg *mdp5_cfg_get_config(struct mdp5_cfg_handler *cfg_handler) | ||
130 | { | ||
131 | return &cfg_handler->config; | ||
132 | } | ||
133 | |||
134 | int mdp5_cfg_get_hw_rev(struct mdp5_cfg_handler *cfg_handler) | ||
135 | { | ||
136 | return cfg_handler->revision; | ||
137 | } | ||
138 | |||
139 | void mdp5_cfg_destroy(struct mdp5_cfg_handler *cfg_handler) | ||
140 | { | ||
141 | kfree(cfg_handler); | ||
142 | } | ||
143 | |||
144 | struct mdp5_cfg_handler *mdp5_cfg_init(struct mdp5_kms *mdp5_kms, | ||
145 | uint32_t major, uint32_t minor) | ||
146 | { | ||
147 | struct drm_device *dev = mdp5_kms->dev; | ||
148 | struct platform_device *pdev = dev->platformdev; | ||
149 | struct mdp5_cfg_handler *cfg_handler; | ||
150 | struct mdp5_cfg_platform *pconfig; | ||
151 | int i, ret = 0; | ||
152 | |||
153 | cfg_handler = kzalloc(sizeof(*cfg_handler), GFP_KERNEL); | ||
154 | if (unlikely(!cfg_handler)) { | ||
155 | ret = -ENOMEM; | ||
156 | goto fail; | ||
157 | } | ||
158 | |||
159 | if (major != 1) { | ||
160 | dev_err(dev->dev, "unexpected MDP major version: v%d.%d\n", | ||
161 | major, minor); | ||
162 | ret = -ENXIO; | ||
163 | goto fail; | ||
164 | } | ||
165 | |||
166 | /* only after mdp5_cfg global pointer's init can we access the hw */ | ||
167 | for (i = 0; i < ARRAY_SIZE(cfg_handlers); i++) { | ||
168 | if (cfg_handlers[i].revision != minor) | ||
169 | continue; | ||
170 | mdp5_cfg = cfg_handlers[i].config.hw; | ||
171 | |||
172 | break; | ||
173 | } | ||
174 | if (unlikely(!mdp5_cfg)) { | ||
175 | dev_err(dev->dev, "unexpected MDP minor revision: v%d.%d\n", | ||
176 | major, minor); | ||
177 | ret = -ENXIO; | ||
178 | goto fail; | ||
179 | } | ||
180 | |||
181 | cfg_handler->revision = minor; | ||
182 | cfg_handler->config.hw = mdp5_cfg; | ||
183 | |||
184 | pconfig = mdp5_get_config(pdev); | ||
185 | memcpy(&cfg_handler->config.platform, pconfig, sizeof(*pconfig)); | ||
186 | |||
187 | DBG("MDP5: %s hw config selected", mdp5_cfg->name); | ||
188 | |||
189 | return cfg_handler; | ||
190 | |||
191 | fail: | ||
192 | if (cfg_handler) | ||
193 | mdp5_cfg_destroy(cfg_handler); | ||
194 | |||
195 | return NULL; | ||
196 | } | ||
197 | |||
198 | static struct mdp5_cfg_platform *mdp5_get_config(struct platform_device *dev) | ||
199 | { | ||
200 | static struct mdp5_cfg_platform config = {}; | ||
201 | #ifdef CONFIG_OF | ||
202 | /* TODO */ | ||
203 | #endif | ||
204 | config.iommu = iommu_domain_alloc(&platform_bus_type); | ||
205 | |||
206 | return &config; | ||
207 | } | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.h b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.h new file mode 100644 index 000000000000..dba4d52cceeb --- /dev/null +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_cfg.h | |||
@@ -0,0 +1,91 @@ | |||
1 | /* | ||
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
3 | * | ||
4 | * This program is free software; you can redistribute it and/or modify | ||
5 | * it under the terms of the GNU General Public License version 2 and | ||
6 | * only version 2 as published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope that it will be useful, | ||
9 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
11 | * GNU General Public License for more details. | ||
12 | */ | ||
13 | |||
14 | #ifndef __MDP5_CFG_H__ | ||
15 | #define __MDP5_CFG_H__ | ||
16 | |||
17 | #include "msm_drv.h" | ||
18 | |||
19 | /* | ||
20 | * mdp5_cfg | ||
21 | * | ||
22 | * This module configures the dynamic offsets used by mdp5.xml.h | ||
23 | * (initialized in mdp5_cfg.c) | ||
24 | */ | ||
25 | extern const struct mdp5_cfg_hw *mdp5_cfg; | ||
26 | |||
27 | #define MAX_CTL 8 | ||
28 | #define MAX_BASES 8 | ||
29 | #define MAX_SMP_BLOCKS 44 | ||
30 | #define MAX_CLIENTS 32 | ||
31 | |||
32 | typedef DECLARE_BITMAP(mdp5_smp_state_t, MAX_SMP_BLOCKS); | ||
33 | |||
34 | #define MDP5_SUB_BLOCK_DEFINITION \ | ||
35 | int count; \ | ||
36 | uint32_t base[MAX_BASES] | ||
37 | |||
38 | struct mdp5_sub_block { | ||
39 | MDP5_SUB_BLOCK_DEFINITION; | ||
40 | }; | ||
41 | |||
42 | struct mdp5_lm_block { | ||
43 | MDP5_SUB_BLOCK_DEFINITION; | ||
44 | uint32_t nb_stages; /* number of stages per blender */ | ||
45 | }; | ||
46 | |||
47 | struct mdp5_smp_block { | ||
48 | int mmb_count; /* number of SMP MMBs */ | ||
49 | int mmb_size; /* MMB: size in bytes */ | ||
50 | mdp5_smp_state_t reserved_state;/* SMP MMBs statically allocated */ | ||
51 | int reserved[MAX_CLIENTS]; /* # of MMBs allocated per client */ | ||
52 | }; | ||
53 | |||
54 | struct mdp5_cfg_hw { | ||
55 | char *name; | ||
56 | |||
57 | struct mdp5_smp_block smp; | ||
58 | struct mdp5_sub_block ctl; | ||
59 | struct mdp5_sub_block pipe_vig; | ||
60 | struct mdp5_sub_block pipe_rgb; | ||
61 | struct mdp5_sub_block pipe_dma; | ||
62 | struct mdp5_lm_block lm; | ||
63 | struct mdp5_sub_block dspp; | ||
64 | struct mdp5_sub_block ad; | ||
65 | struct mdp5_sub_block intf; | ||
66 | |||
67 | uint32_t max_clk; | ||
68 | }; | ||
69 | |||
70 | /* platform config data (ie. from DT, or pdata) */ | ||
71 | struct mdp5_cfg_platform { | ||
72 | struct iommu_domain *iommu; | ||
73 | }; | ||
74 | |||
75 | struct mdp5_cfg { | ||
76 | const struct mdp5_cfg_hw *hw; | ||
77 | struct mdp5_cfg_platform platform; | ||
78 | }; | ||
79 | |||
80 | struct mdp5_kms; | ||
81 | struct mdp5_cfg_handler; | ||
82 | |||
83 | const struct mdp5_cfg_hw *mdp5_cfg_get_hw_config(struct mdp5_cfg_handler *cfg_hnd); | ||
84 | struct mdp5_cfg *mdp5_cfg_get_config(struct mdp5_cfg_handler *cfg_hnd); | ||
85 | int mdp5_cfg_get_hw_rev(struct mdp5_cfg_handler *cfg_hnd); | ||
86 | |||
87 | struct mdp5_cfg_handler *mdp5_cfg_init(struct mdp5_kms *mdp5_kms, | ||
88 | uint32_t major, uint32_t minor); | ||
89 | void mdp5_cfg_destroy(struct mdp5_cfg_handler *cfg_hnd); | ||
90 | |||
91 | #endif /* __MDP5_CFG_H__ */ | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_crtc.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_crtc.c index ebe2e60f3ab1..0e9a2e3a82d7 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_crtc.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_crtc.c | |||
@@ -1,4 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
2 | * Copyright (C) 2013 Red Hat | 3 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 4 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 5 | * |
@@ -17,43 +18,35 @@ | |||
17 | 18 | ||
18 | #include "mdp5_kms.h" | 19 | #include "mdp5_kms.h" |
19 | 20 | ||
21 | #include <linux/sort.h> | ||
20 | #include <drm/drm_mode.h> | 22 | #include <drm/drm_mode.h> |
21 | #include "drm_crtc.h" | 23 | #include "drm_crtc.h" |
22 | #include "drm_crtc_helper.h" | 24 | #include "drm_crtc_helper.h" |
23 | #include "drm_flip_work.h" | 25 | #include "drm_flip_work.h" |
24 | 26 | ||
27 | #define SSPP_MAX (SSPP_RGB3 + 1) /* TODO: Add SSPP_MAX in mdp5.xml.h */ | ||
28 | |||
25 | struct mdp5_crtc { | 29 | struct mdp5_crtc { |
26 | struct drm_crtc base; | 30 | struct drm_crtc base; |
27 | char name[8]; | 31 | char name[8]; |
28 | struct drm_plane *plane; | ||
29 | struct drm_plane *planes[8]; | ||
30 | int id; | 32 | int id; |
31 | bool enabled; | 33 | bool enabled; |
32 | 34 | ||
33 | /* which mixer/encoder we route output to: */ | 35 | /* layer mixer used for this CRTC (+ its lock): */ |
34 | int mixer; | 36 | #define GET_LM_ID(crtc_id) ((crtc_id == 3) ? 5 : crtc_id) |
37 | int lm; | ||
38 | spinlock_t lm_lock; /* protect REG_MDP5_LM_* registers */ | ||
39 | |||
40 | /* CTL used for this CRTC: */ | ||
41 | struct mdp5_ctl *ctl; | ||
35 | 42 | ||
36 | /* if there is a pending flip, these will be non-null: */ | 43 | /* if there is a pending flip, these will be non-null: */ |
37 | struct drm_pending_vblank_event *event; | 44 | struct drm_pending_vblank_event *event; |
38 | struct msm_fence_cb pageflip_cb; | ||
39 | 45 | ||
40 | #define PENDING_CURSOR 0x1 | 46 | #define PENDING_CURSOR 0x1 |
41 | #define PENDING_FLIP 0x2 | 47 | #define PENDING_FLIP 0x2 |
42 | atomic_t pending; | 48 | atomic_t pending; |
43 | 49 | ||
44 | /* the fb that we logically (from PoV of KMS API) hold a ref | ||
45 | * to. Which we may not yet be scanning out (we may still | ||
46 | * be scanning out previous in case of page_flip while waiting | ||
47 | * for gpu rendering to complete: | ||
48 | */ | ||
49 | struct drm_framebuffer *fb; | ||
50 | |||
51 | /* the fb that we currently hold a scanout ref to: */ | ||
52 | struct drm_framebuffer *scanout_fb; | ||
53 | |||
54 | /* for unref'ing framebuffers after scanout completes: */ | ||
55 | struct drm_flip_work unref_fb_work; | ||
56 | |||
57 | struct mdp_irq vblank; | 50 | struct mdp_irq vblank; |
58 | struct mdp_irq err; | 51 | struct mdp_irq err; |
59 | }; | 52 | }; |
@@ -73,67 +66,38 @@ static void request_pending(struct drm_crtc *crtc, uint32_t pending) | |||
73 | mdp_irq_register(&get_kms(crtc)->base, &mdp5_crtc->vblank); | 66 | mdp_irq_register(&get_kms(crtc)->base, &mdp5_crtc->vblank); |
74 | } | 67 | } |
75 | 68 | ||
76 | static void crtc_flush(struct drm_crtc *crtc) | 69 | #define mdp5_lm_get_flush(lm) mdp_ctl_flush_mask_lm(lm) |
77 | { | ||
78 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | ||
79 | struct mdp5_kms *mdp5_kms = get_kms(crtc); | ||
80 | int id = mdp5_crtc->id; | ||
81 | uint32_t i, flush = 0; | ||
82 | |||
83 | for (i = 0; i < ARRAY_SIZE(mdp5_crtc->planes); i++) { | ||
84 | struct drm_plane *plane = mdp5_crtc->planes[i]; | ||
85 | if (plane) { | ||
86 | enum mdp5_pipe pipe = mdp5_plane_pipe(plane); | ||
87 | flush |= pipe2flush(pipe); | ||
88 | } | ||
89 | } | ||
90 | flush |= mixer2flush(mdp5_crtc->id); | ||
91 | flush |= MDP5_CTL_FLUSH_CTL; | ||
92 | |||
93 | DBG("%s: flush=%08x", mdp5_crtc->name, flush); | ||
94 | |||
95 | mdp5_write(mdp5_kms, REG_MDP5_CTL_FLUSH(id), flush); | ||
96 | } | ||
97 | 70 | ||
98 | static void update_fb(struct drm_crtc *crtc, struct drm_framebuffer *new_fb) | 71 | static void crtc_flush(struct drm_crtc *crtc, u32 flush_mask) |
99 | { | 72 | { |
100 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 73 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
101 | struct drm_framebuffer *old_fb = mdp5_crtc->fb; | ||
102 | |||
103 | /* grab reference to incoming scanout fb: */ | ||
104 | drm_framebuffer_reference(new_fb); | ||
105 | mdp5_crtc->base.primary->fb = new_fb; | ||
106 | mdp5_crtc->fb = new_fb; | ||
107 | 74 | ||
108 | if (old_fb) | 75 | DBG("%s: flush=%08x", mdp5_crtc->name, flush_mask); |
109 | drm_flip_work_queue(&mdp5_crtc->unref_fb_work, old_fb); | 76 | mdp5_ctl_commit(mdp5_crtc->ctl, flush_mask); |
110 | } | 77 | } |
111 | 78 | ||
112 | /* unlike update_fb(), take a ref to the new scanout fb *before* updating | 79 | /* |
113 | * plane, then call this. Needed to ensure we don't unref the buffer that | 80 | * flush updates, to make sure hw is updated to new scanout fb, |
114 | * is actually still being scanned out. | 81 | * so that we can safely queue unref to current fb (ie. next |
115 | * | 82 | * vblank we know hw is done w/ previous scanout_fb). |
116 | * Note that this whole thing goes away with atomic.. since we can defer | ||
117 | * calling into driver until rendering is done. | ||
118 | */ | 83 | */ |
119 | static void update_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb) | 84 | static void crtc_flush_all(struct drm_crtc *crtc) |
120 | { | 85 | { |
121 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 86 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
87 | struct drm_plane *plane; | ||
88 | uint32_t flush_mask = 0; | ||
122 | 89 | ||
123 | /* flush updates, to make sure hw is updated to new scanout fb, | 90 | /* we could have already released CTL in the disable path: */ |
124 | * so that we can safely queue unref to current fb (ie. next | 91 | if (!mdp5_crtc->ctl) |
125 | * vblank we know hw is done w/ previous scanout_fb). | 92 | return; |
126 | */ | ||
127 | crtc_flush(crtc); | ||
128 | |||
129 | if (mdp5_crtc->scanout_fb) | ||
130 | drm_flip_work_queue(&mdp5_crtc->unref_fb_work, | ||
131 | mdp5_crtc->scanout_fb); | ||
132 | 93 | ||
133 | mdp5_crtc->scanout_fb = fb; | 94 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
95 | flush_mask |= mdp5_plane_get_flush(plane); | ||
96 | } | ||
97 | flush_mask |= mdp5_ctl_get_flush(mdp5_crtc->ctl); | ||
98 | flush_mask |= mdp5_lm_get_flush(mdp5_crtc->lm); | ||
134 | 99 | ||
135 | /* enable vblank to complete flip: */ | 100 | crtc_flush(crtc, flush_mask); |
136 | request_pending(crtc, PENDING_FLIP); | ||
137 | } | 101 | } |
138 | 102 | ||
139 | /* if file!=NULL, this is preclose potential cancel-flip path */ | 103 | /* if file!=NULL, this is preclose potential cancel-flip path */ |
@@ -142,7 +106,8 @@ static void complete_flip(struct drm_crtc *crtc, struct drm_file *file) | |||
142 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 106 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
143 | struct drm_device *dev = crtc->dev; | 107 | struct drm_device *dev = crtc->dev; |
144 | struct drm_pending_vblank_event *event; | 108 | struct drm_pending_vblank_event *event; |
145 | unsigned long flags, i; | 109 | struct drm_plane *plane; |
110 | unsigned long flags; | ||
146 | 111 | ||
147 | spin_lock_irqsave(&dev->event_lock, flags); | 112 | spin_lock_irqsave(&dev->event_lock, flags); |
148 | event = mdp5_crtc->event; | 113 | event = mdp5_crtc->event; |
@@ -153,50 +118,22 @@ static void complete_flip(struct drm_crtc *crtc, struct drm_file *file) | |||
153 | */ | 118 | */ |
154 | if (!file || (event->base.file_priv == file)) { | 119 | if (!file || (event->base.file_priv == file)) { |
155 | mdp5_crtc->event = NULL; | 120 | mdp5_crtc->event = NULL; |
121 | DBG("%s: send event: %p", mdp5_crtc->name, event); | ||
156 | drm_send_vblank_event(dev, mdp5_crtc->id, event); | 122 | drm_send_vblank_event(dev, mdp5_crtc->id, event); |
157 | } | 123 | } |
158 | } | 124 | } |
159 | spin_unlock_irqrestore(&dev->event_lock, flags); | 125 | spin_unlock_irqrestore(&dev->event_lock, flags); |
160 | 126 | ||
161 | for (i = 0; i < ARRAY_SIZE(mdp5_crtc->planes); i++) { | 127 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
162 | struct drm_plane *plane = mdp5_crtc->planes[i]; | 128 | mdp5_plane_complete_flip(plane); |
163 | if (plane) | ||
164 | mdp5_plane_complete_flip(plane); | ||
165 | } | 129 | } |
166 | } | 130 | } |
167 | 131 | ||
168 | static void pageflip_cb(struct msm_fence_cb *cb) | ||
169 | { | ||
170 | struct mdp5_crtc *mdp5_crtc = | ||
171 | container_of(cb, struct mdp5_crtc, pageflip_cb); | ||
172 | struct drm_crtc *crtc = &mdp5_crtc->base; | ||
173 | struct drm_framebuffer *fb = mdp5_crtc->fb; | ||
174 | |||
175 | if (!fb) | ||
176 | return; | ||
177 | |||
178 | drm_framebuffer_reference(fb); | ||
179 | mdp5_plane_set_scanout(mdp5_crtc->plane, fb); | ||
180 | update_scanout(crtc, fb); | ||
181 | } | ||
182 | |||
183 | static void unref_fb_worker(struct drm_flip_work *work, void *val) | ||
184 | { | ||
185 | struct mdp5_crtc *mdp5_crtc = | ||
186 | container_of(work, struct mdp5_crtc, unref_fb_work); | ||
187 | struct drm_device *dev = mdp5_crtc->base.dev; | ||
188 | |||
189 | mutex_lock(&dev->mode_config.mutex); | ||
190 | drm_framebuffer_unreference(val); | ||
191 | mutex_unlock(&dev->mode_config.mutex); | ||
192 | } | ||
193 | |||
194 | static void mdp5_crtc_destroy(struct drm_crtc *crtc) | 132 | static void mdp5_crtc_destroy(struct drm_crtc *crtc) |
195 | { | 133 | { |
196 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 134 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
197 | 135 | ||
198 | drm_crtc_cleanup(crtc); | 136 | drm_crtc_cleanup(crtc); |
199 | drm_flip_work_cleanup(&mdp5_crtc->unref_fb_work); | ||
200 | 137 | ||
201 | kfree(mdp5_crtc); | 138 | kfree(mdp5_crtc); |
202 | } | 139 | } |
@@ -214,6 +151,8 @@ static void mdp5_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
214 | mdp5_enable(mdp5_kms); | 151 | mdp5_enable(mdp5_kms); |
215 | mdp_irq_register(&mdp5_kms->base, &mdp5_crtc->err); | 152 | mdp_irq_register(&mdp5_kms->base, &mdp5_crtc->err); |
216 | } else { | 153 | } else { |
154 | /* set STAGE_UNUSED for all layers */ | ||
155 | mdp5_ctl_blend(mdp5_crtc->ctl, mdp5_crtc->lm, 0x00000000); | ||
217 | mdp_irq_unregister(&mdp5_kms->base, &mdp5_crtc->err); | 156 | mdp_irq_unregister(&mdp5_kms->base, &mdp5_crtc->err); |
218 | mdp5_disable(mdp5_kms); | 157 | mdp5_disable(mdp5_kms); |
219 | } | 158 | } |
@@ -228,54 +167,78 @@ static bool mdp5_crtc_mode_fixup(struct drm_crtc *crtc, | |||
228 | return true; | 167 | return true; |
229 | } | 168 | } |
230 | 169 | ||
170 | /* | ||
171 | * blend_setup() - blend all the planes of a CRTC | ||
172 | * | ||
173 | * When border is enabled, the border color will ALWAYS be the base layer. | ||
174 | * Therefore, the first plane (private RGB pipe) will start at STAGE0. | ||
175 | * If disabled, the first plane starts at STAGE_BASE. | ||
176 | * | ||
177 | * Note: | ||
178 | * Border is not enabled here because the private plane is exactly | ||
179 | * the CRTC resolution. | ||
180 | */ | ||
231 | static void blend_setup(struct drm_crtc *crtc) | 181 | static void blend_setup(struct drm_crtc *crtc) |
232 | { | 182 | { |
233 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 183 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
234 | struct mdp5_kms *mdp5_kms = get_kms(crtc); | 184 | struct mdp5_kms *mdp5_kms = get_kms(crtc); |
235 | int id = mdp5_crtc->id; | 185 | struct drm_plane *plane; |
186 | const struct mdp5_cfg_hw *hw_cfg; | ||
187 | uint32_t lm = mdp5_crtc->lm, blend_cfg = 0; | ||
188 | unsigned long flags; | ||
189 | #define blender(stage) ((stage) - STAGE_BASE) | ||
236 | 190 | ||
237 | /* | 191 | hw_cfg = mdp5_cfg_get_hw_config(mdp5_kms->cfg); |
238 | * Hard-coded setup for now until I figure out how the | ||
239 | * layer-mixer works | ||
240 | */ | ||
241 | 192 | ||
242 | /* LM[id]: */ | 193 | spin_lock_irqsave(&mdp5_crtc->lm_lock, flags); |
243 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_COLOR_OUT(id), | 194 | |
244 | MDP5_LM_BLEND_COLOR_OUT_STAGE0_FG_ALPHA); | 195 | /* ctl could be released already when we are shutting down: */ |
245 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_OP_MODE(id, 0), | 196 | if (!mdp5_crtc->ctl) |
246 | MDP5_LM_BLEND_OP_MODE_FG_ALPHA(FG_CONST) | | 197 | goto out; |
247 | MDP5_LM_BLEND_OP_MODE_BG_ALPHA(FG_PIXEL) | | ||
248 | MDP5_LM_BLEND_OP_MODE_BG_INV_ALPHA); | ||
249 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_FG_ALPHA(id, 0), 0xff); | ||
250 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_BG_ALPHA(id, 0), 0x00); | ||
251 | |||
252 | /* NOTE: seems that LM[n] and CTL[m], we do not need n==m.. but | ||
253 | * we want to be setting CTL[m].LAYER[n]. Not sure what the | ||
254 | * point of having CTL[m].LAYER[o] (for o!=n).. maybe that is | ||
255 | * used when chaining up mixers for high resolution displays? | ||
256 | */ | ||
257 | 198 | ||
258 | /* CTL[id]: */ | 199 | drm_atomic_crtc_for_each_plane(plane, crtc) { |
259 | mdp5_write(mdp5_kms, REG_MDP5_CTL_LAYER_REG(id, 0), | 200 | enum mdp_mixer_stage_id stage = |
260 | MDP5_CTL_LAYER_REG_RGB0(STAGE0) | | 201 | to_mdp5_plane_state(plane->state)->stage; |
261 | MDP5_CTL_LAYER_REG_BORDER_COLOR); | 202 | |
262 | mdp5_write(mdp5_kms, REG_MDP5_CTL_LAYER_REG(id, 1), 0); | 203 | /* |
263 | mdp5_write(mdp5_kms, REG_MDP5_CTL_LAYER_REG(id, 2), 0); | 204 | * Note: This cannot happen with current implementation but |
264 | mdp5_write(mdp5_kms, REG_MDP5_CTL_LAYER_REG(id, 3), 0); | 205 | * we need to check this condition once z property is added |
265 | mdp5_write(mdp5_kms, REG_MDP5_CTL_LAYER_REG(id, 4), 0); | 206 | */ |
207 | BUG_ON(stage > hw_cfg->lm.nb_stages); | ||
208 | |||
209 | /* LM */ | ||
210 | mdp5_write(mdp5_kms, | ||
211 | REG_MDP5_LM_BLEND_OP_MODE(lm, blender(stage)), | ||
212 | MDP5_LM_BLEND_OP_MODE_FG_ALPHA(FG_CONST) | | ||
213 | MDP5_LM_BLEND_OP_MODE_BG_ALPHA(BG_CONST)); | ||
214 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_FG_ALPHA(lm, | ||
215 | blender(stage)), 0xff); | ||
216 | mdp5_write(mdp5_kms, REG_MDP5_LM_BLEND_BG_ALPHA(lm, | ||
217 | blender(stage)), 0x00); | ||
218 | /* CTL */ | ||
219 | blend_cfg |= mdp_ctl_blend_mask(mdp5_plane_pipe(plane), stage); | ||
220 | DBG("%s: blending pipe %s on stage=%d", mdp5_crtc->name, | ||
221 | pipe2name(mdp5_plane_pipe(plane)), stage); | ||
222 | } | ||
223 | |||
224 | DBG("%s: lm%d: blend config = 0x%08x", mdp5_crtc->name, lm, blend_cfg); | ||
225 | mdp5_ctl_blend(mdp5_crtc->ctl, lm, blend_cfg); | ||
226 | |||
227 | out: | ||
228 | spin_unlock_irqrestore(&mdp5_crtc->lm_lock, flags); | ||
266 | } | 229 | } |
267 | 230 | ||
268 | static int mdp5_crtc_mode_set(struct drm_crtc *crtc, | 231 | static void mdp5_crtc_mode_set_nofb(struct drm_crtc *crtc) |
269 | struct drm_display_mode *mode, | ||
270 | struct drm_display_mode *adjusted_mode, | ||
271 | int x, int y, | ||
272 | struct drm_framebuffer *old_fb) | ||
273 | { | 232 | { |
274 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 233 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
275 | struct mdp5_kms *mdp5_kms = get_kms(crtc); | 234 | struct mdp5_kms *mdp5_kms = get_kms(crtc); |
276 | int ret; | 235 | unsigned long flags; |
236 | struct drm_display_mode *mode; | ||
277 | 237 | ||
278 | mode = adjusted_mode; | 238 | if (WARN_ON(!crtc->state)) |
239 | return; | ||
240 | |||
241 | mode = &crtc->state->adjusted_mode; | ||
279 | 242 | ||
280 | DBG("%s: set mode: %d:\"%s\" %d %d %d %d %d %d %d %d %d %d 0x%x 0x%x", | 243 | DBG("%s: set mode: %d:\"%s\" %d %d %d %d %d %d %d %d %d %d 0x%x 0x%x", |
281 | mdp5_crtc->name, mode->base.id, mode->name, | 244 | mdp5_crtc->name, mode->base.id, mode->name, |
@@ -286,28 +249,11 @@ static int mdp5_crtc_mode_set(struct drm_crtc *crtc, | |||
286 | mode->vsync_end, mode->vtotal, | 249 | mode->vsync_end, mode->vtotal, |
287 | mode->type, mode->flags); | 250 | mode->type, mode->flags); |
288 | 251 | ||
289 | /* grab extra ref for update_scanout() */ | 252 | spin_lock_irqsave(&mdp5_crtc->lm_lock, flags); |
290 | drm_framebuffer_reference(crtc->primary->fb); | 253 | mdp5_write(mdp5_kms, REG_MDP5_LM_OUT_SIZE(mdp5_crtc->lm), |
291 | |||
292 | ret = mdp5_plane_mode_set(mdp5_crtc->plane, crtc, crtc->primary->fb, | ||
293 | 0, 0, mode->hdisplay, mode->vdisplay, | ||
294 | x << 16, y << 16, | ||
295 | mode->hdisplay << 16, mode->vdisplay << 16); | ||
296 | if (ret) { | ||
297 | drm_framebuffer_unreference(crtc->primary->fb); | ||
298 | dev_err(crtc->dev->dev, "%s: failed to set mode on plane: %d\n", | ||
299 | mdp5_crtc->name, ret); | ||
300 | return ret; | ||
301 | } | ||
302 | |||
303 | mdp5_write(mdp5_kms, REG_MDP5_LM_OUT_SIZE(mdp5_crtc->id), | ||
304 | MDP5_LM_OUT_SIZE_WIDTH(mode->hdisplay) | | 254 | MDP5_LM_OUT_SIZE_WIDTH(mode->hdisplay) | |
305 | MDP5_LM_OUT_SIZE_HEIGHT(mode->vdisplay)); | 255 | MDP5_LM_OUT_SIZE_HEIGHT(mode->vdisplay)); |
306 | 256 | spin_unlock_irqrestore(&mdp5_crtc->lm_lock, flags); | |
307 | update_fb(crtc, crtc->primary->fb); | ||
308 | update_scanout(crtc, crtc->primary->fb); | ||
309 | |||
310 | return 0; | ||
311 | } | 257 | } |
312 | 258 | ||
313 | static void mdp5_crtc_prepare(struct drm_crtc *crtc) | 259 | static void mdp5_crtc_prepare(struct drm_crtc *crtc) |
@@ -321,66 +267,119 @@ static void mdp5_crtc_prepare(struct drm_crtc *crtc) | |||
321 | 267 | ||
322 | static void mdp5_crtc_commit(struct drm_crtc *crtc) | 268 | static void mdp5_crtc_commit(struct drm_crtc *crtc) |
323 | { | 269 | { |
270 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | ||
271 | DBG("%s", mdp5_crtc->name); | ||
324 | mdp5_crtc_dpms(crtc, DRM_MODE_DPMS_ON); | 272 | mdp5_crtc_dpms(crtc, DRM_MODE_DPMS_ON); |
325 | crtc_flush(crtc); | 273 | crtc_flush_all(crtc); |
326 | /* drop the ref to mdp clk's that we got in prepare: */ | 274 | /* drop the ref to mdp clk's that we got in prepare: */ |
327 | mdp5_disable(get_kms(crtc)); | 275 | mdp5_disable(get_kms(crtc)); |
328 | } | 276 | } |
329 | 277 | ||
330 | static int mdp5_crtc_mode_set_base(struct drm_crtc *crtc, int x, int y, | 278 | static void mdp5_crtc_load_lut(struct drm_crtc *crtc) |
331 | struct drm_framebuffer *old_fb) | 279 | { |
280 | } | ||
281 | |||
282 | struct plane_state { | ||
283 | struct drm_plane *plane; | ||
284 | struct mdp5_plane_state *state; | ||
285 | }; | ||
286 | |||
287 | static int pstate_cmp(const void *a, const void *b) | ||
288 | { | ||
289 | struct plane_state *pa = (struct plane_state *)a; | ||
290 | struct plane_state *pb = (struct plane_state *)b; | ||
291 | return pa->state->zpos - pb->state->zpos; | ||
292 | } | ||
293 | |||
294 | static int mdp5_crtc_atomic_check(struct drm_crtc *crtc, | ||
295 | struct drm_crtc_state *state) | ||
332 | { | 296 | { |
333 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 297 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
334 | struct drm_plane *plane = mdp5_crtc->plane; | 298 | struct mdp5_kms *mdp5_kms = get_kms(crtc); |
335 | struct drm_display_mode *mode = &crtc->mode; | 299 | struct drm_plane *plane; |
336 | int ret; | 300 | struct drm_device *dev = crtc->dev; |
337 | 301 | struct plane_state pstates[STAGE3 + 1]; | |
338 | /* grab extra ref for update_scanout() */ | 302 | int cnt = 0, i; |
339 | drm_framebuffer_reference(crtc->primary->fb); | 303 | |
340 | 304 | DBG("%s: check", mdp5_crtc->name); | |
341 | ret = mdp5_plane_mode_set(plane, crtc, crtc->primary->fb, | 305 | |
342 | 0, 0, mode->hdisplay, mode->vdisplay, | 306 | if (mdp5_crtc->event) { |
343 | x << 16, y << 16, | 307 | dev_err(dev->dev, "already pending flip!\n"); |
344 | mode->hdisplay << 16, mode->vdisplay << 16); | 308 | return -EBUSY; |
345 | if (ret) { | ||
346 | drm_framebuffer_unreference(crtc->primary->fb); | ||
347 | return ret; | ||
348 | } | 309 | } |
349 | 310 | ||
350 | update_fb(crtc, crtc->primary->fb); | 311 | /* request a free CTL, if none is already allocated for this CRTC */ |
351 | update_scanout(crtc, crtc->primary->fb); | 312 | if (state->enable && !mdp5_crtc->ctl) { |
313 | mdp5_crtc->ctl = mdp5_ctlm_request(mdp5_kms->ctlm, crtc); | ||
314 | if (WARN_ON(!mdp5_crtc->ctl)) | ||
315 | return -EINVAL; | ||
316 | } | ||
317 | |||
318 | /* verify that there are not too many planes attached to crtc | ||
319 | * and that we don't have conflicting mixer stages: | ||
320 | */ | ||
321 | drm_atomic_crtc_state_for_each_plane(plane, state) { | ||
322 | struct drm_plane_state *pstate; | ||
323 | |||
324 | if (cnt >= ARRAY_SIZE(pstates)) { | ||
325 | dev_err(dev->dev, "too many planes!\n"); | ||
326 | return -EINVAL; | ||
327 | } | ||
328 | |||
329 | pstate = state->state->plane_states[drm_plane_index(plane)]; | ||
330 | |||
331 | /* plane might not have changed, in which case take | ||
332 | * current state: | ||
333 | */ | ||
334 | if (!pstate) | ||
335 | pstate = plane->state; | ||
336 | |||
337 | pstates[cnt].plane = plane; | ||
338 | pstates[cnt].state = to_mdp5_plane_state(pstate); | ||
339 | |||
340 | cnt++; | ||
341 | } | ||
342 | |||
343 | sort(pstates, cnt, sizeof(pstates[0]), pstate_cmp, NULL); | ||
344 | |||
345 | for (i = 0; i < cnt; i++) { | ||
346 | pstates[i].state->stage = STAGE_BASE + i; | ||
347 | DBG("%s: assign pipe %s on stage=%d", mdp5_crtc->name, | ||
348 | pipe2name(mdp5_plane_pipe(pstates[i].plane)), | ||
349 | pstates[i].state->stage); | ||
350 | } | ||
352 | 351 | ||
353 | return 0; | 352 | return 0; |
354 | } | 353 | } |
355 | 354 | ||
356 | static void mdp5_crtc_load_lut(struct drm_crtc *crtc) | 355 | static void mdp5_crtc_atomic_begin(struct drm_crtc *crtc) |
357 | { | 356 | { |
357 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | ||
358 | DBG("%s: begin", mdp5_crtc->name); | ||
358 | } | 359 | } |
359 | 360 | ||
360 | static int mdp5_crtc_page_flip(struct drm_crtc *crtc, | 361 | static void mdp5_crtc_atomic_flush(struct drm_crtc *crtc) |
361 | struct drm_framebuffer *new_fb, | ||
362 | struct drm_pending_vblank_event *event, | ||
363 | uint32_t page_flip_flags) | ||
364 | { | 362 | { |
365 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 363 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
366 | struct drm_device *dev = crtc->dev; | 364 | struct drm_device *dev = crtc->dev; |
367 | struct drm_gem_object *obj; | ||
368 | unsigned long flags; | 365 | unsigned long flags; |
369 | 366 | ||
370 | if (mdp5_crtc->event) { | 367 | DBG("%s: flush", mdp5_crtc->name); |
371 | dev_err(dev->dev, "already pending flip!\n"); | ||
372 | return -EBUSY; | ||
373 | } | ||
374 | 368 | ||
375 | obj = msm_framebuffer_bo(new_fb, 0); | 369 | WARN_ON(mdp5_crtc->event); |
376 | 370 | ||
377 | spin_lock_irqsave(&dev->event_lock, flags); | 371 | spin_lock_irqsave(&dev->event_lock, flags); |
378 | mdp5_crtc->event = event; | 372 | mdp5_crtc->event = crtc->state->event; |
379 | spin_unlock_irqrestore(&dev->event_lock, flags); | 373 | spin_unlock_irqrestore(&dev->event_lock, flags); |
380 | 374 | ||
381 | update_fb(crtc, new_fb); | 375 | blend_setup(crtc); |
376 | crtc_flush_all(crtc); | ||
377 | request_pending(crtc, PENDING_FLIP); | ||
382 | 378 | ||
383 | return msm_gem_queue_inactive_cb(obj, &mdp5_crtc->pageflip_cb); | 379 | if (mdp5_crtc->ctl && !crtc->state->enable) { |
380 | mdp5_ctl_release(mdp5_crtc->ctl); | ||
381 | mdp5_crtc->ctl = NULL; | ||
382 | } | ||
384 | } | 383 | } |
385 | 384 | ||
386 | static int mdp5_crtc_set_property(struct drm_crtc *crtc, | 385 | static int mdp5_crtc_set_property(struct drm_crtc *crtc, |
@@ -391,27 +390,33 @@ static int mdp5_crtc_set_property(struct drm_crtc *crtc, | |||
391 | } | 390 | } |
392 | 391 | ||
393 | static const struct drm_crtc_funcs mdp5_crtc_funcs = { | 392 | static const struct drm_crtc_funcs mdp5_crtc_funcs = { |
394 | .set_config = drm_crtc_helper_set_config, | 393 | .set_config = drm_atomic_helper_set_config, |
395 | .destroy = mdp5_crtc_destroy, | 394 | .destroy = mdp5_crtc_destroy, |
396 | .page_flip = mdp5_crtc_page_flip, | 395 | .page_flip = drm_atomic_helper_page_flip, |
397 | .set_property = mdp5_crtc_set_property, | 396 | .set_property = mdp5_crtc_set_property, |
397 | .reset = drm_atomic_helper_crtc_reset, | ||
398 | .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state, | ||
399 | .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state, | ||
398 | }; | 400 | }; |
399 | 401 | ||
400 | static const struct drm_crtc_helper_funcs mdp5_crtc_helper_funcs = { | 402 | static const struct drm_crtc_helper_funcs mdp5_crtc_helper_funcs = { |
401 | .dpms = mdp5_crtc_dpms, | 403 | .dpms = mdp5_crtc_dpms, |
402 | .mode_fixup = mdp5_crtc_mode_fixup, | 404 | .mode_fixup = mdp5_crtc_mode_fixup, |
403 | .mode_set = mdp5_crtc_mode_set, | 405 | .mode_set_nofb = mdp5_crtc_mode_set_nofb, |
406 | .mode_set = drm_helper_crtc_mode_set, | ||
407 | .mode_set_base = drm_helper_crtc_mode_set_base, | ||
404 | .prepare = mdp5_crtc_prepare, | 408 | .prepare = mdp5_crtc_prepare, |
405 | .commit = mdp5_crtc_commit, | 409 | .commit = mdp5_crtc_commit, |
406 | .mode_set_base = mdp5_crtc_mode_set_base, | ||
407 | .load_lut = mdp5_crtc_load_lut, | 410 | .load_lut = mdp5_crtc_load_lut, |
411 | .atomic_check = mdp5_crtc_atomic_check, | ||
412 | .atomic_begin = mdp5_crtc_atomic_begin, | ||
413 | .atomic_flush = mdp5_crtc_atomic_flush, | ||
408 | }; | 414 | }; |
409 | 415 | ||
410 | static void mdp5_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) | 416 | static void mdp5_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) |
411 | { | 417 | { |
412 | struct mdp5_crtc *mdp5_crtc = container_of(irq, struct mdp5_crtc, vblank); | 418 | struct mdp5_crtc *mdp5_crtc = container_of(irq, struct mdp5_crtc, vblank); |
413 | struct drm_crtc *crtc = &mdp5_crtc->base; | 419 | struct drm_crtc *crtc = &mdp5_crtc->base; |
414 | struct msm_drm_private *priv = crtc->dev->dev_private; | ||
415 | unsigned pending; | 420 | unsigned pending; |
416 | 421 | ||
417 | mdp_irq_unregister(&get_kms(crtc)->base, &mdp5_crtc->vblank); | 422 | mdp_irq_unregister(&get_kms(crtc)->base, &mdp5_crtc->vblank); |
@@ -420,16 +425,14 @@ static void mdp5_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus) | |||
420 | 425 | ||
421 | if (pending & PENDING_FLIP) { | 426 | if (pending & PENDING_FLIP) { |
422 | complete_flip(crtc, NULL); | 427 | complete_flip(crtc, NULL); |
423 | drm_flip_work_commit(&mdp5_crtc->unref_fb_work, priv->wq); | ||
424 | } | 428 | } |
425 | } | 429 | } |
426 | 430 | ||
427 | static void mdp5_crtc_err_irq(struct mdp_irq *irq, uint32_t irqstatus) | 431 | static void mdp5_crtc_err_irq(struct mdp_irq *irq, uint32_t irqstatus) |
428 | { | 432 | { |
429 | struct mdp5_crtc *mdp5_crtc = container_of(irq, struct mdp5_crtc, err); | 433 | struct mdp5_crtc *mdp5_crtc = container_of(irq, struct mdp5_crtc, err); |
430 | struct drm_crtc *crtc = &mdp5_crtc->base; | 434 | |
431 | DBG("%s: error: %08x", mdp5_crtc->name, irqstatus); | 435 | DBG("%s: error: %08x", mdp5_crtc->name, irqstatus); |
432 | crtc_flush(crtc); | ||
433 | } | 436 | } |
434 | 437 | ||
435 | uint32_t mdp5_crtc_vblank(struct drm_crtc *crtc) | 438 | uint32_t mdp5_crtc_vblank(struct drm_crtc *crtc) |
@@ -450,10 +453,9 @@ void mdp5_crtc_set_intf(struct drm_crtc *crtc, int intf, | |||
450 | { | 453 | { |
451 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 454 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
452 | struct mdp5_kms *mdp5_kms = get_kms(crtc); | 455 | struct mdp5_kms *mdp5_kms = get_kms(crtc); |
453 | static const enum mdp5_intfnum intfnum[] = { | 456 | uint32_t flush_mask = 0; |
454 | INTF0, INTF1, INTF2, INTF3, | ||
455 | }; | ||
456 | uint32_t intf_sel; | 457 | uint32_t intf_sel; |
458 | unsigned long flags; | ||
457 | 459 | ||
458 | /* now that we know what irq's we want: */ | 460 | /* now that we know what irq's we want: */ |
459 | mdp5_crtc->err.irqmask = intf2err(intf); | 461 | mdp5_crtc->err.irqmask = intf2err(intf); |
@@ -463,6 +465,7 @@ void mdp5_crtc_set_intf(struct drm_crtc *crtc, int intf, | |||
463 | if (!mdp5_kms) | 465 | if (!mdp5_kms) |
464 | return; | 466 | return; |
465 | 467 | ||
468 | spin_lock_irqsave(&mdp5_kms->resource_lock, flags); | ||
466 | intf_sel = mdp5_read(mdp5_kms, REG_MDP5_DISP_INTF_SEL); | 469 | intf_sel = mdp5_read(mdp5_kms, REG_MDP5_DISP_INTF_SEL); |
467 | 470 | ||
468 | switch (intf) { | 471 | switch (intf) { |
@@ -487,45 +490,25 @@ void mdp5_crtc_set_intf(struct drm_crtc *crtc, int intf, | |||
487 | break; | 490 | break; |
488 | } | 491 | } |
489 | 492 | ||
490 | blend_setup(crtc); | 493 | mdp5_write(mdp5_kms, REG_MDP5_DISP_INTF_SEL, intf_sel); |
494 | spin_unlock_irqrestore(&mdp5_kms->resource_lock, flags); | ||
491 | 495 | ||
492 | DBG("%s: intf_sel=%08x", mdp5_crtc->name, intf_sel); | 496 | DBG("%s: intf_sel=%08x", mdp5_crtc->name, intf_sel); |
497 | mdp5_ctl_set_intf(mdp5_crtc->ctl, intf); | ||
498 | flush_mask |= mdp5_ctl_get_flush(mdp5_crtc->ctl); | ||
499 | flush_mask |= mdp5_lm_get_flush(mdp5_crtc->lm); | ||
493 | 500 | ||
494 | mdp5_write(mdp5_kms, REG_MDP5_DISP_INTF_SEL, intf_sel); | 501 | crtc_flush(crtc, flush_mask); |
495 | mdp5_write(mdp5_kms, REG_MDP5_CTL_OP(mdp5_crtc->id), | ||
496 | MDP5_CTL_OP_MODE(MODE_NONE) | | ||
497 | MDP5_CTL_OP_INTF_NUM(intfnum[intf])); | ||
498 | |||
499 | crtc_flush(crtc); | ||
500 | } | 502 | } |
501 | 503 | ||
502 | static void set_attach(struct drm_crtc *crtc, enum mdp5_pipe pipe_id, | 504 | int mdp5_crtc_get_lm(struct drm_crtc *crtc) |
503 | struct drm_plane *plane) | ||
504 | { | 505 | { |
505 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); | 506 | struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc); |
506 | 507 | ||
507 | BUG_ON(pipe_id >= ARRAY_SIZE(mdp5_crtc->planes)); | 508 | if (WARN_ON(!crtc)) |
509 | return -EINVAL; | ||
508 | 510 | ||
509 | if (mdp5_crtc->planes[pipe_id] == plane) | 511 | return mdp5_crtc->lm; |
510 | return; | ||
511 | |||
512 | mdp5_crtc->planes[pipe_id] = plane; | ||
513 | blend_setup(crtc); | ||
514 | if (mdp5_crtc->enabled && (plane != mdp5_crtc->plane)) | ||
515 | crtc_flush(crtc); | ||
516 | } | ||
517 | |||
518 | void mdp5_crtc_attach(struct drm_crtc *crtc, struct drm_plane *plane) | ||
519 | { | ||
520 | set_attach(crtc, mdp5_plane_pipe(plane), plane); | ||
521 | } | ||
522 | |||
523 | void mdp5_crtc_detach(struct drm_crtc *crtc, struct drm_plane *plane) | ||
524 | { | ||
525 | /* don't actually detatch our primary plane: */ | ||
526 | if (to_mdp5_crtc(crtc)->plane == plane) | ||
527 | return; | ||
528 | set_attach(crtc, mdp5_plane_pipe(plane), NULL); | ||
529 | } | 512 | } |
530 | 513 | ||
531 | /* initialize crtc */ | 514 | /* initialize crtc */ |
@@ -534,18 +517,17 @@ struct drm_crtc *mdp5_crtc_init(struct drm_device *dev, | |||
534 | { | 517 | { |
535 | struct drm_crtc *crtc = NULL; | 518 | struct drm_crtc *crtc = NULL; |
536 | struct mdp5_crtc *mdp5_crtc; | 519 | struct mdp5_crtc *mdp5_crtc; |
537 | int ret; | ||
538 | 520 | ||
539 | mdp5_crtc = kzalloc(sizeof(*mdp5_crtc), GFP_KERNEL); | 521 | mdp5_crtc = kzalloc(sizeof(*mdp5_crtc), GFP_KERNEL); |
540 | if (!mdp5_crtc) { | 522 | if (!mdp5_crtc) |
541 | ret = -ENOMEM; | 523 | return ERR_PTR(-ENOMEM); |
542 | goto fail; | ||
543 | } | ||
544 | 524 | ||
545 | crtc = &mdp5_crtc->base; | 525 | crtc = &mdp5_crtc->base; |
546 | 526 | ||
547 | mdp5_crtc->plane = plane; | ||
548 | mdp5_crtc->id = id; | 527 | mdp5_crtc->id = id; |
528 | mdp5_crtc->lm = GET_LM_ID(id); | ||
529 | |||
530 | spin_lock_init(&mdp5_crtc->lm_lock); | ||
549 | 531 | ||
550 | mdp5_crtc->vblank.irq = mdp5_crtc_vblank_irq; | 532 | mdp5_crtc->vblank.irq = mdp5_crtc_vblank_irq; |
551 | mdp5_crtc->err.irq = mdp5_crtc_err_irq; | 533 | mdp5_crtc->err.irq = mdp5_crtc_err_irq; |
@@ -553,23 +535,11 @@ struct drm_crtc *mdp5_crtc_init(struct drm_device *dev, | |||
553 | snprintf(mdp5_crtc->name, sizeof(mdp5_crtc->name), "%s:%d", | 535 | snprintf(mdp5_crtc->name, sizeof(mdp5_crtc->name), "%s:%d", |
554 | pipe2name(mdp5_plane_pipe(plane)), id); | 536 | pipe2name(mdp5_plane_pipe(plane)), id); |
555 | 537 | ||
556 | ret = drm_flip_work_init(&mdp5_crtc->unref_fb_work, 16, | ||
557 | "unref fb", unref_fb_worker); | ||
558 | if (ret) | ||
559 | goto fail; | ||
560 | |||
561 | INIT_FENCE_CB(&mdp5_crtc->pageflip_cb, pageflip_cb); | ||
562 | |||
563 | drm_crtc_init_with_planes(dev, crtc, plane, NULL, &mdp5_crtc_funcs); | 538 | drm_crtc_init_with_planes(dev, crtc, plane, NULL, &mdp5_crtc_funcs); |
564 | drm_crtc_helper_add(crtc, &mdp5_crtc_helper_funcs); | 539 | drm_crtc_helper_add(crtc, &mdp5_crtc_helper_funcs); |
540 | plane->crtc = crtc; | ||
565 | 541 | ||
566 | mdp5_plane_install_properties(mdp5_crtc->plane, &crtc->base); | 542 | mdp5_plane_install_properties(plane, &crtc->base); |
567 | 543 | ||
568 | return crtc; | 544 | return crtc; |
569 | |||
570 | fail: | ||
571 | if (crtc) | ||
572 | mdp5_crtc_destroy(crtc); | ||
573 | |||
574 | return ERR_PTR(ret); | ||
575 | } | 545 | } |
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.c new file mode 100644 index 000000000000..dea4505ac963 --- /dev/null +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.c | |||
@@ -0,0 +1,322 @@ | |||
1 | /* | ||
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
3 | * | ||
4 | * This program is free software; you can redistribute it and/or modify | ||
5 | * it under the terms of the GNU General Public License version 2 and | ||
6 | * only version 2 as published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope that it will be useful, | ||
9 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
11 | * GNU General Public License for more details. | ||
12 | */ | ||
13 | |||
14 | #include "mdp5_kms.h" | ||
15 | #include "mdp5_ctl.h" | ||
16 | |||
17 | /* | ||
18 | * CTL - MDP Control Pool Manager | ||
19 | * | ||
20 | * Controls are shared between all CRTCs. | ||
21 | * | ||
22 | * They are intended to be used for data path configuration. | ||
23 | * The top level register programming describes the complete data path for | ||
24 | * a specific data path ID - REG_MDP5_CTL_*(<id>, ...) | ||
25 | * | ||
26 | * Hardware capabilities determine the number of concurrent data paths | ||
27 | * | ||
28 | * In certain use cases (high-resolution dual pipe), one single CTL can be | ||
29 | * shared across multiple CRTCs. | ||
30 | * | ||
31 | * Because the number of CTLs can be less than the number of CRTCs, | ||
32 | * CTLs are dynamically allocated from a pool of CTLs, only once a CRTC is | ||
33 | * requested by the client (in mdp5_crtc_mode_set()). | ||
34 | */ | ||
35 | |||
36 | struct mdp5_ctl { | ||
37 | struct mdp5_ctl_manager *ctlm; | ||
38 | |||
39 | u32 id; | ||
40 | |||
41 | /* whether this CTL has been allocated or not: */ | ||
42 | bool busy; | ||
43 | |||
44 | /* memory output connection (@see mdp5_ctl_mode): */ | ||
45 | u32 mode; | ||
46 | |||
47 | /* REG_MDP5_CTL_*(<id>) registers access info + lock: */ | ||
48 | spinlock_t hw_lock; | ||
49 | u32 reg_offset; | ||
50 | |||
51 | /* flush mask used to commit CTL registers */ | ||
52 | u32 flush_mask; | ||
53 | |||
54 | bool cursor_on; | ||
55 | |||
56 | struct drm_crtc *crtc; | ||
57 | }; | ||
58 | |||
59 | struct mdp5_ctl_manager { | ||
60 | struct drm_device *dev; | ||
61 | |||
62 | /* number of CTL / Layer Mixers in this hw config: */ | ||
63 | u32 nlm; | ||
64 | u32 nctl; | ||
65 | |||
66 | /* pool of CTLs + lock to protect resource allocation (ctls[i].busy) */ | ||
67 | spinlock_t pool_lock; | ||
68 | struct mdp5_ctl ctls[MAX_CTL]; | ||
69 | }; | ||
70 | |||
71 | static inline | ||
72 | struct mdp5_kms *get_kms(struct mdp5_ctl_manager *ctl_mgr) | ||
73 | { | ||
74 | struct msm_drm_private *priv = ctl_mgr->dev->dev_private; | ||
75 | |||
76 | return to_mdp5_kms(to_mdp_kms(priv->kms)); | ||
77 | } | ||
78 | |||
79 | static inline | ||
80 | void ctl_write(struct mdp5_ctl *ctl, u32 reg, u32 data) | ||
81 | { | ||
82 | struct mdp5_kms *mdp5_kms = get_kms(ctl->ctlm); | ||
83 | |||
84 | (void)ctl->reg_offset; /* TODO use this instead of mdp5_write */ | ||
85 | mdp5_write(mdp5_kms, reg, data); | ||
86 | } | ||
87 | |||
88 | static inline | ||
89 | u32 ctl_read(struct mdp5_ctl *ctl, u32 reg) | ||
90 | { | ||
91 | struct mdp5_kms *mdp5_kms = get_kms(ctl->ctlm); | ||
92 | |||
93 | (void)ctl->reg_offset; /* TODO use this instead of mdp5_write */ | ||
94 | return mdp5_read(mdp5_kms, reg); | ||
95 | } | ||
96 | |||
97 | |||
98 | int mdp5_ctl_set_intf(struct mdp5_ctl *ctl, enum mdp5_intf intf) | ||
99 | { | ||
100 | unsigned long flags; | ||
101 | static const enum mdp5_intfnum intfnum[] = { | ||
102 | INTF0, INTF1, INTF2, INTF3, | ||
103 | }; | ||
104 | |||
105 | spin_lock_irqsave(&ctl->hw_lock, flags); | ||
106 | ctl_write(ctl, REG_MDP5_CTL_OP(ctl->id), | ||
107 | MDP5_CTL_OP_MODE(ctl->mode) | | ||
108 | MDP5_CTL_OP_INTF_NUM(intfnum[intf])); | ||
109 | spin_unlock_irqrestore(&ctl->hw_lock, flags); | ||
110 | |||
111 | return 0; | ||
112 | } | ||
113 | |||
114 | int mdp5_ctl_set_cursor(struct mdp5_ctl *ctl, bool enable) | ||
115 | { | ||
116 | struct mdp5_ctl_manager *ctl_mgr = ctl->ctlm; | ||
117 | unsigned long flags; | ||
118 | u32 blend_cfg; | ||
119 | int lm; | ||
120 | |||
121 | lm = mdp5_crtc_get_lm(ctl->crtc); | ||
122 | if (unlikely(WARN_ON(lm < 0))) { | ||
123 | dev_err(ctl_mgr->dev->dev, "CTL %d cannot find LM: %d", | ||
124 | ctl->id, lm); | ||
125 | return -EINVAL; | ||
126 | } | ||
127 | |||
128 | spin_lock_irqsave(&ctl->hw_lock, flags); | ||
129 | |||
130 | blend_cfg = ctl_read(ctl, REG_MDP5_CTL_LAYER_REG(ctl->id, lm)); | ||
131 | |||
132 | if (enable) | ||
133 | blend_cfg |= MDP5_CTL_LAYER_REG_CURSOR_OUT; | ||
134 | else | ||
135 | blend_cfg &= ~MDP5_CTL_LAYER_REG_CURSOR_OUT; | ||
136 | |||
137 | ctl_write(ctl, REG_MDP5_CTL_LAYER_REG(ctl->id, lm), blend_cfg); | ||
138 | |||
139 | spin_unlock_irqrestore(&ctl->hw_lock, flags); | ||
140 | |||
141 | ctl->cursor_on = enable; | ||
142 | |||
143 | return 0; | ||
144 | } | ||
145 | |||
146 | |||
147 | int mdp5_ctl_blend(struct mdp5_ctl *ctl, u32 lm, u32 blend_cfg) | ||
148 | { | ||
149 | unsigned long flags; | ||
150 | |||
151 | if (ctl->cursor_on) | ||
152 | blend_cfg |= MDP5_CTL_LAYER_REG_CURSOR_OUT; | ||
153 | else | ||
154 | blend_cfg &= ~MDP5_CTL_LAYER_REG_CURSOR_OUT; | ||
155 | |||
156 | spin_lock_irqsave(&ctl->hw_lock, flags); | ||
157 | ctl_write(ctl, REG_MDP5_CTL_LAYER_REG(ctl->id, lm), blend_cfg); | ||
158 | spin_unlock_irqrestore(&ctl->hw_lock, flags); | ||
159 | |||
160 | return 0; | ||
161 | } | ||
162 | |||
163 | int mdp5_ctl_commit(struct mdp5_ctl *ctl, u32 flush_mask) | ||
164 | { | ||
165 | struct mdp5_ctl_manager *ctl_mgr = ctl->ctlm; | ||
166 | unsigned long flags; | ||
167 | |||
168 | if (flush_mask & MDP5_CTL_FLUSH_CURSOR_DUMMY) { | ||
169 | int lm = mdp5_crtc_get_lm(ctl->crtc); | ||
170 | |||
171 | if (unlikely(WARN_ON(lm < 0))) { | ||
172 | dev_err(ctl_mgr->dev->dev, "CTL %d cannot find LM: %d", | ||
173 | ctl->id, lm); | ||
174 | return -EINVAL; | ||
175 | } | ||
176 | |||
177 | /* for current targets, cursor bit is the same as LM bit */ | ||
178 | flush_mask |= mdp_ctl_flush_mask_lm(lm); | ||
179 | } | ||
180 | |||
181 | spin_lock_irqsave(&ctl->hw_lock, flags); | ||
182 | ctl_write(ctl, REG_MDP5_CTL_FLUSH(ctl->id), flush_mask); | ||
183 | spin_unlock_irqrestore(&ctl->hw_lock, flags); | ||
184 | |||
185 | return 0; | ||
186 | } | ||
187 | |||
188 | u32 mdp5_ctl_get_flush(struct mdp5_ctl *ctl) | ||
189 | { | ||
190 | return ctl->flush_mask; | ||
191 | } | ||
192 | |||
193 | void mdp5_ctl_release(struct mdp5_ctl *ctl) | ||
194 | { | ||
195 | struct mdp5_ctl_manager *ctl_mgr = ctl->ctlm; | ||
196 | unsigned long flags; | ||
197 | |||
198 | if (unlikely(WARN_ON(ctl->id >= MAX_CTL) || !ctl->busy)) { | ||
199 | dev_err(ctl_mgr->dev->dev, "CTL %d in bad state (%d)", | ||
200 | ctl->id, ctl->busy); | ||
201 | return; | ||
202 | } | ||
203 | |||
204 | spin_lock_irqsave(&ctl_mgr->pool_lock, flags); | ||
205 | ctl->busy = false; | ||
206 | spin_unlock_irqrestore(&ctl_mgr->pool_lock, flags); | ||
207 | |||
208 | DBG("CTL %d released", ctl->id); | ||
209 | } | ||
210 | |||
211 | /* | ||
212 | * mdp5_ctl_request() - CTL dynamic allocation | ||
213 | * | ||
214 | * Note: Current implementation considers that we can only have one CRTC per CTL | ||
215 | * | ||
216 | * @return first free CTL | ||
217 | */ | ||
218 | struct mdp5_ctl *mdp5_ctlm_request(struct mdp5_ctl_manager *ctl_mgr, | ||
219 | struct drm_crtc *crtc) | ||
220 | { | ||
221 | struct mdp5_ctl *ctl = NULL; | ||
222 | unsigned long flags; | ||
223 | int c; | ||
224 | |||
225 | spin_lock_irqsave(&ctl_mgr->pool_lock, flags); | ||
226 | |||
227 | for (c = 0; c < ctl_mgr->nctl; c++) | ||
228 | if (!ctl_mgr->ctls[c].busy) | ||
229 | break; | ||
230 | |||
231 | if (unlikely(c >= ctl_mgr->nctl)) { | ||
232 | dev_err(ctl_mgr->dev->dev, "No more CTL available!"); | ||
233 | goto unlock; | ||
234 | } | ||
235 | |||
236 | ctl = &ctl_mgr->ctls[c]; | ||
237 | |||
238 | ctl->crtc = crtc; | ||
239 | ctl->busy = true; | ||
240 | DBG("CTL %d allocated", ctl->id); | ||
241 | |||
242 | unlock: | ||
243 | spin_unlock_irqrestore(&ctl_mgr->pool_lock, flags); | ||
244 | return ctl; | ||
245 | } | ||
246 | |||
247 | void mdp5_ctlm_hw_reset(struct mdp5_ctl_manager *ctl_mgr) | ||
248 | { | ||
249 | unsigned long flags; | ||
250 | int c; | ||
251 | |||
252 | for (c = 0; c < ctl_mgr->nctl; c++) { | ||
253 | struct mdp5_ctl *ctl = &ctl_mgr->ctls[c]; | ||
254 | |||
255 | spin_lock_irqsave(&ctl->hw_lock, flags); | ||
256 | ctl_write(ctl, REG_MDP5_CTL_OP(ctl->id), 0); | ||
257 | spin_unlock_irqrestore(&ctl->hw_lock, flags); | ||
258 | } | ||
259 | } | ||
260 | |||
261 | void mdp5_ctlm_destroy(struct mdp5_ctl_manager *ctl_mgr) | ||
262 | { | ||
263 | kfree(ctl_mgr); | ||
264 | } | ||
265 | |||
266 | struct mdp5_ctl_manager *mdp5_ctlm_init(struct drm_device *dev, | ||
267 | void __iomem *mmio_base, const struct mdp5_cfg_hw *hw_cfg) | ||
268 | { | ||
269 | struct mdp5_ctl_manager *ctl_mgr; | ||
270 | const struct mdp5_sub_block *ctl_cfg = &hw_cfg->ctl; | ||
271 | unsigned long flags; | ||
272 | int c, ret; | ||
273 | |||
274 | ctl_mgr = kzalloc(sizeof(*ctl_mgr), GFP_KERNEL); | ||
275 | if (!ctl_mgr) { | ||
276 | dev_err(dev->dev, "failed to allocate CTL manager\n"); | ||
277 | ret = -ENOMEM; | ||
278 | goto fail; | ||
279 | } | ||
280 | |||
281 | if (unlikely(WARN_ON(ctl_cfg->count > MAX_CTL))) { | ||
282 | dev_err(dev->dev, "Increase static pool size to at least %d\n", | ||
283 | ctl_cfg->count); | ||
284 | ret = -ENOSPC; | ||
285 | goto fail; | ||
286 | } | ||
287 | |||
288 | /* initialize the CTL manager: */ | ||
289 | ctl_mgr->dev = dev; | ||
290 | ctl_mgr->nlm = hw_cfg->lm.count; | ||
291 | ctl_mgr->nctl = ctl_cfg->count; | ||
292 | spin_lock_init(&ctl_mgr->pool_lock); | ||
293 | |||
294 | /* initialize each CTL of the pool: */ | ||
295 | spin_lock_irqsave(&ctl_mgr->pool_lock, flags); | ||
296 | for (c = 0; c < ctl_mgr->nctl; c++) { | ||
297 | struct mdp5_ctl *ctl = &ctl_mgr->ctls[c]; | ||
298 | |||
299 | if (WARN_ON(!ctl_cfg->base[c])) { | ||
300 | dev_err(dev->dev, "CTL_%d: base is null!\n", c); | ||
301 | ret = -EINVAL; | ||
302 | goto fail; | ||
303 | } | ||
304 | ctl->ctlm = ctl_mgr; | ||
305 | ctl->id = c; | ||
306 | ctl->mode = MODE_NONE; | ||
307 | ctl->reg_offset = ctl_cfg->base[c]; | ||
308 | ctl->flush_mask = MDP5_CTL_FLUSH_CTL; | ||
309 | ctl->busy = false; | ||
310 | spin_lock_init(&ctl->hw_lock); | ||
311 | } | ||
312 | spin_unlock_irqrestore(&ctl_mgr->pool_lock, flags); | ||
313 | DBG("Pool of %d CTLs created.", ctl_mgr->nctl); | ||
314 | |||
315 | return ctl_mgr; | ||
316 | |||
317 | fail: | ||
318 | if (ctl_mgr) | ||
319 | mdp5_ctlm_destroy(ctl_mgr); | ||
320 | |||
321 | return ERR_PTR(ret); | ||
322 | } | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.h b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.h new file mode 100644 index 000000000000..1018519b6af2 --- /dev/null +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_ctl.h | |||
@@ -0,0 +1,122 @@ | |||
1 | /* | ||
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
3 | * | ||
4 | * This program is free software; you can redistribute it and/or modify | ||
5 | * it under the terms of the GNU General Public License version 2 and | ||
6 | * only version 2 as published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope that it will be useful, | ||
9 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
11 | * GNU General Public License for more details. | ||
12 | */ | ||
13 | |||
14 | #ifndef __MDP5_CTL_H__ | ||
15 | #define __MDP5_CTL_H__ | ||
16 | |||
17 | #include "msm_drv.h" | ||
18 | |||
19 | /* | ||
20 | * CTL Manager prototypes: | ||
21 | * mdp5_ctlm_init() returns a ctlm (CTL Manager) handler, | ||
22 | * which is then used to call the other mdp5_ctlm_*(ctlm, ...) functions. | ||
23 | */ | ||
24 | struct mdp5_ctl_manager; | ||
25 | struct mdp5_ctl_manager *mdp5_ctlm_init(struct drm_device *dev, | ||
26 | void __iomem *mmio_base, const struct mdp5_cfg_hw *hw_cfg); | ||
27 | void mdp5_ctlm_hw_reset(struct mdp5_ctl_manager *ctlm); | ||
28 | void mdp5_ctlm_destroy(struct mdp5_ctl_manager *ctlm); | ||
29 | |||
30 | /* | ||
31 | * CTL prototypes: | ||
32 | * mdp5_ctl_request(ctlm, ...) returns a ctl (CTL resource) handler, | ||
33 | * which is then used to call the other mdp5_ctl_*(ctl, ...) functions. | ||
34 | */ | ||
35 | struct mdp5_ctl *mdp5_ctlm_request(struct mdp5_ctl_manager *ctlm, struct drm_crtc *crtc); | ||
36 | |||
37 | int mdp5_ctl_set_intf(struct mdp5_ctl *ctl, enum mdp5_intf intf); | ||
38 | |||
39 | int mdp5_ctl_set_cursor(struct mdp5_ctl *ctl, bool enable); | ||
40 | |||
41 | /* @blend_cfg: see LM blender config definition below */ | ||
42 | int mdp5_ctl_blend(struct mdp5_ctl *ctl, u32 lm, u32 blend_cfg); | ||
43 | |||
44 | /* @flush_mask: see CTL flush masks definitions below */ | ||
45 | int mdp5_ctl_commit(struct mdp5_ctl *ctl, u32 flush_mask); | ||
46 | u32 mdp5_ctl_get_flush(struct mdp5_ctl *ctl); | ||
47 | |||
48 | void mdp5_ctl_release(struct mdp5_ctl *ctl); | ||
49 | |||
50 | /* | ||
51 | * blend_cfg (LM blender config): | ||
52 | * | ||
53 | * The function below allows the caller of mdp5_ctl_blend() to specify how pipes | ||
54 | * are being blended according to their stage (z-order), through @blend_cfg arg. | ||
55 | */ | ||
56 | static inline u32 mdp_ctl_blend_mask(enum mdp5_pipe pipe, | ||
57 | enum mdp_mixer_stage_id stage) | ||
58 | { | ||
59 | switch (pipe) { | ||
60 | case SSPP_VIG0: return MDP5_CTL_LAYER_REG_VIG0(stage); | ||
61 | case SSPP_VIG1: return MDP5_CTL_LAYER_REG_VIG1(stage); | ||
62 | case SSPP_VIG2: return MDP5_CTL_LAYER_REG_VIG2(stage); | ||
63 | case SSPP_RGB0: return MDP5_CTL_LAYER_REG_RGB0(stage); | ||
64 | case SSPP_RGB1: return MDP5_CTL_LAYER_REG_RGB1(stage); | ||
65 | case SSPP_RGB2: return MDP5_CTL_LAYER_REG_RGB2(stage); | ||
66 | case SSPP_DMA0: return MDP5_CTL_LAYER_REG_DMA0(stage); | ||
67 | case SSPP_DMA1: return MDP5_CTL_LAYER_REG_DMA1(stage); | ||
68 | case SSPP_VIG3: return MDP5_CTL_LAYER_REG_VIG3(stage); | ||
69 | case SSPP_RGB3: return MDP5_CTL_LAYER_REG_RGB3(stage); | ||
70 | default: return 0; | ||
71 | } | ||
72 | } | ||
73 | |||
74 | /* | ||
75 | * flush_mask (CTL flush masks): | ||
76 | * | ||
77 | * The following functions allow each DRM entity to get and store | ||
78 | * their own flush mask. | ||
79 | * Once stored, these masks will then be accessed through each DRM's | ||
80 | * interface and used by the caller of mdp5_ctl_commit() to specify | ||
81 | * which block(s) need to be flushed through @flush_mask parameter. | ||
82 | */ | ||
83 | |||
84 | #define MDP5_CTL_FLUSH_CURSOR_DUMMY 0x80000000 | ||
85 | |||
86 | static inline u32 mdp_ctl_flush_mask_cursor(int cursor_id) | ||
87 | { | ||
88 | /* TODO: use id once multiple cursor support is present */ | ||
89 | (void)cursor_id; | ||
90 | |||
91 | return MDP5_CTL_FLUSH_CURSOR_DUMMY; | ||
92 | } | ||
93 | |||
94 | static inline u32 mdp_ctl_flush_mask_lm(int lm) | ||
95 | { | ||
96 | switch (lm) { | ||
97 | case 0: return MDP5_CTL_FLUSH_LM0; | ||
98 | case 1: return MDP5_CTL_FLUSH_LM1; | ||
99 | case 2: return MDP5_CTL_FLUSH_LM2; | ||
100 | case 5: return MDP5_CTL_FLUSH_LM5; | ||
101 | default: return 0; | ||
102 | } | ||
103 | } | ||
104 | |||
105 | static inline u32 mdp_ctl_flush_mask_pipe(enum mdp5_pipe pipe) | ||
106 | { | ||
107 | switch (pipe) { | ||
108 | case SSPP_VIG0: return MDP5_CTL_FLUSH_VIG0; | ||
109 | case SSPP_VIG1: return MDP5_CTL_FLUSH_VIG1; | ||
110 | case SSPP_VIG2: return MDP5_CTL_FLUSH_VIG2; | ||
111 | case SSPP_RGB0: return MDP5_CTL_FLUSH_RGB0; | ||
112 | case SSPP_RGB1: return MDP5_CTL_FLUSH_RGB1; | ||
113 | case SSPP_RGB2: return MDP5_CTL_FLUSH_RGB2; | ||
114 | case SSPP_DMA0: return MDP5_CTL_FLUSH_DMA0; | ||
115 | case SSPP_DMA1: return MDP5_CTL_FLUSH_DMA1; | ||
116 | case SSPP_VIG3: return MDP5_CTL_FLUSH_VIG3; | ||
117 | case SSPP_RGB3: return MDP5_CTL_FLUSH_RGB3; | ||
118 | default: return 0; | ||
119 | } | ||
120 | } | ||
121 | |||
122 | #endif /* __MDP5_CTL_H__ */ | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_encoder.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_encoder.c index edec7bfaa952..0254bfdeb92f 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_encoder.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_encoder.c | |||
@@ -24,6 +24,7 @@ struct mdp5_encoder { | |||
24 | struct drm_encoder base; | 24 | struct drm_encoder base; |
25 | int intf; | 25 | int intf; |
26 | enum mdp5_intf intf_id; | 26 | enum mdp5_intf intf_id; |
27 | spinlock_t intf_lock; /* protect REG_MDP5_INTF_* registers */ | ||
27 | bool enabled; | 28 | bool enabled; |
28 | uint32_t bsc; | 29 | uint32_t bsc; |
29 | }; | 30 | }; |
@@ -115,6 +116,7 @@ static void mdp5_encoder_dpms(struct drm_encoder *encoder, int mode) | |||
115 | struct mdp5_kms *mdp5_kms = get_kms(encoder); | 116 | struct mdp5_kms *mdp5_kms = get_kms(encoder); |
116 | int intf = mdp5_encoder->intf; | 117 | int intf = mdp5_encoder->intf; |
117 | bool enabled = (mode == DRM_MODE_DPMS_ON); | 118 | bool enabled = (mode == DRM_MODE_DPMS_ON); |
119 | unsigned long flags; | ||
118 | 120 | ||
119 | DBG("mode=%d", mode); | 121 | DBG("mode=%d", mode); |
120 | 122 | ||
@@ -123,9 +125,24 @@ static void mdp5_encoder_dpms(struct drm_encoder *encoder, int mode) | |||
123 | 125 | ||
124 | if (enabled) { | 126 | if (enabled) { |
125 | bs_set(mdp5_encoder, 1); | 127 | bs_set(mdp5_encoder, 1); |
128 | spin_lock_irqsave(&mdp5_encoder->intf_lock, flags); | ||
126 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(intf), 1); | 129 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(intf), 1); |
130 | spin_unlock_irqrestore(&mdp5_encoder->intf_lock, flags); | ||
127 | } else { | 131 | } else { |
132 | spin_lock_irqsave(&mdp5_encoder->intf_lock, flags); | ||
128 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(intf), 0); | 133 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(intf), 0); |
134 | spin_unlock_irqrestore(&mdp5_encoder->intf_lock, flags); | ||
135 | |||
136 | /* | ||
137 | * Wait for a vsync so we know the ENABLE=0 latched before | ||
138 | * the (connector) source of the vsync's gets disabled, | ||
139 | * otherwise we end up in a funny state if we re-enable | ||
140 | * before the disable latches, which results that some of | ||
141 | * the settings changes for the new modeset (like new | ||
142 | * scanout buffer) don't latch properly.. | ||
143 | */ | ||
144 | mdp_irq_wait(&mdp5_kms->base, intf2vblank(intf)); | ||
145 | |||
129 | bs_set(mdp5_encoder, 0); | 146 | bs_set(mdp5_encoder, 0); |
130 | } | 147 | } |
131 | 148 | ||
@@ -150,6 +167,7 @@ static void mdp5_encoder_mode_set(struct drm_encoder *encoder, | |||
150 | uint32_t display_v_start, display_v_end; | 167 | uint32_t display_v_start, display_v_end; |
151 | uint32_t hsync_start_x, hsync_end_x; | 168 | uint32_t hsync_start_x, hsync_end_x; |
152 | uint32_t format; | 169 | uint32_t format; |
170 | unsigned long flags; | ||
153 | 171 | ||
154 | mode = adjusted_mode; | 172 | mode = adjusted_mode; |
155 | 173 | ||
@@ -180,6 +198,8 @@ static void mdp5_encoder_mode_set(struct drm_encoder *encoder, | |||
180 | display_v_start = (mode->vtotal - mode->vsync_start) * mode->htotal + dtv_hsync_skew; | 198 | display_v_start = (mode->vtotal - mode->vsync_start) * mode->htotal + dtv_hsync_skew; |
181 | display_v_end = vsync_period - ((mode->vsync_start - mode->vdisplay) * mode->htotal) + dtv_hsync_skew - 1; | 199 | display_v_end = vsync_period - ((mode->vsync_start - mode->vdisplay) * mode->htotal) + dtv_hsync_skew - 1; |
182 | 200 | ||
201 | spin_lock_irqsave(&mdp5_encoder->intf_lock, flags); | ||
202 | |||
183 | mdp5_write(mdp5_kms, REG_MDP5_INTF_HSYNC_CTL(intf), | 203 | mdp5_write(mdp5_kms, REG_MDP5_INTF_HSYNC_CTL(intf), |
184 | MDP5_INTF_HSYNC_CTL_PULSEW(mode->hsync_end - mode->hsync_start) | | 204 | MDP5_INTF_HSYNC_CTL_PULSEW(mode->hsync_end - mode->hsync_start) | |
185 | MDP5_INTF_HSYNC_CTL_PERIOD(mode->htotal)); | 205 | MDP5_INTF_HSYNC_CTL_PERIOD(mode->htotal)); |
@@ -201,6 +221,8 @@ static void mdp5_encoder_mode_set(struct drm_encoder *encoder, | |||
201 | mdp5_write(mdp5_kms, REG_MDP5_INTF_ACTIVE_VEND_F0(intf), 0); | 221 | mdp5_write(mdp5_kms, REG_MDP5_INTF_ACTIVE_VEND_F0(intf), 0); |
202 | mdp5_write(mdp5_kms, REG_MDP5_INTF_PANEL_FORMAT(intf), format); | 222 | mdp5_write(mdp5_kms, REG_MDP5_INTF_PANEL_FORMAT(intf), format); |
203 | mdp5_write(mdp5_kms, REG_MDP5_INTF_FRAME_LINE_COUNT_EN(intf), 0x3); /* frame+line? */ | 223 | mdp5_write(mdp5_kms, REG_MDP5_INTF_FRAME_LINE_COUNT_EN(intf), 0x3); /* frame+line? */ |
224 | |||
225 | spin_unlock_irqrestore(&mdp5_encoder->intf_lock, flags); | ||
204 | } | 226 | } |
205 | 227 | ||
206 | static void mdp5_encoder_prepare(struct drm_encoder *encoder) | 228 | static void mdp5_encoder_prepare(struct drm_encoder *encoder) |
@@ -242,6 +264,8 @@ struct drm_encoder *mdp5_encoder_init(struct drm_device *dev, int intf, | |||
242 | mdp5_encoder->intf_id = intf_id; | 264 | mdp5_encoder->intf_id = intf_id; |
243 | encoder = &mdp5_encoder->base; | 265 | encoder = &mdp5_encoder->base; |
244 | 266 | ||
267 | spin_lock_init(&mdp5_encoder->intf_lock); | ||
268 | |||
245 | drm_encoder_init(dev, encoder, &mdp5_encoder_funcs, | 269 | drm_encoder_init(dev, encoder, &mdp5_encoder_funcs, |
246 | DRM_MODE_ENCODER_TMDS); | 270 | DRM_MODE_ENCODER_TMDS); |
247 | drm_encoder_helper_add(encoder, &mdp5_encoder_helper_funcs); | 271 | drm_encoder_helper_add(encoder, &mdp5_encoder_helper_funcs); |
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_irq.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_irq.c index f2b985bc2adf..70ac81edd40f 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_irq.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_irq.c | |||
@@ -15,6 +15,8 @@ | |||
15 | * this program. If not, see <http://www.gnu.org/licenses/>. | 15 | * this program. If not, see <http://www.gnu.org/licenses/>. |
16 | */ | 16 | */ |
17 | 17 | ||
18 | #include <linux/irqdomain.h> | ||
19 | #include <linux/irq.h> | ||
18 | 20 | ||
19 | #include "msm_drv.h" | 21 | #include "msm_drv.h" |
20 | #include "mdp5_kms.h" | 22 | #include "mdp5_kms.h" |
@@ -88,11 +90,17 @@ irqreturn_t mdp5_irq(struct msm_kms *kms) | |||
88 | 90 | ||
89 | VERB("intr=%08x", intr); | 91 | VERB("intr=%08x", intr); |
90 | 92 | ||
91 | if (intr & MDP5_HW_INTR_STATUS_INTR_MDP) | 93 | if (intr & MDP5_HW_INTR_STATUS_INTR_MDP) { |
92 | mdp5_irq_mdp(mdp_kms); | 94 | mdp5_irq_mdp(mdp_kms); |
95 | intr &= ~MDP5_HW_INTR_STATUS_INTR_MDP; | ||
96 | } | ||
93 | 97 | ||
94 | if (intr & MDP5_HW_INTR_STATUS_INTR_HDMI) | 98 | while (intr) { |
95 | hdmi_irq(0, mdp5_kms->hdmi); | 99 | irq_hw_number_t hwirq = fls(intr) - 1; |
100 | generic_handle_irq(irq_find_mapping( | ||
101 | mdp5_kms->irqcontroller.domain, hwirq)); | ||
102 | intr &= ~(1 << hwirq); | ||
103 | } | ||
96 | 104 | ||
97 | return IRQ_HANDLED; | 105 | return IRQ_HANDLED; |
98 | } | 106 | } |
@@ -109,3 +117,82 @@ void mdp5_disable_vblank(struct msm_kms *kms, struct drm_crtc *crtc) | |||
109 | mdp_update_vblank_mask(to_mdp_kms(kms), | 117 | mdp_update_vblank_mask(to_mdp_kms(kms), |
110 | mdp5_crtc_vblank(crtc), false); | 118 | mdp5_crtc_vblank(crtc), false); |
111 | } | 119 | } |
120 | |||
121 | /* | ||
122 | * interrupt-controller implementation, so sub-blocks (hdmi/eDP/dsi/etc) | ||
123 | * can register to get their irq's delivered | ||
124 | */ | ||
125 | |||
126 | #define VALID_IRQS (MDP5_HW_INTR_STATUS_INTR_DSI0 | \ | ||
127 | MDP5_HW_INTR_STATUS_INTR_DSI1 | \ | ||
128 | MDP5_HW_INTR_STATUS_INTR_HDMI | \ | ||
129 | MDP5_HW_INTR_STATUS_INTR_EDP) | ||
130 | |||
131 | static void mdp5_hw_mask_irq(struct irq_data *irqd) | ||
132 | { | ||
133 | struct mdp5_kms *mdp5_kms = irq_data_get_irq_chip_data(irqd); | ||
134 | smp_mb__before_atomic(); | ||
135 | clear_bit(irqd->hwirq, &mdp5_kms->irqcontroller.enabled_mask); | ||
136 | smp_mb__after_atomic(); | ||
137 | } | ||
138 | |||
139 | static void mdp5_hw_unmask_irq(struct irq_data *irqd) | ||
140 | { | ||
141 | struct mdp5_kms *mdp5_kms = irq_data_get_irq_chip_data(irqd); | ||
142 | smp_mb__before_atomic(); | ||
143 | set_bit(irqd->hwirq, &mdp5_kms->irqcontroller.enabled_mask); | ||
144 | smp_mb__after_atomic(); | ||
145 | } | ||
146 | |||
147 | static struct irq_chip mdp5_hw_irq_chip = { | ||
148 | .name = "mdp5", | ||
149 | .irq_mask = mdp5_hw_mask_irq, | ||
150 | .irq_unmask = mdp5_hw_unmask_irq, | ||
151 | }; | ||
152 | |||
153 | static int mdp5_hw_irqdomain_map(struct irq_domain *d, | ||
154 | unsigned int irq, irq_hw_number_t hwirq) | ||
155 | { | ||
156 | struct mdp5_kms *mdp5_kms = d->host_data; | ||
157 | |||
158 | if (!(VALID_IRQS & (1 << hwirq))) | ||
159 | return -EPERM; | ||
160 | |||
161 | irq_set_chip_and_handler(irq, &mdp5_hw_irq_chip, handle_level_irq); | ||
162 | irq_set_chip_data(irq, mdp5_kms); | ||
163 | set_irq_flags(irq, IRQF_VALID); | ||
164 | |||
165 | return 0; | ||
166 | } | ||
167 | |||
168 | static struct irq_domain_ops mdp5_hw_irqdomain_ops = { | ||
169 | .map = mdp5_hw_irqdomain_map, | ||
170 | .xlate = irq_domain_xlate_onecell, | ||
171 | }; | ||
172 | |||
173 | |||
174 | int mdp5_irq_domain_init(struct mdp5_kms *mdp5_kms) | ||
175 | { | ||
176 | struct device *dev = mdp5_kms->dev->dev; | ||
177 | struct irq_domain *d; | ||
178 | |||
179 | d = irq_domain_add_linear(dev->of_node, 32, | ||
180 | &mdp5_hw_irqdomain_ops, mdp5_kms); | ||
181 | if (!d) { | ||
182 | dev_err(dev, "mdp5 irq domain add failed\n"); | ||
183 | return -ENXIO; | ||
184 | } | ||
185 | |||
186 | mdp5_kms->irqcontroller.enabled_mask = 0; | ||
187 | mdp5_kms->irqcontroller.domain = d; | ||
188 | |||
189 | return 0; | ||
190 | } | ||
191 | |||
192 | void mdp5_irq_domain_fini(struct mdp5_kms *mdp5_kms) | ||
193 | { | ||
194 | if (mdp5_kms->irqcontroller.domain) { | ||
195 | irq_domain_remove(mdp5_kms->irqcontroller.domain); | ||
196 | mdp5_kms->irqcontroller.domain = NULL; | ||
197 | } | ||
198 | } | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.c index 31a2c6331a1d..a11f1b80c488 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.c | |||
@@ -1,4 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (c) 2014, The Linux Foundation. All rights reserved. | ||
2 | * Copyright (C) 2013 Red Hat | 3 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 4 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 5 | * |
@@ -24,145 +25,11 @@ static const char *iommu_ports[] = { | |||
24 | "mdp_0", | 25 | "mdp_0", |
25 | }; | 26 | }; |
26 | 27 | ||
27 | static struct mdp5_platform_config *mdp5_get_config(struct platform_device *dev); | ||
28 | |||
29 | const struct mdp5_config *mdp5_cfg; | ||
30 | |||
31 | static const struct mdp5_config msm8x74_config = { | ||
32 | .name = "msm8x74", | ||
33 | .ctl = { | ||
34 | .count = 5, | ||
35 | .base = { 0x00600, 0x00700, 0x00800, 0x00900, 0x00a00 }, | ||
36 | }, | ||
37 | .pipe_vig = { | ||
38 | .count = 3, | ||
39 | .base = { 0x01200, 0x01600, 0x01a00 }, | ||
40 | }, | ||
41 | .pipe_rgb = { | ||
42 | .count = 3, | ||
43 | .base = { 0x01e00, 0x02200, 0x02600 }, | ||
44 | }, | ||
45 | .pipe_dma = { | ||
46 | .count = 2, | ||
47 | .base = { 0x02a00, 0x02e00 }, | ||
48 | }, | ||
49 | .lm = { | ||
50 | .count = 5, | ||
51 | .base = { 0x03200, 0x03600, 0x03a00, 0x03e00, 0x04200 }, | ||
52 | }, | ||
53 | .dspp = { | ||
54 | .count = 3, | ||
55 | .base = { 0x04600, 0x04a00, 0x04e00 }, | ||
56 | }, | ||
57 | .ad = { | ||
58 | .count = 2, | ||
59 | .base = { 0x13100, 0x13300 }, /* NOTE: no ad in v1.0 */ | ||
60 | }, | ||
61 | .intf = { | ||
62 | .count = 4, | ||
63 | .base = { 0x12500, 0x12700, 0x12900, 0x12b00 }, | ||
64 | }, | ||
65 | }; | ||
66 | |||
67 | static const struct mdp5_config apq8084_config = { | ||
68 | .name = "apq8084", | ||
69 | .ctl = { | ||
70 | .count = 5, | ||
71 | .base = { 0x00600, 0x00700, 0x00800, 0x00900, 0x00a00 }, | ||
72 | }, | ||
73 | .pipe_vig = { | ||
74 | .count = 4, | ||
75 | .base = { 0x01200, 0x01600, 0x01a00, 0x01e00 }, | ||
76 | }, | ||
77 | .pipe_rgb = { | ||
78 | .count = 4, | ||
79 | .base = { 0x02200, 0x02600, 0x02a00, 0x02e00 }, | ||
80 | }, | ||
81 | .pipe_dma = { | ||
82 | .count = 2, | ||
83 | .base = { 0x03200, 0x03600 }, | ||
84 | }, | ||
85 | .lm = { | ||
86 | .count = 6, | ||
87 | .base = { 0x03a00, 0x03e00, 0x04200, 0x04600, 0x04a00, 0x04e00 }, | ||
88 | }, | ||
89 | .dspp = { | ||
90 | .count = 4, | ||
91 | .base = { 0x05200, 0x05600, 0x05a00, 0x05e00 }, | ||
92 | |||
93 | }, | ||
94 | .ad = { | ||
95 | .count = 3, | ||
96 | .base = { 0x13500, 0x13700, 0x13900 }, | ||
97 | }, | ||
98 | .intf = { | ||
99 | .count = 5, | ||
100 | .base = { 0x12500, 0x12700, 0x12900, 0x12b00, 0x12d00 }, | ||
101 | }, | ||
102 | }; | ||
103 | |||
104 | struct mdp5_config_entry { | ||
105 | int revision; | ||
106 | const struct mdp5_config *config; | ||
107 | }; | ||
108 | |||
109 | static const struct mdp5_config_entry mdp5_configs[] = { | ||
110 | { .revision = 0, .config = &msm8x74_config }, | ||
111 | { .revision = 2, .config = &msm8x74_config }, | ||
112 | { .revision = 3, .config = &apq8084_config }, | ||
113 | }; | ||
114 | |||
115 | static int mdp5_select_hw_cfg(struct msm_kms *kms) | ||
116 | { | ||
117 | struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms)); | ||
118 | struct drm_device *dev = mdp5_kms->dev; | ||
119 | uint32_t version, major, minor; | ||
120 | int i, ret = 0; | ||
121 | |||
122 | mdp5_enable(mdp5_kms); | ||
123 | version = mdp5_read(mdp5_kms, REG_MDP5_MDP_VERSION); | ||
124 | mdp5_disable(mdp5_kms); | ||
125 | |||
126 | major = FIELD(version, MDP5_MDP_VERSION_MAJOR); | ||
127 | minor = FIELD(version, MDP5_MDP_VERSION_MINOR); | ||
128 | |||
129 | DBG("found MDP5 version v%d.%d", major, minor); | ||
130 | |||
131 | if (major != 1) { | ||
132 | dev_err(dev->dev, "unexpected MDP major version: v%d.%d\n", | ||
133 | major, minor); | ||
134 | ret = -ENXIO; | ||
135 | goto out; | ||
136 | } | ||
137 | |||
138 | mdp5_kms->rev = minor; | ||
139 | |||
140 | /* only after mdp5_cfg global pointer's init can we access the hw */ | ||
141 | for (i = 0; i < ARRAY_SIZE(mdp5_configs); i++) { | ||
142 | if (mdp5_configs[i].revision != minor) | ||
143 | continue; | ||
144 | mdp5_kms->hw_cfg = mdp5_cfg = mdp5_configs[i].config; | ||
145 | break; | ||
146 | } | ||
147 | if (unlikely(!mdp5_kms->hw_cfg)) { | ||
148 | dev_err(dev->dev, "unexpected MDP minor revision: v%d.%d\n", | ||
149 | major, minor); | ||
150 | ret = -ENXIO; | ||
151 | goto out; | ||
152 | } | ||
153 | |||
154 | DBG("MDP5: %s config selected", mdp5_kms->hw_cfg->name); | ||
155 | |||
156 | return 0; | ||
157 | out: | ||
158 | return ret; | ||
159 | } | ||
160 | |||
161 | static int mdp5_hw_init(struct msm_kms *kms) | 28 | static int mdp5_hw_init(struct msm_kms *kms) |
162 | { | 29 | { |
163 | struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms)); | 30 | struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms)); |
164 | struct drm_device *dev = mdp5_kms->dev; | 31 | struct drm_device *dev = mdp5_kms->dev; |
165 | int i; | 32 | unsigned long flags; |
166 | 33 | ||
167 | pm_runtime_get_sync(dev->dev); | 34 | pm_runtime_get_sync(dev->dev); |
168 | 35 | ||
@@ -190,10 +57,11 @@ static int mdp5_hw_init(struct msm_kms *kms) | |||
190 | * care. | 57 | * care. |
191 | */ | 58 | */ |
192 | 59 | ||
60 | spin_lock_irqsave(&mdp5_kms->resource_lock, flags); | ||
193 | mdp5_write(mdp5_kms, REG_MDP5_DISP_INTF_SEL, 0); | 61 | mdp5_write(mdp5_kms, REG_MDP5_DISP_INTF_SEL, 0); |
62 | spin_unlock_irqrestore(&mdp5_kms->resource_lock, flags); | ||
194 | 63 | ||
195 | for (i = 0; i < mdp5_kms->hw_cfg->ctl.count; i++) | 64 | mdp5_ctlm_hw_reset(mdp5_kms->ctlm); |
196 | mdp5_write(mdp5_kms, REG_MDP5_CTL_OP(i), 0); | ||
197 | 65 | ||
198 | pm_runtime_put_sync(dev->dev); | 66 | pm_runtime_put_sync(dev->dev); |
199 | 67 | ||
@@ -221,10 +89,20 @@ static void mdp5_destroy(struct msm_kms *kms) | |||
221 | struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms)); | 89 | struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms)); |
222 | struct msm_mmu *mmu = mdp5_kms->mmu; | 90 | struct msm_mmu *mmu = mdp5_kms->mmu; |
223 | 91 | ||
92 | mdp5_irq_domain_fini(mdp5_kms); | ||
93 | |||
224 | if (mmu) { | 94 | if (mmu) { |
225 | mmu->funcs->detach(mmu, iommu_ports, ARRAY_SIZE(iommu_ports)); | 95 | mmu->funcs->detach(mmu, iommu_ports, ARRAY_SIZE(iommu_ports)); |
226 | mmu->funcs->destroy(mmu); | 96 | mmu->funcs->destroy(mmu); |
227 | } | 97 | } |
98 | |||
99 | if (mdp5_kms->ctlm) | ||
100 | mdp5_ctlm_destroy(mdp5_kms->ctlm); | ||
101 | if (mdp5_kms->smp) | ||
102 | mdp5_smp_destroy(mdp5_kms->smp); | ||
103 | if (mdp5_kms->cfg) | ||
104 | mdp5_cfg_destroy(mdp5_kms->cfg); | ||
105 | |||
228 | kfree(mdp5_kms); | 106 | kfree(mdp5_kms); |
229 | } | 107 | } |
230 | 108 | ||
@@ -274,17 +152,31 @@ static int modeset_init(struct mdp5_kms *mdp5_kms) | |||
274 | static const enum mdp5_pipe crtcs[] = { | 152 | static const enum mdp5_pipe crtcs[] = { |
275 | SSPP_RGB0, SSPP_RGB1, SSPP_RGB2, SSPP_RGB3, | 153 | SSPP_RGB0, SSPP_RGB1, SSPP_RGB2, SSPP_RGB3, |
276 | }; | 154 | }; |
155 | static const enum mdp5_pipe pub_planes[] = { | ||
156 | SSPP_VIG0, SSPP_VIG1, SSPP_VIG2, SSPP_VIG3, | ||
157 | }; | ||
277 | struct drm_device *dev = mdp5_kms->dev; | 158 | struct drm_device *dev = mdp5_kms->dev; |
278 | struct msm_drm_private *priv = dev->dev_private; | 159 | struct msm_drm_private *priv = dev->dev_private; |
279 | struct drm_encoder *encoder; | 160 | struct drm_encoder *encoder; |
161 | const struct mdp5_cfg_hw *hw_cfg; | ||
280 | int i, ret; | 162 | int i, ret; |
281 | 163 | ||
282 | /* construct CRTCs: */ | 164 | hw_cfg = mdp5_cfg_get_hw_config(mdp5_kms->cfg); |
283 | for (i = 0; i < mdp5_kms->hw_cfg->pipe_rgb.count; i++) { | 165 | |
166 | /* register our interrupt-controller for hdmi/eDP/dsi/etc | ||
167 | * to use for irqs routed through mdp: | ||
168 | */ | ||
169 | ret = mdp5_irq_domain_init(mdp5_kms); | ||
170 | if (ret) | ||
171 | goto fail; | ||
172 | |||
173 | /* construct CRTCs and their private planes: */ | ||
174 | for (i = 0; i < hw_cfg->pipe_rgb.count; i++) { | ||
284 | struct drm_plane *plane; | 175 | struct drm_plane *plane; |
285 | struct drm_crtc *crtc; | 176 | struct drm_crtc *crtc; |
286 | 177 | ||
287 | plane = mdp5_plane_init(dev, crtcs[i], true); | 178 | plane = mdp5_plane_init(dev, crtcs[i], true, |
179 | hw_cfg->pipe_rgb.base[i]); | ||
288 | if (IS_ERR(plane)) { | 180 | if (IS_ERR(plane)) { |
289 | ret = PTR_ERR(plane); | 181 | ret = PTR_ERR(plane); |
290 | dev_err(dev->dev, "failed to construct plane for %s (%d)\n", | 182 | dev_err(dev->dev, "failed to construct plane for %s (%d)\n", |
@@ -302,6 +194,20 @@ static int modeset_init(struct mdp5_kms *mdp5_kms) | |||
302 | priv->crtcs[priv->num_crtcs++] = crtc; | 194 | priv->crtcs[priv->num_crtcs++] = crtc; |
303 | } | 195 | } |
304 | 196 | ||
197 | /* Construct public planes: */ | ||
198 | for (i = 0; i < hw_cfg->pipe_vig.count; i++) { | ||
199 | struct drm_plane *plane; | ||
200 | |||
201 | plane = mdp5_plane_init(dev, pub_planes[i], false, | ||
202 | hw_cfg->pipe_vig.base[i]); | ||
203 | if (IS_ERR(plane)) { | ||
204 | ret = PTR_ERR(plane); | ||
205 | dev_err(dev->dev, "failed to construct %s plane: %d\n", | ||
206 | pipe2name(pub_planes[i]), ret); | ||
207 | goto fail; | ||
208 | } | ||
209 | } | ||
210 | |||
305 | /* Construct encoder for HDMI: */ | 211 | /* Construct encoder for HDMI: */ |
306 | encoder = mdp5_encoder_init(dev, 3, INTF_HDMI); | 212 | encoder = mdp5_encoder_init(dev, 3, INTF_HDMI); |
307 | if (IS_ERR(encoder)) { | 213 | if (IS_ERR(encoder)) { |
@@ -324,11 +230,12 @@ static int modeset_init(struct mdp5_kms *mdp5_kms) | |||
324 | priv->encoders[priv->num_encoders++] = encoder; | 230 | priv->encoders[priv->num_encoders++] = encoder; |
325 | 231 | ||
326 | /* Construct bridge/connector for HDMI: */ | 232 | /* Construct bridge/connector for HDMI: */ |
327 | mdp5_kms->hdmi = hdmi_init(dev, encoder); | 233 | if (priv->hdmi) { |
328 | if (IS_ERR(mdp5_kms->hdmi)) { | 234 | ret = hdmi_modeset_init(priv->hdmi, dev, encoder); |
329 | ret = PTR_ERR(mdp5_kms->hdmi); | 235 | if (ret) { |
330 | dev_err(dev->dev, "failed to initialize HDMI: %d\n", ret); | 236 | dev_err(dev->dev, "failed to initialize HDMI: %d\n", ret); |
331 | goto fail; | 237 | goto fail; |
238 | } | ||
332 | } | 239 | } |
333 | 240 | ||
334 | return 0; | 241 | return 0; |
@@ -337,6 +244,21 @@ fail: | |||
337 | return ret; | 244 | return ret; |
338 | } | 245 | } |
339 | 246 | ||
247 | static void read_hw_revision(struct mdp5_kms *mdp5_kms, | ||
248 | uint32_t *major, uint32_t *minor) | ||
249 | { | ||
250 | uint32_t version; | ||
251 | |||
252 | mdp5_enable(mdp5_kms); | ||
253 | version = mdp5_read(mdp5_kms, REG_MDP5_MDP_VERSION); | ||
254 | mdp5_disable(mdp5_kms); | ||
255 | |||
256 | *major = FIELD(version, MDP5_MDP_VERSION_MAJOR); | ||
257 | *minor = FIELD(version, MDP5_MDP_VERSION_MINOR); | ||
258 | |||
259 | DBG("MDP5 version v%d.%d", *major, *minor); | ||
260 | } | ||
261 | |||
340 | static int get_clk(struct platform_device *pdev, struct clk **clkp, | 262 | static int get_clk(struct platform_device *pdev, struct clk **clkp, |
341 | const char *name) | 263 | const char *name) |
342 | { | 264 | { |
@@ -353,10 +275,11 @@ static int get_clk(struct platform_device *pdev, struct clk **clkp, | |||
353 | struct msm_kms *mdp5_kms_init(struct drm_device *dev) | 275 | struct msm_kms *mdp5_kms_init(struct drm_device *dev) |
354 | { | 276 | { |
355 | struct platform_device *pdev = dev->platformdev; | 277 | struct platform_device *pdev = dev->platformdev; |
356 | struct mdp5_platform_config *config = mdp5_get_config(pdev); | 278 | struct mdp5_cfg *config; |
357 | struct mdp5_kms *mdp5_kms; | 279 | struct mdp5_kms *mdp5_kms; |
358 | struct msm_kms *kms = NULL; | 280 | struct msm_kms *kms = NULL; |
359 | struct msm_mmu *mmu; | 281 | struct msm_mmu *mmu; |
282 | uint32_t major, minor; | ||
360 | int i, ret; | 283 | int i, ret; |
361 | 284 | ||
362 | mdp5_kms = kzalloc(sizeof(*mdp5_kms), GFP_KERNEL); | 285 | mdp5_kms = kzalloc(sizeof(*mdp5_kms), GFP_KERNEL); |
@@ -366,12 +289,13 @@ struct msm_kms *mdp5_kms_init(struct drm_device *dev) | |||
366 | goto fail; | 289 | goto fail; |
367 | } | 290 | } |
368 | 291 | ||
292 | spin_lock_init(&mdp5_kms->resource_lock); | ||
293 | |||
369 | mdp_kms_init(&mdp5_kms->base, &kms_funcs); | 294 | mdp_kms_init(&mdp5_kms->base, &kms_funcs); |
370 | 295 | ||
371 | kms = &mdp5_kms->base.base; | 296 | kms = &mdp5_kms->base.base; |
372 | 297 | ||
373 | mdp5_kms->dev = dev; | 298 | mdp5_kms->dev = dev; |
374 | mdp5_kms->smp_blk_cnt = config->smp_blk_cnt; | ||
375 | 299 | ||
376 | mdp5_kms->mmio = msm_ioremap(pdev, "mdp_phys", "MDP5"); | 300 | mdp5_kms->mmio = msm_ioremap(pdev, "mdp_phys", "MDP5"); |
377 | if (IS_ERR(mdp5_kms->mmio)) { | 301 | if (IS_ERR(mdp5_kms->mmio)) { |
@@ -416,24 +340,52 @@ struct msm_kms *mdp5_kms_init(struct drm_device *dev) | |||
416 | if (ret) | 340 | if (ret) |
417 | goto fail; | 341 | goto fail; |
418 | 342 | ||
419 | ret = clk_set_rate(mdp5_kms->src_clk, config->max_clk); | 343 | /* we need to set a default rate before enabling. Set a safe |
344 | * rate first, then figure out hw revision, and then set a | ||
345 | * more optimal rate: | ||
346 | */ | ||
347 | clk_set_rate(mdp5_kms->src_clk, 200000000); | ||
348 | |||
349 | read_hw_revision(mdp5_kms, &major, &minor); | ||
420 | 350 | ||
421 | ret = mdp5_select_hw_cfg(kms); | 351 | mdp5_kms->cfg = mdp5_cfg_init(mdp5_kms, major, minor); |
422 | if (ret) | 352 | if (IS_ERR(mdp5_kms->cfg)) { |
353 | ret = PTR_ERR(mdp5_kms->cfg); | ||
354 | mdp5_kms->cfg = NULL; | ||
423 | goto fail; | 355 | goto fail; |
356 | } | ||
357 | |||
358 | config = mdp5_cfg_get_config(mdp5_kms->cfg); | ||
359 | |||
360 | /* TODO: compute core clock rate at runtime */ | ||
361 | clk_set_rate(mdp5_kms->src_clk, config->hw->max_clk); | ||
362 | |||
363 | mdp5_kms->smp = mdp5_smp_init(mdp5_kms->dev, &config->hw->smp); | ||
364 | if (IS_ERR(mdp5_kms->smp)) { | ||
365 | ret = PTR_ERR(mdp5_kms->smp); | ||
366 | mdp5_kms->smp = NULL; | ||
367 | goto fail; | ||
368 | } | ||
369 | |||
370 | mdp5_kms->ctlm = mdp5_ctlm_init(dev, mdp5_kms->mmio, config->hw); | ||
371 | if (IS_ERR(mdp5_kms->ctlm)) { | ||
372 | ret = PTR_ERR(mdp5_kms->ctlm); | ||
373 | mdp5_kms->ctlm = NULL; | ||
374 | goto fail; | ||
375 | } | ||
424 | 376 | ||
425 | /* make sure things are off before attaching iommu (bootloader could | 377 | /* make sure things are off before attaching iommu (bootloader could |
426 | * have left things on, in which case we'll start getting faults if | 378 | * have left things on, in which case we'll start getting faults if |
427 | * we don't disable): | 379 | * we don't disable): |
428 | */ | 380 | */ |
429 | mdp5_enable(mdp5_kms); | 381 | mdp5_enable(mdp5_kms); |
430 | for (i = 0; i < mdp5_kms->hw_cfg->intf.count; i++) | 382 | for (i = 0; i < config->hw->intf.count; i++) |
431 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(i), 0); | 383 | mdp5_write(mdp5_kms, REG_MDP5_INTF_TIMING_ENGINE_EN(i), 0); |
432 | mdp5_disable(mdp5_kms); | 384 | mdp5_disable(mdp5_kms); |
433 | mdelay(16); | 385 | mdelay(16); |
434 | 386 | ||
435 | if (config->iommu) { | 387 | if (config->platform.iommu) { |
436 | mmu = msm_iommu_new(&pdev->dev, config->iommu); | 388 | mmu = msm_iommu_new(&pdev->dev, config->platform.iommu); |
437 | if (IS_ERR(mmu)) { | 389 | if (IS_ERR(mmu)) { |
438 | ret = PTR_ERR(mmu); | 390 | ret = PTR_ERR(mmu); |
439 | dev_err(dev->dev, "failed to init iommu: %d\n", ret); | 391 | dev_err(dev->dev, "failed to init iommu: %d\n", ret); |
@@ -474,18 +426,3 @@ fail: | |||
474 | mdp5_destroy(kms); | 426 | mdp5_destroy(kms); |
475 | return ERR_PTR(ret); | 427 | return ERR_PTR(ret); |
476 | } | 428 | } |
477 | |||
478 | static struct mdp5_platform_config *mdp5_get_config(struct platform_device *dev) | ||
479 | { | ||
480 | static struct mdp5_platform_config config = {}; | ||
481 | #ifdef CONFIG_OF | ||
482 | /* TODO */ | ||
483 | #endif | ||
484 | config.iommu = iommu_domain_alloc(&platform_bus_type); | ||
485 | /* TODO hard-coded in downstream mdss, but should it be? */ | ||
486 | config.max_clk = 200000000; | ||
487 | /* TODO get from DT: */ | ||
488 | config.smp_blk_cnt = 22; | ||
489 | |||
490 | return &config; | ||
491 | } | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.h b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.h index 5bf340dd0f00..dd69c77c0d64 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.h +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_kms.h | |||
@@ -21,25 +21,9 @@ | |||
21 | #include "msm_drv.h" | 21 | #include "msm_drv.h" |
22 | #include "msm_kms.h" | 22 | #include "msm_kms.h" |
23 | #include "mdp/mdp_kms.h" | 23 | #include "mdp/mdp_kms.h" |
24 | /* dynamic offsets used by mdp5.xml.h (initialized in mdp5_kms.c) */ | 24 | #include "mdp5_cfg.h" /* must be included before mdp5.xml.h */ |
25 | #define MDP5_MAX_BASES 8 | ||
26 | struct mdp5_sub_block { | ||
27 | int count; | ||
28 | uint32_t base[MDP5_MAX_BASES]; | ||
29 | }; | ||
30 | struct mdp5_config { | ||
31 | char *name; | ||
32 | struct mdp5_sub_block ctl; | ||
33 | struct mdp5_sub_block pipe_vig; | ||
34 | struct mdp5_sub_block pipe_rgb; | ||
35 | struct mdp5_sub_block pipe_dma; | ||
36 | struct mdp5_sub_block lm; | ||
37 | struct mdp5_sub_block dspp; | ||
38 | struct mdp5_sub_block ad; | ||
39 | struct mdp5_sub_block intf; | ||
40 | }; | ||
41 | extern const struct mdp5_config *mdp5_cfg; | ||
42 | #include "mdp5.xml.h" | 25 | #include "mdp5.xml.h" |
26 | #include "mdp5_ctl.h" | ||
43 | #include "mdp5_smp.h" | 27 | #include "mdp5_smp.h" |
44 | 28 | ||
45 | struct mdp5_kms { | 29 | struct mdp5_kms { |
@@ -47,17 +31,14 @@ struct mdp5_kms { | |||
47 | 31 | ||
48 | struct drm_device *dev; | 32 | struct drm_device *dev; |
49 | 33 | ||
50 | int rev; | 34 | struct mdp5_cfg_handler *cfg; |
51 | const struct mdp5_config *hw_cfg; | ||
52 | 35 | ||
53 | /* mapper-id used to request GEM buffer mapped for scanout: */ | 36 | /* mapper-id used to request GEM buffer mapped for scanout: */ |
54 | int id; | 37 | int id; |
55 | struct msm_mmu *mmu; | 38 | struct msm_mmu *mmu; |
56 | 39 | ||
57 | /* for tracking smp allocation amongst pipes: */ | 40 | struct mdp5_smp *smp; |
58 | mdp5_smp_state_t smp_state; | 41 | struct mdp5_ctl_manager *ctlm; |
59 | struct mdp5_client_smp_state smp_client_state[CID_MAX]; | ||
60 | int smp_blk_cnt; | ||
61 | 42 | ||
62 | /* io/register spaces: */ | 43 | /* io/register spaces: */ |
63 | void __iomem *mmio, *vbif; | 44 | void __iomem *mmio, *vbif; |
@@ -71,18 +52,47 @@ struct mdp5_kms { | |||
71 | struct clk *lut_clk; | 52 | struct clk *lut_clk; |
72 | struct clk *vsync_clk; | 53 | struct clk *vsync_clk; |
73 | 54 | ||
74 | struct hdmi *hdmi; | 55 | /* |
56 | * lock to protect access to global resources: ie., following register: | ||
57 | * - REG_MDP5_DISP_INTF_SEL | ||
58 | */ | ||
59 | spinlock_t resource_lock; | ||
75 | 60 | ||
76 | struct mdp_irq error_handler; | 61 | struct mdp_irq error_handler; |
62 | |||
63 | struct { | ||
64 | volatile unsigned long enabled_mask; | ||
65 | struct irq_domain *domain; | ||
66 | } irqcontroller; | ||
77 | }; | 67 | }; |
78 | #define to_mdp5_kms(x) container_of(x, struct mdp5_kms, base) | 68 | #define to_mdp5_kms(x) container_of(x, struct mdp5_kms, base) |
79 | 69 | ||
80 | /* platform config data (ie. from DT, or pdata) */ | 70 | struct mdp5_plane_state { |
81 | struct mdp5_platform_config { | 71 | struct drm_plane_state base; |
82 | struct iommu_domain *iommu; | 72 | |
83 | uint32_t max_clk; | 73 | /* "virtual" zpos.. we calculate actual mixer-stage at runtime |
84 | int smp_blk_cnt; | 74 | * by sorting the attached planes by zpos and then assigning |
75 | * mixer stage lowest to highest. Private planes get default | ||
76 | * zpos of zero, and public planes a unique value that is | ||
77 | * greater than zero. This way, things work out if a naive | ||
78 | * userspace assigns planes to a crtc without setting zpos. | ||
79 | */ | ||
80 | int zpos; | ||
81 | |||
82 | /* the actual mixer stage, calculated in crtc->atomic_check() | ||
83 | * NOTE: this should move to mdp5_crtc_state, when that exists | ||
84 | */ | ||
85 | enum mdp_mixer_stage_id stage; | ||
86 | |||
87 | /* some additional transactional status to help us know in the | ||
88 | * apply path whether we need to update SMP allocation, and | ||
89 | * whether current update is still pending: | ||
90 | */ | ||
91 | bool mode_changed : 1; | ||
92 | bool pending : 1; | ||
85 | }; | 93 | }; |
94 | #define to_mdp5_plane_state(x) \ | ||
95 | container_of(x, struct mdp5_plane_state, base) | ||
86 | 96 | ||
87 | static inline void mdp5_write(struct mdp5_kms *mdp5_kms, u32 reg, u32 data) | 97 | static inline void mdp5_write(struct mdp5_kms *mdp5_kms, u32 reg, u32 data) |
88 | { | 98 | { |
@@ -107,23 +117,6 @@ static inline const char *pipe2name(enum mdp5_pipe pipe) | |||
107 | return names[pipe]; | 117 | return names[pipe]; |
108 | } | 118 | } |
109 | 119 | ||
110 | static inline uint32_t pipe2flush(enum mdp5_pipe pipe) | ||
111 | { | ||
112 | switch (pipe) { | ||
113 | case SSPP_VIG0: return MDP5_CTL_FLUSH_VIG0; | ||
114 | case SSPP_VIG1: return MDP5_CTL_FLUSH_VIG1; | ||
115 | case SSPP_VIG2: return MDP5_CTL_FLUSH_VIG2; | ||
116 | case SSPP_RGB0: return MDP5_CTL_FLUSH_RGB0; | ||
117 | case SSPP_RGB1: return MDP5_CTL_FLUSH_RGB1; | ||
118 | case SSPP_RGB2: return MDP5_CTL_FLUSH_RGB2; | ||
119 | case SSPP_DMA0: return MDP5_CTL_FLUSH_DMA0; | ||
120 | case SSPP_DMA1: return MDP5_CTL_FLUSH_DMA1; | ||
121 | case SSPP_VIG3: return MDP5_CTL_FLUSH_VIG3; | ||
122 | case SSPP_RGB3: return MDP5_CTL_FLUSH_RGB3; | ||
123 | default: return 0; | ||
124 | } | ||
125 | } | ||
126 | |||
127 | static inline int pipe2nclients(enum mdp5_pipe pipe) | 120 | static inline int pipe2nclients(enum mdp5_pipe pipe) |
128 | { | 121 | { |
129 | switch (pipe) { | 122 | switch (pipe) { |
@@ -137,34 +130,6 @@ static inline int pipe2nclients(enum mdp5_pipe pipe) | |||
137 | } | 130 | } |
138 | } | 131 | } |
139 | 132 | ||
140 | static inline enum mdp5_client_id pipe2client(enum mdp5_pipe pipe, int plane) | ||
141 | { | ||
142 | WARN_ON(plane >= pipe2nclients(pipe)); | ||
143 | switch (pipe) { | ||
144 | case SSPP_VIG0: return CID_VIG0_Y + plane; | ||
145 | case SSPP_VIG1: return CID_VIG1_Y + plane; | ||
146 | case SSPP_VIG2: return CID_VIG2_Y + plane; | ||
147 | case SSPP_RGB0: return CID_RGB0; | ||
148 | case SSPP_RGB1: return CID_RGB1; | ||
149 | case SSPP_RGB2: return CID_RGB2; | ||
150 | case SSPP_DMA0: return CID_DMA0_Y + plane; | ||
151 | case SSPP_DMA1: return CID_DMA1_Y + plane; | ||
152 | case SSPP_VIG3: return CID_VIG3_Y + plane; | ||
153 | case SSPP_RGB3: return CID_RGB3; | ||
154 | default: return CID_UNUSED; | ||
155 | } | ||
156 | } | ||
157 | |||
158 | static inline uint32_t mixer2flush(int lm) | ||
159 | { | ||
160 | switch (lm) { | ||
161 | case 0: return MDP5_CTL_FLUSH_LM0; | ||
162 | case 1: return MDP5_CTL_FLUSH_LM1; | ||
163 | case 2: return MDP5_CTL_FLUSH_LM2; | ||
164 | default: return 0; | ||
165 | } | ||
166 | } | ||
167 | |||
168 | static inline uint32_t intf2err(int intf) | 133 | static inline uint32_t intf2err(int intf) |
169 | { | 134 | { |
170 | switch (intf) { | 135 | switch (intf) { |
@@ -197,6 +162,8 @@ void mdp5_irq_uninstall(struct msm_kms *kms); | |||
197 | irqreturn_t mdp5_irq(struct msm_kms *kms); | 162 | irqreturn_t mdp5_irq(struct msm_kms *kms); |
198 | int mdp5_enable_vblank(struct msm_kms *kms, struct drm_crtc *crtc); | 163 | int mdp5_enable_vblank(struct msm_kms *kms, struct drm_crtc *crtc); |
199 | void mdp5_disable_vblank(struct msm_kms *kms, struct drm_crtc *crtc); | 164 | void mdp5_disable_vblank(struct msm_kms *kms, struct drm_crtc *crtc); |
165 | int mdp5_irq_domain_init(struct mdp5_kms *mdp5_kms); | ||
166 | void mdp5_irq_domain_fini(struct mdp5_kms *mdp5_kms); | ||
200 | 167 | ||
201 | static inline | 168 | static inline |
202 | uint32_t mdp5_get_formats(enum mdp5_pipe pipe, uint32_t *pixel_formats, | 169 | uint32_t mdp5_get_formats(enum mdp5_pipe pipe, uint32_t *pixel_formats, |
@@ -210,26 +177,18 @@ uint32_t mdp5_get_formats(enum mdp5_pipe pipe, uint32_t *pixel_formats, | |||
210 | 177 | ||
211 | void mdp5_plane_install_properties(struct drm_plane *plane, | 178 | void mdp5_plane_install_properties(struct drm_plane *plane, |
212 | struct drm_mode_object *obj); | 179 | struct drm_mode_object *obj); |
213 | void mdp5_plane_set_scanout(struct drm_plane *plane, | 180 | uint32_t mdp5_plane_get_flush(struct drm_plane *plane); |
214 | struct drm_framebuffer *fb); | ||
215 | int mdp5_plane_mode_set(struct drm_plane *plane, | ||
216 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | ||
217 | int crtc_x, int crtc_y, | ||
218 | unsigned int crtc_w, unsigned int crtc_h, | ||
219 | uint32_t src_x, uint32_t src_y, | ||
220 | uint32_t src_w, uint32_t src_h); | ||
221 | void mdp5_plane_complete_flip(struct drm_plane *plane); | 181 | void mdp5_plane_complete_flip(struct drm_plane *plane); |
222 | enum mdp5_pipe mdp5_plane_pipe(struct drm_plane *plane); | 182 | enum mdp5_pipe mdp5_plane_pipe(struct drm_plane *plane); |
223 | struct drm_plane *mdp5_plane_init(struct drm_device *dev, | 183 | struct drm_plane *mdp5_plane_init(struct drm_device *dev, |
224 | enum mdp5_pipe pipe, bool private_plane); | 184 | enum mdp5_pipe pipe, bool private_plane, uint32_t reg_offset); |
225 | 185 | ||
226 | uint32_t mdp5_crtc_vblank(struct drm_crtc *crtc); | 186 | uint32_t mdp5_crtc_vblank(struct drm_crtc *crtc); |
227 | 187 | ||
188 | int mdp5_crtc_get_lm(struct drm_crtc *crtc); | ||
228 | void mdp5_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file); | 189 | void mdp5_crtc_cancel_pending_flip(struct drm_crtc *crtc, struct drm_file *file); |
229 | void mdp5_crtc_set_intf(struct drm_crtc *crtc, int intf, | 190 | void mdp5_crtc_set_intf(struct drm_crtc *crtc, int intf, |
230 | enum mdp5_intf intf_id); | 191 | enum mdp5_intf intf_id); |
231 | void mdp5_crtc_attach(struct drm_crtc *crtc, struct drm_plane *plane); | ||
232 | void mdp5_crtc_detach(struct drm_crtc *crtc, struct drm_plane *plane); | ||
233 | struct drm_crtc *mdp5_crtc_init(struct drm_device *dev, | 192 | struct drm_crtc *mdp5_crtc_init(struct drm_device *dev, |
234 | struct drm_plane *plane, int id); | 193 | struct drm_plane *plane, int id); |
235 | 194 | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_plane.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_plane.c index f3daec4412ad..26e5fdea6594 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_plane.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_plane.c | |||
@@ -1,4 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (c) 2014 The Linux Foundation. All rights reserved. | ||
2 | * Copyright (C) 2013 Red Hat | 3 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 4 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 5 | * |
@@ -17,6 +18,7 @@ | |||
17 | 18 | ||
18 | #include "mdp5_kms.h" | 19 | #include "mdp5_kms.h" |
19 | 20 | ||
21 | #define MAX_PLANE 4 | ||
20 | 22 | ||
21 | struct mdp5_plane { | 23 | struct mdp5_plane { |
22 | struct drm_plane base; | 24 | struct drm_plane base; |
@@ -24,6 +26,11 @@ struct mdp5_plane { | |||
24 | 26 | ||
25 | enum mdp5_pipe pipe; | 27 | enum mdp5_pipe pipe; |
26 | 28 | ||
29 | spinlock_t pipe_lock; /* protect REG_MDP5_PIPE_* registers */ | ||
30 | uint32_t reg_offset; | ||
31 | |||
32 | uint32_t flush_mask; /* used to commit pipe registers */ | ||
33 | |||
27 | uint32_t nformats; | 34 | uint32_t nformats; |
28 | uint32_t formats[32]; | 35 | uint32_t formats[32]; |
29 | 36 | ||
@@ -31,31 +38,24 @@ struct mdp5_plane { | |||
31 | }; | 38 | }; |
32 | #define to_mdp5_plane(x) container_of(x, struct mdp5_plane, base) | 39 | #define to_mdp5_plane(x) container_of(x, struct mdp5_plane, base) |
33 | 40 | ||
41 | static int mdp5_plane_mode_set(struct drm_plane *plane, | ||
42 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | ||
43 | int crtc_x, int crtc_y, | ||
44 | unsigned int crtc_w, unsigned int crtc_h, | ||
45 | uint32_t src_x, uint32_t src_y, | ||
46 | uint32_t src_w, uint32_t src_h); | ||
47 | static void set_scanout_locked(struct drm_plane *plane, | ||
48 | struct drm_framebuffer *fb); | ||
49 | |||
34 | static struct mdp5_kms *get_kms(struct drm_plane *plane) | 50 | static struct mdp5_kms *get_kms(struct drm_plane *plane) |
35 | { | 51 | { |
36 | struct msm_drm_private *priv = plane->dev->dev_private; | 52 | struct msm_drm_private *priv = plane->dev->dev_private; |
37 | return to_mdp5_kms(to_mdp_kms(priv->kms)); | 53 | return to_mdp5_kms(to_mdp_kms(priv->kms)); |
38 | } | 54 | } |
39 | 55 | ||
40 | static int mdp5_plane_update(struct drm_plane *plane, | 56 | static bool plane_enabled(struct drm_plane_state *state) |
41 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | ||
42 | int crtc_x, int crtc_y, | ||
43 | unsigned int crtc_w, unsigned int crtc_h, | ||
44 | uint32_t src_x, uint32_t src_y, | ||
45 | uint32_t src_w, uint32_t src_h) | ||
46 | { | 57 | { |
47 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 58 | return state->fb && state->crtc; |
48 | |||
49 | mdp5_plane->enabled = true; | ||
50 | |||
51 | if (plane->fb) | ||
52 | drm_framebuffer_unreference(plane->fb); | ||
53 | |||
54 | drm_framebuffer_reference(fb); | ||
55 | |||
56 | return mdp5_plane_mode_set(plane, crtc, fb, | ||
57 | crtc_x, crtc_y, crtc_w, crtc_h, | ||
58 | src_x, src_y, src_w, src_h); | ||
59 | } | 59 | } |
60 | 60 | ||
61 | static int mdp5_plane_disable(struct drm_plane *plane) | 61 | static int mdp5_plane_disable(struct drm_plane *plane) |
@@ -63,21 +63,13 @@ static int mdp5_plane_disable(struct drm_plane *plane) | |||
63 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 63 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
64 | struct mdp5_kms *mdp5_kms = get_kms(plane); | 64 | struct mdp5_kms *mdp5_kms = get_kms(plane); |
65 | enum mdp5_pipe pipe = mdp5_plane->pipe; | 65 | enum mdp5_pipe pipe = mdp5_plane->pipe; |
66 | int i; | ||
67 | 66 | ||
68 | DBG("%s: disable", mdp5_plane->name); | 67 | DBG("%s: disable", mdp5_plane->name); |
69 | 68 | ||
70 | /* update our SMP request to zero (release all our blks): */ | 69 | if (mdp5_kms) { |
71 | for (i = 0; i < pipe2nclients(pipe); i++) | 70 | /* Release the memory we requested earlier from the SMP: */ |
72 | mdp5_smp_request(mdp5_kms, pipe2client(pipe, i), 0); | 71 | mdp5_smp_release(mdp5_kms->smp, pipe); |
73 | 72 | } | |
74 | /* TODO detaching now will cause us not to get the last | ||
75 | * vblank and mdp5_smp_commit().. so other planes will | ||
76 | * still see smp blocks previously allocated to us as | ||
77 | * in-use.. | ||
78 | */ | ||
79 | if (plane->crtc) | ||
80 | mdp5_crtc_detach(plane->crtc, plane); | ||
81 | 73 | ||
82 | return 0; | 74 | return 0; |
83 | } | 75 | } |
@@ -85,11 +77,8 @@ static int mdp5_plane_disable(struct drm_plane *plane) | |||
85 | static void mdp5_plane_destroy(struct drm_plane *plane) | 77 | static void mdp5_plane_destroy(struct drm_plane *plane) |
86 | { | 78 | { |
87 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 79 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
88 | struct msm_drm_private *priv = plane->dev->dev_private; | ||
89 | |||
90 | if (priv->kms) | ||
91 | mdp5_plane_disable(plane); | ||
92 | 80 | ||
81 | drm_plane_helper_disable(plane); | ||
93 | drm_plane_cleanup(plane); | 82 | drm_plane_cleanup(plane); |
94 | 83 | ||
95 | kfree(mdp5_plane); | 84 | kfree(mdp5_plane); |
@@ -109,109 +98,186 @@ int mdp5_plane_set_property(struct drm_plane *plane, | |||
109 | return -EINVAL; | 98 | return -EINVAL; |
110 | } | 99 | } |
111 | 100 | ||
101 | static void mdp5_plane_reset(struct drm_plane *plane) | ||
102 | { | ||
103 | struct mdp5_plane_state *mdp5_state; | ||
104 | |||
105 | if (plane->state && plane->state->fb) | ||
106 | drm_framebuffer_unreference(plane->state->fb); | ||
107 | |||
108 | kfree(to_mdp5_plane_state(plane->state)); | ||
109 | mdp5_state = kzalloc(sizeof(*mdp5_state), GFP_KERNEL); | ||
110 | |||
111 | if (plane->type == DRM_PLANE_TYPE_PRIMARY) { | ||
112 | mdp5_state->zpos = 0; | ||
113 | } else { | ||
114 | mdp5_state->zpos = 1 + drm_plane_index(plane); | ||
115 | } | ||
116 | |||
117 | plane->state = &mdp5_state->base; | ||
118 | } | ||
119 | |||
120 | static struct drm_plane_state * | ||
121 | mdp5_plane_duplicate_state(struct drm_plane *plane) | ||
122 | { | ||
123 | struct mdp5_plane_state *mdp5_state; | ||
124 | |||
125 | if (WARN_ON(!plane->state)) | ||
126 | return NULL; | ||
127 | |||
128 | mdp5_state = kmemdup(to_mdp5_plane_state(plane->state), | ||
129 | sizeof(*mdp5_state), GFP_KERNEL); | ||
130 | |||
131 | if (mdp5_state && mdp5_state->base.fb) | ||
132 | drm_framebuffer_reference(mdp5_state->base.fb); | ||
133 | |||
134 | mdp5_state->mode_changed = false; | ||
135 | mdp5_state->pending = false; | ||
136 | |||
137 | return &mdp5_state->base; | ||
138 | } | ||
139 | |||
140 | static void mdp5_plane_destroy_state(struct drm_plane *plane, | ||
141 | struct drm_plane_state *state) | ||
142 | { | ||
143 | if (state->fb) | ||
144 | drm_framebuffer_unreference(state->fb); | ||
145 | |||
146 | kfree(to_mdp5_plane_state(state)); | ||
147 | } | ||
148 | |||
112 | static const struct drm_plane_funcs mdp5_plane_funcs = { | 149 | static const struct drm_plane_funcs mdp5_plane_funcs = { |
113 | .update_plane = mdp5_plane_update, | 150 | .update_plane = drm_atomic_helper_update_plane, |
114 | .disable_plane = mdp5_plane_disable, | 151 | .disable_plane = drm_atomic_helper_disable_plane, |
115 | .destroy = mdp5_plane_destroy, | 152 | .destroy = mdp5_plane_destroy, |
116 | .set_property = mdp5_plane_set_property, | 153 | .set_property = mdp5_plane_set_property, |
154 | .reset = mdp5_plane_reset, | ||
155 | .atomic_duplicate_state = mdp5_plane_duplicate_state, | ||
156 | .atomic_destroy_state = mdp5_plane_destroy_state, | ||
117 | }; | 157 | }; |
118 | 158 | ||
119 | void mdp5_plane_set_scanout(struct drm_plane *plane, | 159 | static int mdp5_plane_prepare_fb(struct drm_plane *plane, |
120 | struct drm_framebuffer *fb) | 160 | struct drm_framebuffer *fb) |
121 | { | 161 | { |
122 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 162 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
123 | struct mdp5_kms *mdp5_kms = get_kms(plane); | 163 | struct mdp5_kms *mdp5_kms = get_kms(plane); |
124 | enum mdp5_pipe pipe = mdp5_plane->pipe; | ||
125 | uint32_t nplanes = drm_format_num_planes(fb->pixel_format); | ||
126 | uint32_t iova[4]; | ||
127 | int i; | ||
128 | |||
129 | for (i = 0; i < nplanes; i++) { | ||
130 | struct drm_gem_object *bo = msm_framebuffer_bo(fb, i); | ||
131 | msm_gem_get_iova(bo, mdp5_kms->id, &iova[i]); | ||
132 | } | ||
133 | for (; i < 4; i++) | ||
134 | iova[i] = 0; | ||
135 | 164 | ||
136 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_STRIDE_A(pipe), | 165 | DBG("%s: prepare: FB[%u]", mdp5_plane->name, fb->base.id); |
137 | MDP5_PIPE_SRC_STRIDE_A_P0(fb->pitches[0]) | | 166 | return msm_framebuffer_prepare(fb, mdp5_kms->id); |
138 | MDP5_PIPE_SRC_STRIDE_A_P1(fb->pitches[1])); | ||
139 | |||
140 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_STRIDE_B(pipe), | ||
141 | MDP5_PIPE_SRC_STRIDE_B_P2(fb->pitches[2]) | | ||
142 | MDP5_PIPE_SRC_STRIDE_B_P3(fb->pitches[3])); | ||
143 | |||
144 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC0_ADDR(pipe), iova[0]); | ||
145 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC1_ADDR(pipe), iova[1]); | ||
146 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC2_ADDR(pipe), iova[2]); | ||
147 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC3_ADDR(pipe), iova[3]); | ||
148 | |||
149 | plane->fb = fb; | ||
150 | } | 167 | } |
151 | 168 | ||
152 | /* NOTE: looks like if horizontal decimation is used (if we supported that) | 169 | static void mdp5_plane_cleanup_fb(struct drm_plane *plane, |
153 | * then the width used to calculate SMP block requirements is the post- | 170 | struct drm_framebuffer *fb) |
154 | * decimated width. Ie. SMP buffering sits downstream of decimation (which | ||
155 | * presumably happens during the dma from scanout buffer). | ||
156 | */ | ||
157 | static int request_smp_blocks(struct drm_plane *plane, uint32_t format, | ||
158 | uint32_t nplanes, uint32_t width) | ||
159 | { | 171 | { |
160 | struct drm_device *dev = plane->dev; | ||
161 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 172 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
162 | struct mdp5_kms *mdp5_kms = get_kms(plane); | 173 | struct mdp5_kms *mdp5_kms = get_kms(plane); |
163 | enum mdp5_pipe pipe = mdp5_plane->pipe; | ||
164 | int i, hsub, nlines, nblks, ret; | ||
165 | 174 | ||
166 | hsub = drm_format_horz_chroma_subsampling(format); | 175 | DBG("%s: cleanup: FB[%u]", mdp5_plane->name, fb->base.id); |
176 | msm_framebuffer_cleanup(fb, mdp5_kms->id); | ||
177 | } | ||
167 | 178 | ||
168 | /* different if BWC (compressed framebuffer?) enabled: */ | 179 | static int mdp5_plane_atomic_check(struct drm_plane *plane, |
169 | nlines = 2; | 180 | struct drm_plane_state *state) |
181 | { | ||
182 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | ||
183 | struct drm_plane_state *old_state = plane->state; | ||
170 | 184 | ||
171 | for (i = 0, nblks = 0; i < nplanes; i++) { | 185 | DBG("%s: check (%d -> %d)", mdp5_plane->name, |
172 | int n, fetch_stride, cpp; | 186 | plane_enabled(old_state), plane_enabled(state)); |
173 | 187 | ||
174 | cpp = drm_format_plane_cpp(format, i); | 188 | if (plane_enabled(state) && plane_enabled(old_state)) { |
175 | fetch_stride = width * cpp / (i ? hsub : 1); | 189 | /* we cannot change SMP block configuration during scanout: */ |
190 | bool full_modeset = false; | ||
191 | if (state->fb->pixel_format != old_state->fb->pixel_format) { | ||
192 | DBG("%s: pixel_format change!", mdp5_plane->name); | ||
193 | full_modeset = true; | ||
194 | } | ||
195 | if (state->src_w != old_state->src_w) { | ||
196 | DBG("%s: src_w change!", mdp5_plane->name); | ||
197 | full_modeset = true; | ||
198 | } | ||
199 | if (to_mdp5_plane_state(old_state)->pending) { | ||
200 | DBG("%s: still pending!", mdp5_plane->name); | ||
201 | full_modeset = true; | ||
202 | } | ||
203 | if (full_modeset) { | ||
204 | struct drm_crtc_state *crtc_state = | ||
205 | drm_atomic_get_crtc_state(state->state, state->crtc); | ||
206 | crtc_state->mode_changed = true; | ||
207 | to_mdp5_plane_state(state)->mode_changed = true; | ||
208 | } | ||
209 | } else { | ||
210 | to_mdp5_plane_state(state)->mode_changed = true; | ||
211 | } | ||
176 | 212 | ||
177 | n = DIV_ROUND_UP(fetch_stride * nlines, SMP_BLK_SIZE); | 213 | return 0; |
214 | } | ||
178 | 215 | ||
179 | /* for hw rev v1.00 */ | 216 | static void mdp5_plane_atomic_update(struct drm_plane *plane, |
180 | if (mdp5_kms->rev == 0) | 217 | struct drm_plane_state *old_state) |
181 | n = roundup_pow_of_two(n); | 218 | { |
219 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | ||
220 | struct drm_plane_state *state = plane->state; | ||
182 | 221 | ||
183 | DBG("%s[%d]: request %d SMP blocks", mdp5_plane->name, i, n); | 222 | DBG("%s: update", mdp5_plane->name); |
184 | ret = mdp5_smp_request(mdp5_kms, pipe2client(pipe, i), n); | ||
185 | if (ret) { | ||
186 | dev_err(dev->dev, "Could not allocate %d SMP blocks: %d\n", | ||
187 | n, ret); | ||
188 | return ret; | ||
189 | } | ||
190 | 223 | ||
191 | nblks += n; | 224 | if (!plane_enabled(state)) { |
225 | to_mdp5_plane_state(state)->pending = true; | ||
226 | mdp5_plane_disable(plane); | ||
227 | } else if (to_mdp5_plane_state(state)->mode_changed) { | ||
228 | int ret; | ||
229 | to_mdp5_plane_state(state)->pending = true; | ||
230 | ret = mdp5_plane_mode_set(plane, | ||
231 | state->crtc, state->fb, | ||
232 | state->crtc_x, state->crtc_y, | ||
233 | state->crtc_w, state->crtc_h, | ||
234 | state->src_x, state->src_y, | ||
235 | state->src_w, state->src_h); | ||
236 | /* atomic_check should have ensured that this doesn't fail */ | ||
237 | WARN_ON(ret < 0); | ||
238 | } else { | ||
239 | unsigned long flags; | ||
240 | spin_lock_irqsave(&mdp5_plane->pipe_lock, flags); | ||
241 | set_scanout_locked(plane, state->fb); | ||
242 | spin_unlock_irqrestore(&mdp5_plane->pipe_lock, flags); | ||
192 | } | 243 | } |
193 | |||
194 | /* in success case, return total # of blocks allocated: */ | ||
195 | return nblks; | ||
196 | } | 244 | } |
197 | 245 | ||
198 | static void set_fifo_thresholds(struct drm_plane *plane, int nblks) | 246 | static const struct drm_plane_helper_funcs mdp5_plane_helper_funcs = { |
247 | .prepare_fb = mdp5_plane_prepare_fb, | ||
248 | .cleanup_fb = mdp5_plane_cleanup_fb, | ||
249 | .atomic_check = mdp5_plane_atomic_check, | ||
250 | .atomic_update = mdp5_plane_atomic_update, | ||
251 | }; | ||
252 | |||
253 | static void set_scanout_locked(struct drm_plane *plane, | ||
254 | struct drm_framebuffer *fb) | ||
199 | { | 255 | { |
200 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | 256 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
201 | struct mdp5_kms *mdp5_kms = get_kms(plane); | 257 | struct mdp5_kms *mdp5_kms = get_kms(plane); |
202 | enum mdp5_pipe pipe = mdp5_plane->pipe; | 258 | enum mdp5_pipe pipe = mdp5_plane->pipe; |
203 | uint32_t val; | ||
204 | 259 | ||
205 | /* 1/4 of SMP pool that is being fetched */ | 260 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_STRIDE_A(pipe), |
206 | val = (nblks * SMP_ENTRIES_PER_BLK) / 4; | 261 | MDP5_PIPE_SRC_STRIDE_A_P0(fb->pitches[0]) | |
262 | MDP5_PIPE_SRC_STRIDE_A_P1(fb->pitches[1])); | ||
207 | 263 | ||
208 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_0(pipe), val * 1); | 264 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_STRIDE_B(pipe), |
209 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_1(pipe), val * 2); | 265 | MDP5_PIPE_SRC_STRIDE_B_P2(fb->pitches[2]) | |
210 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_2(pipe), val * 3); | 266 | MDP5_PIPE_SRC_STRIDE_B_P3(fb->pitches[3])); |
267 | |||
268 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC0_ADDR(pipe), | ||
269 | msm_framebuffer_iova(fb, mdp5_kms->id, 0)); | ||
270 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC1_ADDR(pipe), | ||
271 | msm_framebuffer_iova(fb, mdp5_kms->id, 1)); | ||
272 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC2_ADDR(pipe), | ||
273 | msm_framebuffer_iova(fb, mdp5_kms->id, 2)); | ||
274 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC3_ADDR(pipe), | ||
275 | msm_framebuffer_iova(fb, mdp5_kms->id, 4)); | ||
211 | 276 | ||
277 | plane->fb = fb; | ||
212 | } | 278 | } |
213 | 279 | ||
214 | int mdp5_plane_mode_set(struct drm_plane *plane, | 280 | static int mdp5_plane_mode_set(struct drm_plane *plane, |
215 | struct drm_crtc *crtc, struct drm_framebuffer *fb, | 281 | struct drm_crtc *crtc, struct drm_framebuffer *fb, |
216 | int crtc_x, int crtc_y, | 282 | int crtc_x, int crtc_y, |
217 | unsigned int crtc_w, unsigned int crtc_h, | 283 | unsigned int crtc_w, unsigned int crtc_h, |
@@ -225,7 +291,8 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
225 | uint32_t nplanes, config = 0; | 291 | uint32_t nplanes, config = 0; |
226 | uint32_t phasex_step = 0, phasey_step = 0; | 292 | uint32_t phasex_step = 0, phasey_step = 0; |
227 | uint32_t hdecm = 0, vdecm = 0; | 293 | uint32_t hdecm = 0, vdecm = 0; |
228 | int i, nblks; | 294 | unsigned long flags; |
295 | int ret; | ||
229 | 296 | ||
230 | nplanes = drm_format_num_planes(fb->pixel_format); | 297 | nplanes = drm_format_num_planes(fb->pixel_format); |
231 | 298 | ||
@@ -243,12 +310,11 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
243 | fb->base.id, src_x, src_y, src_w, src_h, | 310 | fb->base.id, src_x, src_y, src_w, src_h, |
244 | crtc->base.id, crtc_x, crtc_y, crtc_w, crtc_h); | 311 | crtc->base.id, crtc_x, crtc_y, crtc_w, crtc_h); |
245 | 312 | ||
246 | /* | 313 | /* Request some memory from the SMP: */ |
247 | * Calculate and request required # of smp blocks: | 314 | ret = mdp5_smp_request(mdp5_kms->smp, |
248 | */ | 315 | mdp5_plane->pipe, fb->pixel_format, src_w); |
249 | nblks = request_smp_blocks(plane, fb->pixel_format, nplanes, src_w); | 316 | if (ret) |
250 | if (nblks < 0) | 317 | return ret; |
251 | return nblks; | ||
252 | 318 | ||
253 | /* | 319 | /* |
254 | * Currently we update the hw for allocations/requests immediately, | 320 | * Currently we update the hw for allocations/requests immediately, |
@@ -256,8 +322,7 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
256 | * would move into atomic->check_plane_state(), while updating the | 322 | * would move into atomic->check_plane_state(), while updating the |
257 | * hw would remain here: | 323 | * hw would remain here: |
258 | */ | 324 | */ |
259 | for (i = 0; i < pipe2nclients(pipe); i++) | 325 | mdp5_smp_configure(mdp5_kms->smp, pipe); |
260 | mdp5_smp_configure(mdp5_kms, pipe2client(pipe, i)); | ||
261 | 326 | ||
262 | if (src_w != crtc_w) { | 327 | if (src_w != crtc_w) { |
263 | config |= MDP5_PIPE_SCALE_CONFIG_SCALEX_EN; | 328 | config |= MDP5_PIPE_SCALE_CONFIG_SCALEX_EN; |
@@ -269,6 +334,8 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
269 | /* TODO calc phasey_step, vdecm */ | 334 | /* TODO calc phasey_step, vdecm */ |
270 | } | 335 | } |
271 | 336 | ||
337 | spin_lock_irqsave(&mdp5_plane->pipe_lock, flags); | ||
338 | |||
272 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_IMG_SIZE(pipe), | 339 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_IMG_SIZE(pipe), |
273 | MDP5_PIPE_SRC_IMG_SIZE_WIDTH(src_w) | | 340 | MDP5_PIPE_SRC_IMG_SIZE_WIDTH(src_w) | |
274 | MDP5_PIPE_SRC_IMG_SIZE_HEIGHT(src_h)); | 341 | MDP5_PIPE_SRC_IMG_SIZE_HEIGHT(src_h)); |
@@ -289,8 +356,6 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
289 | MDP5_PIPE_OUT_XY_X(crtc_x) | | 356 | MDP5_PIPE_OUT_XY_X(crtc_x) | |
290 | MDP5_PIPE_OUT_XY_Y(crtc_y)); | 357 | MDP5_PIPE_OUT_XY_Y(crtc_y)); |
291 | 358 | ||
292 | mdp5_plane_set_scanout(plane, fb); | ||
293 | |||
294 | format = to_mdp_format(msm_framebuffer_format(fb)); | 359 | format = to_mdp_format(msm_framebuffer_format(fb)); |
295 | 360 | ||
296 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_FORMAT(pipe), | 361 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_SRC_FORMAT(pipe), |
@@ -330,22 +395,24 @@ int mdp5_plane_mode_set(struct drm_plane *plane, | |||
330 | MDP5_PIPE_SCALE_CONFIG_SCALEX_MAX_FILTER(SCALE_FILTER_NEAREST) | | 395 | MDP5_PIPE_SCALE_CONFIG_SCALEX_MAX_FILTER(SCALE_FILTER_NEAREST) | |
331 | MDP5_PIPE_SCALE_CONFIG_SCALEY_MAX_FILTER(SCALE_FILTER_NEAREST)); | 396 | MDP5_PIPE_SCALE_CONFIG_SCALEY_MAX_FILTER(SCALE_FILTER_NEAREST)); |
332 | 397 | ||
333 | set_fifo_thresholds(plane, nblks); | 398 | set_scanout_locked(plane, fb); |
334 | 399 | ||
335 | /* TODO detach from old crtc (if we had more than one) */ | 400 | spin_unlock_irqrestore(&mdp5_plane->pipe_lock, flags); |
336 | mdp5_crtc_attach(crtc, plane); | ||
337 | 401 | ||
338 | return 0; | 402 | return ret; |
339 | } | 403 | } |
340 | 404 | ||
341 | void mdp5_plane_complete_flip(struct drm_plane *plane) | 405 | void mdp5_plane_complete_flip(struct drm_plane *plane) |
342 | { | 406 | { |
343 | struct mdp5_kms *mdp5_kms = get_kms(plane); | 407 | struct mdp5_kms *mdp5_kms = get_kms(plane); |
344 | enum mdp5_pipe pipe = to_mdp5_plane(plane)->pipe; | 408 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); |
345 | int i; | 409 | enum mdp5_pipe pipe = mdp5_plane->pipe; |
410 | |||
411 | DBG("%s: complete flip", mdp5_plane->name); | ||
346 | 412 | ||
347 | for (i = 0; i < pipe2nclients(pipe); i++) | 413 | mdp5_smp_commit(mdp5_kms->smp, pipe); |
348 | mdp5_smp_commit(mdp5_kms, pipe2client(pipe, i)); | 414 | |
415 | to_mdp5_plane_state(plane->state)->pending = false; | ||
349 | } | 416 | } |
350 | 417 | ||
351 | enum mdp5_pipe mdp5_plane_pipe(struct drm_plane *plane) | 418 | enum mdp5_pipe mdp5_plane_pipe(struct drm_plane *plane) |
@@ -354,9 +421,16 @@ enum mdp5_pipe mdp5_plane_pipe(struct drm_plane *plane) | |||
354 | return mdp5_plane->pipe; | 421 | return mdp5_plane->pipe; |
355 | } | 422 | } |
356 | 423 | ||
424 | uint32_t mdp5_plane_get_flush(struct drm_plane *plane) | ||
425 | { | ||
426 | struct mdp5_plane *mdp5_plane = to_mdp5_plane(plane); | ||
427 | |||
428 | return mdp5_plane->flush_mask; | ||
429 | } | ||
430 | |||
357 | /* initialize plane */ | 431 | /* initialize plane */ |
358 | struct drm_plane *mdp5_plane_init(struct drm_device *dev, | 432 | struct drm_plane *mdp5_plane_init(struct drm_device *dev, |
359 | enum mdp5_pipe pipe, bool private_plane) | 433 | enum mdp5_pipe pipe, bool private_plane, uint32_t reg_offset) |
360 | { | 434 | { |
361 | struct drm_plane *plane = NULL; | 435 | struct drm_plane *plane = NULL; |
362 | struct mdp5_plane *mdp5_plane; | 436 | struct mdp5_plane *mdp5_plane; |
@@ -377,10 +451,18 @@ struct drm_plane *mdp5_plane_init(struct drm_device *dev, | |||
377 | mdp5_plane->nformats = mdp5_get_formats(pipe, mdp5_plane->formats, | 451 | mdp5_plane->nformats = mdp5_get_formats(pipe, mdp5_plane->formats, |
378 | ARRAY_SIZE(mdp5_plane->formats)); | 452 | ARRAY_SIZE(mdp5_plane->formats)); |
379 | 453 | ||
454 | mdp5_plane->flush_mask = mdp_ctl_flush_mask_pipe(pipe); | ||
455 | mdp5_plane->reg_offset = reg_offset; | ||
456 | spin_lock_init(&mdp5_plane->pipe_lock); | ||
457 | |||
380 | type = private_plane ? DRM_PLANE_TYPE_PRIMARY : DRM_PLANE_TYPE_OVERLAY; | 458 | type = private_plane ? DRM_PLANE_TYPE_PRIMARY : DRM_PLANE_TYPE_OVERLAY; |
381 | drm_universal_plane_init(dev, plane, 0xff, &mdp5_plane_funcs, | 459 | ret = drm_universal_plane_init(dev, plane, 0xff, &mdp5_plane_funcs, |
382 | mdp5_plane->formats, mdp5_plane->nformats, | 460 | mdp5_plane->formats, mdp5_plane->nformats, |
383 | type); | 461 | type); |
462 | if (ret) | ||
463 | goto fail; | ||
464 | |||
465 | drm_plane_helper_add(plane, &mdp5_plane_helper_funcs); | ||
384 | 466 | ||
385 | mdp5_plane_install_properties(plane, &plane->base); | 467 | mdp5_plane_install_properties(plane, &plane->base); |
386 | 468 | ||
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.c b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.c index 2d0236b963a6..bf551885e019 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.c +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.c | |||
@@ -1,4 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (c) 2014, The Linux Foundation. All rights reserved. | ||
2 | * Copyright (C) 2013 Red Hat | 3 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 4 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 5 | * |
@@ -29,8 +30,11 @@ | |||
29 | * Based on the size of the attached scanout buffer, a certain # of | 30 | * Based on the size of the attached scanout buffer, a certain # of |
30 | * blocks must be allocated to that client out of the shared pool. | 31 | * blocks must be allocated to that client out of the shared pool. |
31 | * | 32 | * |
32 | * For each block, it can be either free, or pending/in-use by a | 33 | * In some hw, some blocks are statically allocated for certain pipes |
33 | * client. The updates happen in three steps: | 34 | * and CANNOT be re-allocated (eg: MMB0 and MMB1 both tied to RGB0). |
35 | * | ||
36 | * For each block that can be dynamically allocated, it can be either | ||
37 | * free, or pending/in-use by a client. The updates happen in three steps: | ||
34 | * | 38 | * |
35 | * 1) mdp5_smp_request(): | 39 | * 1) mdp5_smp_request(): |
36 | * When plane scanout is setup, calculate required number of | 40 | * When plane scanout is setup, calculate required number of |
@@ -61,21 +65,68 @@ | |||
61 | * inuse and pending state of all clients.. | 65 | * inuse and pending state of all clients.. |
62 | */ | 66 | */ |
63 | 67 | ||
64 | static DEFINE_SPINLOCK(smp_lock); | 68 | struct mdp5_smp { |
69 | struct drm_device *dev; | ||
70 | |||
71 | int blk_cnt; | ||
72 | int blk_size; | ||
73 | |||
74 | spinlock_t state_lock; | ||
75 | mdp5_smp_state_t state; /* to track smp allocation amongst pipes: */ | ||
76 | |||
77 | struct mdp5_client_smp_state client_state[CID_MAX]; | ||
78 | }; | ||
65 | 79 | ||
80 | static inline | ||
81 | struct mdp5_kms *get_kms(struct mdp5_smp *smp) | ||
82 | { | ||
83 | struct msm_drm_private *priv = smp->dev->dev_private; | ||
84 | |||
85 | return to_mdp5_kms(to_mdp_kms(priv->kms)); | ||
86 | } | ||
87 | |||
88 | static inline enum mdp5_client_id pipe2client(enum mdp5_pipe pipe, int plane) | ||
89 | { | ||
90 | WARN_ON(plane >= pipe2nclients(pipe)); | ||
91 | switch (pipe) { | ||
92 | case SSPP_VIG0: return CID_VIG0_Y + plane; | ||
93 | case SSPP_VIG1: return CID_VIG1_Y + plane; | ||
94 | case SSPP_VIG2: return CID_VIG2_Y + plane; | ||
95 | case SSPP_RGB0: return CID_RGB0; | ||
96 | case SSPP_RGB1: return CID_RGB1; | ||
97 | case SSPP_RGB2: return CID_RGB2; | ||
98 | case SSPP_DMA0: return CID_DMA0_Y + plane; | ||
99 | case SSPP_DMA1: return CID_DMA1_Y + plane; | ||
100 | case SSPP_VIG3: return CID_VIG3_Y + plane; | ||
101 | case SSPP_RGB3: return CID_RGB3; | ||
102 | default: return CID_UNUSED; | ||
103 | } | ||
104 | } | ||
66 | 105 | ||
67 | /* step #1: update # of blocks pending for the client: */ | 106 | /* step #1: update # of blocks pending for the client: */ |
68 | int mdp5_smp_request(struct mdp5_kms *mdp5_kms, | 107 | static int smp_request_block(struct mdp5_smp *smp, |
69 | enum mdp5_client_id cid, int nblks) | 108 | enum mdp5_client_id cid, int nblks) |
70 | { | 109 | { |
71 | struct mdp5_client_smp_state *ps = &mdp5_kms->smp_client_state[cid]; | 110 | struct mdp5_kms *mdp5_kms = get_kms(smp); |
72 | int i, ret, avail, cur_nblks, cnt = mdp5_kms->smp_blk_cnt; | 111 | const struct mdp5_cfg_hw *hw_cfg; |
112 | struct mdp5_client_smp_state *ps = &smp->client_state[cid]; | ||
113 | int i, ret, avail, cur_nblks, cnt = smp->blk_cnt; | ||
114 | int reserved; | ||
73 | unsigned long flags; | 115 | unsigned long flags; |
74 | 116 | ||
75 | spin_lock_irqsave(&smp_lock, flags); | 117 | hw_cfg = mdp5_cfg_get_hw_config(mdp5_kms->cfg); |
118 | reserved = hw_cfg->smp.reserved[cid]; | ||
119 | |||
120 | spin_lock_irqsave(&smp->state_lock, flags); | ||
76 | 121 | ||
77 | avail = cnt - bitmap_weight(mdp5_kms->smp_state, cnt); | 122 | nblks -= reserved; |
123 | if (reserved) | ||
124 | DBG("%d MMBs allocated (%d reserved)", nblks, reserved); | ||
125 | |||
126 | avail = cnt - bitmap_weight(smp->state, cnt); | ||
78 | if (nblks > avail) { | 127 | if (nblks > avail) { |
128 | dev_err(mdp5_kms->dev->dev, "out of blks (req=%d > avail=%d)\n", | ||
129 | nblks, avail); | ||
79 | ret = -ENOSPC; | 130 | ret = -ENOSPC; |
80 | goto fail; | 131 | goto fail; |
81 | } | 132 | } |
@@ -84,9 +135,9 @@ int mdp5_smp_request(struct mdp5_kms *mdp5_kms, | |||
84 | if (nblks > cur_nblks) { | 135 | if (nblks > cur_nblks) { |
85 | /* grow the existing pending reservation: */ | 136 | /* grow the existing pending reservation: */ |
86 | for (i = cur_nblks; i < nblks; i++) { | 137 | for (i = cur_nblks; i < nblks; i++) { |
87 | int blk = find_first_zero_bit(mdp5_kms->smp_state, cnt); | 138 | int blk = find_first_zero_bit(smp->state, cnt); |
88 | set_bit(blk, ps->pending); | 139 | set_bit(blk, ps->pending); |
89 | set_bit(blk, mdp5_kms->smp_state); | 140 | set_bit(blk, smp->state); |
90 | } | 141 | } |
91 | } else { | 142 | } else { |
92 | /* shrink the existing pending reservation: */ | 143 | /* shrink the existing pending reservation: */ |
@@ -98,15 +149,88 @@ int mdp5_smp_request(struct mdp5_kms *mdp5_kms, | |||
98 | } | 149 | } |
99 | 150 | ||
100 | fail: | 151 | fail: |
101 | spin_unlock_irqrestore(&smp_lock, flags); | 152 | spin_unlock_irqrestore(&smp->state_lock, flags); |
153 | return 0; | ||
154 | } | ||
155 | |||
156 | static void set_fifo_thresholds(struct mdp5_smp *smp, | ||
157 | enum mdp5_pipe pipe, int nblks) | ||
158 | { | ||
159 | struct mdp5_kms *mdp5_kms = get_kms(smp); | ||
160 | u32 smp_entries_per_blk = smp->blk_size / (128 / BITS_PER_BYTE); | ||
161 | u32 val; | ||
162 | |||
163 | /* 1/4 of SMP pool that is being fetched */ | ||
164 | val = (nblks * smp_entries_per_blk) / 4; | ||
165 | |||
166 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_0(pipe), val * 1); | ||
167 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_1(pipe), val * 2); | ||
168 | mdp5_write(mdp5_kms, REG_MDP5_PIPE_REQPRIO_FIFO_WM_2(pipe), val * 3); | ||
169 | } | ||
170 | |||
171 | /* | ||
172 | * NOTE: looks like if horizontal decimation is used (if we supported that) | ||
173 | * then the width used to calculate SMP block requirements is the post- | ||
174 | * decimated width. Ie. SMP buffering sits downstream of decimation (which | ||
175 | * presumably happens during the dma from scanout buffer). | ||
176 | */ | ||
177 | int mdp5_smp_request(struct mdp5_smp *smp, enum mdp5_pipe pipe, u32 fmt, u32 width) | ||
178 | { | ||
179 | struct mdp5_kms *mdp5_kms = get_kms(smp); | ||
180 | struct drm_device *dev = mdp5_kms->dev; | ||
181 | int rev = mdp5_cfg_get_hw_rev(mdp5_kms->cfg); | ||
182 | int i, hsub, nplanes, nlines, nblks, ret; | ||
183 | |||
184 | nplanes = drm_format_num_planes(fmt); | ||
185 | hsub = drm_format_horz_chroma_subsampling(fmt); | ||
186 | |||
187 | /* different if BWC (compressed framebuffer?) enabled: */ | ||
188 | nlines = 2; | ||
189 | |||
190 | for (i = 0, nblks = 0; i < nplanes; i++) { | ||
191 | int n, fetch_stride, cpp; | ||
192 | |||
193 | cpp = drm_format_plane_cpp(fmt, i); | ||
194 | fetch_stride = width * cpp / (i ? hsub : 1); | ||
195 | |||
196 | n = DIV_ROUND_UP(fetch_stride * nlines, smp->blk_size); | ||
197 | |||
198 | /* for hw rev v1.00 */ | ||
199 | if (rev == 0) | ||
200 | n = roundup_pow_of_two(n); | ||
201 | |||
202 | DBG("%s[%d]: request %d SMP blocks", pipe2name(pipe), i, n); | ||
203 | ret = smp_request_block(smp, pipe2client(pipe, i), n); | ||
204 | if (ret) { | ||
205 | dev_err(dev->dev, "Cannot allocate %d SMP blocks: %d\n", | ||
206 | n, ret); | ||
207 | return ret; | ||
208 | } | ||
209 | |||
210 | nblks += n; | ||
211 | } | ||
212 | |||
213 | set_fifo_thresholds(smp, pipe, nblks); | ||
214 | |||
102 | return 0; | 215 | return 0; |
103 | } | 216 | } |
104 | 217 | ||
105 | static void update_smp_state(struct mdp5_kms *mdp5_kms, | 218 | /* Release SMP blocks for all clients of the pipe */ |
219 | void mdp5_smp_release(struct mdp5_smp *smp, enum mdp5_pipe pipe) | ||
220 | { | ||
221 | int i, nblks; | ||
222 | |||
223 | for (i = 0, nblks = 0; i < pipe2nclients(pipe); i++) | ||
224 | smp_request_block(smp, pipe2client(pipe, i), 0); | ||
225 | set_fifo_thresholds(smp, pipe, 0); | ||
226 | } | ||
227 | |||
228 | static void update_smp_state(struct mdp5_smp *smp, | ||
106 | enum mdp5_client_id cid, mdp5_smp_state_t *assigned) | 229 | enum mdp5_client_id cid, mdp5_smp_state_t *assigned) |
107 | { | 230 | { |
108 | int cnt = mdp5_kms->smp_blk_cnt; | 231 | struct mdp5_kms *mdp5_kms = get_kms(smp); |
109 | uint32_t blk, val; | 232 | int cnt = smp->blk_cnt; |
233 | u32 blk, val; | ||
110 | 234 | ||
111 | for_each_set_bit(blk, *assigned, cnt) { | 235 | for_each_set_bit(blk, *assigned, cnt) { |
112 | int idx = blk / 3; | 236 | int idx = blk / 3; |
@@ -135,39 +259,80 @@ static void update_smp_state(struct mdp5_kms *mdp5_kms, | |||
135 | } | 259 | } |
136 | 260 | ||
137 | /* step #2: configure hw for union(pending, inuse): */ | 261 | /* step #2: configure hw for union(pending, inuse): */ |
138 | void mdp5_smp_configure(struct mdp5_kms *mdp5_kms, enum mdp5_client_id cid) | 262 | void mdp5_smp_configure(struct mdp5_smp *smp, enum mdp5_pipe pipe) |
139 | { | 263 | { |
140 | struct mdp5_client_smp_state *ps = &mdp5_kms->smp_client_state[cid]; | 264 | int cnt = smp->blk_cnt; |
141 | int cnt = mdp5_kms->smp_blk_cnt; | ||
142 | mdp5_smp_state_t assigned; | 265 | mdp5_smp_state_t assigned; |
266 | int i; | ||
143 | 267 | ||
144 | bitmap_or(assigned, ps->inuse, ps->pending, cnt); | 268 | for (i = 0; i < pipe2nclients(pipe); i++) { |
145 | update_smp_state(mdp5_kms, cid, &assigned); | 269 | enum mdp5_client_id cid = pipe2client(pipe, i); |
270 | struct mdp5_client_smp_state *ps = &smp->client_state[cid]; | ||
271 | |||
272 | bitmap_or(assigned, ps->inuse, ps->pending, cnt); | ||
273 | update_smp_state(smp, cid, &assigned); | ||
274 | } | ||
146 | } | 275 | } |
147 | 276 | ||
148 | /* step #3: after vblank, copy pending -> inuse: */ | 277 | /* step #3: after vblank, copy pending -> inuse: */ |
149 | void mdp5_smp_commit(struct mdp5_kms *mdp5_kms, enum mdp5_client_id cid) | 278 | void mdp5_smp_commit(struct mdp5_smp *smp, enum mdp5_pipe pipe) |
150 | { | 279 | { |
151 | struct mdp5_client_smp_state *ps = &mdp5_kms->smp_client_state[cid]; | 280 | int cnt = smp->blk_cnt; |
152 | int cnt = mdp5_kms->smp_blk_cnt; | ||
153 | mdp5_smp_state_t released; | 281 | mdp5_smp_state_t released; |
282 | int i; | ||
283 | |||
284 | for (i = 0; i < pipe2nclients(pipe); i++) { | ||
285 | enum mdp5_client_id cid = pipe2client(pipe, i); | ||
286 | struct mdp5_client_smp_state *ps = &smp->client_state[cid]; | ||
287 | |||
288 | /* | ||
289 | * Figure out if there are any blocks we where previously | ||
290 | * using, which can be released and made available to other | ||
291 | * clients: | ||
292 | */ | ||
293 | if (bitmap_andnot(released, ps->inuse, ps->pending, cnt)) { | ||
294 | unsigned long flags; | ||
295 | |||
296 | spin_lock_irqsave(&smp->state_lock, flags); | ||
297 | /* clear released blocks: */ | ||
298 | bitmap_andnot(smp->state, smp->state, released, cnt); | ||
299 | spin_unlock_irqrestore(&smp->state_lock, flags); | ||
154 | 300 | ||
155 | /* | 301 | update_smp_state(smp, CID_UNUSED, &released); |
156 | * Figure out if there are any blocks we where previously | 302 | } |
157 | * using, which can be released and made available to other | 303 | |
158 | * clients: | 304 | bitmap_copy(ps->inuse, ps->pending, cnt); |
159 | */ | ||
160 | if (bitmap_andnot(released, ps->inuse, ps->pending, cnt)) { | ||
161 | unsigned long flags; | ||
162 | |||
163 | spin_lock_irqsave(&smp_lock, flags); | ||
164 | /* clear released blocks: */ | ||
165 | bitmap_andnot(mdp5_kms->smp_state, mdp5_kms->smp_state, | ||
166 | released, cnt); | ||
167 | spin_unlock_irqrestore(&smp_lock, flags); | ||
168 | |||
169 | update_smp_state(mdp5_kms, CID_UNUSED, &released); | ||
170 | } | 305 | } |
306 | } | ||
307 | |||
308 | void mdp5_smp_destroy(struct mdp5_smp *smp) | ||
309 | { | ||
310 | kfree(smp); | ||
311 | } | ||
312 | |||
313 | struct mdp5_smp *mdp5_smp_init(struct drm_device *dev, const struct mdp5_smp_block *cfg) | ||
314 | { | ||
315 | struct mdp5_smp *smp = NULL; | ||
316 | int ret; | ||
317 | |||
318 | smp = kzalloc(sizeof(*smp), GFP_KERNEL); | ||
319 | if (unlikely(!smp)) { | ||
320 | ret = -ENOMEM; | ||
321 | goto fail; | ||
322 | } | ||
323 | |||
324 | smp->dev = dev; | ||
325 | smp->blk_cnt = cfg->mmb_count; | ||
326 | smp->blk_size = cfg->mmb_size; | ||
327 | |||
328 | /* statically tied MMBs cannot be re-allocated: */ | ||
329 | bitmap_copy(smp->state, cfg->reserved_state, smp->blk_cnt); | ||
330 | spin_lock_init(&smp->state_lock); | ||
331 | |||
332 | return smp; | ||
333 | fail: | ||
334 | if (smp) | ||
335 | mdp5_smp_destroy(smp); | ||
171 | 336 | ||
172 | bitmap_copy(ps->inuse, ps->pending, cnt); | 337 | return ERR_PTR(ret); |
173 | } | 338 | } |
diff --git a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.h b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.h index 0ab739e1a1dd..e47179f63585 100644 --- a/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.h +++ b/drivers/gpu/drm/msm/mdp/mdp5/mdp5_smp.h | |||
@@ -1,4 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (c) 2014, The Linux Foundation. All rights reserved. | ||
2 | * Copyright (C) 2013 Red Hat | 3 | * Copyright (C) 2013 Red Hat |
3 | * Author: Rob Clark <robdclark@gmail.com> | 4 | * Author: Rob Clark <robdclark@gmail.com> |
4 | * | 5 | * |
@@ -20,22 +21,26 @@ | |||
20 | 21 | ||
21 | #include "msm_drv.h" | 22 | #include "msm_drv.h" |
22 | 23 | ||
23 | #define MAX_SMP_BLOCKS 22 | ||
24 | #define SMP_BLK_SIZE 4096 | ||
25 | #define SMP_ENTRIES_PER_BLK (SMP_BLK_SIZE / 16) | ||
26 | |||
27 | typedef DECLARE_BITMAP(mdp5_smp_state_t, MAX_SMP_BLOCKS); | ||
28 | |||
29 | struct mdp5_client_smp_state { | 24 | struct mdp5_client_smp_state { |
30 | mdp5_smp_state_t inuse; | 25 | mdp5_smp_state_t inuse; |
31 | mdp5_smp_state_t pending; | 26 | mdp5_smp_state_t pending; |
32 | }; | 27 | }; |
33 | 28 | ||
34 | struct mdp5_kms; | 29 | struct mdp5_kms; |
30 | struct mdp5_smp; | ||
31 | |||
32 | /* | ||
33 | * SMP module prototypes: | ||
34 | * mdp5_smp_init() returns a SMP @handler, | ||
35 | * which is then used to call the other mdp5_smp_*(handler, ...) functions. | ||
36 | */ | ||
35 | 37 | ||
36 | int mdp5_smp_request(struct mdp5_kms *mdp5_kms, enum mdp5_client_id cid, int nblks); | 38 | struct mdp5_smp *mdp5_smp_init(struct drm_device *dev, const struct mdp5_smp_block *cfg); |
37 | void mdp5_smp_configure(struct mdp5_kms *mdp5_kms, enum mdp5_client_id cid); | 39 | void mdp5_smp_destroy(struct mdp5_smp *smp); |
38 | void mdp5_smp_commit(struct mdp5_kms *mdp5_kms, enum mdp5_client_id cid); | ||
39 | 40 | ||
41 | int mdp5_smp_request(struct mdp5_smp *smp, enum mdp5_pipe pipe, u32 fmt, u32 width); | ||
42 | void mdp5_smp_configure(struct mdp5_smp *smp, enum mdp5_pipe pipe); | ||
43 | void mdp5_smp_commit(struct mdp5_smp *smp, enum mdp5_pipe pipe); | ||
44 | void mdp5_smp_release(struct mdp5_smp *smp, enum mdp5_pipe pipe); | ||
40 | 45 | ||
41 | #endif /* __MDP5_SMP_H__ */ | 46 | #endif /* __MDP5_SMP_H__ */ |
diff --git a/drivers/gpu/drm/msm/msm_atomic.c b/drivers/gpu/drm/msm/msm_atomic.c new file mode 100644 index 000000000000..f0de412e13dc --- /dev/null +++ b/drivers/gpu/drm/msm/msm_atomic.c | |||
@@ -0,0 +1,163 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2014 Red Hat | ||
3 | * Author: Rob Clark <robdclark@gmail.com> | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or modify it | ||
6 | * under the terms of the GNU General Public License version 2 as published by | ||
7 | * the Free Software Foundation. | ||
8 | * | ||
9 | * This program is distributed in the hope that it will be useful, but WITHOUT | ||
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | ||
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | ||
12 | * more details. | ||
13 | * | ||
14 | * You should have received a copy of the GNU General Public License along with | ||
15 | * this program. If not, see <http://www.gnu.org/licenses/>. | ||
16 | */ | ||
17 | |||
18 | #include "msm_drv.h" | ||
19 | #include "msm_kms.h" | ||
20 | #include "msm_gem.h" | ||
21 | |||
22 | struct msm_commit { | ||
23 | struct drm_atomic_state *state; | ||
24 | uint32_t fence; | ||
25 | struct msm_fence_cb fence_cb; | ||
26 | }; | ||
27 | |||
28 | static void fence_cb(struct msm_fence_cb *cb); | ||
29 | |||
30 | static struct msm_commit *new_commit(struct drm_atomic_state *state) | ||
31 | { | ||
32 | struct msm_commit *c = kzalloc(sizeof(*c), GFP_KERNEL); | ||
33 | |||
34 | if (!c) | ||
35 | return NULL; | ||
36 | |||
37 | c->state = state; | ||
38 | /* TODO we might need a way to indicate to run the cb on a | ||
39 | * different wq so wait_for_vblanks() doesn't block retiring | ||
40 | * bo's.. | ||
41 | */ | ||
42 | INIT_FENCE_CB(&c->fence_cb, fence_cb); | ||
43 | |||
44 | return c; | ||
45 | } | ||
46 | |||
47 | /* The (potentially) asynchronous part of the commit. At this point | ||
48 | * nothing can fail short of armageddon. | ||
49 | */ | ||
50 | static void complete_commit(struct msm_commit *c) | ||
51 | { | ||
52 | struct drm_atomic_state *state = c->state; | ||
53 | struct drm_device *dev = state->dev; | ||
54 | |||
55 | drm_atomic_helper_commit_pre_planes(dev, state); | ||
56 | |||
57 | drm_atomic_helper_commit_planes(dev, state); | ||
58 | |||
59 | drm_atomic_helper_commit_post_planes(dev, state); | ||
60 | |||
61 | drm_atomic_helper_wait_for_vblanks(dev, state); | ||
62 | |||
63 | drm_atomic_helper_cleanup_planes(dev, state); | ||
64 | |||
65 | drm_atomic_state_free(state); | ||
66 | |||
67 | kfree(c); | ||
68 | } | ||
69 | |||
70 | static void fence_cb(struct msm_fence_cb *cb) | ||
71 | { | ||
72 | struct msm_commit *c = | ||
73 | container_of(cb, struct msm_commit, fence_cb); | ||
74 | complete_commit(c); | ||
75 | } | ||
76 | |||
77 | static void add_fb(struct msm_commit *c, struct drm_framebuffer *fb) | ||
78 | { | ||
79 | struct drm_gem_object *obj = msm_framebuffer_bo(fb, 0); | ||
80 | c->fence = max(c->fence, msm_gem_fence(to_msm_bo(obj), MSM_PREP_READ)); | ||
81 | } | ||
82 | |||
83 | |||
84 | /** | ||
85 | * drm_atomic_helper_commit - commit validated state object | ||
86 | * @dev: DRM device | ||
87 | * @state: the driver state object | ||
88 | * @async: asynchronous commit | ||
89 | * | ||
90 | * This function commits a with drm_atomic_helper_check() pre-validated state | ||
91 | * object. This can still fail when e.g. the framebuffer reservation fails. For | ||
92 | * now this doesn't implement asynchronous commits. | ||
93 | * | ||
94 | * RETURNS | ||
95 | * Zero for success or -errno. | ||
96 | */ | ||
97 | int msm_atomic_commit(struct drm_device *dev, | ||
98 | struct drm_atomic_state *state, bool async) | ||
99 | { | ||
100 | struct msm_commit *c; | ||
101 | int nplanes = dev->mode_config.num_total_plane; | ||
102 | int i, ret; | ||
103 | |||
104 | ret = drm_atomic_helper_prepare_planes(dev, state); | ||
105 | if (ret) | ||
106 | return ret; | ||
107 | |||
108 | c = new_commit(state); | ||
109 | |||
110 | /* | ||
111 | * Figure out what fence to wait for: | ||
112 | */ | ||
113 | for (i = 0; i < nplanes; i++) { | ||
114 | struct drm_plane *plane = state->planes[i]; | ||
115 | struct drm_plane_state *new_state = state->plane_states[i]; | ||
116 | |||
117 | if (!plane) | ||
118 | continue; | ||
119 | |||
120 | if ((plane->state->fb != new_state->fb) && new_state->fb) | ||
121 | add_fb(c, new_state->fb); | ||
122 | } | ||
123 | |||
124 | /* | ||
125 | * This is the point of no return - everything below never fails except | ||
126 | * when the hw goes bonghits. Which means we can commit the new state on | ||
127 | * the software side now. | ||
128 | */ | ||
129 | |||
130 | drm_atomic_helper_swap_state(dev, state); | ||
131 | |||
132 | /* | ||
133 | * Everything below can be run asynchronously without the need to grab | ||
134 | * any modeset locks at all under one conditions: It must be guaranteed | ||
135 | * that the asynchronous work has either been cancelled (if the driver | ||
136 | * supports it, which at least requires that the framebuffers get | ||
137 | * cleaned up with drm_atomic_helper_cleanup_planes()) or completed | ||
138 | * before the new state gets committed on the software side with | ||
139 | * drm_atomic_helper_swap_state(). | ||
140 | * | ||
141 | * This scheme allows new atomic state updates to be prepared and | ||
142 | * checked in parallel to the asynchronous completion of the previous | ||
143 | * update. Which is important since compositors need to figure out the | ||
144 | * composition of the next frame right after having submitted the | ||
145 | * current layout. | ||
146 | */ | ||
147 | |||
148 | if (async) { | ||
149 | msm_queue_fence_cb(dev, &c->fence_cb, c->fence); | ||
150 | return 0; | ||
151 | } | ||
152 | |||
153 | ret = msm_wait_fence_interruptable(dev, c->fence, NULL); | ||
154 | if (ret) { | ||
155 | WARN_ON(ret); // TODO unswap state back? or?? | ||
156 | kfree(c); | ||
157 | return ret; | ||
158 | } | ||
159 | |||
160 | complete_commit(c); | ||
161 | |||
162 | return 0; | ||
163 | } | ||
diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c index 42e1c48eef28..c795217e1bfc 100644 --- a/drivers/gpu/drm/msm/msm_drv.c +++ b/drivers/gpu/drm/msm/msm_drv.c | |||
@@ -29,6 +29,8 @@ static void msm_fb_output_poll_changed(struct drm_device *dev) | |||
29 | static const struct drm_mode_config_funcs mode_config_funcs = { | 29 | static const struct drm_mode_config_funcs mode_config_funcs = { |
30 | .fb_create = msm_framebuffer_create, | 30 | .fb_create = msm_framebuffer_create, |
31 | .output_poll_changed = msm_fb_output_poll_changed, | 31 | .output_poll_changed = msm_fb_output_poll_changed, |
32 | .atomic_check = drm_atomic_helper_check, | ||
33 | .atomic_commit = msm_atomic_commit, | ||
32 | }; | 34 | }; |
33 | 35 | ||
34 | int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu) | 36 | int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu) |
@@ -294,6 +296,8 @@ static int msm_load(struct drm_device *dev, unsigned long flags) | |||
294 | goto fail; | 296 | goto fail; |
295 | } | 297 | } |
296 | 298 | ||
299 | drm_mode_config_reset(dev); | ||
300 | |||
297 | #ifdef CONFIG_DRM_MSM_FBDEV | 301 | #ifdef CONFIG_DRM_MSM_FBDEV |
298 | priv->fbdev = msm_fbdev_init(dev); | 302 | priv->fbdev = msm_fbdev_init(dev); |
299 | #endif | 303 | #endif |
@@ -619,6 +623,26 @@ int msm_wait_fence_interruptable(struct drm_device *dev, uint32_t fence, | |||
619 | return ret; | 623 | return ret; |
620 | } | 624 | } |
621 | 625 | ||
626 | int msm_queue_fence_cb(struct drm_device *dev, | ||
627 | struct msm_fence_cb *cb, uint32_t fence) | ||
628 | { | ||
629 | struct msm_drm_private *priv = dev->dev_private; | ||
630 | int ret = 0; | ||
631 | |||
632 | mutex_lock(&dev->struct_mutex); | ||
633 | if (!list_empty(&cb->work.entry)) { | ||
634 | ret = -EINVAL; | ||
635 | } else if (fence > priv->completed_fence) { | ||
636 | cb->fence = fence; | ||
637 | list_add_tail(&cb->work.entry, &priv->fence_cbs); | ||
638 | } else { | ||
639 | queue_work(priv->wq, &cb->work); | ||
640 | } | ||
641 | mutex_unlock(&dev->struct_mutex); | ||
642 | |||
643 | return ret; | ||
644 | } | ||
645 | |||
622 | /* called from workqueue */ | 646 | /* called from workqueue */ |
623 | void msm_update_fence(struct drm_device *dev, uint32_t fence) | 647 | void msm_update_fence(struct drm_device *dev, uint32_t fence) |
624 | { | 648 | { |
@@ -832,6 +856,7 @@ static struct drm_driver msm_driver = { | |||
832 | .gem_prime_import_sg_table = msm_gem_prime_import_sg_table, | 856 | .gem_prime_import_sg_table = msm_gem_prime_import_sg_table, |
833 | .gem_prime_vmap = msm_gem_prime_vmap, | 857 | .gem_prime_vmap = msm_gem_prime_vmap, |
834 | .gem_prime_vunmap = msm_gem_prime_vunmap, | 858 | .gem_prime_vunmap = msm_gem_prime_vunmap, |
859 | .gem_prime_mmap = msm_gem_prime_mmap, | ||
835 | #ifdef CONFIG_DEBUG_FS | 860 | #ifdef CONFIG_DEBUG_FS |
836 | .debugfs_init = msm_debugfs_init, | 861 | .debugfs_init = msm_debugfs_init, |
837 | .debugfs_cleanup = msm_debugfs_cleanup, | 862 | .debugfs_cleanup = msm_debugfs_cleanup, |
diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h index 67f9d0a2332c..136303818436 100644 --- a/drivers/gpu/drm/msm/msm_drv.h +++ b/drivers/gpu/drm/msm/msm_drv.h | |||
@@ -32,15 +32,6 @@ | |||
32 | #include <linux/types.h> | 32 | #include <linux/types.h> |
33 | #include <asm/sizes.h> | 33 | #include <asm/sizes.h> |
34 | 34 | ||
35 | |||
36 | #if defined(CONFIG_COMPILE_TEST) && !defined(CONFIG_ARCH_QCOM) | ||
37 | /* stubs we need for compile-test: */ | ||
38 | static inline struct device *msm_iommu_get_ctx(const char *ctx_name) | ||
39 | { | ||
40 | return NULL; | ||
41 | } | ||
42 | #endif | ||
43 | |||
44 | #ifndef CONFIG_OF | 35 | #ifndef CONFIG_OF |
45 | #include <mach/board.h> | 36 | #include <mach/board.h> |
46 | #include <mach/socinfo.h> | 37 | #include <mach/socinfo.h> |
@@ -48,7 +39,10 @@ static inline struct device *msm_iommu_get_ctx(const char *ctx_name) | |||
48 | #endif | 39 | #endif |
49 | 40 | ||
50 | #include <drm/drmP.h> | 41 | #include <drm/drmP.h> |
42 | #include <drm/drm_atomic.h> | ||
43 | #include <drm/drm_atomic_helper.h> | ||
51 | #include <drm/drm_crtc_helper.h> | 44 | #include <drm/drm_crtc_helper.h> |
45 | #include <drm/drm_plane_helper.h> | ||
52 | #include <drm/drm_fb_helper.h> | 46 | #include <drm/drm_fb_helper.h> |
53 | #include <drm/msm_drm.h> | 47 | #include <drm/msm_drm.h> |
54 | #include <drm/drm_gem.h> | 48 | #include <drm/drm_gem.h> |
@@ -75,7 +69,12 @@ struct msm_drm_private { | |||
75 | struct msm_kms *kms; | 69 | struct msm_kms *kms; |
76 | 70 | ||
77 | /* subordinate devices, if present: */ | 71 | /* subordinate devices, if present: */ |
78 | struct platform_device *hdmi_pdev, *gpu_pdev; | 72 | struct platform_device *gpu_pdev; |
73 | |||
74 | /* possibly this should be in the kms component, but it is | ||
75 | * shared by both mdp4 and mdp5.. | ||
76 | */ | ||
77 | struct hdmi *hdmi; | ||
79 | 78 | ||
80 | /* when we have more than one 'msm_gpu' these need to be an array: */ | 79 | /* when we have more than one 'msm_gpu' these need to be an array: */ |
81 | struct msm_gpu *gpu; | 80 | struct msm_gpu *gpu; |
@@ -145,21 +144,29 @@ void __msm_fence_worker(struct work_struct *work); | |||
145 | (_cb)->func = _func; \ | 144 | (_cb)->func = _func; \ |
146 | } while (0) | 145 | } while (0) |
147 | 146 | ||
147 | int msm_atomic_commit(struct drm_device *dev, | ||
148 | struct drm_atomic_state *state, bool async); | ||
149 | |||
148 | int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu); | 150 | int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu); |
149 | 151 | ||
150 | int msm_wait_fence_interruptable(struct drm_device *dev, uint32_t fence, | 152 | int msm_wait_fence_interruptable(struct drm_device *dev, uint32_t fence, |
151 | struct timespec *timeout); | 153 | struct timespec *timeout); |
154 | int msm_queue_fence_cb(struct drm_device *dev, | ||
155 | struct msm_fence_cb *cb, uint32_t fence); | ||
152 | void msm_update_fence(struct drm_device *dev, uint32_t fence); | 156 | void msm_update_fence(struct drm_device *dev, uint32_t fence); |
153 | 157 | ||
154 | int msm_ioctl_gem_submit(struct drm_device *dev, void *data, | 158 | int msm_ioctl_gem_submit(struct drm_device *dev, void *data, |
155 | struct drm_file *file); | 159 | struct drm_file *file); |
156 | 160 | ||
161 | int msm_gem_mmap_obj(struct drm_gem_object *obj, | ||
162 | struct vm_area_struct *vma); | ||
157 | int msm_gem_mmap(struct file *filp, struct vm_area_struct *vma); | 163 | int msm_gem_mmap(struct file *filp, struct vm_area_struct *vma); |
158 | int msm_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); | 164 | int msm_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); |
159 | uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj); | 165 | uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj); |
160 | int msm_gem_get_iova_locked(struct drm_gem_object *obj, int id, | 166 | int msm_gem_get_iova_locked(struct drm_gem_object *obj, int id, |
161 | uint32_t *iova); | 167 | uint32_t *iova); |
162 | int msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova); | 168 | int msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova); |
169 | uint32_t msm_gem_iova(struct drm_gem_object *obj, int id); | ||
163 | struct page **msm_gem_get_pages(struct drm_gem_object *obj); | 170 | struct page **msm_gem_get_pages(struct drm_gem_object *obj); |
164 | void msm_gem_put_pages(struct drm_gem_object *obj); | 171 | void msm_gem_put_pages(struct drm_gem_object *obj); |
165 | void msm_gem_put_iova(struct drm_gem_object *obj, int id); | 172 | void msm_gem_put_iova(struct drm_gem_object *obj, int id); |
@@ -170,6 +177,7 @@ int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, | |||
170 | struct sg_table *msm_gem_prime_get_sg_table(struct drm_gem_object *obj); | 177 | struct sg_table *msm_gem_prime_get_sg_table(struct drm_gem_object *obj); |
171 | void *msm_gem_prime_vmap(struct drm_gem_object *obj); | 178 | void *msm_gem_prime_vmap(struct drm_gem_object *obj); |
172 | void msm_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr); | 179 | void msm_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr); |
180 | int msm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); | ||
173 | struct drm_gem_object *msm_gem_prime_import_sg_table(struct drm_device *dev, | 181 | struct drm_gem_object *msm_gem_prime_import_sg_table(struct drm_device *dev, |
174 | struct dma_buf_attachment *attach, struct sg_table *sg); | 182 | struct dma_buf_attachment *attach, struct sg_table *sg); |
175 | int msm_gem_prime_pin(struct drm_gem_object *obj); | 183 | int msm_gem_prime_pin(struct drm_gem_object *obj); |
@@ -192,6 +200,9 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev, | |||
192 | struct drm_gem_object *msm_gem_import(struct drm_device *dev, | 200 | struct drm_gem_object *msm_gem_import(struct drm_device *dev, |
193 | uint32_t size, struct sg_table *sgt); | 201 | uint32_t size, struct sg_table *sgt); |
194 | 202 | ||
203 | int msm_framebuffer_prepare(struct drm_framebuffer *fb, int id); | ||
204 | void msm_framebuffer_cleanup(struct drm_framebuffer *fb, int id); | ||
205 | uint32_t msm_framebuffer_iova(struct drm_framebuffer *fb, int id, int plane); | ||
195 | struct drm_gem_object *msm_framebuffer_bo(struct drm_framebuffer *fb, int plane); | 206 | struct drm_gem_object *msm_framebuffer_bo(struct drm_framebuffer *fb, int plane); |
196 | const struct msm_format *msm_framebuffer_format(struct drm_framebuffer *fb); | 207 | const struct msm_format *msm_framebuffer_format(struct drm_framebuffer *fb); |
197 | struct drm_framebuffer *msm_framebuffer_init(struct drm_device *dev, | 208 | struct drm_framebuffer *msm_framebuffer_init(struct drm_device *dev, |
@@ -202,8 +213,8 @@ struct drm_framebuffer *msm_framebuffer_create(struct drm_device *dev, | |||
202 | struct drm_fb_helper *msm_fbdev_init(struct drm_device *dev); | 213 | struct drm_fb_helper *msm_fbdev_init(struct drm_device *dev); |
203 | 214 | ||
204 | struct hdmi; | 215 | struct hdmi; |
205 | struct hdmi *hdmi_init(struct drm_device *dev, struct drm_encoder *encoder); | 216 | int hdmi_modeset_init(struct hdmi *hdmi, struct drm_device *dev, |
206 | irqreturn_t hdmi_irq(int irq, void *dev_id); | 217 | struct drm_encoder *encoder); |
207 | void __init hdmi_register(void); | 218 | void __init hdmi_register(void); |
208 | void __exit hdmi_unregister(void); | 219 | void __exit hdmi_unregister(void); |
209 | 220 | ||
diff --git a/drivers/gpu/drm/msm/msm_fb.c b/drivers/gpu/drm/msm/msm_fb.c index 81bafdf19ab3..84dec161d836 100644 --- a/drivers/gpu/drm/msm/msm_fb.c +++ b/drivers/gpu/drm/msm/msm_fb.c | |||
@@ -24,7 +24,7 @@ | |||
24 | struct msm_framebuffer { | 24 | struct msm_framebuffer { |
25 | struct drm_framebuffer base; | 25 | struct drm_framebuffer base; |
26 | const struct msm_format *format; | 26 | const struct msm_format *format; |
27 | struct drm_gem_object *planes[2]; | 27 | struct drm_gem_object *planes[3]; |
28 | }; | 28 | }; |
29 | #define to_msm_framebuffer(x) container_of(x, struct msm_framebuffer, base) | 29 | #define to_msm_framebuffer(x) container_of(x, struct msm_framebuffer, base) |
30 | 30 | ||
@@ -87,6 +87,44 @@ void msm_framebuffer_describe(struct drm_framebuffer *fb, struct seq_file *m) | |||
87 | } | 87 | } |
88 | #endif | 88 | #endif |
89 | 89 | ||
90 | /* prepare/pin all the fb's bo's for scanout. Note that it is not valid | ||
91 | * to prepare an fb more multiple different initiator 'id's. But that | ||
92 | * should be fine, since only the scanout (mdpN) side of things needs | ||
93 | * this, the gpu doesn't care about fb's. | ||
94 | */ | ||
95 | int msm_framebuffer_prepare(struct drm_framebuffer *fb, int id) | ||
96 | { | ||
97 | struct msm_framebuffer *msm_fb = to_msm_framebuffer(fb); | ||
98 | int ret, i, n = drm_format_num_planes(fb->pixel_format); | ||
99 | uint32_t iova; | ||
100 | |||
101 | for (i = 0; i < n; i++) { | ||
102 | ret = msm_gem_get_iova(msm_fb->planes[i], id, &iova); | ||
103 | DBG("FB[%u]: iova[%d]: %08x (%d)", fb->base.id, i, iova, ret); | ||
104 | if (ret) | ||
105 | return ret; | ||
106 | } | ||
107 | |||
108 | return 0; | ||
109 | } | ||
110 | |||
111 | void msm_framebuffer_cleanup(struct drm_framebuffer *fb, int id) | ||
112 | { | ||
113 | struct msm_framebuffer *msm_fb = to_msm_framebuffer(fb); | ||
114 | int i, n = drm_format_num_planes(fb->pixel_format); | ||
115 | |||
116 | for (i = 0; i < n; i++) | ||
117 | msm_gem_put_iova(msm_fb->planes[i], id); | ||
118 | } | ||
119 | |||
120 | uint32_t msm_framebuffer_iova(struct drm_framebuffer *fb, int id, int plane) | ||
121 | { | ||
122 | struct msm_framebuffer *msm_fb = to_msm_framebuffer(fb); | ||
123 | if (!msm_fb->planes[plane]) | ||
124 | return 0; | ||
125 | return msm_gem_iova(msm_fb->planes[plane], id); | ||
126 | } | ||
127 | |||
90 | struct drm_gem_object *msm_framebuffer_bo(struct drm_framebuffer *fb, int plane) | 128 | struct drm_gem_object *msm_framebuffer_bo(struct drm_framebuffer *fb, int plane) |
91 | { | 129 | { |
92 | struct msm_framebuffer *msm_fb = to_msm_framebuffer(fb); | 130 | struct msm_framebuffer *msm_fb = to_msm_framebuffer(fb); |
@@ -166,6 +204,11 @@ struct drm_framebuffer *msm_framebuffer_init(struct drm_device *dev, | |||
166 | 204 | ||
167 | msm_fb->format = format; | 205 | msm_fb->format = format; |
168 | 206 | ||
207 | if (n > ARRAY_SIZE(msm_fb->planes)) { | ||
208 | ret = -EINVAL; | ||
209 | goto fail; | ||
210 | } | ||
211 | |||
169 | for (i = 0; i < n; i++) { | 212 | for (i = 0; i < n; i++) { |
170 | unsigned int width = mode_cmd->width / (i ? hsub : 1); | 213 | unsigned int width = mode_cmd->width / (i ? hsub : 1); |
171 | unsigned int height = mode_cmd->height / (i ? vsub : 1); | 214 | unsigned int height = mode_cmd->height / (i ? vsub : 1); |
diff --git a/drivers/gpu/drm/msm/msm_fbdev.c b/drivers/gpu/drm/msm/msm_fbdev.c index ab5bfd2d0ebf..94d55e526b4e 100644 --- a/drivers/gpu/drm/msm/msm_fbdev.c +++ b/drivers/gpu/drm/msm/msm_fbdev.c | |||
@@ -93,9 +93,6 @@ static int msm_fbdev_create(struct drm_fb_helper *helper, | |||
93 | uint32_t paddr; | 93 | uint32_t paddr; |
94 | int ret, size; | 94 | int ret, size; |
95 | 95 | ||
96 | sizes->surface_bpp = 32; | ||
97 | sizes->surface_depth = 24; | ||
98 | |||
99 | DBG("create fbdev: %dx%d@%d (%dx%d)", sizes->surface_width, | 96 | DBG("create fbdev: %dx%d@%d (%dx%d)", sizes->surface_width, |
100 | sizes->surface_height, sizes->surface_bpp, | 97 | sizes->surface_height, sizes->surface_bpp, |
101 | sizes->fb_width, sizes->fb_height); | 98 | sizes->fb_width, sizes->fb_height); |
diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c index 4b1b82adabde..4a6f0e49d5b5 100644 --- a/drivers/gpu/drm/msm/msm_gem.c +++ b/drivers/gpu/drm/msm/msm_gem.c | |||
@@ -309,6 +309,7 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj, int id, | |||
309 | return ret; | 309 | return ret; |
310 | } | 310 | } |
311 | 311 | ||
312 | /* get iova, taking a reference. Should have a matching put */ | ||
312 | int msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova) | 313 | int msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova) |
313 | { | 314 | { |
314 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 315 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
@@ -328,6 +329,16 @@ int msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova) | |||
328 | return ret; | 329 | return ret; |
329 | } | 330 | } |
330 | 331 | ||
332 | /* get iova without taking a reference, used in places where you have | ||
333 | * already done a 'msm_gem_get_iova()'. | ||
334 | */ | ||
335 | uint32_t msm_gem_iova(struct drm_gem_object *obj, int id) | ||
336 | { | ||
337 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | ||
338 | WARN_ON(!msm_obj->domain[id].iova); | ||
339 | return msm_obj->domain[id].iova; | ||
340 | } | ||
341 | |||
331 | void msm_gem_put_iova(struct drm_gem_object *obj, int id) | 342 | void msm_gem_put_iova(struct drm_gem_object *obj, int id) |
332 | { | 343 | { |
333 | // XXX TODO .. | 344 | // XXX TODO .. |
@@ -397,23 +408,10 @@ void *msm_gem_vaddr(struct drm_gem_object *obj) | |||
397 | int msm_gem_queue_inactive_cb(struct drm_gem_object *obj, | 408 | int msm_gem_queue_inactive_cb(struct drm_gem_object *obj, |
398 | struct msm_fence_cb *cb) | 409 | struct msm_fence_cb *cb) |
399 | { | 410 | { |
400 | struct drm_device *dev = obj->dev; | ||
401 | struct msm_drm_private *priv = dev->dev_private; | ||
402 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 411 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
403 | int ret = 0; | 412 | uint32_t fence = msm_gem_fence(msm_obj, |
404 | 413 | MSM_PREP_READ | MSM_PREP_WRITE); | |
405 | mutex_lock(&dev->struct_mutex); | 414 | return msm_queue_fence_cb(obj->dev, cb, fence); |
406 | if (!list_empty(&cb->work.entry)) { | ||
407 | ret = -EINVAL; | ||
408 | } else if (is_active(msm_obj)) { | ||
409 | cb->fence = max(msm_obj->read_fence, msm_obj->write_fence); | ||
410 | list_add_tail(&cb->work.entry, &priv->fence_cbs); | ||
411 | } else { | ||
412 | queue_work(priv->wq, &cb->work); | ||
413 | } | ||
414 | mutex_unlock(&dev->struct_mutex); | ||
415 | |||
416 | return ret; | ||
417 | } | 415 | } |
418 | 416 | ||
419 | void msm_gem_move_to_active(struct drm_gem_object *obj, | 417 | void msm_gem_move_to_active(struct drm_gem_object *obj, |
@@ -452,12 +450,8 @@ int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, | |||
452 | int ret = 0; | 450 | int ret = 0; |
453 | 451 | ||
454 | if (is_active(msm_obj)) { | 452 | if (is_active(msm_obj)) { |
455 | uint32_t fence = 0; | 453 | uint32_t fence = msm_gem_fence(msm_obj, op); |
456 | 454 | ||
457 | if (op & MSM_PREP_READ) | ||
458 | fence = msm_obj->write_fence; | ||
459 | if (op & MSM_PREP_WRITE) | ||
460 | fence = max(fence, msm_obj->read_fence); | ||
461 | if (op & MSM_PREP_NOSYNC) | 455 | if (op & MSM_PREP_NOSYNC) |
462 | timeout = NULL; | 456 | timeout = NULL; |
463 | 457 | ||
@@ -525,13 +519,11 @@ void msm_gem_free_object(struct drm_gem_object *obj) | |||
525 | for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) { | 519 | for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) { |
526 | struct msm_mmu *mmu = priv->mmus[id]; | 520 | struct msm_mmu *mmu = priv->mmus[id]; |
527 | if (mmu && msm_obj->domain[id].iova) { | 521 | if (mmu && msm_obj->domain[id].iova) { |
528 | uint32_t offset = (uint32_t)mmap_offset(obj); | 522 | uint32_t offset = msm_obj->domain[id].iova; |
529 | mmu->funcs->unmap(mmu, offset, msm_obj->sgt, obj->size); | 523 | mmu->funcs->unmap(mmu, offset, msm_obj->sgt, obj->size); |
530 | } | 524 | } |
531 | } | 525 | } |
532 | 526 | ||
533 | drm_gem_free_mmap_offset(obj); | ||
534 | |||
535 | if (obj->import_attach) { | 527 | if (obj->import_attach) { |
536 | if (msm_obj->vaddr) | 528 | if (msm_obj->vaddr) |
537 | dma_buf_vunmap(obj->import_attach->dmabuf, msm_obj->vaddr); | 529 | dma_buf_vunmap(obj->import_attach->dmabuf, msm_obj->vaddr); |
diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h index bfb052688f8e..8fbbd0594c46 100644 --- a/drivers/gpu/drm/msm/msm_gem.h +++ b/drivers/gpu/drm/msm/msm_gem.h | |||
@@ -70,6 +70,19 @@ static inline bool is_active(struct msm_gem_object *msm_obj) | |||
70 | return msm_obj->gpu != NULL; | 70 | return msm_obj->gpu != NULL; |
71 | } | 71 | } |
72 | 72 | ||
73 | static inline uint32_t msm_gem_fence(struct msm_gem_object *msm_obj, | ||
74 | uint32_t op) | ||
75 | { | ||
76 | uint32_t fence = 0; | ||
77 | |||
78 | if (op & MSM_PREP_READ) | ||
79 | fence = msm_obj->write_fence; | ||
80 | if (op & MSM_PREP_WRITE) | ||
81 | fence = max(fence, msm_obj->read_fence); | ||
82 | |||
83 | return fence; | ||
84 | } | ||
85 | |||
73 | #define MAX_CMDS 4 | 86 | #define MAX_CMDS 4 |
74 | 87 | ||
75 | /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, | 88 | /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, |
diff --git a/drivers/gpu/drm/msm/msm_gem_prime.c b/drivers/gpu/drm/msm/msm_gem_prime.c index ad772fe36115..dd7a7ab603e2 100644 --- a/drivers/gpu/drm/msm/msm_gem_prime.c +++ b/drivers/gpu/drm/msm/msm_gem_prime.c | |||
@@ -37,6 +37,19 @@ void msm_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr) | |||
37 | /* TODO msm_gem_vunmap() */ | 37 | /* TODO msm_gem_vunmap() */ |
38 | } | 38 | } |
39 | 39 | ||
40 | int msm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) | ||
41 | { | ||
42 | int ret; | ||
43 | |||
44 | mutex_lock(&obj->dev->struct_mutex); | ||
45 | ret = drm_gem_mmap_obj(obj, obj->size, vma); | ||
46 | mutex_unlock(&obj->dev->struct_mutex); | ||
47 | if (ret < 0) | ||
48 | return ret; | ||
49 | |||
50 | return msm_gem_mmap_obj(vma->vm_private_data, vma); | ||
51 | } | ||
52 | |||
40 | struct drm_gem_object *msm_gem_prime_import_sg_table(struct drm_device *dev, | 53 | struct drm_gem_object *msm_gem_prime_import_sg_table(struct drm_device *dev, |
41 | struct dma_buf_attachment *attach, struct sg_table *sg) | 54 | struct dma_buf_attachment *attach, struct sg_table *sg) |
42 | { | 55 | { |