Home | History | Annotate | Line # | Download | only in amdgpu
      1 /*	$NetBSD: amdgpu_mxgpu_vi.c,v 1.2 2021/12/18 23:44:58 riastradh Exp $	*/
      2 
      3 /*
      4  * Copyright 2017 Advanced Micro Devices, Inc.
      5  *
      6  * Permission is hereby granted, free of charge, to any person obtaining a
      7  * copy of this software and associated documentation files (the "Software"),
      8  * to deal in the Software without restriction, including without limitation
      9  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
     10  * and/or sell copies of the Software, and to permit persons to whom the
     11  * Software is furnished to do so, subject to the following conditions:
     12  *
     13  * The above copyright notice and this permission notice shall be included in
     14  * all copies or substantial portions of the Software.
     15  *
     16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     19  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     20  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     21  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     22  * OTHER DEALINGS IN THE SOFTWARE.
     23  *
     24  * Authors: Xiangliang.Yu (at) amd.com
     25  */
     26 
     27 #include <sys/cdefs.h>
     28 __KERNEL_RCSID(0, "$NetBSD: amdgpu_mxgpu_vi.c,v 1.2 2021/12/18 23:44:58 riastradh Exp $");
     29 
     30 #include "amdgpu.h"
     31 #include "vi.h"
     32 #include "bif/bif_5_0_d.h"
     33 #include "bif/bif_5_0_sh_mask.h"
     34 #include "vid.h"
     35 #include "gca/gfx_8_0_d.h"
     36 #include "gca/gfx_8_0_sh_mask.h"
     37 #include "gmc_v8_0.h"
     38 #include "gfx_v8_0.h"
     39 #include "sdma_v3_0.h"
     40 #include "tonga_ih.h"
     41 #include "gmc/gmc_8_2_d.h"
     42 #include "gmc/gmc_8_2_sh_mask.h"
     43 #include "oss/oss_3_0_d.h"
     44 #include "oss/oss_3_0_sh_mask.h"
     45 #include "dce/dce_10_0_d.h"
     46 #include "dce/dce_10_0_sh_mask.h"
     47 #include "smu/smu_7_1_3_d.h"
     48 #include "mxgpu_vi.h"
     49 
     50 /* VI golden setting */
     51 static const u32 xgpu_fiji_mgcg_cgcg_init[] = {
     52 	mmRLC_CGTT_MGCG_OVERRIDE, 0xffffffff, 0xffffffff,
     53 	mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000,
     54 	mmCB_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100,
     55 	mmCGTT_BCI_CLK_CTRL, 0xffffffff, 0x00000100,
     56 	mmCGTT_CP_CLK_CTRL, 0xffffffff, 0x00000100,
     57 	mmCGTT_CPC_CLK_CTRL, 0xffffffff, 0x00000100,
     58 	mmCGTT_CPF_CLK_CTRL, 0xffffffff, 0x40000100,
     59 	mmCGTT_DRM_CLK_CTRL0, 0xffffffff, 0x00600100,
     60 	mmCGTT_GDS_CLK_CTRL, 0xffffffff, 0x00000100,
     61 	mmCGTT_IA_CLK_CTRL, 0xffffffff, 0x06000100,
     62 	mmCGTT_PA_CLK_CTRL, 0xffffffff, 0x00000100,
     63 	mmCGTT_WD_CLK_CTRL, 0xffffffff, 0x06000100,
     64 	mmCGTT_PC_CLK_CTRL, 0xffffffff, 0x00000100,
     65 	mmCGTT_RLC_CLK_CTRL, 0xffffffff, 0x00000100,
     66 	mmCGTT_SC_CLK_CTRL, 0xffffffff, 0x00000100,
     67 	mmCGTT_SPI_CLK_CTRL, 0xffffffff, 0x00000100,
     68 	mmCGTT_SQ_CLK_CTRL, 0xffffffff, 0x00000100,
     69 	mmCGTT_SQG_CLK_CTRL, 0xffffffff, 0x00000100,
     70 	mmCGTT_SX_CLK_CTRL0, 0xffffffff, 0x00000100,
     71 	mmCGTT_SX_CLK_CTRL1, 0xffffffff, 0x00000100,
     72 	mmCGTT_SX_CLK_CTRL2, 0xffffffff, 0x00000100,
     73 	mmCGTT_SX_CLK_CTRL3, 0xffffffff, 0x00000100,
     74 	mmCGTT_SX_CLK_CTRL4, 0xffffffff, 0x00000100,
     75 	mmCGTT_TCI_CLK_CTRL, 0xffffffff, 0x00000100,
     76 	mmCGTT_TCP_CLK_CTRL, 0xffffffff, 0x00000100,
     77 	mmCGTT_VGT_CLK_CTRL, 0xffffffff, 0x06000100,
     78 	mmDB_CGTT_CLK_CTRL_0, 0xffffffff, 0x00000100,
     79 	mmTA_CGTT_CTRL, 0xffffffff, 0x00000100,
     80 	mmTCA_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100,
     81 	mmTCC_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100,
     82 	mmTD_CGTT_CTRL, 0xffffffff, 0x00000100,
     83 	mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000,
     84 	mmCGTS_SM_CTRL_REG, 0xffffffff, 0x96e00200,
     85 	mmCP_RB_WPTR_POLL_CNTL, 0xffffffff, 0x00900100,
     86 	mmRLC_CGCG_CGLS_CTRL, 0xffffffff, 0x0020003c,
     87 	mmPCIE_INDEX, 0xffffffff, 0x0140001c,
     88 	mmPCIE_DATA, 0x000f0000, 0x00000000,
     89 	mmSMC_IND_INDEX_4, 0xffffffff, 0xC060000C,
     90 	mmSMC_IND_DATA_4, 0xc0000fff, 0x00000100,
     91 	mmXDMA_CLOCK_GATING_CNTL, 0xffffffff, 0x00000100,
     92 	mmXDMA_MEM_POWER_CNTL, 0x00000101, 0x00000000,
     93 	mmMC_MEM_POWER_LS, 0xffffffff, 0x00000104,
     94 	mmCGTT_DRM_CLK_CTRL0, 0xff000fff, 0x00000100,
     95 	mmHDP_XDP_CGTT_BLK_CTRL, 0xc0000fff, 0x00000104,
     96 	mmCP_MEM_SLP_CNTL, 0x00000001, 0x00000001,
     97 	mmSDMA0_CLK_CTRL, 0xff000ff0, 0x00000100,
     98 	mmSDMA1_CLK_CTRL, 0xff000ff0, 0x00000100,
     99 };
    100 
    101 static const u32 xgpu_fiji_golden_settings_a10[] = {
    102 	mmCB_HW_CONTROL_3, 0x000001ff, 0x00000040,
    103 	mmDB_DEBUG2, 0xf00fffff, 0x00000400,
    104 	mmDCI_CLK_CNTL, 0x00000080, 0x00000000,
    105 	mmFBC_DEBUG_COMP, 0x000000f0, 0x00000070,
    106 	mmFBC_MISC, 0x1f311fff, 0x12300000,
    107 	mmHDMI_CONTROL, 0x31000111, 0x00000011,
    108 	mmPA_SC_ENHANCE, 0xffffffff, 0x20000001,
    109 	mmPA_SC_LINE_STIPPLE_STATE, 0x0000ff0f, 0x00000000,
    110 	mmSDMA0_CHICKEN_BITS, 0xfc910007, 0x00810007,
    111 	mmSDMA0_GFX_IB_CNTL, 0x800f0111, 0x00000100,
    112 	mmSDMA0_RLC0_IB_CNTL, 0x800f0111, 0x00000100,
    113 	mmSDMA0_RLC1_IB_CNTL, 0x800f0111, 0x00000100,
    114 	mmSDMA1_CHICKEN_BITS, 0xfc910007, 0x00810007,
    115 	mmSDMA1_GFX_IB_CNTL, 0x800f0111, 0x00000100,
    116 	mmSDMA1_RLC0_IB_CNTL, 0x800f0111, 0x00000100,
    117 	mmSDMA1_RLC1_IB_CNTL, 0x800f0111, 0x00000100,
    118 	mmSQ_RANDOM_WAVE_PRI, 0x001fffff, 0x000006fd,
    119 	mmTA_CNTL_AUX, 0x000f000f, 0x000b0000,
    120 	mmTCC_EXE_DISABLE, 0x00000002, 0x00000002,
    121 	mmTCP_ADDR_CONFIG, 0x000003ff, 0x000000ff,
    122 	mmVGT_RESET_DEBUG, 0x00000004, 0x00000004,
    123 	mmVM_PRT_APERTURE0_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    124 	mmVM_PRT_APERTURE1_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    125 	mmVM_PRT_APERTURE2_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    126 	mmVM_PRT_APERTURE3_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    127 };
    128 
    129 static const u32 xgpu_fiji_golden_common_all[] = {
    130 	mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000,
    131 	mmPA_SC_RASTER_CONFIG, 0xffffffff, 0x3a00161a,
    132 	mmPA_SC_RASTER_CONFIG_1, 0xffffffff, 0x0000002e,
    133 	mmGB_ADDR_CONFIG, 0xffffffff, 0x22011003,
    134 	mmSPI_RESOURCE_RESERVE_CU_0, 0xffffffff, 0x00000800,
    135 	mmSPI_RESOURCE_RESERVE_CU_1, 0xffffffff, 0x00000800,
    136 	mmSPI_RESOURCE_RESERVE_EN_CU_0, 0xffffffff, 0x00007FBF,
    137 	mmSPI_RESOURCE_RESERVE_EN_CU_1, 0xffffffff, 0x00007FAF,
    138 	mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000,
    139 	mmSPI_CONFIG_CNTL_1, 0x0000000f, 0x00000009,
    140 };
    141 
    142 static const u32 xgpu_tonga_mgcg_cgcg_init[] = {
    143 	mmRLC_CGTT_MGCG_OVERRIDE,   0xffffffff, 0xffffffff,
    144 	mmGRBM_GFX_INDEX,           0xffffffff, 0xe0000000,
    145 	mmCB_CGTT_SCLK_CTRL,        0xffffffff, 0x00000100,
    146 	mmCGTT_BCI_CLK_CTRL,        0xffffffff, 0x00000100,
    147 	mmCGTT_CP_CLK_CTRL,         0xffffffff, 0x00000100,
    148 	mmCGTT_CPC_CLK_CTRL,        0xffffffff, 0x00000100,
    149 	mmCGTT_CPF_CLK_CTRL,        0xffffffff, 0x40000100,
    150 	mmCGTT_DRM_CLK_CTRL0,       0xffffffff, 0x00600100,
    151 	mmCGTT_GDS_CLK_CTRL,        0xffffffff, 0x00000100,
    152 	mmCGTT_IA_CLK_CTRL,         0xffffffff, 0x06000100,
    153 	mmCGTT_PA_CLK_CTRL,         0xffffffff, 0x00000100,
    154 	mmCGTT_WD_CLK_CTRL,         0xffffffff, 0x06000100,
    155 	mmCGTT_PC_CLK_CTRL,         0xffffffff, 0x00000100,
    156 	mmCGTT_RLC_CLK_CTRL,        0xffffffff, 0x00000100,
    157 	mmCGTT_SC_CLK_CTRL,         0xffffffff, 0x00000100,
    158 	mmCGTT_SPI_CLK_CTRL,        0xffffffff, 0x00000100,
    159 	mmCGTT_SQ_CLK_CTRL,         0xffffffff, 0x00000100,
    160 	mmCGTT_SQG_CLK_CTRL,        0xffffffff, 0x00000100,
    161 	mmCGTT_SX_CLK_CTRL0,        0xffffffff, 0x00000100,
    162 	mmCGTT_SX_CLK_CTRL1,        0xffffffff, 0x00000100,
    163 	mmCGTT_SX_CLK_CTRL2,        0xffffffff, 0x00000100,
    164 	mmCGTT_SX_CLK_CTRL3,        0xffffffff, 0x00000100,
    165 	mmCGTT_SX_CLK_CTRL4,        0xffffffff, 0x00000100,
    166 	mmCGTT_TCI_CLK_CTRL,        0xffffffff, 0x00000100,
    167 	mmCGTT_TCP_CLK_CTRL,        0xffffffff, 0x00000100,
    168 	mmCGTT_VGT_CLK_CTRL,        0xffffffff, 0x06000100,
    169 	mmDB_CGTT_CLK_CTRL_0,       0xffffffff, 0x00000100,
    170 	mmTA_CGTT_CTRL,             0xffffffff, 0x00000100,
    171 	mmTCA_CGTT_SCLK_CTRL,       0xffffffff, 0x00000100,
    172 	mmTCC_CGTT_SCLK_CTRL,       0xffffffff, 0x00000100,
    173 	mmTD_CGTT_CTRL,             0xffffffff, 0x00000100,
    174 	mmGRBM_GFX_INDEX,           0xffffffff, 0xe0000000,
    175 	mmCGTS_CU0_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    176 	mmCGTS_CU0_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    177 	mmCGTS_CU0_TA_SQC_CTRL_REG, 0xffffffff, 0x00040007,
    178 	mmCGTS_CU0_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    179 	mmCGTS_CU0_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    180 	mmCGTS_CU1_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    181 	mmCGTS_CU1_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    182 	mmCGTS_CU1_TA_CTRL_REG,     0xffffffff, 0x00040007,
    183 	mmCGTS_CU1_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    184 	mmCGTS_CU1_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    185 	mmCGTS_CU2_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    186 	mmCGTS_CU2_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    187 	mmCGTS_CU2_TA_CTRL_REG,     0xffffffff, 0x00040007,
    188 	mmCGTS_CU2_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    189 	mmCGTS_CU2_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    190 	mmCGTS_CU3_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    191 	mmCGTS_CU3_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    192 	mmCGTS_CU3_TA_CTRL_REG,     0xffffffff, 0x00040007,
    193 	mmCGTS_CU3_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    194 	mmCGTS_CU3_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    195 	mmCGTS_CU4_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    196 	mmCGTS_CU4_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    197 	mmCGTS_CU4_TA_SQC_CTRL_REG, 0xffffffff, 0x00040007,
    198 	mmCGTS_CU4_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    199 	mmCGTS_CU4_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    200 	mmCGTS_CU5_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    201 	mmCGTS_CU5_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    202 	mmCGTS_CU5_TA_CTRL_REG,     0xffffffff, 0x00040007,
    203 	mmCGTS_CU5_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    204 	mmCGTS_CU5_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    205 	mmCGTS_CU6_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    206 	mmCGTS_CU6_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    207 	mmCGTS_CU6_TA_CTRL_REG,     0xffffffff, 0x00040007,
    208 	mmCGTS_CU6_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    209 	mmCGTS_CU6_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    210 	mmCGTS_CU7_SP0_CTRL_REG,    0xffffffff, 0x00010000,
    211 	mmCGTS_CU7_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002,
    212 	mmCGTS_CU7_TA_CTRL_REG,     0xffffffff, 0x00040007,
    213 	mmCGTS_CU7_SP1_CTRL_REG,    0xffffffff, 0x00060005,
    214 	mmCGTS_CU7_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008,
    215 	mmCGTS_SM_CTRL_REG,         0xffffffff, 0x96e00200,
    216 	mmCP_RB_WPTR_POLL_CNTL,     0xffffffff, 0x00900100,
    217 	mmRLC_CGCG_CGLS_CTRL,       0xffffffff, 0x0020003c,
    218 	mmPCIE_INDEX,               0xffffffff, 0x0140001c,
    219 	mmPCIE_DATA,                0x000f0000, 0x00000000,
    220 	mmSMC_IND_INDEX_4,          0xffffffff, 0xC060000C,
    221 	mmSMC_IND_DATA_4,           0xc0000fff, 0x00000100,
    222 	mmXDMA_CLOCK_GATING_CNTL,   0xffffffff, 0x00000100,
    223 	mmXDMA_MEM_POWER_CNTL,      0x00000101, 0x00000000,
    224 	mmMC_MEM_POWER_LS,          0xffffffff, 0x00000104,
    225 	mmCGTT_DRM_CLK_CTRL0,       0xff000fff, 0x00000100,
    226 	mmHDP_XDP_CGTT_BLK_CTRL,    0xc0000fff, 0x00000104,
    227 	mmCP_MEM_SLP_CNTL,          0x00000001, 0x00000001,
    228 	mmSDMA0_CLK_CTRL,           0xff000ff0, 0x00000100,
    229 	mmSDMA1_CLK_CTRL,           0xff000ff0, 0x00000100,
    230 };
    231 
    232 static const u32 xgpu_tonga_golden_settings_a11[] = {
    233 	mmCB_HW_CONTROL, 0xfffdf3cf, 0x00007208,
    234 	mmCB_HW_CONTROL_3, 0x00000040, 0x00000040,
    235 	mmDB_DEBUG2, 0xf00fffff, 0x00000400,
    236 	mmDCI_CLK_CNTL, 0x00000080, 0x00000000,
    237 	mmFBC_DEBUG_COMP, 0x000000f0, 0x00000070,
    238 	mmFBC_MISC, 0x1f311fff, 0x12300000,
    239 	mmGB_GPU_ID, 0x0000000f, 0x00000000,
    240 	mmHDMI_CONTROL, 0x31000111, 0x00000011,
    241 	mmMC_ARB_WTM_GRPWT_RD, 0x00000003, 0x00000000,
    242 	mmMC_HUB_RDREQ_DMIF_LIMIT, 0x0000007f, 0x00000028,
    243 	mmMC_HUB_WDP_UMC, 0x00007fb6, 0x00000991,
    244 	mmPA_SC_ENHANCE, 0xffffffff, 0x20000001,
    245 	mmPA_SC_FIFO_DEPTH_CNTL, 0x000003ff, 0x000000fc,
    246 	mmPA_SC_LINE_STIPPLE_STATE, 0x0000ff0f, 0x00000000,
    247 	mmRLC_CGCG_CGLS_CTRL, 0x00000003, 0x0000003c,
    248 	mmSDMA0_CHICKEN_BITS, 0xfc910007, 0x00810007,
    249 	mmSDMA0_CLK_CTRL, 0xff000fff, 0x00000000,
    250 	mmSDMA0_GFX_IB_CNTL, 0x800f0111, 0x00000100,
    251 	mmSDMA0_RLC0_IB_CNTL, 0x800f0111, 0x00000100,
    252 	mmSDMA0_RLC1_IB_CNTL, 0x800f0111, 0x00000100,
    253 	mmSDMA1_CHICKEN_BITS, 0xfc910007, 0x00810007,
    254 	mmSDMA1_CLK_CTRL, 0xff000fff, 0x00000000,
    255 	mmSDMA1_GFX_IB_CNTL, 0x800f0111, 0x00000100,
    256 	mmSDMA1_RLC0_IB_CNTL, 0x800f0111, 0x00000100,
    257 	mmSDMA1_RLC1_IB_CNTL, 0x800f0111, 0x00000100,
    258 	mmSQ_RANDOM_WAVE_PRI, 0x001fffff, 0x000006fd,
    259 	mmTA_CNTL_AUX, 0x000f000f, 0x000b0000,
    260 	mmTCC_CTRL, 0x00100000, 0xf31fff7f,
    261 	mmTCC_EXE_DISABLE, 0x00000002, 0x00000002,
    262 	mmTCP_ADDR_CONFIG, 0x000003ff, 0x000002fb,
    263 	mmTCP_CHAN_STEER_HI, 0xffffffff, 0x0000543b,
    264 	mmTCP_CHAN_STEER_LO, 0xffffffff, 0xa9210876,
    265 	mmVGT_RESET_DEBUG, 0x00000004, 0x00000004,
    266 	mmVM_PRT_APERTURE0_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    267 	mmVM_PRT_APERTURE1_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    268 	mmVM_PRT_APERTURE2_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    269 	mmVM_PRT_APERTURE3_LOW_ADDR, 0x0fffffff, 0x0fffffff,
    270 };
    271 
    272 static const u32 xgpu_tonga_golden_common_all[] = {
    273 	mmGRBM_GFX_INDEX,               0xffffffff, 0xe0000000,
    274 	mmPA_SC_RASTER_CONFIG,          0xffffffff, 0x16000012,
    275 	mmPA_SC_RASTER_CONFIG_1,        0xffffffff, 0x0000002A,
    276 	mmGB_ADDR_CONFIG,               0xffffffff, 0x22011002,
    277 	mmSPI_RESOURCE_RESERVE_CU_0,    0xffffffff, 0x00000800,
    278 	mmSPI_RESOURCE_RESERVE_CU_1,    0xffffffff, 0x00000800,
    279 	mmSPI_RESOURCE_RESERVE_EN_CU_0, 0xffffffff, 0x00007FBF,
    280 };
    281 
    282 void xgpu_vi_init_golden_registers(struct amdgpu_device *adev)
    283 {
    284 	switch (adev->asic_type) {
    285 	case CHIP_FIJI:
    286 		amdgpu_device_program_register_sequence(adev,
    287 							xgpu_fiji_mgcg_cgcg_init,
    288 							ARRAY_SIZE(
    289 								xgpu_fiji_mgcg_cgcg_init));
    290 		amdgpu_device_program_register_sequence(adev,
    291 							xgpu_fiji_golden_settings_a10,
    292 							ARRAY_SIZE(
    293 								xgpu_fiji_golden_settings_a10));
    294 		amdgpu_device_program_register_sequence(adev,
    295 							xgpu_fiji_golden_common_all,
    296 							ARRAY_SIZE(
    297 								xgpu_fiji_golden_common_all));
    298 		break;
    299 	case CHIP_TONGA:
    300 		amdgpu_device_program_register_sequence(adev,
    301 							xgpu_tonga_mgcg_cgcg_init,
    302 							ARRAY_SIZE(
    303 								xgpu_tonga_mgcg_cgcg_init));
    304 		amdgpu_device_program_register_sequence(adev,
    305 							xgpu_tonga_golden_settings_a11,
    306 							ARRAY_SIZE(
    307 								xgpu_tonga_golden_settings_a11));
    308 		amdgpu_device_program_register_sequence(adev,
    309 							xgpu_tonga_golden_common_all,
    310 							ARRAY_SIZE(
    311 								xgpu_tonga_golden_common_all));
    312 		break;
    313 	default:
    314 		BUG_ON("Doesn't support chip type.\n");
    315 		break;
    316 	}
    317 }
    318 
    319 /*
    320  * Mailbox communication between GPU hypervisor and VFs
    321  */
    322 static void xgpu_vi_mailbox_send_ack(struct amdgpu_device *adev)
    323 {
    324 	u32 reg;
    325 	int timeout = VI_MAILBOX_TIMEDOUT;
    326 	u32 mask = REG_FIELD_MASK(MAILBOX_CONTROL, RCV_MSG_VALID);
    327 
    328 	reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    329 	reg = REG_SET_FIELD(reg, MAILBOX_CONTROL, RCV_MSG_ACK, 1);
    330 	WREG32_NO_KIQ(mmMAILBOX_CONTROL, reg);
    331 
    332 	/*Wait for RCV_MSG_VALID to be 0*/
    333 	reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    334 	while (reg & mask) {
    335 		if (timeout <= 0) {
    336 			pr_err("RCV_MSG_VALID is not cleared\n");
    337 			break;
    338 		}
    339 		mdelay(1);
    340 		timeout -=1;
    341 
    342 		reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    343 	}
    344 }
    345 
    346 static void xgpu_vi_mailbox_set_valid(struct amdgpu_device *adev, bool val)
    347 {
    348 	u32 reg;
    349 
    350 	reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    351 	reg = REG_SET_FIELD(reg, MAILBOX_CONTROL,
    352 			    TRN_MSG_VALID, val ? 1 : 0);
    353 	WREG32_NO_KIQ(mmMAILBOX_CONTROL, reg);
    354 }
    355 
    356 static void xgpu_vi_mailbox_trans_msg(struct amdgpu_device *adev,
    357 				      enum idh_request req)
    358 {
    359 	u32 reg;
    360 
    361 	reg = RREG32_NO_KIQ(mmMAILBOX_MSGBUF_TRN_DW0);
    362 	reg = REG_SET_FIELD(reg, MAILBOX_MSGBUF_TRN_DW0,
    363 			    MSGBUF_DATA, req);
    364 	WREG32_NO_KIQ(mmMAILBOX_MSGBUF_TRN_DW0, reg);
    365 
    366 	xgpu_vi_mailbox_set_valid(adev, true);
    367 }
    368 
    369 static int xgpu_vi_mailbox_rcv_msg(struct amdgpu_device *adev,
    370 				   enum idh_event event)
    371 {
    372 	u32 reg;
    373 	u32 mask = REG_FIELD_MASK(MAILBOX_CONTROL, RCV_MSG_VALID);
    374 
    375 	/* workaround: host driver doesn't set VALID for CMPL now */
    376 	if (event != IDH_FLR_NOTIFICATION_CMPL) {
    377 		reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    378 		if (!(reg & mask))
    379 			return -ENOENT;
    380 	}
    381 
    382 	reg = RREG32_NO_KIQ(mmMAILBOX_MSGBUF_RCV_DW0);
    383 	if (reg != event)
    384 		return -ENOENT;
    385 
    386 	/* send ack to PF */
    387 	xgpu_vi_mailbox_send_ack(adev);
    388 
    389 	return 0;
    390 }
    391 
    392 static int xgpu_vi_poll_ack(struct amdgpu_device *adev)
    393 {
    394 	int r = 0, timeout = VI_MAILBOX_TIMEDOUT;
    395 	u32 mask = REG_FIELD_MASK(MAILBOX_CONTROL, TRN_MSG_ACK);
    396 	u32 reg;
    397 
    398 	reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    399 	while (!(reg & mask)) {
    400 		if (timeout <= 0) {
    401 			pr_err("Doesn't get ack from pf.\n");
    402 			r = -ETIME;
    403 			break;
    404 		}
    405 		mdelay(5);
    406 		timeout -= 5;
    407 
    408 		reg = RREG32_NO_KIQ(mmMAILBOX_CONTROL);
    409 	}
    410 
    411 	return r;
    412 }
    413 
    414 static int xgpu_vi_poll_msg(struct amdgpu_device *adev, enum idh_event event)
    415 {
    416 	int r = 0, timeout = VI_MAILBOX_TIMEDOUT;
    417 
    418 	r = xgpu_vi_mailbox_rcv_msg(adev, event);
    419 	while (r) {
    420 		if (timeout <= 0) {
    421 			pr_err("Doesn't get ack from pf.\n");
    422 			r = -ETIME;
    423 			break;
    424 		}
    425 		mdelay(5);
    426 		timeout -= 5;
    427 
    428 		r = xgpu_vi_mailbox_rcv_msg(adev, event);
    429 	}
    430 
    431 	return r;
    432 }
    433 
    434 static int xgpu_vi_send_access_requests(struct amdgpu_device *adev,
    435 					enum idh_request request)
    436 {
    437 	int r;
    438 
    439 	xgpu_vi_mailbox_trans_msg(adev, request);
    440 
    441 	/* start to poll ack */
    442 	r = xgpu_vi_poll_ack(adev);
    443 	if (r)
    444 		return r;
    445 
    446 	xgpu_vi_mailbox_set_valid(adev, false);
    447 
    448 	/* start to check msg if request is idh_req_gpu_init_access */
    449 	if (request == IDH_REQ_GPU_INIT_ACCESS ||
    450 		request == IDH_REQ_GPU_FINI_ACCESS ||
    451 		request == IDH_REQ_GPU_RESET_ACCESS) {
    452 		r = xgpu_vi_poll_msg(adev, IDH_READY_TO_ACCESS_GPU);
    453 		if (r) {
    454 			pr_err("Doesn't get ack from pf, give up\n");
    455 			return r;
    456 		}
    457 	}
    458 
    459 	return 0;
    460 }
    461 
    462 static int xgpu_vi_request_reset(struct amdgpu_device *adev)
    463 {
    464 	return xgpu_vi_send_access_requests(adev, IDH_REQ_GPU_RESET_ACCESS);
    465 }
    466 
    467 static int xgpu_vi_wait_reset_cmpl(struct amdgpu_device *adev)
    468 {
    469 	return xgpu_vi_poll_msg(adev, IDH_FLR_NOTIFICATION_CMPL);
    470 }
    471 
    472 static int xgpu_vi_request_full_gpu_access(struct amdgpu_device *adev,
    473 					   bool init)
    474 {
    475 	enum idh_request req;
    476 
    477 	req = init ? IDH_REQ_GPU_INIT_ACCESS : IDH_REQ_GPU_FINI_ACCESS;
    478 	return xgpu_vi_send_access_requests(adev, req);
    479 }
    480 
    481 static int xgpu_vi_release_full_gpu_access(struct amdgpu_device *adev,
    482 					   bool init)
    483 {
    484 	enum idh_request req;
    485 	int r = 0;
    486 
    487 	req = init ? IDH_REL_GPU_INIT_ACCESS : IDH_REL_GPU_FINI_ACCESS;
    488 	r = xgpu_vi_send_access_requests(adev, req);
    489 
    490 	return r;
    491 }
    492 
    493 /* add support mailbox interrupts */
    494 static int xgpu_vi_mailbox_ack_irq(struct amdgpu_device *adev,
    495 				   struct amdgpu_irq_src *source,
    496 				   struct amdgpu_iv_entry *entry)
    497 {
    498 	DRM_DEBUG("get ack intr and do nothing.\n");
    499 	return 0;
    500 }
    501 
    502 static int xgpu_vi_set_mailbox_ack_irq(struct amdgpu_device *adev,
    503 				       struct amdgpu_irq_src *src,
    504 				       unsigned type,
    505 				       enum amdgpu_interrupt_state state)
    506 {
    507 	u32 tmp = RREG32_NO_KIQ(mmMAILBOX_INT_CNTL);
    508 
    509 	tmp = REG_SET_FIELD(tmp, MAILBOX_INT_CNTL, ACK_INT_EN,
    510 			    (state == AMDGPU_IRQ_STATE_ENABLE) ? 1 : 0);
    511 	WREG32_NO_KIQ(mmMAILBOX_INT_CNTL, tmp);
    512 
    513 	return 0;
    514 }
    515 
    516 static void xgpu_vi_mailbox_flr_work(struct work_struct *work)
    517 {
    518 	struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work);
    519 	struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt);
    520 
    521 	/* wait until RCV_MSG become 3 */
    522 	if (xgpu_vi_poll_msg(adev, IDH_FLR_NOTIFICATION_CMPL)) {
    523 		pr_err("failed to receive FLR_CMPL\n");
    524 		return;
    525 	}
    526 
    527 	/* Trigger recovery due to world switch failure */
    528 	if (amdgpu_device_should_recover_gpu(adev))
    529 		amdgpu_device_gpu_recover(adev, NULL);
    530 }
    531 
    532 static int xgpu_vi_set_mailbox_rcv_irq(struct amdgpu_device *adev,
    533 				       struct amdgpu_irq_src *src,
    534 				       unsigned type,
    535 				       enum amdgpu_interrupt_state state)
    536 {
    537 	u32 tmp = RREG32_NO_KIQ(mmMAILBOX_INT_CNTL);
    538 
    539 	tmp = REG_SET_FIELD(tmp, MAILBOX_INT_CNTL, VALID_INT_EN,
    540 			    (state == AMDGPU_IRQ_STATE_ENABLE) ? 1 : 0);
    541 	WREG32_NO_KIQ(mmMAILBOX_INT_CNTL, tmp);
    542 
    543 	return 0;
    544 }
    545 
    546 static int xgpu_vi_mailbox_rcv_irq(struct amdgpu_device *adev,
    547 				   struct amdgpu_irq_src *source,
    548 				   struct amdgpu_iv_entry *entry)
    549 {
    550 	int r;
    551 
    552 	/* trigger gpu-reset by hypervisor only if TDR disbaled */
    553 	if (!amdgpu_gpu_recovery) {
    554 		/* see what event we get */
    555 		r = xgpu_vi_mailbox_rcv_msg(adev, IDH_FLR_NOTIFICATION);
    556 
    557 		/* only handle FLR_NOTIFY now */
    558 		if (!r)
    559 			schedule_work(&adev->virt.flr_work);
    560 	}
    561 
    562 	return 0;
    563 }
    564 
    565 static const struct amdgpu_irq_src_funcs xgpu_vi_mailbox_ack_irq_funcs = {
    566 	.set = xgpu_vi_set_mailbox_ack_irq,
    567 	.process = xgpu_vi_mailbox_ack_irq,
    568 };
    569 
    570 static const struct amdgpu_irq_src_funcs xgpu_vi_mailbox_rcv_irq_funcs = {
    571 	.set = xgpu_vi_set_mailbox_rcv_irq,
    572 	.process = xgpu_vi_mailbox_rcv_irq,
    573 };
    574 
    575 void xgpu_vi_mailbox_set_irq_funcs(struct amdgpu_device *adev)
    576 {
    577 	adev->virt.ack_irq.num_types = 1;
    578 	adev->virt.ack_irq.funcs = &xgpu_vi_mailbox_ack_irq_funcs;
    579 	adev->virt.rcv_irq.num_types = 1;
    580 	adev->virt.rcv_irq.funcs = &xgpu_vi_mailbox_rcv_irq_funcs;
    581 }
    582 
    583 int xgpu_vi_mailbox_add_irq_id(struct amdgpu_device *adev)
    584 {
    585 	int r;
    586 
    587 	r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 135, &adev->virt.rcv_irq);
    588 	if (r)
    589 		return r;
    590 
    591 	r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 138, &adev->virt.ack_irq);
    592 	if (r) {
    593 		amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0);
    594 		return r;
    595 	}
    596 
    597 	return 0;
    598 }
    599 
    600 int xgpu_vi_mailbox_get_irq(struct amdgpu_device *adev)
    601 {
    602 	int r;
    603 
    604 	r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0);
    605 	if (r)
    606 		return r;
    607 	r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0);
    608 	if (r) {
    609 		amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0);
    610 		return r;
    611 	}
    612 
    613 	INIT_WORK(&adev->virt.flr_work, xgpu_vi_mailbox_flr_work);
    614 
    615 	return 0;
    616 }
    617 
    618 void xgpu_vi_mailbox_put_irq(struct amdgpu_device *adev)
    619 {
    620 	amdgpu_irq_put(adev, &adev->virt.ack_irq, 0);
    621 	amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0);
    622 }
    623 
    624 const struct amdgpu_virt_ops xgpu_vi_virt_ops = {
    625 	.req_full_gpu		= xgpu_vi_request_full_gpu_access,
    626 	.rel_full_gpu		= xgpu_vi_release_full_gpu_access,
    627 	.reset_gpu		= xgpu_vi_request_reset,
    628 	.wait_reset             = xgpu_vi_wait_reset_cmpl,
    629 	.trans_msg		= NULL, /* Does not need to trans VF errors to host. */
    630 };
    631