Home | History | Annotate | Line # | Download | only in radeon
      1 /*	$NetBSD: radeon_ci_dpm.c,v 1.7 2023/09/30 10:46:45 mrg Exp $	*/
      2 
      3 /*
      4  * Copyright 2013 Advanced Micro Devices, Inc.
      5  *
      6  * Permission is hereby granted, free of charge, to any person obtaining a
      7  * copy of this software and associated documentation files (the "Software"),
      8  * to deal in the Software without restriction, including without limitation
      9  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
     10  * and/or sell copies of the Software, and to permit persons to whom the
     11  * Software is furnished to do so, subject to the following conditions:
     12  *
     13  * The above copyright notice and this permission notice shall be included in
     14  * all copies or substantial portions of the Software.
     15  *
     16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     19  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     20  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     21  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     22  * OTHER DEALINGS IN THE SOFTWARE.
     23  *
     24  */
     25 
     26 #include <sys/cdefs.h>
     27 __KERNEL_RCSID(0, "$NetBSD: radeon_ci_dpm.c,v 1.7 2023/09/30 10:46:45 mrg Exp $");
     28 
     29 #include <linux/firmware.h>
     30 #include <linux/pci.h>
     31 #include <linux/seq_file.h>
     32 
     33 #include "atom.h"
     34 #include "ci_dpm.h"
     35 #include "cikd.h"
     36 #include "r600_dpm.h"
     37 #include "radeon.h"
     38 #include "radeon_asic.h"
     39 #include "radeon_ucode.h"
     40 
     41 #define MC_CG_ARB_FREQ_F0           0x0a
     42 #define MC_CG_ARB_FREQ_F1           0x0b
     43 #define MC_CG_ARB_FREQ_F2           0x0c
     44 #define MC_CG_ARB_FREQ_F3           0x0d
     45 
     46 #define SMC_RAM_END 0x40000
     47 
     48 #define VOLTAGE_SCALE               4
     49 #define VOLTAGE_VID_OFFSET_SCALE1    625
     50 #define VOLTAGE_VID_OFFSET_SCALE2    100
     51 
     52 static const struct ci_pt_defaults defaults_hawaii_xt =
     53 {
     54 	1, 0xF, 0xFD, 0x19, 5, 0x14, 0, 0xB0000,
     55 	{ 0x2E,  0x00,  0x00,  0x88,  0x00,  0x00,  0x72,  0x60,  0x51,  0xA7,  0x79,  0x6B,  0x90,  0xBD,  0x79  },
     56 	{ 0x217, 0x217, 0x217, 0x242, 0x242, 0x242, 0x269, 0x269, 0x269, 0x2A1, 0x2A1, 0x2A1, 0x2C9, 0x2C9, 0x2C9 }
     57 };
     58 
     59 static const struct ci_pt_defaults defaults_hawaii_pro =
     60 {
     61 	1, 0xF, 0xFD, 0x19, 5, 0x14, 0, 0x65062,
     62 	{ 0x2E,  0x00,  0x00,  0x88,  0x00,  0x00,  0x72,  0x60,  0x51,  0xA7,  0x79,  0x6B,  0x90,  0xBD,  0x79  },
     63 	{ 0x217, 0x217, 0x217, 0x242, 0x242, 0x242, 0x269, 0x269, 0x269, 0x2A1, 0x2A1, 0x2A1, 0x2C9, 0x2C9, 0x2C9 }
     64 };
     65 
     66 static const struct ci_pt_defaults defaults_bonaire_xt =
     67 {
     68 	1, 0xF, 0xFD, 0x19, 5, 45, 0, 0xB0000,
     69 	{ 0x79,  0x253, 0x25D, 0xAE,  0x72,  0x80,  0x83,  0x86,  0x6F,  0xC8,  0xC9,  0xC9,  0x2F,  0x4D,  0x61  },
     70 	{ 0x17C, 0x172, 0x180, 0x1BC, 0x1B3, 0x1BD, 0x206, 0x200, 0x203, 0x25D, 0x25A, 0x255, 0x2C3, 0x2C5, 0x2B4 }
     71 };
     72 
     73 static const struct ci_pt_defaults defaults_bonaire_pro __unused =
     74 {
     75 	1, 0xF, 0xFD, 0x19, 5, 45, 0, 0x65062,
     76 	{ 0x8C,  0x23F, 0x244, 0xA6,  0x83,  0x85,  0x86,  0x86,  0x83,  0xDB,  0xDB,  0xDA,  0x67,  0x60,  0x5F  },
     77 	{ 0x187, 0x193, 0x193, 0x1C7, 0x1D1, 0x1D1, 0x210, 0x219, 0x219, 0x266, 0x26C, 0x26C, 0x2C9, 0x2CB, 0x2CB }
     78 };
     79 
     80 static const struct ci_pt_defaults defaults_saturn_xt =
     81 {
     82 	1, 0xF, 0xFD, 0x19, 5, 55, 0, 0x70000,
     83 	{ 0x8C,  0x247, 0x249, 0xA6,  0x80,  0x81,  0x8B,  0x89,  0x86,  0xC9,  0xCA,  0xC9,  0x4D,  0x4D,  0x4D  },
     84 	{ 0x187, 0x187, 0x187, 0x1C7, 0x1C7, 0x1C7, 0x210, 0x210, 0x210, 0x266, 0x266, 0x266, 0x2C9, 0x2C9, 0x2C9 }
     85 };
     86 
     87 static const struct ci_pt_defaults defaults_saturn_pro __unused =
     88 {
     89 	1, 0xF, 0xFD, 0x19, 5, 55, 0, 0x30000,
     90 	{ 0x96,  0x21D, 0x23B, 0xA1,  0x85,  0x87,  0x83,  0x84,  0x81,  0xE6,  0xE6,  0xE6,  0x71,  0x6A,  0x6A  },
     91 	{ 0x193, 0x19E, 0x19E, 0x1D2, 0x1DC, 0x1DC, 0x21A, 0x223, 0x223, 0x26E, 0x27E, 0x274, 0x2CF, 0x2D2, 0x2D2 }
     92 };
     93 
     94 static const struct ci_pt_config_reg didt_config_ci[] =
     95 {
     96 	{ 0x10, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
     97 	{ 0x10, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
     98 	{ 0x10, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
     99 	{ 0x10, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    100 	{ 0x11, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    101 	{ 0x11, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    102 	{ 0x11, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    103 	{ 0x11, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    104 	{ 0x12, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    105 	{ 0x12, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    106 	{ 0x12, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    107 	{ 0x12, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    108 	{ 0x2, 0x00003fff, 0, 0x4, CISLANDS_CONFIGREG_DIDT_IND },
    109 	{ 0x2, 0x03ff0000, 16, 0x80, CISLANDS_CONFIGREG_DIDT_IND },
    110 	{ 0x2, 0x78000000, 27, 0x3, CISLANDS_CONFIGREG_DIDT_IND },
    111 	{ 0x1, 0x0000ffff, 0, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    112 	{ 0x1, 0xffff0000, 16, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    113 	{ 0x0, 0x00000001, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    114 	{ 0x30, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    115 	{ 0x30, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    116 	{ 0x30, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    117 	{ 0x30, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    118 	{ 0x31, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    119 	{ 0x31, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    120 	{ 0x31, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    121 	{ 0x31, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    122 	{ 0x32, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    123 	{ 0x32, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    124 	{ 0x32, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    125 	{ 0x32, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    126 	{ 0x22, 0x00003fff, 0, 0x4, CISLANDS_CONFIGREG_DIDT_IND },
    127 	{ 0x22, 0x03ff0000, 16, 0x80, CISLANDS_CONFIGREG_DIDT_IND },
    128 	{ 0x22, 0x78000000, 27, 0x3, CISLANDS_CONFIGREG_DIDT_IND },
    129 	{ 0x21, 0x0000ffff, 0, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    130 	{ 0x21, 0xffff0000, 16, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    131 	{ 0x20, 0x00000001, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    132 	{ 0x50, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    133 	{ 0x50, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    134 	{ 0x50, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    135 	{ 0x50, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    136 	{ 0x51, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    137 	{ 0x51, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    138 	{ 0x51, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    139 	{ 0x51, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    140 	{ 0x52, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    141 	{ 0x52, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    142 	{ 0x52, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    143 	{ 0x52, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    144 	{ 0x42, 0x00003fff, 0, 0x4, CISLANDS_CONFIGREG_DIDT_IND },
    145 	{ 0x42, 0x03ff0000, 16, 0x80, CISLANDS_CONFIGREG_DIDT_IND },
    146 	{ 0x42, 0x78000000, 27, 0x3, CISLANDS_CONFIGREG_DIDT_IND },
    147 	{ 0x41, 0x0000ffff, 0, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    148 	{ 0x41, 0xffff0000, 16, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    149 	{ 0x40, 0x00000001, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    150 	{ 0x70, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    151 	{ 0x70, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    152 	{ 0x70, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    153 	{ 0x70, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    154 	{ 0x71, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    155 	{ 0x71, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    156 	{ 0x71, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    157 	{ 0x71, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    158 	{ 0x72, 0x000000ff, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    159 	{ 0x72, 0x0000ff00, 8, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    160 	{ 0x72, 0x00ff0000, 16, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    161 	{ 0x72, 0xff000000, 24, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    162 	{ 0x62, 0x00003fff, 0, 0x4, CISLANDS_CONFIGREG_DIDT_IND },
    163 	{ 0x62, 0x03ff0000, 16, 0x80, CISLANDS_CONFIGREG_DIDT_IND },
    164 	{ 0x62, 0x78000000, 27, 0x3, CISLANDS_CONFIGREG_DIDT_IND },
    165 	{ 0x61, 0x0000ffff, 0, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    166 	{ 0x61, 0xffff0000, 16, 0x3FFF, CISLANDS_CONFIGREG_DIDT_IND },
    167 	{ 0x60, 0x00000001, 0, 0x0, CISLANDS_CONFIGREG_DIDT_IND },
    168 	{ 0xFFFFFFFF }
    169 };
    170 
    171 extern u8 rv770_get_memory_module_index(struct radeon_device *rdev);
    172 extern int ni_copy_and_switch_arb_sets(struct radeon_device *rdev,
    173 				       u32 arb_freq_src, u32 arb_freq_dest);
    174 extern u8 si_get_ddr3_mclk_frequency_ratio(u32 memory_clock);
    175 extern u8 si_get_mclk_frequency_ratio(u32 memory_clock, bool strobe_mode);
    176 extern void si_trim_voltage_table_to_fit_state_table(struct radeon_device *rdev,
    177 						     u32 max_voltage_steps,
    178 						     struct atom_voltage_table *voltage_table);
    179 extern void cik_enter_rlc_safe_mode(struct radeon_device *rdev);
    180 extern void cik_exit_rlc_safe_mode(struct radeon_device *rdev);
    181 extern int ci_mc_load_microcode(struct radeon_device *rdev);
    182 extern void cik_update_cg(struct radeon_device *rdev,
    183 			  u32 block, bool enable);
    184 
    185 static int ci_get_std_voltage_value_sidd(struct radeon_device *rdev,
    186 					 struct atom_voltage_table_entry *voltage_table,
    187 					 u16 *std_voltage_hi_sidd, u16 *std_voltage_lo_sidd);
    188 static int ci_set_power_limit(struct radeon_device *rdev, u32 n);
    189 static int ci_set_overdrive_target_tdp(struct radeon_device *rdev,
    190 				       u32 target_tdp);
    191 static int ci_update_uvd_dpm(struct radeon_device *rdev, bool gate);
    192 
    193 static PPSMC_Result ci_send_msg_to_smc(struct radeon_device *rdev, PPSMC_Msg msg);
    194 static PPSMC_Result ci_send_msg_to_smc_with_parameter(struct radeon_device *rdev,
    195 						      PPSMC_Msg msg, u32 parameter);
    196 
    197 static void ci_thermal_start_smc_fan_control(struct radeon_device *rdev);
    198 static void ci_fan_ctrl_set_default_mode(struct radeon_device *rdev);
    199 
    200 static struct ci_power_info *ci_get_pi(struct radeon_device *rdev)
    201 {
    202 	struct ci_power_info *pi = rdev->pm.dpm.priv;
    203 
    204 	return pi;
    205 }
    206 
    207 static struct ci_ps *ci_get_ps(struct radeon_ps *rps)
    208 {
    209 	struct ci_ps *ps = rps->ps_priv;
    210 
    211 	return ps;
    212 }
    213 
    214 static void ci_initialize_powertune_defaults(struct radeon_device *rdev)
    215 {
    216 	struct ci_power_info *pi = ci_get_pi(rdev);
    217 
    218 	switch (rdev->pdev->device) {
    219 	case 0x6649:
    220 	case 0x6650:
    221 	case 0x6651:
    222 	case 0x6658:
    223 	case 0x665C:
    224 	case 0x665D:
    225 	default:
    226 		pi->powertune_defaults = &defaults_bonaire_xt;
    227 		break;
    228 	case 0x6640:
    229 	case 0x6641:
    230 	case 0x6646:
    231 	case 0x6647:
    232 		pi->powertune_defaults = &defaults_saturn_xt;
    233 		break;
    234 	case 0x67B8:
    235 	case 0x67B0:
    236 		pi->powertune_defaults = &defaults_hawaii_xt;
    237 		break;
    238 	case 0x67BA:
    239 	case 0x67B1:
    240 		pi->powertune_defaults = &defaults_hawaii_pro;
    241 		break;
    242 	case 0x67A0:
    243 	case 0x67A1:
    244 	case 0x67A2:
    245 	case 0x67A8:
    246 	case 0x67A9:
    247 	case 0x67AA:
    248 	case 0x67B9:
    249 	case 0x67BE:
    250 		pi->powertune_defaults = &defaults_bonaire_xt;
    251 		break;
    252 	}
    253 
    254 	pi->dte_tj_offset = 0;
    255 
    256 	pi->caps_power_containment = true;
    257 	pi->caps_cac = false;
    258 	pi->caps_sq_ramping = false;
    259 	pi->caps_db_ramping = false;
    260 	pi->caps_td_ramping = false;
    261 	pi->caps_tcp_ramping = false;
    262 
    263 	if (pi->caps_power_containment) {
    264 		pi->caps_cac = true;
    265 		if (rdev->family == CHIP_HAWAII)
    266 			pi->enable_bapm_feature = false;
    267 		else
    268 			pi->enable_bapm_feature = true;
    269 		pi->enable_tdc_limit_feature = true;
    270 		pi->enable_pkg_pwr_tracking_feature = true;
    271 	}
    272 }
    273 
    274 static u8 ci_convert_to_vid(u16 vddc)
    275 {
    276 	return (6200 - (vddc * VOLTAGE_SCALE)) / 25;
    277 }
    278 
    279 static int ci_populate_bapm_vddc_vid_sidd(struct radeon_device *rdev)
    280 {
    281 	struct ci_power_info *pi = ci_get_pi(rdev);
    282 	u8 *hi_vid = pi->smc_powertune_table.BapmVddCVidHiSidd;
    283 	u8 *lo_vid = pi->smc_powertune_table.BapmVddCVidLoSidd;
    284 	u8 *hi2_vid = pi->smc_powertune_table.BapmVddCVidHiSidd2;
    285 	u32 i;
    286 
    287 	if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries == NULL)
    288 		return -EINVAL;
    289 	if (rdev->pm.dpm.dyn_state.cac_leakage_table.count > 8)
    290 		return -EINVAL;
    291 	if (rdev->pm.dpm.dyn_state.cac_leakage_table.count !=
    292 	    rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.count)
    293 		return -EINVAL;
    294 
    295 	for (i = 0; i < rdev->pm.dpm.dyn_state.cac_leakage_table.count; i++) {
    296 		if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_EVV) {
    297 			lo_vid[i] = ci_convert_to_vid(rdev->pm.dpm.dyn_state.cac_leakage_table.entries[i].vddc1);
    298 			hi_vid[i] = ci_convert_to_vid(rdev->pm.dpm.dyn_state.cac_leakage_table.entries[i].vddc2);
    299 			hi2_vid[i] = ci_convert_to_vid(rdev->pm.dpm.dyn_state.cac_leakage_table.entries[i].vddc3);
    300 		} else {
    301 			lo_vid[i] = ci_convert_to_vid(rdev->pm.dpm.dyn_state.cac_leakage_table.entries[i].vddc);
    302 			hi_vid[i] = ci_convert_to_vid((u16)rdev->pm.dpm.dyn_state.cac_leakage_table.entries[i].leakage);
    303 		}
    304 	}
    305 	return 0;
    306 }
    307 
    308 static int ci_populate_vddc_vid(struct radeon_device *rdev)
    309 {
    310 	struct ci_power_info *pi = ci_get_pi(rdev);
    311 	u8 *vid = pi->smc_powertune_table.VddCVid;
    312 	u32 i;
    313 
    314 	if (pi->vddc_voltage_table.count > 8)
    315 		return -EINVAL;
    316 
    317 	for (i = 0; i < pi->vddc_voltage_table.count; i++)
    318 		vid[i] = ci_convert_to_vid(pi->vddc_voltage_table.entries[i].value);
    319 
    320 	return 0;
    321 }
    322 
    323 static int ci_populate_svi_load_line(struct radeon_device *rdev)
    324 {
    325 	struct ci_power_info *pi = ci_get_pi(rdev);
    326 	const struct ci_pt_defaults *pt_defaults = pi->powertune_defaults;
    327 
    328 	pi->smc_powertune_table.SviLoadLineEn = pt_defaults->svi_load_line_en;
    329 	pi->smc_powertune_table.SviLoadLineVddC = pt_defaults->svi_load_line_vddc;
    330 	pi->smc_powertune_table.SviLoadLineTrimVddC = 3;
    331 	pi->smc_powertune_table.SviLoadLineOffsetVddC = 0;
    332 
    333 	return 0;
    334 }
    335 
    336 static int ci_populate_tdc_limit(struct radeon_device *rdev)
    337 {
    338 	struct ci_power_info *pi = ci_get_pi(rdev);
    339 	const struct ci_pt_defaults *pt_defaults = pi->powertune_defaults;
    340 	u16 tdc_limit;
    341 
    342 	tdc_limit = rdev->pm.dpm.dyn_state.cac_tdp_table->tdc * 256;
    343 	pi->smc_powertune_table.TDC_VDDC_PkgLimit = cpu_to_be16(tdc_limit);
    344 	pi->smc_powertune_table.TDC_VDDC_ThrottleReleaseLimitPerc =
    345 		pt_defaults->tdc_vddc_throttle_release_limit_perc;
    346 	pi->smc_powertune_table.TDC_MAWt = pt_defaults->tdc_mawt;
    347 
    348 	return 0;
    349 }
    350 
    351 static int ci_populate_dw8(struct radeon_device *rdev)
    352 {
    353 	struct ci_power_info *pi = ci_get_pi(rdev);
    354 	const struct ci_pt_defaults *pt_defaults = pi->powertune_defaults;
    355 	int ret;
    356 
    357 	ret = ci_read_smc_sram_dword(rdev,
    358 				     SMU7_FIRMWARE_HEADER_LOCATION +
    359 				     offsetof(SMU7_Firmware_Header, PmFuseTable) +
    360 				     offsetof(SMU7_Discrete_PmFuses, TdcWaterfallCtl),
    361 				     (u32 *)&pi->smc_powertune_table.TdcWaterfallCtl,
    362 				     pi->sram_end);
    363 	if (ret)
    364 		return -EINVAL;
    365 	else
    366 		pi->smc_powertune_table.TdcWaterfallCtl = pt_defaults->tdc_waterfall_ctl;
    367 
    368 	return 0;
    369 }
    370 
    371 static int ci_populate_fuzzy_fan(struct radeon_device *rdev)
    372 {
    373 	struct ci_power_info *pi = ci_get_pi(rdev);
    374 
    375 	if ((rdev->pm.dpm.fan.fan_output_sensitivity & (1 << 15)) ||
    376 	    (rdev->pm.dpm.fan.fan_output_sensitivity == 0))
    377 		rdev->pm.dpm.fan.fan_output_sensitivity =
    378 			rdev->pm.dpm.fan.default_fan_output_sensitivity;
    379 
    380 	pi->smc_powertune_table.FuzzyFan_PwmSetDelta =
    381 		cpu_to_be16(rdev->pm.dpm.fan.fan_output_sensitivity);
    382 
    383 	return 0;
    384 }
    385 
    386 static int ci_min_max_v_gnbl_pm_lid_from_bapm_vddc(struct radeon_device *rdev)
    387 {
    388 	struct ci_power_info *pi = ci_get_pi(rdev);
    389 	u8 *hi_vid = pi->smc_powertune_table.BapmVddCVidHiSidd;
    390 	u8 *lo_vid = pi->smc_powertune_table.BapmVddCVidLoSidd;
    391 	int i, min, max;
    392 
    393 	min = max = hi_vid[0];
    394 	for (i = 0; i < 8; i++) {
    395 		if (0 != hi_vid[i]) {
    396 			if (min > hi_vid[i])
    397 				min = hi_vid[i];
    398 			if (max < hi_vid[i])
    399 				max = hi_vid[i];
    400 		}
    401 
    402 		if (0 != lo_vid[i]) {
    403 			if (min > lo_vid[i])
    404 				min = lo_vid[i];
    405 			if (max < lo_vid[i])
    406 				max = lo_vid[i];
    407 		}
    408 	}
    409 
    410 	if ((min == 0) || (max == 0))
    411 		return -EINVAL;
    412 	pi->smc_powertune_table.GnbLPMLMaxVid = (u8)max;
    413 	pi->smc_powertune_table.GnbLPMLMinVid = (u8)min;
    414 
    415 	return 0;
    416 }
    417 
    418 static int ci_populate_bapm_vddc_base_leakage_sidd(struct radeon_device *rdev)
    419 {
    420 	struct ci_power_info *pi = ci_get_pi(rdev);
    421 	u16 hi_sidd = pi->smc_powertune_table.BapmVddCBaseLeakageHiSidd;
    422 	u16 lo_sidd = pi->smc_powertune_table.BapmVddCBaseLeakageLoSidd;
    423 	struct radeon_cac_tdp_table *cac_tdp_table =
    424 		rdev->pm.dpm.dyn_state.cac_tdp_table;
    425 
    426 	hi_sidd = cac_tdp_table->high_cac_leakage / 100 * 256;
    427 	lo_sidd = cac_tdp_table->low_cac_leakage / 100 * 256;
    428 
    429 	pi->smc_powertune_table.BapmVddCBaseLeakageHiSidd = cpu_to_be16(hi_sidd);
    430 	pi->smc_powertune_table.BapmVddCBaseLeakageLoSidd = cpu_to_be16(lo_sidd);
    431 
    432 	return 0;
    433 }
    434 
    435 static int ci_populate_bapm_parameters_in_dpm_table(struct radeon_device *rdev)
    436 {
    437 	struct ci_power_info *pi = ci_get_pi(rdev);
    438 	const struct ci_pt_defaults *pt_defaults = pi->powertune_defaults;
    439 	SMU7_Discrete_DpmTable  *dpm_table = &pi->smc_state_table;
    440 	struct radeon_cac_tdp_table *cac_tdp_table =
    441 		rdev->pm.dpm.dyn_state.cac_tdp_table;
    442 	struct radeon_ppm_table *ppm = rdev->pm.dpm.dyn_state.ppm_table;
    443 	int i, j, k;
    444 	const u16 *def1;
    445 	const u16 *def2;
    446 
    447 	dpm_table->DefaultTdp = cac_tdp_table->tdp * 256;
    448 	dpm_table->TargetTdp = cac_tdp_table->configurable_tdp * 256;
    449 
    450 	dpm_table->DTETjOffset = (u8)pi->dte_tj_offset;
    451 	dpm_table->GpuTjMax =
    452 		(u8)(pi->thermal_temp_setting.temperature_high / 1000);
    453 	dpm_table->GpuTjHyst = 8;
    454 
    455 	dpm_table->DTEAmbientTempBase = pt_defaults->dte_ambient_temp_base;
    456 
    457 	if (ppm) {
    458 		dpm_table->PPM_PkgPwrLimit = cpu_to_be16((u16)ppm->dgpu_tdp * 256 / 1000);
    459 		dpm_table->PPM_TemperatureLimit = cpu_to_be16((u16)ppm->tj_max * 256);
    460 	} else {
    461 		dpm_table->PPM_PkgPwrLimit = cpu_to_be16(0);
    462 		dpm_table->PPM_TemperatureLimit = cpu_to_be16(0);
    463 	}
    464 
    465 	dpm_table->BAPM_TEMP_GRADIENT = cpu_to_be32(pt_defaults->bapm_temp_gradient);
    466 	def1 = pt_defaults->bapmti_r;
    467 	def2 = pt_defaults->bapmti_rc;
    468 
    469 	for (i = 0; i < SMU7_DTE_ITERATIONS; i++) {
    470 		for (j = 0; j < SMU7_DTE_SOURCES; j++) {
    471 			for (k = 0; k < SMU7_DTE_SINKS; k++) {
    472 				dpm_table->BAPMTI_R[i][j][k] = cpu_to_be16(*def1);
    473 				dpm_table->BAPMTI_RC[i][j][k] = cpu_to_be16(*def2);
    474 				def1++;
    475 				def2++;
    476 			}
    477 		}
    478 	}
    479 
    480 	return 0;
    481 }
    482 
    483 static int ci_populate_pm_base(struct radeon_device *rdev)
    484 {
    485 	struct ci_power_info *pi = ci_get_pi(rdev);
    486 	u32 pm_fuse_table_offset;
    487 	int ret;
    488 
    489 	if (pi->caps_power_containment) {
    490 		ret = ci_read_smc_sram_dword(rdev,
    491 					     SMU7_FIRMWARE_HEADER_LOCATION +
    492 					     offsetof(SMU7_Firmware_Header, PmFuseTable),
    493 					     &pm_fuse_table_offset, pi->sram_end);
    494 		if (ret)
    495 			return ret;
    496 		ret = ci_populate_bapm_vddc_vid_sidd(rdev);
    497 		if (ret)
    498 			return ret;
    499 		ret = ci_populate_vddc_vid(rdev);
    500 		if (ret)
    501 			return ret;
    502 		ret = ci_populate_svi_load_line(rdev);
    503 		if (ret)
    504 			return ret;
    505 		ret = ci_populate_tdc_limit(rdev);
    506 		if (ret)
    507 			return ret;
    508 		ret = ci_populate_dw8(rdev);
    509 		if (ret)
    510 			return ret;
    511 		ret = ci_populate_fuzzy_fan(rdev);
    512 		if (ret)
    513 			return ret;
    514 		ret = ci_min_max_v_gnbl_pm_lid_from_bapm_vddc(rdev);
    515 		if (ret)
    516 			return ret;
    517 		ret = ci_populate_bapm_vddc_base_leakage_sidd(rdev);
    518 		if (ret)
    519 			return ret;
    520 		ret = ci_copy_bytes_to_smc(rdev, pm_fuse_table_offset,
    521 					   (u8 *)&pi->smc_powertune_table,
    522 					   sizeof(SMU7_Discrete_PmFuses), pi->sram_end);
    523 		if (ret)
    524 			return ret;
    525 	}
    526 
    527 	return 0;
    528 }
    529 
    530 static void ci_do_enable_didt(struct radeon_device *rdev, const bool enable)
    531 {
    532 	struct ci_power_info *pi = ci_get_pi(rdev);
    533 	u32 data;
    534 
    535 	if (pi->caps_sq_ramping) {
    536 		data = RREG32_DIDT(DIDT_SQ_CTRL0);
    537 		if (enable)
    538 			data |= DIDT_CTRL_EN;
    539 		else
    540 			data &= ~DIDT_CTRL_EN;
    541 		WREG32_DIDT(DIDT_SQ_CTRL0, data);
    542 	}
    543 
    544 	if (pi->caps_db_ramping) {
    545 		data = RREG32_DIDT(DIDT_DB_CTRL0);
    546 		if (enable)
    547 			data |= DIDT_CTRL_EN;
    548 		else
    549 			data &= ~DIDT_CTRL_EN;
    550 		WREG32_DIDT(DIDT_DB_CTRL0, data);
    551 	}
    552 
    553 	if (pi->caps_td_ramping) {
    554 		data = RREG32_DIDT(DIDT_TD_CTRL0);
    555 		if (enable)
    556 			data |= DIDT_CTRL_EN;
    557 		else
    558 			data &= ~DIDT_CTRL_EN;
    559 		WREG32_DIDT(DIDT_TD_CTRL0, data);
    560 	}
    561 
    562 	if (pi->caps_tcp_ramping) {
    563 		data = RREG32_DIDT(DIDT_TCP_CTRL0);
    564 		if (enable)
    565 			data |= DIDT_CTRL_EN;
    566 		else
    567 			data &= ~DIDT_CTRL_EN;
    568 		WREG32_DIDT(DIDT_TCP_CTRL0, data);
    569 	}
    570 }
    571 
    572 static int ci_program_pt_config_registers(struct radeon_device *rdev,
    573 					  const struct ci_pt_config_reg *cac_config_regs)
    574 {
    575 	const struct ci_pt_config_reg *config_regs = cac_config_regs;
    576 	u32 data;
    577 	u32 cache = 0;
    578 
    579 	if (config_regs == NULL)
    580 		return -EINVAL;
    581 
    582 	while (config_regs->offset != 0xFFFFFFFF) {
    583 		if (config_regs->type == CISLANDS_CONFIGREG_CACHE) {
    584 			cache |= ((config_regs->value << config_regs->shift) & config_regs->mask);
    585 		} else {
    586 			switch (config_regs->type) {
    587 			case CISLANDS_CONFIGREG_SMC_IND:
    588 				data = RREG32_SMC(config_regs->offset);
    589 				break;
    590 			case CISLANDS_CONFIGREG_DIDT_IND:
    591 				data = RREG32_DIDT(config_regs->offset);
    592 				break;
    593 			default:
    594 				data = RREG32(config_regs->offset << 2);
    595 				break;
    596 			}
    597 
    598 			data &= ~config_regs->mask;
    599 			data |= ((config_regs->value << config_regs->shift) & config_regs->mask);
    600 			data |= cache;
    601 
    602 			switch (config_regs->type) {
    603 			case CISLANDS_CONFIGREG_SMC_IND:
    604 				WREG32_SMC(config_regs->offset, data);
    605 				break;
    606 			case CISLANDS_CONFIGREG_DIDT_IND:
    607 				WREG32_DIDT(config_regs->offset, data);
    608 				break;
    609 			default:
    610 				WREG32(config_regs->offset << 2, data);
    611 				break;
    612 			}
    613 			cache = 0;
    614 		}
    615 		config_regs++;
    616 	}
    617 	return 0;
    618 }
    619 
    620 static int ci_enable_didt(struct radeon_device *rdev, bool enable)
    621 {
    622 	struct ci_power_info *pi = ci_get_pi(rdev);
    623 	int ret;
    624 
    625 	if (pi->caps_sq_ramping || pi->caps_db_ramping ||
    626 	    pi->caps_td_ramping || pi->caps_tcp_ramping) {
    627 		cik_enter_rlc_safe_mode(rdev);
    628 
    629 		if (enable) {
    630 			ret = ci_program_pt_config_registers(rdev, didt_config_ci);
    631 			if (ret) {
    632 				cik_exit_rlc_safe_mode(rdev);
    633 				return ret;
    634 			}
    635 		}
    636 
    637 		ci_do_enable_didt(rdev, enable);
    638 
    639 		cik_exit_rlc_safe_mode(rdev);
    640 	}
    641 
    642 	return 0;
    643 }
    644 
    645 static int ci_enable_power_containment(struct radeon_device *rdev, bool enable)
    646 {
    647 	struct ci_power_info *pi = ci_get_pi(rdev);
    648 	PPSMC_Result smc_result;
    649 	int ret = 0;
    650 
    651 	if (enable) {
    652 		pi->power_containment_features = 0;
    653 		if (pi->caps_power_containment) {
    654 			if (pi->enable_bapm_feature) {
    655 				smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_EnableDTE);
    656 				if (smc_result != PPSMC_Result_OK)
    657 					ret = -EINVAL;
    658 				else
    659 					pi->power_containment_features |= POWERCONTAINMENT_FEATURE_BAPM;
    660 			}
    661 
    662 			if (pi->enable_tdc_limit_feature) {
    663 				smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_TDCLimitEnable);
    664 				if (smc_result != PPSMC_Result_OK)
    665 					ret = -EINVAL;
    666 				else
    667 					pi->power_containment_features |= POWERCONTAINMENT_FEATURE_TDCLimit;
    668 			}
    669 
    670 			if (pi->enable_pkg_pwr_tracking_feature) {
    671 				smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_PkgPwrLimitEnable);
    672 				if (smc_result != PPSMC_Result_OK) {
    673 					ret = -EINVAL;
    674 				} else {
    675 					struct radeon_cac_tdp_table *cac_tdp_table =
    676 						rdev->pm.dpm.dyn_state.cac_tdp_table;
    677 					u32 default_pwr_limit =
    678 						(u32)(cac_tdp_table->maximum_power_delivery_limit * 256);
    679 
    680 					pi->power_containment_features |= POWERCONTAINMENT_FEATURE_PkgPwrLimit;
    681 
    682 					ci_set_power_limit(rdev, default_pwr_limit);
    683 				}
    684 			}
    685 		}
    686 	} else {
    687 		if (pi->caps_power_containment && pi->power_containment_features) {
    688 			if (pi->power_containment_features & POWERCONTAINMENT_FEATURE_TDCLimit)
    689 				ci_send_msg_to_smc(rdev, PPSMC_MSG_TDCLimitDisable);
    690 
    691 			if (pi->power_containment_features & POWERCONTAINMENT_FEATURE_BAPM)
    692 				ci_send_msg_to_smc(rdev, PPSMC_MSG_DisableDTE);
    693 
    694 			if (pi->power_containment_features & POWERCONTAINMENT_FEATURE_PkgPwrLimit)
    695 				ci_send_msg_to_smc(rdev, PPSMC_MSG_PkgPwrLimitDisable);
    696 			pi->power_containment_features = 0;
    697 		}
    698 	}
    699 
    700 	return ret;
    701 }
    702 
    703 static int ci_enable_smc_cac(struct radeon_device *rdev, bool enable)
    704 {
    705 	struct ci_power_info *pi = ci_get_pi(rdev);
    706 	PPSMC_Result smc_result;
    707 	int ret = 0;
    708 
    709 	if (pi->caps_cac) {
    710 		if (enable) {
    711 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_EnableCac);
    712 			if (smc_result != PPSMC_Result_OK) {
    713 				ret = -EINVAL;
    714 				pi->cac_enabled = false;
    715 			} else {
    716 				pi->cac_enabled = true;
    717 			}
    718 		} else if (pi->cac_enabled) {
    719 			ci_send_msg_to_smc(rdev, PPSMC_MSG_DisableCac);
    720 			pi->cac_enabled = false;
    721 		}
    722 	}
    723 
    724 	return ret;
    725 }
    726 
    727 static int ci_enable_thermal_based_sclk_dpm(struct radeon_device *rdev,
    728 					    bool enable)
    729 {
    730 	struct ci_power_info *pi = ci_get_pi(rdev);
    731 	PPSMC_Result smc_result = PPSMC_Result_OK;
    732 
    733 	if (pi->thermal_sclk_dpm_enabled) {
    734 		if (enable)
    735 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_ENABLE_THERMAL_DPM);
    736 		else
    737 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_DISABLE_THERMAL_DPM);
    738 	}
    739 
    740 	if (smc_result == PPSMC_Result_OK)
    741 		return 0;
    742 	else
    743 		return -EINVAL;
    744 }
    745 
    746 static int ci_power_control_set_level(struct radeon_device *rdev)
    747 {
    748 	struct ci_power_info *pi = ci_get_pi(rdev);
    749 	struct radeon_cac_tdp_table *cac_tdp_table =
    750 		rdev->pm.dpm.dyn_state.cac_tdp_table;
    751 	s32 adjust_percent;
    752 	s32 target_tdp;
    753 	int ret = 0;
    754 	bool adjust_polarity = false; /* ??? */
    755 
    756 	if (pi->caps_power_containment) {
    757 		adjust_percent = adjust_polarity ?
    758 			rdev->pm.dpm.tdp_adjustment : (-1 * rdev->pm.dpm.tdp_adjustment);
    759 		target_tdp = ((100 + adjust_percent) *
    760 			      (s32)cac_tdp_table->configurable_tdp) / 100;
    761 
    762 		ret = ci_set_overdrive_target_tdp(rdev, (u32)target_tdp);
    763 	}
    764 
    765 	return ret;
    766 }
    767 
    768 void ci_dpm_powergate_uvd(struct radeon_device *rdev, bool gate)
    769 {
    770 	struct ci_power_info *pi = ci_get_pi(rdev);
    771 
    772 	if (pi->uvd_power_gated == gate)
    773 		return;
    774 
    775 	pi->uvd_power_gated = gate;
    776 
    777 	ci_update_uvd_dpm(rdev, gate);
    778 }
    779 
    780 bool ci_dpm_vblank_too_short(struct radeon_device *rdev)
    781 {
    782 	struct ci_power_info *pi = ci_get_pi(rdev);
    783 	u32 vblank_time = r600_dpm_get_vblank_time(rdev);
    784 	u32 switch_limit = pi->mem_gddr5 ? 450 : 300;
    785 
    786 	/* disable mclk switching if the refresh is >120Hz, even if the
    787         * blanking period would allow it
    788         */
    789 	if (r600_dpm_get_vrefresh(rdev) > 120)
    790 		return true;
    791 
    792 	if (vblank_time < switch_limit)
    793 		return true;
    794 	else
    795 		return false;
    796 
    797 }
    798 
    799 static void ci_apply_state_adjust_rules(struct radeon_device *rdev,
    800 					struct radeon_ps *rps)
    801 {
    802 	struct ci_ps *ps = ci_get_ps(rps);
    803 	struct ci_power_info *pi = ci_get_pi(rdev);
    804 	struct radeon_clock_and_voltage_limits *max_limits;
    805 	bool disable_mclk_switching;
    806 	u32 sclk, mclk;
    807 	int i;
    808 
    809 	if (rps->vce_active) {
    810 		rps->evclk = rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].evclk;
    811 		rps->ecclk = rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].ecclk;
    812 	} else {
    813 		rps->evclk = 0;
    814 		rps->ecclk = 0;
    815 	}
    816 
    817 	if ((rdev->pm.dpm.new_active_crtc_count > 1) ||
    818 	    ci_dpm_vblank_too_short(rdev))
    819 		disable_mclk_switching = true;
    820 	else
    821 		disable_mclk_switching = false;
    822 
    823 	if ((rps->class & ATOM_PPLIB_CLASSIFICATION_UI_MASK) == ATOM_PPLIB_CLASSIFICATION_UI_BATTERY)
    824 		pi->battery_state = true;
    825 	else
    826 		pi->battery_state = false;
    827 
    828 	if (rdev->pm.dpm.ac_power)
    829 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
    830 	else
    831 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc;
    832 
    833 	if (rdev->pm.dpm.ac_power == false) {
    834 		for (i = 0; i < ps->performance_level_count; i++) {
    835 			if (ps->performance_levels[i].mclk > max_limits->mclk)
    836 				ps->performance_levels[i].mclk = max_limits->mclk;
    837 			if (ps->performance_levels[i].sclk > max_limits->sclk)
    838 				ps->performance_levels[i].sclk = max_limits->sclk;
    839 		}
    840 	}
    841 
    842 	/* XXX validate the min clocks required for display */
    843 
    844 	if (disable_mclk_switching) {
    845 		mclk  = ps->performance_levels[ps->performance_level_count - 1].mclk;
    846 		sclk = ps->performance_levels[0].sclk;
    847 	} else {
    848 		mclk = ps->performance_levels[0].mclk;
    849 		sclk = ps->performance_levels[0].sclk;
    850 	}
    851 
    852 	if (rps->vce_active) {
    853 		if (sclk < rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].sclk)
    854 			sclk = rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].sclk;
    855 		if (mclk < rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].mclk)
    856 			mclk = rdev->pm.dpm.vce_states[rdev->pm.dpm.vce_level].mclk;
    857 	}
    858 
    859 	ps->performance_levels[0].sclk = sclk;
    860 	ps->performance_levels[0].mclk = mclk;
    861 
    862 	if (ps->performance_levels[1].sclk < ps->performance_levels[0].sclk)
    863 		ps->performance_levels[1].sclk = ps->performance_levels[0].sclk;
    864 
    865 	if (disable_mclk_switching) {
    866 		if (ps->performance_levels[0].mclk < ps->performance_levels[1].mclk)
    867 			ps->performance_levels[0].mclk = ps->performance_levels[1].mclk;
    868 	} else {
    869 		if (ps->performance_levels[1].mclk < ps->performance_levels[0].mclk)
    870 			ps->performance_levels[1].mclk = ps->performance_levels[0].mclk;
    871 	}
    872 }
    873 
    874 static int ci_thermal_set_temperature_range(struct radeon_device *rdev,
    875 					    int min_temp, int max_temp)
    876 {
    877 	int low_temp = 0 * 1000;
    878 	int high_temp = 255 * 1000;
    879 	u32 tmp;
    880 
    881 	if (low_temp < min_temp)
    882 		low_temp = min_temp;
    883 	if (high_temp > max_temp)
    884 		high_temp = max_temp;
    885 	if (high_temp < low_temp) {
    886 		DRM_ERROR("invalid thermal range: %d - %d\n", low_temp, high_temp);
    887 		return -EINVAL;
    888 	}
    889 
    890 	tmp = RREG32_SMC(CG_THERMAL_INT);
    891 	tmp &= ~(CI_DIG_THERM_INTH_MASK | CI_DIG_THERM_INTL_MASK);
    892 	tmp |= CI_DIG_THERM_INTH(high_temp / 1000) |
    893 		CI_DIG_THERM_INTL(low_temp / 1000);
    894 	WREG32_SMC(CG_THERMAL_INT, tmp);
    895 
    896 #if 0
    897 	/* XXX: need to figure out how to handle this properly */
    898 	tmp = RREG32_SMC(CG_THERMAL_CTRL);
    899 	tmp &= DIG_THERM_DPM_MASK;
    900 	tmp |= DIG_THERM_DPM(high_temp / 1000);
    901 	WREG32_SMC(CG_THERMAL_CTRL, tmp);
    902 #endif
    903 
    904 	rdev->pm.dpm.thermal.min_temp = low_temp;
    905 	rdev->pm.dpm.thermal.max_temp = high_temp;
    906 
    907 	return 0;
    908 }
    909 
    910 static int ci_thermal_enable_alert(struct radeon_device *rdev,
    911 				   bool enable)
    912 {
    913 	u32 thermal_int = RREG32_SMC(CG_THERMAL_INT);
    914 	PPSMC_Result result;
    915 
    916 	if (enable) {
    917 		thermal_int &= ~(THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW);
    918 		WREG32_SMC(CG_THERMAL_INT, thermal_int);
    919 		rdev->irq.dpm_thermal = false;
    920 		result = ci_send_msg_to_smc(rdev, PPSMC_MSG_Thermal_Cntl_Enable);
    921 		if (result != PPSMC_Result_OK) {
    922 			DRM_DEBUG_KMS("Could not enable thermal interrupts.\n");
    923 			return -EINVAL;
    924 		}
    925 	} else {
    926 		thermal_int |= THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW;
    927 		WREG32_SMC(CG_THERMAL_INT, thermal_int);
    928 		rdev->irq.dpm_thermal = true;
    929 		result = ci_send_msg_to_smc(rdev, PPSMC_MSG_Thermal_Cntl_Disable);
    930 		if (result != PPSMC_Result_OK) {
    931 			DRM_DEBUG_KMS("Could not disable thermal interrupts.\n");
    932 			return -EINVAL;
    933 		}
    934 	}
    935 
    936 	return 0;
    937 }
    938 
    939 static void ci_fan_ctrl_set_static_mode(struct radeon_device *rdev, u32 mode)
    940 {
    941 	struct ci_power_info *pi = ci_get_pi(rdev);
    942 	u32 tmp;
    943 
    944 	if (pi->fan_ctrl_is_in_default_mode) {
    945 		tmp = (RREG32_SMC(CG_FDO_CTRL2) & FDO_PWM_MODE_MASK) >> FDO_PWM_MODE_SHIFT;
    946 		pi->fan_ctrl_default_mode = tmp;
    947 		tmp = (RREG32_SMC(CG_FDO_CTRL2) & TMIN_MASK) >> TMIN_SHIFT;
    948 		pi->t_min = tmp;
    949 		pi->fan_ctrl_is_in_default_mode = false;
    950 	}
    951 
    952 	tmp = RREG32_SMC(CG_FDO_CTRL2) & ~TMIN_MASK;
    953 	tmp |= TMIN(0);
    954 	WREG32_SMC(CG_FDO_CTRL2, tmp);
    955 
    956 	tmp = RREG32_SMC(CG_FDO_CTRL2) & ~FDO_PWM_MODE_MASK;
    957 	tmp |= FDO_PWM_MODE(mode);
    958 	WREG32_SMC(CG_FDO_CTRL2, tmp);
    959 }
    960 
    961 static int ci_thermal_setup_fan_table(struct radeon_device *rdev)
    962 {
    963 	struct ci_power_info *pi = ci_get_pi(rdev);
    964 	SMU7_Discrete_FanTable fan_table = { FDO_MODE_HARDWARE };
    965 	u32 duty100;
    966 	u32 t_diff1, t_diff2, pwm_diff1, pwm_diff2;
    967 	u16 fdo_min, slope1, slope2;
    968 	u32 reference_clock, tmp;
    969 	int ret;
    970 	u64 tmp64;
    971 
    972 	if (!pi->fan_table_start) {
    973 		rdev->pm.dpm.fan.ucode_fan_control = false;
    974 		return 0;
    975 	}
    976 
    977 	duty100 = (RREG32_SMC(CG_FDO_CTRL1) & FMAX_DUTY100_MASK) >> FMAX_DUTY100_SHIFT;
    978 
    979 	if (duty100 == 0) {
    980 		rdev->pm.dpm.fan.ucode_fan_control = false;
    981 		return 0;
    982 	}
    983 
    984 	tmp64 = (u64)rdev->pm.dpm.fan.pwm_min * duty100;
    985 	do_div(tmp64, 10000);
    986 	fdo_min = (u16)tmp64;
    987 
    988 	t_diff1 = rdev->pm.dpm.fan.t_med - rdev->pm.dpm.fan.t_min;
    989 	t_diff2 = rdev->pm.dpm.fan.t_high - rdev->pm.dpm.fan.t_med;
    990 
    991 	pwm_diff1 = rdev->pm.dpm.fan.pwm_med - rdev->pm.dpm.fan.pwm_min;
    992 	pwm_diff2 = rdev->pm.dpm.fan.pwm_high - rdev->pm.dpm.fan.pwm_med;
    993 
    994 	slope1 = (u16)((50 + ((16 * duty100 * pwm_diff1) / t_diff1)) / 100);
    995 	slope2 = (u16)((50 + ((16 * duty100 * pwm_diff2) / t_diff2)) / 100);
    996 
    997 	fan_table.TempMin = cpu_to_be16((50 + rdev->pm.dpm.fan.t_min) / 100);
    998 	fan_table.TempMed = cpu_to_be16((50 + rdev->pm.dpm.fan.t_med) / 100);
    999 	fan_table.TempMax = cpu_to_be16((50 + rdev->pm.dpm.fan.t_max) / 100);
   1000 
   1001 	fan_table.Slope1 = cpu_to_be16(slope1);
   1002 	fan_table.Slope2 = cpu_to_be16(slope2);
   1003 
   1004 	fan_table.FdoMin = cpu_to_be16(fdo_min);
   1005 
   1006 	fan_table.HystDown = cpu_to_be16(rdev->pm.dpm.fan.t_hyst);
   1007 
   1008 	fan_table.HystUp = cpu_to_be16(1);
   1009 
   1010 	fan_table.HystSlope = cpu_to_be16(1);
   1011 
   1012 	fan_table.TempRespLim = cpu_to_be16(5);
   1013 
   1014 	reference_clock = radeon_get_xclk(rdev);
   1015 
   1016 	fan_table.RefreshPeriod = cpu_to_be32((rdev->pm.dpm.fan.cycle_delay *
   1017 					       reference_clock) / 1600);
   1018 
   1019 	fan_table.FdoMax = cpu_to_be16((u16)duty100);
   1020 
   1021 	tmp = (RREG32_SMC(CG_MULT_THERMAL_CTRL) & TEMP_SEL_MASK) >> TEMP_SEL_SHIFT;
   1022 	fan_table.TempSrc = (uint8_t)tmp;
   1023 
   1024 	ret = ci_copy_bytes_to_smc(rdev,
   1025 				   pi->fan_table_start,
   1026 				   (u8 *)(&fan_table),
   1027 				   sizeof(fan_table),
   1028 				   pi->sram_end);
   1029 
   1030 	if (ret) {
   1031 		DRM_ERROR("Failed to load fan table to the SMC.");
   1032 		rdev->pm.dpm.fan.ucode_fan_control = false;
   1033 	}
   1034 
   1035 	return 0;
   1036 }
   1037 
   1038 static int ci_fan_ctrl_start_smc_fan_control(struct radeon_device *rdev)
   1039 {
   1040 	struct ci_power_info *pi = ci_get_pi(rdev);
   1041 	PPSMC_Result ret;
   1042 
   1043 	if (pi->caps_od_fuzzy_fan_control_support) {
   1044 		ret = ci_send_msg_to_smc_with_parameter(rdev,
   1045 							PPSMC_StartFanControl,
   1046 							FAN_CONTROL_FUZZY);
   1047 		if (ret != PPSMC_Result_OK)
   1048 			return -EINVAL;
   1049 		ret = ci_send_msg_to_smc_with_parameter(rdev,
   1050 							PPSMC_MSG_SetFanPwmMax,
   1051 							rdev->pm.dpm.fan.default_max_fan_pwm);
   1052 		if (ret != PPSMC_Result_OK)
   1053 			return -EINVAL;
   1054 	} else {
   1055 		ret = ci_send_msg_to_smc_with_parameter(rdev,
   1056 							PPSMC_StartFanControl,
   1057 							FAN_CONTROL_TABLE);
   1058 		if (ret != PPSMC_Result_OK)
   1059 			return -EINVAL;
   1060 	}
   1061 
   1062 	pi->fan_is_controlled_by_smc = true;
   1063 	return 0;
   1064 }
   1065 
   1066 static int ci_fan_ctrl_stop_smc_fan_control(struct radeon_device *rdev)
   1067 {
   1068 	PPSMC_Result ret;
   1069 	struct ci_power_info *pi = ci_get_pi(rdev);
   1070 
   1071 	ret = ci_send_msg_to_smc(rdev, PPSMC_StopFanControl);
   1072 	if (ret == PPSMC_Result_OK) {
   1073 		pi->fan_is_controlled_by_smc = false;
   1074 		return 0;
   1075 	} else
   1076 		return -EINVAL;
   1077 }
   1078 
   1079 int ci_fan_ctrl_get_fan_speed_percent(struct radeon_device *rdev,
   1080 					     u32 *speed)
   1081 {
   1082 	u32 duty, duty100;
   1083 	u64 tmp64;
   1084 
   1085 	if (rdev->pm.no_fan)
   1086 		return -ENOENT;
   1087 
   1088 	duty100 = (RREG32_SMC(CG_FDO_CTRL1) & FMAX_DUTY100_MASK) >> FMAX_DUTY100_SHIFT;
   1089 	duty = (RREG32_SMC(CG_THERMAL_STATUS) & FDO_PWM_DUTY_MASK) >> FDO_PWM_DUTY_SHIFT;
   1090 
   1091 	if (duty100 == 0)
   1092 		return -EINVAL;
   1093 
   1094 	tmp64 = (u64)duty * 100;
   1095 	do_div(tmp64, duty100);
   1096 	*speed = (u32)tmp64;
   1097 
   1098 	if (*speed > 100)
   1099 		*speed = 100;
   1100 
   1101 	return 0;
   1102 }
   1103 
   1104 int ci_fan_ctrl_set_fan_speed_percent(struct radeon_device *rdev,
   1105 					     u32 speed)
   1106 {
   1107 	u32 tmp;
   1108 	u32 duty, duty100;
   1109 	u64 tmp64;
   1110 	struct ci_power_info *pi = ci_get_pi(rdev);
   1111 
   1112 	if (rdev->pm.no_fan)
   1113 		return -ENOENT;
   1114 
   1115 	if (pi->fan_is_controlled_by_smc)
   1116 		return -EINVAL;
   1117 
   1118 	if (speed > 100)
   1119 		return -EINVAL;
   1120 
   1121 	duty100 = (RREG32_SMC(CG_FDO_CTRL1) & FMAX_DUTY100_MASK) >> FMAX_DUTY100_SHIFT;
   1122 
   1123 	if (duty100 == 0)
   1124 		return -EINVAL;
   1125 
   1126 	tmp64 = (u64)speed * duty100;
   1127 	do_div(tmp64, 100);
   1128 	duty = (u32)tmp64;
   1129 
   1130 	tmp = RREG32_SMC(CG_FDO_CTRL0) & ~FDO_STATIC_DUTY_MASK;
   1131 	tmp |= FDO_STATIC_DUTY(duty);
   1132 	WREG32_SMC(CG_FDO_CTRL0, tmp);
   1133 
   1134 	return 0;
   1135 }
   1136 
   1137 void ci_fan_ctrl_set_mode(struct radeon_device *rdev, u32 mode)
   1138 {
   1139 	if (mode) {
   1140 		/* stop auto-manage */
   1141 		if (rdev->pm.dpm.fan.ucode_fan_control)
   1142 			ci_fan_ctrl_stop_smc_fan_control(rdev);
   1143 		ci_fan_ctrl_set_static_mode(rdev, mode);
   1144 	} else {
   1145 		/* restart auto-manage */
   1146 		if (rdev->pm.dpm.fan.ucode_fan_control)
   1147 			ci_thermal_start_smc_fan_control(rdev);
   1148 		else
   1149 			ci_fan_ctrl_set_default_mode(rdev);
   1150 	}
   1151 }
   1152 
   1153 u32 ci_fan_ctrl_get_mode(struct radeon_device *rdev)
   1154 {
   1155 	struct ci_power_info *pi = ci_get_pi(rdev);
   1156 	u32 tmp;
   1157 
   1158 	if (pi->fan_is_controlled_by_smc)
   1159 		return 0;
   1160 
   1161 	tmp = RREG32_SMC(CG_FDO_CTRL2) & FDO_PWM_MODE_MASK;
   1162 	return (tmp >> FDO_PWM_MODE_SHIFT);
   1163 }
   1164 
   1165 #if 0
   1166 static int ci_fan_ctrl_get_fan_speed_rpm(struct radeon_device *rdev,
   1167 					 u32 *speed)
   1168 {
   1169 	u32 tach_period;
   1170 	u32 xclk = radeon_get_xclk(rdev);
   1171 
   1172 	if (rdev->pm.no_fan)
   1173 		return -ENOENT;
   1174 
   1175 	if (rdev->pm.fan_pulses_per_revolution == 0)
   1176 		return -ENOENT;
   1177 
   1178 	tach_period = (RREG32_SMC(CG_TACH_STATUS) & TACH_PERIOD_MASK) >> TACH_PERIOD_SHIFT;
   1179 	if (tach_period == 0)
   1180 		return -ENOENT;
   1181 
   1182 	*speed = 60 * xclk * 10000 / tach_period;
   1183 
   1184 	return 0;
   1185 }
   1186 
   1187 static int ci_fan_ctrl_set_fan_speed_rpm(struct radeon_device *rdev,
   1188 					 u32 speed)
   1189 {
   1190 	u32 tach_period, tmp;
   1191 	u32 xclk = radeon_get_xclk(rdev);
   1192 
   1193 	if (rdev->pm.no_fan)
   1194 		return -ENOENT;
   1195 
   1196 	if (rdev->pm.fan_pulses_per_revolution == 0)
   1197 		return -ENOENT;
   1198 
   1199 	if ((speed < rdev->pm.fan_min_rpm) ||
   1200 	    (speed > rdev->pm.fan_max_rpm))
   1201 		return -EINVAL;
   1202 
   1203 	if (rdev->pm.dpm.fan.ucode_fan_control)
   1204 		ci_fan_ctrl_stop_smc_fan_control(rdev);
   1205 
   1206 	tach_period = 60 * xclk * 10000 / (8 * speed);
   1207 	tmp = RREG32_SMC(CG_TACH_CTRL) & ~TARGET_PERIOD_MASK;
   1208 	tmp |= TARGET_PERIOD(tach_period);
   1209 	WREG32_SMC(CG_TACH_CTRL, tmp);
   1210 
   1211 	ci_fan_ctrl_set_static_mode(rdev, FDO_PWM_MODE_STATIC_RPM);
   1212 
   1213 	return 0;
   1214 }
   1215 #endif
   1216 
   1217 static void ci_fan_ctrl_set_default_mode(struct radeon_device *rdev)
   1218 {
   1219 	struct ci_power_info *pi = ci_get_pi(rdev);
   1220 	u32 tmp;
   1221 
   1222 	if (!pi->fan_ctrl_is_in_default_mode) {
   1223 		tmp = RREG32_SMC(CG_FDO_CTRL2) & ~FDO_PWM_MODE_MASK;
   1224 		tmp |= FDO_PWM_MODE(pi->fan_ctrl_default_mode);
   1225 		WREG32_SMC(CG_FDO_CTRL2, tmp);
   1226 
   1227 		tmp = RREG32_SMC(CG_FDO_CTRL2) & ~TMIN_MASK;
   1228 		tmp |= TMIN(pi->t_min);
   1229 		WREG32_SMC(CG_FDO_CTRL2, tmp);
   1230 		pi->fan_ctrl_is_in_default_mode = true;
   1231 	}
   1232 }
   1233 
   1234 static void ci_thermal_start_smc_fan_control(struct radeon_device *rdev)
   1235 {
   1236 	if (rdev->pm.dpm.fan.ucode_fan_control) {
   1237 		ci_fan_ctrl_start_smc_fan_control(rdev);
   1238 		ci_fan_ctrl_set_static_mode(rdev, FDO_PWM_MODE_STATIC);
   1239 	}
   1240 }
   1241 
   1242 static void ci_thermal_initialize(struct radeon_device *rdev)
   1243 {
   1244 	u32 tmp;
   1245 
   1246 	if (rdev->pm.fan_pulses_per_revolution) {
   1247 		tmp = RREG32_SMC(CG_TACH_CTRL) & ~EDGE_PER_REV_MASK;
   1248 		tmp |= EDGE_PER_REV(rdev->pm.fan_pulses_per_revolution -1);
   1249 		WREG32_SMC(CG_TACH_CTRL, tmp);
   1250 	}
   1251 
   1252 	tmp = RREG32_SMC(CG_FDO_CTRL2) & ~TACH_PWM_RESP_RATE_MASK;
   1253 	tmp |= TACH_PWM_RESP_RATE(0x28);
   1254 	WREG32_SMC(CG_FDO_CTRL2, tmp);
   1255 }
   1256 
   1257 static int ci_thermal_start_thermal_controller(struct radeon_device *rdev)
   1258 {
   1259 	int ret;
   1260 
   1261 	ci_thermal_initialize(rdev);
   1262 	ret = ci_thermal_set_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX);
   1263 	if (ret)
   1264 		return ret;
   1265 	ret = ci_thermal_enable_alert(rdev, true);
   1266 	if (ret)
   1267 		return ret;
   1268 	if (rdev->pm.dpm.fan.ucode_fan_control) {
   1269 		ret = ci_thermal_setup_fan_table(rdev);
   1270 		if (ret)
   1271 			return ret;
   1272 		ci_thermal_start_smc_fan_control(rdev);
   1273 	}
   1274 
   1275 	return 0;
   1276 }
   1277 
   1278 static void ci_thermal_stop_thermal_controller(struct radeon_device *rdev)
   1279 {
   1280 	if (!rdev->pm.no_fan)
   1281 		ci_fan_ctrl_set_default_mode(rdev);
   1282 }
   1283 
   1284 #if 0
   1285 static int ci_read_smc_soft_register(struct radeon_device *rdev,
   1286 				     u16 reg_offset, u32 *value)
   1287 {
   1288 	struct ci_power_info *pi = ci_get_pi(rdev);
   1289 
   1290 	return ci_read_smc_sram_dword(rdev,
   1291 				      pi->soft_regs_start + reg_offset,
   1292 				      value, pi->sram_end);
   1293 }
   1294 #endif
   1295 
   1296 static int ci_write_smc_soft_register(struct radeon_device *rdev,
   1297 				      u16 reg_offset, u32 value)
   1298 {
   1299 	struct ci_power_info *pi = ci_get_pi(rdev);
   1300 
   1301 	return ci_write_smc_sram_dword(rdev,
   1302 				       pi->soft_regs_start + reg_offset,
   1303 				       value, pi->sram_end);
   1304 }
   1305 
   1306 static void ci_init_fps_limits(struct radeon_device *rdev)
   1307 {
   1308 	struct ci_power_info *pi = ci_get_pi(rdev);
   1309 	SMU7_Discrete_DpmTable *table = &pi->smc_state_table;
   1310 
   1311 	if (pi->caps_fps) {
   1312 		u16 tmp;
   1313 
   1314 		tmp = 45;
   1315 		table->FpsHighT = cpu_to_be16(tmp);
   1316 
   1317 		tmp = 30;
   1318 		table->FpsLowT = cpu_to_be16(tmp);
   1319 	}
   1320 }
   1321 
   1322 static int ci_update_sclk_t(struct radeon_device *rdev)
   1323 {
   1324 	struct ci_power_info *pi = ci_get_pi(rdev);
   1325 	int ret = 0;
   1326 	u32 low_sclk_interrupt_t = 0;
   1327 
   1328 	if (pi->caps_sclk_throttle_low_notification) {
   1329 		low_sclk_interrupt_t = cpu_to_be32(pi->low_sclk_interrupt_t);
   1330 
   1331 		ret = ci_copy_bytes_to_smc(rdev,
   1332 					   pi->dpm_table_start +
   1333 					   offsetof(SMU7_Discrete_DpmTable, LowSclkInterruptT),
   1334 					   (u8 *)&low_sclk_interrupt_t,
   1335 					   sizeof(u32), pi->sram_end);
   1336 
   1337 	}
   1338 
   1339 	return ret;
   1340 }
   1341 
   1342 static void ci_get_leakage_voltages(struct radeon_device *rdev)
   1343 {
   1344 	struct ci_power_info *pi = ci_get_pi(rdev);
   1345 	u16 leakage_id, virtual_voltage_id;
   1346 	u16 vddc, vddci;
   1347 	int i;
   1348 
   1349 	pi->vddc_leakage.count = 0;
   1350 	pi->vddci_leakage.count = 0;
   1351 
   1352 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_EVV) {
   1353 		for (i = 0; i < CISLANDS_MAX_LEAKAGE_COUNT; i++) {
   1354 			virtual_voltage_id = ATOM_VIRTUAL_VOLTAGE_ID0 + i;
   1355 			if (radeon_atom_get_voltage_evv(rdev, virtual_voltage_id, &vddc) != 0)
   1356 				continue;
   1357 			if (vddc != 0 && vddc != virtual_voltage_id) {
   1358 				pi->vddc_leakage.actual_voltage[pi->vddc_leakage.count] = vddc;
   1359 				pi->vddc_leakage.leakage_id[pi->vddc_leakage.count] = virtual_voltage_id;
   1360 				pi->vddc_leakage.count++;
   1361 			}
   1362 		}
   1363 	} else if (radeon_atom_get_leakage_id_from_vbios(rdev, &leakage_id) == 0) {
   1364 		for (i = 0; i < CISLANDS_MAX_LEAKAGE_COUNT; i++) {
   1365 			virtual_voltage_id = ATOM_VIRTUAL_VOLTAGE_ID0 + i;
   1366 			if (radeon_atom_get_leakage_vddc_based_on_leakage_params(rdev, &vddc, &vddci,
   1367 										 virtual_voltage_id,
   1368 										 leakage_id) == 0) {
   1369 				if (vddc != 0 && vddc != virtual_voltage_id) {
   1370 					pi->vddc_leakage.actual_voltage[pi->vddc_leakage.count] = vddc;
   1371 					pi->vddc_leakage.leakage_id[pi->vddc_leakage.count] = virtual_voltage_id;
   1372 					pi->vddc_leakage.count++;
   1373 				}
   1374 				if (vddci != 0 && vddci != virtual_voltage_id) {
   1375 					pi->vddci_leakage.actual_voltage[pi->vddci_leakage.count] = vddci;
   1376 					pi->vddci_leakage.leakage_id[pi->vddci_leakage.count] = virtual_voltage_id;
   1377 					pi->vddci_leakage.count++;
   1378 				}
   1379 			}
   1380 		}
   1381 	}
   1382 }
   1383 
   1384 static void ci_set_dpm_event_sources(struct radeon_device *rdev, u32 sources)
   1385 {
   1386 	struct ci_power_info *pi = ci_get_pi(rdev);
   1387 	bool want_thermal_protection;
   1388 	enum radeon_dpm_event_src dpm_event_src;
   1389 	u32 tmp;
   1390 
   1391 	switch (sources) {
   1392 	case 0:
   1393 	default:
   1394 		want_thermal_protection = false;
   1395 		break;
   1396 	case (1 << RADEON_DPM_AUTO_THROTTLE_SRC_THERMAL):
   1397 		want_thermal_protection = true;
   1398 		dpm_event_src = RADEON_DPM_EVENT_SRC_DIGITAL;
   1399 		break;
   1400 	case (1 << RADEON_DPM_AUTO_THROTTLE_SRC_EXTERNAL):
   1401 		want_thermal_protection = true;
   1402 		dpm_event_src = RADEON_DPM_EVENT_SRC_EXTERNAL;
   1403 		break;
   1404 	case ((1 << RADEON_DPM_AUTO_THROTTLE_SRC_EXTERNAL) |
   1405 	      (1 << RADEON_DPM_AUTO_THROTTLE_SRC_THERMAL)):
   1406 		want_thermal_protection = true;
   1407 		dpm_event_src = RADEON_DPM_EVENT_SRC_DIGIAL_OR_EXTERNAL;
   1408 		break;
   1409 	}
   1410 
   1411 	if (want_thermal_protection) {
   1412 #if 0
   1413 		/* XXX: need to figure out how to handle this properly */
   1414 		tmp = RREG32_SMC(CG_THERMAL_CTRL);
   1415 		tmp &= DPM_EVENT_SRC_MASK;
   1416 		tmp |= DPM_EVENT_SRC(dpm_event_src);
   1417 		WREG32_SMC(CG_THERMAL_CTRL, tmp);
   1418 #else
   1419 		(void)dpm_event_src;
   1420 #endif
   1421 
   1422 		tmp = RREG32_SMC(GENERAL_PWRMGT);
   1423 		if (pi->thermal_protection)
   1424 			tmp &= ~THERMAL_PROTECTION_DIS;
   1425 		else
   1426 			tmp |= THERMAL_PROTECTION_DIS;
   1427 		WREG32_SMC(GENERAL_PWRMGT, tmp);
   1428 	} else {
   1429 		tmp = RREG32_SMC(GENERAL_PWRMGT);
   1430 		tmp |= THERMAL_PROTECTION_DIS;
   1431 		WREG32_SMC(GENERAL_PWRMGT, tmp);
   1432 	}
   1433 }
   1434 
   1435 static void ci_enable_auto_throttle_source(struct radeon_device *rdev,
   1436 					   enum radeon_dpm_auto_throttle_src source,
   1437 					   bool enable)
   1438 {
   1439 	struct ci_power_info *pi = ci_get_pi(rdev);
   1440 
   1441 	if (enable) {
   1442 		if (!(pi->active_auto_throttle_sources & (1 << source))) {
   1443 			pi->active_auto_throttle_sources |= 1 << source;
   1444 			ci_set_dpm_event_sources(rdev, pi->active_auto_throttle_sources);
   1445 		}
   1446 	} else {
   1447 		if (pi->active_auto_throttle_sources & (1 << source)) {
   1448 			pi->active_auto_throttle_sources &= ~(1 << source);
   1449 			ci_set_dpm_event_sources(rdev, pi->active_auto_throttle_sources);
   1450 		}
   1451 	}
   1452 }
   1453 
   1454 static void ci_enable_vr_hot_gpio_interrupt(struct radeon_device *rdev)
   1455 {
   1456 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_REGULATOR_HOT)
   1457 		ci_send_msg_to_smc(rdev, PPSMC_MSG_EnableVRHotGPIOInterrupt);
   1458 }
   1459 
   1460 static int ci_unfreeze_sclk_mclk_dpm(struct radeon_device *rdev)
   1461 {
   1462 	struct ci_power_info *pi = ci_get_pi(rdev);
   1463 	PPSMC_Result smc_result;
   1464 
   1465 	if (!pi->need_update_smu7_dpm_table)
   1466 		return 0;
   1467 
   1468 	if ((!pi->sclk_dpm_key_disabled) &&
   1469 	    (pi->need_update_smu7_dpm_table & (DPMTABLE_OD_UPDATE_SCLK | DPMTABLE_UPDATE_SCLK))) {
   1470 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_SCLKDPM_UnfreezeLevel);
   1471 		if (smc_result != PPSMC_Result_OK)
   1472 			return -EINVAL;
   1473 	}
   1474 
   1475 	if ((!pi->mclk_dpm_key_disabled) &&
   1476 	    (pi->need_update_smu7_dpm_table & DPMTABLE_OD_UPDATE_MCLK)) {
   1477 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_MCLKDPM_UnfreezeLevel);
   1478 		if (smc_result != PPSMC_Result_OK)
   1479 			return -EINVAL;
   1480 	}
   1481 
   1482 	pi->need_update_smu7_dpm_table = 0;
   1483 	return 0;
   1484 }
   1485 
   1486 static int ci_enable_sclk_mclk_dpm(struct radeon_device *rdev, bool enable)
   1487 {
   1488 	struct ci_power_info *pi = ci_get_pi(rdev);
   1489 	PPSMC_Result smc_result;
   1490 
   1491 	if (enable) {
   1492 		if (!pi->sclk_dpm_key_disabled) {
   1493 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_DPM_Enable);
   1494 			if (smc_result != PPSMC_Result_OK)
   1495 				return -EINVAL;
   1496 		}
   1497 
   1498 		if (!pi->mclk_dpm_key_disabled) {
   1499 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_MCLKDPM_Enable);
   1500 			if (smc_result != PPSMC_Result_OK)
   1501 				return -EINVAL;
   1502 
   1503 			WREG32_P(MC_SEQ_CNTL_3, CAC_EN, ~CAC_EN);
   1504 
   1505 			WREG32_SMC(LCAC_MC0_CNTL, 0x05);
   1506 			WREG32_SMC(LCAC_MC1_CNTL, 0x05);
   1507 			WREG32_SMC(LCAC_CPL_CNTL, 0x100005);
   1508 
   1509 			udelay(10);
   1510 
   1511 			WREG32_SMC(LCAC_MC0_CNTL, 0x400005);
   1512 			WREG32_SMC(LCAC_MC1_CNTL, 0x400005);
   1513 			WREG32_SMC(LCAC_CPL_CNTL, 0x500005);
   1514 		}
   1515 	} else {
   1516 		if (!pi->sclk_dpm_key_disabled) {
   1517 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_DPM_Disable);
   1518 			if (smc_result != PPSMC_Result_OK)
   1519 				return -EINVAL;
   1520 		}
   1521 
   1522 		if (!pi->mclk_dpm_key_disabled) {
   1523 			smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_MCLKDPM_Disable);
   1524 			if (smc_result != PPSMC_Result_OK)
   1525 				return -EINVAL;
   1526 		}
   1527 	}
   1528 
   1529 	return 0;
   1530 }
   1531 
   1532 static int ci_start_dpm(struct radeon_device *rdev)
   1533 {
   1534 	struct ci_power_info *pi = ci_get_pi(rdev);
   1535 	PPSMC_Result smc_result;
   1536 	int ret;
   1537 	u32 tmp;
   1538 
   1539 	tmp = RREG32_SMC(GENERAL_PWRMGT);
   1540 	tmp |= GLOBAL_PWRMGT_EN;
   1541 	WREG32_SMC(GENERAL_PWRMGT, tmp);
   1542 
   1543 	tmp = RREG32_SMC(SCLK_PWRMGT_CNTL);
   1544 	tmp |= DYNAMIC_PM_EN;
   1545 	WREG32_SMC(SCLK_PWRMGT_CNTL, tmp);
   1546 
   1547 	ci_write_smc_soft_register(rdev, offsetof(SMU7_SoftRegisters, VoltageChangeTimeout), 0x1000);
   1548 
   1549 	WREG32_P(BIF_LNCNT_RESET, 0, ~RESET_LNCNT_EN);
   1550 
   1551 	smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_Voltage_Cntl_Enable);
   1552 	if (smc_result != PPSMC_Result_OK)
   1553 		return -EINVAL;
   1554 
   1555 	ret = ci_enable_sclk_mclk_dpm(rdev, true);
   1556 	if (ret)
   1557 		return ret;
   1558 
   1559 	if (!pi->pcie_dpm_key_disabled) {
   1560 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_PCIeDPM_Enable);
   1561 		if (smc_result != PPSMC_Result_OK)
   1562 			return -EINVAL;
   1563 	}
   1564 
   1565 	return 0;
   1566 }
   1567 
   1568 static int ci_freeze_sclk_mclk_dpm(struct radeon_device *rdev)
   1569 {
   1570 	struct ci_power_info *pi = ci_get_pi(rdev);
   1571 	PPSMC_Result smc_result;
   1572 
   1573 	if (!pi->need_update_smu7_dpm_table)
   1574 		return 0;
   1575 
   1576 	if ((!pi->sclk_dpm_key_disabled) &&
   1577 	    (pi->need_update_smu7_dpm_table & (DPMTABLE_OD_UPDATE_SCLK | DPMTABLE_UPDATE_SCLK))) {
   1578 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_SCLKDPM_FreezeLevel);
   1579 		if (smc_result != PPSMC_Result_OK)
   1580 			return -EINVAL;
   1581 	}
   1582 
   1583 	if ((!pi->mclk_dpm_key_disabled) &&
   1584 	    (pi->need_update_smu7_dpm_table & DPMTABLE_OD_UPDATE_MCLK)) {
   1585 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_MCLKDPM_FreezeLevel);
   1586 		if (smc_result != PPSMC_Result_OK)
   1587 			return -EINVAL;
   1588 	}
   1589 
   1590 	return 0;
   1591 }
   1592 
   1593 static int ci_stop_dpm(struct radeon_device *rdev)
   1594 {
   1595 	struct ci_power_info *pi = ci_get_pi(rdev);
   1596 	PPSMC_Result smc_result;
   1597 	int ret;
   1598 	u32 tmp;
   1599 
   1600 	tmp = RREG32_SMC(GENERAL_PWRMGT);
   1601 	tmp &= ~GLOBAL_PWRMGT_EN;
   1602 	WREG32_SMC(GENERAL_PWRMGT, tmp);
   1603 
   1604 	tmp = RREG32_SMC(SCLK_PWRMGT_CNTL);
   1605 	tmp &= ~DYNAMIC_PM_EN;
   1606 	WREG32_SMC(SCLK_PWRMGT_CNTL, tmp);
   1607 
   1608 	if (!pi->pcie_dpm_key_disabled) {
   1609 		smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_PCIeDPM_Disable);
   1610 		if (smc_result != PPSMC_Result_OK)
   1611 			return -EINVAL;
   1612 	}
   1613 
   1614 	ret = ci_enable_sclk_mclk_dpm(rdev, false);
   1615 	if (ret)
   1616 		return ret;
   1617 
   1618 	smc_result = ci_send_msg_to_smc(rdev, PPSMC_MSG_Voltage_Cntl_Disable);
   1619 	if (smc_result != PPSMC_Result_OK)
   1620 		return -EINVAL;
   1621 
   1622 	return 0;
   1623 }
   1624 
   1625 static void ci_enable_sclk_control(struct radeon_device *rdev, bool enable)
   1626 {
   1627 	u32 tmp = RREG32_SMC(SCLK_PWRMGT_CNTL);
   1628 
   1629 	if (enable)
   1630 		tmp &= ~SCLK_PWRMGT_OFF;
   1631 	else
   1632 		tmp |= SCLK_PWRMGT_OFF;
   1633 	WREG32_SMC(SCLK_PWRMGT_CNTL, tmp);
   1634 }
   1635 
   1636 #if 0
   1637 static int ci_notify_hw_of_power_source(struct radeon_device *rdev,
   1638 					bool ac_power)
   1639 {
   1640 	struct ci_power_info *pi = ci_get_pi(rdev);
   1641 	struct radeon_cac_tdp_table *cac_tdp_table =
   1642 		rdev->pm.dpm.dyn_state.cac_tdp_table;
   1643 	u32 power_limit;
   1644 
   1645 	if (ac_power)
   1646 		power_limit = (u32)(cac_tdp_table->maximum_power_delivery_limit * 256);
   1647 	else
   1648 		power_limit = (u32)(cac_tdp_table->battery_power_limit * 256);
   1649 
   1650 	ci_set_power_limit(rdev, power_limit);
   1651 
   1652 	if (pi->caps_automatic_dc_transition) {
   1653 		if (ac_power)
   1654 			ci_send_msg_to_smc(rdev, PPSMC_MSG_RunningOnAC);
   1655 		else
   1656 			ci_send_msg_to_smc(rdev, PPSMC_MSG_Remove_DC_Clamp);
   1657 	}
   1658 
   1659 	return 0;
   1660 }
   1661 #endif
   1662 
   1663 static PPSMC_Result ci_send_msg_to_smc(struct radeon_device *rdev, PPSMC_Msg msg)
   1664 {
   1665 	u32 tmp;
   1666 	int i;
   1667 
   1668 	if (!ci_is_smc_running(rdev))
   1669 		return PPSMC_Result_Failed;
   1670 
   1671 	WREG32(SMC_MESSAGE_0, msg);
   1672 
   1673 	for (i = 0; i < rdev->usec_timeout; i++) {
   1674 		tmp = RREG32(SMC_RESP_0);
   1675 		if (tmp != 0)
   1676 			break;
   1677 		udelay(1);
   1678 	}
   1679 	tmp = RREG32(SMC_RESP_0);
   1680 
   1681 	return (PPSMC_Result)tmp;
   1682 }
   1683 
   1684 static PPSMC_Result ci_send_msg_to_smc_with_parameter(struct radeon_device *rdev,
   1685 						      PPSMC_Msg msg, u32 parameter)
   1686 {
   1687 	WREG32(SMC_MSG_ARG_0, parameter);
   1688 	return ci_send_msg_to_smc(rdev, msg);
   1689 }
   1690 
   1691 static PPSMC_Result ci_send_msg_to_smc_return_parameter(struct radeon_device *rdev,
   1692 							PPSMC_Msg msg, u32 *parameter)
   1693 {
   1694 	PPSMC_Result smc_result;
   1695 
   1696 	smc_result = ci_send_msg_to_smc(rdev, msg);
   1697 
   1698 	if ((smc_result == PPSMC_Result_OK) && parameter)
   1699 		*parameter = RREG32(SMC_MSG_ARG_0);
   1700 
   1701 	return smc_result;
   1702 }
   1703 
   1704 static int ci_dpm_force_state_sclk(struct radeon_device *rdev, u32 n)
   1705 {
   1706 	struct ci_power_info *pi = ci_get_pi(rdev);
   1707 
   1708 	if (!pi->sclk_dpm_key_disabled) {
   1709 		PPSMC_Result smc_result =
   1710 			ci_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_SCLKDPM_SetEnabledMask, 1 << n);
   1711 		if (smc_result != PPSMC_Result_OK)
   1712 			return -EINVAL;
   1713 	}
   1714 
   1715 	return 0;
   1716 }
   1717 
   1718 static int ci_dpm_force_state_mclk(struct radeon_device *rdev, u32 n)
   1719 {
   1720 	struct ci_power_info *pi = ci_get_pi(rdev);
   1721 
   1722 	if (!pi->mclk_dpm_key_disabled) {
   1723 		PPSMC_Result smc_result =
   1724 			ci_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_MCLKDPM_SetEnabledMask, 1 << n);
   1725 		if (smc_result != PPSMC_Result_OK)
   1726 			return -EINVAL;
   1727 	}
   1728 
   1729 	return 0;
   1730 }
   1731 
   1732 static int ci_dpm_force_state_pcie(struct radeon_device *rdev, u32 n)
   1733 {
   1734 	struct ci_power_info *pi = ci_get_pi(rdev);
   1735 
   1736 	if (!pi->pcie_dpm_key_disabled) {
   1737 		PPSMC_Result smc_result =
   1738 			ci_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_PCIeDPM_ForceLevel, n);
   1739 		if (smc_result != PPSMC_Result_OK)
   1740 			return -EINVAL;
   1741 	}
   1742 
   1743 	return 0;
   1744 }
   1745 
   1746 static int ci_set_power_limit(struct radeon_device *rdev, u32 n)
   1747 {
   1748 	struct ci_power_info *pi = ci_get_pi(rdev);
   1749 
   1750 	if (pi->power_containment_features & POWERCONTAINMENT_FEATURE_PkgPwrLimit) {
   1751 		PPSMC_Result smc_result =
   1752 			ci_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_PkgPwrSetLimit, n);
   1753 		if (smc_result != PPSMC_Result_OK)
   1754 			return -EINVAL;
   1755 	}
   1756 
   1757 	return 0;
   1758 }
   1759 
   1760 static int ci_set_overdrive_target_tdp(struct radeon_device *rdev,
   1761 				       u32 target_tdp)
   1762 {
   1763 	PPSMC_Result smc_result =
   1764 		ci_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_OverDriveSetTargetTdp, target_tdp);
   1765 	if (smc_result != PPSMC_Result_OK)
   1766 		return -EINVAL;
   1767 	return 0;
   1768 }
   1769 
   1770 #if 0
   1771 static int ci_set_boot_state(struct radeon_device *rdev)
   1772 {
   1773 	return ci_enable_sclk_mclk_dpm(rdev, false);
   1774 }
   1775 #endif
   1776 
   1777 static u32 ci_get_average_sclk_freq(struct radeon_device *rdev)
   1778 {
   1779 	u32 sclk_freq;
   1780 	PPSMC_Result smc_result =
   1781 		ci_send_msg_to_smc_return_parameter(rdev,
   1782 						    PPSMC_MSG_API_GetSclkFrequency,
   1783 						    &sclk_freq);
   1784 	if (smc_result != PPSMC_Result_OK)
   1785 		sclk_freq = 0;
   1786 
   1787 	return sclk_freq;
   1788 }
   1789 
   1790 static u32 ci_get_average_mclk_freq(struct radeon_device *rdev)
   1791 {
   1792 	u32 mclk_freq;
   1793 	PPSMC_Result smc_result =
   1794 		ci_send_msg_to_smc_return_parameter(rdev,
   1795 						    PPSMC_MSG_API_GetMclkFrequency,
   1796 						    &mclk_freq);
   1797 	if (smc_result != PPSMC_Result_OK)
   1798 		mclk_freq = 0;
   1799 
   1800 	return mclk_freq;
   1801 }
   1802 
   1803 static void ci_dpm_start_smc(struct radeon_device *rdev)
   1804 {
   1805 	int i;
   1806 
   1807 	ci_program_jump_on_start(rdev);
   1808 	ci_start_smc_clock(rdev);
   1809 	ci_start_smc(rdev);
   1810 	for (i = 0; i < rdev->usec_timeout; i++) {
   1811 		if (RREG32_SMC(FIRMWARE_FLAGS) & INTERRUPTS_ENABLED)
   1812 			break;
   1813 	}
   1814 }
   1815 
   1816 static void ci_dpm_stop_smc(struct radeon_device *rdev)
   1817 {
   1818 	ci_reset_smc(rdev);
   1819 	ci_stop_smc_clock(rdev);
   1820 }
   1821 
   1822 static int ci_process_firmware_header(struct radeon_device *rdev)
   1823 {
   1824 	struct ci_power_info *pi = ci_get_pi(rdev);
   1825 	u32 tmp;
   1826 	int ret;
   1827 
   1828 	ret = ci_read_smc_sram_dword(rdev,
   1829 				     SMU7_FIRMWARE_HEADER_LOCATION +
   1830 				     offsetof(SMU7_Firmware_Header, DpmTable),
   1831 				     &tmp, pi->sram_end);
   1832 	if (ret)
   1833 		return ret;
   1834 
   1835 	pi->dpm_table_start = tmp;
   1836 
   1837 	ret = ci_read_smc_sram_dword(rdev,
   1838 				     SMU7_FIRMWARE_HEADER_LOCATION +
   1839 				     offsetof(SMU7_Firmware_Header, SoftRegisters),
   1840 				     &tmp, pi->sram_end);
   1841 	if (ret)
   1842 		return ret;
   1843 
   1844 	pi->soft_regs_start = tmp;
   1845 
   1846 	ret = ci_read_smc_sram_dword(rdev,
   1847 				     SMU7_FIRMWARE_HEADER_LOCATION +
   1848 				     offsetof(SMU7_Firmware_Header, mcRegisterTable),
   1849 				     &tmp, pi->sram_end);
   1850 	if (ret)
   1851 		return ret;
   1852 
   1853 	pi->mc_reg_table_start = tmp;
   1854 
   1855 	ret = ci_read_smc_sram_dword(rdev,
   1856 				     SMU7_FIRMWARE_HEADER_LOCATION +
   1857 				     offsetof(SMU7_Firmware_Header, FanTable),
   1858 				     &tmp, pi->sram_end);
   1859 	if (ret)
   1860 		return ret;
   1861 
   1862 	pi->fan_table_start = tmp;
   1863 
   1864 	ret = ci_read_smc_sram_dword(rdev,
   1865 				     SMU7_FIRMWARE_HEADER_LOCATION +
   1866 				     offsetof(SMU7_Firmware_Header, mcArbDramTimingTable),
   1867 				     &tmp, pi->sram_end);
   1868 	if (ret)
   1869 		return ret;
   1870 
   1871 	pi->arb_table_start = tmp;
   1872 
   1873 	return 0;
   1874 }
   1875 
   1876 static void ci_read_clock_registers(struct radeon_device *rdev)
   1877 {
   1878 	struct ci_power_info *pi = ci_get_pi(rdev);
   1879 
   1880 	pi->clock_registers.cg_spll_func_cntl =
   1881 		RREG32_SMC(CG_SPLL_FUNC_CNTL);
   1882 	pi->clock_registers.cg_spll_func_cntl_2 =
   1883 		RREG32_SMC(CG_SPLL_FUNC_CNTL_2);
   1884 	pi->clock_registers.cg_spll_func_cntl_3 =
   1885 		RREG32_SMC(CG_SPLL_FUNC_CNTL_3);
   1886 	pi->clock_registers.cg_spll_func_cntl_4 =
   1887 		RREG32_SMC(CG_SPLL_FUNC_CNTL_4);
   1888 	pi->clock_registers.cg_spll_spread_spectrum =
   1889 		RREG32_SMC(CG_SPLL_SPREAD_SPECTRUM);
   1890 	pi->clock_registers.cg_spll_spread_spectrum_2 =
   1891 		RREG32_SMC(CG_SPLL_SPREAD_SPECTRUM_2);
   1892 	pi->clock_registers.dll_cntl = RREG32(DLL_CNTL);
   1893 	pi->clock_registers.mclk_pwrmgt_cntl = RREG32(MCLK_PWRMGT_CNTL);
   1894 	pi->clock_registers.mpll_ad_func_cntl = RREG32(MPLL_AD_FUNC_CNTL);
   1895 	pi->clock_registers.mpll_dq_func_cntl = RREG32(MPLL_DQ_FUNC_CNTL);
   1896 	pi->clock_registers.mpll_func_cntl = RREG32(MPLL_FUNC_CNTL);
   1897 	pi->clock_registers.mpll_func_cntl_1 = RREG32(MPLL_FUNC_CNTL_1);
   1898 	pi->clock_registers.mpll_func_cntl_2 = RREG32(MPLL_FUNC_CNTL_2);
   1899 	pi->clock_registers.mpll_ss1 = RREG32(MPLL_SS1);
   1900 	pi->clock_registers.mpll_ss2 = RREG32(MPLL_SS2);
   1901 }
   1902 
   1903 static void ci_init_sclk_t(struct radeon_device *rdev)
   1904 {
   1905 	struct ci_power_info *pi = ci_get_pi(rdev);
   1906 
   1907 	pi->low_sclk_interrupt_t = 0;
   1908 }
   1909 
   1910 static void ci_enable_thermal_protection(struct radeon_device *rdev,
   1911 					 bool enable)
   1912 {
   1913 	u32 tmp = RREG32_SMC(GENERAL_PWRMGT);
   1914 
   1915 	if (enable)
   1916 		tmp &= ~THERMAL_PROTECTION_DIS;
   1917 	else
   1918 		tmp |= THERMAL_PROTECTION_DIS;
   1919 	WREG32_SMC(GENERAL_PWRMGT, tmp);
   1920 }
   1921 
   1922 static void ci_enable_acpi_power_management(struct radeon_device *rdev)
   1923 {
   1924 	u32 tmp = RREG32_SMC(GENERAL_PWRMGT);
   1925 
   1926 	tmp |= STATIC_PM_EN;
   1927 
   1928 	WREG32_SMC(GENERAL_PWRMGT, tmp);
   1929 }
   1930 
   1931 #if 0
   1932 static int ci_enter_ulp_state(struct radeon_device *rdev)
   1933 {
   1934 
   1935 	WREG32(SMC_MESSAGE_0, PPSMC_MSG_SwitchToMinimumPower);
   1936 
   1937 	udelay(25000);
   1938 
   1939 	return 0;
   1940 }
   1941 
   1942 static int ci_exit_ulp_state(struct radeon_device *rdev)
   1943 {
   1944 	int i;
   1945 
   1946 	WREG32(SMC_MESSAGE_0, PPSMC_MSG_ResumeFromMinimumPower);
   1947 
   1948 	udelay(7000);
   1949 
   1950 	for (i = 0; i < rdev->usec_timeout; i++) {
   1951 		if (RREG32(SMC_RESP_0) == 1)
   1952 			break;
   1953 		udelay(1000);
   1954 	}
   1955 
   1956 	return 0;
   1957 }
   1958 #endif
   1959 
   1960 static int ci_notify_smc_display_change(struct radeon_device *rdev,
   1961 					bool has_display)
   1962 {
   1963 	PPSMC_Msg msg = has_display ? PPSMC_MSG_HasDisplay : PPSMC_MSG_NoDisplay;
   1964 
   1965 	return (ci_send_msg_to_smc(rdev, msg) == PPSMC_Result_OK) ?  0 : -EINVAL;
   1966 }
   1967 
   1968 static int ci_enable_ds_master_switch(struct radeon_device *rdev,
   1969 				      bool enable)
   1970 {
   1971 	struct ci_power_info *pi = ci_get_pi(rdev);
   1972 
   1973 	if (enable) {
   1974 		if (pi->caps_sclk_ds) {
   1975 			if (ci_send_msg_to_smc(rdev, PPSMC_MSG_MASTER_DeepSleep_ON) != PPSMC_Result_OK)
   1976 				return -EINVAL;
   1977 		} else {
   1978 			if (ci_send_msg_to_smc(rdev, PPSMC_MSG_MASTER_DeepSleep_OFF) != PPSMC_Result_OK)
   1979 				return -EINVAL;
   1980 		}
   1981 	} else {
   1982 		if (pi->caps_sclk_ds) {
   1983 			if (ci_send_msg_to_smc(rdev, PPSMC_MSG_MASTER_DeepSleep_OFF) != PPSMC_Result_OK)
   1984 				return -EINVAL;
   1985 		}
   1986 	}
   1987 
   1988 	return 0;
   1989 }
   1990 
   1991 static void ci_program_display_gap(struct radeon_device *rdev)
   1992 {
   1993 	u32 tmp = RREG32_SMC(CG_DISPLAY_GAP_CNTL);
   1994 	u32 pre_vbi_time_in_us;
   1995 	u32 frame_time_in_us;
   1996 	u32 ref_clock = rdev->clock.spll.reference_freq;
   1997 	u32 refresh_rate = r600_dpm_get_vrefresh(rdev);
   1998 	u32 vblank_time = r600_dpm_get_vblank_time(rdev);
   1999 
   2000 	tmp &= ~DISP_GAP_MASK;
   2001 	if (rdev->pm.dpm.new_active_crtc_count > 0)
   2002 		tmp |= DISP_GAP(R600_PM_DISPLAY_GAP_VBLANK_OR_WM);
   2003 	else
   2004 		tmp |= DISP_GAP(R600_PM_DISPLAY_GAP_IGNORE);
   2005 	WREG32_SMC(CG_DISPLAY_GAP_CNTL, tmp);
   2006 
   2007 	if (refresh_rate == 0)
   2008 		refresh_rate = 60;
   2009 	if (vblank_time == 0xffffffff)
   2010 		vblank_time = 500;
   2011 	frame_time_in_us = 1000000 / refresh_rate;
   2012 	pre_vbi_time_in_us =
   2013 		frame_time_in_us - 200 - vblank_time;
   2014 	tmp = pre_vbi_time_in_us * (ref_clock / 100);
   2015 
   2016 	WREG32_SMC(CG_DISPLAY_GAP_CNTL2, tmp);
   2017 	ci_write_smc_soft_register(rdev, offsetof(SMU7_SoftRegisters, PreVBlankGap), 0x64);
   2018 	ci_write_smc_soft_register(rdev, offsetof(SMU7_SoftRegisters, VBlankTimeout), (frame_time_in_us - pre_vbi_time_in_us));
   2019 
   2020 
   2021 	ci_notify_smc_display_change(rdev, (rdev->pm.dpm.new_active_crtc_count == 1));
   2022 
   2023 }
   2024 
   2025 static void ci_enable_spread_spectrum(struct radeon_device *rdev, bool enable)
   2026 {
   2027 	struct ci_power_info *pi = ci_get_pi(rdev);
   2028 	u32 tmp;
   2029 
   2030 	if (enable) {
   2031 		if (pi->caps_sclk_ss_support) {
   2032 			tmp = RREG32_SMC(GENERAL_PWRMGT);
   2033 			tmp |= DYN_SPREAD_SPECTRUM_EN;
   2034 			WREG32_SMC(GENERAL_PWRMGT, tmp);
   2035 		}
   2036 	} else {
   2037 		tmp = RREG32_SMC(CG_SPLL_SPREAD_SPECTRUM);
   2038 		tmp &= ~SSEN;
   2039 		WREG32_SMC(CG_SPLL_SPREAD_SPECTRUM, tmp);
   2040 
   2041 		tmp = RREG32_SMC(GENERAL_PWRMGT);
   2042 		tmp &= ~DYN_SPREAD_SPECTRUM_EN;
   2043 		WREG32_SMC(GENERAL_PWRMGT, tmp);
   2044 	}
   2045 }
   2046 
   2047 static void ci_program_sstp(struct radeon_device *rdev)
   2048 {
   2049 	WREG32_SMC(CG_SSP, (SSTU(R600_SSTU_DFLT) | SST(R600_SST_DFLT)));
   2050 }
   2051 
   2052 static void ci_enable_display_gap(struct radeon_device *rdev)
   2053 {
   2054 	u32 tmp = RREG32_SMC(CG_DISPLAY_GAP_CNTL);
   2055 
   2056 	tmp &= ~(DISP_GAP_MASK | DISP_GAP_MCHG_MASK);
   2057 	tmp |= (DISP_GAP(R600_PM_DISPLAY_GAP_IGNORE) |
   2058 		DISP_GAP_MCHG(R600_PM_DISPLAY_GAP_VBLANK));
   2059 
   2060 	WREG32_SMC(CG_DISPLAY_GAP_CNTL, tmp);
   2061 }
   2062 
   2063 static void ci_program_vc(struct radeon_device *rdev)
   2064 {
   2065 	u32 tmp;
   2066 
   2067 	tmp = RREG32_SMC(SCLK_PWRMGT_CNTL);
   2068 	tmp &= ~(RESET_SCLK_CNT | RESET_BUSY_CNT);
   2069 	WREG32_SMC(SCLK_PWRMGT_CNTL, tmp);
   2070 
   2071 	WREG32_SMC(CG_FTV_0, CISLANDS_VRC_DFLT0);
   2072 	WREG32_SMC(CG_FTV_1, CISLANDS_VRC_DFLT1);
   2073 	WREG32_SMC(CG_FTV_2, CISLANDS_VRC_DFLT2);
   2074 	WREG32_SMC(CG_FTV_3, CISLANDS_VRC_DFLT3);
   2075 	WREG32_SMC(CG_FTV_4, CISLANDS_VRC_DFLT4);
   2076 	WREG32_SMC(CG_FTV_5, CISLANDS_VRC_DFLT5);
   2077 	WREG32_SMC(CG_FTV_6, CISLANDS_VRC_DFLT6);
   2078 	WREG32_SMC(CG_FTV_7, CISLANDS_VRC_DFLT7);
   2079 }
   2080 
   2081 static void ci_clear_vc(struct radeon_device *rdev)
   2082 {
   2083 	u32 tmp;
   2084 
   2085 	tmp = RREG32_SMC(SCLK_PWRMGT_CNTL);
   2086 	tmp |= (RESET_SCLK_CNT | RESET_BUSY_CNT);
   2087 	WREG32_SMC(SCLK_PWRMGT_CNTL, tmp);
   2088 
   2089 	WREG32_SMC(CG_FTV_0, 0);
   2090 	WREG32_SMC(CG_FTV_1, 0);
   2091 	WREG32_SMC(CG_FTV_2, 0);
   2092 	WREG32_SMC(CG_FTV_3, 0);
   2093 	WREG32_SMC(CG_FTV_4, 0);
   2094 	WREG32_SMC(CG_FTV_5, 0);
   2095 	WREG32_SMC(CG_FTV_6, 0);
   2096 	WREG32_SMC(CG_FTV_7, 0);
   2097 }
   2098 
   2099 static int ci_upload_firmware(struct radeon_device *rdev)
   2100 {
   2101 	struct ci_power_info *pi = ci_get_pi(rdev);
   2102 	int i, ret;
   2103 
   2104 	for (i = 0; i < rdev->usec_timeout; i++) {
   2105 		if (RREG32_SMC(RCU_UC_EVENTS) & BOOT_SEQ_DONE)
   2106 			break;
   2107 	}
   2108 	WREG32_SMC(SMC_SYSCON_MISC_CNTL, 1);
   2109 
   2110 	ci_stop_smc_clock(rdev);
   2111 	ci_reset_smc(rdev);
   2112 
   2113 	ret = ci_load_smc_ucode(rdev, pi->sram_end);
   2114 
   2115 	return ret;
   2116 
   2117 }
   2118 
   2119 static int ci_get_svi2_voltage_table(struct radeon_device *rdev,
   2120 				     struct radeon_clock_voltage_dependency_table *voltage_dependency_table,
   2121 				     struct atom_voltage_table *voltage_table)
   2122 {
   2123 	u32 i;
   2124 
   2125 	if (voltage_dependency_table == NULL)
   2126 		return -EINVAL;
   2127 
   2128 	voltage_table->mask_low = 0;
   2129 	voltage_table->phase_delay = 0;
   2130 
   2131 	voltage_table->count = voltage_dependency_table->count;
   2132 	for (i = 0; i < voltage_table->count; i++) {
   2133 		voltage_table->entries[i].value = voltage_dependency_table->entries[i].v;
   2134 		voltage_table->entries[i].smio_low = 0;
   2135 	}
   2136 
   2137 	return 0;
   2138 }
   2139 
   2140 static int ci_construct_voltage_tables(struct radeon_device *rdev)
   2141 {
   2142 	struct ci_power_info *pi = ci_get_pi(rdev);
   2143 	int ret;
   2144 
   2145 	if (pi->voltage_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO) {
   2146 		ret = radeon_atom_get_voltage_table(rdev, VOLTAGE_TYPE_VDDC,
   2147 						    VOLTAGE_OBJ_GPIO_LUT,
   2148 						    &pi->vddc_voltage_table);
   2149 		if (ret)
   2150 			return ret;
   2151 	} else if (pi->voltage_control == CISLANDS_VOLTAGE_CONTROL_BY_SVID2) {
   2152 		ret = ci_get_svi2_voltage_table(rdev,
   2153 						&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk,
   2154 						&pi->vddc_voltage_table);
   2155 		if (ret)
   2156 			return ret;
   2157 	}
   2158 
   2159 	if (pi->vddc_voltage_table.count > SMU7_MAX_LEVELS_VDDC)
   2160 		si_trim_voltage_table_to_fit_state_table(rdev, SMU7_MAX_LEVELS_VDDC,
   2161 							 &pi->vddc_voltage_table);
   2162 
   2163 	if (pi->vddci_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO) {
   2164 		ret = radeon_atom_get_voltage_table(rdev, VOLTAGE_TYPE_VDDCI,
   2165 						    VOLTAGE_OBJ_GPIO_LUT,
   2166 						    &pi->vddci_voltage_table);
   2167 		if (ret)
   2168 			return ret;
   2169 	} else if (pi->vddci_control == CISLANDS_VOLTAGE_CONTROL_BY_SVID2) {
   2170 		ret = ci_get_svi2_voltage_table(rdev,
   2171 						&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk,
   2172 						&pi->vddci_voltage_table);
   2173 		if (ret)
   2174 			return ret;
   2175 	}
   2176 
   2177 	if (pi->vddci_voltage_table.count > SMU7_MAX_LEVELS_VDDCI)
   2178 		si_trim_voltage_table_to_fit_state_table(rdev, SMU7_MAX_LEVELS_VDDCI,
   2179 							 &pi->vddci_voltage_table);
   2180 
   2181 	if (pi->mvdd_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO) {
   2182 		ret = radeon_atom_get_voltage_table(rdev, VOLTAGE_TYPE_MVDDC,
   2183 						    VOLTAGE_OBJ_GPIO_LUT,
   2184 						    &pi->mvdd_voltage_table);
   2185 		if (ret)
   2186 			return ret;
   2187 	} else if (pi->mvdd_control == CISLANDS_VOLTAGE_CONTROL_BY_SVID2) {
   2188 		ret = ci_get_svi2_voltage_table(rdev,
   2189 						&rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk,
   2190 						&pi->mvdd_voltage_table);
   2191 		if (ret)
   2192 			return ret;
   2193 	}
   2194 
   2195 	if (pi->mvdd_voltage_table.count > SMU7_MAX_LEVELS_MVDD)
   2196 		si_trim_voltage_table_to_fit_state_table(rdev, SMU7_MAX_LEVELS_MVDD,
   2197 							 &pi->mvdd_voltage_table);
   2198 
   2199 	return 0;
   2200 }
   2201 
   2202 static void ci_populate_smc_voltage_table(struct radeon_device *rdev,
   2203 					  struct atom_voltage_table_entry *voltage_table,
   2204 					  SMU7_Discrete_VoltageLevel *smc_voltage_table)
   2205 {
   2206 	int ret;
   2207 
   2208 	ret = ci_get_std_voltage_value_sidd(rdev, voltage_table,
   2209 					    &smc_voltage_table->StdVoltageHiSidd,
   2210 					    &smc_voltage_table->StdVoltageLoSidd);
   2211 
   2212 	if (ret) {
   2213 		smc_voltage_table->StdVoltageHiSidd = voltage_table->value * VOLTAGE_SCALE;
   2214 		smc_voltage_table->StdVoltageLoSidd = voltage_table->value * VOLTAGE_SCALE;
   2215 	}
   2216 
   2217 	smc_voltage_table->Voltage = cpu_to_be16(voltage_table->value * VOLTAGE_SCALE);
   2218 	smc_voltage_table->StdVoltageHiSidd =
   2219 		cpu_to_be16(smc_voltage_table->StdVoltageHiSidd);
   2220 	smc_voltage_table->StdVoltageLoSidd =
   2221 		cpu_to_be16(smc_voltage_table->StdVoltageLoSidd);
   2222 }
   2223 
   2224 static int ci_populate_smc_vddc_table(struct radeon_device *rdev,
   2225 				      SMU7_Discrete_DpmTable *table)
   2226 {
   2227 	struct ci_power_info *pi = ci_get_pi(rdev);
   2228 	unsigned int count;
   2229 
   2230 	table->VddcLevelCount = pi->vddc_voltage_table.count;
   2231 	for (count = 0; count < table->VddcLevelCount; count++) {
   2232 		ci_populate_smc_voltage_table(rdev,
   2233 					      &pi->vddc_voltage_table.entries[count],
   2234 					      &table->VddcLevel[count]);
   2235 
   2236 		if (pi->voltage_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO)
   2237 			table->VddcLevel[count].Smio |=
   2238 				pi->vddc_voltage_table.entries[count].smio_low;
   2239 		else
   2240 			table->VddcLevel[count].Smio = 0;
   2241 	}
   2242 	table->VddcLevelCount = cpu_to_be32(table->VddcLevelCount);
   2243 
   2244 	return 0;
   2245 }
   2246 
   2247 static int ci_populate_smc_vddci_table(struct radeon_device *rdev,
   2248 				       SMU7_Discrete_DpmTable *table)
   2249 {
   2250 	unsigned int count;
   2251 	struct ci_power_info *pi = ci_get_pi(rdev);
   2252 
   2253 	table->VddciLevelCount = pi->vddci_voltage_table.count;
   2254 	for (count = 0; count < table->VddciLevelCount; count++) {
   2255 		ci_populate_smc_voltage_table(rdev,
   2256 					      &pi->vddci_voltage_table.entries[count],
   2257 					      &table->VddciLevel[count]);
   2258 
   2259 		if (pi->vddci_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO)
   2260 			table->VddciLevel[count].Smio |=
   2261 				pi->vddci_voltage_table.entries[count].smio_low;
   2262 		else
   2263 			table->VddciLevel[count].Smio = 0;
   2264 	}
   2265 	table->VddciLevelCount = cpu_to_be32(table->VddciLevelCount);
   2266 
   2267 	return 0;
   2268 }
   2269 
   2270 static int ci_populate_smc_mvdd_table(struct radeon_device *rdev,
   2271 				      SMU7_Discrete_DpmTable *table)
   2272 {
   2273 	struct ci_power_info *pi = ci_get_pi(rdev);
   2274 	unsigned int count;
   2275 
   2276 	table->MvddLevelCount = pi->mvdd_voltage_table.count;
   2277 	for (count = 0; count < table->MvddLevelCount; count++) {
   2278 		ci_populate_smc_voltage_table(rdev,
   2279 					      &pi->mvdd_voltage_table.entries[count],
   2280 					      &table->MvddLevel[count]);
   2281 
   2282 		if (pi->mvdd_control == CISLANDS_VOLTAGE_CONTROL_BY_GPIO)
   2283 			table->MvddLevel[count].Smio |=
   2284 				pi->mvdd_voltage_table.entries[count].smio_low;
   2285 		else
   2286 			table->MvddLevel[count].Smio = 0;
   2287 	}
   2288 	table->MvddLevelCount = cpu_to_be32(table->MvddLevelCount);
   2289 
   2290 	return 0;
   2291 }
   2292 
   2293 static int ci_populate_smc_voltage_tables(struct radeon_device *rdev,
   2294 					  SMU7_Discrete_DpmTable *table)
   2295 {
   2296 	int ret;
   2297 
   2298 	ret = ci_populate_smc_vddc_table(rdev, table);
   2299 	if (ret)
   2300 		return ret;
   2301 
   2302 	ret = ci_populate_smc_vddci_table(rdev, table);
   2303 	if (ret)
   2304 		return ret;
   2305 
   2306 	ret = ci_populate_smc_mvdd_table(rdev, table);
   2307 	if (ret)
   2308 		return ret;
   2309 
   2310 	return 0;
   2311 }
   2312 
   2313 static int ci_populate_mvdd_value(struct radeon_device *rdev, u32 mclk,
   2314 				  SMU7_Discrete_VoltageLevel *voltage)
   2315 {
   2316 	struct ci_power_info *pi = ci_get_pi(rdev);
   2317 	u32 i = 0;
   2318 
   2319 	if (pi->mvdd_control != CISLANDS_VOLTAGE_CONTROL_NONE) {
   2320 		for (i = 0; i < rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.count; i++) {
   2321 			if (mclk <= rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries[i].clk) {
   2322 				voltage->Voltage = pi->mvdd_voltage_table.entries[i].value;
   2323 				break;
   2324 			}
   2325 		}
   2326 
   2327 		if (i >= rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.count)
   2328 			return -EINVAL;
   2329 	}
   2330 
   2331 	return -EINVAL;
   2332 }
   2333 
   2334 static int ci_get_std_voltage_value_sidd(struct radeon_device *rdev,
   2335 					 struct atom_voltage_table_entry *voltage_table,
   2336 					 u16 *std_voltage_hi_sidd, u16 *std_voltage_lo_sidd)
   2337 {
   2338 	u16 v_index, idx;
   2339 	bool voltage_found = false;
   2340 	*std_voltage_hi_sidd = voltage_table->value * VOLTAGE_SCALE;
   2341 	*std_voltage_lo_sidd = voltage_table->value * VOLTAGE_SCALE;
   2342 
   2343 	if (rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries == NULL)
   2344 		return -EINVAL;
   2345 
   2346 	if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries) {
   2347 		for (v_index = 0; (u32)v_index < rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.count; v_index++) {
   2348 			if (voltage_table->value ==
   2349 			    rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[v_index].v) {
   2350 				voltage_found = true;
   2351 				if ((u32)v_index < rdev->pm.dpm.dyn_state.cac_leakage_table.count)
   2352 					idx = v_index;
   2353 				else
   2354 					idx = rdev->pm.dpm.dyn_state.cac_leakage_table.count - 1;
   2355 				*std_voltage_lo_sidd =
   2356 					rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].vddc * VOLTAGE_SCALE;
   2357 				*std_voltage_hi_sidd =
   2358 					rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].leakage * VOLTAGE_SCALE;
   2359 				break;
   2360 			}
   2361 		}
   2362 
   2363 		if (!voltage_found) {
   2364 			for (v_index = 0; (u32)v_index < rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.count; v_index++) {
   2365 				if (voltage_table->value <=
   2366 				    rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[v_index].v) {
   2367 					voltage_found = true;
   2368 					if ((u32)v_index < rdev->pm.dpm.dyn_state.cac_leakage_table.count)
   2369 						idx = v_index;
   2370 					else
   2371 						idx = rdev->pm.dpm.dyn_state.cac_leakage_table.count - 1;
   2372 					*std_voltage_lo_sidd =
   2373 						rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].vddc * VOLTAGE_SCALE;
   2374 					*std_voltage_hi_sidd =
   2375 						rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].leakage * VOLTAGE_SCALE;
   2376 					break;
   2377 				}
   2378 			}
   2379 		}
   2380 	}
   2381 
   2382 	return 0;
   2383 }
   2384 
   2385 static void ci_populate_phase_value_based_on_sclk(struct radeon_device *rdev,
   2386 						  const struct radeon_phase_shedding_limits_table *limits,
   2387 						  u32 sclk,
   2388 						  u32 *phase_shedding)
   2389 {
   2390 	unsigned int i;
   2391 
   2392 	*phase_shedding = 1;
   2393 
   2394 	for (i = 0; i < limits->count; i++) {
   2395 		if (sclk < limits->entries[i].sclk) {
   2396 			*phase_shedding = i;
   2397 			break;
   2398 		}
   2399 	}
   2400 }
   2401 
   2402 static void ci_populate_phase_value_based_on_mclk(struct radeon_device *rdev,
   2403 						  const struct radeon_phase_shedding_limits_table *limits,
   2404 						  u32 mclk,
   2405 						  u32 *phase_shedding)
   2406 {
   2407 	unsigned int i;
   2408 
   2409 	*phase_shedding = 1;
   2410 
   2411 	for (i = 0; i < limits->count; i++) {
   2412 		if (mclk < limits->entries[i].mclk) {
   2413 			*phase_shedding = i;
   2414 			break;
   2415 		}
   2416 	}
   2417 }
   2418 
   2419 static int ci_init_arb_table_index(struct radeon_device *rdev)
   2420 {
   2421 	struct ci_power_info *pi = ci_get_pi(rdev);
   2422 	u32 tmp;
   2423 	int ret;
   2424 
   2425 	ret = ci_read_smc_sram_dword(rdev, pi->arb_table_start,
   2426 				     &tmp, pi->sram_end);
   2427 	if (ret)
   2428 		return ret;
   2429 
   2430 	tmp &= 0x00FFFFFF;
   2431 	tmp |= MC_CG_ARB_FREQ_F1 << 24;
   2432 
   2433 	return ci_write_smc_sram_dword(rdev, pi->arb_table_start,
   2434 				       tmp, pi->sram_end);
   2435 }
   2436 
   2437 static int ci_get_dependency_volt_by_clk(struct radeon_device *rdev,
   2438 					 struct radeon_clock_voltage_dependency_table *allowed_clock_voltage_table,
   2439 					 u32 clock, u32 *voltage)
   2440 {
   2441 	u32 i = 0;
   2442 
   2443 	if (allowed_clock_voltage_table->count == 0)
   2444 		return -EINVAL;
   2445 
   2446 	for (i = 0; i < allowed_clock_voltage_table->count; i++) {
   2447 		if (allowed_clock_voltage_table->entries[i].clk >= clock) {
   2448 			*voltage = allowed_clock_voltage_table->entries[i].v;
   2449 			return 0;
   2450 		}
   2451 	}
   2452 
   2453 	*voltage = allowed_clock_voltage_table->entries[i-1].v;
   2454 
   2455 	return 0;
   2456 }
   2457 
   2458 static u8 ci_get_sleep_divider_id_from_clock(struct radeon_device *rdev,
   2459 					     u32 sclk, u32 min_sclk_in_sr)
   2460 {
   2461 	u32 i;
   2462 	u32 tmp;
   2463 	u32 min = (min_sclk_in_sr > CISLAND_MINIMUM_ENGINE_CLOCK) ?
   2464 		min_sclk_in_sr : CISLAND_MINIMUM_ENGINE_CLOCK;
   2465 
   2466 	if (sclk < min)
   2467 		return 0;
   2468 
   2469 	for (i = CISLAND_MAX_DEEPSLEEP_DIVIDER_ID;  ; i--) {
   2470 		tmp = sclk / (1 << i);
   2471 		if (tmp >= min || i == 0)
   2472 			break;
   2473 	}
   2474 
   2475 	return (u8)i;
   2476 }
   2477 
   2478 static int ci_initial_switch_from_arb_f0_to_f1(struct radeon_device *rdev)
   2479 {
   2480 	return ni_copy_and_switch_arb_sets(rdev, MC_CG_ARB_FREQ_F0, MC_CG_ARB_FREQ_F1);
   2481 }
   2482 
   2483 static int ci_reset_to_default(struct radeon_device *rdev)
   2484 {
   2485 	return (ci_send_msg_to_smc(rdev, PPSMC_MSG_ResetToDefaults) == PPSMC_Result_OK) ?
   2486 		0 : -EINVAL;
   2487 }
   2488 
   2489 static int ci_force_switch_to_arb_f0(struct radeon_device *rdev)
   2490 {
   2491 	u32 tmp;
   2492 
   2493 	tmp = (RREG32_SMC(SMC_SCRATCH9) & 0x0000ff00) >> 8;
   2494 
   2495 	if (tmp == MC_CG_ARB_FREQ_F0)
   2496 		return 0;
   2497 
   2498 	return ni_copy_and_switch_arb_sets(rdev, tmp, MC_CG_ARB_FREQ_F0);
   2499 }
   2500 
   2501 static void ci_register_patching_mc_arb(struct radeon_device *rdev,
   2502 					const u32 engine_clock,
   2503 					const u32 memory_clock,
   2504 					u32 *dram_timimg2)
   2505 {
   2506 	bool patch;
   2507 	u32 tmp, tmp2;
   2508 
   2509 	tmp = RREG32(MC_SEQ_MISC0);
   2510 	patch = ((tmp & 0x0000f00) == 0x300) ? true : false;
   2511 
   2512 	if (patch &&
   2513 	    ((rdev->pdev->device == 0x67B0) ||
   2514 	     (rdev->pdev->device == 0x67B1))) {
   2515 		if ((memory_clock > 100000) && (memory_clock <= 125000)) {
   2516 			tmp2 = (((0x31 * engine_clock) / 125000) - 1) & 0xff;
   2517 			*dram_timimg2 &= ~0x00ff0000;
   2518 			*dram_timimg2 |= tmp2 << 16;
   2519 		} else if ((memory_clock > 125000) && (memory_clock <= 137500)) {
   2520 			tmp2 = (((0x36 * engine_clock) / 137500) - 1) & 0xff;
   2521 			*dram_timimg2 &= ~0x00ff0000;
   2522 			*dram_timimg2 |= tmp2 << 16;
   2523 		}
   2524 	}
   2525 }
   2526 
   2527 
   2528 static int ci_populate_memory_timing_parameters(struct radeon_device *rdev,
   2529 						u32 sclk,
   2530 						u32 mclk,
   2531 						SMU7_Discrete_MCArbDramTimingTableEntry *arb_regs)
   2532 {
   2533 	u32 dram_timing;
   2534 	u32 dram_timing2;
   2535 	u32 burst_time;
   2536 
   2537 	radeon_atom_set_engine_dram_timings(rdev, sclk, mclk);
   2538 
   2539 	dram_timing  = RREG32(MC_ARB_DRAM_TIMING);
   2540 	dram_timing2 = RREG32(MC_ARB_DRAM_TIMING2);
   2541 	burst_time = RREG32(MC_ARB_BURST_TIME) & STATE0_MASK;
   2542 
   2543 	ci_register_patching_mc_arb(rdev, sclk, mclk, &dram_timing2);
   2544 
   2545 	arb_regs->McArbDramTiming  = cpu_to_be32(dram_timing);
   2546 	arb_regs->McArbDramTiming2 = cpu_to_be32(dram_timing2);
   2547 	arb_regs->McArbBurstTime = (u8)burst_time;
   2548 
   2549 	return 0;
   2550 }
   2551 
   2552 static int ci_do_program_memory_timing_parameters(struct radeon_device *rdev)
   2553 {
   2554 	struct ci_power_info *pi = ci_get_pi(rdev);
   2555 	SMU7_Discrete_MCArbDramTimingTable arb_regs;
   2556 	u32 i, j;
   2557 	int ret =  0;
   2558 
   2559 	memset(&arb_regs, 0, sizeof(SMU7_Discrete_MCArbDramTimingTable));
   2560 
   2561 	for (i = 0; i < pi->dpm_table.sclk_table.count; i++) {
   2562 		for (j = 0; j < pi->dpm_table.mclk_table.count; j++) {
   2563 			ret = ci_populate_memory_timing_parameters(rdev,
   2564 								   pi->dpm_table.sclk_table.dpm_levels[i].value,
   2565 								   pi->dpm_table.mclk_table.dpm_levels[j].value,
   2566 								   &arb_regs.entries[i][j]);
   2567 			if (ret)
   2568 				break;
   2569 		}
   2570 	}
   2571 
   2572 	if (ret == 0)
   2573 		ret = ci_copy_bytes_to_smc(rdev,
   2574 					   pi->arb_table_start,
   2575 					   (u8 *)&arb_regs,
   2576 					   sizeof(SMU7_Discrete_MCArbDramTimingTable),
   2577 					   pi->sram_end);
   2578 
   2579 	return ret;
   2580 }
   2581 
   2582 static int ci_program_memory_timing_parameters(struct radeon_device *rdev)
   2583 {
   2584 	struct ci_power_info *pi = ci_get_pi(rdev);
   2585 
   2586 	if (pi->need_update_smu7_dpm_table == 0)
   2587 		return 0;
   2588 
   2589 	return ci_do_program_memory_timing_parameters(rdev);
   2590 }
   2591 
   2592 static void ci_populate_smc_initial_state(struct radeon_device *rdev,
   2593 					  struct radeon_ps *radeon_boot_state)
   2594 {
   2595 	struct ci_ps *boot_state = ci_get_ps(radeon_boot_state);
   2596 	struct ci_power_info *pi = ci_get_pi(rdev);
   2597 	u32 level = 0;
   2598 
   2599 	for (level = 0; level < rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.count; level++) {
   2600 		if (rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[level].clk >=
   2601 		    boot_state->performance_levels[0].sclk) {
   2602 			pi->smc_state_table.GraphicsBootLevel = level;
   2603 			break;
   2604 		}
   2605 	}
   2606 
   2607 	for (level = 0; level < rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.count; level++) {
   2608 		if (rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries[level].clk >=
   2609 		    boot_state->performance_levels[0].mclk) {
   2610 			pi->smc_state_table.MemoryBootLevel = level;
   2611 			break;
   2612 		}
   2613 	}
   2614 }
   2615 
   2616 static u32 ci_get_dpm_level_enable_mask_value(struct ci_single_dpm_table *dpm_table)
   2617 {
   2618 	u32 i;
   2619 	u32 mask_value = 0;
   2620 
   2621 	for (i = dpm_table->count; i > 0; i--) {
   2622 		mask_value = mask_value << 1;
   2623 		if (dpm_table->dpm_levels[i-1].enabled)
   2624 			mask_value |= 0x1;
   2625 		else
   2626 			mask_value &= 0xFFFFFFFE;
   2627 	}
   2628 
   2629 	return mask_value;
   2630 }
   2631 
   2632 static void ci_populate_smc_link_level(struct radeon_device *rdev,
   2633 				       SMU7_Discrete_DpmTable *table)
   2634 {
   2635 	struct ci_power_info *pi = ci_get_pi(rdev);
   2636 	struct ci_dpm_table *dpm_table = &pi->dpm_table;
   2637 	u32 i;
   2638 
   2639 	for (i = 0; i < dpm_table->pcie_speed_table.count; i++) {
   2640 		table->LinkLevel[i].PcieGenSpeed =
   2641 			(u8)dpm_table->pcie_speed_table.dpm_levels[i].value;
   2642 		table->LinkLevel[i].PcieLaneCount =
   2643 			r600_encode_pci_lane_width(dpm_table->pcie_speed_table.dpm_levels[i].param1);
   2644 		table->LinkLevel[i].EnabledForActivity = 1;
   2645 		table->LinkLevel[i].DownT = cpu_to_be32(5);
   2646 		table->LinkLevel[i].UpT = cpu_to_be32(30);
   2647 	}
   2648 
   2649 	pi->smc_state_table.LinkLevelCount = (u8)dpm_table->pcie_speed_table.count;
   2650 	pi->dpm_level_enable_mask.pcie_dpm_enable_mask =
   2651 		ci_get_dpm_level_enable_mask_value(&dpm_table->pcie_speed_table);
   2652 }
   2653 
   2654 static int ci_populate_smc_uvd_level(struct radeon_device *rdev,
   2655 				     SMU7_Discrete_DpmTable *table)
   2656 {
   2657 	u32 count;
   2658 	struct atom_clock_dividers dividers;
   2659 	int ret = -EINVAL;
   2660 
   2661 	table->UvdLevelCount =
   2662 		rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.count;
   2663 
   2664 	for (count = 0; count < table->UvdLevelCount; count++) {
   2665 		table->UvdLevel[count].VclkFrequency =
   2666 			rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[count].vclk;
   2667 		table->UvdLevel[count].DclkFrequency =
   2668 			rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[count].dclk;
   2669 		table->UvdLevel[count].MinVddc =
   2670 			rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[count].v * VOLTAGE_SCALE;
   2671 		table->UvdLevel[count].MinVddcPhases = 1;
   2672 
   2673 		ret = radeon_atom_get_clock_dividers(rdev,
   2674 						     COMPUTE_GPUCLK_INPUT_FLAG_DEFAULT_GPUCLK,
   2675 						     table->UvdLevel[count].VclkFrequency, false, &dividers);
   2676 		if (ret)
   2677 			return ret;
   2678 
   2679 		table->UvdLevel[count].VclkDivider = (u8)dividers.post_divider;
   2680 
   2681 		ret = radeon_atom_get_clock_dividers(rdev,
   2682 						     COMPUTE_GPUCLK_INPUT_FLAG_DEFAULT_GPUCLK,
   2683 						     table->UvdLevel[count].DclkFrequency, false, &dividers);
   2684 		if (ret)
   2685 			return ret;
   2686 
   2687 		table->UvdLevel[count].DclkDivider = (u8)dividers.post_divider;
   2688 
   2689 		table->UvdLevel[count].VclkFrequency = cpu_to_be32(table->UvdLevel[count].VclkFrequency);
   2690 		table->UvdLevel[count].DclkFrequency = cpu_to_be32(table->UvdLevel[count].DclkFrequency);
   2691 		table->UvdLevel[count].MinVddc = cpu_to_be16(table->UvdLevel[count].MinVddc);
   2692 	}
   2693 
   2694 	return ret;
   2695 }
   2696 
   2697 static int ci_populate_smc_vce_level(struct radeon_device *rdev,
   2698 				     SMU7_Discrete_DpmTable *table)
   2699 {
   2700 	u32 count;
   2701 	struct atom_clock_dividers dividers;
   2702 	int ret = -EINVAL;
   2703 
   2704 	table->VceLevelCount =
   2705 		rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.count;
   2706 
   2707 	for (count = 0; count < table->VceLevelCount; count++) {
   2708 		table->VceLevel[count].Frequency =
   2709 			rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries[count].evclk;
   2710 		table->VceLevel[count].MinVoltage =
   2711 			(u16)rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries[count].v * VOLTAGE_SCALE;
   2712 		table->VceLevel[count].MinPhases = 1;
   2713 
   2714 		ret = radeon_atom_get_clock_dividers(rdev,
   2715 						     COMPUTE_GPUCLK_INPUT_FLAG_DEFAULT_GPUCLK,
   2716 						     table->VceLevel[count].Frequency, false, &dividers);
   2717 		if (ret)
   2718 			return ret;
   2719 
   2720 		table->VceLevel[count].Divider = (u8)dividers.post_divider;
   2721 
   2722 		table->VceLevel[count].Frequency = cpu_to_be32(table->VceLevel[count].Frequency);
   2723 		table->VceLevel[count].MinVoltage = cpu_to_be16(table->VceLevel[count].MinVoltage);
   2724 	}
   2725 
   2726 	return ret;
   2727 
   2728 }
   2729 
   2730 static int ci_populate_smc_acp_level(struct radeon_device *rdev,
   2731 				     SMU7_Discrete_DpmTable *table)
   2732 {
   2733 	u32 count;
   2734 	struct atom_clock_dividers dividers;
   2735 	int ret = -EINVAL;
   2736 
   2737 	table->AcpLevelCount = (u8)
   2738 		(rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.count);
   2739 
   2740 	for (count = 0; count < table->AcpLevelCount; count++) {
   2741 		table->AcpLevel[count].Frequency =
   2742 			rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries[count].clk;
   2743 		table->AcpLevel[count].MinVoltage =
   2744 			rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries[count].v;
   2745 		table->AcpLevel[count].MinPhases = 1;
   2746 
   2747 		ret = radeon_atom_get_clock_dividers(rdev,
   2748 						     COMPUTE_GPUCLK_INPUT_FLAG_DEFAULT_GPUCLK,
   2749 						     table->AcpLevel[count].Frequency, false, &dividers);
   2750 		if (ret)
   2751 			return ret;
   2752 
   2753 		table->AcpLevel[count].Divider = (u8)dividers.post_divider;
   2754 
   2755 		table->AcpLevel[count].Frequency = cpu_to_be32(table->AcpLevel[count].Frequency);
   2756 		table->AcpLevel[count].MinVoltage = cpu_to_be16(table->AcpLevel[count].MinVoltage);
   2757 	}
   2758 
   2759 	return ret;
   2760 }
   2761 
   2762 static int ci_populate_smc_samu_level(struct radeon_device *rdev,
   2763 				      SMU7_Discrete_DpmTable *table)
   2764 {
   2765 	u32 count;
   2766 	struct atom_clock_dividers dividers;
   2767 	int ret = -EINVAL;
   2768 
   2769 	table->SamuLevelCount =
   2770 		rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.count;
   2771 
   2772 	for (count = 0; count < table->SamuLevelCount; count++) {
   2773 		table->SamuLevel[count].Frequency =
   2774 			rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries[count].clk;
   2775 		table->SamuLevel[count].MinVoltage =
   2776 			rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries[count].v * VOLTAGE_SCALE;
   2777 		table->SamuLevel[count].MinPhases = 1;
   2778 
   2779 		ret = radeon_atom_get_clock_dividers(rdev,
   2780 						     COMPUTE_GPUCLK_INPUT_FLAG_DEFAULT_GPUCLK,
   2781 						     table->SamuLevel[count].Frequency, false, &dividers);
   2782 		if (ret)
   2783 			return ret;
   2784 
   2785 		table->SamuLevel[count].Divider = (u8)dividers.post_divider;
   2786 
   2787 		table->SamuLevel[count].Frequency = cpu_to_be32(table->SamuLevel[count].Frequency);
   2788 		table->SamuLevel[count].MinVoltage = cpu_to_be16(table->SamuLevel[count].MinVoltage);
   2789 	}
   2790 
   2791 	return ret;
   2792 }
   2793 
   2794 static int ci_calculate_mclk_params(struct radeon_device *rdev,
   2795 				    u32 memory_clock,
   2796 				    SMU7_Discrete_MemoryLevel *mclk,
   2797 				    bool strobe_mode,
   2798 				    bool dll_state_on)
   2799 {
   2800 	struct ci_power_info *pi = ci_get_pi(rdev);
   2801 	u32  dll_cntl = pi->clock_registers.dll_cntl;
   2802 	u32  mclk_pwrmgt_cntl = pi->clock_registers.mclk_pwrmgt_cntl;
   2803 	u32  mpll_ad_func_cntl = pi->clock_registers.mpll_ad_func_cntl;
   2804 	u32  mpll_dq_func_cntl = pi->clock_registers.mpll_dq_func_cntl;
   2805 	u32  mpll_func_cntl = pi->clock_registers.mpll_func_cntl;
   2806 	u32  mpll_func_cntl_1 = pi->clock_registers.mpll_func_cntl_1;
   2807 	u32  mpll_func_cntl_2 = pi->clock_registers.mpll_func_cntl_2;
   2808 	u32  mpll_ss1 = pi->clock_registers.mpll_ss1;
   2809 	u32  mpll_ss2 = pi->clock_registers.mpll_ss2;
   2810 	struct atom_mpll_param mpll_param;
   2811 	int ret;
   2812 
   2813 	ret = radeon_atom_get_memory_pll_dividers(rdev, memory_clock, strobe_mode, &mpll_param);
   2814 	if (ret)
   2815 		return ret;
   2816 
   2817 	mpll_func_cntl &= ~BWCTRL_MASK;
   2818 	mpll_func_cntl |= BWCTRL(mpll_param.bwcntl);
   2819 
   2820 	mpll_func_cntl_1 &= ~(CLKF_MASK | CLKFRAC_MASK | VCO_MODE_MASK);
   2821 	mpll_func_cntl_1 |= CLKF(mpll_param.clkf) |
   2822 		CLKFRAC(mpll_param.clkfrac) | VCO_MODE(mpll_param.vco_mode);
   2823 
   2824 	mpll_ad_func_cntl &= ~YCLK_POST_DIV_MASK;
   2825 	mpll_ad_func_cntl |= YCLK_POST_DIV(mpll_param.post_div);
   2826 
   2827 	if (pi->mem_gddr5) {
   2828 		mpll_dq_func_cntl &= ~(YCLK_SEL_MASK | YCLK_POST_DIV_MASK);
   2829 		mpll_dq_func_cntl |= YCLK_SEL(mpll_param.yclk_sel) |
   2830 			YCLK_POST_DIV(mpll_param.post_div);
   2831 	}
   2832 
   2833 	if (pi->caps_mclk_ss_support) {
   2834 		struct radeon_atom_ss ss;
   2835 		u32 freq_nom;
   2836 		u32 tmp;
   2837 		u32 reference_clock = rdev->clock.mpll.reference_freq;
   2838 
   2839 		if (mpll_param.qdr == 1)
   2840 			freq_nom = memory_clock * 4 * (1 << mpll_param.post_div);
   2841 		else
   2842 			freq_nom = memory_clock * 2 * (1 << mpll_param.post_div);
   2843 
   2844 		tmp = (freq_nom / reference_clock);
   2845 		tmp = tmp * tmp;
   2846 		if (radeon_atombios_get_asic_ss_info(rdev, &ss,
   2847 						     ASIC_INTERNAL_MEMORY_SS, freq_nom)) {
   2848 			u32 clks = reference_clock * 5 / ss.rate;
   2849 			u32 clkv = (u32)((((131 * ss.percentage * ss.rate) / 100) * tmp) / freq_nom);
   2850 
   2851 			mpll_ss1 &= ~CLKV_MASK;
   2852 			mpll_ss1 |= CLKV(clkv);
   2853 
   2854 			mpll_ss2 &= ~CLKS_MASK;
   2855 			mpll_ss2 |= CLKS(clks);
   2856 		}
   2857 	}
   2858 
   2859 	mclk_pwrmgt_cntl &= ~DLL_SPEED_MASK;
   2860 	mclk_pwrmgt_cntl |= DLL_SPEED(mpll_param.dll_speed);
   2861 
   2862 	if (dll_state_on)
   2863 		mclk_pwrmgt_cntl |= MRDCK0_PDNB | MRDCK1_PDNB;
   2864 	else
   2865 		mclk_pwrmgt_cntl &= ~(MRDCK0_PDNB | MRDCK1_PDNB);
   2866 
   2867 	mclk->MclkFrequency = memory_clock;
   2868 	mclk->MpllFuncCntl = mpll_func_cntl;
   2869 	mclk->MpllFuncCntl_1 = mpll_func_cntl_1;
   2870 	mclk->MpllFuncCntl_2 = mpll_func_cntl_2;
   2871 	mclk->MpllAdFuncCntl = mpll_ad_func_cntl;
   2872 	mclk->MpllDqFuncCntl = mpll_dq_func_cntl;
   2873 	mclk->MclkPwrmgtCntl = mclk_pwrmgt_cntl;
   2874 	mclk->DllCntl = dll_cntl;
   2875 	mclk->MpllSs1 = mpll_ss1;
   2876 	mclk->MpllSs2 = mpll_ss2;
   2877 
   2878 	return 0;
   2879 }
   2880 
   2881 static int ci_populate_single_memory_level(struct radeon_device *rdev,
   2882 					   u32 memory_clock,
   2883 					   SMU7_Discrete_MemoryLevel *memory_level)
   2884 {
   2885 	struct ci_power_info *pi = ci_get_pi(rdev);
   2886 	int ret;
   2887 	bool dll_state_on;
   2888 
   2889 	if (rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries) {
   2890 		ret = ci_get_dependency_volt_by_clk(rdev,
   2891 						    &rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk,
   2892 						    memory_clock, &memory_level->MinVddc);
   2893 		if (ret)
   2894 			return ret;
   2895 	}
   2896 
   2897 	if (rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries) {
   2898 		ret = ci_get_dependency_volt_by_clk(rdev,
   2899 						    &rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk,
   2900 						    memory_clock, &memory_level->MinVddci);
   2901 		if (ret)
   2902 			return ret;
   2903 	}
   2904 
   2905 	if (rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries) {
   2906 		ret = ci_get_dependency_volt_by_clk(rdev,
   2907 						    &rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk,
   2908 						    memory_clock, &memory_level->MinMvdd);
   2909 		if (ret)
   2910 			return ret;
   2911 	}
   2912 
   2913 	memory_level->MinVddcPhases = 1;
   2914 
   2915 	if (pi->vddc_phase_shed_control)
   2916 		ci_populate_phase_value_based_on_mclk(rdev,
   2917 						      &rdev->pm.dpm.dyn_state.phase_shedding_limits_table,
   2918 						      memory_clock,
   2919 						      &memory_level->MinVddcPhases);
   2920 
   2921 	memory_level->EnabledForThrottle = 1;
   2922 	memory_level->UpH = 0;
   2923 	memory_level->DownH = 100;
   2924 	memory_level->VoltageDownH = 0;
   2925 	memory_level->ActivityLevel = (u16)pi->mclk_activity_target;
   2926 
   2927 	memory_level->StutterEnable = false;
   2928 	memory_level->StrobeEnable = false;
   2929 	memory_level->EdcReadEnable = false;
   2930 	memory_level->EdcWriteEnable = false;
   2931 	memory_level->RttEnable = false;
   2932 
   2933 	memory_level->DisplayWatermark = PPSMC_DISPLAY_WATERMARK_LOW;
   2934 
   2935 	if (pi->mclk_stutter_mode_threshold &&
   2936 	    (memory_clock <= pi->mclk_stutter_mode_threshold) &&
   2937 	    (pi->uvd_enabled == false) &&
   2938 	    (RREG32(DPG_PIPE_STUTTER_CONTROL) & STUTTER_ENABLE) &&
   2939 	    (rdev->pm.dpm.new_active_crtc_count <= 2))
   2940 		memory_level->StutterEnable = true;
   2941 
   2942 	if (pi->mclk_strobe_mode_threshold &&
   2943 	    (memory_clock <= pi->mclk_strobe_mode_threshold))
   2944 		memory_level->StrobeEnable = 1;
   2945 
   2946 	if (pi->mem_gddr5) {
   2947 		memory_level->StrobeRatio =
   2948 			si_get_mclk_frequency_ratio(memory_clock, memory_level->StrobeEnable);
   2949 		if (pi->mclk_edc_enable_threshold &&
   2950 		    (memory_clock > pi->mclk_edc_enable_threshold))
   2951 			memory_level->EdcReadEnable = true;
   2952 
   2953 		if (pi->mclk_edc_wr_enable_threshold &&
   2954 		    (memory_clock > pi->mclk_edc_wr_enable_threshold))
   2955 			memory_level->EdcWriteEnable = true;
   2956 
   2957 		if (memory_level->StrobeEnable) {
   2958 			if (si_get_mclk_frequency_ratio(memory_clock, true) >=
   2959 			    ((RREG32(MC_SEQ_MISC7) >> 16) & 0xf))
   2960 				dll_state_on = ((RREG32(MC_SEQ_MISC5) >> 1) & 0x1) ? true : false;
   2961 			else
   2962 				dll_state_on = ((RREG32(MC_SEQ_MISC6) >> 1) & 0x1) ? true : false;
   2963 		} else {
   2964 			dll_state_on = pi->dll_default_on;
   2965 		}
   2966 	} else {
   2967 		memory_level->StrobeRatio = si_get_ddr3_mclk_frequency_ratio(memory_clock);
   2968 		dll_state_on = ((RREG32(MC_SEQ_MISC5) >> 1) & 0x1) ? true : false;
   2969 	}
   2970 
   2971 	ret = ci_calculate_mclk_params(rdev, memory_clock, memory_level, memory_level->StrobeEnable, dll_state_on);
   2972 	if (ret)
   2973 		return ret;
   2974 
   2975 	memory_level->MinVddc = cpu_to_be32(memory_level->MinVddc * VOLTAGE_SCALE);
   2976 	memory_level->MinVddcPhases = cpu_to_be32(memory_level->MinVddcPhases);
   2977 	memory_level->MinVddci = cpu_to_be32(memory_level->MinVddci * VOLTAGE_SCALE);
   2978 	memory_level->MinMvdd = cpu_to_be32(memory_level->MinMvdd * VOLTAGE_SCALE);
   2979 
   2980 	memory_level->MclkFrequency = cpu_to_be32(memory_level->MclkFrequency);
   2981 	memory_level->ActivityLevel = cpu_to_be16(memory_level->ActivityLevel);
   2982 	memory_level->MpllFuncCntl = cpu_to_be32(memory_level->MpllFuncCntl);
   2983 	memory_level->MpllFuncCntl_1 = cpu_to_be32(memory_level->MpllFuncCntl_1);
   2984 	memory_level->MpllFuncCntl_2 = cpu_to_be32(memory_level->MpllFuncCntl_2);
   2985 	memory_level->MpllAdFuncCntl = cpu_to_be32(memory_level->MpllAdFuncCntl);
   2986 	memory_level->MpllDqFuncCntl = cpu_to_be32(memory_level->MpllDqFuncCntl);
   2987 	memory_level->MclkPwrmgtCntl = cpu_to_be32(memory_level->MclkPwrmgtCntl);
   2988 	memory_level->DllCntl = cpu_to_be32(memory_level->DllCntl);
   2989 	memory_level->MpllSs1 = cpu_to_be32(memory_level->MpllSs1);
   2990 	memory_level->MpllSs2 = cpu_to_be32(memory_level->MpllSs2);
   2991 
   2992 	return 0;
   2993 }
   2994 
   2995 static int ci_populate_smc_acpi_level(struct radeon_device *rdev,
   2996 				      SMU7_Discrete_DpmTable *table)
   2997 {
   2998 	struct ci_power_info *pi = ci_get_pi(rdev);
   2999 	struct atom_clock_dividers dividers;
   3000 	SMU7_Discrete_VoltageLevel voltage_level;
   3001 	u32 spll_func_cntl = pi->clock_registers.cg_spll_func_cntl;
   3002 	u32 spll_func_cntl_2 = pi->clock_registers.cg_spll_func_cntl_2;
   3003 	u32 dll_cntl = pi->clock_registers.dll_cntl;
   3004 	u32 mclk_pwrmgt_cntl = pi->clock_registers.mclk_pwrmgt_cntl;
   3005 	int ret;
   3006 
   3007 	table->ACPILevel.Flags &= ~PPSMC_SWSTATE_FLAG_DC;
   3008 
   3009 	if (pi->acpi_vddc)
   3010 		table->ACPILevel.MinVddc = cpu_to_be32(pi->acpi_vddc * VOLTAGE_SCALE);
   3011 	else
   3012 		table->ACPILevel.MinVddc = cpu_to_be32(pi->min_vddc_in_pp_table * VOLTAGE_SCALE);
   3013 
   3014 	table->ACPILevel.MinVddcPhases = pi->vddc_phase_shed_control ? 0 : 1;
   3015 
   3016 	table->ACPILevel.SclkFrequency = rdev->clock.spll.reference_freq;
   3017 
   3018 	ret = radeon_atom_get_clock_dividers(rdev,
   3019 					     COMPUTE_GPUCLK_INPUT_FLAG_SCLK,
   3020 					     table->ACPILevel.SclkFrequency, false, &dividers);
   3021 	if (ret)
   3022 		return ret;
   3023 
   3024 	table->ACPILevel.SclkDid = (u8)dividers.post_divider;
   3025 	table->ACPILevel.DisplayWatermark = PPSMC_DISPLAY_WATERMARK_LOW;
   3026 	table->ACPILevel.DeepSleepDivId = 0;
   3027 
   3028 	spll_func_cntl &= ~SPLL_PWRON;
   3029 	spll_func_cntl |= SPLL_RESET;
   3030 
   3031 	spll_func_cntl_2 &= ~SCLK_MUX_SEL_MASK;
   3032 	spll_func_cntl_2 |= SCLK_MUX_SEL(4);
   3033 
   3034 	table->ACPILevel.CgSpllFuncCntl = spll_func_cntl;
   3035 	table->ACPILevel.CgSpllFuncCntl2 = spll_func_cntl_2;
   3036 	table->ACPILevel.CgSpllFuncCntl3 = pi->clock_registers.cg_spll_func_cntl_3;
   3037 	table->ACPILevel.CgSpllFuncCntl4 = pi->clock_registers.cg_spll_func_cntl_4;
   3038 	table->ACPILevel.SpllSpreadSpectrum = pi->clock_registers.cg_spll_spread_spectrum;
   3039 	table->ACPILevel.SpllSpreadSpectrum2 = pi->clock_registers.cg_spll_spread_spectrum_2;
   3040 	table->ACPILevel.CcPwrDynRm = 0;
   3041 	table->ACPILevel.CcPwrDynRm1 = 0;
   3042 
   3043 	table->ACPILevel.Flags = cpu_to_be32(table->ACPILevel.Flags);
   3044 	table->ACPILevel.MinVddcPhases = cpu_to_be32(table->ACPILevel.MinVddcPhases);
   3045 	table->ACPILevel.SclkFrequency = cpu_to_be32(table->ACPILevel.SclkFrequency);
   3046 	table->ACPILevel.CgSpllFuncCntl = cpu_to_be32(table->ACPILevel.CgSpllFuncCntl);
   3047 	table->ACPILevel.CgSpllFuncCntl2 = cpu_to_be32(table->ACPILevel.CgSpllFuncCntl2);
   3048 	table->ACPILevel.CgSpllFuncCntl3 = cpu_to_be32(table->ACPILevel.CgSpllFuncCntl3);
   3049 	table->ACPILevel.CgSpllFuncCntl4 = cpu_to_be32(table->ACPILevel.CgSpllFuncCntl4);
   3050 	table->ACPILevel.SpllSpreadSpectrum = cpu_to_be32(table->ACPILevel.SpllSpreadSpectrum);
   3051 	table->ACPILevel.SpllSpreadSpectrum2 = cpu_to_be32(table->ACPILevel.SpllSpreadSpectrum2);
   3052 	table->ACPILevel.CcPwrDynRm = cpu_to_be32(table->ACPILevel.CcPwrDynRm);
   3053 	table->ACPILevel.CcPwrDynRm1 = cpu_to_be32(table->ACPILevel.CcPwrDynRm1);
   3054 
   3055 	table->MemoryACPILevel.MinVddc = table->ACPILevel.MinVddc;
   3056 	table->MemoryACPILevel.MinVddcPhases = table->ACPILevel.MinVddcPhases;
   3057 
   3058 	if (pi->vddci_control != CISLANDS_VOLTAGE_CONTROL_NONE) {
   3059 		if (pi->acpi_vddci)
   3060 			table->MemoryACPILevel.MinVddci =
   3061 				cpu_to_be32(pi->acpi_vddci * VOLTAGE_SCALE);
   3062 		else
   3063 			table->MemoryACPILevel.MinVddci =
   3064 				cpu_to_be32(pi->min_vddci_in_pp_table * VOLTAGE_SCALE);
   3065 	}
   3066 
   3067 	if (ci_populate_mvdd_value(rdev, 0, &voltage_level))
   3068 		table->MemoryACPILevel.MinMvdd = 0;
   3069 	else
   3070 		table->MemoryACPILevel.MinMvdd =
   3071 			cpu_to_be32(voltage_level.Voltage * VOLTAGE_SCALE);
   3072 
   3073 	mclk_pwrmgt_cntl |= MRDCK0_RESET | MRDCK1_RESET;
   3074 	mclk_pwrmgt_cntl &= ~(MRDCK0_PDNB | MRDCK1_PDNB);
   3075 
   3076 	dll_cntl &= ~(MRDCK0_BYPASS | MRDCK1_BYPASS);
   3077 
   3078 	table->MemoryACPILevel.DllCntl = cpu_to_be32(dll_cntl);
   3079 	table->MemoryACPILevel.MclkPwrmgtCntl = cpu_to_be32(mclk_pwrmgt_cntl);
   3080 	table->MemoryACPILevel.MpllAdFuncCntl =
   3081 		cpu_to_be32(pi->clock_registers.mpll_ad_func_cntl);
   3082 	table->MemoryACPILevel.MpllDqFuncCntl =
   3083 		cpu_to_be32(pi->clock_registers.mpll_dq_func_cntl);
   3084 	table->MemoryACPILevel.MpllFuncCntl =
   3085 		cpu_to_be32(pi->clock_registers.mpll_func_cntl);
   3086 	table->MemoryACPILevel.MpllFuncCntl_1 =
   3087 		cpu_to_be32(pi->clock_registers.mpll_func_cntl_1);
   3088 	table->MemoryACPILevel.MpllFuncCntl_2 =
   3089 		cpu_to_be32(pi->clock_registers.mpll_func_cntl_2);
   3090 	table->MemoryACPILevel.MpllSs1 = cpu_to_be32(pi->clock_registers.mpll_ss1);
   3091 	table->MemoryACPILevel.MpllSs2 = cpu_to_be32(pi->clock_registers.mpll_ss2);
   3092 
   3093 	table->MemoryACPILevel.EnabledForThrottle = 0;
   3094 	table->MemoryACPILevel.EnabledForActivity = 0;
   3095 	table->MemoryACPILevel.UpH = 0;
   3096 	table->MemoryACPILevel.DownH = 100;
   3097 	table->MemoryACPILevel.VoltageDownH = 0;
   3098 	table->MemoryACPILevel.ActivityLevel =
   3099 		cpu_to_be16((u16)pi->mclk_activity_target);
   3100 
   3101 	table->MemoryACPILevel.StutterEnable = false;
   3102 	table->MemoryACPILevel.StrobeEnable = false;
   3103 	table->MemoryACPILevel.EdcReadEnable = false;
   3104 	table->MemoryACPILevel.EdcWriteEnable = false;
   3105 	table->MemoryACPILevel.RttEnable = false;
   3106 
   3107 	return 0;
   3108 }
   3109 
   3110 
   3111 static int ci_enable_ulv(struct radeon_device *rdev, bool enable)
   3112 {
   3113 	struct ci_power_info *pi = ci_get_pi(rdev);
   3114 	struct ci_ulv_parm *ulv = &pi->ulv;
   3115 
   3116 	if (ulv->supported) {
   3117 		if (enable)
   3118 			return (ci_send_msg_to_smc(rdev, PPSMC_MSG_EnableULV) == PPSMC_Result_OK) ?
   3119 				0 : -EINVAL;
   3120 		else
   3121 			return (ci_send_msg_to_smc(rdev, PPSMC_MSG_DisableULV) == PPSMC_Result_OK) ?
   3122 				0 : -EINVAL;
   3123 	}
   3124 
   3125 	return 0;
   3126 }
   3127 
   3128 static int ci_populate_ulv_level(struct radeon_device *rdev,
   3129 				 SMU7_Discrete_Ulv *state)
   3130 {
   3131 	struct ci_power_info *pi = ci_get_pi(rdev);
   3132 	u16 ulv_voltage = rdev->pm.dpm.backbias_response_time;
   3133 
   3134 	state->CcPwrDynRm = 0;
   3135 	state->CcPwrDynRm1 = 0;
   3136 
   3137 	if (ulv_voltage == 0) {
   3138 		pi->ulv.supported = false;
   3139 		return 0;
   3140 	}
   3141 
   3142 	if (pi->voltage_control != CISLANDS_VOLTAGE_CONTROL_BY_SVID2) {
   3143 		if (ulv_voltage > rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[0].v)
   3144 			state->VddcOffset = 0;
   3145 		else
   3146 			state->VddcOffset =
   3147 				rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[0].v - ulv_voltage;
   3148 	} else {
   3149 		if (ulv_voltage > rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[0].v)
   3150 			state->VddcOffsetVid = 0;
   3151 		else
   3152 			state->VddcOffsetVid = (u8)
   3153 				((rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries[0].v - ulv_voltage) *
   3154 				 VOLTAGE_VID_OFFSET_SCALE2 / VOLTAGE_VID_OFFSET_SCALE1);
   3155 	}
   3156 	state->VddcPhase = pi->vddc_phase_shed_control ? 0 : 1;
   3157 
   3158 	state->CcPwrDynRm = cpu_to_be32(state->CcPwrDynRm);
   3159 	state->CcPwrDynRm1 = cpu_to_be32(state->CcPwrDynRm1);
   3160 	state->VddcOffset = cpu_to_be16(state->VddcOffset);
   3161 
   3162 	return 0;
   3163 }
   3164 
   3165 static int ci_calculate_sclk_params(struct radeon_device *rdev,
   3166 				    u32 engine_clock,
   3167 				    SMU7_Discrete_GraphicsLevel *sclk)
   3168 {
   3169 	struct ci_power_info *pi = ci_get_pi(rdev);
   3170 	struct atom_clock_dividers dividers;
   3171 	u32 spll_func_cntl_3 = pi->clock_registers.cg_spll_func_cntl_3;
   3172 	u32 spll_func_cntl_4 = pi->clock_registers.cg_spll_func_cntl_4;
   3173 	u32 cg_spll_spread_spectrum = pi->clock_registers.cg_spll_spread_spectrum;
   3174 	u32 cg_spll_spread_spectrum_2 = pi->clock_registers.cg_spll_spread_spectrum_2;
   3175 	u32 reference_clock = rdev->clock.spll.reference_freq;
   3176 	u32 reference_divider;
   3177 	u32 fbdiv;
   3178 	int ret;
   3179 
   3180 	ret = radeon_atom_get_clock_dividers(rdev,
   3181 					     COMPUTE_GPUCLK_INPUT_FLAG_SCLK,
   3182 					     engine_clock, false, &dividers);
   3183 	if (ret)
   3184 		return ret;
   3185 
   3186 	reference_divider = 1 + dividers.ref_div;
   3187 	fbdiv = dividers.fb_div & 0x3FFFFFF;
   3188 
   3189 	spll_func_cntl_3 &= ~SPLL_FB_DIV_MASK;
   3190 	spll_func_cntl_3 |= SPLL_FB_DIV(fbdiv);
   3191 	spll_func_cntl_3 |= SPLL_DITHEN;
   3192 
   3193 	if (pi->caps_sclk_ss_support) {
   3194 		struct radeon_atom_ss ss;
   3195 		u32 vco_freq = engine_clock * dividers.post_div;
   3196 
   3197 		if (radeon_atombios_get_asic_ss_info(rdev, &ss,
   3198 						     ASIC_INTERNAL_ENGINE_SS, vco_freq)) {
   3199 			u32 clk_s = reference_clock * 5 / (reference_divider * ss.rate);
   3200 			u32 clk_v = 4 * ss.percentage * fbdiv / (clk_s * 10000);
   3201 
   3202 			cg_spll_spread_spectrum &= ~CLK_S_MASK;
   3203 			cg_spll_spread_spectrum |= CLK_S(clk_s);
   3204 			cg_spll_spread_spectrum |= SSEN;
   3205 
   3206 			cg_spll_spread_spectrum_2 &= ~CLK_V_MASK;
   3207 			cg_spll_spread_spectrum_2 |= CLK_V(clk_v);
   3208 		}
   3209 	}
   3210 
   3211 	sclk->SclkFrequency = engine_clock;
   3212 	sclk->CgSpllFuncCntl3 = spll_func_cntl_3;
   3213 	sclk->CgSpllFuncCntl4 = spll_func_cntl_4;
   3214 	sclk->SpllSpreadSpectrum = cg_spll_spread_spectrum;
   3215 	sclk->SpllSpreadSpectrum2  = cg_spll_spread_spectrum_2;
   3216 	sclk->SclkDid = (u8)dividers.post_divider;
   3217 
   3218 	return 0;
   3219 }
   3220 
   3221 static int ci_populate_single_graphic_level(struct radeon_device *rdev,
   3222 					    u32 engine_clock,
   3223 					    u16 sclk_activity_level_t,
   3224 					    SMU7_Discrete_GraphicsLevel *graphic_level)
   3225 {
   3226 	struct ci_power_info *pi = ci_get_pi(rdev);
   3227 	int ret;
   3228 
   3229 	ret = ci_calculate_sclk_params(rdev, engine_clock, graphic_level);
   3230 	if (ret)
   3231 		return ret;
   3232 
   3233 	ret = ci_get_dependency_volt_by_clk(rdev,
   3234 					    &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk,
   3235 					    engine_clock, &graphic_level->MinVddc);
   3236 	if (ret)
   3237 		return ret;
   3238 
   3239 	graphic_level->SclkFrequency = engine_clock;
   3240 
   3241 	graphic_level->Flags =  0;
   3242 	graphic_level->MinVddcPhases = 1;
   3243 
   3244 	if (pi->vddc_phase_shed_control)
   3245 		ci_populate_phase_value_based_on_sclk(rdev,
   3246 						      &rdev->pm.dpm.dyn_state.phase_shedding_limits_table,
   3247 						      engine_clock,
   3248 						      &graphic_level->MinVddcPhases);
   3249 
   3250 	graphic_level->ActivityLevel = sclk_activity_level_t;
   3251 
   3252 	graphic_level->CcPwrDynRm = 0;
   3253 	graphic_level->CcPwrDynRm1 = 0;
   3254 	graphic_level->EnabledForThrottle = 1;
   3255 	graphic_level->UpH = 0;
   3256 	graphic_level->DownH = 0;
   3257 	graphic_level->VoltageDownH = 0;
   3258 	graphic_level->PowerThrottle = 0;
   3259 
   3260 	if (pi->caps_sclk_ds)
   3261 		graphic_level->DeepSleepDivId = ci_get_sleep_divider_id_from_clock(rdev,
   3262 										   engine_clock,
   3263 										   CISLAND_MINIMUM_ENGINE_CLOCK);
   3264 
   3265 	graphic_level->DisplayWatermark = PPSMC_DISPLAY_WATERMARK_LOW;
   3266 
   3267 	graphic_level->Flags = cpu_to_be32(graphic_level->Flags);
   3268 	graphic_level->MinVddc = cpu_to_be32(graphic_level->MinVddc * VOLTAGE_SCALE);
   3269 	graphic_level->MinVddcPhases = cpu_to_be32(graphic_level->MinVddcPhases);
   3270 	graphic_level->SclkFrequency = cpu_to_be32(graphic_level->SclkFrequency);
   3271 	graphic_level->ActivityLevel = cpu_to_be16(graphic_level->ActivityLevel);
   3272 	graphic_level->CgSpllFuncCntl3 = cpu_to_be32(graphic_level->CgSpllFuncCntl3);
   3273 	graphic_level->CgSpllFuncCntl4 = cpu_to_be32(graphic_level->CgSpllFuncCntl4);
   3274 	graphic_level->SpllSpreadSpectrum = cpu_to_be32(graphic_level->SpllSpreadSpectrum);
   3275 	graphic_level->SpllSpreadSpectrum2 = cpu_to_be32(graphic_level->SpllSpreadSpectrum2);
   3276 	graphic_level->CcPwrDynRm = cpu_to_be32(graphic_level->CcPwrDynRm);
   3277 	graphic_level->CcPwrDynRm1 = cpu_to_be32(graphic_level->CcPwrDynRm1);
   3278 
   3279 	return 0;
   3280 }
   3281 
   3282 static int ci_populate_all_graphic_levels(struct radeon_device *rdev)
   3283 {
   3284 	struct ci_power_info *pi = ci_get_pi(rdev);
   3285 	struct ci_dpm_table *dpm_table = &pi->dpm_table;
   3286 	u32 level_array_address = pi->dpm_table_start +
   3287 		offsetof(SMU7_Discrete_DpmTable, GraphicsLevel);
   3288 	u32 level_array_size = sizeof(SMU7_Discrete_GraphicsLevel) *
   3289 		SMU7_MAX_LEVELS_GRAPHICS;
   3290 	SMU7_Discrete_GraphicsLevel *levels = pi->smc_state_table.GraphicsLevel;
   3291 	u32 i, ret;
   3292 
   3293 	memset(levels, 0, level_array_size);
   3294 
   3295 	for (i = 0; i < dpm_table->sclk_table.count; i++) {
   3296 		ret = ci_populate_single_graphic_level(rdev,
   3297 						       dpm_table->sclk_table.dpm_levels[i].value,
   3298 						       (u16)pi->activity_target[i],
   3299 						       &pi->smc_state_table.GraphicsLevel[i]);
   3300 		if (ret)
   3301 			return ret;
   3302 		if (i > 1)
   3303 			pi->smc_state_table.GraphicsLevel[i].DeepSleepDivId = 0;
   3304 		if (i == (dpm_table->sclk_table.count - 1))
   3305 			pi->smc_state_table.GraphicsLevel[i].DisplayWatermark =
   3306 				PPSMC_DISPLAY_WATERMARK_HIGH;
   3307 	}
   3308 	pi->smc_state_table.GraphicsLevel[0].EnabledForActivity = 1;
   3309 
   3310 	pi->smc_state_table.GraphicsDpmLevelCount = (u8)dpm_table->sclk_table.count;
   3311 	pi->dpm_level_enable_mask.sclk_dpm_enable_mask =
   3312 		ci_get_dpm_level_enable_mask_value(&dpm_table->sclk_table);
   3313 
   3314 	ret = ci_copy_bytes_to_smc(rdev, level_array_address,
   3315 				   (u8 *)levels, level_array_size,
   3316 				   pi->sram_end);
   3317 	if (ret)
   3318 		return ret;
   3319 
   3320 	return 0;
   3321 }
   3322 
   3323 static int ci_populate_ulv_state(struct radeon_device *rdev,
   3324 				 SMU7_Discrete_Ulv *ulv_level)
   3325 {
   3326 	return ci_populate_ulv_level(rdev, ulv_level);
   3327 }
   3328 
   3329 static int ci_populate_all_memory_levels(struct radeon_device *rdev)
   3330 {
   3331 	struct ci_power_info *pi = ci_get_pi(rdev);
   3332 	struct ci_dpm_table *dpm_table = &pi->dpm_table;
   3333 	u32 level_array_address = pi->dpm_table_start +
   3334 		offsetof(SMU7_Discrete_DpmTable, MemoryLevel);
   3335 	u32 level_array_size = sizeof(SMU7_Discrete_MemoryLevel) *
   3336 		SMU7_MAX_LEVELS_MEMORY;
   3337 	SMU7_Discrete_MemoryLevel *levels = pi->smc_state_table.MemoryLevel;
   3338 	u32 i, ret;
   3339 
   3340 	memset(levels, 0, level_array_size);
   3341 
   3342 	for (i = 0; i < dpm_table->mclk_table.count; i++) {
   3343 		if (dpm_table->mclk_table.dpm_levels[i].value == 0)
   3344 			return -EINVAL;
   3345 		ret = ci_populate_single_memory_level(rdev,
   3346 						      dpm_table->mclk_table.dpm_levels[i].value,
   3347 						      &pi->smc_state_table.MemoryLevel[i]);
   3348 		if (ret)
   3349 			return ret;
   3350 	}
   3351 
   3352 	pi->smc_state_table.MemoryLevel[0].EnabledForActivity = 1;
   3353 
   3354 	if ((dpm_table->mclk_table.count >= 2) &&
   3355 	    ((rdev->pdev->device == 0x67B0) || (rdev->pdev->device == 0x67B1))) {
   3356 		pi->smc_state_table.MemoryLevel[1].MinVddc =
   3357 			pi->smc_state_table.MemoryLevel[0].MinVddc;
   3358 		pi->smc_state_table.MemoryLevel[1].MinVddcPhases =
   3359 			pi->smc_state_table.MemoryLevel[0].MinVddcPhases;
   3360 	}
   3361 
   3362 	pi->smc_state_table.MemoryLevel[0].ActivityLevel = cpu_to_be16(0x1F);
   3363 
   3364 	pi->smc_state_table.MemoryDpmLevelCount = (u8)dpm_table->mclk_table.count;
   3365 	pi->dpm_level_enable_mask.mclk_dpm_enable_mask =
   3366 		ci_get_dpm_level_enable_mask_value(&dpm_table->mclk_table);
   3367 
   3368 	pi->smc_state_table.MemoryLevel[dpm_table->mclk_table.count - 1].DisplayWatermark =
   3369 		PPSMC_DISPLAY_WATERMARK_HIGH;
   3370 
   3371 	ret = ci_copy_bytes_to_smc(rdev, level_array_address,
   3372 				   (u8 *)levels, level_array_size,
   3373 				   pi->sram_end);
   3374 	if (ret)
   3375 		return ret;
   3376 
   3377 	return 0;
   3378 }
   3379 
   3380 static void ci_reset_single_dpm_table(struct radeon_device *rdev,
   3381 				      struct ci_single_dpm_table* dpm_table,
   3382 				      u32 count)
   3383 {
   3384 	u32 i;
   3385 
   3386 	dpm_table->count = count;
   3387 	for (i = 0; i < MAX_REGULAR_DPM_NUMBER; i++)
   3388 		dpm_table->dpm_levels[i].enabled = false;
   3389 }
   3390 
   3391 static void ci_setup_pcie_table_entry(struct ci_single_dpm_table* dpm_table,
   3392 				      u32 index, u32 pcie_gen, u32 pcie_lanes)
   3393 {
   3394 	dpm_table->dpm_levels[index].value = pcie_gen;
   3395 	dpm_table->dpm_levels[index].param1 = pcie_lanes;
   3396 	dpm_table->dpm_levels[index].enabled = true;
   3397 }
   3398 
   3399 static int ci_setup_default_pcie_tables(struct radeon_device *rdev)
   3400 {
   3401 	struct ci_power_info *pi = ci_get_pi(rdev);
   3402 
   3403 	if (!pi->use_pcie_performance_levels && !pi->use_pcie_powersaving_levels)
   3404 		return -EINVAL;
   3405 
   3406 	if (pi->use_pcie_performance_levels && !pi->use_pcie_powersaving_levels) {
   3407 		pi->pcie_gen_powersaving = pi->pcie_gen_performance;
   3408 		pi->pcie_lane_powersaving = pi->pcie_lane_performance;
   3409 	} else if (!pi->use_pcie_performance_levels && pi->use_pcie_powersaving_levels) {
   3410 		pi->pcie_gen_performance = pi->pcie_gen_powersaving;
   3411 		pi->pcie_lane_performance = pi->pcie_lane_powersaving;
   3412 	}
   3413 
   3414 	ci_reset_single_dpm_table(rdev,
   3415 				  &pi->dpm_table.pcie_speed_table,
   3416 				  SMU7_MAX_LEVELS_LINK);
   3417 
   3418 	if (rdev->family == CHIP_BONAIRE)
   3419 		ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 0,
   3420 					  pi->pcie_gen_powersaving.min,
   3421 					  pi->pcie_lane_powersaving.max);
   3422 	else
   3423 		ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 0,
   3424 					  pi->pcie_gen_powersaving.min,
   3425 					  pi->pcie_lane_powersaving.min);
   3426 	ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 1,
   3427 				  pi->pcie_gen_performance.min,
   3428 				  pi->pcie_lane_performance.min);
   3429 	ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 2,
   3430 				  pi->pcie_gen_powersaving.min,
   3431 				  pi->pcie_lane_powersaving.max);
   3432 	ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 3,
   3433 				  pi->pcie_gen_performance.min,
   3434 				  pi->pcie_lane_performance.max);
   3435 	ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 4,
   3436 				  pi->pcie_gen_powersaving.max,
   3437 				  pi->pcie_lane_powersaving.max);
   3438 	ci_setup_pcie_table_entry(&pi->dpm_table.pcie_speed_table, 5,
   3439 				  pi->pcie_gen_performance.max,
   3440 				  pi->pcie_lane_performance.max);
   3441 
   3442 	pi->dpm_table.pcie_speed_table.count = 6;
   3443 
   3444 	return 0;
   3445 }
   3446 
   3447 static int ci_setup_default_dpm_tables(struct radeon_device *rdev)
   3448 {
   3449 	struct ci_power_info *pi = ci_get_pi(rdev);
   3450 	struct radeon_clock_voltage_dependency_table *allowed_sclk_vddc_table =
   3451 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk;
   3452 	struct radeon_clock_voltage_dependency_table *allowed_mclk_table =
   3453 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk;
   3454 	struct radeon_cac_leakage_table *std_voltage_table =
   3455 		&rdev->pm.dpm.dyn_state.cac_leakage_table;
   3456 	u32 i;
   3457 
   3458 	if (allowed_sclk_vddc_table == NULL)
   3459 		return -EINVAL;
   3460 	if (allowed_sclk_vddc_table->count < 1)
   3461 		return -EINVAL;
   3462 	if (allowed_mclk_table == NULL)
   3463 		return -EINVAL;
   3464 	if (allowed_mclk_table->count < 1)
   3465 		return -EINVAL;
   3466 
   3467 	memset(&pi->dpm_table, 0, sizeof(struct ci_dpm_table));
   3468 
   3469 	ci_reset_single_dpm_table(rdev,
   3470 				  &pi->dpm_table.sclk_table,
   3471 				  SMU7_MAX_LEVELS_GRAPHICS);
   3472 	ci_reset_single_dpm_table(rdev,
   3473 				  &pi->dpm_table.mclk_table,
   3474 				  SMU7_MAX_LEVELS_MEMORY);
   3475 	ci_reset_single_dpm_table(rdev,
   3476 				  &pi->dpm_table.vddc_table,
   3477 				  SMU7_MAX_LEVELS_VDDC);
   3478 	ci_reset_single_dpm_table(rdev,
   3479 				  &pi->dpm_table.vddci_table,
   3480 				  SMU7_MAX_LEVELS_VDDCI);
   3481 	ci_reset_single_dpm_table(rdev,
   3482 				  &pi->dpm_table.mvdd_table,
   3483 				  SMU7_MAX_LEVELS_MVDD);
   3484 
   3485 	pi->dpm_table.sclk_table.count = 0;
   3486 	for (i = 0; i < allowed_sclk_vddc_table->count; i++) {
   3487 		if ((i == 0) ||
   3488 		    (pi->dpm_table.sclk_table.dpm_levels[pi->dpm_table.sclk_table.count-1].value !=
   3489 		     allowed_sclk_vddc_table->entries[i].clk)) {
   3490 			pi->dpm_table.sclk_table.dpm_levels[pi->dpm_table.sclk_table.count].value =
   3491 				allowed_sclk_vddc_table->entries[i].clk;
   3492 			pi->dpm_table.sclk_table.dpm_levels[pi->dpm_table.sclk_table.count].enabled =
   3493 				(i == 0) ? true : false;
   3494 			pi->dpm_table.sclk_table.count++;
   3495 		}
   3496 	}
   3497 
   3498 	pi->dpm_table.mclk_table.count = 0;
   3499 	for (i = 0; i < allowed_mclk_table->count; i++) {
   3500 		if ((i == 0) ||
   3501 		    (pi->dpm_table.mclk_table.dpm_levels[pi->dpm_table.mclk_table.count-1].value !=
   3502 		     allowed_mclk_table->entries[i].clk)) {
   3503 			pi->dpm_table.mclk_table.dpm_levels[pi->dpm_table.mclk_table.count].value =
   3504 				allowed_mclk_table->entries[i].clk;
   3505 			pi->dpm_table.mclk_table.dpm_levels[pi->dpm_table.mclk_table.count].enabled =
   3506 				(i == 0) ? true : false;
   3507 			pi->dpm_table.mclk_table.count++;
   3508 		}
   3509 	}
   3510 
   3511 	for (i = 0; i < allowed_sclk_vddc_table->count; i++) {
   3512 		pi->dpm_table.vddc_table.dpm_levels[i].value =
   3513 			allowed_sclk_vddc_table->entries[i].v;
   3514 		pi->dpm_table.vddc_table.dpm_levels[i].param1 =
   3515 			std_voltage_table->entries[i].leakage;
   3516 		pi->dpm_table.vddc_table.dpm_levels[i].enabled = true;
   3517 	}
   3518 	pi->dpm_table.vddc_table.count = allowed_sclk_vddc_table->count;
   3519 
   3520 	allowed_mclk_table = &rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk;
   3521 	if (allowed_mclk_table) {
   3522 		for (i = 0; i < allowed_mclk_table->count; i++) {
   3523 			pi->dpm_table.vddci_table.dpm_levels[i].value =
   3524 				allowed_mclk_table->entries[i].v;
   3525 			pi->dpm_table.vddci_table.dpm_levels[i].enabled = true;
   3526 		}
   3527 		pi->dpm_table.vddci_table.count = allowed_mclk_table->count;
   3528 	}
   3529 
   3530 	allowed_mclk_table = &rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk;
   3531 	if (allowed_mclk_table) {
   3532 		for (i = 0; i < allowed_mclk_table->count; i++) {
   3533 			pi->dpm_table.mvdd_table.dpm_levels[i].value =
   3534 				allowed_mclk_table->entries[i].v;
   3535 			pi->dpm_table.mvdd_table.dpm_levels[i].enabled = true;
   3536 		}
   3537 		pi->dpm_table.mvdd_table.count = allowed_mclk_table->count;
   3538 	}
   3539 
   3540 	ci_setup_default_pcie_tables(rdev);
   3541 
   3542 	return 0;
   3543 }
   3544 
   3545 static int ci_find_boot_level(struct ci_single_dpm_table *table,
   3546 			      u32 value, u32 *boot_level)
   3547 {
   3548 	u32 i;
   3549 	int ret = -EINVAL;
   3550 
   3551 	for(i = 0; i < table->count; i++) {
   3552 		if (value == table->dpm_levels[i].value) {
   3553 			*boot_level = i;
   3554 			ret = 0;
   3555 		}
   3556 	}
   3557 
   3558 	return ret;
   3559 }
   3560 
   3561 static int ci_init_smc_table(struct radeon_device *rdev)
   3562 {
   3563 	struct ci_power_info *pi = ci_get_pi(rdev);
   3564 	struct ci_ulv_parm *ulv = &pi->ulv;
   3565 	struct radeon_ps *radeon_boot_state = rdev->pm.dpm.boot_ps;
   3566 	SMU7_Discrete_DpmTable *table = &pi->smc_state_table;
   3567 	int ret;
   3568 
   3569 	ret = ci_setup_default_dpm_tables(rdev);
   3570 	if (ret)
   3571 		return ret;
   3572 
   3573 	if (pi->voltage_control != CISLANDS_VOLTAGE_CONTROL_NONE)
   3574 		ci_populate_smc_voltage_tables(rdev, table);
   3575 
   3576 	ci_init_fps_limits(rdev);
   3577 
   3578 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_HARDWAREDC)
   3579 		table->SystemFlags |= PPSMC_SYSTEMFLAG_GPIO_DC;
   3580 
   3581 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_STEPVDDC)
   3582 		table->SystemFlags |= PPSMC_SYSTEMFLAG_STEPVDDC;
   3583 
   3584 	if (pi->mem_gddr5)
   3585 		table->SystemFlags |= PPSMC_SYSTEMFLAG_GDDR5;
   3586 
   3587 	if (ulv->supported) {
   3588 		ret = ci_populate_ulv_state(rdev, &pi->smc_state_table.Ulv);
   3589 		if (ret)
   3590 			return ret;
   3591 		WREG32_SMC(CG_ULV_PARAMETER, ulv->cg_ulv_parameter);
   3592 	}
   3593 
   3594 	ret = ci_populate_all_graphic_levels(rdev);
   3595 	if (ret)
   3596 		return ret;
   3597 
   3598 	ret = ci_populate_all_memory_levels(rdev);
   3599 	if (ret)
   3600 		return ret;
   3601 
   3602 	ci_populate_smc_link_level(rdev, table);
   3603 
   3604 	ret = ci_populate_smc_acpi_level(rdev, table);
   3605 	if (ret)
   3606 		return ret;
   3607 
   3608 	ret = ci_populate_smc_vce_level(rdev, table);
   3609 	if (ret)
   3610 		return ret;
   3611 
   3612 	ret = ci_populate_smc_acp_level(rdev, table);
   3613 	if (ret)
   3614 		return ret;
   3615 
   3616 	ret = ci_populate_smc_samu_level(rdev, table);
   3617 	if (ret)
   3618 		return ret;
   3619 
   3620 	ret = ci_do_program_memory_timing_parameters(rdev);
   3621 	if (ret)
   3622 		return ret;
   3623 
   3624 	ret = ci_populate_smc_uvd_level(rdev, table);
   3625 	if (ret)
   3626 		return ret;
   3627 
   3628 	table->UvdBootLevel  = 0;
   3629 	table->VceBootLevel  = 0;
   3630 	table->AcpBootLevel  = 0;
   3631 	table->SamuBootLevel  = 0;
   3632 	table->GraphicsBootLevel  = 0;
   3633 	table->MemoryBootLevel  = 0;
   3634 
   3635 	ret = ci_find_boot_level(&pi->dpm_table.sclk_table,
   3636 				 pi->vbios_boot_state.sclk_bootup_value,
   3637 				 (u32 *)&pi->smc_state_table.GraphicsBootLevel);
   3638 
   3639 	ret = ci_find_boot_level(&pi->dpm_table.mclk_table,
   3640 				 pi->vbios_boot_state.mclk_bootup_value,
   3641 				 (u32 *)&pi->smc_state_table.MemoryBootLevel);
   3642 
   3643 	table->BootVddc = pi->vbios_boot_state.vddc_bootup_value;
   3644 	table->BootVddci = pi->vbios_boot_state.vddci_bootup_value;
   3645 	table->BootMVdd = pi->vbios_boot_state.mvdd_bootup_value;
   3646 
   3647 	ci_populate_smc_initial_state(rdev, radeon_boot_state);
   3648 
   3649 	ret = ci_populate_bapm_parameters_in_dpm_table(rdev);
   3650 	if (ret)
   3651 		return ret;
   3652 
   3653 	table->UVDInterval = 1;
   3654 	table->VCEInterval = 1;
   3655 	table->ACPInterval = 1;
   3656 	table->SAMUInterval = 1;
   3657 	table->GraphicsVoltageChangeEnable = 1;
   3658 	table->GraphicsThermThrottleEnable = 1;
   3659 	table->GraphicsInterval = 1;
   3660 	table->VoltageInterval = 1;
   3661 	table->ThermalInterval = 1;
   3662 	table->TemperatureLimitHigh = (u16)((pi->thermal_temp_setting.temperature_high *
   3663 					     CISLANDS_Q88_FORMAT_CONVERSION_UNIT) / 1000);
   3664 	table->TemperatureLimitLow = (u16)((pi->thermal_temp_setting.temperature_low *
   3665 					    CISLANDS_Q88_FORMAT_CONVERSION_UNIT) / 1000);
   3666 	table->MemoryVoltageChangeEnable = 1;
   3667 	table->MemoryInterval = 1;
   3668 	table->VoltageResponseTime = 0;
   3669 	table->VddcVddciDelta = 4000;
   3670 	table->PhaseResponseTime = 0;
   3671 	table->MemoryThermThrottleEnable = 1;
   3672 	table->PCIeBootLinkLevel = pi->dpm_table.pcie_speed_table.count - 1;
   3673 	table->PCIeGenInterval = 1;
   3674 	if (pi->voltage_control == CISLANDS_VOLTAGE_CONTROL_BY_SVID2)
   3675 		table->SVI2Enable  = 1;
   3676 	else
   3677 		table->SVI2Enable  = 0;
   3678 
   3679 	table->ThermGpio = 17;
   3680 	table->SclkStepSize = 0x4000;
   3681 
   3682 	table->SystemFlags = cpu_to_be32(table->SystemFlags);
   3683 	table->SmioMaskVddcVid = cpu_to_be32(table->SmioMaskVddcVid);
   3684 	table->SmioMaskVddcPhase = cpu_to_be32(table->SmioMaskVddcPhase);
   3685 	table->SmioMaskVddciVid = cpu_to_be32(table->SmioMaskVddciVid);
   3686 	table->SmioMaskMvddVid = cpu_to_be32(table->SmioMaskMvddVid);
   3687 	table->SclkStepSize = cpu_to_be32(table->SclkStepSize);
   3688 	table->TemperatureLimitHigh = cpu_to_be16(table->TemperatureLimitHigh);
   3689 	table->TemperatureLimitLow = cpu_to_be16(table->TemperatureLimitLow);
   3690 	table->VddcVddciDelta = cpu_to_be16(table->VddcVddciDelta);
   3691 	table->VoltageResponseTime = cpu_to_be16(table->VoltageResponseTime);
   3692 	table->PhaseResponseTime = cpu_to_be16(table->PhaseResponseTime);
   3693 	table->BootVddc = cpu_to_be16(table->BootVddc * VOLTAGE_SCALE);
   3694 	table->BootVddci = cpu_to_be16(table->BootVddci * VOLTAGE_SCALE);
   3695 	table->BootMVdd = cpu_to_be16(table->BootMVdd * VOLTAGE_SCALE);
   3696 
   3697 	ret = ci_copy_bytes_to_smc(rdev,
   3698 				   pi->dpm_table_start +
   3699 				   offsetof(SMU7_Discrete_DpmTable, SystemFlags),
   3700 				   (u8 *)&table->SystemFlags,
   3701 				   sizeof(SMU7_Discrete_DpmTable) - 3 * sizeof(SMU7_PIDController),
   3702 				   pi->sram_end);
   3703 	if (ret)
   3704 		return ret;
   3705 
   3706 	return 0;
   3707 }
   3708 
   3709 static void ci_trim_single_dpm_states(struct radeon_device *rdev,
   3710 				      struct ci_single_dpm_table *dpm_table,
   3711 				      u32 low_limit, u32 high_limit)
   3712 {
   3713 	u32 i;
   3714 
   3715 	for (i = 0; i < dpm_table->count; i++) {
   3716 		if ((dpm_table->dpm_levels[i].value < low_limit) ||
   3717 		    (dpm_table->dpm_levels[i].value > high_limit))
   3718 			dpm_table->dpm_levels[i].enabled = false;
   3719 		else
   3720 			dpm_table->dpm_levels[i].enabled = true;
   3721 	}
   3722 }
   3723 
   3724 static void ci_trim_pcie_dpm_states(struct radeon_device *rdev,
   3725 				    u32 speed_low, u32 lanes_low,
   3726 				    u32 speed_high, u32 lanes_high)
   3727 {
   3728 	struct ci_power_info *pi = ci_get_pi(rdev);
   3729 	struct ci_single_dpm_table *pcie_table = &pi->dpm_table.pcie_speed_table;
   3730 	u32 i, j;
   3731 
   3732 	for (i = 0; i < pcie_table->count; i++) {
   3733 		if ((pcie_table->dpm_levels[i].value < speed_low) ||
   3734 		    (pcie_table->dpm_levels[i].param1 < lanes_low) ||
   3735 		    (pcie_table->dpm_levels[i].value > speed_high) ||
   3736 		    (pcie_table->dpm_levels[i].param1 > lanes_high))
   3737 			pcie_table->dpm_levels[i].enabled = false;
   3738 		else
   3739 			pcie_table->dpm_levels[i].enabled = true;
   3740 	}
   3741 
   3742 	for (i = 0; i < pcie_table->count; i++) {
   3743 		if (pcie_table->dpm_levels[i].enabled) {
   3744 			for (j = i + 1; j < pcie_table->count; j++) {
   3745 				if (pcie_table->dpm_levels[j].enabled) {
   3746 					if ((pcie_table->dpm_levels[i].value == pcie_table->dpm_levels[j].value) &&
   3747 					    (pcie_table->dpm_levels[i].param1 == pcie_table->dpm_levels[j].param1))
   3748 						pcie_table->dpm_levels[j].enabled = false;
   3749 				}
   3750 			}
   3751 		}
   3752 	}
   3753 }
   3754 
   3755 static int ci_trim_dpm_states(struct radeon_device *rdev,
   3756 			      struct radeon_ps *radeon_state)
   3757 {
   3758 	struct ci_ps *state = ci_get_ps(radeon_state);
   3759 	struct ci_power_info *pi = ci_get_pi(rdev);
   3760 	u32 high_limit_count;
   3761 
   3762 	if (state->performance_level_count < 1)
   3763 		return -EINVAL;
   3764 
   3765 	if (state->performance_level_count == 1)
   3766 		high_limit_count = 0;
   3767 	else
   3768 		high_limit_count = 1;
   3769 
   3770 	ci_trim_single_dpm_states(rdev,
   3771 				  &pi->dpm_table.sclk_table,
   3772 				  state->performance_levels[0].sclk,
   3773 				  state->performance_levels[high_limit_count].sclk);
   3774 
   3775 	ci_trim_single_dpm_states(rdev,
   3776 				  &pi->dpm_table.mclk_table,
   3777 				  state->performance_levels[0].mclk,
   3778 				  state->performance_levels[high_limit_count].mclk);
   3779 
   3780 	ci_trim_pcie_dpm_states(rdev,
   3781 				state->performance_levels[0].pcie_gen,
   3782 				state->performance_levels[0].pcie_lane,
   3783 				state->performance_levels[high_limit_count].pcie_gen,
   3784 				state->performance_levels[high_limit_count].pcie_lane);
   3785 
   3786 	return 0;
   3787 }
   3788 
   3789 static int ci_apply_disp_minimum_voltage_request(struct radeon_device *rdev)
   3790 {
   3791 	struct radeon_clock_voltage_dependency_table *disp_voltage_table =
   3792 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk;
   3793 	struct radeon_clock_voltage_dependency_table *vddc_table =
   3794 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk;
   3795 	u32 requested_voltage = 0;
   3796 	u32 i;
   3797 
   3798 	if (disp_voltage_table == NULL)
   3799 		return -EINVAL;
   3800 	if (!disp_voltage_table->count)
   3801 		return -EINVAL;
   3802 
   3803 	for (i = 0; i < disp_voltage_table->count; i++) {
   3804 		if (rdev->clock.current_dispclk == disp_voltage_table->entries[i].clk)
   3805 			requested_voltage = disp_voltage_table->entries[i].v;
   3806 	}
   3807 
   3808 	for (i = 0; i < vddc_table->count; i++) {
   3809 		if (requested_voltage <= vddc_table->entries[i].v) {
   3810 			requested_voltage = vddc_table->entries[i].v;
   3811 			return (ci_send_msg_to_smc_with_parameter(rdev,
   3812 								  PPSMC_MSG_VddC_Request,
   3813 								  requested_voltage * VOLTAGE_SCALE) == PPSMC_Result_OK) ?
   3814 				0 : -EINVAL;
   3815 		}
   3816 	}
   3817 
   3818 	return -EINVAL;
   3819 }
   3820 
   3821 static int ci_upload_dpm_level_enable_mask(struct radeon_device *rdev)
   3822 {
   3823 	struct ci_power_info *pi = ci_get_pi(rdev);
   3824 	PPSMC_Result result;
   3825 
   3826 	ci_apply_disp_minimum_voltage_request(rdev);
   3827 
   3828 	if (!pi->sclk_dpm_key_disabled) {
   3829 		if (pi->dpm_level_enable_mask.sclk_dpm_enable_mask) {
   3830 			result = ci_send_msg_to_smc_with_parameter(rdev,
   3831 								   PPSMC_MSG_SCLKDPM_SetEnabledMask,
   3832 								   pi->dpm_level_enable_mask.sclk_dpm_enable_mask);
   3833 			if (result != PPSMC_Result_OK)
   3834 				return -EINVAL;
   3835 		}
   3836 	}
   3837 
   3838 	if (!pi->mclk_dpm_key_disabled) {
   3839 		if (pi->dpm_level_enable_mask.mclk_dpm_enable_mask) {
   3840 			result = ci_send_msg_to_smc_with_parameter(rdev,
   3841 								   PPSMC_MSG_MCLKDPM_SetEnabledMask,
   3842 								   pi->dpm_level_enable_mask.mclk_dpm_enable_mask);
   3843 			if (result != PPSMC_Result_OK)
   3844 				return -EINVAL;
   3845 		}
   3846 	}
   3847 #if 0
   3848 	if (!pi->pcie_dpm_key_disabled) {
   3849 		if (pi->dpm_level_enable_mask.pcie_dpm_enable_mask) {
   3850 			result = ci_send_msg_to_smc_with_parameter(rdev,
   3851 								   PPSMC_MSG_PCIeDPM_SetEnabledMask,
   3852 								   pi->dpm_level_enable_mask.pcie_dpm_enable_mask);
   3853 			if (result != PPSMC_Result_OK)
   3854 				return -EINVAL;
   3855 		}
   3856 	}
   3857 #endif
   3858 	return 0;
   3859 }
   3860 
   3861 static void ci_find_dpm_states_clocks_in_dpm_table(struct radeon_device *rdev,
   3862 						   struct radeon_ps *radeon_state)
   3863 {
   3864 	struct ci_power_info *pi = ci_get_pi(rdev);
   3865 	struct ci_ps *state = ci_get_ps(radeon_state);
   3866 	struct ci_single_dpm_table *sclk_table = &pi->dpm_table.sclk_table;
   3867 	u32 sclk = state->performance_levels[state->performance_level_count-1].sclk;
   3868 	struct ci_single_dpm_table *mclk_table = &pi->dpm_table.mclk_table;
   3869 	u32 mclk = state->performance_levels[state->performance_level_count-1].mclk;
   3870 	u32 i;
   3871 
   3872 	pi->need_update_smu7_dpm_table = 0;
   3873 
   3874 	for (i = 0; i < sclk_table->count; i++) {
   3875 		if (sclk == sclk_table->dpm_levels[i].value)
   3876 			break;
   3877 	}
   3878 
   3879 	if (i >= sclk_table->count) {
   3880 		pi->need_update_smu7_dpm_table |= DPMTABLE_OD_UPDATE_SCLK;
   3881 	} else {
   3882 		/* XXX The current code always reprogrammed the sclk levels,
   3883 		 * but we don't currently handle disp sclk requirements
   3884 		 * so just skip it.
   3885 		 */
   3886 		if (CISLAND_MINIMUM_ENGINE_CLOCK != CISLAND_MINIMUM_ENGINE_CLOCK)
   3887 			pi->need_update_smu7_dpm_table |= DPMTABLE_UPDATE_SCLK;
   3888 	}
   3889 
   3890 	for (i = 0; i < mclk_table->count; i++) {
   3891 		if (mclk == mclk_table->dpm_levels[i].value)
   3892 			break;
   3893 	}
   3894 
   3895 	if (i >= mclk_table->count)
   3896 		pi->need_update_smu7_dpm_table |= DPMTABLE_OD_UPDATE_MCLK;
   3897 
   3898 	if (rdev->pm.dpm.current_active_crtc_count !=
   3899 	    rdev->pm.dpm.new_active_crtc_count)
   3900 		pi->need_update_smu7_dpm_table |= DPMTABLE_UPDATE_MCLK;
   3901 }
   3902 
   3903 static int ci_populate_and_upload_sclk_mclk_dpm_levels(struct radeon_device *rdev,
   3904 						       struct radeon_ps *radeon_state)
   3905 {
   3906 	struct ci_power_info *pi = ci_get_pi(rdev);
   3907 	struct ci_ps *state = ci_get_ps(radeon_state);
   3908 	u32 sclk = state->performance_levels[state->performance_level_count-1].sclk;
   3909 	u32 mclk = state->performance_levels[state->performance_level_count-1].mclk;
   3910 	struct ci_dpm_table *dpm_table = &pi->dpm_table;
   3911 	int ret;
   3912 
   3913 	if (!pi->need_update_smu7_dpm_table)
   3914 		return 0;
   3915 
   3916 	if (pi->need_update_smu7_dpm_table & DPMTABLE_OD_UPDATE_SCLK)
   3917 		dpm_table->sclk_table.dpm_levels[dpm_table->sclk_table.count-1].value = sclk;
   3918 
   3919 	if (pi->need_update_smu7_dpm_table & DPMTABLE_OD_UPDATE_MCLK)
   3920 		dpm_table->mclk_table.dpm_levels[dpm_table->mclk_table.count-1].value = mclk;
   3921 
   3922 	if (pi->need_update_smu7_dpm_table & (DPMTABLE_OD_UPDATE_SCLK | DPMTABLE_UPDATE_SCLK)) {
   3923 		ret = ci_populate_all_graphic_levels(rdev);
   3924 		if (ret)
   3925 			return ret;
   3926 	}
   3927 
   3928 	if (pi->need_update_smu7_dpm_table & (DPMTABLE_OD_UPDATE_MCLK | DPMTABLE_UPDATE_MCLK)) {
   3929 		ret = ci_populate_all_memory_levels(rdev);
   3930 		if (ret)
   3931 			return ret;
   3932 	}
   3933 
   3934 	return 0;
   3935 }
   3936 
   3937 static int ci_enable_uvd_dpm(struct radeon_device *rdev, bool enable)
   3938 {
   3939 	struct ci_power_info *pi = ci_get_pi(rdev);
   3940 	const struct radeon_clock_and_voltage_limits *max_limits;
   3941 	int i;
   3942 
   3943 	if (rdev->pm.dpm.ac_power)
   3944 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
   3945 	else
   3946 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc;
   3947 
   3948 	if (enable) {
   3949 		pi->dpm_level_enable_mask.uvd_dpm_enable_mask = 0;
   3950 
   3951 		for (i = rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.count - 1; i >= 0; i--) {
   3952 			if (rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].v <= max_limits->vddc) {
   3953 				pi->dpm_level_enable_mask.uvd_dpm_enable_mask |= 1 << i;
   3954 
   3955 				if (!pi->caps_uvd_dpm)
   3956 					break;
   3957 			}
   3958 		}
   3959 
   3960 		ci_send_msg_to_smc_with_parameter(rdev,
   3961 						  PPSMC_MSG_UVDDPM_SetEnabledMask,
   3962 						  pi->dpm_level_enable_mask.uvd_dpm_enable_mask);
   3963 
   3964 		if (pi->last_mclk_dpm_enable_mask & 0x1) {
   3965 			pi->uvd_enabled = true;
   3966 			pi->dpm_level_enable_mask.mclk_dpm_enable_mask &= 0xFFFFFFFE;
   3967 			ci_send_msg_to_smc_with_parameter(rdev,
   3968 							  PPSMC_MSG_MCLKDPM_SetEnabledMask,
   3969 							  pi->dpm_level_enable_mask.mclk_dpm_enable_mask);
   3970 		}
   3971 	} else {
   3972 		if (pi->last_mclk_dpm_enable_mask & 0x1) {
   3973 			pi->uvd_enabled = false;
   3974 			pi->dpm_level_enable_mask.mclk_dpm_enable_mask |= 1;
   3975 			ci_send_msg_to_smc_with_parameter(rdev,
   3976 							  PPSMC_MSG_MCLKDPM_SetEnabledMask,
   3977 							  pi->dpm_level_enable_mask.mclk_dpm_enable_mask);
   3978 		}
   3979 	}
   3980 
   3981 	return (ci_send_msg_to_smc(rdev, enable ?
   3982 				   PPSMC_MSG_UVDDPM_Enable : PPSMC_MSG_UVDDPM_Disable) == PPSMC_Result_OK) ?
   3983 		0 : -EINVAL;
   3984 }
   3985 
   3986 static int ci_enable_vce_dpm(struct radeon_device *rdev, bool enable)
   3987 {
   3988 	struct ci_power_info *pi = ci_get_pi(rdev);
   3989 	const struct radeon_clock_and_voltage_limits *max_limits;
   3990 	int i;
   3991 
   3992 	if (rdev->pm.dpm.ac_power)
   3993 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
   3994 	else
   3995 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc;
   3996 
   3997 	if (enable) {
   3998 		pi->dpm_level_enable_mask.vce_dpm_enable_mask = 0;
   3999 		for (i = rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.count - 1; i >= 0; i--) {
   4000 			if (rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries[i].v <= max_limits->vddc) {
   4001 				pi->dpm_level_enable_mask.vce_dpm_enable_mask |= 1 << i;
   4002 
   4003 				if (!pi->caps_vce_dpm)
   4004 					break;
   4005 			}
   4006 		}
   4007 
   4008 		ci_send_msg_to_smc_with_parameter(rdev,
   4009 						  PPSMC_MSG_VCEDPM_SetEnabledMask,
   4010 						  pi->dpm_level_enable_mask.vce_dpm_enable_mask);
   4011 	}
   4012 
   4013 	return (ci_send_msg_to_smc(rdev, enable ?
   4014 				   PPSMC_MSG_VCEDPM_Enable : PPSMC_MSG_VCEDPM_Disable) == PPSMC_Result_OK) ?
   4015 		0 : -EINVAL;
   4016 }
   4017 
   4018 #if 0
   4019 static int ci_enable_samu_dpm(struct radeon_device *rdev, bool enable)
   4020 {
   4021 	struct ci_power_info *pi = ci_get_pi(rdev);
   4022 	const struct radeon_clock_and_voltage_limits *max_limits;
   4023 	int i;
   4024 
   4025 	if (rdev->pm.dpm.ac_power)
   4026 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
   4027 	else
   4028 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc;
   4029 
   4030 	if (enable) {
   4031 		pi->dpm_level_enable_mask.samu_dpm_enable_mask = 0;
   4032 		for (i = rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.count - 1; i >= 0; i--) {
   4033 			if (rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries[i].v <= max_limits->vddc) {
   4034 				pi->dpm_level_enable_mask.samu_dpm_enable_mask |= 1 << i;
   4035 
   4036 				if (!pi->caps_samu_dpm)
   4037 					break;
   4038 			}
   4039 		}
   4040 
   4041 		ci_send_msg_to_smc_with_parameter(rdev,
   4042 						  PPSMC_MSG_SAMUDPM_SetEnabledMask,
   4043 						  pi->dpm_level_enable_mask.samu_dpm_enable_mask);
   4044 	}
   4045 	return (ci_send_msg_to_smc(rdev, enable ?
   4046 				   PPSMC_MSG_SAMUDPM_Enable : PPSMC_MSG_SAMUDPM_Disable) == PPSMC_Result_OK) ?
   4047 		0 : -EINVAL;
   4048 }
   4049 
   4050 static int ci_enable_acp_dpm(struct radeon_device *rdev, bool enable)
   4051 {
   4052 	struct ci_power_info *pi = ci_get_pi(rdev);
   4053 	const struct radeon_clock_and_voltage_limits *max_limits;
   4054 	int i;
   4055 
   4056 	if (rdev->pm.dpm.ac_power)
   4057 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
   4058 	else
   4059 		max_limits = &rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc;
   4060 
   4061 	if (enable) {
   4062 		pi->dpm_level_enable_mask.acp_dpm_enable_mask = 0;
   4063 		for (i = rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.count - 1; i >= 0; i--) {
   4064 			if (rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries[i].v <= max_limits->vddc) {
   4065 				pi->dpm_level_enable_mask.acp_dpm_enable_mask |= 1 << i;
   4066 
   4067 				if (!pi->caps_acp_dpm)
   4068 					break;
   4069 			}
   4070 		}
   4071 
   4072 		ci_send_msg_to_smc_with_parameter(rdev,
   4073 						  PPSMC_MSG_ACPDPM_SetEnabledMask,
   4074 						  pi->dpm_level_enable_mask.acp_dpm_enable_mask);
   4075 	}
   4076 
   4077 	return (ci_send_msg_to_smc(rdev, enable ?
   4078 				   PPSMC_MSG_ACPDPM_Enable : PPSMC_MSG_ACPDPM_Disable) == PPSMC_Result_OK) ?
   4079 		0 : -EINVAL;
   4080 }
   4081 #endif
   4082 
   4083 static int ci_update_uvd_dpm(struct radeon_device *rdev, bool gate)
   4084 {
   4085 	struct ci_power_info *pi = ci_get_pi(rdev);
   4086 	u32 tmp;
   4087 
   4088 	if (!gate) {
   4089 		if (pi->caps_uvd_dpm ||
   4090 		    (rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.count <= 0))
   4091 			pi->smc_state_table.UvdBootLevel = 0;
   4092 		else
   4093 			pi->smc_state_table.UvdBootLevel =
   4094 				rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.count - 1;
   4095 
   4096 		tmp = RREG32_SMC(DPM_TABLE_475);
   4097 		tmp &= ~UvdBootLevel_MASK;
   4098 		tmp |= UvdBootLevel(pi->smc_state_table.UvdBootLevel);
   4099 		WREG32_SMC(DPM_TABLE_475, tmp);
   4100 	}
   4101 
   4102 	return ci_enable_uvd_dpm(rdev, !gate);
   4103 }
   4104 
   4105 static u8 ci_get_vce_boot_level(struct radeon_device *rdev)
   4106 {
   4107 	u8 i;
   4108 	u32 min_evclk = 30000; /* ??? */
   4109 	struct radeon_vce_clock_voltage_dependency_table *table =
   4110 		&rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table;
   4111 
   4112 	for (i = 0; i < table->count; i++) {
   4113 		if (table->entries[i].evclk >= min_evclk)
   4114 			return i;
   4115 	}
   4116 
   4117 	return table->count - 1;
   4118 }
   4119 
   4120 static int ci_update_vce_dpm(struct radeon_device *rdev,
   4121 			     struct radeon_ps *radeon_new_state,
   4122 			     struct radeon_ps *radeon_current_state)
   4123 {
   4124 	struct ci_power_info *pi = ci_get_pi(rdev);
   4125 	int ret = 0;
   4126 	u32 tmp;
   4127 
   4128 	if (radeon_current_state->evclk != radeon_new_state->evclk) {
   4129 		if (radeon_new_state->evclk) {
   4130 			/* turn the clocks on when encoding */
   4131 			cik_update_cg(rdev, RADEON_CG_BLOCK_VCE, false);
   4132 
   4133 			pi->smc_state_table.VceBootLevel = ci_get_vce_boot_level(rdev);
   4134 			tmp = RREG32_SMC(DPM_TABLE_475);
   4135 			tmp &= ~VceBootLevel_MASK;
   4136 			tmp |= VceBootLevel(pi->smc_state_table.VceBootLevel);
   4137 			WREG32_SMC(DPM_TABLE_475, tmp);
   4138 
   4139 			ret = ci_enable_vce_dpm(rdev, true);
   4140 		} else {
   4141 			/* turn the clocks off when not encoding */
   4142 			cik_update_cg(rdev, RADEON_CG_BLOCK_VCE, true);
   4143 
   4144 			ret = ci_enable_vce_dpm(rdev, false);
   4145 		}
   4146 	}
   4147 	return ret;
   4148 }
   4149 
   4150 #if 0
   4151 static int ci_update_samu_dpm(struct radeon_device *rdev, bool gate)
   4152 {
   4153 	return ci_enable_samu_dpm(rdev, gate);
   4154 }
   4155 
   4156 static int ci_update_acp_dpm(struct radeon_device *rdev, bool gate)
   4157 {
   4158 	struct ci_power_info *pi = ci_get_pi(rdev);
   4159 	u32 tmp;
   4160 
   4161 	if (!gate) {
   4162 		pi->smc_state_table.AcpBootLevel = 0;
   4163 
   4164 		tmp = RREG32_SMC(DPM_TABLE_475);
   4165 		tmp &= ~AcpBootLevel_MASK;
   4166 		tmp |= AcpBootLevel(pi->smc_state_table.AcpBootLevel);
   4167 		WREG32_SMC(DPM_TABLE_475, tmp);
   4168 	}
   4169 
   4170 	return ci_enable_acp_dpm(rdev, !gate);
   4171 }
   4172 #endif
   4173 
   4174 static int ci_generate_dpm_level_enable_mask(struct radeon_device *rdev,
   4175 					     struct radeon_ps *radeon_state)
   4176 {
   4177 	struct ci_power_info *pi = ci_get_pi(rdev);
   4178 	int ret;
   4179 
   4180 	ret = ci_trim_dpm_states(rdev, radeon_state);
   4181 	if (ret)
   4182 		return ret;
   4183 
   4184 	pi->dpm_level_enable_mask.sclk_dpm_enable_mask =
   4185 		ci_get_dpm_level_enable_mask_value(&pi->dpm_table.sclk_table);
   4186 	pi->dpm_level_enable_mask.mclk_dpm_enable_mask =
   4187 		ci_get_dpm_level_enable_mask_value(&pi->dpm_table.mclk_table);
   4188 	pi->last_mclk_dpm_enable_mask =
   4189 		pi->dpm_level_enable_mask.mclk_dpm_enable_mask;
   4190 	if (pi->uvd_enabled) {
   4191 		if (pi->dpm_level_enable_mask.mclk_dpm_enable_mask & 1)
   4192 			pi->dpm_level_enable_mask.mclk_dpm_enable_mask &= 0xFFFFFFFE;
   4193 	}
   4194 	pi->dpm_level_enable_mask.pcie_dpm_enable_mask =
   4195 		ci_get_dpm_level_enable_mask_value(&pi->dpm_table.pcie_speed_table);
   4196 
   4197 	return 0;
   4198 }
   4199 
   4200 static u32 ci_get_lowest_enabled_level(struct radeon_device *rdev,
   4201 				       u32 level_mask)
   4202 {
   4203 	u32 level = 0;
   4204 
   4205 	while ((level_mask & (1 << level)) == 0)
   4206 		level++;
   4207 
   4208 	return level;
   4209 }
   4210 
   4211 
   4212 int ci_dpm_force_performance_level(struct radeon_device *rdev,
   4213 				   enum radeon_dpm_forced_level level)
   4214 {
   4215 	struct ci_power_info *pi = ci_get_pi(rdev);
   4216 	u32 tmp, levels, i;
   4217 	int ret;
   4218 
   4219 	if (level == RADEON_DPM_FORCED_LEVEL_HIGH) {
   4220 		if ((!pi->pcie_dpm_key_disabled) &&
   4221 		    pi->dpm_level_enable_mask.pcie_dpm_enable_mask) {
   4222 			levels = 0;
   4223 			tmp = pi->dpm_level_enable_mask.pcie_dpm_enable_mask;
   4224 			while (tmp >>= 1)
   4225 				levels++;
   4226 			if (levels) {
   4227 				ret = ci_dpm_force_state_pcie(rdev, level);
   4228 				if (ret)
   4229 					return ret;
   4230 				for (i = 0; i < rdev->usec_timeout; i++) {
   4231 					tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX_1) &
   4232 					       CURR_PCIE_INDEX_MASK) >> CURR_PCIE_INDEX_SHIFT;
   4233 					if (tmp == levels)
   4234 						break;
   4235 					udelay(1);
   4236 				}
   4237 			}
   4238 		}
   4239 		if ((!pi->sclk_dpm_key_disabled) &&
   4240 		    pi->dpm_level_enable_mask.sclk_dpm_enable_mask) {
   4241 			levels = 0;
   4242 			tmp = pi->dpm_level_enable_mask.sclk_dpm_enable_mask;
   4243 			while (tmp >>= 1)
   4244 				levels++;
   4245 			if (levels) {
   4246 				ret = ci_dpm_force_state_sclk(rdev, levels);
   4247 				if (ret)
   4248 					return ret;
   4249 				for (i = 0; i < rdev->usec_timeout; i++) {
   4250 					tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX) &
   4251 					       CURR_SCLK_INDEX_MASK) >> CURR_SCLK_INDEX_SHIFT;
   4252 					if (tmp == levels)
   4253 						break;
   4254 					udelay(1);
   4255 				}
   4256 			}
   4257 		}
   4258 		if ((!pi->mclk_dpm_key_disabled) &&
   4259 		    pi->dpm_level_enable_mask.mclk_dpm_enable_mask) {
   4260 			levels = 0;
   4261 			tmp = pi->dpm_level_enable_mask.mclk_dpm_enable_mask;
   4262 			while (tmp >>= 1)
   4263 				levels++;
   4264 			if (levels) {
   4265 				ret = ci_dpm_force_state_mclk(rdev, levels);
   4266 				if (ret)
   4267 					return ret;
   4268 				for (i = 0; i < rdev->usec_timeout; i++) {
   4269 					tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX) &
   4270 					       CURR_MCLK_INDEX_MASK) >> CURR_MCLK_INDEX_SHIFT;
   4271 					if (tmp == levels)
   4272 						break;
   4273 					udelay(1);
   4274 				}
   4275 			}
   4276 		}
   4277 	} else if (level == RADEON_DPM_FORCED_LEVEL_LOW) {
   4278 		if ((!pi->sclk_dpm_key_disabled) &&
   4279 		    pi->dpm_level_enable_mask.sclk_dpm_enable_mask) {
   4280 			levels = ci_get_lowest_enabled_level(rdev,
   4281 							     pi->dpm_level_enable_mask.sclk_dpm_enable_mask);
   4282 			ret = ci_dpm_force_state_sclk(rdev, levels);
   4283 			if (ret)
   4284 				return ret;
   4285 			for (i = 0; i < rdev->usec_timeout; i++) {
   4286 				tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX) &
   4287 				       CURR_SCLK_INDEX_MASK) >> CURR_SCLK_INDEX_SHIFT;
   4288 				if (tmp == levels)
   4289 					break;
   4290 				udelay(1);
   4291 			}
   4292 		}
   4293 		if ((!pi->mclk_dpm_key_disabled) &&
   4294 		    pi->dpm_level_enable_mask.mclk_dpm_enable_mask) {
   4295 			levels = ci_get_lowest_enabled_level(rdev,
   4296 							     pi->dpm_level_enable_mask.mclk_dpm_enable_mask);
   4297 			ret = ci_dpm_force_state_mclk(rdev, levels);
   4298 			if (ret)
   4299 				return ret;
   4300 			for (i = 0; i < rdev->usec_timeout; i++) {
   4301 				tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX) &
   4302 				       CURR_MCLK_INDEX_MASK) >> CURR_MCLK_INDEX_SHIFT;
   4303 				if (tmp == levels)
   4304 					break;
   4305 				udelay(1);
   4306 			}
   4307 		}
   4308 		if ((!pi->pcie_dpm_key_disabled) &&
   4309 		    pi->dpm_level_enable_mask.pcie_dpm_enable_mask) {
   4310 			levels = ci_get_lowest_enabled_level(rdev,
   4311 							     pi->dpm_level_enable_mask.pcie_dpm_enable_mask);
   4312 			ret = ci_dpm_force_state_pcie(rdev, levels);
   4313 			if (ret)
   4314 				return ret;
   4315 			for (i = 0; i < rdev->usec_timeout; i++) {
   4316 				tmp = (RREG32_SMC(TARGET_AND_CURRENT_PROFILE_INDEX_1) &
   4317 				       CURR_PCIE_INDEX_MASK) >> CURR_PCIE_INDEX_SHIFT;
   4318 				if (tmp == levels)
   4319 					break;
   4320 				udelay(1);
   4321 			}
   4322 		}
   4323 	} else if (level == RADEON_DPM_FORCED_LEVEL_AUTO) {
   4324 		if (!pi->pcie_dpm_key_disabled) {
   4325 			PPSMC_Result smc_result;
   4326 
   4327 			smc_result = ci_send_msg_to_smc(rdev,
   4328 							PPSMC_MSG_PCIeDPM_UnForceLevel);
   4329 			if (smc_result != PPSMC_Result_OK)
   4330 				return -EINVAL;
   4331 		}
   4332 		ret = ci_upload_dpm_level_enable_mask(rdev);
   4333 		if (ret)
   4334 			return ret;
   4335 	}
   4336 
   4337 	rdev->pm.dpm.forced_level = level;
   4338 
   4339 	return 0;
   4340 }
   4341 
   4342 static int ci_set_mc_special_registers(struct radeon_device *rdev,
   4343 				       struct ci_mc_reg_table *table)
   4344 {
   4345 	struct ci_power_info *pi = ci_get_pi(rdev);
   4346 	u8 i, j, k;
   4347 	u32 temp_reg;
   4348 
   4349 	for (i = 0, j = table->last; i < table->last; i++) {
   4350 		if (j >= SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4351 			return -EINVAL;
   4352 		switch(table->mc_reg_address[i].s1 << 2) {
   4353 		case MC_SEQ_MISC1:
   4354 			temp_reg = RREG32(MC_PMG_CMD_EMRS);
   4355 			table->mc_reg_address[j].s1 = MC_PMG_CMD_EMRS >> 2;
   4356 			table->mc_reg_address[j].s0 = MC_SEQ_PMG_CMD_EMRS_LP >> 2;
   4357 			for (k = 0; k < table->num_entries; k++) {
   4358 				table->mc_reg_table_entry[k].mc_data[j] =
   4359 					((temp_reg & 0xffff0000)) | ((table->mc_reg_table_entry[k].mc_data[i] & 0xffff0000) >> 16);
   4360 			}
   4361 			j++;
   4362 			if (j >= SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4363 				return -EINVAL;
   4364 
   4365 			temp_reg = RREG32(MC_PMG_CMD_MRS);
   4366 			table->mc_reg_address[j].s1 = MC_PMG_CMD_MRS >> 2;
   4367 			table->mc_reg_address[j].s0 = MC_SEQ_PMG_CMD_MRS_LP >> 2;
   4368 			for (k = 0; k < table->num_entries; k++) {
   4369 				table->mc_reg_table_entry[k].mc_data[j] =
   4370 					(temp_reg & 0xffff0000) | (table->mc_reg_table_entry[k].mc_data[i] & 0x0000ffff);
   4371 				if (!pi->mem_gddr5)
   4372 					table->mc_reg_table_entry[k].mc_data[j] |= 0x100;
   4373 			}
   4374 			j++;
   4375 			if (j > SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4376 				return -EINVAL;
   4377 
   4378 			if (!pi->mem_gddr5) {
   4379 				table->mc_reg_address[j].s1 = MC_PMG_AUTO_CMD >> 2;
   4380 				table->mc_reg_address[j].s0 = MC_PMG_AUTO_CMD >> 2;
   4381 				for (k = 0; k < table->num_entries; k++) {
   4382 					table->mc_reg_table_entry[k].mc_data[j] =
   4383 						(table->mc_reg_table_entry[k].mc_data[i] & 0xffff0000) >> 16;
   4384 				}
   4385 				j++;
   4386 				if (j > SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4387 					return -EINVAL;
   4388 			}
   4389 			break;
   4390 		case MC_SEQ_RESERVE_M:
   4391 			temp_reg = RREG32(MC_PMG_CMD_MRS1);
   4392 			table->mc_reg_address[j].s1 = MC_PMG_CMD_MRS1 >> 2;
   4393 			table->mc_reg_address[j].s0 = MC_SEQ_PMG_CMD_MRS1_LP >> 2;
   4394 			for (k = 0; k < table->num_entries; k++) {
   4395 				table->mc_reg_table_entry[k].mc_data[j] =
   4396 					(temp_reg & 0xffff0000) | (table->mc_reg_table_entry[k].mc_data[i] & 0x0000ffff);
   4397 			}
   4398 			j++;
   4399 			if (j > SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4400 				return -EINVAL;
   4401 			break;
   4402 		default:
   4403 			break;
   4404 		}
   4405 
   4406 	}
   4407 
   4408 	table->last = j;
   4409 
   4410 	return 0;
   4411 }
   4412 
   4413 static bool ci_check_s0_mc_reg_index(u16 in_reg, u16 *out_reg)
   4414 {
   4415 	bool result = true;
   4416 
   4417 	switch(in_reg) {
   4418 	case MC_SEQ_RAS_TIMING >> 2:
   4419 		*out_reg = MC_SEQ_RAS_TIMING_LP >> 2;
   4420 		break;
   4421 	case MC_SEQ_DLL_STBY >> 2:
   4422 		*out_reg = MC_SEQ_DLL_STBY_LP >> 2;
   4423 		break;
   4424 	case MC_SEQ_G5PDX_CMD0 >> 2:
   4425 		*out_reg = MC_SEQ_G5PDX_CMD0_LP >> 2;
   4426 		break;
   4427 	case MC_SEQ_G5PDX_CMD1 >> 2:
   4428 		*out_reg = MC_SEQ_G5PDX_CMD1_LP >> 2;
   4429 		break;
   4430 	case MC_SEQ_G5PDX_CTRL >> 2:
   4431 		*out_reg = MC_SEQ_G5PDX_CTRL_LP >> 2;
   4432 		break;
   4433 	case MC_SEQ_CAS_TIMING >> 2:
   4434 		*out_reg = MC_SEQ_CAS_TIMING_LP >> 2;
   4435 		break;
   4436 	case MC_SEQ_MISC_TIMING >> 2:
   4437 		*out_reg = MC_SEQ_MISC_TIMING_LP >> 2;
   4438 		break;
   4439 	case MC_SEQ_MISC_TIMING2 >> 2:
   4440 		*out_reg = MC_SEQ_MISC_TIMING2_LP >> 2;
   4441 		break;
   4442 	case MC_SEQ_PMG_DVS_CMD >> 2:
   4443 		*out_reg = MC_SEQ_PMG_DVS_CMD_LP >> 2;
   4444 		break;
   4445 	case MC_SEQ_PMG_DVS_CTL >> 2:
   4446 		*out_reg = MC_SEQ_PMG_DVS_CTL_LP >> 2;
   4447 		break;
   4448 	case MC_SEQ_RD_CTL_D0 >> 2:
   4449 		*out_reg = MC_SEQ_RD_CTL_D0_LP >> 2;
   4450 		break;
   4451 	case MC_SEQ_RD_CTL_D1 >> 2:
   4452 		*out_reg = MC_SEQ_RD_CTL_D1_LP >> 2;
   4453 		break;
   4454 	case MC_SEQ_WR_CTL_D0 >> 2:
   4455 		*out_reg = MC_SEQ_WR_CTL_D0_LP >> 2;
   4456 		break;
   4457 	case MC_SEQ_WR_CTL_D1 >> 2:
   4458 		*out_reg = MC_SEQ_WR_CTL_D1_LP >> 2;
   4459 		break;
   4460 	case MC_PMG_CMD_EMRS >> 2:
   4461 		*out_reg = MC_SEQ_PMG_CMD_EMRS_LP >> 2;
   4462 		break;
   4463 	case MC_PMG_CMD_MRS >> 2:
   4464 		*out_reg = MC_SEQ_PMG_CMD_MRS_LP >> 2;
   4465 		break;
   4466 	case MC_PMG_CMD_MRS1 >> 2:
   4467 		*out_reg = MC_SEQ_PMG_CMD_MRS1_LP >> 2;
   4468 		break;
   4469 	case MC_SEQ_PMG_TIMING >> 2:
   4470 		*out_reg = MC_SEQ_PMG_TIMING_LP >> 2;
   4471 		break;
   4472 	case MC_PMG_CMD_MRS2 >> 2:
   4473 		*out_reg = MC_SEQ_PMG_CMD_MRS2_LP >> 2;
   4474 		break;
   4475 	case MC_SEQ_WR_CTL_2 >> 2:
   4476 		*out_reg = MC_SEQ_WR_CTL_2_LP >> 2;
   4477 		break;
   4478 	default:
   4479 		result = false;
   4480 		break;
   4481 	}
   4482 
   4483 	return result;
   4484 }
   4485 
   4486 static void ci_set_valid_flag(struct ci_mc_reg_table *table)
   4487 {
   4488 	u8 i, j;
   4489 
   4490 	for (i = 0; i < table->last; i++) {
   4491 		for (j = 1; j < table->num_entries; j++) {
   4492 			if (table->mc_reg_table_entry[j-1].mc_data[i] !=
   4493 			    table->mc_reg_table_entry[j].mc_data[i]) {
   4494 				table->valid_flag |= 1 << i;
   4495 				break;
   4496 			}
   4497 		}
   4498 	}
   4499 }
   4500 
   4501 static void ci_set_s0_mc_reg_index(struct ci_mc_reg_table *table)
   4502 {
   4503 	u32 i;
   4504 	u16 address;
   4505 
   4506 	for (i = 0; i < table->last; i++) {
   4507 		table->mc_reg_address[i].s0 =
   4508 			ci_check_s0_mc_reg_index(table->mc_reg_address[i].s1, &address) ?
   4509 			address : table->mc_reg_address[i].s1;
   4510 	}
   4511 }
   4512 
   4513 static int ci_copy_vbios_mc_reg_table(const struct atom_mc_reg_table *table,
   4514 				      struct ci_mc_reg_table *ci_table)
   4515 {
   4516 	u8 i, j;
   4517 
   4518 	if (table->last > SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4519 		return -EINVAL;
   4520 	if (table->num_entries > MAX_AC_TIMING_ENTRIES)
   4521 		return -EINVAL;
   4522 
   4523 	for (i = 0; i < table->last; i++)
   4524 		ci_table->mc_reg_address[i].s1 = table->mc_reg_address[i].s1;
   4525 
   4526 	ci_table->last = table->last;
   4527 
   4528 	for (i = 0; i < table->num_entries; i++) {
   4529 		ci_table->mc_reg_table_entry[i].mclk_max =
   4530 			table->mc_reg_table_entry[i].mclk_max;
   4531 		for (j = 0; j < table->last; j++)
   4532 			ci_table->mc_reg_table_entry[i].mc_data[j] =
   4533 				table->mc_reg_table_entry[i].mc_data[j];
   4534 	}
   4535 	ci_table->num_entries = table->num_entries;
   4536 
   4537 	return 0;
   4538 }
   4539 
   4540 static int ci_register_patching_mc_seq(struct radeon_device *rdev,
   4541 				       struct ci_mc_reg_table *table)
   4542 {
   4543 	u8 i, k;
   4544 	u32 tmp;
   4545 	bool patch;
   4546 
   4547 	tmp = RREG32(MC_SEQ_MISC0);
   4548 	patch = ((tmp & 0x0000f00) == 0x300) ? true : false;
   4549 
   4550 	if (patch &&
   4551 	    ((rdev->pdev->device == 0x67B0) ||
   4552 	     (rdev->pdev->device == 0x67B1))) {
   4553 		for (i = 0; i < table->last; i++) {
   4554 			if (table->last >= SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4555 				return -EINVAL;
   4556 			switch(table->mc_reg_address[i].s1 >> 2) {
   4557 			case MC_SEQ_MISC1:
   4558 				for (k = 0; k < table->num_entries; k++) {
   4559 					if ((table->mc_reg_table_entry[k].mclk_max == 125000) ||
   4560 					    (table->mc_reg_table_entry[k].mclk_max == 137500))
   4561 						table->mc_reg_table_entry[k].mc_data[i] =
   4562 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFFFFFF8) |
   4563 							0x00000007;
   4564 				}
   4565 				break;
   4566 			case MC_SEQ_WR_CTL_D0:
   4567 				for (k = 0; k < table->num_entries; k++) {
   4568 					if ((table->mc_reg_table_entry[k].mclk_max == 125000) ||
   4569 					    (table->mc_reg_table_entry[k].mclk_max == 137500))
   4570 						table->mc_reg_table_entry[k].mc_data[i] =
   4571 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFFF0F00) |
   4572 							0x0000D0DD;
   4573 				}
   4574 				break;
   4575 			case MC_SEQ_WR_CTL_D1:
   4576 				for (k = 0; k < table->num_entries; k++) {
   4577 					if ((table->mc_reg_table_entry[k].mclk_max == 125000) ||
   4578 					    (table->mc_reg_table_entry[k].mclk_max == 137500))
   4579 						table->mc_reg_table_entry[k].mc_data[i] =
   4580 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFFF0F00) |
   4581 							0x0000D0DD;
   4582 				}
   4583 				break;
   4584 			case MC_SEQ_WR_CTL_2:
   4585 				for (k = 0; k < table->num_entries; k++) {
   4586 					if ((table->mc_reg_table_entry[k].mclk_max == 125000) ||
   4587 					    (table->mc_reg_table_entry[k].mclk_max == 137500))
   4588 						table->mc_reg_table_entry[k].mc_data[i] = 0;
   4589 				}
   4590 				break;
   4591 			case MC_SEQ_CAS_TIMING:
   4592 				for (k = 0; k < table->num_entries; k++) {
   4593 					if (table->mc_reg_table_entry[k].mclk_max == 125000)
   4594 						table->mc_reg_table_entry[k].mc_data[i] =
   4595 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFE0FE0F) |
   4596 							0x000C0140;
   4597 					else if (table->mc_reg_table_entry[k].mclk_max == 137500)
   4598 						table->mc_reg_table_entry[k].mc_data[i] =
   4599 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFE0FE0F) |
   4600 							0x000C0150;
   4601 				}
   4602 				break;
   4603 			case MC_SEQ_MISC_TIMING:
   4604 				for (k = 0; k < table->num_entries; k++) {
   4605 					if (table->mc_reg_table_entry[k].mclk_max == 125000)
   4606 						table->mc_reg_table_entry[k].mc_data[i] =
   4607 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFFFFFE0) |
   4608 							0x00000030;
   4609 					else if (table->mc_reg_table_entry[k].mclk_max == 137500)
   4610 						table->mc_reg_table_entry[k].mc_data[i] =
   4611 							(table->mc_reg_table_entry[k].mc_data[i] & 0xFFFFFFE0) |
   4612 							0x00000035;
   4613 				}
   4614 				break;
   4615 			default:
   4616 				break;
   4617 			}
   4618 		}
   4619 
   4620 		WREG32(MC_SEQ_IO_DEBUG_INDEX, 3);
   4621 		tmp = RREG32(MC_SEQ_IO_DEBUG_DATA);
   4622 		tmp = (tmp & 0xFFF8FFFF) | (1 << 16);
   4623 		WREG32(MC_SEQ_IO_DEBUG_INDEX, 3);
   4624 		WREG32(MC_SEQ_IO_DEBUG_DATA, tmp);
   4625 	}
   4626 
   4627 	return 0;
   4628 }
   4629 
   4630 static int ci_initialize_mc_reg_table(struct radeon_device *rdev)
   4631 {
   4632 	struct ci_power_info *pi = ci_get_pi(rdev);
   4633 	struct atom_mc_reg_table *table;
   4634 	struct ci_mc_reg_table *ci_table = &pi->mc_reg_table;
   4635 	u8 module_index = rv770_get_memory_module_index(rdev);
   4636 	int ret;
   4637 
   4638 	table = kzalloc(sizeof(struct atom_mc_reg_table), GFP_KERNEL);
   4639 	if (!table)
   4640 		return -ENOMEM;
   4641 
   4642 	WREG32(MC_SEQ_RAS_TIMING_LP, RREG32(MC_SEQ_RAS_TIMING));
   4643 	WREG32(MC_SEQ_CAS_TIMING_LP, RREG32(MC_SEQ_CAS_TIMING));
   4644 	WREG32(MC_SEQ_DLL_STBY_LP, RREG32(MC_SEQ_DLL_STBY));
   4645 	WREG32(MC_SEQ_G5PDX_CMD0_LP, RREG32(MC_SEQ_G5PDX_CMD0));
   4646 	WREG32(MC_SEQ_G5PDX_CMD1_LP, RREG32(MC_SEQ_G5PDX_CMD1));
   4647 	WREG32(MC_SEQ_G5PDX_CTRL_LP, RREG32(MC_SEQ_G5PDX_CTRL));
   4648 	WREG32(MC_SEQ_PMG_DVS_CMD_LP, RREG32(MC_SEQ_PMG_DVS_CMD));
   4649 	WREG32(MC_SEQ_PMG_DVS_CTL_LP, RREG32(MC_SEQ_PMG_DVS_CTL));
   4650 	WREG32(MC_SEQ_MISC_TIMING_LP, RREG32(MC_SEQ_MISC_TIMING));
   4651 	WREG32(MC_SEQ_MISC_TIMING2_LP, RREG32(MC_SEQ_MISC_TIMING2));
   4652 	WREG32(MC_SEQ_PMG_CMD_EMRS_LP, RREG32(MC_PMG_CMD_EMRS));
   4653 	WREG32(MC_SEQ_PMG_CMD_MRS_LP, RREG32(MC_PMG_CMD_MRS));
   4654 	WREG32(MC_SEQ_PMG_CMD_MRS1_LP, RREG32(MC_PMG_CMD_MRS1));
   4655 	WREG32(MC_SEQ_WR_CTL_D0_LP, RREG32(MC_SEQ_WR_CTL_D0));
   4656 	WREG32(MC_SEQ_WR_CTL_D1_LP, RREG32(MC_SEQ_WR_CTL_D1));
   4657 	WREG32(MC_SEQ_RD_CTL_D0_LP, RREG32(MC_SEQ_RD_CTL_D0));
   4658 	WREG32(MC_SEQ_RD_CTL_D1_LP, RREG32(MC_SEQ_RD_CTL_D1));
   4659 	WREG32(MC_SEQ_PMG_TIMING_LP, RREG32(MC_SEQ_PMG_TIMING));
   4660 	WREG32(MC_SEQ_PMG_CMD_MRS2_LP, RREG32(MC_PMG_CMD_MRS2));
   4661 	WREG32(MC_SEQ_WR_CTL_2_LP, RREG32(MC_SEQ_WR_CTL_2));
   4662 
   4663 	ret = radeon_atom_init_mc_reg_table(rdev, module_index, table);
   4664 	if (ret)
   4665 		goto init_mc_done;
   4666 
   4667 	ret = ci_copy_vbios_mc_reg_table(table, ci_table);
   4668 	if (ret)
   4669 		goto init_mc_done;
   4670 
   4671 	ci_set_s0_mc_reg_index(ci_table);
   4672 
   4673 	ret = ci_register_patching_mc_seq(rdev, ci_table);
   4674 	if (ret)
   4675 		goto init_mc_done;
   4676 
   4677 	ret = ci_set_mc_special_registers(rdev, ci_table);
   4678 	if (ret)
   4679 		goto init_mc_done;
   4680 
   4681 	ci_set_valid_flag(ci_table);
   4682 
   4683 init_mc_done:
   4684 	kfree(table);
   4685 
   4686 	return ret;
   4687 }
   4688 
   4689 static int ci_populate_mc_reg_addresses(struct radeon_device *rdev,
   4690 					SMU7_Discrete_MCRegisters *mc_reg_table)
   4691 {
   4692 	struct ci_power_info *pi = ci_get_pi(rdev);
   4693 	u32 i, j;
   4694 
   4695 	for (i = 0, j = 0; j < pi->mc_reg_table.last; j++) {
   4696 		if (pi->mc_reg_table.valid_flag & (1 << j)) {
   4697 			if (i >= SMU7_DISCRETE_MC_REGISTER_ARRAY_SIZE)
   4698 				return -EINVAL;
   4699 			mc_reg_table->address[i].s0 = cpu_to_be16(pi->mc_reg_table.mc_reg_address[j].s0);
   4700 			mc_reg_table->address[i].s1 = cpu_to_be16(pi->mc_reg_table.mc_reg_address[j].s1);
   4701 			i++;
   4702 		}
   4703 	}
   4704 
   4705 	mc_reg_table->last = (u8)i;
   4706 
   4707 	return 0;
   4708 }
   4709 
   4710 static void ci_convert_mc_registers(const struct ci_mc_reg_entry *entry,
   4711 				    SMU7_Discrete_MCRegisterSet *data,
   4712 				    u32 num_entries, u32 valid_flag)
   4713 {
   4714 	u32 i, j;
   4715 
   4716 	for (i = 0, j = 0; j < num_entries; j++) {
   4717 		if (valid_flag & (1 << j)) {
   4718 			data->value[i] = cpu_to_be32(entry->mc_data[j]);
   4719 			i++;
   4720 		}
   4721 	}
   4722 }
   4723 
   4724 static void ci_convert_mc_reg_table_entry_to_smc(struct radeon_device *rdev,
   4725 						 const u32 memory_clock,
   4726 						 SMU7_Discrete_MCRegisterSet *mc_reg_table_data)
   4727 {
   4728 	struct ci_power_info *pi = ci_get_pi(rdev);
   4729 	u32 i = 0;
   4730 
   4731 	for(i = 0; i < pi->mc_reg_table.num_entries; i++) {
   4732 		if (memory_clock <= pi->mc_reg_table.mc_reg_table_entry[i].mclk_max)
   4733 			break;
   4734 	}
   4735 
   4736 	if ((i == pi->mc_reg_table.num_entries) && (i > 0))
   4737 		--i;
   4738 
   4739 	ci_convert_mc_registers(&pi->mc_reg_table.mc_reg_table_entry[i],
   4740 				mc_reg_table_data, pi->mc_reg_table.last,
   4741 				pi->mc_reg_table.valid_flag);
   4742 }
   4743 
   4744 static void ci_convert_mc_reg_table_to_smc(struct radeon_device *rdev,
   4745 					   SMU7_Discrete_MCRegisters *mc_reg_table)
   4746 {
   4747 	struct ci_power_info *pi = ci_get_pi(rdev);
   4748 	u32 i;
   4749 
   4750 	for (i = 0; i < pi->dpm_table.mclk_table.count; i++)
   4751 		ci_convert_mc_reg_table_entry_to_smc(rdev,
   4752 						     pi->dpm_table.mclk_table.dpm_levels[i].value,
   4753 						     &mc_reg_table->data[i]);
   4754 }
   4755 
   4756 static int ci_populate_initial_mc_reg_table(struct radeon_device *rdev)
   4757 {
   4758 	struct ci_power_info *pi = ci_get_pi(rdev);
   4759 	int ret;
   4760 
   4761 	memset(&pi->smc_mc_reg_table, 0, sizeof(SMU7_Discrete_MCRegisters));
   4762 
   4763 	ret = ci_populate_mc_reg_addresses(rdev, &pi->smc_mc_reg_table);
   4764 	if (ret)
   4765 		return ret;
   4766 	ci_convert_mc_reg_table_to_smc(rdev, &pi->smc_mc_reg_table);
   4767 
   4768 	return ci_copy_bytes_to_smc(rdev,
   4769 				    pi->mc_reg_table_start,
   4770 				    (u8 *)&pi->smc_mc_reg_table,
   4771 				    sizeof(SMU7_Discrete_MCRegisters),
   4772 				    pi->sram_end);
   4773 }
   4774 
   4775 static int ci_update_and_upload_mc_reg_table(struct radeon_device *rdev)
   4776 {
   4777 	struct ci_power_info *pi = ci_get_pi(rdev);
   4778 
   4779 	if (!(pi->need_update_smu7_dpm_table & DPMTABLE_OD_UPDATE_MCLK))
   4780 		return 0;
   4781 
   4782 	memset(&pi->smc_mc_reg_table, 0, sizeof(SMU7_Discrete_MCRegisters));
   4783 
   4784 	ci_convert_mc_reg_table_to_smc(rdev, &pi->smc_mc_reg_table);
   4785 
   4786 	return ci_copy_bytes_to_smc(rdev,
   4787 				    pi->mc_reg_table_start +
   4788 				    offsetof(SMU7_Discrete_MCRegisters, data[0]),
   4789 				    (u8 *)&pi->smc_mc_reg_table.data[0],
   4790 				    sizeof(SMU7_Discrete_MCRegisterSet) *
   4791 				    pi->dpm_table.mclk_table.count,
   4792 				    pi->sram_end);
   4793 }
   4794 
   4795 static void ci_enable_voltage_control(struct radeon_device *rdev)
   4796 {
   4797 	u32 tmp = RREG32_SMC(GENERAL_PWRMGT);
   4798 
   4799 	tmp |= VOLT_PWRMGT_EN;
   4800 	WREG32_SMC(GENERAL_PWRMGT, tmp);
   4801 }
   4802 
   4803 static enum radeon_pcie_gen ci_get_maximum_link_speed(struct radeon_device *rdev,
   4804 						      struct radeon_ps *radeon_state)
   4805 {
   4806 	struct ci_ps *state = ci_get_ps(radeon_state);
   4807 	int i;
   4808 	u16 pcie_speed, max_speed = 0;
   4809 
   4810 	for (i = 0; i < state->performance_level_count; i++) {
   4811 		pcie_speed = state->performance_levels[i].pcie_gen;
   4812 		if (max_speed < pcie_speed)
   4813 			max_speed = pcie_speed;
   4814 	}
   4815 
   4816 	return max_speed;
   4817 }
   4818 
   4819 static u16 ci_get_current_pcie_speed(struct radeon_device *rdev)
   4820 {
   4821 	u32 speed_cntl = 0;
   4822 
   4823 	speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL) & LC_CURRENT_DATA_RATE_MASK;
   4824 	speed_cntl >>= LC_CURRENT_DATA_RATE_SHIFT;
   4825 
   4826 	return (u16)speed_cntl;
   4827 }
   4828 
   4829 static int ci_get_current_pcie_lane_number(struct radeon_device *rdev)
   4830 {
   4831 	u32 link_width = 0;
   4832 
   4833 	link_width = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL) & LC_LINK_WIDTH_RD_MASK;
   4834 	link_width >>= LC_LINK_WIDTH_RD_SHIFT;
   4835 
   4836 	switch (link_width) {
   4837 	case RADEON_PCIE_LC_LINK_WIDTH_X1:
   4838 		return 1;
   4839 	case RADEON_PCIE_LC_LINK_WIDTH_X2:
   4840 		return 2;
   4841 	case RADEON_PCIE_LC_LINK_WIDTH_X4:
   4842 		return 4;
   4843 	case RADEON_PCIE_LC_LINK_WIDTH_X8:
   4844 		return 8;
   4845 	case RADEON_PCIE_LC_LINK_WIDTH_X12:
   4846 		/* not actually supported */
   4847 		return 12;
   4848 	case RADEON_PCIE_LC_LINK_WIDTH_X0:
   4849 	case RADEON_PCIE_LC_LINK_WIDTH_X16:
   4850 	default:
   4851 		return 16;
   4852 	}
   4853 }
   4854 
   4855 static void ci_request_link_speed_change_before_state_change(struct radeon_device *rdev,
   4856 							     struct radeon_ps *radeon_new_state,
   4857 							     struct radeon_ps *radeon_current_state)
   4858 {
   4859 	struct ci_power_info *pi = ci_get_pi(rdev);
   4860 	enum radeon_pcie_gen target_link_speed =
   4861 		ci_get_maximum_link_speed(rdev, radeon_new_state);
   4862 	enum radeon_pcie_gen current_link_speed;
   4863 
   4864 	if (pi->force_pcie_gen == RADEON_PCIE_GEN_INVALID)
   4865 		current_link_speed = ci_get_maximum_link_speed(rdev, radeon_current_state);
   4866 	else
   4867 		current_link_speed = pi->force_pcie_gen;
   4868 
   4869 	pi->force_pcie_gen = RADEON_PCIE_GEN_INVALID;
   4870 	pi->pspp_notify_required = false;
   4871 	if (target_link_speed > current_link_speed) {
   4872 		switch (target_link_speed) {
   4873 #ifdef CONFIG_ACPI
   4874 		case RADEON_PCIE_GEN3:
   4875 			if (radeon_acpi_pcie_performance_request(rdev, PCIE_PERF_REQ_PECI_GEN3, false) == 0)
   4876 				break;
   4877 			pi->force_pcie_gen = RADEON_PCIE_GEN2;
   4878 			if (current_link_speed == RADEON_PCIE_GEN2)
   4879 				break;
   4880 			/* fall through */
   4881 		case RADEON_PCIE_GEN2:
   4882 			if (radeon_acpi_pcie_performance_request(rdev, PCIE_PERF_REQ_PECI_GEN2, false) == 0)
   4883 				break;
   4884 #endif
   4885 			/* fall through */
   4886 		default:
   4887 			pi->force_pcie_gen = ci_get_current_pcie_speed(rdev);
   4888 			break;
   4889 		}
   4890 	} else {
   4891 		if (target_link_speed < current_link_speed)
   4892 			pi->pspp_notify_required = true;
   4893 	}
   4894 }
   4895 
   4896 static void ci_notify_link_speed_change_after_state_change(struct radeon_device *rdev,
   4897 							   struct radeon_ps *radeon_new_state,
   4898 							   struct radeon_ps *radeon_current_state)
   4899 {
   4900 	struct ci_power_info *pi = ci_get_pi(rdev);
   4901 	enum radeon_pcie_gen target_link_speed =
   4902 		ci_get_maximum_link_speed(rdev, radeon_new_state);
   4903 	u8 request;
   4904 
   4905 	if (pi->pspp_notify_required) {
   4906 		if (target_link_speed == RADEON_PCIE_GEN3)
   4907 			request = PCIE_PERF_REQ_PECI_GEN3;
   4908 		else if (target_link_speed == RADEON_PCIE_GEN2)
   4909 			request = PCIE_PERF_REQ_PECI_GEN2;
   4910 		else
   4911 			request = PCIE_PERF_REQ_PECI_GEN1;
   4912 
   4913 		if ((request == PCIE_PERF_REQ_PECI_GEN1) &&
   4914 		    (ci_get_current_pcie_speed(rdev) > 0))
   4915 			return;
   4916 
   4917 #ifdef CONFIG_ACPI
   4918 		radeon_acpi_pcie_performance_request(rdev, request, false);
   4919 #endif
   4920 	}
   4921 }
   4922 
   4923 static int ci_set_private_data_variables_based_on_pptable(struct radeon_device *rdev)
   4924 {
   4925 	struct ci_power_info *pi = ci_get_pi(rdev);
   4926 	struct radeon_clock_voltage_dependency_table *allowed_sclk_vddc_table =
   4927 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk;
   4928 	struct radeon_clock_voltage_dependency_table *allowed_mclk_vddc_table =
   4929 		&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk;
   4930 	struct radeon_clock_voltage_dependency_table *allowed_mclk_vddci_table =
   4931 		&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk;
   4932 
   4933 	if (allowed_sclk_vddc_table == NULL)
   4934 		return -EINVAL;
   4935 	if (allowed_sclk_vddc_table->count < 1)
   4936 		return -EINVAL;
   4937 	if (allowed_mclk_vddc_table == NULL)
   4938 		return -EINVAL;
   4939 	if (allowed_mclk_vddc_table->count < 1)
   4940 		return -EINVAL;
   4941 	if (allowed_mclk_vddci_table == NULL)
   4942 		return -EINVAL;
   4943 	if (allowed_mclk_vddci_table->count < 1)
   4944 		return -EINVAL;
   4945 
   4946 	pi->min_vddc_in_pp_table = allowed_sclk_vddc_table->entries[0].v;
   4947 	pi->max_vddc_in_pp_table =
   4948 		allowed_sclk_vddc_table->entries[allowed_sclk_vddc_table->count - 1].v;
   4949 
   4950 	pi->min_vddci_in_pp_table = allowed_mclk_vddci_table->entries[0].v;
   4951 	pi->max_vddci_in_pp_table =
   4952 		allowed_mclk_vddci_table->entries[allowed_mclk_vddci_table->count - 1].v;
   4953 
   4954 	rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac.sclk =
   4955 		allowed_sclk_vddc_table->entries[allowed_sclk_vddc_table->count - 1].clk;
   4956 	rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac.mclk =
   4957 		allowed_mclk_vddc_table->entries[allowed_sclk_vddc_table->count - 1].clk;
   4958 	rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac.vddc =
   4959 		allowed_sclk_vddc_table->entries[allowed_sclk_vddc_table->count - 1].v;
   4960 	rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac.vddci =
   4961 		allowed_mclk_vddci_table->entries[allowed_mclk_vddci_table->count - 1].v;
   4962 
   4963 	return 0;
   4964 }
   4965 
   4966 static void ci_patch_with_vddc_leakage(struct radeon_device *rdev, u16 *vddc)
   4967 {
   4968 	struct ci_power_info *pi = ci_get_pi(rdev);
   4969 	struct ci_leakage_voltage *leakage_table = &pi->vddc_leakage;
   4970 	u32 leakage_index;
   4971 
   4972 	for (leakage_index = 0; leakage_index < leakage_table->count; leakage_index++) {
   4973 		if (leakage_table->leakage_id[leakage_index] == *vddc) {
   4974 			*vddc = leakage_table->actual_voltage[leakage_index];
   4975 			break;
   4976 		}
   4977 	}
   4978 }
   4979 
   4980 static void ci_patch_with_vddci_leakage(struct radeon_device *rdev, u16 *vddci)
   4981 {
   4982 	struct ci_power_info *pi = ci_get_pi(rdev);
   4983 	struct ci_leakage_voltage *leakage_table = &pi->vddci_leakage;
   4984 	u32 leakage_index;
   4985 
   4986 	for (leakage_index = 0; leakage_index < leakage_table->count; leakage_index++) {
   4987 		if (leakage_table->leakage_id[leakage_index] == *vddci) {
   4988 			*vddci = leakage_table->actual_voltage[leakage_index];
   4989 			break;
   4990 		}
   4991 	}
   4992 }
   4993 
   4994 static void ci_patch_clock_voltage_dependency_table_with_vddc_leakage(struct radeon_device *rdev,
   4995 								      struct radeon_clock_voltage_dependency_table *table)
   4996 {
   4997 	u32 i;
   4998 
   4999 	if (table) {
   5000 		for (i = 0; i < table->count; i++)
   5001 			ci_patch_with_vddc_leakage(rdev, &table->entries[i].v);
   5002 	}
   5003 }
   5004 
   5005 static void ci_patch_clock_voltage_dependency_table_with_vddci_leakage(struct radeon_device *rdev,
   5006 								       struct radeon_clock_voltage_dependency_table *table)
   5007 {
   5008 	u32 i;
   5009 
   5010 	if (table) {
   5011 		for (i = 0; i < table->count; i++)
   5012 			ci_patch_with_vddci_leakage(rdev, &table->entries[i].v);
   5013 	}
   5014 }
   5015 
   5016 static void ci_patch_vce_clock_voltage_dependency_table_with_vddc_leakage(struct radeon_device *rdev,
   5017 									  struct radeon_vce_clock_voltage_dependency_table *table)
   5018 {
   5019 	u32 i;
   5020 
   5021 	if (table) {
   5022 		for (i = 0; i < table->count; i++)
   5023 			ci_patch_with_vddc_leakage(rdev, &table->entries[i].v);
   5024 	}
   5025 }
   5026 
   5027 static void ci_patch_uvd_clock_voltage_dependency_table_with_vddc_leakage(struct radeon_device *rdev,
   5028 									  struct radeon_uvd_clock_voltage_dependency_table *table)
   5029 {
   5030 	u32 i;
   5031 
   5032 	if (table) {
   5033 		for (i = 0; i < table->count; i++)
   5034 			ci_patch_with_vddc_leakage(rdev, &table->entries[i].v);
   5035 	}
   5036 }
   5037 
   5038 static void ci_patch_vddc_phase_shed_limit_table_with_vddc_leakage(struct radeon_device *rdev,
   5039 								   struct radeon_phase_shedding_limits_table *table)
   5040 {
   5041 	u32 i;
   5042 
   5043 	if (table) {
   5044 		for (i = 0; i < table->count; i++)
   5045 			ci_patch_with_vddc_leakage(rdev, &table->entries[i].voltage);
   5046 	}
   5047 }
   5048 
   5049 static void ci_patch_clock_voltage_limits_with_vddc_leakage(struct radeon_device *rdev,
   5050 							    struct radeon_clock_and_voltage_limits *table)
   5051 {
   5052 	if (table) {
   5053 		ci_patch_with_vddc_leakage(rdev, (u16 *)&table->vddc);
   5054 		ci_patch_with_vddci_leakage(rdev, (u16 *)&table->vddci);
   5055 	}
   5056 }
   5057 
   5058 static void ci_patch_cac_leakage_table_with_vddc_leakage(struct radeon_device *rdev,
   5059 							 struct radeon_cac_leakage_table *table)
   5060 {
   5061 	u32 i;
   5062 
   5063 	if (table) {
   5064 		for (i = 0; i < table->count; i++)
   5065 			ci_patch_with_vddc_leakage(rdev, &table->entries[i].vddc);
   5066 	}
   5067 }
   5068 
   5069 static void ci_patch_dependency_tables_with_leakage(struct radeon_device *rdev)
   5070 {
   5071 
   5072 	ci_patch_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5073 								  &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk);
   5074 	ci_patch_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5075 								  &rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk);
   5076 	ci_patch_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5077 								  &rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk);
   5078 	ci_patch_clock_voltage_dependency_table_with_vddci_leakage(rdev,
   5079 								   &rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk);
   5080 	ci_patch_vce_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5081 								      &rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table);
   5082 	ci_patch_uvd_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5083 								      &rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table);
   5084 	ci_patch_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5085 								  &rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table);
   5086 	ci_patch_clock_voltage_dependency_table_with_vddc_leakage(rdev,
   5087 								  &rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table);
   5088 	ci_patch_vddc_phase_shed_limit_table_with_vddc_leakage(rdev,
   5089 							       &rdev->pm.dpm.dyn_state.phase_shedding_limits_table);
   5090 	ci_patch_clock_voltage_limits_with_vddc_leakage(rdev,
   5091 							&rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac);
   5092 	ci_patch_clock_voltage_limits_with_vddc_leakage(rdev,
   5093 							&rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc);
   5094 	ci_patch_cac_leakage_table_with_vddc_leakage(rdev,
   5095 						     &rdev->pm.dpm.dyn_state.cac_leakage_table);
   5096 
   5097 }
   5098 
   5099 static void ci_get_memory_type(struct radeon_device *rdev)
   5100 {
   5101 	struct ci_power_info *pi = ci_get_pi(rdev);
   5102 	u32 tmp;
   5103 
   5104 	tmp = RREG32(MC_SEQ_MISC0);
   5105 
   5106 	if (((tmp & MC_SEQ_MISC0_GDDR5_MASK) >> MC_SEQ_MISC0_GDDR5_SHIFT) ==
   5107 	    MC_SEQ_MISC0_GDDR5_VALUE)
   5108 		pi->mem_gddr5 = true;
   5109 	else
   5110 		pi->mem_gddr5 = false;
   5111 
   5112 }
   5113 
   5114 static void ci_update_current_ps(struct radeon_device *rdev,
   5115 				 struct radeon_ps *rps)
   5116 {
   5117 	struct ci_ps *new_ps = ci_get_ps(rps);
   5118 	struct ci_power_info *pi = ci_get_pi(rdev);
   5119 
   5120 	pi->current_rps = *rps;
   5121 	pi->current_ps = *new_ps;
   5122 	pi->current_rps.ps_priv = &pi->current_ps;
   5123 }
   5124 
   5125 static void ci_update_requested_ps(struct radeon_device *rdev,
   5126 				   struct radeon_ps *rps)
   5127 {
   5128 	struct ci_ps *new_ps = ci_get_ps(rps);
   5129 	struct ci_power_info *pi = ci_get_pi(rdev);
   5130 
   5131 	pi->requested_rps = *rps;
   5132 	pi->requested_ps = *new_ps;
   5133 	pi->requested_rps.ps_priv = &pi->requested_ps;
   5134 }
   5135 
   5136 int ci_dpm_pre_set_power_state(struct radeon_device *rdev)
   5137 {
   5138 	struct ci_power_info *pi = ci_get_pi(rdev);
   5139 	struct radeon_ps requested_ps = *rdev->pm.dpm.requested_ps;
   5140 	struct radeon_ps *new_ps = &requested_ps;
   5141 
   5142 	ci_update_requested_ps(rdev, new_ps);
   5143 
   5144 	ci_apply_state_adjust_rules(rdev, &pi->requested_rps);
   5145 
   5146 	return 0;
   5147 }
   5148 
   5149 void ci_dpm_post_set_power_state(struct radeon_device *rdev)
   5150 {
   5151 	struct ci_power_info *pi = ci_get_pi(rdev);
   5152 	struct radeon_ps *new_ps = &pi->requested_rps;
   5153 
   5154 	ci_update_current_ps(rdev, new_ps);
   5155 }
   5156 
   5157 
   5158 void ci_dpm_setup_asic(struct radeon_device *rdev)
   5159 {
   5160 	int r;
   5161 
   5162 	r = ci_mc_load_microcode(rdev);
   5163 	if (r)
   5164 		DRM_ERROR("Failed to load MC firmware!\n");
   5165 	ci_read_clock_registers(rdev);
   5166 	ci_get_memory_type(rdev);
   5167 	ci_enable_acpi_power_management(rdev);
   5168 	ci_init_sclk_t(rdev);
   5169 }
   5170 
   5171 int ci_dpm_enable(struct radeon_device *rdev)
   5172 {
   5173 	struct ci_power_info *pi = ci_get_pi(rdev);
   5174 	struct radeon_ps *boot_ps = rdev->pm.dpm.boot_ps;
   5175 	int ret;
   5176 
   5177 	if (ci_is_smc_running(rdev))
   5178 		return -EINVAL;
   5179 	if (pi->voltage_control != CISLANDS_VOLTAGE_CONTROL_NONE) {
   5180 		ci_enable_voltage_control(rdev);
   5181 		ret = ci_construct_voltage_tables(rdev);
   5182 		if (ret) {
   5183 			DRM_ERROR("ci_construct_voltage_tables failed\n");
   5184 			return ret;
   5185 		}
   5186 	}
   5187 	if (pi->caps_dynamic_ac_timing) {
   5188 		ret = ci_initialize_mc_reg_table(rdev);
   5189 		if (ret)
   5190 			pi->caps_dynamic_ac_timing = false;
   5191 	}
   5192 	if (pi->dynamic_ss)
   5193 		ci_enable_spread_spectrum(rdev, true);
   5194 	if (pi->thermal_protection)
   5195 		ci_enable_thermal_protection(rdev, true);
   5196 	ci_program_sstp(rdev);
   5197 	ci_enable_display_gap(rdev);
   5198 	ci_program_vc(rdev);
   5199 	ret = ci_upload_firmware(rdev);
   5200 	if (ret) {
   5201 		DRM_ERROR("ci_upload_firmware failed\n");
   5202 		return ret;
   5203 	}
   5204 	ret = ci_process_firmware_header(rdev);
   5205 	if (ret) {
   5206 		DRM_ERROR("ci_process_firmware_header failed\n");
   5207 		return ret;
   5208 	}
   5209 	ret = ci_initial_switch_from_arb_f0_to_f1(rdev);
   5210 	if (ret) {
   5211 		DRM_ERROR("ci_initial_switch_from_arb_f0_to_f1 failed\n");
   5212 		return ret;
   5213 	}
   5214 	ret = ci_init_smc_table(rdev);
   5215 	if (ret) {
   5216 		DRM_ERROR("ci_init_smc_table failed\n");
   5217 		return ret;
   5218 	}
   5219 	ret = ci_init_arb_table_index(rdev);
   5220 	if (ret) {
   5221 		DRM_ERROR("ci_init_arb_table_index failed\n");
   5222 		return ret;
   5223 	}
   5224 	if (pi->caps_dynamic_ac_timing) {
   5225 		ret = ci_populate_initial_mc_reg_table(rdev);
   5226 		if (ret) {
   5227 			DRM_ERROR("ci_populate_initial_mc_reg_table failed\n");
   5228 			return ret;
   5229 		}
   5230 	}
   5231 	ret = ci_populate_pm_base(rdev);
   5232 	if (ret) {
   5233 		DRM_ERROR("ci_populate_pm_base failed\n");
   5234 		return ret;
   5235 	}
   5236 	ci_dpm_start_smc(rdev);
   5237 	ci_enable_vr_hot_gpio_interrupt(rdev);
   5238 	ret = ci_notify_smc_display_change(rdev, false);
   5239 	if (ret) {
   5240 		DRM_ERROR("ci_notify_smc_display_change failed\n");
   5241 		return ret;
   5242 	}
   5243 	ci_enable_sclk_control(rdev, true);
   5244 	ret = ci_enable_ulv(rdev, true);
   5245 	if (ret) {
   5246 		DRM_ERROR("ci_enable_ulv failed\n");
   5247 		return ret;
   5248 	}
   5249 	ret = ci_enable_ds_master_switch(rdev, true);
   5250 	if (ret) {
   5251 		DRM_ERROR("ci_enable_ds_master_switch failed\n");
   5252 		return ret;
   5253 	}
   5254 	ret = ci_start_dpm(rdev);
   5255 	if (ret) {
   5256 		DRM_ERROR("ci_start_dpm failed\n");
   5257 		return ret;
   5258 	}
   5259 	ret = ci_enable_didt(rdev, true);
   5260 	if (ret) {
   5261 		DRM_ERROR("ci_enable_didt failed\n");
   5262 		return ret;
   5263 	}
   5264 	ret = ci_enable_smc_cac(rdev, true);
   5265 	if (ret) {
   5266 		DRM_ERROR("ci_enable_smc_cac failed\n");
   5267 		return ret;
   5268 	}
   5269 	ret = ci_enable_power_containment(rdev, true);
   5270 	if (ret) {
   5271 		DRM_ERROR("ci_enable_power_containment failed\n");
   5272 		return ret;
   5273 	}
   5274 
   5275 	ret = ci_power_control_set_level(rdev);
   5276 	if (ret) {
   5277 		DRM_ERROR("ci_power_control_set_level failed\n");
   5278 		return ret;
   5279 	}
   5280 
   5281 	ci_enable_auto_throttle_source(rdev, RADEON_DPM_AUTO_THROTTLE_SRC_THERMAL, true);
   5282 
   5283 	ret = ci_enable_thermal_based_sclk_dpm(rdev, true);
   5284 	if (ret) {
   5285 		DRM_ERROR("ci_enable_thermal_based_sclk_dpm failed\n");
   5286 		return ret;
   5287 	}
   5288 
   5289 	ci_thermal_start_thermal_controller(rdev);
   5290 
   5291 	ci_update_current_ps(rdev, boot_ps);
   5292 
   5293 	return 0;
   5294 }
   5295 
   5296 static int ci_set_temperature_range(struct radeon_device *rdev)
   5297 {
   5298 	int ret;
   5299 
   5300 	ret = ci_thermal_enable_alert(rdev, false);
   5301 	if (ret)
   5302 		return ret;
   5303 	ret = ci_thermal_set_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX);
   5304 	if (ret)
   5305 		return ret;
   5306 	ret = ci_thermal_enable_alert(rdev, true);
   5307 	if (ret)
   5308 		return ret;
   5309 
   5310 	return ret;
   5311 }
   5312 
   5313 int ci_dpm_late_enable(struct radeon_device *rdev)
   5314 {
   5315 	int ret;
   5316 
   5317 	ret = ci_set_temperature_range(rdev);
   5318 	if (ret)
   5319 		return ret;
   5320 
   5321 	ci_dpm_powergate_uvd(rdev, true);
   5322 
   5323 	return 0;
   5324 }
   5325 
   5326 void ci_dpm_disable(struct radeon_device *rdev)
   5327 {
   5328 	struct ci_power_info *pi = ci_get_pi(rdev);
   5329 	struct radeon_ps *boot_ps = rdev->pm.dpm.boot_ps;
   5330 
   5331 	ci_dpm_powergate_uvd(rdev, false);
   5332 
   5333 	if (!ci_is_smc_running(rdev))
   5334 		return;
   5335 
   5336 	ci_thermal_stop_thermal_controller(rdev);
   5337 
   5338 	if (pi->thermal_protection)
   5339 		ci_enable_thermal_protection(rdev, false);
   5340 	ci_enable_power_containment(rdev, false);
   5341 	ci_enable_smc_cac(rdev, false);
   5342 	ci_enable_didt(rdev, false);
   5343 	ci_enable_spread_spectrum(rdev, false);
   5344 	ci_enable_auto_throttle_source(rdev, RADEON_DPM_AUTO_THROTTLE_SRC_THERMAL, false);
   5345 	ci_stop_dpm(rdev);
   5346 	ci_enable_ds_master_switch(rdev, false);
   5347 	ci_enable_ulv(rdev, false);
   5348 	ci_clear_vc(rdev);
   5349 	ci_reset_to_default(rdev);
   5350 	ci_dpm_stop_smc(rdev);
   5351 	ci_force_switch_to_arb_f0(rdev);
   5352 	ci_enable_thermal_based_sclk_dpm(rdev, false);
   5353 
   5354 	ci_update_current_ps(rdev, boot_ps);
   5355 }
   5356 
   5357 int ci_dpm_set_power_state(struct radeon_device *rdev)
   5358 {
   5359 	struct ci_power_info *pi = ci_get_pi(rdev);
   5360 	struct radeon_ps *new_ps = &pi->requested_rps;
   5361 	struct radeon_ps *old_ps = &pi->current_rps;
   5362 	int ret;
   5363 
   5364 	ci_find_dpm_states_clocks_in_dpm_table(rdev, new_ps);
   5365 	if (pi->pcie_performance_request)
   5366 		ci_request_link_speed_change_before_state_change(rdev, new_ps, old_ps);
   5367 	ret = ci_freeze_sclk_mclk_dpm(rdev);
   5368 	if (ret) {
   5369 		DRM_ERROR("ci_freeze_sclk_mclk_dpm failed\n");
   5370 		return ret;
   5371 	}
   5372 	ret = ci_populate_and_upload_sclk_mclk_dpm_levels(rdev, new_ps);
   5373 	if (ret) {
   5374 		DRM_ERROR("ci_populate_and_upload_sclk_mclk_dpm_levels failed\n");
   5375 		return ret;
   5376 	}
   5377 	ret = ci_generate_dpm_level_enable_mask(rdev, new_ps);
   5378 	if (ret) {
   5379 		DRM_ERROR("ci_generate_dpm_level_enable_mask failed\n");
   5380 		return ret;
   5381 	}
   5382 
   5383 	ret = ci_update_vce_dpm(rdev, new_ps, old_ps);
   5384 	if (ret) {
   5385 		DRM_ERROR("ci_update_vce_dpm failed\n");
   5386 		return ret;
   5387 	}
   5388 
   5389 	ret = ci_update_sclk_t(rdev);
   5390 	if (ret) {
   5391 		DRM_ERROR("ci_update_sclk_t failed\n");
   5392 		return ret;
   5393 	}
   5394 	if (pi->caps_dynamic_ac_timing) {
   5395 		ret = ci_update_and_upload_mc_reg_table(rdev);
   5396 		if (ret) {
   5397 			DRM_ERROR("ci_update_and_upload_mc_reg_table failed\n");
   5398 			return ret;
   5399 		}
   5400 	}
   5401 	ret = ci_program_memory_timing_parameters(rdev);
   5402 	if (ret) {
   5403 		DRM_ERROR("ci_program_memory_timing_parameters failed\n");
   5404 		return ret;
   5405 	}
   5406 	ret = ci_unfreeze_sclk_mclk_dpm(rdev);
   5407 	if (ret) {
   5408 		DRM_ERROR("ci_unfreeze_sclk_mclk_dpm failed\n");
   5409 		return ret;
   5410 	}
   5411 	ret = ci_upload_dpm_level_enable_mask(rdev);
   5412 	if (ret) {
   5413 		DRM_ERROR("ci_upload_dpm_level_enable_mask failed\n");
   5414 		return ret;
   5415 	}
   5416 	if (pi->pcie_performance_request)
   5417 		ci_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps);
   5418 
   5419 	return 0;
   5420 }
   5421 
   5422 #if 0
   5423 void ci_dpm_reset_asic(struct radeon_device *rdev)
   5424 {
   5425 	ci_set_boot_state(rdev);
   5426 }
   5427 #endif
   5428 
   5429 void ci_dpm_display_configuration_changed(struct radeon_device *rdev)
   5430 {
   5431 	ci_program_display_gap(rdev);
   5432 }
   5433 
   5434 union power_info {
   5435 	struct _ATOM_POWERPLAY_INFO info;
   5436 	struct _ATOM_POWERPLAY_INFO_V2 info_2;
   5437 	struct _ATOM_POWERPLAY_INFO_V3 info_3;
   5438 	struct _ATOM_PPLIB_POWERPLAYTABLE pplib;
   5439 	struct _ATOM_PPLIB_POWERPLAYTABLE2 pplib2;
   5440 	struct _ATOM_PPLIB_POWERPLAYTABLE3 pplib3;
   5441 };
   5442 
   5443 union pplib_clock_info {
   5444 	struct _ATOM_PPLIB_R600_CLOCK_INFO r600;
   5445 	struct _ATOM_PPLIB_RS780_CLOCK_INFO rs780;
   5446 	struct _ATOM_PPLIB_EVERGREEN_CLOCK_INFO evergreen;
   5447 	struct _ATOM_PPLIB_SUMO_CLOCK_INFO sumo;
   5448 	struct _ATOM_PPLIB_SI_CLOCK_INFO si;
   5449 	struct _ATOM_PPLIB_CI_CLOCK_INFO ci;
   5450 };
   5451 
   5452 union pplib_power_state {
   5453 	struct _ATOM_PPLIB_STATE v1;
   5454 	struct _ATOM_PPLIB_STATE_V2 v2;
   5455 };
   5456 
   5457 static void ci_parse_pplib_non_clock_info(struct radeon_device *rdev,
   5458 					  struct radeon_ps *rps,
   5459 					  struct _ATOM_PPLIB_NONCLOCK_INFO *non_clock_info,
   5460 					  u8 table_rev)
   5461 {
   5462 	rps->caps = le32_to_cpu(non_clock_info->ulCapsAndSettings);
   5463 	rps->class = le16_to_cpu(non_clock_info->usClassification);
   5464 	rps->class2 = le16_to_cpu(non_clock_info->usClassification2);
   5465 
   5466 	if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) {
   5467 		rps->vclk = le32_to_cpu(non_clock_info->ulVCLK);
   5468 		rps->dclk = le32_to_cpu(non_clock_info->ulDCLK);
   5469 	} else {
   5470 		rps->vclk = 0;
   5471 		rps->dclk = 0;
   5472 	}
   5473 
   5474 	if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT)
   5475 		rdev->pm.dpm.boot_ps = rps;
   5476 	if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE)
   5477 		rdev->pm.dpm.uvd_ps = rps;
   5478 }
   5479 
   5480 static void ci_parse_pplib_clock_info(struct radeon_device *rdev,
   5481 				      struct radeon_ps *rps, int index,
   5482 				      union pplib_clock_info *clock_info)
   5483 {
   5484 	struct ci_power_info *pi = ci_get_pi(rdev);
   5485 	struct ci_ps *ps = ci_get_ps(rps);
   5486 	struct ci_pl *pl = &ps->performance_levels[index];
   5487 
   5488 	ps->performance_level_count = index + 1;
   5489 
   5490 	pl->sclk = le16_to_cpu(clock_info->ci.usEngineClockLow);
   5491 	pl->sclk |= clock_info->ci.ucEngineClockHigh << 16;
   5492 	pl->mclk = le16_to_cpu(clock_info->ci.usMemoryClockLow);
   5493 	pl->mclk |= clock_info->ci.ucMemoryClockHigh << 16;
   5494 
   5495 	pl->pcie_gen = r600_get_pcie_gen_support(rdev,
   5496 						 pi->sys_pcie_mask,
   5497 						 pi->vbios_boot_state.pcie_gen_bootup_value,
   5498 						 clock_info->ci.ucPCIEGen);
   5499 	pl->pcie_lane = r600_get_pcie_lane_support(rdev,
   5500 						   pi->vbios_boot_state.pcie_lane_bootup_value,
   5501 						   le16_to_cpu(clock_info->ci.usPCIELane));
   5502 
   5503 	if (rps->class & ATOM_PPLIB_CLASSIFICATION_ACPI) {
   5504 		pi->acpi_pcie_gen = pl->pcie_gen;
   5505 	}
   5506 
   5507 	if (rps->class2 & ATOM_PPLIB_CLASSIFICATION2_ULV) {
   5508 		pi->ulv.supported = true;
   5509 		pi->ulv.pl = *pl;
   5510 		pi->ulv.cg_ulv_parameter = CISLANDS_CGULVPARAMETER_DFLT;
   5511 	}
   5512 
   5513 	/* patch up boot state */
   5514 	if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) {
   5515 		pl->mclk = pi->vbios_boot_state.mclk_bootup_value;
   5516 		pl->sclk = pi->vbios_boot_state.sclk_bootup_value;
   5517 		pl->pcie_gen = pi->vbios_boot_state.pcie_gen_bootup_value;
   5518 		pl->pcie_lane = pi->vbios_boot_state.pcie_lane_bootup_value;
   5519 	}
   5520 
   5521 	switch (rps->class & ATOM_PPLIB_CLASSIFICATION_UI_MASK) {
   5522 	case ATOM_PPLIB_CLASSIFICATION_UI_BATTERY:
   5523 		pi->use_pcie_powersaving_levels = true;
   5524 		if (pi->pcie_gen_powersaving.max < pl->pcie_gen)
   5525 			pi->pcie_gen_powersaving.max = pl->pcie_gen;
   5526 		if (pi->pcie_gen_powersaving.min > pl->pcie_gen)
   5527 			pi->pcie_gen_powersaving.min = pl->pcie_gen;
   5528 		if (pi->pcie_lane_powersaving.max < pl->pcie_lane)
   5529 			pi->pcie_lane_powersaving.max = pl->pcie_lane;
   5530 		if (pi->pcie_lane_powersaving.min > pl->pcie_lane)
   5531 			pi->pcie_lane_powersaving.min = pl->pcie_lane;
   5532 		break;
   5533 	case ATOM_PPLIB_CLASSIFICATION_UI_PERFORMANCE:
   5534 		pi->use_pcie_performance_levels = true;
   5535 		if (pi->pcie_gen_performance.max < pl->pcie_gen)
   5536 			pi->pcie_gen_performance.max = pl->pcie_gen;
   5537 		if (pi->pcie_gen_performance.min > pl->pcie_gen)
   5538 			pi->pcie_gen_performance.min = pl->pcie_gen;
   5539 		if (pi->pcie_lane_performance.max < pl->pcie_lane)
   5540 			pi->pcie_lane_performance.max = pl->pcie_lane;
   5541 		if (pi->pcie_lane_performance.min > pl->pcie_lane)
   5542 			pi->pcie_lane_performance.min = pl->pcie_lane;
   5543 		break;
   5544 	default:
   5545 		break;
   5546 	}
   5547 }
   5548 
   5549 static int ci_parse_power_table(struct radeon_device *rdev)
   5550 {
   5551 	struct radeon_mode_info *mode_info = &rdev->mode_info;
   5552 	struct _ATOM_PPLIB_NONCLOCK_INFO *non_clock_info;
   5553 	union pplib_power_state *power_state;
   5554 	int i, j, k, non_clock_array_index, clock_array_index;
   5555 	union pplib_clock_info *clock_info;
   5556 	struct _StateArray *state_array;
   5557 	struct _ClockInfoArray *clock_info_array;
   5558 	struct _NonClockInfoArray *non_clock_info_array;
   5559 	union power_info *power_info;
   5560 	int index = GetIndexIntoMasterTable(DATA, PowerPlayInfo);
   5561 	u16 data_offset;
   5562 	u8 frev, crev;
   5563 	u8 *power_state_offset;
   5564 	struct ci_ps *ps;
   5565 
   5566 	if (!atom_parse_data_header(mode_info->atom_context, index, NULL,
   5567 				   &frev, &crev, &data_offset))
   5568 		return -EINVAL;
   5569 	power_info = (union power_info *)(mode_info->atom_context->bios + data_offset);
   5570 
   5571 	state_array = (struct _StateArray *)
   5572 		(mode_info->atom_context->bios + data_offset +
   5573 		 le16_to_cpu(power_info->pplib.usStateArrayOffset));
   5574 	clock_info_array = (struct _ClockInfoArray *)
   5575 		(mode_info->atom_context->bios + data_offset +
   5576 		 le16_to_cpu(power_info->pplib.usClockInfoArrayOffset));
   5577 	non_clock_info_array = (struct _NonClockInfoArray *)
   5578 		(mode_info->atom_context->bios + data_offset +
   5579 		 le16_to_cpu(power_info->pplib.usNonClockInfoArrayOffset));
   5580 
   5581 	rdev->pm.dpm.ps = kcalloc(state_array->ucNumEntries,
   5582 				  sizeof(struct radeon_ps),
   5583 				  GFP_KERNEL);
   5584 	if (!rdev->pm.dpm.ps)
   5585 		return -ENOMEM;
   5586 	power_state_offset = (u8 *)state_array->states;
   5587 	for (i = 0; i < state_array->ucNumEntries; i++) {
   5588 		u8 *idx;
   5589 		power_state = (union pplib_power_state *)power_state_offset;
   5590 		non_clock_array_index = power_state->v2.nonClockInfoIndex;
   5591 		non_clock_info = (struct _ATOM_PPLIB_NONCLOCK_INFO *)
   5592 			&non_clock_info_array->nonClockInfo[non_clock_array_index];
   5593 		if (!rdev->pm.power_state[i].clock_info)
   5594 			return -EINVAL;
   5595 		ps = kzalloc(sizeof(struct ci_ps), GFP_KERNEL);
   5596 		if (ps == NULL) {
   5597 			kfree(rdev->pm.dpm.ps);
   5598 			return -ENOMEM;
   5599 		}
   5600 		rdev->pm.dpm.ps[i].ps_priv = ps;
   5601 		ci_parse_pplib_non_clock_info(rdev, &rdev->pm.dpm.ps[i],
   5602 					      non_clock_info,
   5603 					      non_clock_info_array->ucEntrySize);
   5604 		k = 0;
   5605 		idx = (u8 *)&power_state->v2.clockInfoIndex[0];
   5606 		for (j = 0; j < power_state->v2.ucNumDPMLevels; j++) {
   5607 			clock_array_index = idx[j];
   5608 			if (clock_array_index >= clock_info_array->ucNumEntries)
   5609 				continue;
   5610 			if (k >= CISLANDS_MAX_HARDWARE_POWERLEVELS)
   5611 				break;
   5612 			clock_info = (union pplib_clock_info *)
   5613 				((u8 *)&clock_info_array->clockInfo[0] +
   5614 				 (clock_array_index * clock_info_array->ucEntrySize));
   5615 			ci_parse_pplib_clock_info(rdev,
   5616 						  &rdev->pm.dpm.ps[i], k,
   5617 						  clock_info);
   5618 			k++;
   5619 		}
   5620 		power_state_offset += 2 + power_state->v2.ucNumDPMLevels;
   5621 	}
   5622 	rdev->pm.dpm.num_ps = state_array->ucNumEntries;
   5623 
   5624 	/* fill in the vce power states */
   5625 	for (i = 0; i < RADEON_MAX_VCE_LEVELS; i++) {
   5626 		u32 sclk, mclk;
   5627 		clock_array_index = rdev->pm.dpm.vce_states[i].clk_idx;
   5628 		clock_info = (union pplib_clock_info *)
   5629 			&clock_info_array->clockInfo[clock_array_index * clock_info_array->ucEntrySize];
   5630 		sclk = le16_to_cpu(clock_info->ci.usEngineClockLow);
   5631 		sclk |= clock_info->ci.ucEngineClockHigh << 16;
   5632 		mclk = le16_to_cpu(clock_info->ci.usMemoryClockLow);
   5633 		mclk |= clock_info->ci.ucMemoryClockHigh << 16;
   5634 		rdev->pm.dpm.vce_states[i].sclk = sclk;
   5635 		rdev->pm.dpm.vce_states[i].mclk = mclk;
   5636 	}
   5637 
   5638 	return 0;
   5639 }
   5640 
   5641 static int ci_get_vbios_boot_values(struct radeon_device *rdev,
   5642 				    struct ci_vbios_boot_state *boot_state)
   5643 {
   5644 	struct radeon_mode_info *mode_info = &rdev->mode_info;
   5645 	int index = GetIndexIntoMasterTable(DATA, FirmwareInfo);
   5646 	ATOM_FIRMWARE_INFO_V2_2 *firmware_info;
   5647 	u8 frev, crev;
   5648 	u16 data_offset;
   5649 
   5650 	if (atom_parse_data_header(mode_info->atom_context, index, NULL,
   5651 				   &frev, &crev, &data_offset)) {
   5652 		firmware_info =
   5653 			(ATOM_FIRMWARE_INFO_V2_2 *)(mode_info->atom_context->bios +
   5654 						    data_offset);
   5655 		boot_state->mvdd_bootup_value = le16_to_cpu(firmware_info->usBootUpMVDDCVoltage);
   5656 		boot_state->vddc_bootup_value = le16_to_cpu(firmware_info->usBootUpVDDCVoltage);
   5657 		boot_state->vddci_bootup_value = le16_to_cpu(firmware_info->usBootUpVDDCIVoltage);
   5658 		boot_state->pcie_gen_bootup_value = ci_get_current_pcie_speed(rdev);
   5659 		boot_state->pcie_lane_bootup_value = ci_get_current_pcie_lane_number(rdev);
   5660 		boot_state->sclk_bootup_value = le32_to_cpu(firmware_info->ulDefaultEngineClock);
   5661 		boot_state->mclk_bootup_value = le32_to_cpu(firmware_info->ulDefaultMemoryClock);
   5662 
   5663 		return 0;
   5664 	}
   5665 	return -EINVAL;
   5666 }
   5667 
   5668 void ci_dpm_fini(struct radeon_device *rdev)
   5669 {
   5670 	int i;
   5671 
   5672 	for (i = 0; i < rdev->pm.dpm.num_ps; i++) {
   5673 		kfree(rdev->pm.dpm.ps[i].ps_priv);
   5674 	}
   5675 	kfree(rdev->pm.dpm.ps);
   5676 	kfree(rdev->pm.dpm.priv);
   5677 	kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries);
   5678 	r600_free_extended_power_table(rdev);
   5679 }
   5680 
   5681 int ci_dpm_init(struct radeon_device *rdev)
   5682 {
   5683 	int index = GetIndexIntoMasterTable(DATA, ASIC_InternalSS_Info);
   5684 	SMU7_Discrete_DpmTable  *dpm_table;
   5685 	struct radeon_gpio_rec gpio;
   5686 	u16 data_offset, size;
   5687 	u8 frev, crev;
   5688 	struct ci_power_info *pi;
   5689 	enum pci_bus_speed speed_cap = PCI_SPEED_UNKNOWN;
   5690 	struct pci_dev *root = rdev->pdev->bus->self;
   5691 	int ret;
   5692 
   5693 	pi = kzalloc(sizeof(struct ci_power_info), GFP_KERNEL);
   5694 	if (pi == NULL)
   5695 		return -ENOMEM;
   5696 	rdev->pm.dpm.priv = pi;
   5697 
   5698 	if (!pci_is_root_bus(rdev->pdev->bus))
   5699 		speed_cap = pcie_get_speed_cap(root);
   5700 	if (speed_cap == PCI_SPEED_UNKNOWN) {
   5701 		pi->sys_pcie_mask = 0;
   5702 	} else {
   5703 		if (speed_cap == PCIE_SPEED_8_0GT)
   5704 			pi->sys_pcie_mask = RADEON_PCIE_SPEED_25 |
   5705 				RADEON_PCIE_SPEED_50 |
   5706 				RADEON_PCIE_SPEED_80;
   5707 		else if (speed_cap == PCIE_SPEED_5_0GT)
   5708 			pi->sys_pcie_mask = RADEON_PCIE_SPEED_25 |
   5709 				RADEON_PCIE_SPEED_50;
   5710 		else
   5711 			pi->sys_pcie_mask = RADEON_PCIE_SPEED_25;
   5712 	}
   5713 	pi->force_pcie_gen = RADEON_PCIE_GEN_INVALID;
   5714 
   5715 	pi->pcie_gen_performance.max = RADEON_PCIE_GEN1;
   5716 	pi->pcie_gen_performance.min = RADEON_PCIE_GEN3;
   5717 	pi->pcie_gen_powersaving.max = RADEON_PCIE_GEN1;
   5718 	pi->pcie_gen_powersaving.min = RADEON_PCIE_GEN3;
   5719 
   5720 	pi->pcie_lane_performance.max = 0;
   5721 	pi->pcie_lane_performance.min = 16;
   5722 	pi->pcie_lane_powersaving.max = 0;
   5723 	pi->pcie_lane_powersaving.min = 16;
   5724 
   5725 	ret = ci_get_vbios_boot_values(rdev, &pi->vbios_boot_state);
   5726 	if (ret) {
   5727 		ci_dpm_fini(rdev);
   5728 		return ret;
   5729 	}
   5730 
   5731 	ret = r600_get_platform_caps(rdev);
   5732 	if (ret) {
   5733 		ci_dpm_fini(rdev);
   5734 		return ret;
   5735 	}
   5736 
   5737 	ret = r600_parse_extended_power_table(rdev);
   5738 	if (ret) {
   5739 		ci_dpm_fini(rdev);
   5740 		return ret;
   5741 	}
   5742 
   5743 	ret = ci_parse_power_table(rdev);
   5744 	if (ret) {
   5745 		ci_dpm_fini(rdev);
   5746 		return ret;
   5747 	}
   5748 
   5749 	pi->dll_default_on = false;
   5750 	pi->sram_end = SMC_RAM_END;
   5751 
   5752 	pi->activity_target[0] = CISLAND_TARGETACTIVITY_DFLT;
   5753 	pi->activity_target[1] = CISLAND_TARGETACTIVITY_DFLT;
   5754 	pi->activity_target[2] = CISLAND_TARGETACTIVITY_DFLT;
   5755 	pi->activity_target[3] = CISLAND_TARGETACTIVITY_DFLT;
   5756 	pi->activity_target[4] = CISLAND_TARGETACTIVITY_DFLT;
   5757 	pi->activity_target[5] = CISLAND_TARGETACTIVITY_DFLT;
   5758 	pi->activity_target[6] = CISLAND_TARGETACTIVITY_DFLT;
   5759 	pi->activity_target[7] = CISLAND_TARGETACTIVITY_DFLT;
   5760 
   5761 	pi->mclk_activity_target = CISLAND_MCLK_TARGETACTIVITY_DFLT;
   5762 
   5763 	pi->sclk_dpm_key_disabled = 0;
   5764 	pi->mclk_dpm_key_disabled = 0;
   5765 	pi->pcie_dpm_key_disabled = 0;
   5766 	pi->thermal_sclk_dpm_enabled = 0;
   5767 
   5768 	/* mclk dpm is unstable on some R7 260X cards with the old mc ucode */
   5769 	if ((rdev->pdev->device == 0x6658) &&
   5770 	    (rdev->mc_fw->size == (BONAIRE_MC_UCODE_SIZE * 4))) {
   5771 		pi->mclk_dpm_key_disabled = 1;
   5772 	}
   5773 
   5774 	pi->caps_sclk_ds = true;
   5775 
   5776 	pi->mclk_strobe_mode_threshold = 40000;
   5777 	pi->mclk_stutter_mode_threshold = 40000;
   5778 	pi->mclk_edc_enable_threshold = 40000;
   5779 	pi->mclk_edc_wr_enable_threshold = 40000;
   5780 
   5781 	ci_initialize_powertune_defaults(rdev);
   5782 
   5783 	pi->caps_fps = false;
   5784 
   5785 	pi->caps_sclk_throttle_low_notification = false;
   5786 
   5787 	pi->caps_uvd_dpm = true;
   5788 	pi->caps_vce_dpm = true;
   5789 
   5790 	ci_get_leakage_voltages(rdev);
   5791 	ci_patch_dependency_tables_with_leakage(rdev);
   5792 	ci_set_private_data_variables_based_on_pptable(rdev);
   5793 
   5794 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries =
   5795 		kcalloc(4,
   5796 			sizeof(struct radeon_clock_voltage_dependency_entry),
   5797 			GFP_KERNEL);
   5798 	if (!rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries) {
   5799 		ci_dpm_fini(rdev);
   5800 		return -ENOMEM;
   5801 	}
   5802 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.count = 4;
   5803 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[0].clk = 0;
   5804 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[0].v = 0;
   5805 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[1].clk = 36000;
   5806 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[1].v = 720;
   5807 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[2].clk = 54000;
   5808 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[2].v = 810;
   5809 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[3].clk = 72000;
   5810 	rdev->pm.dpm.dyn_state.vddc_dependency_on_dispclk.entries[3].v = 900;
   5811 
   5812 	rdev->pm.dpm.dyn_state.mclk_sclk_ratio = 4;
   5813 	rdev->pm.dpm.dyn_state.sclk_mclk_delta = 15000;
   5814 	rdev->pm.dpm.dyn_state.vddc_vddci_delta = 200;
   5815 
   5816 	rdev->pm.dpm.dyn_state.valid_sclk_values.count = 0;
   5817 	rdev->pm.dpm.dyn_state.valid_sclk_values.values = NULL;
   5818 	rdev->pm.dpm.dyn_state.valid_mclk_values.count = 0;
   5819 	rdev->pm.dpm.dyn_state.valid_mclk_values.values = NULL;
   5820 
   5821 	if (rdev->family == CHIP_HAWAII) {
   5822 		pi->thermal_temp_setting.temperature_low = 94500;
   5823 		pi->thermal_temp_setting.temperature_high = 95000;
   5824 		pi->thermal_temp_setting.temperature_shutdown = 104000;
   5825 	} else {
   5826 		pi->thermal_temp_setting.temperature_low = 99500;
   5827 		pi->thermal_temp_setting.temperature_high = 100000;
   5828 		pi->thermal_temp_setting.temperature_shutdown = 104000;
   5829 	}
   5830 
   5831 	pi->uvd_enabled = false;
   5832 
   5833 	dpm_table = &pi->smc_state_table;
   5834 
   5835 	gpio = radeon_atombios_lookup_gpio(rdev, VDDC_VRHOT_GPIO_PINID);
   5836 	if (gpio.valid) {
   5837 		dpm_table->VRHotGpio = gpio.shift;
   5838 		rdev->pm.dpm.platform_caps |= ATOM_PP_PLATFORM_CAP_REGULATOR_HOT;
   5839 	} else {
   5840 		dpm_table->VRHotGpio = CISLANDS_UNUSED_GPIO_PIN;
   5841 		rdev->pm.dpm.platform_caps &= ~ATOM_PP_PLATFORM_CAP_REGULATOR_HOT;
   5842 	}
   5843 
   5844 	gpio = radeon_atombios_lookup_gpio(rdev, PP_AC_DC_SWITCH_GPIO_PINID);
   5845 	if (gpio.valid) {
   5846 		dpm_table->AcDcGpio = gpio.shift;
   5847 		rdev->pm.dpm.platform_caps |= ATOM_PP_PLATFORM_CAP_HARDWAREDC;
   5848 	} else {
   5849 		dpm_table->AcDcGpio = CISLANDS_UNUSED_GPIO_PIN;
   5850 		rdev->pm.dpm.platform_caps &= ~ATOM_PP_PLATFORM_CAP_HARDWAREDC;
   5851 	}
   5852 
   5853 	gpio = radeon_atombios_lookup_gpio(rdev, VDDC_PCC_GPIO_PINID);
   5854 	if (gpio.valid) {
   5855 		u32 tmp = RREG32_SMC(CNB_PWRMGT_CNTL);
   5856 
   5857 		switch (gpio.shift) {
   5858 		case 0:
   5859 			tmp &= ~GNB_SLOW_MODE_MASK;
   5860 			tmp |= GNB_SLOW_MODE(1);
   5861 			break;
   5862 		case 1:
   5863 			tmp &= ~GNB_SLOW_MODE_MASK;
   5864 			tmp |= GNB_SLOW_MODE(2);
   5865 			break;
   5866 		case 2:
   5867 			tmp |= GNB_SLOW;
   5868 			break;
   5869 		case 3:
   5870 			tmp |= FORCE_NB_PS1;
   5871 			break;
   5872 		case 4:
   5873 			tmp |= DPM_ENABLED;
   5874 			break;
   5875 		default:
   5876 			DRM_DEBUG("Invalid PCC GPIO: %u!\n", gpio.shift);
   5877 			break;
   5878 		}
   5879 		WREG32_SMC(CNB_PWRMGT_CNTL, tmp);
   5880 	}
   5881 
   5882 	pi->voltage_control = CISLANDS_VOLTAGE_CONTROL_NONE;
   5883 	pi->vddci_control = CISLANDS_VOLTAGE_CONTROL_NONE;
   5884 	pi->mvdd_control = CISLANDS_VOLTAGE_CONTROL_NONE;
   5885 	if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_VDDC, VOLTAGE_OBJ_GPIO_LUT))
   5886 		pi->voltage_control = CISLANDS_VOLTAGE_CONTROL_BY_GPIO;
   5887 	else if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_VDDC, VOLTAGE_OBJ_SVID2))
   5888 		pi->voltage_control = CISLANDS_VOLTAGE_CONTROL_BY_SVID2;
   5889 
   5890 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_VDDCI_CONTROL) {
   5891 		if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_VDDCI, VOLTAGE_OBJ_GPIO_LUT))
   5892 			pi->vddci_control = CISLANDS_VOLTAGE_CONTROL_BY_GPIO;
   5893 		else if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_VDDCI, VOLTAGE_OBJ_SVID2))
   5894 			pi->vddci_control = CISLANDS_VOLTAGE_CONTROL_BY_SVID2;
   5895 		else
   5896 			rdev->pm.dpm.platform_caps &= ~ATOM_PP_PLATFORM_CAP_VDDCI_CONTROL;
   5897 	}
   5898 
   5899 	if (rdev->pm.dpm.platform_caps & ATOM_PP_PLATFORM_CAP_MVDDCONTROL) {
   5900 		if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_MVDDC, VOLTAGE_OBJ_GPIO_LUT))
   5901 			pi->mvdd_control = CISLANDS_VOLTAGE_CONTROL_BY_GPIO;
   5902 		else if (radeon_atom_is_voltage_gpio(rdev, VOLTAGE_TYPE_MVDDC, VOLTAGE_OBJ_SVID2))
   5903 			pi->mvdd_control = CISLANDS_VOLTAGE_CONTROL_BY_SVID2;
   5904 		else
   5905 			rdev->pm.dpm.platform_caps &= ~ATOM_PP_PLATFORM_CAP_MVDDCONTROL;
   5906 	}
   5907 
   5908 	pi->vddc_phase_shed_control = true;
   5909 
   5910 #if defined(CONFIG_ACPI)
   5911 	pi->pcie_performance_request =
   5912 		radeon_acpi_is_pcie_performance_request_supported(rdev);
   5913 #else
   5914 	pi->pcie_performance_request = false;
   5915 #endif
   5916 
   5917 	if (atom_parse_data_header(rdev->mode_info.atom_context, index, &size,
   5918 				   &frev, &crev, &data_offset)) {
   5919 		pi->caps_sclk_ss_support = true;
   5920 		pi->caps_mclk_ss_support = true;
   5921 		pi->dynamic_ss = true;
   5922 	} else {
   5923 		pi->caps_sclk_ss_support = false;
   5924 		pi->caps_mclk_ss_support = false;
   5925 		pi->dynamic_ss = true;
   5926 	}
   5927 
   5928 	if (rdev->pm.int_thermal_type != THERMAL_TYPE_NONE)
   5929 		pi->thermal_protection = true;
   5930 	else
   5931 		pi->thermal_protection = false;
   5932 
   5933 	pi->caps_dynamic_ac_timing = true;
   5934 
   5935 	pi->uvd_power_gated = false;
   5936 
   5937 	/* make sure dc limits are valid */
   5938 	if ((rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc.sclk == 0) ||
   5939 	    (rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc.mclk == 0))
   5940 		rdev->pm.dpm.dyn_state.max_clock_voltage_on_dc =
   5941 			rdev->pm.dpm.dyn_state.max_clock_voltage_on_ac;
   5942 
   5943 	pi->fan_ctrl_is_in_default_mode = true;
   5944 
   5945 	return 0;
   5946 }
   5947 
   5948 #ifdef CONFIG_DEBUG_FS
   5949 void ci_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev,
   5950 						    struct seq_file *m)
   5951 {
   5952 	struct ci_power_info *pi = ci_get_pi(rdev);
   5953 	struct radeon_ps *rps = &pi->current_rps;
   5954 	u32 sclk = ci_get_average_sclk_freq(rdev);
   5955 	u32 mclk = ci_get_average_mclk_freq(rdev);
   5956 
   5957 	seq_printf(m, "uvd    %sabled\n", pi->uvd_enabled ? "en" : "dis");
   5958 	seq_printf(m, "vce    %sabled\n", rps->vce_active ? "en" : "dis");
   5959 	seq_printf(m, "power level avg    sclk: %u mclk: %u\n",
   5960 		   sclk, mclk);
   5961 }
   5962 #endif
   5963 
   5964 void ci_dpm_print_power_state(struct radeon_device *rdev,
   5965 			      struct radeon_ps *rps)
   5966 {
   5967 	struct ci_ps *ps = ci_get_ps(rps);
   5968 	struct ci_pl *pl;
   5969 	int i;
   5970 
   5971 	r600_dpm_print_class_info(rps->class, rps->class2);
   5972 	r600_dpm_print_cap_info(rps->caps);
   5973 	printk("\tuvd    vclk: %d dclk: %d\n", rps->vclk, rps->dclk);
   5974 	for (i = 0; i < ps->performance_level_count; i++) {
   5975 		pl = &ps->performance_levels[i];
   5976 		printk("\t\tpower level %d    sclk: %u mclk: %u pcie gen: %u pcie lanes: %u\n",
   5977 		       i, pl->sclk, pl->mclk, pl->pcie_gen + 1, pl->pcie_lane);
   5978 	}
   5979 	r600_dpm_print_ps_status(rdev, rps);
   5980 }
   5981 
   5982 u32 ci_dpm_get_current_sclk(struct radeon_device *rdev)
   5983 {
   5984 	u32 sclk = ci_get_average_sclk_freq(rdev);
   5985 
   5986 	return sclk;
   5987 }
   5988 
   5989 u32 ci_dpm_get_current_mclk(struct radeon_device *rdev)
   5990 {
   5991 	u32 mclk = ci_get_average_mclk_freq(rdev);
   5992 
   5993 	return mclk;
   5994 }
   5995 
   5996 u32 ci_dpm_get_sclk(struct radeon_device *rdev, bool low)
   5997 {
   5998 	struct ci_power_info *pi = ci_get_pi(rdev);
   5999 	struct ci_ps *requested_state = ci_get_ps(&pi->requested_rps);
   6000 
   6001 	if (low)
   6002 		return requested_state->performance_levels[0].sclk;
   6003 	else
   6004 		return requested_state->performance_levels[requested_state->performance_level_count - 1].sclk;
   6005 }
   6006 
   6007 u32 ci_dpm_get_mclk(struct radeon_device *rdev, bool low)
   6008 {
   6009 	struct ci_power_info *pi = ci_get_pi(rdev);
   6010 	struct ci_ps *requested_state = ci_get_ps(&pi->requested_rps);
   6011 
   6012 	if (low)
   6013 		return requested_state->performance_levels[0].mclk;
   6014 	else
   6015 		return requested_state->performance_levels[requested_state->performance_level_count - 1].mclk;
   6016 }
   6017