Home | History | Annotate | Line # | Download | only in clk
      1 /*	$NetBSD: nouveau_nvkm_subdev_clk_base.c,v 1.5 2021/12/18 23:45:39 riastradh Exp $	*/
      2 
      3 /*
      4  * Copyright 2013 Red Hat Inc.
      5  *
      6  * Permission is hereby granted, free of charge, to any person obtaining a
      7  * copy of this software and associated documentation files (the "Software"),
      8  * to deal in the Software without restriction, including without limitation
      9  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
     10  * and/or sell copies of the Software, and to permit persons to whom the
     11  * Software is furnished to do so, subject to the following conditions:
     12  *
     13  * The above copyright notice and this permission notice shall be included in
     14  * all copies or substantial portions of the Software.
     15  *
     16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     19  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     20  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     21  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     22  * OTHER DEALINGS IN THE SOFTWARE.
     23  *
     24  * Authors: Ben Skeggs
     25  */
     26 #include <sys/cdefs.h>
     27 __KERNEL_RCSID(0, "$NetBSD: nouveau_nvkm_subdev_clk_base.c,v 1.5 2021/12/18 23:45:39 riastradh Exp $");
     28 
     29 #include "priv.h"
     30 
     31 #include <subdev/bios.h>
     32 #include <subdev/bios/boost.h>
     33 #include <subdev/bios/cstep.h>
     34 #include <subdev/bios/perf.h>
     35 #include <subdev/bios/vpstate.h>
     36 #include <subdev/fb.h>
     37 #include <subdev/therm.h>
     38 #include <subdev/volt.h>
     39 
     40 #include <core/option.h>
     41 
     42 /******************************************************************************
     43  * misc
     44  *****************************************************************************/
     45 static u32
     46 nvkm_clk_adjust(struct nvkm_clk *clk, bool adjust,
     47 		u8 pstate, u8 domain, u32 input)
     48 {
     49 	struct nvkm_bios *bios = clk->subdev.device->bios;
     50 	struct nvbios_boostE boostE;
     51 	u8  ver, hdr, cnt, len;
     52 	u32 data;
     53 
     54 	data = nvbios_boostEm(bios, pstate, &ver, &hdr, &cnt, &len, &boostE);
     55 	if (data) {
     56 		struct nvbios_boostS boostS;
     57 		u8  idx = 0, sver, shdr;
     58 		u32 subd;
     59 
     60 		input = max(boostE.min, input);
     61 		input = min(boostE.max, input);
     62 		do {
     63 			sver = ver;
     64 			shdr = hdr;
     65 			subd = nvbios_boostSp(bios, idx++, data, &sver, &shdr,
     66 					      cnt, len, &boostS);
     67 			if (subd && boostS.domain == domain) {
     68 				if (adjust)
     69 					input = input * boostS.percent / 100;
     70 				input = max(boostS.min, input);
     71 				input = min(boostS.max, input);
     72 				break;
     73 			}
     74 		} while (subd);
     75 	}
     76 
     77 	return input;
     78 }
     79 
     80 /******************************************************************************
     81  * C-States
     82  *****************************************************************************/
     83 static bool
     84 nvkm_cstate_valid(struct nvkm_clk *clk, struct nvkm_cstate *cstate,
     85 		  u32 max_volt, int temp)
     86 {
     87 	const struct nvkm_domain *domain = clk->domains;
     88 	struct nvkm_volt *volt = clk->subdev.device->volt;
     89 	int voltage;
     90 
     91 	while (domain && domain->name != nv_clk_src_max) {
     92 		if (domain->flags & NVKM_CLK_DOM_FLAG_VPSTATE) {
     93 			u32 freq = cstate->domain[domain->name];
     94 			switch (clk->boost_mode) {
     95 			case NVKM_CLK_BOOST_NONE:
     96 				if (clk->base_khz && freq > clk->base_khz)
     97 					return false;
     98 				/* fall through */
     99 			case NVKM_CLK_BOOST_BIOS:
    100 				if (clk->boost_khz && freq > clk->boost_khz)
    101 					return false;
    102 			}
    103 		}
    104 		domain++;
    105 	}
    106 
    107 	if (!volt)
    108 		return true;
    109 
    110 	voltage = nvkm_volt_map(volt, cstate->voltage, temp);
    111 	if (voltage < 0)
    112 		return false;
    113 	return voltage <= min(max_volt, volt->max_uv);
    114 }
    115 
    116 static struct nvkm_cstate *
    117 nvkm_cstate_find_best(struct nvkm_clk *clk, struct nvkm_pstate *pstate,
    118 		      struct nvkm_cstate *cstate)
    119 {
    120 	struct nvkm_device *device = clk->subdev.device;
    121 	struct nvkm_volt *volt = device->volt;
    122 	int max_volt;
    123 
    124 	if (!pstate || !cstate)
    125 		return NULL;
    126 
    127 	if (!volt)
    128 		return cstate;
    129 
    130 	max_volt = volt->max_uv;
    131 	if (volt->max0_id != 0xff)
    132 		max_volt = min(max_volt,
    133 			       nvkm_volt_map(volt, volt->max0_id, clk->temp));
    134 	if (volt->max1_id != 0xff)
    135 		max_volt = min(max_volt,
    136 			       nvkm_volt_map(volt, volt->max1_id, clk->temp));
    137 	if (volt->max2_id != 0xff)
    138 		max_volt = min(max_volt,
    139 			       nvkm_volt_map(volt, volt->max2_id, clk->temp));
    140 
    141 	list_for_each_entry_from_reverse(cstate, &pstate->list, head) {
    142 		if (nvkm_cstate_valid(clk, cstate, max_volt, clk->temp))
    143 			break;
    144 	}
    145 
    146 	return cstate;
    147 }
    148 
    149 static struct nvkm_cstate *
    150 nvkm_cstate_get(struct nvkm_clk *clk, struct nvkm_pstate *pstate, int cstatei)
    151 {
    152 	struct nvkm_cstate *cstate;
    153 	if (cstatei == NVKM_CLK_CSTATE_HIGHEST)
    154 		return list_last_entry(&pstate->list, typeof(*cstate), head);
    155 	else {
    156 		list_for_each_entry(cstate, &pstate->list, head) {
    157 			if (cstate->id == cstatei)
    158 				return cstate;
    159 		}
    160 	}
    161 	return NULL;
    162 }
    163 
    164 static int
    165 nvkm_cstate_prog(struct nvkm_clk *clk, struct nvkm_pstate *pstate, int cstatei)
    166 {
    167 	struct nvkm_subdev *subdev = &clk->subdev;
    168 	struct nvkm_device *device = subdev->device;
    169 	struct nvkm_therm *therm = device->therm;
    170 	struct nvkm_volt *volt = device->volt;
    171 	struct nvkm_cstate *cstate;
    172 	int ret;
    173 
    174 	if (!list_empty(&pstate->list)) {
    175 		cstate = nvkm_cstate_get(clk, pstate, cstatei);
    176 		cstate = nvkm_cstate_find_best(clk, pstate, cstate);
    177 	} else {
    178 		cstate = &pstate->base;
    179 	}
    180 
    181 	if (therm) {
    182 		ret = nvkm_therm_cstate(therm, pstate->fanspeed, +1);
    183 		if (ret && ret != -ENODEV) {
    184 			nvkm_error(subdev, "failed to raise fan speed: %d\n", ret);
    185 			return ret;
    186 		}
    187 	}
    188 
    189 	if (volt) {
    190 		ret = nvkm_volt_set_id(volt, cstate->voltage,
    191 				       pstate->base.voltage, clk->temp, +1);
    192 		if (ret && ret != -ENODEV) {
    193 			nvkm_error(subdev, "failed to raise voltage: %d\n", ret);
    194 			return ret;
    195 		}
    196 	}
    197 
    198 	ret = clk->func->calc(clk, cstate);
    199 	if (ret == 0) {
    200 		ret = clk->func->prog(clk);
    201 		clk->func->tidy(clk);
    202 	}
    203 
    204 	if (volt) {
    205 		ret = nvkm_volt_set_id(volt, cstate->voltage,
    206 				       pstate->base.voltage, clk->temp, -1);
    207 		if (ret && ret != -ENODEV)
    208 			nvkm_error(subdev, "failed to lower voltage: %d\n", ret);
    209 	}
    210 
    211 	if (therm) {
    212 		ret = nvkm_therm_cstate(therm, pstate->fanspeed, -1);
    213 		if (ret && ret != -ENODEV)
    214 			nvkm_error(subdev, "failed to lower fan speed: %d\n", ret);
    215 	}
    216 
    217 	return ret;
    218 }
    219 
    220 static void
    221 nvkm_cstate_del(struct nvkm_cstate *cstate)
    222 {
    223 	list_del(&cstate->head);
    224 	kfree(cstate);
    225 }
    226 
    227 static int
    228 nvkm_cstate_new(struct nvkm_clk *clk, int idx, struct nvkm_pstate *pstate)
    229 {
    230 	struct nvkm_bios *bios = clk->subdev.device->bios;
    231 	struct nvkm_volt *volt = clk->subdev.device->volt;
    232 	const struct nvkm_domain *domain = clk->domains;
    233 	struct nvkm_cstate *cstate = NULL;
    234 	struct nvbios_cstepX cstepX;
    235 	u8  ver, hdr;
    236 	u32 data;
    237 
    238 	data = nvbios_cstepXp(bios, idx, &ver, &hdr, &cstepX);
    239 	if (!data)
    240 		return -ENOENT;
    241 
    242 	if (volt && nvkm_volt_map_min(volt, cstepX.voltage) > volt->max_uv)
    243 		return -EINVAL;
    244 
    245 	cstate = kzalloc(sizeof(*cstate), GFP_KERNEL);
    246 	if (!cstate)
    247 		return -ENOMEM;
    248 
    249 	*cstate = pstate->base;
    250 	cstate->voltage = cstepX.voltage;
    251 	cstate->id = idx;
    252 
    253 	while (domain && domain->name != nv_clk_src_max) {
    254 		if (domain->flags & NVKM_CLK_DOM_FLAG_CORE) {
    255 			u32 freq = nvkm_clk_adjust(clk, true, pstate->pstate,
    256 						   domain->bios, cstepX.freq);
    257 			cstate->domain[domain->name] = freq;
    258 		}
    259 		domain++;
    260 	}
    261 
    262 	list_add(&cstate->head, &pstate->list);
    263 	return 0;
    264 }
    265 
    266 /******************************************************************************
    267  * P-States
    268  *****************************************************************************/
    269 static int
    270 nvkm_pstate_prog(struct nvkm_clk *clk, int pstatei)
    271 {
    272 	struct nvkm_subdev *subdev = &clk->subdev;
    273 	struct nvkm_fb *fb = subdev->device->fb;
    274 	struct nvkm_pci *pci = subdev->device->pci;
    275 	struct nvkm_pstate *pstate;
    276 	int ret, idx = 0;
    277 
    278 	list_for_each_entry(pstate, &clk->states, head) {
    279 		if (idx++ == pstatei)
    280 			break;
    281 	}
    282 
    283 	nvkm_debug(subdev, "setting performance state %d\n", pstatei);
    284 	clk->pstate = pstatei;
    285 
    286 	nvkm_pcie_set_link(pci, pstate->pcie_speed, pstate->pcie_width);
    287 
    288 	if (fb && fb->ram && fb->ram->func->calc) {
    289 		struct nvkm_ram *ram = fb->ram;
    290 		int khz = pstate->base.domain[nv_clk_src_mem];
    291 		do {
    292 			ret = ram->func->calc(ram, khz);
    293 			if (ret == 0)
    294 				ret = ram->func->prog(ram);
    295 		} while (ret > 0);
    296 		ram->func->tidy(ram);
    297 	}
    298 
    299 	return nvkm_cstate_prog(clk, pstate, NVKM_CLK_CSTATE_HIGHEST);
    300 }
    301 
    302 static void
    303 nvkm_pstate_work(struct work_struct *work)
    304 {
    305 	struct nvkm_clk *clk = container_of(work, typeof(*clk), work);
    306 	struct nvkm_subdev *subdev = &clk->subdev;
    307 	int pstate;
    308 
    309 	spin_lock(&clk->lock);
    310 	if (!atomic_xchg(&clk->waiting, 0)) {
    311 		spin_unlock(&clk->lock);
    312 		return;
    313 	}
    314 
    315 	clk->pwrsrc = power_supply_is_system_supplied();
    316 
    317 	nvkm_trace(subdev, "P %d PWR %d U(AC) %d U(DC) %d A %d T %dC D %d\n",
    318 		   clk->pstate, clk->pwrsrc, clk->ustate_ac, clk->ustate_dc,
    319 		   clk->astate, clk->temp, clk->dstate);
    320 
    321 	pstate = clk->pwrsrc ? clk->ustate_ac : clk->ustate_dc;
    322 	if (clk->state_nr && pstate != -1) {
    323 		pstate = (pstate < 0) ? clk->astate : pstate;
    324 		pstate = min(pstate, clk->state_nr - 1);
    325 		pstate = max(pstate, clk->dstate);
    326 	} else {
    327 		pstate = clk->pstate = -1;
    328 	}
    329 
    330 	nvkm_trace(subdev, "-> %d\n", pstate);
    331 	if (pstate != clk->pstate) {
    332 		int ret = nvkm_pstate_prog(clk, pstate);
    333 		if (ret) {
    334 			nvkm_error(subdev, "error setting pstate %d: %d\n",
    335 				   pstate, ret);
    336 		}
    337 	}
    338 
    339 #ifdef __NetBSD__
    340 	DRM_SPIN_WAKEUP_ALL(&clk->wait, &clk->lock);
    341 #else
    342 	wake_up_all(&clk->wait);
    343 #endif
    344 	spin_unlock(&clk->lock);
    345 	nvkm_notify_get(&clk->pwrsrc_ntfy);
    346 }
    347 
    348 static int
    349 nvkm_pstate_calc(struct nvkm_clk *clk, bool wait)
    350 {
    351 	atomic_set(&clk->waiting, 1);
    352 	schedule_work(&clk->work);
    353 	if (wait) {
    354 #ifdef __NetBSD__
    355 		if (cold) {
    356 			unsigned timo = 1000;
    357 			while (timo --> 0) {
    358 				if (atomic_read(&clk->waiting))
    359 					return 0;
    360 				DELAY(100);
    361 			}
    362 			return -ETIMEDOUT;
    363 		}
    364 		int ret;
    365 		spin_lock(&clk->lock);
    366 		DRM_SPIN_WAIT_NOINTR_UNTIL(ret, &clk->wait, &clk->lock,
    367 		    !atomic_read(&clk->waiting));
    368 		spin_unlock(&clk->lock);
    369 		KASSERT(ret == 0);
    370 #else
    371 		wait_event(clk->wait, !atomic_read(&clk->waiting));
    372 #endif
    373 	}
    374 	return 0;
    375 }
    376 
    377 static void
    378 nvkm_pstate_info(struct nvkm_clk *clk, struct nvkm_pstate *pstate)
    379 {
    380 	const struct nvkm_domain *clock = clk->domains - 1;
    381 	struct nvkm_cstate *cstate;
    382 	struct nvkm_subdev *subdev = &clk->subdev;
    383 	char info[3][32] = { "", "", "" };
    384 	char name[4] = "--";
    385 	int i = -1;
    386 
    387 	if (pstate->pstate != 0xff)
    388 		snprintf(name, sizeof(name), "%02x", pstate->pstate);
    389 
    390 	while ((++clock)->name != nv_clk_src_max) {
    391 		u32 lo = pstate->base.domain[clock->name];
    392 		u32 hi = lo;
    393 		if (hi == 0)
    394 			continue;
    395 
    396 		nvkm_debug(subdev, "%02x: %10d KHz\n", clock->name, lo);
    397 		list_for_each_entry(cstate, &pstate->list, head) {
    398 			u32 freq = cstate->domain[clock->name];
    399 			lo = min(lo, freq);
    400 			hi = max(hi, freq);
    401 			nvkm_debug(subdev, "%10d KHz\n", freq);
    402 		}
    403 
    404 		if (clock->mname && ++i < ARRAY_SIZE(info)) {
    405 			lo /= clock->mdiv;
    406 			hi /= clock->mdiv;
    407 			if (lo == hi) {
    408 				snprintf(info[i], sizeof(info[i]), "%s %d MHz",
    409 					 clock->mname, lo);
    410 			} else {
    411 				snprintf(info[i], sizeof(info[i]),
    412 					 "%s %d-%d MHz", clock->mname, lo, hi);
    413 			}
    414 		}
    415 	}
    416 
    417 	nvkm_debug(subdev, "%s: %s %s %s\n", name, info[0], info[1], info[2]);
    418 }
    419 
    420 static void
    421 nvkm_pstate_del(struct nvkm_pstate *pstate)
    422 {
    423 	struct nvkm_cstate *cstate, *temp;
    424 
    425 	list_for_each_entry_safe(cstate, temp, &pstate->list, head) {
    426 		nvkm_cstate_del(cstate);
    427 	}
    428 
    429 	list_del(&pstate->head);
    430 	kfree(pstate);
    431 }
    432 
    433 static int
    434 nvkm_pstate_new(struct nvkm_clk *clk, int idx)
    435 {
    436 	struct nvkm_bios *bios = clk->subdev.device->bios;
    437 	const struct nvkm_domain *domain = clk->domains - 1;
    438 	struct nvkm_pstate *pstate;
    439 	struct nvkm_cstate *cstate;
    440 	struct nvbios_cstepE cstepE;
    441 	struct nvbios_perfE perfE;
    442 	u8  ver, hdr, cnt, len;
    443 	u32 data;
    444 
    445 	data = nvbios_perfEp(bios, idx, &ver, &hdr, &cnt, &len, &perfE);
    446 	if (!data)
    447 		return -EINVAL;
    448 	if (perfE.pstate == 0xff)
    449 		return 0;
    450 
    451 	pstate = kzalloc(sizeof(*pstate), GFP_KERNEL);
    452 	cstate = &pstate->base;
    453 	if (!pstate)
    454 		return -ENOMEM;
    455 
    456 	INIT_LIST_HEAD(&pstate->list);
    457 
    458 	pstate->pstate = perfE.pstate;
    459 	pstate->fanspeed = perfE.fanspeed;
    460 	pstate->pcie_speed = perfE.pcie_speed;
    461 	pstate->pcie_width = perfE.pcie_width;
    462 	cstate->voltage = perfE.voltage;
    463 	cstate->domain[nv_clk_src_core] = perfE.core;
    464 	cstate->domain[nv_clk_src_shader] = perfE.shader;
    465 	cstate->domain[nv_clk_src_mem] = perfE.memory;
    466 	cstate->domain[nv_clk_src_vdec] = perfE.vdec;
    467 	cstate->domain[nv_clk_src_dom6] = perfE.disp;
    468 
    469 	while (ver >= 0x40 && (++domain)->name != nv_clk_src_max) {
    470 		struct nvbios_perfS perfS;
    471 		u8  sver = ver, shdr = hdr;
    472 		u32 perfSe = nvbios_perfSp(bios, data, domain->bios,
    473 					  &sver, &shdr, cnt, len, &perfS);
    474 		if (perfSe == 0 || sver != 0x40)
    475 			continue;
    476 
    477 		if (domain->flags & NVKM_CLK_DOM_FLAG_CORE) {
    478 			perfS.v40.freq = nvkm_clk_adjust(clk, false,
    479 							 pstate->pstate,
    480 							 domain->bios,
    481 							 perfS.v40.freq);
    482 		}
    483 
    484 		cstate->domain[domain->name] = perfS.v40.freq;
    485 	}
    486 
    487 	data = nvbios_cstepEm(bios, pstate->pstate, &ver, &hdr, &cstepE);
    488 	if (data) {
    489 		int idx = cstepE.index;
    490 		do {
    491 			nvkm_cstate_new(clk, idx, pstate);
    492 		} while(idx--);
    493 	}
    494 
    495 	nvkm_pstate_info(clk, pstate);
    496 	list_add_tail(&pstate->head, &clk->states);
    497 	clk->state_nr++;
    498 	return 0;
    499 }
    500 
    501 /******************************************************************************
    502  * Adjustment triggers
    503  *****************************************************************************/
    504 static int
    505 nvkm_clk_ustate_update(struct nvkm_clk *clk, int req)
    506 {
    507 	struct nvkm_pstate *pstate;
    508 	int i = 0;
    509 
    510 	if (!clk->allow_reclock)
    511 		return -ENOSYS;
    512 
    513 	if (req != -1 && req != -2) {
    514 		list_for_each_entry(pstate, &clk->states, head) {
    515 			if (pstate->pstate == req)
    516 				break;
    517 			i++;
    518 		}
    519 
    520 		if (pstate->pstate != req)
    521 			return -EINVAL;
    522 		req = i;
    523 	}
    524 
    525 	return req + 2;
    526 }
    527 
    528 static int
    529 nvkm_clk_nstate(struct nvkm_clk *clk, const char *mode, int arglen)
    530 {
    531 	int ret = 1;
    532 
    533 	if (clk->allow_reclock && !strncasecmpz(mode, "auto", arglen))
    534 		return -2;
    535 
    536 	if (strncasecmpz(mode, "disabled", arglen)) {
    537 		char *m = kstrndup(mode, arglen, GFP_KERNEL);
    538 		long v;
    539 
    540 		if (!kstrtol(m, 0, &v)) {
    541 			ret = nvkm_clk_ustate_update(clk, v);
    542 			if (ret < 0)
    543 				ret = 1;
    544 		}
    545 		kfree(m);
    546 	}
    547 
    548 	return ret - 2;
    549 }
    550 
    551 int
    552 nvkm_clk_ustate(struct nvkm_clk *clk, int req, int pwr)
    553 {
    554 	int ret = nvkm_clk_ustate_update(clk, req);
    555 	if (ret >= 0) {
    556 		if (ret -= 2, pwr) clk->ustate_ac = ret;
    557 		else		   clk->ustate_dc = ret;
    558 		return nvkm_pstate_calc(clk, true);
    559 	}
    560 	return ret;
    561 }
    562 
    563 int
    564 nvkm_clk_astate(struct nvkm_clk *clk, int req, int rel, bool wait)
    565 {
    566 	if (!rel) clk->astate  = req;
    567 	if ( rel) clk->astate += rel;
    568 	clk->astate = min(clk->astate, clk->state_nr - 1);
    569 	clk->astate = max(clk->astate, 0);
    570 	return nvkm_pstate_calc(clk, wait);
    571 }
    572 
    573 int
    574 nvkm_clk_tstate(struct nvkm_clk *clk, u8 temp)
    575 {
    576 	if (clk->temp == temp)
    577 		return 0;
    578 	clk->temp = temp;
    579 	return nvkm_pstate_calc(clk, false);
    580 }
    581 
    582 int
    583 nvkm_clk_dstate(struct nvkm_clk *clk, int req, int rel)
    584 {
    585 	if (!rel) clk->dstate  = req;
    586 	if ( rel) clk->dstate += rel;
    587 	clk->dstate = min(clk->dstate, clk->state_nr - 1);
    588 	clk->dstate = max(clk->dstate, 0);
    589 	return nvkm_pstate_calc(clk, true);
    590 }
    591 
    592 static int
    593 nvkm_clk_pwrsrc(struct nvkm_notify *notify)
    594 {
    595 	struct nvkm_clk *clk =
    596 		container_of(notify, typeof(*clk), pwrsrc_ntfy);
    597 	nvkm_pstate_calc(clk, false);
    598 	return NVKM_NOTIFY_DROP;
    599 }
    600 
    601 /******************************************************************************
    602  * subdev base class implementation
    603  *****************************************************************************/
    604 
    605 int
    606 nvkm_clk_read(struct nvkm_clk *clk, enum nv_clk_src src)
    607 {
    608 	return clk->func->read(clk, src);
    609 }
    610 
    611 static int
    612 nvkm_clk_fini(struct nvkm_subdev *subdev, bool suspend)
    613 {
    614 	struct nvkm_clk *clk = nvkm_clk(subdev);
    615 	nvkm_notify_put(&clk->pwrsrc_ntfy);
    616 	flush_work(&clk->work);
    617 	if (clk->func->fini)
    618 		clk->func->fini(clk);
    619 	return 0;
    620 }
    621 
    622 static int
    623 nvkm_clk_init(struct nvkm_subdev *subdev)
    624 {
    625 	struct nvkm_clk *clk = nvkm_clk(subdev);
    626 	const struct nvkm_domain *clock = clk->domains;
    627 	int ret;
    628 
    629 	memset(&clk->bstate, 0x00, sizeof(clk->bstate));
    630 	INIT_LIST_HEAD(&clk->bstate.list);
    631 	clk->bstate.pstate = 0xff;
    632 
    633 	while (clock->name != nv_clk_src_max) {
    634 		ret = nvkm_clk_read(clk, clock->name);
    635 		if (ret < 0) {
    636 			nvkm_error(subdev, "%02x freq unknown\n", clock->name);
    637 			return ret;
    638 		}
    639 		clk->bstate.base.domain[clock->name] = ret;
    640 		clock++;
    641 	}
    642 
    643 	nvkm_pstate_info(clk, &clk->bstate);
    644 
    645 	if (clk->func->init)
    646 		return clk->func->init(clk);
    647 
    648 	clk->astate = clk->state_nr - 1;
    649 	clk->dstate = 0;
    650 	clk->pstate = -1;
    651 	clk->temp = 90; /* reasonable default value */
    652 	nvkm_pstate_calc(clk, true);
    653 	return 0;
    654 }
    655 
    656 static void *
    657 nvkm_clk_dtor(struct nvkm_subdev *subdev)
    658 {
    659 	struct nvkm_clk *clk = nvkm_clk(subdev);
    660 	struct nvkm_pstate *pstate, *temp;
    661 
    662 	nvkm_notify_fini(&clk->pwrsrc_ntfy);
    663 
    664 	/* Early return if the pstates have been provided statically */
    665 	if (clk->func->pstates)
    666 		return clk;
    667 
    668 	list_for_each_entry_safe(pstate, temp, &clk->states, head) {
    669 		nvkm_pstate_del(pstate);
    670 	}
    671 
    672 #ifdef __NetBSD__
    673 	DRM_DESTROY_WAITQUEUE(&clk->wait);
    674 	spin_lock_destroy(&clk->lock);
    675 #endif
    676 
    677 	return clk;
    678 }
    679 
    680 static const struct nvkm_subdev_func
    681 nvkm_clk = {
    682 	.dtor = nvkm_clk_dtor,
    683 	.init = nvkm_clk_init,
    684 	.fini = nvkm_clk_fini,
    685 };
    686 
    687 int
    688 nvkm_clk_ctor(const struct nvkm_clk_func *func, struct nvkm_device *device,
    689 	      int index, bool allow_reclock, struct nvkm_clk *clk)
    690 {
    691 	struct nvkm_subdev *subdev = &clk->subdev;
    692 	struct nvkm_bios *bios = device->bios;
    693 	int ret, idx, arglen;
    694 	const char *mode;
    695 	struct nvbios_vpstate_header h;
    696 
    697 	nvkm_subdev_ctor(&nvkm_clk, device, index, subdev);
    698 
    699 	if (bios && !nvbios_vpstate_parse(bios, &h)) {
    700 		struct nvbios_vpstate_entry base, boost;
    701 		if (!nvbios_vpstate_entry(bios, &h, h.boost_id, &boost))
    702 			clk->boost_khz = boost.clock_mhz * 1000;
    703 		if (!nvbios_vpstate_entry(bios, &h, h.base_id, &base))
    704 			clk->base_khz = base.clock_mhz * 1000;
    705 	}
    706 
    707 	clk->func = func;
    708 	INIT_LIST_HEAD(&clk->states);
    709 	clk->domains = func->domains;
    710 	clk->ustate_ac = -1;
    711 	clk->ustate_dc = -1;
    712 	clk->allow_reclock = allow_reclock;
    713 
    714 	INIT_WORK(&clk->work, nvkm_pstate_work);
    715 #ifdef __NetBSD__
    716 	spin_lock_init(&clk->lock);
    717 	DRM_INIT_WAITQUEUE(&clk->wait, "nvclk");
    718 #else
    719 	init_waitqueue_head(&clk->wait);
    720 #endif
    721 	atomic_set(&clk->waiting, 0);
    722 
    723 	/* If no pstates are provided, try and fetch them from the BIOS */
    724 	if (!func->pstates) {
    725 		idx = 0;
    726 		do {
    727 			ret = nvkm_pstate_new(clk, idx++);
    728 		} while (ret == 0);
    729 	} else {
    730 		for (idx = 0; idx < func->nr_pstates; idx++)
    731 			list_add_tail(&func->pstates[idx].head, &clk->states);
    732 		clk->state_nr = func->nr_pstates;
    733 	}
    734 
    735 	ret = nvkm_notify_init(NULL, &device->event, nvkm_clk_pwrsrc, true,
    736 			       NULL, 0, 0, &clk->pwrsrc_ntfy);
    737 	if (ret)
    738 		return ret;
    739 
    740 	mode = nvkm_stropt(device->cfgopt, "NvClkMode", &arglen);
    741 	if (mode) {
    742 		clk->ustate_ac = nvkm_clk_nstate(clk, mode, arglen);
    743 		clk->ustate_dc = nvkm_clk_nstate(clk, mode, arglen);
    744 	}
    745 
    746 	mode = nvkm_stropt(device->cfgopt, "NvClkModeAC", &arglen);
    747 	if (mode)
    748 		clk->ustate_ac = nvkm_clk_nstate(clk, mode, arglen);
    749 
    750 	mode = nvkm_stropt(device->cfgopt, "NvClkModeDC", &arglen);
    751 	if (mode)
    752 		clk->ustate_dc = nvkm_clk_nstate(clk, mode, arglen);
    753 
    754 	clk->boost_mode = nvkm_longopt(device->cfgopt, "NvBoost",
    755 				       NVKM_CLK_BOOST_NONE);
    756 	return 0;
    757 }
    758 
    759 int
    760 nvkm_clk_new_(const struct nvkm_clk_func *func, struct nvkm_device *device,
    761 	      int index, bool allow_reclock, struct nvkm_clk **pclk)
    762 {
    763 	if (!(*pclk = kzalloc(sizeof(**pclk), GFP_KERNEL)))
    764 		return -ENOMEM;
    765 	return nvkm_clk_ctor(func, device, index, allow_reclock, *pclk);
    766 }
    767