Lines Matching refs:cf
64 cpufreq_register(struct cpufreq *cf)
72 KASSERT(cf != NULL);
74 KASSERT(cf->cf_get_freq != NULL);
75 KASSERT(cf->cf_set_freq != NULL);
76 KASSERT(cf->cf_state_count > 0);
77 KASSERT(cf->cf_state_count < CPUFREQ_STATE_MAX);
87 cf_backend->cf_mp = cf->cf_mp;
88 cf_backend->cf_cookie = cf->cf_cookie;
89 cf_backend->cf_get_freq = cf->cf_get_freq;
90 cf_backend->cf_set_freq = cf->cf_set_freq;
92 (void)strlcpy(cf_backend->cf_name, cf->cf_name, sizeof(cf->cf_name));
97 for (c = i = 0; i < cf->cf_state_count; i++) {
102 if (cf->cf_state[i].cfs_freq == 0)
105 if (cf->cf_state[i].cfs_freq > 9999 &&
106 cf->cf_state[i].cfs_freq != CPUFREQ_STATE_ENABLED &&
107 cf->cf_state[i].cfs_freq != CPUFREQ_STATE_DISABLED)
112 if (cf->cf_state[i].cfs_freq >=
113 cf->cf_state[j].cfs_freq) {
123 cf_backend->cf_state[c].cfs_freq = cf->cf_state[i].cfs_freq;
124 cf_backend->cf_state[c].cfs_power = cf->cf_state[i].cfs_power;
164 struct cpufreq *cf = cf_backend;
177 for (i = 0; i < cf->cf_state_count; i++) {
192 cpufreq_set_all_raw(cf->cf_state[i].cfs_freq);
214 cf->cf_state[i].cfs_latency = s / n;
223 struct cpufreq *cf = cf_backend;
228 if (cf->cf_init != true) {
237 cf->cf_state_saved = s;
245 struct cpufreq *cf = cf_backend;
249 if (cf->cf_init != true || cf->cf_state_saved == 0) {
254 cpufreq_set_raw(ci, cf->cf_state_saved);
261 struct cpufreq *cf = cf_backend;
266 if (cf->cf_init != true) {
280 struct cpufreq *cf = cf_backend;
282 KASSERT(cf->cf_init != false);
285 return cf->cf_state[0].cfs_freq;
291 struct cpufreq *cf = cf_backend;
293 KASSERT(cf->cf_init != false);
296 return cf->cf_state[cf->cf_state_count - 1].cfs_freq;
302 struct cpufreq *cf = cf_backend;
306 KASSERT(cf->cf_init != false);
309 xc = xc_unicast(0, (*cf->cf_get_freq), cf->cf_cookie, &freq, ci);
318 struct cpufreq *cf = cf_backend;
322 if (cf->cf_init != true || dst == NULL) {
327 memcpy(dst, cf, sizeof(*cf));
336 struct cpufreq *cf = cf_backend;
340 if (cf->cf_init != true || cfs == NULL) {
354 struct cpufreq *cf = cf_backend;
358 if (cf->cf_init != true || cfs == NULL) {
363 if (index >= cf->cf_state_count) {
368 memcpy(cfs, &cf->cf_state[index], sizeof(*cfs));
377 struct cpufreq *cf = cf_backend;
381 KASSERT(cf->cf_init != false);
384 hi = cf->cf_state_count;
389 f = cf->cf_state[i].cfs_freq;
400 memcpy(cfs, &cf->cf_state[i], sizeof(*cfs));
406 struct cpufreq *cf = cf_backend;
410 if (__predict_false(cf->cf_init != true)) {
422 struct cpufreq *cf = cf_backend;
425 KASSERT(cf->cf_init != false);
428 xc = xc_unicast(0, (*cf->cf_set_freq), cf->cf_cookie, &freq, ci);
435 struct cpufreq *cf = cf_backend;
439 if (__predict_false(cf->cf_init != true)) {
451 struct cpufreq *cf = cf_backend;
454 KASSERT(cf->cf_init != false);
457 xc = xc_broadcast(0, (*cf->cf_set_freq), cf->cf_cookie, &freq);
477 struct cpufreq *cf = cf_backend;
484 if (__predict_false(cf->cf_init != true)) {
499 if (index < 0 || index >= (int32_t)cf->cf_state_count) {
504 cpufreq_set_raw(ci, cf->cf_state[index].cfs_freq);