blob: ae9522c2258433ed37c5ccdcdb12456d0fd7d94d [file] [log] [blame]
CC Ma7d116dc2015-04-13 14:47:57 +08001/*
2 * Copyright (c) 2013-2015, ARM Limited and Contributors. All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions are met:
6 *
7 * Redistributions of source code must retain the above copyright notice, this
8 * list of conditions and the following disclaimer.
9 *
10 * Redistributions in binary form must reproduce the above copyright notice,
11 * this list of conditions and the following disclaimer in the documentation
12 * and/or other materials provided with the distribution.
13 *
14 * Neither the name of ARM nor the names of its contributors may be used
15 * to endorse or promote products derived from this software without specific
16 * prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
19 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
22 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
28 * POSSIBILITY OF SUCH DAMAGE.
29 */
30
31#include <arch_helpers.h>
32#include <arm_gic.h>
33#include <assert.h>
34#include <bakery_lock.h>
35#include <cci.h>
36#include <console.h>
37#include <debug.h>
38#include <errno.h>
CC Ma7d116dc2015-04-13 14:47:57 +080039#include <mcucfg.h>
40#include <mmio.h>
41#include <mt8173_def.h>
42#include <mt_cpuxgpt.h> /* generic_timer_backup() */
43#include <plat_private.h>
44#include <power_tracer.h>
45#include <psci.h>
46#include <rtc.h>
47#include <scu.h>
48#include <spm_hotplug.h>
49#include <spm_mcdi.h>
50#include <spm_suspend.h>
51
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +080052#if !ENABLE_PLAT_COMPAT
53#define MTK_PWR_LVL0 0
54#define MTK_PWR_LVL1 1
55#define MTK_PWR_LVL2 2
56
57/* Macros to read the MTK power domain state */
58#define MTK_CORE_PWR_STATE(state) (state)->pwr_domain_state[MTK_PWR_LVL0]
59#define MTK_CLUSTER_PWR_STATE(state) (state)->pwr_domain_state[MTK_PWR_LVL1]
60#define MTK_SYSTEM_PWR_STATE(state) ((PLAT_MAX_PWR_LVL > MTK_PWR_LVL1) ?\
61 (state)->pwr_domain_state[MTK_PWR_LVL2] : 0)
62#endif
63
CC Ma7d116dc2015-04-13 14:47:57 +080064struct core_context {
65 unsigned long timer_data[8];
66 unsigned int count;
67 unsigned int rst;
68 unsigned int abt;
69 unsigned int brk;
70};
71
72struct cluster_context {
73 struct core_context core[PLATFORM_MAX_CPUS_PER_CLUSTER];
74};
75
76/*
77 * Top level structure to hold the complete context of a multi cluster system
78 */
79struct system_context {
80 struct cluster_context cluster[PLATFORM_CLUSTER_COUNT];
81};
82
83/*
84 * Top level structure which encapsulates the context of the entire system
85 */
86static struct system_context dormant_data[1];
87
88static inline struct cluster_context *system_cluster(
89 struct system_context *system,
90 uint32_t clusterid)
91{
92 return &system->cluster[clusterid];
93}
94
95static inline struct core_context *cluster_core(struct cluster_context *cluster,
96 uint32_t cpuid)
97{
98 return &cluster->core[cpuid];
99}
100
101static struct cluster_context *get_cluster_data(unsigned long mpidr)
102{
103 uint32_t clusterid;
104
105 clusterid = (mpidr & MPIDR_CLUSTER_MASK) >> MPIDR_AFFINITY_BITS;
106
107 return system_cluster(dormant_data, clusterid);
108}
109
110static struct core_context *get_core_data(unsigned long mpidr)
111{
112 struct cluster_context *cluster;
113 uint32_t cpuid;
114
115 cluster = get_cluster_data(mpidr);
116 cpuid = mpidr & MPIDR_CPU_MASK;
117
118 return cluster_core(cluster, cpuid);
119}
120
121static void mt_save_generic_timer(unsigned long *container)
122{
123 uint64_t ctl;
124 uint64_t val;
125
126 __asm__ volatile("mrs %x0, cntkctl_el1\n\t"
127 "mrs %x1, cntp_cval_el0\n\t"
128 "stp %x0, %x1, [%2, #0]"
129 : "=&r" (ctl), "=&r" (val)
130 : "r" (container)
131 : "memory");
132
133 __asm__ volatile("mrs %x0, cntp_tval_el0\n\t"
134 "mrs %x1, cntp_ctl_el0\n\t"
135 "stp %x0, %x1, [%2, #16]"
136 : "=&r" (val), "=&r" (ctl)
137 : "r" (container)
138 : "memory");
139
140 __asm__ volatile("mrs %x0, cntv_tval_el0\n\t"
141 "mrs %x1, cntv_ctl_el0\n\t"
142 "stp %x0, %x1, [%2, #32]"
143 : "=&r" (val), "=&r" (ctl)
144 : "r" (container)
145 : "memory");
146}
147
148static void mt_restore_generic_timer(unsigned long *container)
149{
150 uint64_t ctl;
151 uint64_t val;
152
153 __asm__ volatile("ldp %x0, %x1, [%2, #0]\n\t"
154 "msr cntkctl_el1, %x0\n\t"
155 "msr cntp_cval_el0, %x1"
156 : "=&r" (ctl), "=&r" (val)
157 : "r" (container)
158 : "memory");
159
160 __asm__ volatile("ldp %x0, %x1, [%2, #16]\n\t"
161 "msr cntp_tval_el0, %x0\n\t"
162 "msr cntp_ctl_el0, %x1"
163 : "=&r" (val), "=&r" (ctl)
164 : "r" (container)
165 : "memory");
166
167 __asm__ volatile("ldp %x0, %x1, [%2, #32]\n\t"
168 "msr cntv_tval_el0, %x0\n\t"
169 "msr cntv_ctl_el0, %x1"
170 : "=&r" (val), "=&r" (ctl)
171 : "r" (container)
172 : "memory");
173}
174
175static inline uint64_t read_cntpctl(void)
176{
177 uint64_t cntpctl;
178
179 __asm__ volatile("mrs %x0, cntp_ctl_el0"
180 : "=r" (cntpctl) : : "memory");
181
182 return cntpctl;
183}
184
185static inline void write_cntpctl(uint64_t cntpctl)
186{
187 __asm__ volatile("msr cntp_ctl_el0, %x0" : : "r"(cntpctl));
188}
189
190static void stop_generic_timer(void)
191{
192 /*
193 * Disable the timer and mask the irq to prevent
194 * suprious interrupts on this cpu interface. It
195 * will bite us when we come back if we don't. It
196 * will be replayed on the inbound cluster.
197 */
198 uint64_t cntpctl = read_cntpctl();
199
200 write_cntpctl(clr_cntp_ctl_enable(cntpctl));
201}
202
203static void mt_cpu_save(unsigned long mpidr)
204{
205 struct core_context *core;
206
207 core = get_core_data(mpidr);
208 mt_save_generic_timer(core->timer_data);
209
210 /* disable timer irq, and upper layer should enable it again. */
211 stop_generic_timer();
212}
213
214static void mt_cpu_restore(unsigned long mpidr)
215{
216 struct core_context *core;
217
218 core = get_core_data(mpidr);
219 mt_restore_generic_timer(core->timer_data);
220}
221
222static void mt_platform_save_context(unsigned long mpidr)
223{
224 /* mcusys_save_context: */
225 mt_cpu_save(mpidr);
226}
227
228static void mt_platform_restore_context(unsigned long mpidr)
229{
230 /* mcusys_restore_context: */
231 mt_cpu_restore(mpidr);
232}
233
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800234#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800235/*******************************************************************************
236* Private function which is used to determine if any platform actions
237* should be performed for the specified affinity instance given its
238* state. Nothing needs to be done if the 'state' is not off or if this is not
239* the highest affinity level which will enter the 'state'.
240*******************************************************************************/
241static int32_t plat_do_plat_actions(unsigned int afflvl, unsigned int state)
242{
243 unsigned int max_phys_off_afflvl;
244
245 assert(afflvl <= MPIDR_AFFLVL2);
246
247 if (state != PSCI_STATE_OFF)
248 return -EAGAIN;
249
250 /*
251 * Find the highest affinity level which will be suspended and postpone
252 * all the platform specific actions until that level is hit.
253 */
254 max_phys_off_afflvl = psci_get_max_phys_off_afflvl();
255 assert(max_phys_off_afflvl != PSCI_INVALID_DATA);
256 if (afflvl != max_phys_off_afflvl)
257 return -EAGAIN;
258
259 return 0;
260}
261
262/*******************************************************************************
263 * MTK_platform handler called when an affinity instance is about to enter
264 * standby.
265 ******************************************************************************/
266static void plat_affinst_standby(unsigned int power_state)
267{
268 unsigned int target_afflvl;
269
270 /* Sanity check the requested state */
271 target_afflvl = psci_get_pstate_afflvl(power_state);
272
273 /*
274 * It's possible to enter standby only on affinity level 0 i.e. a cpu
275 * on the MTK_platform. Ignore any other affinity level.
276 */
277 if (target_afflvl == MPIDR_AFFLVL0) {
278 /*
279 * Enter standby state. dsb is good practice before using wfi
280 * to enter low power states.
281 */
282 dsb();
283 wfi();
284 }
285}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800286#else
287static void plat_cpu_standby(plat_local_state_t cpu_state)
288{
289 unsigned int scr;
290
291 scr = read_scr_el3();
292 write_scr_el3(scr | SCR_IRQ_BIT);
293 isb();
294 dsb();
295 wfi();
296 write_scr_el3(scr);
297}
298#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800299
300/*******************************************************************************
301 * MTK_platform handler called when an affinity instance is about to be turned
302 * on. The level and mpidr determine the affinity instance.
303 ******************************************************************************/
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800304#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800305static int plat_affinst_on(unsigned long mpidr,
306 unsigned long sec_entrypoint,
307 unsigned int afflvl,
308 unsigned int state)
309{
310 int rc = PSCI_E_SUCCESS;
311 unsigned long cpu_id;
312 unsigned long cluster_id;
313 uintptr_t rv;
314
315 /*
316 * It's possible to turn on only affinity level 0 i.e. a cpu
317 * on the MTK_platform. Ignore any other affinity level.
318 */
319 if (afflvl != MPIDR_AFFLVL0)
320 return rc;
321
322 cpu_id = mpidr & MPIDR_CPU_MASK;
323 cluster_id = mpidr & MPIDR_CLUSTER_MASK;
324
325 if (cluster_id)
326 rv = (uintptr_t)&mt8173_mcucfg->mp1_rv_addr[cpu_id].rv_addr_lw;
327 else
328 rv = (uintptr_t)&mt8173_mcucfg->mp0_rv_addr[cpu_id].rv_addr_lw;
329
330 mmio_write_32(rv, sec_entrypoint);
331 INFO("mt_on[%ld:%ld], entry %x\n",
332 cluster_id, cpu_id, mmio_read_32(rv));
333
334 spm_hotplug_on(mpidr);
335
336 return rc;
337}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800338#else
339static uintptr_t secure_entrypoint;
340
341static int plat_power_domain_on(unsigned long mpidr)
342{
343 int rc = PSCI_E_SUCCESS;
344 unsigned long cpu_id;
345 unsigned long cluster_id;
346 uintptr_t rv;
347
348 cpu_id = mpidr & MPIDR_CPU_MASK;
349 cluster_id = mpidr & MPIDR_CLUSTER_MASK;
350
351 if (cluster_id)
352 rv = (uintptr_t)&mt8173_mcucfg->mp1_rv_addr[cpu_id].rv_addr_lw;
353 else
354 rv = (uintptr_t)&mt8173_mcucfg->mp0_rv_addr[cpu_id].rv_addr_lw;
355
356 mmio_write_32(rv, secure_entrypoint);
357 INFO("mt_on[%ld:%ld], entry %x\n",
358 cluster_id, cpu_id, mmio_read_32(rv));
359
360 spm_hotplug_on(mpidr);
361 return rc;
362}
363#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800364
365/*******************************************************************************
366 * MTK_platform handler called when an affinity instance is about to be turned
367 * off. The level and mpidr determine the affinity instance. The 'state' arg.
368 * allows the platform to decide whether the cluster is being turned off and
369 * take apt actions.
370 *
371 * CAUTION: This function is called with coherent stacks so that caches can be
372 * turned off, flushed and coherency disabled. There is no guarantee that caches
373 * will remain turned on across calls to this function as each affinity level is
374 * dealt with. So do not write & read global variables across calls. It will be
375 * wise to do flush a write to the global to prevent unpredictable results.
376 ******************************************************************************/
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800377#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800378static void plat_affinst_off(unsigned int afflvl, unsigned int state)
379{
380 unsigned long mpidr = read_mpidr_el1();
381
382 /* Determine if any platform actions need to be executed. */
383 if (plat_do_plat_actions(afflvl, state) == -EAGAIN)
384 return;
385
386 /* Prevent interrupts from spuriously waking up this cpu */
387 arm_gic_cpuif_deactivate();
388
389 spm_hotplug_off(mpidr);
390
391 trace_power_flow(mpidr, CPU_DOWN);
392
393 if (afflvl != MPIDR_AFFLVL0) {
394 /* Disable coherency if this cluster is to be turned off */
395 plat_cci_disable();
396
397 trace_power_flow(mpidr, CLUSTER_DOWN);
398 }
399}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800400#else
401static void plat_power_domain_off(const psci_power_state_t *state)
402{
403 unsigned long mpidr = read_mpidr_el1();
404
405 /* Prevent interrupts from spuriously waking up this cpu */
406 arm_gic_cpuif_deactivate();
407
408 spm_hotplug_off(mpidr);
409
410 trace_power_flow(mpidr, CPU_DOWN);
411
412 if (MTK_CLUSTER_PWR_STATE(state) == MTK_LOCAL_STATE_OFF) {
413 /* Disable coherency if this cluster is to be turned off */
414 plat_cci_disable();
415
416 trace_power_flow(mpidr, CLUSTER_DOWN);
417 }
418}
419#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800420
421/*******************************************************************************
422 * MTK_platform handler called when an affinity instance is about to be
423 * suspended. The level and mpidr determine the affinity instance. The 'state'
424 * arg. allows the platform to decide whether the cluster is being turned off
425 * and take apt actions.
426 *
427 * CAUTION: This function is called with coherent stacks so that caches can be
428 * turned off, flushed and coherency disabled. There is no guarantee that caches
429 * will remain turned on across calls to this function as each affinity level is
430 * dealt with. So do not write & read global variables across calls. It will be
431 * wise to do flush a write to the global to prevent unpredictable results.
432 ******************************************************************************/
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800433#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800434static void plat_affinst_suspend(unsigned long sec_entrypoint,
435 unsigned int afflvl,
436 unsigned int state)
437{
438 unsigned long mpidr = read_mpidr_el1();
439 unsigned long cluster_id;
440 unsigned long cpu_id;
441 uintptr_t rv;
442
443 /* Determine if any platform actions need to be executed. */
444 if (plat_do_plat_actions(afflvl, state) == -EAGAIN)
445 return;
446
447 cpu_id = mpidr & MPIDR_CPU_MASK;
448 cluster_id = mpidr & MPIDR_CLUSTER_MASK;
449
450 if (cluster_id)
451 rv = (uintptr_t)&mt8173_mcucfg->mp1_rv_addr[cpu_id].rv_addr_lw;
452 else
453 rv = (uintptr_t)&mt8173_mcucfg->mp0_rv_addr[cpu_id].rv_addr_lw;
454
455 mmio_write_32(rv, sec_entrypoint);
456
Jimmy Huang8e53ec52015-11-16 14:33:35 +0800457 if (afflvl < MPIDR_AFFLVL2)
458 spm_mcdi_prepare_for_off_state(mpidr, afflvl);
CC Ma7d116dc2015-04-13 14:47:57 +0800459
460 if (afflvl >= MPIDR_AFFLVL0)
461 mt_platform_save_context(mpidr);
462
463 /* Perform the common cluster specific operations */
464 if (afflvl >= MPIDR_AFFLVL1) {
465 /* Disable coherency if this cluster is to be turned off */
466 plat_cci_disable();
CC Ma7d116dc2015-04-13 14:47:57 +0800467 }
468
469 if (afflvl >= MPIDR_AFFLVL2) {
Jimmy Huang8e53ec52015-11-16 14:33:35 +0800470 disable_scu(mpidr);
CC Ma7d116dc2015-04-13 14:47:57 +0800471 generic_timer_backup();
472 spm_system_suspend();
473 /* Prevent interrupts from spuriously waking up this cpu */
474 arm_gic_cpuif_deactivate();
475 }
476}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800477#else
478static void plat_power_domain_suspend(const psci_power_state_t *state)
479{
480 unsigned long mpidr = read_mpidr_el1();
481 unsigned long cluster_id;
482 unsigned long cpu_id;
483 uintptr_t rv;
484
485 cpu_id = mpidr & MPIDR_CPU_MASK;
486 cluster_id = mpidr & MPIDR_CLUSTER_MASK;
487
488 if (cluster_id)
489 rv = (uintptr_t)&mt8173_mcucfg->mp1_rv_addr[cpu_id].rv_addr_lw;
490 else
491 rv = (uintptr_t)&mt8173_mcucfg->mp0_rv_addr[cpu_id].rv_addr_lw;
492
493 mmio_write_32(rv, secure_entrypoint);
494
495 if (MTK_SYSTEM_PWR_STATE(state) != MTK_LOCAL_STATE_OFF) {
496 spm_mcdi_prepare_for_off_state(mpidr, MTK_PWR_LVL0);
497 if (MTK_CLUSTER_PWR_STATE(state) == MTK_LOCAL_STATE_OFF)
498 spm_mcdi_prepare_for_off_state(mpidr, MTK_PWR_LVL1);
499 }
500
501 mt_platform_save_context(mpidr);
502
503 /* Perform the common cluster specific operations */
504 if (MTK_CLUSTER_PWR_STATE(state) == MTK_LOCAL_STATE_OFF) {
505 /* Disable coherency if this cluster is to be turned off */
506 plat_cci_disable();
507 }
508
509 if (MTK_SYSTEM_PWR_STATE(state) == MTK_LOCAL_STATE_OFF) {
510 disable_scu(mpidr);
511 generic_timer_backup();
512 spm_system_suspend();
513 /* Prevent interrupts from spuriously waking up this cpu */
514 arm_gic_cpuif_deactivate();
515 }
516}
517#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800518
519/*******************************************************************************
520 * MTK_platform handler called when an affinity instance has just been powered
521 * on after being turned off earlier. The level and mpidr determine the affinity
522 * instance. The 'state' arg. allows the platform to decide whether the cluster
523 * was turned off prior to wakeup and do what's necessary to setup it up
524 * correctly.
525 ******************************************************************************/
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800526#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800527static void plat_affinst_on_finish(unsigned int afflvl, unsigned int state)
528{
529 unsigned long mpidr = read_mpidr_el1();
530
531 /* Determine if any platform actions need to be executed. */
532 if (plat_do_plat_actions(afflvl, state) == -EAGAIN)
533 return;
534
535 /* Perform the common cluster specific operations */
536 if (afflvl >= MPIDR_AFFLVL1) {
CC Ma7d116dc2015-04-13 14:47:57 +0800537 /* Enable coherency if this cluster was off */
538 plat_cci_enable();
539 trace_power_flow(mpidr, CLUSTER_UP);
540 }
541
542 /* Enable the gic cpu interface */
543 arm_gic_cpuif_setup();
544 arm_gic_pcpu_distif_setup();
545 trace_power_flow(mpidr, CPU_UP);
546}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800547#else
548void mtk_system_pwr_domain_resume(void);
549
550static void plat_power_domain_on_finish(const psci_power_state_t *state)
551{
552 unsigned long mpidr = read_mpidr_el1();
553
554 assert(state->pwr_domain_state[MPIDR_AFFLVL0] == MTK_LOCAL_STATE_OFF);
555
556 if ((PLAT_MAX_PWR_LVL > MTK_PWR_LVL1) &&
557 (state->pwr_domain_state[MTK_PWR_LVL2] == MTK_LOCAL_STATE_OFF))
558 mtk_system_pwr_domain_resume();
559
560 if (state->pwr_domain_state[MPIDR_AFFLVL1] == MTK_LOCAL_STATE_OFF) {
561 plat_cci_enable();
562 trace_power_flow(mpidr, CLUSTER_UP);
563 }
564
565 if ((PLAT_MAX_PWR_LVL > MTK_PWR_LVL1) &&
566 (state->pwr_domain_state[MTK_PWR_LVL2] == MTK_LOCAL_STATE_OFF))
567 return;
568
569 /* Enable the gic cpu interface */
570 arm_gic_cpuif_setup();
571 arm_gic_pcpu_distif_setup();
572 trace_power_flow(mpidr, CPU_UP);
573}
574#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800575
576/*******************************************************************************
577 * MTK_platform handler called when an affinity instance has just been powered
578 * on after having been suspended earlier. The level and mpidr determine the
579 * affinity instance.
580 ******************************************************************************/
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800581#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800582static void plat_affinst_suspend_finish(unsigned int afflvl, unsigned int state)
583{
584 unsigned long mpidr = read_mpidr_el1();
585
586 /* Determine if any platform actions need to be executed. */
587 if (plat_do_plat_actions(afflvl, state) == -EAGAIN)
588 return;
589
590 if (afflvl >= MPIDR_AFFLVL2) {
591 /* Enable the gic cpu interface */
592 arm_gic_setup();
593 arm_gic_cpuif_setup();
594 spm_system_suspend_finish();
Jimmy Huang8e53ec52015-11-16 14:33:35 +0800595 enable_scu(mpidr);
CC Ma7d116dc2015-04-13 14:47:57 +0800596 }
597
598 /* Perform the common cluster specific operations */
599 if (afflvl >= MPIDR_AFFLVL1) {
CC Ma7d116dc2015-04-13 14:47:57 +0800600 /* Enable coherency if this cluster was off */
601 plat_cci_enable();
CC Ma7d116dc2015-04-13 14:47:57 +0800602 }
603
604 if (afflvl >= MPIDR_AFFLVL0)
605 mt_platform_restore_context(mpidr);
606
Jimmy Huang8e53ec52015-11-16 14:33:35 +0800607 if (afflvl < MPIDR_AFFLVL2)
608 spm_mcdi_finish_for_on_state(mpidr, afflvl);
CC Ma7d116dc2015-04-13 14:47:57 +0800609
610 arm_gic_pcpu_distif_setup();
611}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800612#else
613static void plat_power_domain_suspend_finish(const psci_power_state_t *state)
614{
615 unsigned long mpidr = read_mpidr_el1();
CC Ma7d116dc2015-04-13 14:47:57 +0800616
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800617 if (state->pwr_domain_state[MTK_PWR_LVL0] == MTK_LOCAL_STATE_RET)
618 return;
619
620 if (MTK_SYSTEM_PWR_STATE(state) == MTK_LOCAL_STATE_OFF) {
621 /* Enable the gic cpu interface */
622 arm_gic_setup();
623 arm_gic_cpuif_setup();
624 spm_system_suspend_finish();
625 enable_scu(mpidr);
626 }
627
628 /* Perform the common cluster specific operations */
629 if (MTK_CLUSTER_PWR_STATE(state) == MTK_LOCAL_STATE_OFF) {
630 /* Enable coherency if this cluster was off */
631 plat_cci_enable();
632 }
633
634 mt_platform_restore_context(mpidr);
635
636 if (MTK_SYSTEM_PWR_STATE(state) != MTK_LOCAL_STATE_OFF) {
637 spm_mcdi_finish_for_on_state(mpidr, MTK_PWR_LVL0);
638 if (MTK_CLUSTER_PWR_STATE(state) == MTK_LOCAL_STATE_OFF)
639 spm_mcdi_finish_for_on_state(mpidr, MTK_PWR_LVL1);
640 }
641
642 arm_gic_pcpu_distif_setup();
643}
644#endif
645
646#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800647static unsigned int plat_get_sys_suspend_power_state(void)
648{
649 /* StateID: 0, StateType: 1(power down), PowerLevel: 2(system) */
650 return psci_make_powerstate(0, 1, 2);
651}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800652#else
653static void plat_get_sys_suspend_power_state(psci_power_state_t *req_state)
654{
655 assert(PLAT_MAX_PWR_LVL >= 2);
656
657 for (int i = MPIDR_AFFLVL0; i <= PLAT_MAX_PWR_LVL; i++)
658 req_state->pwr_domain_state[i] = MTK_LOCAL_STATE_OFF;
659}
660#endif
CC Ma7d116dc2015-04-13 14:47:57 +0800661
662/*******************************************************************************
663 * MTK handlers to shutdown/reboot the system
664 ******************************************************************************/
665static void __dead2 plat_system_off(void)
666{
667 INFO("MTK System Off\n");
668
CC Ma7d116dc2015-04-13 14:47:57 +0800669 rtc_bbpu_power_down();
670
671 wfi();
672 ERROR("MTK System Off: operation not handled.\n");
673 panic();
674}
675
676static void __dead2 plat_system_reset(void)
677{
678 /* Write the System Configuration Control Register */
679 INFO("MTK System Reset\n");
680
Jimmy Huang2bab3d52015-09-04 10:30:57 +0800681 mmio_clrsetbits_32(MTK_WDT_BASE,
682 (MTK_WDT_MODE_DUAL_MODE | MTK_WDT_MODE_IRQ),
683 MTK_WDT_MODE_KEY);
CC Ma7d116dc2015-04-13 14:47:57 +0800684 mmio_setbits_32(MTK_WDT_BASE, (MTK_WDT_MODE_KEY | MTK_WDT_MODE_EXTEN));
685 mmio_setbits_32(MTK_WDT_SWRST, MTK_WDT_SWRST_KEY);
686
687 wfi();
688 ERROR("MTK System Reset: operation not handled.\n");
689 panic();
690}
691
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800692#if !ENABLE_PLAT_COMPAT
693static int plat_validate_power_state(unsigned int power_state,
694 psci_power_state_t *req_state)
695{
696 int pstate = psci_get_pstate_type(power_state);
697 int pwr_lvl = psci_get_pstate_pwrlvl(power_state);
698 int i;
699
700 assert(req_state);
701
702 if (pwr_lvl > PLAT_MAX_PWR_LVL)
703 return PSCI_E_INVALID_PARAMS;
704
705 /* Sanity check the requested state */
706 if (pstate == PSTATE_TYPE_STANDBY) {
707 /*
708 * It's possible to enter standby only on power level 0
709 * Ignore any other power level.
710 */
711 if (pwr_lvl != 0)
712 return PSCI_E_INVALID_PARAMS;
713
714 req_state->pwr_domain_state[MTK_PWR_LVL0] =
715 MTK_LOCAL_STATE_RET;
716 } else {
717 for (i = 0; i <= pwr_lvl; i++)
718 req_state->pwr_domain_state[i] =
719 MTK_LOCAL_STATE_OFF;
720 }
721
722 /*
723 * We expect the 'state id' to be zero.
724 */
725 if (psci_get_pstate_id(power_state))
726 return PSCI_E_INVALID_PARAMS;
727
728 return PSCI_E_SUCCESS;
729}
730
731void mtk_system_pwr_domain_resume(void)
732{
733 console_init(MT8173_UART0_BASE, MT8173_UART_CLOCK, MT8173_BAUDRATE);
734
735 /* Assert system power domain is available on the platform */
736 assert(PLAT_MAX_PWR_LVL >= MTK_PWR_LVL2);
737
738 arm_gic_cpuif_setup();
739 arm_gic_pcpu_distif_setup();
740}
741#endif
742
743#if ENABLE_PLAT_COMPAT
CC Ma7d116dc2015-04-13 14:47:57 +0800744/*******************************************************************************
745 * Export the platform handlers to enable psci to invoke them
746 ******************************************************************************/
747static const plat_pm_ops_t plat_plat_pm_ops = {
748 .affinst_standby = plat_affinst_standby,
749 .affinst_on = plat_affinst_on,
750 .affinst_off = plat_affinst_off,
751 .affinst_suspend = plat_affinst_suspend,
752 .affinst_on_finish = plat_affinst_on_finish,
753 .affinst_suspend_finish = plat_affinst_suspend_finish,
754 .system_off = plat_system_off,
755 .system_reset = plat_system_reset,
756 .get_sys_suspend_power_state = plat_get_sys_suspend_power_state,
757};
758
759/*******************************************************************************
760 * Export the platform specific power ops & initialize the mtk_platform power
761 * controller
762 ******************************************************************************/
763int platform_setup_pm(const plat_pm_ops_t **plat_ops)
764{
765 *plat_ops = &plat_plat_pm_ops;
766 return 0;
767}
Koan-Sin Tan3fc26aa2017-01-19 16:43:49 +0800768#else
769static const plat_psci_ops_t plat_plat_pm_ops = {
770 .cpu_standby = plat_cpu_standby,
771 .pwr_domain_on = plat_power_domain_on,
772 .pwr_domain_on_finish = plat_power_domain_on_finish,
773 .pwr_domain_off = plat_power_domain_off,
774 .pwr_domain_suspend = plat_power_domain_suspend,
775 .pwr_domain_suspend_finish = plat_power_domain_suspend_finish,
776 .system_off = plat_system_off,
777 .system_reset = plat_system_reset,
778 .validate_power_state = plat_validate_power_state,
779 .get_sys_suspend_power_state = plat_get_sys_suspend_power_state,
780};
781
782int plat_setup_psci_ops(uintptr_t sec_entrypoint,
783 const plat_psci_ops_t **psci_ops)
784{
785 *psci_ops = &plat_plat_pm_ops;
786 secure_entrypoint = sec_entrypoint;
787 return 0;
788}
789
790/*
791 * The PSCI generic code uses this API to let the platform participate in state
792 * coordination during a power management operation. It compares the platform
793 * specific local power states requested by each cpu for a given power domain
794 * and returns the coordinated target power state that the domain should
795 * enter. A platform assigns a number to a local power state. This default
796 * implementation assumes that the platform assigns these numbers in order of
797 * increasing depth of the power state i.e. for two power states X & Y, if X < Y
798 * then X represents a shallower power state than Y. As a result, the
799 * coordinated target local power state for a power domain will be the minimum
800 * of the requested local power states.
801 */
802plat_local_state_t plat_get_target_pwr_state(unsigned int lvl,
803 const plat_local_state_t *states,
804 unsigned int ncpu)
805{
806 plat_local_state_t target = PLAT_MAX_OFF_STATE, temp;
807
808 assert(ncpu);
809
810 do {
811 temp = *states++;
812 if (temp < target)
813 target = temp;
814 } while (--ncpu);
815
816 return target;
817}
818#endif