blob: 69e3e0b556bf7fd9bcd3533c3e7c2502e8224b45 [file] [log] [blame]
Thomas Gleixner2874c5f2019-05-27 08:55:01 +02001// SPDX-License-Identifier: GPL-2.0-or-later
Paul Burton9c38cf42014-01-15 10:31:52 +00002/*
3 * Copyright (C) 2013 Imagination Technologies
Paul Burtonfb615d62017-10-25 17:04:33 -07004 * Author: Paul Burton <paul.burton@mips.com>
Paul Burton9c38cf42014-01-15 10:31:52 +00005 */
6
7#include <linux/errno.h>
Paul Burton245a7862014-04-14 12:04:27 +01008#include <linux/percpu.h>
Paul Burton791412d2018-01-19 16:40:49 +01009#include <linux/of.h>
10#include <linux/of_address.h>
Paul Burton245a7862014-04-14 12:04:27 +010011#include <linux/spinlock.h>
Paul Burton9c38cf42014-01-15 10:31:52 +000012
Paul Burtone83f7e02017-08-12 19:49:41 -070013#include <asm/mips-cps.h>
Paul Burton9c38cf42014-01-15 10:31:52 +000014
15void __iomem *mips_cpc_base;
16
Paul Burton76ae6582014-02-14 09:28:06 +000017static DEFINE_PER_CPU_ALIGNED(spinlock_t, cpc_core_lock);
18
19static DEFINE_PER_CPU_ALIGNED(unsigned long, cpc_core_lock_flags);
20
Paul Burton682c1e52016-10-15 23:03:43 +010021phys_addr_t __weak mips_cpc_default_phys_base(void)
22{
Paul Burton791412d2018-01-19 16:40:49 +010023 struct device_node *cpc_node;
24 struct resource res;
25 int err;
26
27 cpc_node = of_find_compatible_node(of_root, NULL, "mti,mips-cpc");
28 if (cpc_node) {
29 err = of_address_to_resource(cpc_node, 0, &res);
30 if (!err)
31 return res.start;
32 }
33
Paul Burton682c1e52016-10-15 23:03:43 +010034 return 0;
35}
36
Bjorn Helgaas8dedde62015-07-12 18:10:56 -050037/**
38 * mips_cpc_phys_base - retrieve the physical base address of the CPC
39 *
40 * This function returns the physical base address of the Cluster Power
41 * Controller memory mapped registers, or 0 if no Cluster Power Controller
42 * is present.
43 */
44static phys_addr_t mips_cpc_phys_base(void)
Paul Burton9c38cf42014-01-15 10:31:52 +000045{
Markos Chandras391057d2015-07-09 10:40:46 +010046 unsigned long cpc_base;
Paul Burton9c38cf42014-01-15 10:31:52 +000047
48 if (!mips_cm_present())
49 return 0;
50
Paul Burton93c5bba52017-08-12 19:49:27 -070051 if (!(read_gcr_cpc_status() & CM_GCR_CPC_STATUS_EX))
Paul Burton9c38cf42014-01-15 10:31:52 +000052 return 0;
53
54 /* If the CPC is already enabled, leave it so */
55 cpc_base = read_gcr_cpc_base();
Paul Burton93c5bba52017-08-12 19:49:27 -070056 if (cpc_base & CM_GCR_CPC_BASE_CPCEN)
57 return cpc_base & CM_GCR_CPC_BASE_CPCBASE;
Paul Burton9c38cf42014-01-15 10:31:52 +000058
Paul Burton682c1e52016-10-15 23:03:43 +010059 /* Otherwise, use the default address */
Paul Burton9c38cf42014-01-15 10:31:52 +000060 cpc_base = mips_cpc_default_phys_base();
Paul Burton682c1e52016-10-15 23:03:43 +010061 if (!cpc_base)
62 return cpc_base;
63
64 /* Enable the CPC, mapped at the default address */
Paul Burton93c5bba52017-08-12 19:49:27 -070065 write_gcr_cpc_base(cpc_base | CM_GCR_CPC_BASE_CPCEN);
Paul Burton9c38cf42014-01-15 10:31:52 +000066 return cpc_base;
67}
68
69int mips_cpc_probe(void)
70{
Ralf Baechle15d45cc2014-11-22 00:22:09 +010071 phys_addr_t addr;
Matt Redfearn6b89d222016-09-07 10:45:09 +010072 unsigned int cpu;
Paul Burton76ae6582014-02-14 09:28:06 +000073
74 for_each_possible_cpu(cpu)
75 spin_lock_init(&per_cpu(cpc_core_lock, cpu));
Paul Burton9c38cf42014-01-15 10:31:52 +000076
77 addr = mips_cpc_phys_base();
78 if (!addr)
79 return -ENODEV;
80
81 mips_cpc_base = ioremap_nocache(addr, 0x8000);
82 if (!mips_cpc_base)
83 return -ENXIO;
84
85 return 0;
86}
Paul Burton76ae6582014-02-14 09:28:06 +000087
88void mips_cpc_lock_other(unsigned int core)
89{
Matt Redfearn6b89d222016-09-07 10:45:09 +010090 unsigned int curr_core;
Matt Redfearnd6219422016-09-07 10:45:10 +010091
92 if (mips_cm_revision() >= CM_REV_CM3)
93 /* Systems with CM >= 3 lock the CPC via mips_cm_lock_other */
94 return;
95
Paul Burton76ae6582014-02-14 09:28:06 +000096 preempt_disable();
Paul Burtonf875a8322017-08-12 19:49:35 -070097 curr_core = cpu_core(&current_cpu_data);
Paul Burton76ae6582014-02-14 09:28:06 +000098 spin_lock_irqsave(&per_cpu(cpc_core_lock, curr_core),
99 per_cpu(cpc_core_lock_flags, curr_core));
Paul Burton829ca2be2017-08-12 19:49:29 -0700100 write_cpc_cl_other(core << __ffs(CPC_Cx_OTHER_CORENUM));
Paul Burton78a54c42015-09-22 11:12:18 -0700101
102 /*
103 * Ensure the core-other region reflects the appropriate core &
104 * VP before any accesses to it occur.
105 */
106 mb();
Paul Burton76ae6582014-02-14 09:28:06 +0000107}
108
109void mips_cpc_unlock_other(void)
110{
Matt Redfearnd6219422016-09-07 10:45:10 +0100111 unsigned int curr_core;
112
113 if (mips_cm_revision() >= CM_REV_CM3)
114 /* Systems with CM >= 3 lock the CPC via mips_cm_lock_other */
115 return;
116
Paul Burtonf875a8322017-08-12 19:49:35 -0700117 curr_core = cpu_core(&current_cpu_data);
Paul Burton76ae6582014-02-14 09:28:06 +0000118 spin_unlock_irqrestore(&per_cpu(cpc_core_lock, curr_core),
119 per_cpu(cpc_core_lock_flags, curr_core));
120 preempt_enable();
121}