xref: /freebsd/sys/powerpc/aim/mp_cpudep.c (revision a0ee8cc6)
1 /*-
2  * Copyright (c) 2008 Marcel Moolenaar
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  *
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25  */
26 
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29 
30 #include <sys/param.h>
31 #include <sys/systm.h>
32 #include <sys/kernel.h>
33 #include <sys/bus.h>
34 #include <sys/pcpu.h>
35 #include <sys/proc.h>
36 #include <sys/smp.h>
37 
38 #include <machine/bus.h>
39 #include <machine/cpu.h>
40 #include <machine/hid.h>
41 #include <machine/intr_machdep.h>
42 #include <machine/pcb.h>
43 #include <machine/psl.h>
44 #include <machine/smp.h>
45 #include <machine/spr.h>
46 #include <machine/trap.h>
47 
48 #include <dev/ofw/openfirm.h>
49 #include <machine/ofw_machdep.h>
50 
51 void *ap_pcpu;
52 
53 static register_t bsp_state[8] __aligned(8);
54 
55 static void cpudep_save_config(void *dummy);
56 SYSINIT(cpu_save_config, SI_SUB_CPU, SI_ORDER_ANY, cpudep_save_config, NULL);
57 
58 void
59 cpudep_ap_early_bootstrap(void)
60 {
61 #ifndef __powerpc64__
62 	register_t reg;
63 #endif
64 
65 	__asm __volatile("mtsprg 0, %0" :: "r"(ap_pcpu));
66 	powerpc_sync();
67 
68 	switch (mfpvr() >> 16) {
69 	case IBM970:
70 	case IBM970FX:
71 	case IBM970MP:
72 		/* Restore HID4 and HID5, which are necessary for the MMU */
73 
74 #ifdef __powerpc64__
75 		mtspr(SPR_HID4, bsp_state[2]); powerpc_sync(); isync();
76 		mtspr(SPR_HID5, bsp_state[3]); powerpc_sync(); isync();
77 #else
78 		__asm __volatile("ld %0, 16(%2); sync; isync;	\
79 		    mtspr %1, %0; sync; isync;"
80 		    : "=r"(reg) : "K"(SPR_HID4), "b"(bsp_state));
81 		__asm __volatile("ld %0, 24(%2); sync; isync;	\
82 		    mtspr %1, %0; sync; isync;"
83 		    : "=r"(reg) : "K"(SPR_HID5), "b"(bsp_state));
84 #endif
85 		powerpc_sync();
86 		break;
87 	}
88 }
89 
90 uintptr_t
91 cpudep_ap_bootstrap(void)
92 {
93 	register_t msr, sp;
94 
95 	msr = PSL_KERNSET & ~PSL_EE;
96 	mtmsr(msr);
97 
98 	pcpup->pc_curthread = pcpup->pc_idlethread;
99 #ifdef __powerpc64__
100 	__asm __volatile("mr 13,%0" :: "r"(pcpup->pc_curthread));
101 #else
102 	__asm __volatile("mr 2,%0" :: "r"(pcpup->pc_curthread));
103 #endif
104 	pcpup->pc_curpcb = pcpup->pc_curthread->td_pcb;
105 	sp = pcpup->pc_curpcb->pcb_sp;
106 
107 	return (sp);
108 }
109 
110 static register_t
111 mpc74xx_l2_enable(register_t l2cr_config)
112 {
113 	register_t ccr, bit;
114 	uint16_t	vers;
115 
116 	vers = mfpvr() >> 16;
117 	switch (vers) {
118 	case MPC7400:
119 	case MPC7410:
120 		bit = L2CR_L2IP;
121 		break;
122 	default:
123 		bit = L2CR_L2I;
124 		break;
125 	}
126 
127 	ccr = mfspr(SPR_L2CR);
128 	if (ccr & L2CR_L2E)
129 		return (ccr);
130 
131 	/* Configure L2 cache. */
132 	ccr = l2cr_config & ~L2CR_L2E;
133 	mtspr(SPR_L2CR, ccr | L2CR_L2I);
134 	do {
135 		ccr = mfspr(SPR_L2CR);
136 	} while (ccr & bit);
137 	powerpc_sync();
138 	mtspr(SPR_L2CR, l2cr_config);
139 	powerpc_sync();
140 
141 	return (l2cr_config);
142 }
143 
144 static register_t
145 mpc745x_l3_enable(register_t l3cr_config)
146 {
147 	register_t ccr;
148 
149 	ccr = mfspr(SPR_L3CR);
150 	if (ccr & L3CR_L3E)
151 		return (ccr);
152 
153 	/* Configure L3 cache. */
154 	ccr = l3cr_config & ~(L3CR_L3E | L3CR_L3I | L3CR_L3PE | L3CR_L3CLKEN);
155 	mtspr(SPR_L3CR, ccr);
156 	ccr |= 0x4000000;       /* Magic, but documented. */
157 	mtspr(SPR_L3CR, ccr);
158 	ccr |= L3CR_L3CLKEN;
159 	mtspr(SPR_L3CR, ccr);
160 	mtspr(SPR_L3CR, ccr | L3CR_L3I);
161 	while (mfspr(SPR_L3CR) & L3CR_L3I)
162 		;
163 	mtspr(SPR_L3CR, ccr & ~L3CR_L3CLKEN);
164 	powerpc_sync();
165 	DELAY(100);
166 	mtspr(SPR_L3CR, ccr);
167 	powerpc_sync();
168 	DELAY(100);
169 	ccr |= L3CR_L3E;
170 	mtspr(SPR_L3CR, ccr);
171 	powerpc_sync();
172 
173 	return(ccr);
174 }
175 
176 static register_t
177 mpc74xx_l1d_enable(void)
178 {
179 	register_t hid;
180 
181 	hid = mfspr(SPR_HID0);
182 	if (hid & HID0_DCE)
183 		return (hid);
184 
185 	/* Enable L1 D-cache */
186 	hid |= HID0_DCE;
187 	powerpc_sync();
188 	mtspr(SPR_HID0, hid | HID0_DCFI);
189 	powerpc_sync();
190 
191 	return (hid);
192 }
193 
194 static register_t
195 mpc74xx_l1i_enable(void)
196 {
197 	register_t hid;
198 
199 	hid = mfspr(SPR_HID0);
200 	if (hid & HID0_ICE)
201 		return (hid);
202 
203 	/* Enable L1 I-cache */
204 	hid |= HID0_ICE;
205 	isync();
206 	mtspr(SPR_HID0, hid | HID0_ICFI);
207 	isync();
208 
209 	return (hid);
210 }
211 
212 static void
213 cpudep_save_config(void *dummy)
214 {
215 	uint16_t	vers;
216 
217 	vers = mfpvr() >> 16;
218 
219 	switch(vers) {
220 	case IBM970:
221 	case IBM970FX:
222 	case IBM970MP:
223 		#ifdef __powerpc64__
224 		bsp_state[0] = mfspr(SPR_HID0);
225 		bsp_state[1] = mfspr(SPR_HID1);
226 		bsp_state[2] = mfspr(SPR_HID4);
227 		bsp_state[3] = mfspr(SPR_HID5);
228 		#else
229 		__asm __volatile ("mfspr %0,%2; mr %1,%0; srdi %0,%0,32"
230 		    : "=r" (bsp_state[0]),"=r" (bsp_state[1]) : "K" (SPR_HID0));
231 		__asm __volatile ("mfspr %0,%2; mr %1,%0; srdi %0,%0,32"
232 		    : "=r" (bsp_state[2]),"=r" (bsp_state[3]) : "K" (SPR_HID1));
233 		__asm __volatile ("mfspr %0,%2; mr %1,%0; srdi %0,%0,32"
234 		    : "=r" (bsp_state[4]),"=r" (bsp_state[5]) : "K" (SPR_HID4));
235 		__asm __volatile ("mfspr %0,%2; mr %1,%0; srdi %0,%0,32"
236 		    : "=r" (bsp_state[6]),"=r" (bsp_state[7]) : "K" (SPR_HID5));
237 		#endif
238 
239 		powerpc_sync();
240 
241 		break;
242 	case IBMCELLBE:
243 		#ifdef NOTYET /* Causes problems if in instruction stream on 970 */
244 		if (mfmsr() & PSL_HV) {
245 			bsp_state[0] = mfspr(SPR_HID0);
246 			bsp_state[1] = mfspr(SPR_HID1);
247 			bsp_state[2] = mfspr(SPR_HID4);
248 			bsp_state[3] = mfspr(SPR_HID6);
249 
250 			bsp_state[4] = mfspr(SPR_CELL_TSCR);
251 		}
252 		#endif
253 
254 		bsp_state[5] = mfspr(SPR_CELL_TSRL);
255 
256 		break;
257 	case MPC7450:
258 	case MPC7455:
259 	case MPC7457:
260 		/* Only MPC745x CPUs have an L3 cache. */
261 		bsp_state[3] = mfspr(SPR_L3CR);
262 
263 		/* Fallthrough */
264 	case MPC7400:
265 	case MPC7410:
266 	case MPC7447A:
267 	case MPC7448:
268 		bsp_state[2] = mfspr(SPR_L2CR);
269 		bsp_state[1] = mfspr(SPR_HID1);
270 		bsp_state[0] = mfspr(SPR_HID0);
271 		break;
272 	}
273 }
274 
275 void
276 cpudep_ap_setup()
277 {
278 	register_t	reg;
279 	uint16_t	vers;
280 
281 	vers = mfpvr() >> 16;
282 
283 	/* The following is needed for restoring from sleep. */
284 #ifdef __powerpc64__
285 	/* Writing to the time base register is hypervisor-privileged */
286 	if (mfmsr() & PSL_HV)
287 		mttb(0);
288 #else
289 	mttb(0);
290 #endif
291 	switch(vers) {
292 	case IBM970:
293 	case IBM970FX:
294 	case IBM970MP:
295 		/* Set HIOR to 0 */
296 		__asm __volatile("mtspr 311,%0" :: "r"(0));
297 		powerpc_sync();
298 
299 		/*
300 		 * The 970 has strange rules about how to update HID registers.
301 		 * See Table 2-3, 970MP manual
302 		 *
303 		 * Note: HID4 and HID5 restored already in
304 		 * cpudep_ap_early_bootstrap()
305 		 */
306 
307 		__asm __volatile("mtasr %0; sync" :: "r"(0));
308 	#ifdef __powerpc64__
309 		__asm __volatile(" \
310 			sync; isync;					\
311 			mtspr	%1, %0;					\
312 			mfspr	%0, %1;	mfspr	%0, %1;	mfspr	%0, %1;	\
313 			mfspr	%0, %1;	mfspr	%0, %1;	mfspr	%0, %1; \
314 			sync; isync"
315 		    :: "r"(bsp_state[0]), "K"(SPR_HID0));
316 		__asm __volatile("sync; isync;	\
317 		    mtspr %1, %0; mtspr %1, %0; sync; isync"
318 		    :: "r"(bsp_state[1]), "K"(SPR_HID1));
319 	#else
320 		__asm __volatile(" \
321 			ld	%0,0(%2);				\
322 			sync; isync;					\
323 			mtspr	%1, %0;					\
324 			mfspr	%0, %1;	mfspr	%0, %1;	mfspr	%0, %1;	\
325 			mfspr	%0, %1;	mfspr	%0, %1;	mfspr	%0, %1; \
326 			sync; isync"
327 		    : "=r"(reg) : "K"(SPR_HID0), "b"(bsp_state));
328 		__asm __volatile("ld %0, 8(%2); sync; isync;	\
329 		    mtspr %1, %0; mtspr %1, %0; sync; isync"
330 		    : "=r"(reg) : "K"(SPR_HID1), "b"(bsp_state));
331 	#endif
332 
333 		powerpc_sync();
334 		break;
335 	case IBMCELLBE:
336 		#ifdef NOTYET /* Causes problems if in instruction stream on 970 */
337 		if (mfmsr() & PSL_HV) {
338 			mtspr(SPR_HID0, bsp_state[0]);
339 			mtspr(SPR_HID1, bsp_state[1]);
340 			mtspr(SPR_HID4, bsp_state[2]);
341 			mtspr(SPR_HID6, bsp_state[3]);
342 
343 			mtspr(SPR_CELL_TSCR, bsp_state[4]);
344 		}
345 		#endif
346 
347 		mtspr(SPR_CELL_TSRL, bsp_state[5]);
348 
349 		break;
350 	case MPC7400:
351 	case MPC7410:
352 	case MPC7447A:
353 	case MPC7448:
354 	case MPC7450:
355 	case MPC7455:
356 	case MPC7457:
357 		/* XXX: Program the CPU ID into PIR */
358 		__asm __volatile("mtspr 1023,%0" :: "r"(PCPU_GET(cpuid)));
359 
360 		powerpc_sync();
361 		isync();
362 
363 		mtspr(SPR_HID0, bsp_state[0]); isync();
364 		mtspr(SPR_HID1, bsp_state[1]); isync();
365 
366 		/* Now enable the L3 cache. */
367 		switch (vers) {
368 		case MPC7450:
369 		case MPC7455:
370 		case MPC7457:
371 			/* Only MPC745x CPUs have an L3 cache. */
372 			reg = mpc745x_l3_enable(bsp_state[3]);
373 		default:
374 			break;
375 		}
376 
377 		reg = mpc74xx_l2_enable(bsp_state[2]);
378 		reg = mpc74xx_l1d_enable();
379 		reg = mpc74xx_l1i_enable();
380 
381 		break;
382 	default:
383 #ifdef __powerpc64__
384 		if (!(mfmsr() & PSL_HV)) /* Rely on HV to have set things up */
385 			break;
386 #endif
387 		printf("WARNING: Unknown CPU type. Cache performace may be "
388 		    "suboptimal.\n");
389 		break;
390 	}
391 }
392 
393