xref: /openbmc/linux/arch/sparc/kernel/setup_64.c (revision 29c37341)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *  linux/arch/sparc64/kernel/setup.c
4  *
5  *  Copyright (C) 1995,1996  David S. Miller (davem@caip.rutgers.edu)
6  *  Copyright (C) 1997       Jakub Jelinek (jj@sunsite.mff.cuni.cz)
7  */
8 
9 #include <linux/errno.h>
10 #include <linux/sched.h>
11 #include <linux/kernel.h>
12 #include <linux/mm.h>
13 #include <linux/stddef.h>
14 #include <linux/unistd.h>
15 #include <linux/ptrace.h>
16 #include <asm/smp.h>
17 #include <linux/user.h>
18 #include <linux/screen_info.h>
19 #include <linux/delay.h>
20 #include <linux/fs.h>
21 #include <linux/seq_file.h>
22 #include <linux/syscalls.h>
23 #include <linux/kdev_t.h>
24 #include <linux/major.h>
25 #include <linux/string.h>
26 #include <linux/init.h>
27 #include <linux/inet.h>
28 #include <linux/console.h>
29 #include <linux/root_dev.h>
30 #include <linux/interrupt.h>
31 #include <linux/cpu.h>
32 #include <linux/initrd.h>
33 #include <linux/module.h>
34 #include <linux/start_kernel.h>
35 #include <linux/memblock.h>
36 #include <uapi/linux/mount.h>
37 
38 #include <asm/io.h>
39 #include <asm/processor.h>
40 #include <asm/oplib.h>
41 #include <asm/page.h>
42 #include <asm/idprom.h>
43 #include <asm/head.h>
44 #include <asm/starfire.h>
45 #include <asm/mmu_context.h>
46 #include <asm/timer.h>
47 #include <asm/sections.h>
48 #include <asm/setup.h>
49 #include <asm/mmu.h>
50 #include <asm/ns87303.h>
51 #include <asm/btext.h>
52 #include <asm/elf.h>
53 #include <asm/mdesc.h>
54 #include <asm/cacheflush.h>
55 #include <asm/dma.h>
56 #include <asm/irq.h>
57 
58 #ifdef CONFIG_IP_PNP
59 #include <net/ipconfig.h>
60 #endif
61 
62 #include "entry.h"
63 #include "kernel.h"
64 
65 /* Used to synchronize accesses to NatSemi SUPER I/O chip configure
66  * operations in asm/ns87303.h
67  */
68 DEFINE_SPINLOCK(ns87303_lock);
69 EXPORT_SYMBOL(ns87303_lock);
70 
71 struct screen_info screen_info = {
72 	0, 0,			/* orig-x, orig-y */
73 	0,			/* unused */
74 	0,			/* orig-video-page */
75 	0,			/* orig-video-mode */
76 	128,			/* orig-video-cols */
77 	0, 0, 0,		/* unused, ega_bx, unused */
78 	54,			/* orig-video-lines */
79 	0,                      /* orig-video-isVGA */
80 	16                      /* orig-video-points */
81 };
82 
83 static void
84 prom_console_write(struct console *con, const char *s, unsigned int n)
85 {
86 	prom_write(s, n);
87 }
88 
89 /* Exported for mm/init.c:paging_init. */
90 unsigned long cmdline_memory_size = 0;
91 
92 static struct console prom_early_console = {
93 	.name =		"earlyprom",
94 	.write =	prom_console_write,
95 	.flags =	CON_PRINTBUFFER | CON_BOOT | CON_ANYTIME,
96 	.index =	-1,
97 };
98 
99 /*
100  * Process kernel command line switches that are specific to the
101  * SPARC or that require special low-level processing.
102  */
103 static void __init process_switch(char c)
104 {
105 	switch (c) {
106 	case 'd':
107 	case 's':
108 		break;
109 	case 'h':
110 		prom_printf("boot_flags_init: Halt!\n");
111 		prom_halt();
112 		break;
113 	case 'p':
114 		prom_early_console.flags &= ~CON_BOOT;
115 		break;
116 	case 'P':
117 		/* Force UltraSPARC-III P-Cache on. */
118 		if (tlb_type != cheetah) {
119 			printk("BOOT: Ignoring P-Cache force option.\n");
120 			break;
121 		}
122 		cheetah_pcache_forced_on = 1;
123 		add_taint(TAINT_MACHINE_CHECK, LOCKDEP_NOW_UNRELIABLE);
124 		cheetah_enable_pcache();
125 		break;
126 
127 	default:
128 		printk("Unknown boot switch (-%c)\n", c);
129 		break;
130 	}
131 }
132 
133 static void __init boot_flags_init(char *commands)
134 {
135 	while (*commands) {
136 		/* Move to the start of the next "argument". */
137 		while (*commands == ' ')
138 			commands++;
139 
140 		/* Process any command switches, otherwise skip it. */
141 		if (*commands == '\0')
142 			break;
143 		if (*commands == '-') {
144 			commands++;
145 			while (*commands && *commands != ' ')
146 				process_switch(*commands++);
147 			continue;
148 		}
149 		if (!strncmp(commands, "mem=", 4))
150 			cmdline_memory_size = memparse(commands + 4, &commands);
151 
152 		while (*commands && *commands != ' ')
153 			commands++;
154 	}
155 }
156 
157 extern unsigned short root_flags;
158 extern unsigned short root_dev;
159 extern unsigned short ram_flags;
160 #define RAMDISK_IMAGE_START_MASK	0x07FF
161 #define RAMDISK_PROMPT_FLAG		0x8000
162 #define RAMDISK_LOAD_FLAG		0x4000
163 
164 extern int root_mountflags;
165 
166 char reboot_command[COMMAND_LINE_SIZE];
167 
168 static struct pt_regs fake_swapper_regs = { { 0, }, 0, 0, 0, 0 };
169 
170 static void __init per_cpu_patch(void)
171 {
172 	struct cpuid_patch_entry *p;
173 	unsigned long ver;
174 	int is_jbus;
175 
176 	if (tlb_type == spitfire && !this_is_starfire)
177 		return;
178 
179 	is_jbus = 0;
180 	if (tlb_type != hypervisor) {
181 		__asm__ ("rdpr %%ver, %0" : "=r" (ver));
182 		is_jbus = ((ver >> 32UL) == __JALAPENO_ID ||
183 			   (ver >> 32UL) == __SERRANO_ID);
184 	}
185 
186 	p = &__cpuid_patch;
187 	while (p < &__cpuid_patch_end) {
188 		unsigned long addr = p->addr;
189 		unsigned int *insns;
190 
191 		switch (tlb_type) {
192 		case spitfire:
193 			insns = &p->starfire[0];
194 			break;
195 		case cheetah:
196 		case cheetah_plus:
197 			if (is_jbus)
198 				insns = &p->cheetah_jbus[0];
199 			else
200 				insns = &p->cheetah_safari[0];
201 			break;
202 		case hypervisor:
203 			insns = &p->sun4v[0];
204 			break;
205 		default:
206 			prom_printf("Unknown cpu type, halting.\n");
207 			prom_halt();
208 		}
209 
210 		*(unsigned int *) (addr +  0) = insns[0];
211 		wmb();
212 		__asm__ __volatile__("flush	%0" : : "r" (addr +  0));
213 
214 		*(unsigned int *) (addr +  4) = insns[1];
215 		wmb();
216 		__asm__ __volatile__("flush	%0" : : "r" (addr +  4));
217 
218 		*(unsigned int *) (addr +  8) = insns[2];
219 		wmb();
220 		__asm__ __volatile__("flush	%0" : : "r" (addr +  8));
221 
222 		*(unsigned int *) (addr + 12) = insns[3];
223 		wmb();
224 		__asm__ __volatile__("flush	%0" : : "r" (addr + 12));
225 
226 		p++;
227 	}
228 }
229 
230 void sun4v_patch_1insn_range(struct sun4v_1insn_patch_entry *start,
231 			     struct sun4v_1insn_patch_entry *end)
232 {
233 	while (start < end) {
234 		unsigned long addr = start->addr;
235 
236 		*(unsigned int *) (addr +  0) = start->insn;
237 		wmb();
238 		__asm__ __volatile__("flush	%0" : : "r" (addr +  0));
239 
240 		start++;
241 	}
242 }
243 
244 void sun4v_patch_2insn_range(struct sun4v_2insn_patch_entry *start,
245 			     struct sun4v_2insn_patch_entry *end)
246 {
247 	while (start < end) {
248 		unsigned long addr = start->addr;
249 
250 		*(unsigned int *) (addr +  0) = start->insns[0];
251 		wmb();
252 		__asm__ __volatile__("flush	%0" : : "r" (addr +  0));
253 
254 		*(unsigned int *) (addr +  4) = start->insns[1];
255 		wmb();
256 		__asm__ __volatile__("flush	%0" : : "r" (addr +  4));
257 
258 		start++;
259 	}
260 }
261 
262 void sun_m7_patch_2insn_range(struct sun4v_2insn_patch_entry *start,
263 			     struct sun4v_2insn_patch_entry *end)
264 {
265 	while (start < end) {
266 		unsigned long addr = start->addr;
267 
268 		*(unsigned int *) (addr +  0) = start->insns[0];
269 		wmb();
270 		__asm__ __volatile__("flush	%0" : : "r" (addr +  0));
271 
272 		*(unsigned int *) (addr +  4) = start->insns[1];
273 		wmb();
274 		__asm__ __volatile__("flush	%0" : : "r" (addr +  4));
275 
276 		start++;
277 	}
278 }
279 
280 static void __init sun4v_patch(void)
281 {
282 	extern void sun4v_hvapi_init(void);
283 
284 	if (tlb_type != hypervisor)
285 		return;
286 
287 	sun4v_patch_1insn_range(&__sun4v_1insn_patch,
288 				&__sun4v_1insn_patch_end);
289 
290 	sun4v_patch_2insn_range(&__sun4v_2insn_patch,
291 				&__sun4v_2insn_patch_end);
292 
293 	switch (sun4v_chip_type) {
294 	case SUN4V_CHIP_SPARC_M7:
295 	case SUN4V_CHIP_SPARC_M8:
296 	case SUN4V_CHIP_SPARC_SN:
297 		sun4v_patch_1insn_range(&__sun_m7_1insn_patch,
298 					&__sun_m7_1insn_patch_end);
299 		sun_m7_patch_2insn_range(&__sun_m7_2insn_patch,
300 					 &__sun_m7_2insn_patch_end);
301 		break;
302 	default:
303 		break;
304 	}
305 
306 	if (sun4v_chip_type != SUN4V_CHIP_NIAGARA1) {
307 		sun4v_patch_1insn_range(&__fast_win_ctrl_1insn_patch,
308 					&__fast_win_ctrl_1insn_patch_end);
309 	}
310 
311 	sun4v_hvapi_init();
312 }
313 
314 static void __init popc_patch(void)
315 {
316 	struct popc_3insn_patch_entry *p3;
317 	struct popc_6insn_patch_entry *p6;
318 
319 	p3 = &__popc_3insn_patch;
320 	while (p3 < &__popc_3insn_patch_end) {
321 		unsigned long i, addr = p3->addr;
322 
323 		for (i = 0; i < 3; i++) {
324 			*(unsigned int *) (addr +  (i * 4)) = p3->insns[i];
325 			wmb();
326 			__asm__ __volatile__("flush	%0"
327 					     : : "r" (addr +  (i * 4)));
328 		}
329 
330 		p3++;
331 	}
332 
333 	p6 = &__popc_6insn_patch;
334 	while (p6 < &__popc_6insn_patch_end) {
335 		unsigned long i, addr = p6->addr;
336 
337 		for (i = 0; i < 6; i++) {
338 			*(unsigned int *) (addr +  (i * 4)) = p6->insns[i];
339 			wmb();
340 			__asm__ __volatile__("flush	%0"
341 					     : : "r" (addr +  (i * 4)));
342 		}
343 
344 		p6++;
345 	}
346 }
347 
348 static void __init pause_patch(void)
349 {
350 	struct pause_patch_entry *p;
351 
352 	p = &__pause_3insn_patch;
353 	while (p < &__pause_3insn_patch_end) {
354 		unsigned long i, addr = p->addr;
355 
356 		for (i = 0; i < 3; i++) {
357 			*(unsigned int *) (addr +  (i * 4)) = p->insns[i];
358 			wmb();
359 			__asm__ __volatile__("flush	%0"
360 					     : : "r" (addr +  (i * 4)));
361 		}
362 
363 		p++;
364 	}
365 }
366 
367 void __init start_early_boot(void)
368 {
369 	int cpu;
370 
371 	check_if_starfire();
372 	per_cpu_patch();
373 	sun4v_patch();
374 	smp_init_cpu_poke();
375 
376 	cpu = hard_smp_processor_id();
377 	if (cpu >= NR_CPUS) {
378 		prom_printf("Serious problem, boot cpu id (%d) >= NR_CPUS (%d)\n",
379 			    cpu, NR_CPUS);
380 		prom_halt();
381 	}
382 	current_thread_info()->cpu = cpu;
383 
384 	time_init_early();
385 	prom_init_report();
386 	start_kernel();
387 }
388 
389 /* On Ultra, we support all of the v8 capabilities. */
390 unsigned long sparc64_elf_hwcap = (HWCAP_SPARC_FLUSH | HWCAP_SPARC_STBAR |
391 				   HWCAP_SPARC_SWAP | HWCAP_SPARC_MULDIV |
392 				   HWCAP_SPARC_V9);
393 EXPORT_SYMBOL(sparc64_elf_hwcap);
394 
395 static const char *hwcaps[] = {
396 	"flush", "stbar", "swap", "muldiv", "v9",
397 	"ultra3", "blkinit", "n2",
398 
399 	/* These strings are as they appear in the machine description
400 	 * 'hwcap-list' property for cpu nodes.
401 	 */
402 	"mul32", "div32", "fsmuld", "v8plus", "popc", "vis", "vis2",
403 	"ASIBlkInit", "fmaf", "vis3", "hpc", "random", "trans", "fjfmau",
404 	"ima", "cspare", "pause", "cbcond", NULL /*reserved for crypto */,
405 	"adp",
406 };
407 
408 static const char *crypto_hwcaps[] = {
409 	"aes", "des", "kasumi", "camellia", "md5", "sha1", "sha256",
410 	"sha512", "mpmul", "montmul", "montsqr", "crc32c",
411 };
412 
413 void cpucap_info(struct seq_file *m)
414 {
415 	unsigned long caps = sparc64_elf_hwcap;
416 	int i, printed = 0;
417 
418 	seq_puts(m, "cpucaps\t\t: ");
419 	for (i = 0; i < ARRAY_SIZE(hwcaps); i++) {
420 		unsigned long bit = 1UL << i;
421 		if (hwcaps[i] && (caps & bit)) {
422 			seq_printf(m, "%s%s",
423 				   printed ? "," : "", hwcaps[i]);
424 			printed++;
425 		}
426 	}
427 	if (caps & HWCAP_SPARC_CRYPTO) {
428 		unsigned long cfr;
429 
430 		__asm__ __volatile__("rd %%asr26, %0" : "=r" (cfr));
431 		for (i = 0; i < ARRAY_SIZE(crypto_hwcaps); i++) {
432 			unsigned long bit = 1UL << i;
433 			if (cfr & bit) {
434 				seq_printf(m, "%s%s",
435 					   printed ? "," : "", crypto_hwcaps[i]);
436 				printed++;
437 			}
438 		}
439 	}
440 	seq_putc(m, '\n');
441 }
442 
443 static void __init report_one_hwcap(int *printed, const char *name)
444 {
445 	if ((*printed) == 0)
446 		printk(KERN_INFO "CPU CAPS: [");
447 	printk(KERN_CONT "%s%s",
448 	       (*printed) ? "," : "", name);
449 	if (++(*printed) == 8) {
450 		printk(KERN_CONT "]\n");
451 		*printed = 0;
452 	}
453 }
454 
455 static void __init report_crypto_hwcaps(int *printed)
456 {
457 	unsigned long cfr;
458 	int i;
459 
460 	__asm__ __volatile__("rd %%asr26, %0" : "=r" (cfr));
461 
462 	for (i = 0; i < ARRAY_SIZE(crypto_hwcaps); i++) {
463 		unsigned long bit = 1UL << i;
464 		if (cfr & bit)
465 			report_one_hwcap(printed, crypto_hwcaps[i]);
466 	}
467 }
468 
469 static void __init report_hwcaps(unsigned long caps)
470 {
471 	int i, printed = 0;
472 
473 	for (i = 0; i < ARRAY_SIZE(hwcaps); i++) {
474 		unsigned long bit = 1UL << i;
475 		if (hwcaps[i] && (caps & bit))
476 			report_one_hwcap(&printed, hwcaps[i]);
477 	}
478 	if (caps & HWCAP_SPARC_CRYPTO)
479 		report_crypto_hwcaps(&printed);
480 	if (printed != 0)
481 		printk(KERN_CONT "]\n");
482 }
483 
484 static unsigned long __init mdesc_cpu_hwcap_list(void)
485 {
486 	struct mdesc_handle *hp;
487 	unsigned long caps = 0;
488 	const char *prop;
489 	int len;
490 	u64 pn;
491 
492 	hp = mdesc_grab();
493 	if (!hp)
494 		return 0;
495 
496 	pn = mdesc_node_by_name(hp, MDESC_NODE_NULL, "cpu");
497 	if (pn == MDESC_NODE_NULL)
498 		goto out;
499 
500 	prop = mdesc_get_property(hp, pn, "hwcap-list", &len);
501 	if (!prop)
502 		goto out;
503 
504 	while (len) {
505 		int i, plen;
506 
507 		for (i = 0; i < ARRAY_SIZE(hwcaps); i++) {
508 			unsigned long bit = 1UL << i;
509 
510 			if (hwcaps[i] && !strcmp(prop, hwcaps[i])) {
511 				caps |= bit;
512 				break;
513 			}
514 		}
515 		for (i = 0; i < ARRAY_SIZE(crypto_hwcaps); i++) {
516 			if (!strcmp(prop, crypto_hwcaps[i]))
517 				caps |= HWCAP_SPARC_CRYPTO;
518 		}
519 
520 		plen = strlen(prop) + 1;
521 		prop += plen;
522 		len -= plen;
523 	}
524 
525 out:
526 	mdesc_release(hp);
527 	return caps;
528 }
529 
530 /* This yields a mask that user programs can use to figure out what
531  * instruction set this cpu supports.
532  */
533 static void __init init_sparc64_elf_hwcap(void)
534 {
535 	unsigned long cap = sparc64_elf_hwcap;
536 	unsigned long mdesc_caps;
537 
538 	if (tlb_type == cheetah || tlb_type == cheetah_plus)
539 		cap |= HWCAP_SPARC_ULTRA3;
540 	else if (tlb_type == hypervisor) {
541 		if (sun4v_chip_type == SUN4V_CHIP_NIAGARA1 ||
542 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA2 ||
543 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA3 ||
544 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA4 ||
545 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA5 ||
546 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M6 ||
547 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M7 ||
548 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M8 ||
549 		    sun4v_chip_type == SUN4V_CHIP_SPARC_SN ||
550 		    sun4v_chip_type == SUN4V_CHIP_SPARC64X)
551 			cap |= HWCAP_SPARC_BLKINIT;
552 		if (sun4v_chip_type == SUN4V_CHIP_NIAGARA2 ||
553 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA3 ||
554 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA4 ||
555 		    sun4v_chip_type == SUN4V_CHIP_NIAGARA5 ||
556 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M6 ||
557 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M7 ||
558 		    sun4v_chip_type == SUN4V_CHIP_SPARC_M8 ||
559 		    sun4v_chip_type == SUN4V_CHIP_SPARC_SN ||
560 		    sun4v_chip_type == SUN4V_CHIP_SPARC64X)
561 			cap |= HWCAP_SPARC_N2;
562 	}
563 
564 	cap |= (AV_SPARC_MUL32 | AV_SPARC_DIV32 | AV_SPARC_V8PLUS);
565 
566 	mdesc_caps = mdesc_cpu_hwcap_list();
567 	if (!mdesc_caps) {
568 		if (tlb_type == spitfire)
569 			cap |= AV_SPARC_VIS;
570 		if (tlb_type == cheetah || tlb_type == cheetah_plus)
571 			cap |= AV_SPARC_VIS | AV_SPARC_VIS2;
572 		if (tlb_type == cheetah_plus) {
573 			unsigned long impl, ver;
574 
575 			__asm__ __volatile__("rdpr %%ver, %0" : "=r" (ver));
576 			impl = ((ver >> 32) & 0xffff);
577 			if (impl == PANTHER_IMPL)
578 				cap |= AV_SPARC_POPC;
579 		}
580 		if (tlb_type == hypervisor) {
581 			if (sun4v_chip_type == SUN4V_CHIP_NIAGARA1)
582 				cap |= AV_SPARC_ASI_BLK_INIT;
583 			if (sun4v_chip_type == SUN4V_CHIP_NIAGARA2 ||
584 			    sun4v_chip_type == SUN4V_CHIP_NIAGARA3 ||
585 			    sun4v_chip_type == SUN4V_CHIP_NIAGARA4 ||
586 			    sun4v_chip_type == SUN4V_CHIP_NIAGARA5 ||
587 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M6 ||
588 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M7 ||
589 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M8 ||
590 			    sun4v_chip_type == SUN4V_CHIP_SPARC_SN ||
591 			    sun4v_chip_type == SUN4V_CHIP_SPARC64X)
592 				cap |= (AV_SPARC_VIS | AV_SPARC_VIS2 |
593 					AV_SPARC_ASI_BLK_INIT |
594 					AV_SPARC_POPC);
595 			if (sun4v_chip_type == SUN4V_CHIP_NIAGARA3 ||
596 			    sun4v_chip_type == SUN4V_CHIP_NIAGARA4 ||
597 			    sun4v_chip_type == SUN4V_CHIP_NIAGARA5 ||
598 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M6 ||
599 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M7 ||
600 			    sun4v_chip_type == SUN4V_CHIP_SPARC_M8 ||
601 			    sun4v_chip_type == SUN4V_CHIP_SPARC_SN ||
602 			    sun4v_chip_type == SUN4V_CHIP_SPARC64X)
603 				cap |= (AV_SPARC_VIS3 | AV_SPARC_HPC |
604 					AV_SPARC_FMAF);
605 		}
606 	}
607 	sparc64_elf_hwcap = cap | mdesc_caps;
608 
609 	report_hwcaps(sparc64_elf_hwcap);
610 
611 	if (sparc64_elf_hwcap & AV_SPARC_POPC)
612 		popc_patch();
613 	if (sparc64_elf_hwcap & AV_SPARC_PAUSE)
614 		pause_patch();
615 }
616 
617 void __init alloc_irqstack_bootmem(void)
618 {
619 	unsigned int i, node;
620 
621 	for_each_possible_cpu(i) {
622 		node = cpu_to_node(i);
623 
624 		softirq_stack[i] = memblock_alloc_node(THREAD_SIZE,
625 						       THREAD_SIZE, node);
626 		if (!softirq_stack[i])
627 			panic("%s: Failed to allocate %lu bytes align=%lx nid=%d\n",
628 			      __func__, THREAD_SIZE, THREAD_SIZE, node);
629 		hardirq_stack[i] = memblock_alloc_node(THREAD_SIZE,
630 						       THREAD_SIZE, node);
631 		if (!hardirq_stack[i])
632 			panic("%s: Failed to allocate %lu bytes align=%lx nid=%d\n",
633 			      __func__, THREAD_SIZE, THREAD_SIZE, node);
634 	}
635 }
636 
637 void __init setup_arch(char **cmdline_p)
638 {
639 	/* Initialize PROM console and command line. */
640 	*cmdline_p = prom_getbootargs();
641 	strlcpy(boot_command_line, *cmdline_p, COMMAND_LINE_SIZE);
642 	parse_early_param();
643 
644 	boot_flags_init(*cmdline_p);
645 #ifdef CONFIG_EARLYFB
646 	if (btext_find_display())
647 #endif
648 		register_console(&prom_early_console);
649 
650 	if (tlb_type == hypervisor)
651 		pr_info("ARCH: SUN4V\n");
652 	else
653 		pr_info("ARCH: SUN4U\n");
654 
655 	idprom_init();
656 
657 	if (!root_flags)
658 		root_mountflags &= ~MS_RDONLY;
659 	ROOT_DEV = old_decode_dev(root_dev);
660 #ifdef CONFIG_BLK_DEV_RAM
661 	rd_image_start = ram_flags & RAMDISK_IMAGE_START_MASK;
662 #endif
663 
664 	task_thread_info(&init_task)->kregs = &fake_swapper_regs;
665 
666 #ifdef CONFIG_IP_PNP
667 	if (!ic_set_manually) {
668 		phandle chosen = prom_finddevice("/chosen");
669 		u32 cl, sv, gw;
670 
671 		cl = prom_getintdefault (chosen, "client-ip", 0);
672 		sv = prom_getintdefault (chosen, "server-ip", 0);
673 		gw = prom_getintdefault (chosen, "gateway-ip", 0);
674 		if (cl && sv) {
675 			ic_myaddr = cl;
676 			ic_servaddr = sv;
677 			if (gw)
678 				ic_gateway = gw;
679 #if defined(CONFIG_IP_PNP_BOOTP) || defined(CONFIG_IP_PNP_RARP)
680 			ic_proto_enabled = 0;
681 #endif
682 		}
683 	}
684 #endif
685 
686 	/* Get boot processor trap_block[] setup.  */
687 	init_cur_cpu_trap(current_thread_info());
688 
689 	paging_init();
690 	init_sparc64_elf_hwcap();
691 	smp_fill_in_cpu_possible_map();
692 	/*
693 	 * Once the OF device tree and MDESC have been setup and nr_cpus has
694 	 * been parsed, we know the list of possible cpus.  Therefore we can
695 	 * allocate the IRQ stacks.
696 	 */
697 	alloc_irqstack_bootmem();
698 }
699 
700 extern int stop_a_enabled;
701 
702 void sun_do_break(void)
703 {
704 	if (!stop_a_enabled)
705 		return;
706 
707 	prom_printf("\n");
708 	flush_user_windows();
709 
710 	prom_cmdline();
711 }
712 EXPORT_SYMBOL(sun_do_break);
713 
714 int stop_a_enabled = 1;
715 EXPORT_SYMBOL(stop_a_enabled);
716