xref: /openbmc/linux/arch/arm64/kernel/machine_kexec_file.c (revision 4f2c0a4acffbec01079c28f839422e64ddeff004)
1  // SPDX-License-Identifier: GPL-2.0
2  /*
3   * kexec_file for arm64
4   *
5   * Copyright (C) 2018 Linaro Limited
6   * Author: AKASHI Takahiro <takahiro.akashi@linaro.org>
7   *
8   * Most code is derived from arm64 port of kexec-tools
9   */
10  
11  #define pr_fmt(fmt) "kexec_file: " fmt
12  
13  #include <linux/ioport.h>
14  #include <linux/kernel.h>
15  #include <linux/kexec.h>
16  #include <linux/libfdt.h>
17  #include <linux/memblock.h>
18  #include <linux/of.h>
19  #include <linux/of_fdt.h>
20  #include <linux/slab.h>
21  #include <linux/string.h>
22  #include <linux/types.h>
23  #include <linux/vmalloc.h>
24  
25  const struct kexec_file_ops * const kexec_file_loaders[] = {
26  	&kexec_image_ops,
27  	NULL
28  };
29  
arch_kimage_file_post_load_cleanup(struct kimage * image)30  int arch_kimage_file_post_load_cleanup(struct kimage *image)
31  {
32  	kvfree(image->arch.dtb);
33  	image->arch.dtb = NULL;
34  
35  	vfree(image->elf_headers);
36  	image->elf_headers = NULL;
37  	image->elf_headers_sz = 0;
38  
39  	return kexec_image_post_load_cleanup_default(image);
40  }
41  
prepare_elf_headers(void ** addr,unsigned long * sz)42  static int prepare_elf_headers(void **addr, unsigned long *sz)
43  {
44  	struct crash_mem *cmem;
45  	unsigned int nr_ranges;
46  	int ret;
47  	u64 i;
48  	phys_addr_t start, end;
49  
50  	nr_ranges = 2; /* for exclusion of crashkernel region */
51  	for_each_mem_range(i, &start, &end)
52  		nr_ranges++;
53  
54  	cmem = kmalloc(struct_size(cmem, ranges, nr_ranges), GFP_KERNEL);
55  	if (!cmem)
56  		return -ENOMEM;
57  
58  	cmem->max_nr_ranges = nr_ranges;
59  	cmem->nr_ranges = 0;
60  	for_each_mem_range(i, &start, &end) {
61  		cmem->ranges[cmem->nr_ranges].start = start;
62  		cmem->ranges[cmem->nr_ranges].end = end - 1;
63  		cmem->nr_ranges++;
64  	}
65  
66  	/* Exclude crashkernel region */
67  	ret = crash_exclude_mem_range(cmem, crashk_res.start, crashk_res.end);
68  	if (ret)
69  		goto out;
70  
71  	if (crashk_low_res.end) {
72  		ret = crash_exclude_mem_range(cmem, crashk_low_res.start, crashk_low_res.end);
73  		if (ret)
74  			goto out;
75  	}
76  
77  	ret = crash_prepare_elf64_headers(cmem, true, addr, sz);
78  
79  out:
80  	kfree(cmem);
81  	return ret;
82  }
83  
84  /*
85   * Tries to add the initrd and DTB to the image. If it is not possible to find
86   * valid locations, this function will undo changes to the image and return non
87   * zero.
88   */
load_other_segments(struct kimage * image,unsigned long kernel_load_addr,unsigned long kernel_size,char * initrd,unsigned long initrd_len,char * cmdline)89  int load_other_segments(struct kimage *image,
90  			unsigned long kernel_load_addr,
91  			unsigned long kernel_size,
92  			char *initrd, unsigned long initrd_len,
93  			char *cmdline)
94  {
95  	struct kexec_buf kbuf;
96  	void *headers, *dtb = NULL;
97  	unsigned long headers_sz, initrd_load_addr = 0, dtb_len,
98  		      orig_segments = image->nr_segments;
99  	int ret = 0;
100  
101  	kbuf.image = image;
102  	/* not allocate anything below the kernel */
103  	kbuf.buf_min = kernel_load_addr + kernel_size;
104  
105  	/* load elf core header */
106  	if (image->type == KEXEC_TYPE_CRASH) {
107  		ret = prepare_elf_headers(&headers, &headers_sz);
108  		if (ret) {
109  			pr_err("Preparing elf core header failed\n");
110  			goto out_err;
111  		}
112  
113  		kbuf.buffer = headers;
114  		kbuf.bufsz = headers_sz;
115  		kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
116  		kbuf.memsz = headers_sz;
117  		kbuf.buf_align = SZ_64K; /* largest supported page size */
118  		kbuf.buf_max = ULONG_MAX;
119  		kbuf.top_down = true;
120  
121  		ret = kexec_add_buffer(&kbuf);
122  		if (ret) {
123  			vfree(headers);
124  			goto out_err;
125  		}
126  		image->elf_headers = headers;
127  		image->elf_load_addr = kbuf.mem;
128  		image->elf_headers_sz = headers_sz;
129  
130  		pr_debug("Loaded elf core header at 0x%lx bufsz=0x%lx memsz=0x%lx\n",
131  			 image->elf_load_addr, kbuf.bufsz, kbuf.memsz);
132  	}
133  
134  	/* load initrd */
135  	if (initrd) {
136  		kbuf.buffer = initrd;
137  		kbuf.bufsz = initrd_len;
138  		kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
139  		kbuf.memsz = initrd_len;
140  		kbuf.buf_align = 0;
141  		/* within 1GB-aligned window of up to 32GB in size */
142  		kbuf.buf_max = round_down(kernel_load_addr, SZ_1G)
143  						+ (unsigned long)SZ_1G * 32;
144  		kbuf.top_down = false;
145  
146  		ret = kexec_add_buffer(&kbuf);
147  		if (ret)
148  			goto out_err;
149  		initrd_load_addr = kbuf.mem;
150  
151  		pr_debug("Loaded initrd at 0x%lx bufsz=0x%lx memsz=0x%lx\n",
152  				initrd_load_addr, kbuf.bufsz, kbuf.memsz);
153  	}
154  
155  	/* load dtb */
156  	dtb = of_kexec_alloc_and_setup_fdt(image, initrd_load_addr,
157  					   initrd_len, cmdline, 0);
158  	if (!dtb) {
159  		pr_err("Preparing for new dtb failed\n");
160  		ret = -EINVAL;
161  		goto out_err;
162  	}
163  
164  	/* trim it */
165  	fdt_pack(dtb);
166  	dtb_len = fdt_totalsize(dtb);
167  	kbuf.buffer = dtb;
168  	kbuf.bufsz = dtb_len;
169  	kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
170  	kbuf.memsz = dtb_len;
171  	/* not across 2MB boundary */
172  	kbuf.buf_align = SZ_2M;
173  	kbuf.buf_max = ULONG_MAX;
174  	kbuf.top_down = true;
175  
176  	ret = kexec_add_buffer(&kbuf);
177  	if (ret)
178  		goto out_err;
179  	image->arch.dtb = dtb;
180  	image->arch.dtb_mem = kbuf.mem;
181  
182  	pr_debug("Loaded dtb at 0x%lx bufsz=0x%lx memsz=0x%lx\n",
183  			kbuf.mem, kbuf.bufsz, kbuf.memsz);
184  
185  	return 0;
186  
187  out_err:
188  	image->nr_segments = orig_segments;
189  	kvfree(dtb);
190  	return ret;
191  }
192