/* * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. * * Support for Kernel relocation at boot time * * Copyright (C) 2015, Imagination Technologies Ltd. * Authors: Matt Redfearn (matt.redfearn@imgtec.com) */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #define RELOCATED(x) ((void *)((long)x + offset)) extern u32 _relocation_start[]; /* End kernel image / start relocation table */ extern u32 _relocation_end[]; /* End relocation table */ extern long __start___ex_table; /* Start exception table */ extern long __stop___ex_table; /* End exception table */ static inline u32 __init get_synci_step(void) { u32 res; __asm__("rdhwr %0, $1" : "=r" (res)); return res; } static void __init sync_icache(void *kbase, unsigned long kernel_length) { void *kend = kbase + kernel_length; u32 step = get_synci_step(); do { __asm__ __volatile__( "synci 0(%0)" : /* no output */ : "r" (kbase)); kbase += step; } while (kbase < kend); /* Completion barrier */ __sync(); } static int __init apply_r_mips_64_rel(u32 *loc_orig, u32 *loc_new, long offset) { *(u64 *)loc_new += offset; return 0; } static int __init apply_r_mips_32_rel(u32 *loc_orig, u32 *loc_new, long offset) { *loc_new += offset; return 0; } static int __init apply_r_mips_26_rel(u32 *loc_orig, u32 *loc_new, long offset) { unsigned long target_addr = (*loc_orig) & 0x03ffffff; if (offset % 4) { pr_err("Dangerous R_MIPS_26 REL relocation\n"); return -ENOEXEC; } /* Original target address */ target_addr <<= 2; target_addr += (unsigned long)loc_orig & ~0x0fffffff; /* Get the new target address */ target_addr += offset; if ((target_addr & 0xf0000000) != ((unsigned long)loc_new & 0xf0000000)) { pr_err("R_MIPS_26 REL relocation overflow\n"); return -ENOEXEC; } target_addr -= (unsigned long)loc_new & ~0x0fffffff; target_addr >>= 2; *loc_new = (*loc_new & ~0x03ffffff) | (target_addr & 0x03ffffff); return 0; } static int __init apply_r_mips_hi16_rel(u32 *loc_orig, u32 *loc_new, long offset) { unsigned long insn = *loc_orig; unsigned long target = (insn & 0xffff) << 16; /* high 16bits of target */ target += offset; *loc_new = (insn & ~0xffff) | ((target >> 16) & 0xffff); return 0; } static int (*reloc_handlers_rel[]) (u32 *, u32 *, long) __initdata = { [R_MIPS_64] = apply_r_mips_64_rel, [R_MIPS_32] = apply_r_mips_32_rel, [R_MIPS_26] = apply_r_mips_26_rel, [R_MIPS_HI16] = apply_r_mips_hi16_rel, }; int __init do_relocations(void *kbase_old, void *kbase_new, long offset) { u32 *r; u32 *loc_orig; u32 *loc_new; int type; int res; for (r = _relocation_start; r < _relocation_end; r++) { /* Sentinel for last relocation */ if (*r == 0) break; type = (*r >> 24) & 0xff; loc_orig = (void *)(kbase_old + ((*r & 0x00ffffff) << 2)); loc_new = RELOCATED(loc_orig); if (reloc_handlers_rel[type] == NULL) { /* Unsupported relocation */ pr_err("Unhandled relocation type %d at 0x%pK\n", type, loc_orig); return -ENOEXEC; } res = reloc_handlers_rel[type](loc_orig, loc_new, offset); if (res) return res; } return 0; } /* * The exception table is filled in by the relocs tool after vmlinux is linked. * It must be relocated separately since there will not be any relocation * information for it filled in by the linker. */ static int __init relocate_exception_table(long offset) { unsigned long *etable_start, *etable_end, *e; etable_start = RELOCATED(&__start___ex_table); etable_end = RELOCATED(&__stop___ex_table); for (e = etable_start; e < etable_end; e++) *e += offset; return 0; } #ifdef CONFIG_RANDOMIZE_BASE static inline __init unsigned long rotate_xor(unsigned long hash, const void *area, size_t size) { const typeof(hash) *ptr = PTR_ALIGN(area, sizeof(hash)); size_t diff, i; diff = (void *)ptr - area; if (unlikely(size < diff + sizeof(hash))) return hash; size = ALIGN_DOWN(size - diff, sizeof(hash)); for (i = 0; i < size / sizeof(hash); i++) { /* Rotate by odd number of bits and XOR. */ hash = (hash << ((sizeof(hash) * 8) - 7)) | (hash >> 7); hash ^= ptr[i]; } return hash; } static inline __init unsigned long get_random_boot(void) { unsigned long entropy = random_get_entropy(); unsigned long hash = 0; /* Attempt to create a simple but unpredictable starting entropy. */ hash = rotate_xor(hash, linux_banner, strlen(linux_banner)); /* Add in any runtime entropy we can get */ hash = rotate_xor(hash, &entropy, sizeof(entropy)); /* Add in entropy from the kernel rng, as it may already be * initialized from early sources. */ entropy = get_random_long(); hash = rotate_xor(hash, &entropy, sizeof(entropy)); #if defined(CONFIG_USE_OF) /* Get any additional entropy passed in device tree */ if (initial_boot_params) { int node, len; u64 *prop; node = fdt_path_offset(initial_boot_params, "/chosen"); if (node >= 0) { prop = fdt_getprop_w(initial_boot_params, node, "kaslr-seed", &len); if (prop && (len == sizeof(u64))) hash = rotate_xor(hash, prop, sizeof(*prop)); } } #endif /* CONFIG_USE_OF */ return hash; } static inline __init bool kaslr_disabled(void) { char *str; #if defined(CONFIG_CMDLINE_BOOL) const char *builtin_cmdline = CONFIG_CMDLINE; str = strstr(builtin_cmdline, "nokaslr"); if (str == builtin_cmdline || (str > builtin_cmdline && *(str - 1) == ' ')) return true; #endif str = strstr(arcs_cmdline, "nokaslr"); if (str == arcs_cmdline || (str > arcs_cmdline && *(str - 1) == ' ')) return true; return false; } static inline int __init relocation_addr_valid(void *loc_new) { unsigned long kernel_length, text_length; if ((unsigned long)loc_new & 0x0000ffff) { /* Inappropriately aligned new location */ return 0; } kernel_length = (long)_end - (long)(&_text); text_length = (long)(&_etext) - (long)(&_text); /* only relocate to addresses that ensure R_MIPS_26 relocations work */ if (((unsigned long)loc_new & 0xf0000000) != (((unsigned long)loc_new + text_length) & 0xf0000000)) { pr_err("location is not R_MIPS_26 safe for 0x%p\n", loc_new); return 0; } /* do not relocate into reserved space */ if (!early_init_dt_is_available_memory(__pa(loc_new), kernel_length)) { pr_err("location overlaps with reserved memory for 0x%p\n", loc_new); return 0; } /* do not relocate into the initrd */ if (initrd_start) { if (!(initrd_end < (unsigned long)loc_new || initrd_start >= (unsigned long)loc_new + kernel_length)) { pr_err("location overlaps with the initrd for 0x%p\n", loc_new); return 0; } } /* do not overwrite our mtdram */ if (avm_mtdram_check_overlap(arcs_cmdline, __pa(loc_new), kernel_length)) { pr_err("location overlaps with an mtdram for 0x%p\n", loc_new); return 0; } /* enough space before kernel */ if ((unsigned long)loc_new + kernel_length < (unsigned long)&_text) return 1; /* behind the original kernel */ if ((unsigned long)loc_new > (unsigned long)&_end) return 1; /* New location overlaps original kernel */ pr_err("location overlaps with the old kernel for 0x%p\n", loc_new); return 0; } static inline void __init *determine_relocation_address(void) { /* Choose a new address for the kernel */ unsigned long kernel_length; void *dest = &_text; unsigned long offset = 0; int count = 10; if (CONFIG_RANDOMIZE_BASE_MIN_ADDR != 0) dest = (void *)CONFIG_RANDOMIZE_BASE_MIN_ADDR; if (kaslr_disabled()) return dest; kernel_length = (long)_end - (long)(&_text); do { offset = get_random_boot() << 16; offset &= (CONFIG_RANDOMIZE_BASE_MAX_OFFSET - 1); if (relocation_addr_valid(RELOCATED(dest))) return RELOCATED(dest); pr_debug("Relocation by 0x%lX not possible\n", offset); } while (--count); if (dest == &_text) pr_err("Relocation not possible, did not find a working offset\n"); else pr_err("Relocation to fixed offset\n"); return dest; } #else static inline void __init *determine_relocation_address(void) { /* * Choose a new address for the kernel * For now we'll hard code the destination */ return (void *)0xffffffff81000000; } #endif static void __init relocate_clear_old(void) { void *kernel_start = (void *)VMLINUX_LOAD_ADDRESS; unsigned long kernel_length = (long)(&_relocation_start) - (long)(&_text); u32 *cur; if (kernel_start == &_text) { start_kernel(); return; } for (cur = kernel_start; (char *)cur < (char *)kernel_start + kernel_length; ++cur) *cur = 0xdeadbeef; start_kernel(); unreachable(); } void *__init relocate_kernel(void) { void *loc_new; unsigned long kernel_length; unsigned long bss_length; long offset = 0; int res = 1; /* Default to original kernel entry point */ void *kernel_entry = start_kernel; bool avm_use_embedded_dtb = false; /* Get the command line */ fw_init_cmdline(); #if defined(CONFIG_USE_OF) /* * DTB was passed via firmware info so simulate * it was passed from the firmware */ if (!fw_passed_dtb && avm_fw_embedded_dtb()) { avm_use_embedded_dtb = true; fw_passed_dtb = (unsigned long)avm_fw_embedded_dtb(); } /* Deal with the device tree */ early_init_dt_scan(plat_get_fdt()); if (boot_command_line[0]) { /* Boot command line was passed in device tree */ strlcpy(arcs_cmdline, boot_command_line, COMMAND_LINE_SIZE); } #endif /* CONFIG_USE_OF */ early_init_shm_rng_seed(); kernel_length = (long)(&_relocation_start) - (long)(&_text); bss_length = (long)&__bss_stop - (long)&__bss_start; loc_new = determine_relocation_address(); offset = (unsigned long)loc_new - (unsigned long)(&_text); /* Reset the command line now so we don't end up with a duplicate */ arcs_cmdline[0] = '\0'; if (offset) { /* Copy the kernel to it's new location */ memcpy(loc_new, &_text, kernel_length); /* Perform relocations on the new kernel */ res = do_relocations(&_text, loc_new, offset); if (res < 0) goto out; /* Sync the caches ready for execution of new kernel */ sync_icache(loc_new, kernel_length); res = relocate_exception_table(offset); if (res < 0) goto out; #ifdef CONFIG_USE_OF #ifdef CONFIG_MIPS_RAW_APPENDED_DTB /* * The appended DTB is not copied above, as it is located after * "_relocation_start" (see vmlinux.lds.S) * Copy the dtb block (size as specified in lds) and update the * pointer "fw_passed_dtb". The copy of .bss will update the * variable at the new location later. */ memcpy((void *)fw_passed_dtb + offset, (void *)fw_passed_dtb, 0x100000); fw_passed_dtb += offset; /* * Reset the cached pointer from fdt to ensure the new address * is retrieved again, instead of continuing with the original * address. */ initial_boot_params = NULL; #endif #endif avm_fw_info_embed_relocate(offset); if (avm_use_embedded_dtb) { fw_passed_dtb = 0; initial_boot_params = NULL; } /* * The original .bss has already been cleared, and * some variables such as command line parameters * stored to it so make a copy in the new location. */ memcpy(RELOCATED(&__bss_start), &__bss_start, bss_length); /* The current thread is now within the relocated image */ __current_thread_info = RELOCATED(&init_thread_union); /* Return the new kernel's entry point */ kernel_entry = RELOCATED(relocate_clear_old); } out: return kernel_entry; } /* * Show relocation information on panic. */ void show_kernel_relocation(const char *level) { long offset; offset = __pa_symbol(_text) - __pa_symbol(VMLINUX_LOAD_ADDRESS); if (IS_ENABLED(CONFIG_RELOCATABLE) && offset != 0) { printk(level); pr_cont("Kernel relocated by 0x%pK\n", (void *)offset); pr_cont(" .text @ 0x%pK\n", _text); pr_cont(" .data @ 0x%pK\n", _sdata); pr_cont(" .bss @ 0x%pK\n", __bss_start); } } static int kernel_location_notifier_fn(struct notifier_block *self, unsigned long v, void *p) { show_kernel_relocation(KERN_EMERG); return NOTIFY_DONE; } static struct notifier_block kernel_location_notifier = { .notifier_call = kernel_location_notifier_fn }; static int __init register_kernel_offset_dumper(void) { atomic_notifier_chain_register(&panic_notifier_list, &kernel_location_notifier); return 0; } __initcall(register_kernel_offset_dumper);