13d083395SSteven Rostedt /* 23d083395SSteven Rostedt * Code for replacing ftrace calls with jumps. 33d083395SSteven Rostedt * 43d083395SSteven Rostedt * Copyright (C) 2007-2008 Steven Rostedt <srostedt@redhat.com> 53d083395SSteven Rostedt * 63d083395SSteven Rostedt * Thanks goes to Ingo Molnar, for suggesting the idea. 73d083395SSteven Rostedt * Mathieu Desnoyers, for suggesting postponing the modifications. 83d083395SSteven Rostedt * Arjan van de Ven, for keeping me straight, and explaining to me 93d083395SSteven Rostedt * the dangers of modifying code on the run. 103d083395SSteven Rostedt */ 113d083395SSteven Rostedt 123d083395SSteven Rostedt #include <linux/spinlock.h> 133d083395SSteven Rostedt #include <linux/hardirq.h> 146f93fc07SSteven Rostedt #include <linux/uaccess.h> 153d083395SSteven Rostedt #include <linux/ftrace.h> 163d083395SSteven Rostedt #include <linux/percpu.h> 173d083395SSteven Rostedt #include <linux/init.h> 183d083395SSteven Rostedt #include <linux/list.h> 193d083395SSteven Rostedt 20395a59d0SAbhishek Sagar #include <asm/ftrace.h> 21732f3ca7SSteven Rostedt #include <asm/nops.h> 22dfa60abaSSteven Rostedt 233d083395SSteven Rostedt 24dfa60abaSSteven Rostedt /* Long is fine, even if it is only 4 bytes ;-) */ 2537a52f5eSHarvey Harrison static unsigned long *ftrace_nop; 263d083395SSteven Rostedt 273d083395SSteven Rostedt union ftrace_code_union { 28395a59d0SAbhishek Sagar char code[MCOUNT_INSN_SIZE]; 293d083395SSteven Rostedt struct { 303d083395SSteven Rostedt char e8; 313d083395SSteven Rostedt int offset; 323d083395SSteven Rostedt } __attribute__((packed)); 333d083395SSteven Rostedt }; 343d083395SSteven Rostedt 35395a59d0SAbhishek Sagar 363c1720f0SSteven Rostedt static int notrace ftrace_calc_offset(long ip, long addr) 373c1720f0SSteven Rostedt { 383c1720f0SSteven Rostedt return (int)(addr - ip); 393d083395SSteven Rostedt } 403d083395SSteven Rostedt 413c1720f0SSteven Rostedt notrace unsigned char *ftrace_nop_replace(void) 423c1720f0SSteven Rostedt { 433c1720f0SSteven Rostedt return (char *)ftrace_nop; 443c1720f0SSteven Rostedt } 453c1720f0SSteven Rostedt 463c1720f0SSteven Rostedt notrace unsigned char *ftrace_call_replace(unsigned long ip, unsigned long addr) 473c1720f0SSteven Rostedt { 483c1720f0SSteven Rostedt static union ftrace_code_union calc; 493c1720f0SSteven Rostedt 503c1720f0SSteven Rostedt calc.e8 = 0xe8; 51395a59d0SAbhishek Sagar calc.offset = ftrace_calc_offset(ip + MCOUNT_INSN_SIZE, addr); 523c1720f0SSteven Rostedt 533c1720f0SSteven Rostedt /* 543c1720f0SSteven Rostedt * No locking needed, this must be called via kstop_machine 553c1720f0SSteven Rostedt * which in essence is like running on a uniprocessor machine. 563c1720f0SSteven Rostedt */ 573c1720f0SSteven Rostedt return calc.code; 583c1720f0SSteven Rostedt } 593c1720f0SSteven Rostedt 603c1720f0SSteven Rostedt notrace int 613d083395SSteven Rostedt ftrace_modify_code(unsigned long ip, unsigned char *old_code, 623d083395SSteven Rostedt unsigned char *new_code) 633d083395SSteven Rostedt { 646f93fc07SSteven Rostedt unsigned char replaced[MCOUNT_INSN_SIZE]; 653d083395SSteven Rostedt 663d083395SSteven Rostedt /* 673d083395SSteven Rostedt * Note: Due to modules and __init, code can 683d083395SSteven Rostedt * disappear and change, we need to protect against faulting 693d083395SSteven Rostedt * as well as code changing. 703d083395SSteven Rostedt * 713d083395SSteven Rostedt * No real locking needed, this code is run through 726f93fc07SSteven Rostedt * kstop_machine, or before SMP starts. 733d083395SSteven Rostedt */ 74*593eb8a2SSteven Rostedt if (__copy_from_user_inatomic(replaced, (char __user *)ip, 75*593eb8a2SSteven Rostedt MCOUNT_INSN_SIZE)) 76*593eb8a2SSteven Rostedt return -EFAULT; 776f93fc07SSteven Rostedt 786f93fc07SSteven Rostedt if (memcmp(replaced, old_code, MCOUNT_INSN_SIZE) != 0) 79*593eb8a2SSteven Rostedt return -EINVAL; 806f93fc07SSteven Rostedt 81*593eb8a2SSteven Rostedt if (__copy_to_user_inatomic((char __user *)ip, new_code, 82*593eb8a2SSteven Rostedt MCOUNT_INSN_SIZE)) 83*593eb8a2SSteven Rostedt return -EPERM; 846f93fc07SSteven Rostedt 853d083395SSteven Rostedt sync_core(); 863d083395SSteven Rostedt 876f93fc07SSteven Rostedt return 0; 883d083395SSteven Rostedt } 893d083395SSteven Rostedt 90d61f82d0SSteven Rostedt notrace int ftrace_update_ftrace_func(ftrace_func_t func) 91d61f82d0SSteven Rostedt { 92d61f82d0SSteven Rostedt unsigned long ip = (unsigned long)(&ftrace_call); 93395a59d0SAbhishek Sagar unsigned char old[MCOUNT_INSN_SIZE], *new; 94d61f82d0SSteven Rostedt int ret; 95d61f82d0SSteven Rostedt 96395a59d0SAbhishek Sagar memcpy(old, &ftrace_call, MCOUNT_INSN_SIZE); 97d61f82d0SSteven Rostedt new = ftrace_call_replace(ip, (unsigned long)func); 98d61f82d0SSteven Rostedt ret = ftrace_modify_code(ip, old, new); 99d61f82d0SSteven Rostedt 100d61f82d0SSteven Rostedt return ret; 101d61f82d0SSteven Rostedt } 102d61f82d0SSteven Rostedt 103d61f82d0SSteven Rostedt notrace int ftrace_mcount_set(unsigned long *data) 104d61f82d0SSteven Rostedt { 1050a37605cSSteven Rostedt /* mcount is initialized as a nop */ 1060a37605cSSteven Rostedt *data = 0; 107d61f82d0SSteven Rostedt return 0; 108d61f82d0SSteven Rostedt } 109d61f82d0SSteven Rostedt 110d61f82d0SSteven Rostedt int __init ftrace_dyn_arch_init(void *data) 1113d083395SSteven Rostedt { 112732f3ca7SSteven Rostedt extern const unsigned char ftrace_test_p6nop[]; 113732f3ca7SSteven Rostedt extern const unsigned char ftrace_test_nop5[]; 114732f3ca7SSteven Rostedt extern const unsigned char ftrace_test_jmp[]; 115732f3ca7SSteven Rostedt int faulted = 0; 1163d083395SSteven Rostedt 117732f3ca7SSteven Rostedt /* 118732f3ca7SSteven Rostedt * There is no good nop for all x86 archs. 119732f3ca7SSteven Rostedt * We will default to using the P6_NOP5, but first we 120732f3ca7SSteven Rostedt * will test to make sure that the nop will actually 121732f3ca7SSteven Rostedt * work on this CPU. If it faults, we will then 122732f3ca7SSteven Rostedt * go to a lesser efficient 5 byte nop. If that fails 123732f3ca7SSteven Rostedt * we then just use a jmp as our nop. This isn't the most 124732f3ca7SSteven Rostedt * efficient nop, but we can not use a multi part nop 125732f3ca7SSteven Rostedt * since we would then risk being preempted in the middle 126732f3ca7SSteven Rostedt * of that nop, and if we enabled tracing then, it might 127732f3ca7SSteven Rostedt * cause a system crash. 128732f3ca7SSteven Rostedt * 129732f3ca7SSteven Rostedt * TODO: check the cpuid to determine the best nop. 130732f3ca7SSteven Rostedt */ 131732f3ca7SSteven Rostedt asm volatile ( 132732f3ca7SSteven Rostedt "jmp ftrace_test_jmp\n" 133732f3ca7SSteven Rostedt /* This code needs to stay around */ 134732f3ca7SSteven Rostedt ".section .text, \"ax\"\n" 135732f3ca7SSteven Rostedt "ftrace_test_jmp:" 136732f3ca7SSteven Rostedt "jmp ftrace_test_p6nop\n" 1378b27386aSAnders Kaseorg "nop\n" 1388b27386aSAnders Kaseorg "nop\n" 1398b27386aSAnders Kaseorg "nop\n" /* 2 byte jmp + 3 bytes */ 140732f3ca7SSteven Rostedt "ftrace_test_p6nop:" 141732f3ca7SSteven Rostedt P6_NOP5 142732f3ca7SSteven Rostedt "jmp 1f\n" 143732f3ca7SSteven Rostedt "ftrace_test_nop5:" 144732f3ca7SSteven Rostedt ".byte 0x66,0x66,0x66,0x66,0x90\n" 145732f3ca7SSteven Rostedt "jmp 1f\n" 146732f3ca7SSteven Rostedt ".previous\n" 147732f3ca7SSteven Rostedt "1:" 148732f3ca7SSteven Rostedt ".section .fixup, \"ax\"\n" 149732f3ca7SSteven Rostedt "2: movl $1, %0\n" 150732f3ca7SSteven Rostedt " jmp ftrace_test_nop5\n" 151732f3ca7SSteven Rostedt "3: movl $2, %0\n" 152732f3ca7SSteven Rostedt " jmp 1b\n" 153732f3ca7SSteven Rostedt ".previous\n" 154732f3ca7SSteven Rostedt _ASM_EXTABLE(ftrace_test_p6nop, 2b) 155732f3ca7SSteven Rostedt _ASM_EXTABLE(ftrace_test_nop5, 3b) 156732f3ca7SSteven Rostedt : "=r"(faulted) : "0" (faulted)); 157d61f82d0SSteven Rostedt 158732f3ca7SSteven Rostedt switch (faulted) { 159732f3ca7SSteven Rostedt case 0: 160732f3ca7SSteven Rostedt pr_info("ftrace: converting mcount calls to 0f 1f 44 00 00\n"); 161732f3ca7SSteven Rostedt ftrace_nop = (unsigned long *)ftrace_test_p6nop; 162732f3ca7SSteven Rostedt break; 163732f3ca7SSteven Rostedt case 1: 164732f3ca7SSteven Rostedt pr_info("ftrace: converting mcount calls to 66 66 66 66 90\n"); 165732f3ca7SSteven Rostedt ftrace_nop = (unsigned long *)ftrace_test_nop5; 166732f3ca7SSteven Rostedt break; 167732f3ca7SSteven Rostedt case 2: 1688b27386aSAnders Kaseorg pr_info("ftrace: converting mcount calls to jmp . + 5\n"); 169732f3ca7SSteven Rostedt ftrace_nop = (unsigned long *)ftrace_test_jmp; 170732f3ca7SSteven Rostedt break; 171732f3ca7SSteven Rostedt } 172d61f82d0SSteven Rostedt 173732f3ca7SSteven Rostedt /* The return code is retured via data */ 174732f3ca7SSteven Rostedt *(unsigned long *)data = 0; 175dfa60abaSSteven Rostedt 1763d083395SSteven Rostedt return 0; 1773d083395SSteven Rostedt } 178