| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * trace binary printk |
| * |
| * Copyright (C) 2008 Lai Jiangshan <laijs@cn.fujitsu.com> |
| * |
| */ |
| #include <linux/seq_file.h> |
| #include <linux/security.h> |
| #include <linux/uaccess.h> |
| #include <linux/kernel.h> |
| #include <linux/ftrace.h> |
| #include <linux/string.h> |
| #include <linux/module.h> |
| #include <linux/mutex.h> |
| #include <linux/ctype.h> |
| #include <linux/list.h> |
| #include <linux/slab.h> |
| |
| #include "trace.h" |
| |
| #ifdef CONFIG_MODULES |
| |
| /* |
| * modules trace_printk()'s formats are autosaved in struct trace_bprintk_fmt |
| * which are queued on trace_bprintk_fmt_list. |
| */ |
| static LIST_HEAD(trace_bprintk_fmt_list); |
| |
| /* serialize accesses to trace_bprintk_fmt_list */ |
| static DEFINE_MUTEX(btrace_mutex); |
| |
| struct trace_bprintk_fmt { |
| struct list_head list; |
| const char *fmt; |
| }; |
| |
| static inline struct trace_bprintk_fmt *lookup_format(const char *fmt) |
| { |
| struct trace_bprintk_fmt *pos; |
| |
| if (!fmt) |
| return ERR_PTR(-EINVAL); |
| |
| list_for_each_entry(pos, &trace_bprintk_fmt_list, list) { |
| if (!strcmp(pos->fmt, fmt)) |
| return pos; |
| } |
| return NULL; |
| } |
| |
| static |
| void hold_module_trace_bprintk_format(const char **start, const char **end) |
| { |
| const char **iter; |
| char *fmt; |
| |
| /* allocate the trace_printk per cpu buffers */ |
| if (start != end) |
| trace_printk_init_buffers(); |
| |
| mutex_lock(&btrace_mutex); |
| for (iter = start; iter < end; iter++) { |
| struct trace_bprintk_fmt *tb_fmt = lookup_format(*iter); |
| if (tb_fmt) { |
| if (!IS_ERR(tb_fmt)) |
| *iter = tb_fmt->fmt; |
| continue; |
| } |
| |
| fmt = NULL; |
| tb_fmt = kmalloc_obj(*tb_fmt); |
| if (tb_fmt) { |
| fmt = kmalloc(strlen(*iter) + 1, GFP_KERNEL); |
| if (fmt) { |
| list_add_tail(&tb_fmt->list, &trace_bprintk_fmt_list); |
| strcpy(fmt, *iter); |
| tb_fmt->fmt = fmt; |
| } else |
| kfree(tb_fmt); |
| } |
| *iter = fmt; |
| |
| } |
| mutex_unlock(&btrace_mutex); |
| } |
| |
| static int module_trace_bprintk_format_notify(struct notifier_block *self, |
| unsigned long val, void *data) |
| { |
| struct module *mod = data; |
| if (mod->num_trace_bprintk_fmt) { |
| const char **start = mod->trace_bprintk_fmt_start; |
| const char **end = start + mod->num_trace_bprintk_fmt; |
| |
| if (val == MODULE_STATE_COMING) |
| hold_module_trace_bprintk_format(start, end); |
| } |
| return NOTIFY_OK; |
| } |
| |
| /* |
| * The debugfs/tracing/printk_formats file maps the addresses with |
| * the ASCII formats that are used in the bprintk events in the |
| * buffer. For userspace tools to be able to decode the events from |
| * the buffer, they need to be able to map the address with the format. |
| * |
| * The addresses of the bprintk formats are in their own section |
| * __trace_printk_fmt. But for modules we copy them into a link list. |
| * The code to print the formats and their addresses passes around the |
| * address of the fmt string. If the fmt address passed into the seq |
| * functions is within the kernel core __trace_printk_fmt section, then |
| * it simply uses the next pointer in the list. |
| * |
| * When the fmt pointer is outside the kernel core __trace_printk_fmt |
| * section, then we need to read the link list pointers. The trick is |
| * we pass the address of the string to the seq function just like |
| * we do for the kernel core formats. To get back the structure that |
| * holds the format, we simply use container_of() and then go to the |
| * next format in the list. |
| */ |
| static const char ** |
| find_next_mod_format(int start_index, void *v, const char **fmt, loff_t *pos) |
| { |
| struct trace_bprintk_fmt *mod_fmt; |
| |
| if (list_empty(&trace_bprintk_fmt_list)) |
| return NULL; |
| |
| /* |
| * v will point to the address of the fmt record from t_next |
| * v will be NULL from t_start. |
| * If this is the first pointer or called from start |
| * then we need to walk the list. |
| */ |
| if (!v || start_index == *pos) { |
| struct trace_bprintk_fmt *p; |
| |
| /* search the module list */ |
| list_for_each_entry(p, &trace_bprintk_fmt_list, list) { |
| if (start_index == *pos) |
| return &p->fmt; |
| start_index++; |
| } |
| /* pos > index */ |
| return NULL; |
| } |
| |
| /* |
| * v points to the address of the fmt field in the mod list |
| * structure that holds the module print format. |
| */ |
| mod_fmt = container_of(v, typeof(*mod_fmt), fmt); |
| if (mod_fmt->list.next == &trace_bprintk_fmt_list) |
| return NULL; |
| |
| mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list); |
| |
| return &mod_fmt->fmt; |
| } |
| |
| static void format_mod_start(void) |
| { |
| mutex_lock(&btrace_mutex); |
| } |
| |
| static void format_mod_stop(void) |
| { |
| mutex_unlock(&btrace_mutex); |
| } |
| |
| #else /* !CONFIG_MODULES */ |
| __init static int |
| module_trace_bprintk_format_notify(struct notifier_block *self, |
| unsigned long val, void *data) |
| { |
| return NOTIFY_OK; |
| } |
| static inline const char ** |
| find_next_mod_format(int start_index, void *v, const char **fmt, loff_t *pos) |
| { |
| return NULL; |
| } |
| static inline void format_mod_start(void) { } |
| static inline void format_mod_stop(void) { } |
| #endif /* CONFIG_MODULES */ |
| |
| static bool __read_mostly trace_printk_enabled = true; |
| |
| void trace_printk_control(bool enabled) |
| { |
| trace_printk_enabled = enabled; |
| } |
| |
| __initdata_or_module static |
| struct notifier_block module_trace_bprintk_format_nb = { |
| .notifier_call = module_trace_bprintk_format_notify, |
| }; |
| |
| int __trace_bprintk(unsigned long ip, const char *fmt, ...) |
| { |
| int ret; |
| va_list ap; |
| |
| if (unlikely(!fmt)) |
| return 0; |
| |
| if (!trace_printk_enabled) |
| return 0; |
| |
| va_start(ap, fmt); |
| ret = trace_vbprintk(ip, fmt, ap); |
| va_end(ap); |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(__trace_bprintk); |
| |
| int __ftrace_vbprintk(unsigned long ip, const char *fmt, va_list ap) |
| { |
| if (unlikely(!fmt)) |
| return 0; |
| |
| if (!trace_printk_enabled) |
| return 0; |
| |
| return trace_vbprintk(ip, fmt, ap); |
| } |
| EXPORT_SYMBOL_GPL(__ftrace_vbprintk); |
| |
| int __trace_printk(unsigned long ip, const char *fmt, ...) |
| { |
| int ret; |
| va_list ap; |
| |
| if (!trace_printk_enabled) |
| return 0; |
| |
| va_start(ap, fmt); |
| ret = trace_vprintk(ip, fmt, ap); |
| va_end(ap); |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(__trace_printk); |
| |
| int __ftrace_vprintk(unsigned long ip, const char *fmt, va_list ap) |
| { |
| if (!trace_printk_enabled) |
| return 0; |
| |
| return trace_vprintk(ip, fmt, ap); |
| } |
| EXPORT_SYMBOL_GPL(__ftrace_vprintk); |
| |
| bool trace_is_tracepoint_string(const char *str) |
| { |
| const char **ptr = __start___tracepoint_str; |
| |
| for (ptr = __start___tracepoint_str; ptr < __stop___tracepoint_str; ptr++) { |
| if (str == *ptr) |
| return true; |
| } |
| return false; |
| } |
| |
| static const char **find_next(void *v, loff_t *pos) |
| { |
| const char **fmt = v; |
| int start_index; |
| int last_index; |
| |
| start_index = __stop___trace_bprintk_fmt - __start___trace_bprintk_fmt; |
| |
| if (*pos < start_index) |
| return __start___trace_bprintk_fmt + *pos; |
| |
| /* |
| * The __tracepoint_str section is treated the same as the |
| * __trace_printk_fmt section. The difference is that the |
| * __trace_printk_fmt section should only be used by trace_printk() |
| * in a debugging environment, as if anything exists in that section |
| * the trace_prink() helper buffers are allocated, which would just |
| * waste space in a production environment. |
| * |
| * The __tracepoint_str sections on the other hand are used by |
| * tracepoints which need to map pointers to their strings to |
| * the ASCII text for userspace. |
| */ |
| last_index = start_index; |
| start_index = __stop___tracepoint_str - __start___tracepoint_str; |
| |
| if (*pos < last_index + start_index) |
| return __start___tracepoint_str + (*pos - last_index); |
| |
| start_index += last_index; |
| return find_next_mod_format(start_index, v, fmt, pos); |
| } |
| |
| static void * |
| t_start(struct seq_file *m, loff_t *pos) |
| { |
| format_mod_start(); |
| return find_next(NULL, pos); |
| } |
| |
| static void *t_next(struct seq_file *m, void * v, loff_t *pos) |
| { |
| (*pos)++; |
| return find_next(v, pos); |
| } |
| |
| static int t_show(struct seq_file *m, void *v) |
| { |
| const char **fmt = v; |
| const char *str = *fmt; |
| int i; |
| |
| if (!*fmt) |
| return 0; |
| |
| seq_printf(m, "0x%lx : \"", *(unsigned long *)fmt); |
| |
| /* |
| * Tabs and new lines need to be converted. |
| */ |
| for (i = 0; str[i]; i++) { |
| switch (str[i]) { |
| case '\n': |
| seq_puts(m, "\\n"); |
| break; |
| case '\t': |
| seq_puts(m, "\\t"); |
| break; |
| case '\\': |
| seq_putc(m, '\\'); |
| break; |
| case '"': |
| seq_puts(m, "\\\""); |
| break; |
| default: |
| seq_putc(m, str[i]); |
| } |
| } |
| seq_puts(m, "\"\n"); |
| |
| return 0; |
| } |
| |
| static void t_stop(struct seq_file *m, void *p) |
| { |
| format_mod_stop(); |
| } |
| |
| static const struct seq_operations show_format_seq_ops = { |
| .start = t_start, |
| .next = t_next, |
| .show = t_show, |
| .stop = t_stop, |
| }; |
| |
| static int |
| ftrace_formats_open(struct inode *inode, struct file *file) |
| { |
| int ret; |
| |
| ret = security_locked_down(LOCKDOWN_TRACEFS); |
| if (ret) |
| return ret; |
| |
| return seq_open(file, &show_format_seq_ops); |
| } |
| |
| static const struct file_operations ftrace_formats_fops = { |
| .open = ftrace_formats_open, |
| .read = seq_read, |
| .llseek = seq_lseek, |
| .release = seq_release, |
| }; |
| |
| static __always_inline bool printk_binsafe(struct trace_array *tr) |
| { |
| /* |
| * The binary format of traceprintk can cause a crash if used |
| * by a buffer from another boot. Force the use of the |
| * non binary version of trace_printk if the trace_printk |
| * buffer is a boot mapped ring buffer. |
| */ |
| return !(tr->flags & TRACE_ARRAY_FL_BOOT); |
| } |
| |
| int __trace_array_puts(struct trace_array *tr, unsigned long ip, |
| const char *str, int size) |
| { |
| struct ring_buffer_event *event; |
| struct trace_buffer *buffer; |
| struct print_entry *entry; |
| unsigned int trace_ctx; |
| int alloc; |
| |
| if (!(tr->trace_flags & TRACE_ITER(PRINTK))) |
| return 0; |
| |
| if (unlikely(tracing_selftest_running && |
| (tr->flags & TRACE_ARRAY_FL_GLOBAL))) |
| return 0; |
| |
| if (unlikely(tracing_disabled)) |
| return 0; |
| |
| alloc = sizeof(*entry) + size + 2; /* possible \n added */ |
| |
| trace_ctx = tracing_gen_ctx(); |
| buffer = tr->array_buffer.buffer; |
| guard(ring_buffer_nest)(buffer); |
| event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc, |
| trace_ctx); |
| if (!event) |
| return 0; |
| |
| entry = ring_buffer_event_data(event); |
| entry->ip = ip; |
| |
| memcpy(&entry->buf, str, size); |
| |
| /* Add a newline if necessary */ |
| if (entry->buf[size - 1] != '\n') { |
| entry->buf[size] = '\n'; |
| entry->buf[size + 1] = '\0'; |
| } else |
| entry->buf[size] = '\0'; |
| |
| __buffer_unlock_commit(buffer, event); |
| ftrace_trace_stack(tr, buffer, trace_ctx, 4, NULL); |
| return size; |
| } |
| EXPORT_SYMBOL_GPL(__trace_array_puts); |
| |
| /** |
| * __trace_puts - write a constant string into the trace buffer. |
| * @ip: The address of the caller |
| * @str: The constant string to write |
| */ |
| int __trace_puts(unsigned long ip, const char *str) |
| { |
| return __trace_array_puts(printk_trace, ip, str, strlen(str)); |
| } |
| EXPORT_SYMBOL_GPL(__trace_puts); |
| |
| /** |
| * __trace_bputs - write the pointer to a constant string into trace buffer |
| * @ip: The address of the caller |
| * @str: The constant string to write to the buffer to |
| */ |
| int __trace_bputs(unsigned long ip, const char *str) |
| { |
| struct trace_array *tr = READ_ONCE(printk_trace); |
| struct ring_buffer_event *event; |
| struct trace_buffer *buffer; |
| struct bputs_entry *entry; |
| unsigned int trace_ctx; |
| int size = sizeof(struct bputs_entry); |
| |
| if (!printk_binsafe(tr)) |
| return __trace_puts(ip, str); |
| |
| if (!(tr->trace_flags & TRACE_ITER(PRINTK))) |
| return 0; |
| |
| if (unlikely(tracing_selftest_running || tracing_disabled)) |
| return 0; |
| |
| trace_ctx = tracing_gen_ctx(); |
| buffer = tr->array_buffer.buffer; |
| |
| guard(ring_buffer_nest)(buffer); |
| event = __trace_buffer_lock_reserve(buffer, TRACE_BPUTS, size, |
| trace_ctx); |
| if (!event) |
| return 0; |
| |
| entry = ring_buffer_event_data(event); |
| entry->ip = ip; |
| entry->str = str; |
| |
| __buffer_unlock_commit(buffer, event); |
| ftrace_trace_stack(tr, buffer, trace_ctx, 4, NULL); |
| |
| return 1; |
| } |
| EXPORT_SYMBOL_GPL(__trace_bputs); |
| |
| /* created for use with alloc_percpu */ |
| struct trace_buffer_struct { |
| int nesting; |
| char buffer[4][TRACE_BUF_SIZE]; |
| }; |
| |
| static struct trace_buffer_struct __percpu *trace_percpu_buffer; |
| |
| /* |
| * This allows for lockless recording. If we're nested too deeply, then |
| * this returns NULL. |
| */ |
| static char *get_trace_buf(void) |
| { |
| struct trace_buffer_struct *buffer = this_cpu_ptr(trace_percpu_buffer); |
| |
| if (!trace_percpu_buffer || buffer->nesting >= 4) |
| return NULL; |
| |
| buffer->nesting++; |
| |
| /* Interrupts must see nesting incremented before we use the buffer */ |
| barrier(); |
| return &buffer->buffer[buffer->nesting - 1][0]; |
| } |
| |
| static void put_trace_buf(void) |
| { |
| /* Don't let the decrement of nesting leak before this */ |
| barrier(); |
| this_cpu_dec(trace_percpu_buffer->nesting); |
| } |
| |
| static int alloc_percpu_trace_buffer(void) |
| { |
| struct trace_buffer_struct __percpu *buffers; |
| |
| if (trace_percpu_buffer) |
| return 0; |
| |
| buffers = alloc_percpu(struct trace_buffer_struct); |
| if (MEM_FAIL(!buffers, "Could not allocate percpu trace_printk buffer")) |
| return -ENOMEM; |
| |
| trace_percpu_buffer = buffers; |
| return 0; |
| } |
| |
| static int buffers_allocated; |
| |
| void trace_printk_init_buffers(void) |
| { |
| if (buffers_allocated) |
| return; |
| |
| if (alloc_percpu_trace_buffer()) |
| return; |
| |
| /* trace_printk() is for debug use only. Don't use it in production. */ |
| |
| pr_warn("\n"); |
| pr_warn("**********************************************************\n"); |
| pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); |
| pr_warn("** **\n"); |
| pr_warn("** trace_printk() being used. Allocating extra memory. **\n"); |
| pr_warn("** **\n"); |
| pr_warn("** This means that this is a DEBUG kernel and it is **\n"); |
| pr_warn("** unsafe for production use. **\n"); |
| pr_warn("** **\n"); |
| pr_warn("** If you see this message and you are not debugging **\n"); |
| pr_warn("** the kernel, report this immediately to your vendor! **\n"); |
| pr_warn("** **\n"); |
| pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); |
| pr_warn("**********************************************************\n"); |
| |
| /* Expand the buffers to set size */ |
| if (tracing_update_buffers(NULL) < 0) |
| pr_err("Failed to expand tracing buffers for trace_printk() calls\n"); |
| else |
| buffers_allocated = 1; |
| |
| /* |
| * trace_printk_init_buffers() can be called by modules. |
| * If that happens, then we need to start cmdline recording |
| * directly here. |
| */ |
| if (system_state == SYSTEM_RUNNING) |
| tracing_start_cmdline_record(); |
| } |
| EXPORT_SYMBOL_GPL(trace_printk_init_buffers); |
| |
| void trace_printk_start_comm(void) |
| { |
| /* Start tracing comms if trace printk is set */ |
| if (!buffers_allocated) |
| return; |
| tracing_start_cmdline_record(); |
| } |
| |
| void trace_printk_start_stop_comm(int enabled) |
| { |
| if (!buffers_allocated) |
| return; |
| |
| if (enabled) |
| tracing_start_cmdline_record(); |
| else |
| tracing_stop_cmdline_record(); |
| } |
| |
| /** |
| * trace_vbprintk - write binary msg to tracing buffer |
| * @ip: The address of the caller |
| * @fmt: The string format to write to the buffer |
| * @args: Arguments for @fmt |
| */ |
| int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) |
| { |
| struct ring_buffer_event *event; |
| struct trace_buffer *buffer; |
| struct trace_array *tr = READ_ONCE(printk_trace); |
| struct bprint_entry *entry; |
| unsigned int trace_ctx; |
| char *tbuffer; |
| int len = 0, size; |
| |
| if (!printk_binsafe(tr)) |
| return trace_vprintk(ip, fmt, args); |
| |
| if (unlikely(tracing_selftest_running || tracing_disabled)) |
| return 0; |
| |
| /* Don't pollute graph traces with trace_vprintk internals */ |
| pause_graph_tracing(); |
| |
| trace_ctx = tracing_gen_ctx(); |
| guard(preempt_notrace)(); |
| |
| tbuffer = get_trace_buf(); |
| if (!tbuffer) { |
| len = 0; |
| goto out_nobuffer; |
| } |
| |
| len = vbin_printf((u32 *)tbuffer, TRACE_BUF_SIZE/sizeof(int), fmt, args); |
| |
| if (len > TRACE_BUF_SIZE/sizeof(int) || len < 0) |
| goto out_put; |
| |
| size = sizeof(*entry) + sizeof(u32) * len; |
| buffer = tr->array_buffer.buffer; |
| scoped_guard(ring_buffer_nest, buffer) { |
| event = __trace_buffer_lock_reserve(buffer, TRACE_BPRINT, size, |
| trace_ctx); |
| if (!event) |
| goto out_put; |
| entry = ring_buffer_event_data(event); |
| entry->ip = ip; |
| entry->fmt = fmt; |
| |
| memcpy(entry->buf, tbuffer, sizeof(u32) * len); |
| __buffer_unlock_commit(buffer, event); |
| ftrace_trace_stack(tr, buffer, trace_ctx, 6, NULL); |
| } |
| out_put: |
| put_trace_buf(); |
| |
| out_nobuffer: |
| unpause_graph_tracing(); |
| |
| return len; |
| } |
| EXPORT_SYMBOL_GPL(trace_vbprintk); |
| |
| static __printf(3, 0) |
| int __trace_array_vprintk(struct trace_buffer *buffer, |
| unsigned long ip, const char *fmt, va_list args) |
| { |
| struct ring_buffer_event *event; |
| int len = 0, size; |
| struct print_entry *entry; |
| unsigned int trace_ctx; |
| char *tbuffer; |
| |
| if (unlikely(tracing_disabled)) |
| return 0; |
| |
| /* Don't pollute graph traces with trace_vprintk internals */ |
| pause_graph_tracing(); |
| |
| trace_ctx = tracing_gen_ctx(); |
| guard(preempt_notrace)(); |
| |
| |
| tbuffer = get_trace_buf(); |
| if (!tbuffer) { |
| len = 0; |
| goto out_nobuffer; |
| } |
| |
| len = vscnprintf(tbuffer, TRACE_BUF_SIZE, fmt, args); |
| |
| size = sizeof(*entry) + len + 1; |
| scoped_guard(ring_buffer_nest, buffer) { |
| event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, |
| trace_ctx); |
| if (!event) |
| goto out; |
| entry = ring_buffer_event_data(event); |
| entry->ip = ip; |
| |
| memcpy(&entry->buf, tbuffer, len + 1); |
| __buffer_unlock_commit(buffer, event); |
| ftrace_trace_stack(printk_trace, buffer, trace_ctx, 6, NULL); |
| } |
| out: |
| put_trace_buf(); |
| |
| out_nobuffer: |
| unpause_graph_tracing(); |
| |
| return len; |
| } |
| |
| int trace_array_vprintk(struct trace_array *tr, |
| unsigned long ip, const char *fmt, va_list args) |
| { |
| if (tracing_selftest_running && (tr->flags & TRACE_ARRAY_FL_GLOBAL)) |
| return 0; |
| |
| return __trace_array_vprintk(tr->array_buffer.buffer, ip, fmt, args); |
| } |
| |
| /** |
| * trace_array_printk - Print a message to a specific instance |
| * @tr: The instance trace_array descriptor |
| * @ip: The instruction pointer that this is called from. |
| * @fmt: The format to print (printf format) |
| * |
| * If a subsystem sets up its own instance, they have the right to |
| * printk strings into their tracing instance buffer using this |
| * function. Note, this function will not write into the top level |
| * buffer (use trace_printk() for that), as writing into the top level |
| * buffer should only have events that can be individually disabled. |
| * trace_printk() is only used for debugging a kernel, and should not |
| * be ever incorporated in normal use. |
| * |
| * trace_array_printk() can be used, as it will not add noise to the |
| * top level tracing buffer. |
| * |
| * Note, trace_array_init_printk() must be called on @tr before this |
| * can be used. |
| */ |
| int trace_array_printk(struct trace_array *tr, |
| unsigned long ip, const char *fmt, ...) |
| { |
| int ret; |
| va_list ap; |
| |
| if (!tr) |
| return -ENOENT; |
| |
| /* This is only allowed for created instances */ |
| if (tr->flags & TRACE_ARRAY_FL_GLOBAL) |
| return 0; |
| |
| if (!(tr->trace_flags & TRACE_ITER(PRINTK))) |
| return 0; |
| |
| va_start(ap, fmt); |
| ret = trace_array_vprintk(tr, ip, fmt, ap); |
| va_end(ap); |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(trace_array_printk); |
| |
| /** |
| * trace_array_init_printk - Initialize buffers for trace_array_printk() |
| * @tr: The trace array to initialize the buffers for |
| * |
| * As trace_array_printk() only writes into instances, they are OK to |
| * have in the kernel (unlike trace_printk()). This needs to be called |
| * before trace_array_printk() can be used on a trace_array. |
| */ |
| int trace_array_init_printk(struct trace_array *tr) |
| { |
| if (!tr) |
| return -ENOENT; |
| |
| /* This is only allowed for created instances */ |
| if (tr->flags & TRACE_ARRAY_FL_GLOBAL) |
| return -EINVAL; |
| |
| return alloc_percpu_trace_buffer(); |
| } |
| EXPORT_SYMBOL_GPL(trace_array_init_printk); |
| |
| int trace_array_printk_buf(struct trace_buffer *buffer, |
| unsigned long ip, const char *fmt, ...) |
| { |
| int ret; |
| va_list ap; |
| |
| if (!(printk_trace->trace_flags & TRACE_ITER(PRINTK))) |
| return 0; |
| |
| va_start(ap, fmt); |
| ret = __trace_array_vprintk(buffer, ip, fmt, ap); |
| va_end(ap); |
| return ret; |
| } |
| |
| int trace_vprintk(unsigned long ip, const char *fmt, va_list args) |
| { |
| return trace_array_vprintk(printk_trace, ip, fmt, args); |
| } |
| EXPORT_SYMBOL_GPL(trace_vprintk); |
| |
| static __init int init_trace_printk_function_export(void) |
| { |
| int ret; |
| |
| ret = tracing_init_dentry(); |
| if (ret) |
| return 0; |
| |
| trace_create_file("printk_formats", TRACE_MODE_READ, NULL, |
| NULL, &ftrace_formats_fops); |
| |
| return 0; |
| } |
| |
| fs_initcall(init_trace_printk_function_export); |
| |
| static __init int init_trace_printk(void) |
| { |
| return register_module_notifier(&module_trace_bprintk_format_nb); |
| } |
| |
| early_initcall(init_trace_printk); |