/* * This file is part of the SPL: Solaris Porting Layer. * * This file was originally part of Lustre, http://www.lustre.org. * but has subsequently been adapted for use in the SPL in * accordance with the GPL. * * Copyright (C) 2004 Cluster File Systems, Inc. * Copyright (c) 2008 Lawrence Livermore National Security, LLC. * Produced at Lawrence Livermore National Laboratory * Written by: * Zach Brown * Phil Schwan * Brian Behlendorf , * Herb Wartens , * Jim Garlick * UCRL-CODE-235197 * * This is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This is distributed in the hope that it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License * for more details. * * You should have received a copy of the GNU General Public License along * with this program; if not, write to the Free Software Foundation, Inc., * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef DEBUG_SUBSYSTEM #undef DEBUG_SUBSYSTEM #endif #define DEBUG_SUBSYSTEM S_DEBUG unsigned long spl_debug_subsys = ~0; EXPORT_SYMBOL(spl_debug_subsys); module_param(spl_debug_subsys, long, 0644); MODULE_PARM_DESC(spl_debug_subsys, "Subsystem debugging level mask."); unsigned long spl_debug_mask = (D_EMERG | D_ERROR | D_WARNING | D_CONSOLE); EXPORT_SYMBOL(spl_debug_mask); module_param(spl_debug_mask, long, 0644); MODULE_PARM_DESC(spl_debug_mask, "Debugging level mask."); unsigned long spl_debug_printk = D_CANTMASK; EXPORT_SYMBOL(spl_debug_printk); module_param(spl_debug_printk, long, 0644); MODULE_PARM_DESC(spl_debug_printk, "Console printk level mask."); int spl_debug_mb = -1; EXPORT_SYMBOL(spl_debug_mb); module_param(spl_debug_mb, int, 0644); MODULE_PARM_DESC(spl_debug_mb, "Total debug buffer size."); unsigned int spl_debug_binary = 1; EXPORT_SYMBOL(spl_debug_binary); unsigned int spl_debug_catastrophe; EXPORT_SYMBOL(spl_debug_catastrophe); unsigned int spl_debug_panic_on_bug = 1; EXPORT_SYMBOL(spl_debug_panic_on_bug); module_param(spl_debug_panic_on_bug, int, 0644); MODULE_PARM_DESC(spl_debug_panic_on_bug, "Panic on BUG"); static char spl_debug_file_name[PATH_MAX]; char spl_debug_file_path[PATH_MAX] = "/var/dumps/spl-log"; unsigned int spl_console_ratelimit = 1; EXPORT_SYMBOL(spl_console_ratelimit); long spl_console_max_delay; EXPORT_SYMBOL(spl_console_max_delay); long spl_console_min_delay; EXPORT_SYMBOL(spl_console_min_delay); unsigned int spl_console_backoff = SPL_DEFAULT_BACKOFF; EXPORT_SYMBOL(spl_console_backoff); unsigned int spl_debug_stack; EXPORT_SYMBOL(spl_debug_stack); static int spl_panic_in_progress; union trace_data_union (*trace_data[TCD_TYPE_MAX])[NR_CPUS] __cacheline_aligned; char *trace_console_buffers[NR_CPUS][3]; struct rw_semaphore trace_sem; atomic_t trace_tage_allocated = ATOMIC_INIT(0); static int panic_notifier(struct notifier_block *, unsigned long, void *); static int spl_debug_dump_all_pages(dumplog_priv_t *dp, char *); static void trace_fini(void); /* Memory percentage breakdown by type */ static unsigned int pages_factor[TCD_TYPE_MAX] = { 80, /* 80% pages for TCD_TYPE_PROC */ 10, /* 10% pages for TCD_TYPE_SOFTIRQ */ 10 /* 10% pages for TCD_TYPE_IRQ */ }; static struct notifier_block spl_panic_notifier = { notifier_call: panic_notifier, next: NULL, priority: 10000 }; const char * spl_debug_subsys2str(int subsys) { switch (subsys) { default: return NULL; case S_UNDEFINED: return "undefined"; case S_ATOMIC: return "atomic"; case S_KOBJ: return "kobj"; case S_VNODE: return "vnode"; case S_TIME: return "time"; case S_RWLOCK: return "rwlock"; case S_THREAD: return "thread"; case S_CONDVAR: return "condvar"; case S_MUTEX: return "mutex"; case S_RNG: return "rng"; case S_TASKQ: return "taskq"; case S_KMEM: return "kmem"; case S_DEBUG: return "debug"; case S_GENERIC: return "generic"; case S_PROC: return "proc"; case S_MODULE: return "module"; } } const char * spl_debug_dbg2str(int debug) { switch (debug) { default: return NULL; case D_TRACE: return "trace"; case D_INFO: return "info"; case D_WARNING: return "warning"; case D_ERROR: return "error"; case D_EMERG: return "emerg"; case D_CONSOLE: return "console"; case D_IOCTL: return "ioctl"; case D_DPRINTF: return "dprintf"; case D_OTHER: return "other"; } } int spl_debug_mask2str(char *str, int size, unsigned long mask, int is_subsys) { const char *(*fn)(int bit) = is_subsys ? spl_debug_subsys2str : spl_debug_dbg2str; const char *token; int i, bit, len = 0; if (mask == 0) { /* "0" */ if (size > 0) str[0] = '0'; len = 1; } else { /* space-separated tokens */ for (i = 0; i < 32; i++) { bit = 1 << i; if ((mask & bit) == 0) continue; token = fn(bit); if (token == NULL) /* unused bit */ continue; if (len > 0) { /* separator? */ if (len < size) str[len] = ' '; len++; } while (*token != 0) { if (len < size) str[len] = *token; token++; len++; } } } /* terminate 'str' */ if (len < size) str[len] = 0; else str[size - 1] = 0; return len; } static int spl_debug_token2mask(int *mask, const char *str, int len, int is_subsys) { const char *(*fn)(int bit) = is_subsys ? spl_debug_subsys2str : spl_debug_dbg2str; const char *token; int i, j, bit; /* match against known tokens */ for (i = 0; i < 32; i++) { bit = 1 << i; token = fn(bit); if (token == NULL) /* unused? */ continue; /* strcasecmp */ for (j = 0; ; j++) { if (j == len) { /* end of token */ if (token[j] == 0) { *mask = bit; return 0; } break; } if (token[j] == 0) break; if (str[j] == token[j]) continue; if (str[j] < 'A' || 'Z' < str[j]) break; if (str[j] - 'A' + 'a' != token[j]) break; } } return -EINVAL; /* no match */ } int spl_debug_str2mask(unsigned long *mask, const char *str, int is_subsys) { char op = 0; int m = 0, matched, n, t; /* Allow a number for backwards compatibility */ for (n = strlen(str); n > 0; n--) if (!isspace(str[n-1])) break; matched = n; if ((t = sscanf(str, "%i%n", &m, &matched)) >= 1 && matched == n) { *mask = m; return 0; } /* must be a list of debug tokens or numbers separated by * whitespace and optionally an operator ('+' or '-'). If an operator * appears first in , '*mask' is used as the starting point * (relative), otherwise 0 is used (absolute). An operator applies to * all following tokens up to the next operator. */ matched = 0; while (*str != 0) { while (isspace(*str)) /* skip whitespace */ str++; if (*str == 0) break; if (*str == '+' || *str == '-') { op = *str++; /* op on first token == relative */ if (!matched) m = *mask; while (isspace(*str)) /* skip whitespace */ str++; if (*str == 0) /* trailing op */ return -EINVAL; } /* find token length */ for (n = 0; str[n] != 0 && !isspace(str[n]); n++); /* match token */ if (spl_debug_token2mask(&t, str, n, is_subsys) != 0) return -EINVAL; matched = 1; if (op == '-') m &= ~t; else m |= t; str += n; } if (!matched) return -EINVAL; *mask = m; return 0; } static void spl_debug_dumplog_internal(dumplog_priv_t *dp) { void *journal_info; journal_info = current->journal_info; current->journal_info = NULL; snprintf(spl_debug_file_name, sizeof(spl_debug_file_path) - 1, "%s.%ld.%ld", spl_debug_file_path, get_seconds(), (long)dp->dp_pid); printk(KERN_ALERT "SPL: dumping log to %s\n", spl_debug_file_name); spl_debug_dump_all_pages(dp, spl_debug_file_name); current->journal_info = journal_info; } static int spl_debug_dumplog_thread(void *arg) { dumplog_priv_t *dp = (dumplog_priv_t *)arg; spl_debug_dumplog_internal(dp); atomic_set(&dp->dp_done, 1); wake_up(&dp->dp_waitq); do_exit(0); return 0; /* Unreachable */ } /* When flag is set do not use a new thread for the debug dump */ int spl_debug_dumplog(int flags) { struct task_struct *tsk; dumplog_priv_t dp; init_waitqueue_head(&dp.dp_waitq); dp.dp_pid = current->pid; dp.dp_flags = flags; atomic_set(&dp.dp_done, 0); if (dp.dp_flags & DL_NOTHREAD) { spl_debug_dumplog_internal(&dp); } else { tsk = kthread_create(spl_debug_dumplog_thread,(void *)&dp,"spl_debug"); if (tsk == NULL) return -ENOMEM; wake_up_process(tsk); wait_event(dp.dp_waitq, atomic_read(&dp.dp_done)); } return 0; } EXPORT_SYMBOL(spl_debug_dumplog); static char * trace_get_console_buffer(void) { int cpu = get_cpu(); int idx; if (in_irq()) { idx = 0; } else if (in_softirq()) { idx = 1; } else { idx = 2; } return trace_console_buffers[cpu][idx]; } static void trace_put_console_buffer(char *buffer) { put_cpu(); } static struct trace_cpu_data * trace_get_tcd(void) { int cpu; cpu = get_cpu(); if (in_irq()) return &(*trace_data[TCD_TYPE_IRQ])[cpu].tcd; else if (in_softirq()) return &(*trace_data[TCD_TYPE_SOFTIRQ])[cpu].tcd; return &(*trace_data[TCD_TYPE_PROC])[cpu].tcd; } static void trace_put_tcd (struct trace_cpu_data *tcd) { put_cpu(); } static int trace_lock_tcd(struct trace_cpu_data *tcd) { __ASSERT(tcd->tcd_type < TCD_TYPE_MAX); if (tcd->tcd_type == TCD_TYPE_IRQ) local_irq_disable(); else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ) local_bh_disable(); return 1; } static void trace_unlock_tcd(struct trace_cpu_data *tcd) { __ASSERT(tcd->tcd_type < TCD_TYPE_MAX); if (tcd->tcd_type == TCD_TYPE_IRQ) local_irq_enable(); else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ) local_bh_enable(); } static void trace_set_debug_header(struct spl_debug_header *header, int subsys, int mask, const int line, unsigned long stack) { struct timeval tv; do_gettimeofday(&tv); header->ph_subsys = subsys; header->ph_mask = mask; header->ph_cpu_id = smp_processor_id(); header->ph_sec = (__u32)tv.tv_sec; header->ph_usec = tv.tv_usec; header->ph_stack = stack; header->ph_pid = current->pid; header->ph_line_num = line; return; } static void trace_print_to_console(struct spl_debug_header *hdr, int mask, const char *buf, int len, const char *file, const char *fn) { char *prefix = "SPL", *ptype = NULL; if ((mask & D_EMERG) != 0) { prefix = "SPLError"; ptype = KERN_EMERG; } else if ((mask & D_ERROR) != 0) { prefix = "SPLError"; ptype = KERN_ERR; } else if ((mask & D_WARNING) != 0) { prefix = "SPL"; ptype = KERN_WARNING; } else if ((mask & (D_CONSOLE | spl_debug_printk)) != 0) { prefix = "SPL"; ptype = KERN_INFO; } if ((mask & D_CONSOLE) != 0) { printk("%s%s: %.*s", ptype, prefix, len, buf); } else { printk("%s%s: %d:%d:(%s:%d:%s()) %.*s", ptype, prefix, hdr->ph_pid, hdr->ph_stack, file, hdr->ph_line_num, fn, len, buf); } return; } static int trace_max_debug_mb(void) { return MAX(512, ((num_physpages >> (20 - PAGE_SHIFT)) * 80) / 100); } static void trace_call_on_all_cpus(void (*fn)(void *arg), void *arg) { cpumask_t mask, cpus_allowed = current->cpus_allowed; int cpu; for_each_online_cpu(cpu) { cpus_clear(mask); cpu_set(cpu, mask); set_cpus_allowed(current, mask); fn(arg); set_cpus_allowed(current, cpus_allowed); } } static struct trace_page * tage_alloc(int gfp) { struct page *page; struct trace_page *tage; page = alloc_pages(gfp | __GFP_NOWARN, 0); if (page == NULL) return NULL; tage = kmalloc(sizeof(*tage), gfp); if (tage == NULL) { __free_pages(page, 0); return NULL; } tage->page = page; atomic_inc(&trace_tage_allocated); return tage; } static void tage_free(struct trace_page *tage) { __ASSERT(tage != NULL); __ASSERT(tage->page != NULL); __free_pages(tage->page, 0); kfree(tage); atomic_dec(&trace_tage_allocated); } static struct trace_page * tage_from_list(struct list_head *list) { return list_entry(list, struct trace_page, linkage); } static void tage_to_tail(struct trace_page *tage, struct list_head *queue) { __ASSERT(tage != NULL); __ASSERT(queue != NULL); list_move_tail(&tage->linkage, queue); } /* try to return a page that has 'len' bytes left at the end */ static struct trace_page * trace_get_tage_try(struct trace_cpu_data *tcd, unsigned long len) { struct trace_page *tage; if (tcd->tcd_cur_pages > 0) { __ASSERT(!list_empty(&tcd->tcd_pages)); tage = tage_from_list(tcd->tcd_pages.prev); if (tage->used + len <= PAGE_SIZE) return tage; } if (tcd->tcd_cur_pages < tcd->tcd_max_pages) { if (tcd->tcd_cur_stock_pages > 0) { tage = tage_from_list(tcd->tcd_stock_pages.prev); tcd->tcd_cur_stock_pages--; list_del_init(&tage->linkage); } else { tage = tage_alloc(GFP_ATOMIC); if (tage == NULL) { printk(KERN_WARNING "failure to allocate a tage (%ld)\n", tcd->tcd_cur_pages); return NULL; } } tage->used = 0; tage->cpu = smp_processor_id(); tage->type = tcd->tcd_type; list_add_tail(&tage->linkage, &tcd->tcd_pages); tcd->tcd_cur_pages++; return tage; } return NULL; } /* return a page that has 'len' bytes left at the end */ static struct trace_page * trace_get_tage(struct trace_cpu_data *tcd, unsigned long len) { struct trace_page *tage; __ASSERT(len <= PAGE_SIZE); tage = trace_get_tage_try(tcd, len); if (tage) return tage; if (tcd->tcd_cur_pages > 0) { tage = tage_from_list(tcd->tcd_pages.next); tage->used = 0; tage_to_tail(tage, &tcd->tcd_pages); } return tage; } int spl_debug_vmsg(spl_debug_limit_state_t *cdls, int subsys, int mask, const char *file, const char *fn, const int line, const char *format1, va_list args, const char *format2, ...) { struct trace_cpu_data *tcd = NULL; struct spl_debug_header header; struct trace_page *tage; /* string_buf is used only if tcd != NULL, and is always set then */ char *string_buf = NULL; char *debug_buf; int known_size; int needed = 85; /* average message length */ int max_nob; va_list ap; int i; int remain; if (strchr(file, '/')) file = strrchr(file, '/') + 1; trace_set_debug_header(&header, subsys, mask, line, CDEBUG_STACK()); tcd = trace_get_tcd(); if (tcd == NULL) goto console; if (tcd->tcd_shutting_down) { trace_put_tcd(tcd); tcd = NULL; goto console; } known_size = strlen(file) + 1; if (fn) known_size += strlen(fn) + 1; if (spl_debug_binary) known_size += sizeof(header); /* '2' used because vsnprintf returns real size required for output * _without_ terminating NULL. */ for (i = 0; i < 2; i++) { tage = trace_get_tage(tcd, needed + known_size + 1); if (tage == NULL) { if (needed + known_size > PAGE_SIZE) mask |= D_ERROR; trace_put_tcd(tcd); tcd = NULL; goto console; } string_buf = (char *)page_address(tage->page) + tage->used + known_size; max_nob = PAGE_SIZE - tage->used - known_size; if (max_nob <= 0) { printk(KERN_EMERG "negative max_nob: %i\n", max_nob); mask |= D_ERROR; trace_put_tcd(tcd); tcd = NULL; goto console; } needed = 0; if (format1) { va_copy(ap, args); needed = vsnprintf(string_buf, max_nob, format1, ap); va_end(ap); } if (format2) { remain = max_nob - needed; if (remain < 0) remain = 0; va_start(ap, format2); needed += vsnprintf(string_buf+needed, remain, format2, ap); va_end(ap); } if (needed < max_nob) break; } if (unlikely(*(string_buf + needed - 1) != '\n')) printk(KERN_INFO "format at %s:%d:%s doesn't end in newline\n", file, line, fn); header.ph_len = known_size + needed; debug_buf = (char *)page_address(tage->page) + tage->used; if (spl_debug_binary) { memcpy(debug_buf, &header, sizeof(header)); tage->used += sizeof(header); debug_buf += sizeof(header); } strcpy(debug_buf, file); tage->used += strlen(file) + 1; debug_buf += strlen(file) + 1; if (fn) { strcpy(debug_buf, fn); tage->used += strlen(fn) + 1; debug_buf += strlen(fn) + 1; } __ASSERT(debug_buf == string_buf); tage->used += needed; __ASSERT (tage->used <= PAGE_SIZE); console: if ((mask & spl_debug_printk) == 0) { /* no console output requested */ if (tcd != NULL) trace_put_tcd(tcd); return 1; } if (cdls != NULL) { if (spl_console_ratelimit && cdls->cdls_next != 0 && !time_before(cdls->cdls_next, jiffies)) { /* skipping a console message */ cdls->cdls_count++; if (tcd != NULL) trace_put_tcd(tcd); return 1; } if (time_before(cdls->cdls_next + spl_console_max_delay + (10 * HZ), jiffies)) { /* last timeout was a long time ago */ cdls->cdls_delay /= spl_console_backoff * 4; } else { cdls->cdls_delay *= spl_console_backoff; if (cdls->cdls_delay < spl_console_min_delay) cdls->cdls_delay = spl_console_min_delay; else if (cdls->cdls_delay > spl_console_max_delay) cdls->cdls_delay = spl_console_max_delay; } /* ensure cdls_next is never zero after it's been seen */ cdls->cdls_next = (jiffies + cdls->cdls_delay) | 1; } if (tcd != NULL) { trace_print_to_console(&header, mask, string_buf, needed, file, fn); trace_put_tcd(tcd); } else { string_buf = trace_get_console_buffer(); needed = 0; if (format1 != NULL) { va_copy(ap, args); needed = vsnprintf(string_buf, TRACE_CONSOLE_BUFFER_SIZE, format1, ap); va_end(ap); } if (format2 != NULL) { remain = TRACE_CONSOLE_BUFFER_SIZE - needed; if (remain > 0) { va_start(ap, format2); needed += vsnprintf(string_buf+needed, remain, format2, ap); va_end(ap); } } trace_print_to_console(&header, mask, string_buf, needed, file, fn); trace_put_console_buffer(string_buf); } if (cdls != NULL && cdls->cdls_count != 0) { string_buf = trace_get_console_buffer(); needed = snprintf(string_buf, TRACE_CONSOLE_BUFFER_SIZE, "Skipped %d previous similar message%s\n", cdls->cdls_count, (cdls->cdls_count > 1) ? "s" : ""); trace_print_to_console(&header, mask, string_buf, needed, file, fn); trace_put_console_buffer(string_buf); cdls->cdls_count = 0; } return 0; } EXPORT_SYMBOL(spl_debug_vmsg); /* Do the collect_pages job on a single CPU: assumes that all other * CPUs have been stopped during a panic. If this isn't true for * some arch, this will have to be implemented separately in each arch. */ static void collect_pages_from_single_cpu(struct page_collection *pc) { struct trace_cpu_data *tcd; int i, j; tcd_for_each(tcd, i, j) { list_splice_init(&tcd->tcd_pages, &pc->pc_pages); tcd->tcd_cur_pages = 0; } } static void collect_pages_on_cpu(void *info) { struct trace_cpu_data *tcd; struct page_collection *pc = info; int i; spin_lock(&pc->pc_lock); tcd_for_each_type_lock(tcd, i) { list_splice_init(&tcd->tcd_pages, &pc->pc_pages); tcd->tcd_cur_pages = 0; } spin_unlock(&pc->pc_lock); } static void collect_pages(dumplog_priv_t *dp, struct page_collection *pc) { INIT_LIST_HEAD(&pc->pc_pages); if (spl_panic_in_progress || dp->dp_flags & DL_SINGLE_CPU) collect_pages_from_single_cpu(pc); else trace_call_on_all_cpus(collect_pages_on_cpu, pc); } static void put_pages_back_on_cpu(void *info) { struct page_collection *pc = info; struct trace_cpu_data *tcd; struct list_head *cur_head; struct trace_page *tage; struct trace_page *tmp; int i; spin_lock(&pc->pc_lock); tcd_for_each_type_lock(tcd, i) { cur_head = tcd->tcd_pages.next; list_for_each_entry_safe(tage, tmp, &pc->pc_pages, linkage) { __ASSERT_TAGE_INVARIANT(tage); if (tage->cpu != smp_processor_id() || tage->type != i) continue; tage_to_tail(tage, cur_head); tcd->tcd_cur_pages++; } } spin_unlock(&pc->pc_lock); } static void put_pages_back(struct page_collection *pc) { if (!spl_panic_in_progress) trace_call_on_all_cpus(put_pages_back_on_cpu, pc); } static struct file * trace_filp_open (const char *name, int flags, int mode, int *err) { struct file *filp = NULL; int rc; filp = filp_open(name, flags, mode); if (IS_ERR(filp)) { rc = PTR_ERR(filp); printk(KERN_ERR "SPL: Can't open %s file: %d\n", name, rc); if (err) *err = rc; filp = NULL; } return filp; } #define trace_filp_write(fp, b, s, p) (fp)->f_op->write((fp), (b), (s), p) #define trace_filp_fsync(fp) (fp)->f_op->fsync((fp),(fp)->f_dentry,1) #define trace_filp_close(f) filp_close(f, NULL) #define trace_filp_poff(f) (&(f)->f_pos) static int spl_debug_dump_all_pages(dumplog_priv_t *dp, char *filename) { struct page_collection pc; struct file *filp; struct trace_page *tage; struct trace_page *tmp; mm_segment_t oldfs; int rc = 0; down_write(&trace_sem); filp = trace_filp_open(filename, O_CREAT|O_EXCL|O_WRONLY|O_LARGEFILE, 0600, &rc); if (filp == NULL) { if (rc != -EEXIST) printk(KERN_ERR "SPL: Can't open %s for dump: %d\n", filename, rc); goto out; } spin_lock_init(&pc.pc_lock); collect_pages(dp, &pc); if (list_empty(&pc.pc_pages)) { rc = 0; goto close; } oldfs = get_fs(); set_fs(get_ds()); list_for_each_entry_safe(tage, tmp, &pc.pc_pages, linkage) { __ASSERT_TAGE_INVARIANT(tage); rc = trace_filp_write(filp, page_address(tage->page), tage->used, trace_filp_poff(filp)); if (rc != (int)tage->used) { printk(KERN_WARNING "SPL: Wanted to write %u " "but wrote %d\n", tage->used, rc); put_pages_back(&pc); __ASSERT(list_empty(&pc.pc_pages)); break; } list_del(&tage->linkage); tage_free(tage); } set_fs(oldfs); rc = trace_filp_fsync(filp); if (rc) printk(KERN_ERR "SPL: Unable to sync: %d\n", rc); close: trace_filp_close(filp); out: up_write(&trace_sem); return rc; } static void spl_debug_flush_pages(void) { dumplog_priv_t dp; struct page_collection pc; struct trace_page *tage; struct trace_page *tmp; spin_lock_init(&pc.pc_lock); init_waitqueue_head(&dp.dp_waitq); dp.dp_pid = current->pid; dp.dp_flags = 0; atomic_set(&dp.dp_done, 0); collect_pages(&dp, &pc); list_for_each_entry_safe(tage, tmp, &pc.pc_pages, linkage) { __ASSERT_TAGE_INVARIANT(tage); list_del(&tage->linkage); tage_free(tage); } } unsigned long spl_debug_set_mask(unsigned long mask) { spl_debug_mask = mask; return 0; } EXPORT_SYMBOL(spl_debug_set_mask); unsigned long spl_debug_get_mask(void) { return spl_debug_mask; } EXPORT_SYMBOL(spl_debug_get_mask); unsigned long spl_debug_set_subsys(unsigned long subsys) { spl_debug_subsys = subsys; return 0; } EXPORT_SYMBOL(spl_debug_set_subsys); unsigned long spl_debug_get_subsys(void) { return spl_debug_subsys; } EXPORT_SYMBOL(spl_debug_get_subsys); int spl_debug_set_mb(int mb) { int i, j, pages; int limit = trace_max_debug_mb(); struct trace_cpu_data *tcd; if (mb < num_possible_cpus()) { printk(KERN_ERR "SPL: Refusing to set debug buffer size to " "%dMB - lower limit is %d\n", mb, num_possible_cpus()); return -EINVAL; } if (mb > limit) { printk(KERN_ERR "SPL: Refusing to set debug buffer size to " "%dMB - upper limit is %d\n", mb, limit); return -EINVAL; } mb /= num_possible_cpus(); pages = mb << (20 - PAGE_SHIFT); down_write(&trace_sem); tcd_for_each(tcd, i, j) tcd->tcd_max_pages = (pages * tcd->tcd_pages_factor) / 100; up_write(&trace_sem); return 0; } EXPORT_SYMBOL(spl_debug_set_mb); int spl_debug_get_mb(void) { int i, j; struct trace_cpu_data *tcd; int total_pages = 0; down_read(&trace_sem); tcd_for_each(tcd, i, j) total_pages += tcd->tcd_max_pages; up_read(&trace_sem); return (total_pages >> (20 - PAGE_SHIFT)) + 1; } EXPORT_SYMBOL(spl_debug_get_mb); void spl_debug_dumpstack(struct task_struct *tsk) { extern void show_task(struct task_struct *); if (tsk == NULL) tsk = current; printk(KERN_ERR "SPL: Showing stack for process %d\n", tsk->pid); dump_stack(); } EXPORT_SYMBOL(spl_debug_dumpstack); void spl_debug_bug(char *file, const char *func, const int line, int flags) { spl_debug_catastrophe = 1; spl_debug_msg(NULL, 0, D_EMERG, file, func, line, "SBUG\n"); if (in_interrupt()) { panic("SBUG in interrupt.\n"); /* not reached */ } /* Ensure all debug pages and dumped by current cpu */ if (spl_debug_panic_on_bug) spl_panic_in_progress = 1; spl_debug_dumpstack(NULL); spl_debug_dumplog(flags); if (spl_debug_panic_on_bug) panic("SBUG"); set_task_state(current, TASK_UNINTERRUPTIBLE); while (1) schedule(); } EXPORT_SYMBOL(spl_debug_bug); int spl_debug_clear_buffer(void) { spl_debug_flush_pages(); return 0; } EXPORT_SYMBOL(spl_debug_clear_buffer); int spl_debug_mark_buffer(char *text) { CDEBUG(D_WARNING, "*************************************\n"); CDEBUG(D_WARNING, "DEBUG MARKER: %s\n", text); CDEBUG(D_WARNING, "*************************************\n"); return 0; } EXPORT_SYMBOL(spl_debug_mark_buffer); static int panic_notifier(struct notifier_block *self, unsigned long unused1, void *unused2) { if (spl_panic_in_progress) return 0; spl_panic_in_progress = 1; mb(); if (!in_interrupt()) { while (current->lock_depth >= 0) unlock_kernel(); spl_debug_dumplog(DL_NOTHREAD | DL_SINGLE_CPU); } return 0; } static int trace_init(int max_pages) { struct trace_cpu_data *tcd; int i, j; init_rwsem(&trace_sem); /* initialize trace_data */ memset(trace_data, 0, sizeof(trace_data)); for (i = 0; i < TCD_TYPE_MAX; i++) { trace_data[i] = kmalloc(sizeof(union trace_data_union) * NR_CPUS, GFP_KERNEL); if (trace_data[i] == NULL) goto out; } tcd_for_each(tcd, i, j) { tcd->tcd_pages_factor = pages_factor[i]; tcd->tcd_type = i; tcd->tcd_cpu = j; INIT_LIST_HEAD(&tcd->tcd_pages); INIT_LIST_HEAD(&tcd->tcd_stock_pages); tcd->tcd_cur_pages = 0; tcd->tcd_cur_stock_pages = 0; tcd->tcd_max_pages = (max_pages * pages_factor[i]) / 100; tcd->tcd_shutting_down = 0; } for (i = 0; i < num_possible_cpus(); i++) { for (j = 0; j < 3; j++) { trace_console_buffers[i][j] = kmalloc(TRACE_CONSOLE_BUFFER_SIZE, GFP_KERNEL); if (trace_console_buffers[i][j] == NULL) goto out; } } return 0; out: trace_fini(); printk(KERN_ERR "SPL: Insufficient memory for debug logs\n"); return -ENOMEM; } int debug_init(void) { int rc, max = spl_debug_mb; spl_console_max_delay = SPL_DEFAULT_MAX_DELAY; spl_console_min_delay = SPL_DEFAULT_MIN_DELAY; /* If spl_debug_mb is set to an invalid value or uninitialized * then just make the total buffers smp_num_cpus TCD_MAX_PAGES */ if (max > (num_physpages >> (20 - 2 - PAGE_SHIFT)) / 5 || max >= 512 || max < 0) { max = TCD_MAX_PAGES; } else { max = (max / num_online_cpus()) << (20 - PAGE_SHIFT); } rc = trace_init(max); if (rc) return rc; #ifdef HAVE_ATOMIC_PANIC_NOTIFIER atomic_notifier_chain_register(&panic_notifier_list, &spl_panic_notifier); #else notifier_chain_register(&panic_notifier_list, &spl_panic_notifier); #endif return rc; } static void trace_cleanup_on_cpu(void *info) { struct trace_cpu_data *tcd; struct trace_page *tage; struct trace_page *tmp; int i; tcd_for_each_type_lock(tcd, i) { tcd->tcd_shutting_down = 1; list_for_each_entry_safe(tage, tmp, &tcd->tcd_pages, linkage) { __ASSERT_TAGE_INVARIANT(tage); list_del(&tage->linkage); tage_free(tage); } tcd->tcd_cur_pages = 0; } } static void trace_fini(void) { int i, j; trace_call_on_all_cpus(trace_cleanup_on_cpu, NULL); for (i = 0; i < num_possible_cpus(); i++) { for (j = 0; j < 3; j++) { if (trace_console_buffers[i][j] != NULL) { kfree(trace_console_buffers[i][j]); trace_console_buffers[i][j] = NULL; } } } for (i = 0; trace_data[i] != NULL; i++) { kfree(trace_data[i]); trace_data[i] = NULL; } } void debug_fini(void) { #ifdef HAVE_ATOMIC_PANIC_NOTIFIER atomic_notifier_chain_unregister(&panic_notifier_list, &spl_panic_notifier); #else notifier_chain_unregister(&panic_notifier_list, &spl_panic_notifier); #endif trace_fini(); return; }