forked from Minki/linux
Merge branch 'for-5.2-pf-removal' into for-linus
This commit is contained in:
commit
0f46c78391
@ -237,7 +237,7 @@ static int pci_dac_dma_supported(struct pci_dev *dev, u64 mask)
|
||||
ok = 0;
|
||||
|
||||
/* If both conditions above are met, we are fine. */
|
||||
DBGA("pci_dac_dma_supported %s from %pf\n",
|
||||
DBGA("pci_dac_dma_supported %s from %ps\n",
|
||||
ok ? "yes" : "no", __builtin_return_address(0));
|
||||
|
||||
return ok;
|
||||
@ -269,7 +269,7 @@ pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
|
||||
&& paddr + size <= __direct_map_size) {
|
||||
ret = paddr + __direct_map_base;
|
||||
|
||||
DBGA2("pci_map_single: [%p,%zx] -> direct %llx from %pf\n",
|
||||
DBGA2("pci_map_single: [%p,%zx] -> direct %llx from %ps\n",
|
||||
cpu_addr, size, ret, __builtin_return_address(0));
|
||||
|
||||
return ret;
|
||||
@ -280,7 +280,7 @@ pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
|
||||
if (dac_allowed) {
|
||||
ret = paddr + alpha_mv.pci_dac_offset;
|
||||
|
||||
DBGA2("pci_map_single: [%p,%zx] -> DAC %llx from %pf\n",
|
||||
DBGA2("pci_map_single: [%p,%zx] -> DAC %llx from %ps\n",
|
||||
cpu_addr, size, ret, __builtin_return_address(0));
|
||||
|
||||
return ret;
|
||||
@ -317,7 +317,7 @@ pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
|
||||
ret = arena->dma_base + dma_ofs * PAGE_SIZE;
|
||||
ret += (unsigned long)cpu_addr & ~PAGE_MASK;
|
||||
|
||||
DBGA2("pci_map_single: [%p,%zx] np %ld -> sg %llx from %pf\n",
|
||||
DBGA2("pci_map_single: [%p,%zx] np %ld -> sg %llx from %ps\n",
|
||||
cpu_addr, size, npages, ret, __builtin_return_address(0));
|
||||
|
||||
return ret;
|
||||
@ -384,14 +384,14 @@ static void alpha_pci_unmap_page(struct device *dev, dma_addr_t dma_addr,
|
||||
&& dma_addr < __direct_map_base + __direct_map_size) {
|
||||
/* Nothing to do. */
|
||||
|
||||
DBGA2("pci_unmap_single: direct [%llx,%zx] from %pf\n",
|
||||
DBGA2("pci_unmap_single: direct [%llx,%zx] from %ps\n",
|
||||
dma_addr, size, __builtin_return_address(0));
|
||||
|
||||
return;
|
||||
}
|
||||
|
||||
if (dma_addr > 0xffffffff) {
|
||||
DBGA2("pci64_unmap_single: DAC [%llx,%zx] from %pf\n",
|
||||
DBGA2("pci64_unmap_single: DAC [%llx,%zx] from %ps\n",
|
||||
dma_addr, size, __builtin_return_address(0));
|
||||
return;
|
||||
}
|
||||
@ -423,7 +423,7 @@ static void alpha_pci_unmap_page(struct device *dev, dma_addr_t dma_addr,
|
||||
|
||||
spin_unlock_irqrestore(&arena->lock, flags);
|
||||
|
||||
DBGA2("pci_unmap_single: sg [%llx,%zx] np %ld from %pf\n",
|
||||
DBGA2("pci_unmap_single: sg [%llx,%zx] np %ld from %ps\n",
|
||||
dma_addr, size, npages, __builtin_return_address(0));
|
||||
}
|
||||
|
||||
@ -446,7 +446,7 @@ try_again:
|
||||
cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order);
|
||||
if (! cpu_addr) {
|
||||
printk(KERN_INFO "pci_alloc_consistent: "
|
||||
"get_free_pages failed from %pf\n",
|
||||
"get_free_pages failed from %ps\n",
|
||||
__builtin_return_address(0));
|
||||
/* ??? Really atomic allocation? Otherwise we could play
|
||||
with vmalloc and sg if we can't find contiguous memory. */
|
||||
@ -465,7 +465,7 @@ try_again:
|
||||
goto try_again;
|
||||
}
|
||||
|
||||
DBGA2("pci_alloc_consistent: %zx -> [%p,%llx] from %pf\n",
|
||||
DBGA2("pci_alloc_consistent: %zx -> [%p,%llx] from %ps\n",
|
||||
size, cpu_addr, *dma_addrp, __builtin_return_address(0));
|
||||
|
||||
return cpu_addr;
|
||||
@ -485,7 +485,7 @@ static void alpha_pci_free_coherent(struct device *dev, size_t size,
|
||||
pci_unmap_single(pdev, dma_addr, size, PCI_DMA_BIDIRECTIONAL);
|
||||
free_pages((unsigned long)cpu_addr, get_order(size));
|
||||
|
||||
DBGA2("pci_free_consistent: [%llx,%zx] from %pf\n",
|
||||
DBGA2("pci_free_consistent: [%llx,%zx] from %ps\n",
|
||||
dma_addr, size, __builtin_return_address(0));
|
||||
}
|
||||
|
||||
|
@ -631,7 +631,7 @@ static void imx6_pm_stby_poweroff(void)
|
||||
static int imx6_pm_stby_poweroff_probe(void)
|
||||
{
|
||||
if (pm_power_off) {
|
||||
pr_warn("%s: pm_power_off already claimed %p %pf!\n",
|
||||
pr_warn("%s: pm_power_off already claimed %p %ps!\n",
|
||||
__func__, pm_power_off, pm_power_off);
|
||||
return -EBUSY;
|
||||
}
|
||||
|
@ -133,7 +133,7 @@ static const char *usermode_action[] = {
|
||||
static int alignment_proc_show(struct seq_file *m, void *v)
|
||||
{
|
||||
seq_printf(m, "User:\t\t%lu\n", ai_user);
|
||||
seq_printf(m, "System:\t\t%lu (%pF)\n", ai_sys, ai_sys_last_pc);
|
||||
seq_printf(m, "System:\t\t%lu (%pS)\n", ai_sys, ai_sys_last_pc);
|
||||
seq_printf(m, "Skipped:\t%lu\n", ai_skipped);
|
||||
seq_printf(m, "Half:\t\t%lu\n", ai_half);
|
||||
seq_printf(m, "Word:\t\t%lu\n", ai_word);
|
||||
|
@ -147,7 +147,7 @@ void float_raise(signed char flags)
|
||||
#ifdef CONFIG_DEBUG_USER
|
||||
if (flags & debug)
|
||||
printk(KERN_DEBUG
|
||||
"NWFPE: %s[%d] takes exception %08x at %pf from %08lx\n",
|
||||
"NWFPE: %s[%d] takes exception %08x at %ps from %08lx\n",
|
||||
current->comm, current->pid, flags,
|
||||
__builtin_return_address(0), GET_USERREG()->ARM_pc);
|
||||
#endif
|
||||
|
@ -75,7 +75,7 @@ static void __iomem *__ioremap(phys_addr_t addr, unsigned long size,
|
||||
p >= memory_start && p < virt_to_phys(high_memory) &&
|
||||
!(p >= __virt_to_phys((phys_addr_t)__bss_stop) &&
|
||||
p < __virt_to_phys((phys_addr_t)__bss_stop))) {
|
||||
pr_warn("__ioremap(): phys addr "PTE_FMT" is RAM lr %pf\n",
|
||||
pr_warn("__ioremap(): phys addr "PTE_FMT" is RAM lr %ps\n",
|
||||
(unsigned long)p, __builtin_return_address(0));
|
||||
return NULL;
|
||||
}
|
||||
|
@ -876,7 +876,7 @@ void ldom_power_off(void)
|
||||
|
||||
static void ds_conn_reset(struct ds_info *dp)
|
||||
{
|
||||
printk(KERN_ERR "ds-%llu: ds_conn_reset() from %pf\n",
|
||||
printk(KERN_ERR "ds-%llu: ds_conn_reset() from %ps\n",
|
||||
dp->id, __builtin_return_address(0));
|
||||
}
|
||||
|
||||
|
@ -20,7 +20,7 @@
|
||||
|
||||
static void _print_addr(void *data, unsigned long address, int reliable)
|
||||
{
|
||||
pr_info(" [<%08lx>] %s%pF\n", address, reliable ? "" : "? ",
|
||||
pr_info(" [<%08lx>] %s%pS\n", address, reliable ? "" : "? ",
|
||||
(void *)address);
|
||||
}
|
||||
|
||||
|
@ -30,7 +30,7 @@ DECLARE_EVENT_CLASS(x86_exceptions,
|
||||
__entry->error_code = error_code;
|
||||
),
|
||||
|
||||
TP_printk("address=%pf ip=%pf error_code=0x%lx",
|
||||
TP_printk("address=%ps ip=%ps error_code=0x%lx",
|
||||
(void *)__entry->address, (void *)__entry->ip,
|
||||
__entry->error_code) );
|
||||
|
||||
|
@ -58,7 +58,7 @@ static inline void stack_overflow_check(struct pt_regs *regs)
|
||||
if (regs->sp >= estack_top && regs->sp <= estack_bottom)
|
||||
return;
|
||||
|
||||
WARN_ONCE(1, "do_IRQ(): %s has overflown the kernel stack (cur:%Lx,sp:%lx,irq stk top-bottom:%Lx-%Lx,exception stk top-bottom:%Lx-%Lx,ip:%pF)\n",
|
||||
WARN_ONCE(1, "do_IRQ(): %s has overflown the kernel stack (cur:%Lx,sp:%lx,irq stk top-bottom:%Lx-%Lx,exception stk top-bottom:%Lx-%Lx,ip:%pS)\n",
|
||||
current->comm, curbase, regs->sp,
|
||||
irq_stack_top, irq_stack_bottom,
|
||||
estack_top, estack_bottom, (void *)regs->ip);
|
||||
|
@ -145,7 +145,7 @@ __visible bool ex_handler_rdmsr_unsafe(const struct exception_table_entry *fixup
|
||||
unsigned long error_code,
|
||||
unsigned long fault_addr)
|
||||
{
|
||||
if (pr_warn_once("unchecked MSR access error: RDMSR from 0x%x at rIP: 0x%lx (%pF)\n",
|
||||
if (pr_warn_once("unchecked MSR access error: RDMSR from 0x%x at rIP: 0x%lx (%pS)\n",
|
||||
(unsigned int)regs->cx, regs->ip, (void *)regs->ip))
|
||||
show_stack_regs(regs);
|
||||
|
||||
@ -162,7 +162,7 @@ __visible bool ex_handler_wrmsr_unsafe(const struct exception_table_entry *fixup
|
||||
unsigned long error_code,
|
||||
unsigned long fault_addr)
|
||||
{
|
||||
if (pr_warn_once("unchecked MSR access error: WRMSR to 0x%x (tried to write 0x%08x%08x) at rIP: 0x%lx (%pF)\n",
|
||||
if (pr_warn_once("unchecked MSR access error: WRMSR to 0x%x (tried to write 0x%08x%08x) at rIP: 0x%lx (%pS)\n",
|
||||
(unsigned int)regs->cx, (unsigned int)regs->dx,
|
||||
(unsigned int)regs->ax, regs->ip, (void *)regs->ip))
|
||||
show_stack_regs(regs);
|
||||
|
@ -105,7 +105,7 @@ void xen_mc_flush(void)
|
||||
for (i = 0; i < b->mcidx; i++) {
|
||||
if (b->entries[i].result < 0) {
|
||||
#if MC_DEBUG
|
||||
pr_err(" call %2d: op=%lu arg=[%lx] result=%ld\t%pF\n",
|
||||
pr_err(" call %2d: op=%lu arg=[%lx] result=%ld\t%pS\n",
|
||||
i + 1,
|
||||
b->debug[i].op,
|
||||
b->debug[i].args[0],
|
||||
|
@ -414,7 +414,7 @@ static void acpi_pm_notify_handler(acpi_handle handle, u32 val, void *not_used)
|
||||
if (adev->wakeup.flags.notifier_present) {
|
||||
pm_wakeup_ws_event(adev->wakeup.ws, 0, acpi_s2idle_wakeup());
|
||||
if (adev->wakeup.context.func) {
|
||||
acpi_handle_debug(handle, "Running %pF for %s\n",
|
||||
acpi_handle_debug(handle, "Running %pS for %s\n",
|
||||
adev->wakeup.context.func,
|
||||
dev_name(adev->wakeup.context.dev));
|
||||
adev->wakeup.context.func(&adev->wakeup.context);
|
||||
|
@ -205,7 +205,7 @@ static ktime_t initcall_debug_start(struct device *dev, void *cb)
|
||||
if (!pm_print_times_enabled)
|
||||
return 0;
|
||||
|
||||
dev_info(dev, "calling %pF @ %i, parent: %s\n", cb,
|
||||
dev_info(dev, "calling %pS @ %i, parent: %s\n", cb,
|
||||
task_pid_nr(current),
|
||||
dev->parent ? dev_name(dev->parent) : "none");
|
||||
return ktime_get();
|
||||
@ -223,7 +223,7 @@ static void initcall_debug_report(struct device *dev, ktime_t calltime,
|
||||
rettime = ktime_get();
|
||||
nsecs = (s64) ktime_to_ns(ktime_sub(rettime, calltime));
|
||||
|
||||
dev_info(dev, "%pF returned %d after %Ld usecs\n", cb, error,
|
||||
dev_info(dev, "%pS returned %d after %Ld usecs\n", cb, error,
|
||||
(unsigned long long)nsecs >> 10);
|
||||
}
|
||||
|
||||
@ -2062,7 +2062,7 @@ EXPORT_SYMBOL_GPL(dpm_suspend_start);
|
||||
void __suspend_report_result(const char *function, void *fn, int ret)
|
||||
{
|
||||
if (ret)
|
||||
printk(KERN_ERR "%s(): %pF returns %d\n", function, fn, ret);
|
||||
printk(KERN_ERR "%s(): %pS returns %d\n", function, fn, ret);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(__suspend_report_result);
|
||||
|
||||
|
@ -62,19 +62,19 @@ int syscore_suspend(void)
|
||||
list_for_each_entry_reverse(ops, &syscore_ops_list, node)
|
||||
if (ops->suspend) {
|
||||
if (initcall_debug)
|
||||
pr_info("PM: Calling %pF\n", ops->suspend);
|
||||
pr_info("PM: Calling %pS\n", ops->suspend);
|
||||
ret = ops->suspend();
|
||||
if (ret)
|
||||
goto err_out;
|
||||
WARN_ONCE(!irqs_disabled(),
|
||||
"Interrupts enabled after %pF\n", ops->suspend);
|
||||
"Interrupts enabled after %pS\n", ops->suspend);
|
||||
}
|
||||
|
||||
trace_suspend_resume(TPS("syscore_suspend"), 0, false);
|
||||
return 0;
|
||||
|
||||
err_out:
|
||||
pr_err("PM: System core suspend callback %pF failed.\n", ops->suspend);
|
||||
pr_err("PM: System core suspend callback %pS failed.\n", ops->suspend);
|
||||
|
||||
list_for_each_entry_continue(ops, &syscore_ops_list, node)
|
||||
if (ops->resume)
|
||||
@ -100,10 +100,10 @@ void syscore_resume(void)
|
||||
list_for_each_entry(ops, &syscore_ops_list, node)
|
||||
if (ops->resume) {
|
||||
if (initcall_debug)
|
||||
pr_info("PM: Calling %pF\n", ops->resume);
|
||||
pr_info("PM: Calling %pS\n", ops->resume);
|
||||
ops->resume();
|
||||
WARN_ONCE(!irqs_disabled(),
|
||||
"Interrupts enabled after %pF\n", ops->resume);
|
||||
"Interrupts enabled after %pS\n", ops->resume);
|
||||
}
|
||||
trace_suspend_resume(TPS("syscore_resume"), 0, false);
|
||||
}
|
||||
@ -122,7 +122,7 @@ void syscore_shutdown(void)
|
||||
list_for_each_entry_reverse(ops, &syscore_ops_list, node)
|
||||
if (ops->shutdown) {
|
||||
if (initcall_debug)
|
||||
pr_info("PM: Calling %pF\n", ops->shutdown);
|
||||
pr_info("PM: Calling %pS\n", ops->shutdown);
|
||||
ops->shutdown();
|
||||
}
|
||||
|
||||
|
@ -6116,7 +6116,7 @@ int drbd_ack_receiver(struct drbd_thread *thi)
|
||||
|
||||
err = cmd->fn(connection, &pi);
|
||||
if (err) {
|
||||
drbd_err(connection, "%pf failed\n", cmd->fn);
|
||||
drbd_err(connection, "%ps failed\n", cmd->fn);
|
||||
goto reconnect;
|
||||
}
|
||||
|
||||
|
@ -1693,7 +1693,7 @@ irqreturn_t floppy_interrupt(int irq, void *dev_id)
|
||||
/* we don't even know which FDC is the culprit */
|
||||
pr_info("DOR0=%x\n", fdc_state[0].dor);
|
||||
pr_info("floppy interrupt on bizarre fdc %d\n", fdc);
|
||||
pr_info("handler=%pf\n", handler);
|
||||
pr_info("handler=%ps\n", handler);
|
||||
is_alive(__func__, "bizarre fdc");
|
||||
return IRQ_NONE;
|
||||
}
|
||||
@ -1752,7 +1752,7 @@ static void reset_interrupt(void)
|
||||
debugt(__func__, "");
|
||||
result(); /* get the status ready for set_fdc */
|
||||
if (FDCS->reset) {
|
||||
pr_info("reset set in interrupt, calling %pf\n", cont->error);
|
||||
pr_info("reset set in interrupt, calling %ps\n", cont->error);
|
||||
cont->error(); /* a reset just after a reset. BAD! */
|
||||
}
|
||||
cont->redo();
|
||||
@ -1793,7 +1793,7 @@ static void show_floppy(void)
|
||||
pr_info("\n");
|
||||
pr_info("floppy driver state\n");
|
||||
pr_info("-------------------\n");
|
||||
pr_info("now=%lu last interrupt=%lu diff=%lu last called handler=%pf\n",
|
||||
pr_info("now=%lu last interrupt=%lu diff=%lu last called handler=%ps\n",
|
||||
jiffies, interruptjiffies, jiffies - interruptjiffies,
|
||||
lasthandler);
|
||||
|
||||
@ -1812,9 +1812,9 @@ static void show_floppy(void)
|
||||
pr_info("status=%x\n", fd_inb(FD_STATUS));
|
||||
pr_info("fdc_busy=%lu\n", fdc_busy);
|
||||
if (do_floppy)
|
||||
pr_info("do_floppy=%pf\n", do_floppy);
|
||||
pr_info("do_floppy=%ps\n", do_floppy);
|
||||
if (work_pending(&floppy_work))
|
||||
pr_info("floppy_work.func=%pf\n", floppy_work.func);
|
||||
pr_info("floppy_work.func=%ps\n", floppy_work.func);
|
||||
if (delayed_work_pending(&fd_timer))
|
||||
pr_info("delayed work.function=%p expires=%ld\n",
|
||||
fd_timer.work.func,
|
||||
|
@ -432,7 +432,7 @@ static void cpufreq_list_transition_notifiers(void)
|
||||
mutex_lock(&cpufreq_transition_notifier_list.mutex);
|
||||
|
||||
for (nb = cpufreq_transition_notifier_list.head; nb; nb = nb->next)
|
||||
pr_info("%pF\n", nb->notifier_call);
|
||||
pr_info("%pS\n", nb->notifier_call);
|
||||
|
||||
mutex_unlock(&cpufreq_transition_notifier_list.mutex);
|
||||
}
|
||||
|
@ -159,7 +159,7 @@ static inline void mmc_fixup_device(struct mmc_card *card,
|
||||
(f->ext_csd_rev == EXT_CSD_REV_ANY ||
|
||||
f->ext_csd_rev == card->ext_csd.rev) &&
|
||||
rev >= f->rev_start && rev <= f->rev_end) {
|
||||
dev_dbg(&card->dev, "calling %pf\n", f->vendor_fixup);
|
||||
dev_dbg(&card->dev, "calling %ps\n", f->vendor_fixup);
|
||||
f->vendor_fixup(card, f->data);
|
||||
}
|
||||
}
|
||||
|
@ -581,7 +581,7 @@ int __nd_driver_register(struct nd_device_driver *nd_drv, struct module *owner,
|
||||
struct device_driver *drv = &nd_drv->drv;
|
||||
|
||||
if (!nd_drv->type) {
|
||||
pr_debug("driver type bitmask not set (%pf)\n",
|
||||
pr_debug("driver type bitmask not set (%ps)\n",
|
||||
__builtin_return_address(0));
|
||||
return -EINVAL;
|
||||
}
|
||||
|
@ -53,7 +53,7 @@ static int validate_dimm(struct nvdimm_drvdata *ndd)
|
||||
|
||||
rc = nvdimm_check_config_data(ndd->dev);
|
||||
if (rc)
|
||||
dev_dbg(ndd->dev, "%pf: %s error: %d\n",
|
||||
dev_dbg(ndd->dev, "%ps: %s error: %d\n",
|
||||
__builtin_return_address(0), __func__, rc);
|
||||
return rc;
|
||||
}
|
||||
|
@ -578,7 +578,7 @@ static int pci_legacy_suspend(struct device *dev, pm_message_t state)
|
||||
if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0
|
||||
&& pci_dev->current_state != PCI_UNKNOWN) {
|
||||
WARN_ONCE(pci_dev->current_state != prev,
|
||||
"PCI PM: Device state not saved by %pF\n",
|
||||
"PCI PM: Device state not saved by %pS\n",
|
||||
drv->suspend);
|
||||
}
|
||||
}
|
||||
@ -605,7 +605,7 @@ static int pci_legacy_suspend_late(struct device *dev, pm_message_t state)
|
||||
if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0
|
||||
&& pci_dev->current_state != PCI_UNKNOWN) {
|
||||
WARN_ONCE(pci_dev->current_state != prev,
|
||||
"PCI PM: Device state not saved by %pF\n",
|
||||
"PCI PM: Device state not saved by %pS\n",
|
||||
drv->suspend_late);
|
||||
goto Fixup;
|
||||
}
|
||||
@ -773,7 +773,7 @@ static int pci_pm_suspend(struct device *dev)
|
||||
if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0
|
||||
&& pci_dev->current_state != PCI_UNKNOWN) {
|
||||
WARN_ONCE(pci_dev->current_state != prev,
|
||||
"PCI PM: State of device not saved by %pF\n",
|
||||
"PCI PM: State of device not saved by %pS\n",
|
||||
pm->suspend);
|
||||
}
|
||||
}
|
||||
@ -821,7 +821,7 @@ static int pci_pm_suspend_noirq(struct device *dev)
|
||||
if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0
|
||||
&& pci_dev->current_state != PCI_UNKNOWN) {
|
||||
WARN_ONCE(pci_dev->current_state != prev,
|
||||
"PCI PM: State of device not saved by %pF\n",
|
||||
"PCI PM: State of device not saved by %pS\n",
|
||||
pm->suspend_noirq);
|
||||
goto Fixup;
|
||||
}
|
||||
@ -1260,11 +1260,11 @@ static int pci_pm_runtime_suspend(struct device *dev)
|
||||
* log level.
|
||||
*/
|
||||
if (error == -EBUSY || error == -EAGAIN) {
|
||||
dev_dbg(dev, "can't suspend now (%pf returned %d)\n",
|
||||
dev_dbg(dev, "can't suspend now (%ps returned %d)\n",
|
||||
pm->runtime_suspend, error);
|
||||
return error;
|
||||
} else if (error) {
|
||||
dev_err(dev, "can't suspend (%pf returned %d)\n",
|
||||
dev_err(dev, "can't suspend (%ps returned %d)\n",
|
||||
pm->runtime_suspend, error);
|
||||
return error;
|
||||
}
|
||||
@ -1276,7 +1276,7 @@ static int pci_pm_runtime_suspend(struct device *dev)
|
||||
&& !pci_dev->state_saved && pci_dev->current_state != PCI_D0
|
||||
&& pci_dev->current_state != PCI_UNKNOWN) {
|
||||
WARN_ONCE(pci_dev->current_state != prev,
|
||||
"PCI PM: State of device not saved by %pF\n",
|
||||
"PCI PM: State of device not saved by %pS\n",
|
||||
pm->runtime_suspend);
|
||||
return 0;
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ static ktime_t fixup_debug_start(struct pci_dev *dev,
|
||||
void (*fn)(struct pci_dev *dev))
|
||||
{
|
||||
if (initcall_debug)
|
||||
pci_info(dev, "calling %pF @ %i\n", fn, task_pid_nr(current));
|
||||
pci_info(dev, "calling %pS @ %i\n", fn, task_pid_nr(current));
|
||||
|
||||
return ktime_get();
|
||||
}
|
||||
@ -51,7 +51,7 @@ static void fixup_debug_report(struct pci_dev *dev, ktime_t calltime,
|
||||
delta = ktime_sub(rettime, calltime);
|
||||
duration = (unsigned long long) ktime_to_ns(delta) >> 10;
|
||||
if (initcall_debug || duration > 10000)
|
||||
pci_info(dev, "%pF took %lld usecs\n", fn, duration);
|
||||
pci_info(dev, "%pS took %lld usecs\n", fn, duration);
|
||||
}
|
||||
|
||||
static void pci_do_fixups(struct pci_dev *dev, struct pci_fixup *f,
|
||||
|
@ -458,7 +458,7 @@ void pnp_fixup_device(struct pnp_dev *dev)
|
||||
for (f = pnp_fixups; *f->id; f++) {
|
||||
if (!compare_pnp_id(dev->id, f->id))
|
||||
continue;
|
||||
pnp_dbg(&dev->dev, "%s: calling %pF\n", f->id,
|
||||
pnp_dbg(&dev->dev, "%s: calling %pS\n", f->id,
|
||||
f->quirk_function);
|
||||
f->quirk_function(dev);
|
||||
}
|
||||
|
@ -1031,7 +1031,7 @@ static int esp_check_spur_intr(struct esp *esp)
|
||||
|
||||
static void esp_schedule_reset(struct esp *esp)
|
||||
{
|
||||
esp_log_reset("esp_schedule_reset() from %pf\n",
|
||||
esp_log_reset("esp_schedule_reset() from %ps\n",
|
||||
__builtin_return_address(0));
|
||||
esp->flags |= ESP_FLAG_RESETTING;
|
||||
esp_event(esp, ESP_EVENT_RESET);
|
||||
|
@ -539,7 +539,7 @@ static int run_test_both_formats(test_func_t test_func, u32 sectorsize,
|
||||
ret = run_test(test_func, 0, sectorsize, nodesize, alignment);
|
||||
if (ret) {
|
||||
test_err(
|
||||
"%pf failed with extents, sectorsize=%u, nodesize=%u, alignment=%u",
|
||||
"%ps failed with extents, sectorsize=%u, nodesize=%u, alignment=%u",
|
||||
test_func, sectorsize, nodesize, alignment);
|
||||
test_ret = ret;
|
||||
}
|
||||
@ -547,7 +547,7 @@ static int run_test_both_formats(test_func_t test_func, u32 sectorsize,
|
||||
ret = run_test(test_func, 1, sectorsize, nodesize, alignment);
|
||||
if (ret) {
|
||||
test_err(
|
||||
"%pf failed with bitmaps, sectorsize=%u, nodesize=%u, alignment=%u",
|
||||
"%ps failed with bitmaps, sectorsize=%u, nodesize=%u, alignment=%u",
|
||||
test_func, sectorsize, nodesize, alignment);
|
||||
test_ret = ret;
|
||||
}
|
||||
|
@ -1337,7 +1337,7 @@ struct f2fs_private_dio {
|
||||
|
||||
#ifdef CONFIG_F2FS_FAULT_INJECTION
|
||||
#define f2fs_show_injection_info(type) \
|
||||
printk_ratelimited("%sF2FS-fs : inject %s in %s of %pF\n", \
|
||||
printk_ratelimited("%sF2FS-fs : inject %s in %s of %pS\n", \
|
||||
KERN_INFO, f2fs_fault_name[type], \
|
||||
__func__, __builtin_return_address(0))
|
||||
static inline bool time_to_inject(struct f2fs_sb_info *sbi, int type)
|
||||
|
@ -115,7 +115,7 @@ static int pstore_ftrace_seq_show(struct seq_file *s, void *v)
|
||||
|
||||
rec = (struct pstore_ftrace_record *)(ps->record->buf + data->off);
|
||||
|
||||
seq_printf(s, "CPU:%d ts:%llu %08lx %08lx %pf <- %pF\n",
|
||||
seq_printf(s, "CPU:%d ts:%llu %08lx %08lx %ps <- %pS\n",
|
||||
pstore_ftrace_decode_cpu(rec),
|
||||
pstore_ftrace_read_timestamp(rec),
|
||||
rec->ip, rec->parent_ip, (void *)rec->ip,
|
||||
|
@ -1345,7 +1345,7 @@ DECLARE_EVENT_CLASS(btrfs__work,
|
||||
__entry->normal_work = &work->normal_work;
|
||||
),
|
||||
|
||||
TP_printk_btrfs("work=%p (normal_work=%p) wq=%p func=%pf ordered_func=%p "
|
||||
TP_printk_btrfs("work=%p (normal_work=%p) wq=%p func=%ps ordered_func=%p "
|
||||
"ordered_free=%p",
|
||||
__entry->work, __entry->normal_work, __entry->wq,
|
||||
__entry->func, __entry->ordered_func, __entry->ordered_free)
|
||||
|
@ -30,7 +30,7 @@ TRACE_EVENT(cpuhp_enter,
|
||||
__entry->fun = fun;
|
||||
),
|
||||
|
||||
TP_printk("cpu: %04u target: %3d step: %3d (%pf)",
|
||||
TP_printk("cpu: %04u target: %3d step: %3d (%ps)",
|
||||
__entry->cpu, __entry->target, __entry->idx, __entry->fun)
|
||||
);
|
||||
|
||||
@ -58,7 +58,7 @@ TRACE_EVENT(cpuhp_multi_enter,
|
||||
__entry->fun = fun;
|
||||
),
|
||||
|
||||
TP_printk("cpu: %04u target: %3d step: %3d (%pf)",
|
||||
TP_printk("cpu: %04u target: %3d step: %3d (%ps)",
|
||||
__entry->cpu, __entry->target, __entry->idx, __entry->fun)
|
||||
);
|
||||
|
||||
|
@ -27,7 +27,7 @@ DECLARE_EVENT_CLASS(preemptirq_template,
|
||||
__entry->parent_offs = (u32)(parent_ip - (unsigned long)_stext);
|
||||
),
|
||||
|
||||
TP_printk("caller=%pF parent=%pF",
|
||||
TP_printk("caller=%pS parent=%pS",
|
||||
(void *)((unsigned long)(_stext) + __entry->caller_offs),
|
||||
(void *)((unsigned long)(_stext) + __entry->parent_offs))
|
||||
);
|
||||
|
@ -491,7 +491,7 @@ TRACE_EVENT(rcu_callback,
|
||||
__entry->qlen = qlen;
|
||||
),
|
||||
|
||||
TP_printk("%s rhp=%p func=%pf %ld/%ld",
|
||||
TP_printk("%s rhp=%p func=%ps %ld/%ld",
|
||||
__entry->rcuname, __entry->rhp, __entry->func,
|
||||
__entry->qlen_lazy, __entry->qlen)
|
||||
);
|
||||
@ -587,7 +587,7 @@ TRACE_EVENT(rcu_invoke_callback,
|
||||
__entry->func = rhp->func;
|
||||
),
|
||||
|
||||
TP_printk("%s rhp=%p func=%pf",
|
||||
TP_printk("%s rhp=%p func=%ps",
|
||||
__entry->rcuname, __entry->rhp, __entry->func)
|
||||
);
|
||||
|
||||
|
@ -102,7 +102,7 @@ DECLARE_EVENT_CLASS(rpc_task_running,
|
||||
__entry->flags = task->tk_flags;
|
||||
),
|
||||
|
||||
TP_printk("task:%u@%d flags=%4.4x state=%4.4lx status=%d action=%pf",
|
||||
TP_printk("task:%u@%d flags=%4.4x state=%4.4lx status=%d action=%ps",
|
||||
__entry->task_id, __entry->client_id,
|
||||
__entry->flags,
|
||||
__entry->runstate,
|
||||
|
@ -226,7 +226,7 @@ TRACE_EVENT(mm_shrink_slab_start,
|
||||
__entry->priority = priority;
|
||||
),
|
||||
|
||||
TP_printk("%pF %p: nid: %d objects to shrink %ld gfp_flags %s cache items %ld delta %lld total_scan %ld priority %d",
|
||||
TP_printk("%pS %p: nid: %d objects to shrink %ld gfp_flags %s cache items %ld delta %lld total_scan %ld priority %d",
|
||||
__entry->shrink,
|
||||
__entry->shr,
|
||||
__entry->nid,
|
||||
@ -265,7 +265,7 @@ TRACE_EVENT(mm_shrink_slab_end,
|
||||
__entry->total_scan = total_scan;
|
||||
),
|
||||
|
||||
TP_printk("%pF %p: nid: %d unused scan count %ld new scan count %ld total_scan %ld last shrinker return val %d",
|
||||
TP_printk("%pS %p: nid: %d unused scan count %ld new scan count %ld total_scan %ld last shrinker return val %d",
|
||||
__entry->shrink,
|
||||
__entry->shr,
|
||||
__entry->nid,
|
||||
|
@ -60,7 +60,7 @@ TRACE_EVENT(workqueue_queue_work,
|
||||
__entry->cpu = pwq->pool->cpu;
|
||||
),
|
||||
|
||||
TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u",
|
||||
TP_printk("work struct=%p function=%ps workqueue=%p req_cpu=%u cpu=%u",
|
||||
__entry->work, __entry->function, __entry->workqueue,
|
||||
__entry->req_cpu, __entry->cpu)
|
||||
);
|
||||
@ -102,7 +102,7 @@ TRACE_EVENT(workqueue_execute_start,
|
||||
__entry->function = work->func;
|
||||
),
|
||||
|
||||
TP_printk("work struct %p: function %pf", __entry->work, __entry->function)
|
||||
TP_printk("work struct %p: function %ps", __entry->work, __entry->function)
|
||||
);
|
||||
|
||||
/**
|
||||
|
@ -73,7 +73,7 @@ TRACE_EVENT(xen_mc_callback,
|
||||
__entry->fn = fn;
|
||||
__entry->data = data;
|
||||
),
|
||||
TP_printk("callback %pf, data %p",
|
||||
TP_printk("callback %ps, data %p",
|
||||
__entry->fn, __entry->data)
|
||||
);
|
||||
|
||||
|
@ -826,7 +826,7 @@ trace_initcall_start_cb(void *data, initcall_t fn)
|
||||
{
|
||||
ktime_t *calltime = (ktime_t *)data;
|
||||
|
||||
printk(KERN_DEBUG "calling %pF @ %i\n", fn, task_pid_nr(current));
|
||||
printk(KERN_DEBUG "calling %pS @ %i\n", fn, task_pid_nr(current));
|
||||
*calltime = ktime_get();
|
||||
}
|
||||
|
||||
@ -840,7 +840,7 @@ trace_initcall_finish_cb(void *data, initcall_t fn, int ret)
|
||||
rettime = ktime_get();
|
||||
delta = ktime_sub(rettime, *calltime);
|
||||
duration = (unsigned long long) ktime_to_ns(delta) >> 10;
|
||||
printk(KERN_DEBUG "initcall %pF returned %d after %lld usecs\n",
|
||||
printk(KERN_DEBUG "initcall %pS returned %d after %lld usecs\n",
|
||||
fn, ret, duration);
|
||||
}
|
||||
|
||||
@ -897,7 +897,7 @@ int __init_or_module do_one_initcall(initcall_t fn)
|
||||
strlcat(msgbuf, "disabled interrupts ", sizeof(msgbuf));
|
||||
local_irq_enable();
|
||||
}
|
||||
WARN(msgbuf[0], "initcall %pF returned with %s\n", fn, msgbuf);
|
||||
WARN(msgbuf[0], "initcall %pS returned with %s\n", fn, msgbuf);
|
||||
|
||||
add_latent_entropy();
|
||||
return ret;
|
||||
|
@ -119,7 +119,7 @@ static void async_run_entry_fn(struct work_struct *work)
|
||||
|
||||
/* 1) run (and print duration) */
|
||||
if (initcall_debug && system_state < SYSTEM_RUNNING) {
|
||||
pr_debug("calling %lli_%pF @ %i\n",
|
||||
pr_debug("calling %lli_%pS @ %i\n",
|
||||
(long long)entry->cookie,
|
||||
entry->func, task_pid_nr(current));
|
||||
calltime = ktime_get();
|
||||
@ -128,7 +128,7 @@ static void async_run_entry_fn(struct work_struct *work)
|
||||
if (initcall_debug && system_state < SYSTEM_RUNNING) {
|
||||
rettime = ktime_get();
|
||||
delta = ktime_sub(rettime, calltime);
|
||||
pr_debug("initcall %lli_%pF returned 0 after %lld usecs\n",
|
||||
pr_debug("initcall %lli_%pS returned 0 after %lld usecs\n",
|
||||
(long long)entry->cookie,
|
||||
entry->func,
|
||||
(long long)ktime_to_ns(delta) >> 10);
|
||||
|
@ -2028,7 +2028,7 @@ static void handler_chain(struct uprobe *uprobe, struct pt_regs *regs)
|
||||
if (uc->handler) {
|
||||
rc = uc->handler(uc, regs);
|
||||
WARN(rc & ~UPROBE_HANDLER_MASK,
|
||||
"bad rc=0x%x from %pf()\n", rc, uc->handler);
|
||||
"bad rc=0x%x from %ps()\n", rc, uc->handler);
|
||||
}
|
||||
|
||||
if (uc->ret_handler)
|
||||
|
@ -210,7 +210,7 @@ static int fei_seq_show(struct seq_file *m, void *v)
|
||||
{
|
||||
struct fei_attr *attr = list_entry(v, struct fei_attr, list);
|
||||
|
||||
seq_printf(m, "%pf\n", attr->kp.addr);
|
||||
seq_printf(m, "%ps\n", attr->kp.addr);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -152,7 +152,7 @@ static int irq_debug_show(struct seq_file *m, void *p)
|
||||
|
||||
raw_spin_lock_irq(&desc->lock);
|
||||
data = irq_desc_get_irq_data(desc);
|
||||
seq_printf(m, "handler: %pf\n", desc->handle_irq);
|
||||
seq_printf(m, "handler: %ps\n", desc->handle_irq);
|
||||
seq_printf(m, "device: %s\n", desc->dev_name);
|
||||
seq_printf(m, "status: 0x%08x\n", desc->status_use_accessors);
|
||||
irq_debug_show_bits(m, 0, desc->status_use_accessors, irqdesc_states,
|
||||
|
@ -149,7 +149,7 @@ irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc, unsigned int *flags
|
||||
res = action->handler(irq, action->dev_id);
|
||||
trace_irq_handler_exit(irq, action, res);
|
||||
|
||||
if (WARN_ONCE(!irqs_disabled(),"irq %u handler %pF enabled interrupts\n",
|
||||
if (WARN_ONCE(!irqs_disabled(),"irq %u handler %pS enabled interrupts\n",
|
||||
irq, action->handler))
|
||||
local_irq_disable();
|
||||
|
||||
|
@ -778,7 +778,7 @@ int __irq_set_trigger(struct irq_desc *desc, unsigned long flags)
|
||||
ret = 0;
|
||||
break;
|
||||
default:
|
||||
pr_err("Setting trigger mode %lu for irq %u failed (%pF)\n",
|
||||
pr_err("Setting trigger mode %lu for irq %u failed (%pS)\n",
|
||||
flags, irq_desc_get_irq(desc), chip->irq_set_type);
|
||||
}
|
||||
if (unmask)
|
||||
|
@ -212,9 +212,9 @@ static void __report_bad_irq(struct irq_desc *desc, irqreturn_t action_ret)
|
||||
*/
|
||||
raw_spin_lock_irqsave(&desc->lock, flags);
|
||||
for_each_action_of_desc(desc, action) {
|
||||
printk(KERN_ERR "[<%p>] %pf", action->handler, action->handler);
|
||||
printk(KERN_ERR "[<%p>] %ps", action->handler, action->handler);
|
||||
if (action->thread_fn)
|
||||
printk(KERN_CONT " threaded [<%p>] %pf",
|
||||
printk(KERN_CONT " threaded [<%p>] %ps",
|
||||
action->thread_fn, action->thread_fn);
|
||||
printk(KERN_CONT "\n");
|
||||
}
|
||||
|
@ -2870,7 +2870,7 @@ __call_rcu(struct rcu_head *head, rcu_callback_t func, int cpu, bool lazy)
|
||||
* Use rcu:rcu_callback trace event to find the previous
|
||||
* time callback was passed to __call_rcu().
|
||||
*/
|
||||
WARN_ONCE(1, "__call_rcu(): Double-freed CB %p->%pF()!!!\n",
|
||||
WARN_ONCE(1, "__call_rcu(): Double-freed CB %p->%pS()!!!\n",
|
||||
head, head->func);
|
||||
WRITE_ONCE(head->func, rcu_leak_callback);
|
||||
return;
|
||||
|
@ -513,7 +513,7 @@ repeat:
|
||||
}
|
||||
preempt_count_dec();
|
||||
WARN_ONCE(preempt_count(),
|
||||
"cpu_stop: %pf(%p) leaked preempt count\n", fn, arg);
|
||||
"cpu_stop: %ps(%p) leaked preempt count\n", fn, arg);
|
||||
goto repeat;
|
||||
}
|
||||
}
|
||||
|
@ -231,7 +231,7 @@ sched_clock_register(u64 (*read)(void), int bits, unsigned long rate)
|
||||
if (irqtime > 0 || (irqtime == -1 && rate >= 1000000))
|
||||
enable_sched_clock_irqtime();
|
||||
|
||||
pr_debug("Registered %pF as sched_clock source\n", read);
|
||||
pr_debug("Registered %pS as sched_clock source\n", read);
|
||||
}
|
||||
|
||||
void __init generic_sched_clock_init(void)
|
||||
|
@ -1328,7 +1328,7 @@ static void call_timer_fn(struct timer_list *timer, void (*fn)(struct timer_list
|
||||
lock_map_release(&lockdep_map);
|
||||
|
||||
if (count != preempt_count()) {
|
||||
WARN_ONCE(1, "timer: %pF preempt leak: %08x -> %08x\n",
|
||||
WARN_ONCE(1, "timer: %pS preempt leak: %08x -> %08x\n",
|
||||
fn, count, preempt_count());
|
||||
/*
|
||||
* Restore the preempt count. That gives us a decent
|
||||
|
@ -2277,7 +2277,7 @@ __acquires(&pool->lock)
|
||||
|
||||
if (unlikely(in_atomic() || lockdep_depth(current) > 0)) {
|
||||
pr_err("BUG: workqueue leaked lock or atomic: %s/0x%08x/%d\n"
|
||||
" last function: %pf\n",
|
||||
" last function: %ps\n",
|
||||
current->comm, preempt_count(), task_pid_nr(current),
|
||||
worker->current_func);
|
||||
debug_show_held_locks(current);
|
||||
@ -2596,11 +2596,11 @@ static void check_flush_dependency(struct workqueue_struct *target_wq,
|
||||
worker = current_wq_worker();
|
||||
|
||||
WARN_ONCE(current->flags & PF_MEMALLOC,
|
||||
"workqueue: PF_MEMALLOC task %d(%s) is flushing !WQ_MEM_RECLAIM %s:%pf",
|
||||
"workqueue: PF_MEMALLOC task %d(%s) is flushing !WQ_MEM_RECLAIM %s:%ps",
|
||||
current->pid, current->comm, target_wq->name, target_func);
|
||||
WARN_ONCE(worker && ((worker->current_pwq->wq->flags &
|
||||
(WQ_MEM_RECLAIM | __WQ_LEGACY)) == WQ_MEM_RECLAIM),
|
||||
"workqueue: WQ_MEM_RECLAIM %s:%pf is flushing !WQ_MEM_RECLAIM %s:%pf",
|
||||
"workqueue: WQ_MEM_RECLAIM %s:%ps is flushing !WQ_MEM_RECLAIM %s:%ps",
|
||||
worker->current_pwq->wq->name, worker->current_func,
|
||||
target_wq->name, target_func);
|
||||
}
|
||||
@ -4582,7 +4582,7 @@ void print_worker_info(const char *log_lvl, struct task_struct *task)
|
||||
probe_kernel_read(desc, worker->desc, sizeof(desc) - 1);
|
||||
|
||||
if (fn || name[0] || desc[0]) {
|
||||
printk("%sWorkqueue: %s %pf", log_lvl, name, fn);
|
||||
printk("%sWorkqueue: %s %ps", log_lvl, name, fn);
|
||||
if (strcmp(name, desc))
|
||||
pr_cont(" (%s)", desc);
|
||||
pr_cont("\n");
|
||||
@ -4607,7 +4607,7 @@ static void pr_cont_work(bool comma, struct work_struct *work)
|
||||
pr_cont("%s BAR(%d)", comma ? "," : "",
|
||||
task_pid_nr(barr->task));
|
||||
} else {
|
||||
pr_cont("%s %pf", comma ? "," : "", work->func);
|
||||
pr_cont("%s %ps", comma ? "," : "", work->func);
|
||||
}
|
||||
}
|
||||
|
||||
@ -4639,7 +4639,7 @@ static void show_pwq(struct pool_workqueue *pwq)
|
||||
if (worker->current_pwq != pwq)
|
||||
continue;
|
||||
|
||||
pr_cont("%s %d%s:%pf", comma ? "," : "",
|
||||
pr_cont("%s %d%s:%ps", comma ? "," : "",
|
||||
task_pid_nr(worker->task),
|
||||
worker == pwq->wq->rescuer ? "(RESCUER)" : "",
|
||||
worker->current_func);
|
||||
|
@ -189,7 +189,7 @@ static int ei_seq_show(struct seq_file *m, void *v)
|
||||
{
|
||||
struct ei_entry *ent = list_entry(v, struct ei_entry, list);
|
||||
|
||||
seq_printf(m, "%pf\t%s\n", (void *)ent->start_addr,
|
||||
seq_printf(m, "%ps\t%s\n", (void *)ent->start_addr,
|
||||
error_type_string(ent->etype));
|
||||
return 0;
|
||||
}
|
||||
|
@ -151,7 +151,7 @@ static void percpu_ref_switch_to_atomic_rcu(struct rcu_head *rcu)
|
||||
atomic_long_add((long)count - PERCPU_COUNT_BIAS, &ref->count);
|
||||
|
||||
WARN_ONCE(atomic_long_read(&ref->count) <= 0,
|
||||
"percpu ref (%pf) <= 0 (%ld) after switching to atomic",
|
||||
"percpu ref (%ps) <= 0 (%ld) after switching to atomic",
|
||||
ref->release, atomic_long_read(&ref->count));
|
||||
|
||||
/* @ref is viewed as dead on all CPUs, send out switch confirmation */
|
||||
@ -333,7 +333,7 @@ void percpu_ref_kill_and_confirm(struct percpu_ref *ref,
|
||||
spin_lock_irqsave(&percpu_ref_switch_lock, flags);
|
||||
|
||||
WARN_ONCE(ref->percpu_count_ptr & __PERCPU_REF_DEAD,
|
||||
"%s called more than once on %pf!", __func__, ref->release);
|
||||
"%s called more than once on %ps!", __func__, ref->release);
|
||||
|
||||
ref->percpu_count_ptr |= __PERCPU_REF_DEAD;
|
||||
__percpu_ref_switch_mode(ref, confirm_kill);
|
||||
|
@ -701,7 +701,7 @@ int __init_memblock memblock_add(phys_addr_t base, phys_addr_t size)
|
||||
{
|
||||
phys_addr_t end = base + size - 1;
|
||||
|
||||
memblock_dbg("memblock_add: [%pa-%pa] %pF\n",
|
||||
memblock_dbg("memblock_add: [%pa-%pa] %pS\n",
|
||||
&base, &end, (void *)_RET_IP_);
|
||||
|
||||
return memblock_add_range(&memblock.memory, base, size, MAX_NUMNODES, 0);
|
||||
@ -820,7 +820,7 @@ int __init_memblock memblock_free(phys_addr_t base, phys_addr_t size)
|
||||
{
|
||||
phys_addr_t end = base + size - 1;
|
||||
|
||||
memblock_dbg(" memblock_free: [%pa-%pa] %pF\n",
|
||||
memblock_dbg(" memblock_free: [%pa-%pa] %pS\n",
|
||||
&base, &end, (void *)_RET_IP_);
|
||||
|
||||
kmemleak_free_part_phys(base, size);
|
||||
@ -831,7 +831,7 @@ int __init_memblock memblock_reserve(phys_addr_t base, phys_addr_t size)
|
||||
{
|
||||
phys_addr_t end = base + size - 1;
|
||||
|
||||
memblock_dbg("memblock_reserve: [%pa-%pa] %pF\n",
|
||||
memblock_dbg("memblock_reserve: [%pa-%pa] %pS\n",
|
||||
&base, &end, (void *)_RET_IP_);
|
||||
|
||||
return memblock_add_range(&memblock.reserved, base, size, MAX_NUMNODES, 0);
|
||||
@ -1466,7 +1466,7 @@ void * __init memblock_alloc_try_nid_raw(
|
||||
{
|
||||
void *ptr;
|
||||
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pF\n",
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pS\n",
|
||||
__func__, (u64)size, (u64)align, nid, &min_addr,
|
||||
&max_addr, (void *)_RET_IP_);
|
||||
|
||||
@ -1502,7 +1502,7 @@ void * __init memblock_alloc_try_nid_nopanic(
|
||||
{
|
||||
void *ptr;
|
||||
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pF\n",
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pS\n",
|
||||
__func__, (u64)size, (u64)align, nid, &min_addr,
|
||||
&max_addr, (void *)_RET_IP_);
|
||||
|
||||
@ -1538,7 +1538,7 @@ void * __init memblock_alloc_try_nid(
|
||||
{
|
||||
void *ptr;
|
||||
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pF\n",
|
||||
memblock_dbg("%s: %llu bytes align=0x%llx nid=%d from=%pa max_addr=%pa %pS\n",
|
||||
__func__, (u64)size, (u64)align, nid, &min_addr,
|
||||
&max_addr, (void *)_RET_IP_);
|
||||
ptr = memblock_alloc_internal(size, align,
|
||||
@ -1567,7 +1567,7 @@ void __init __memblock_free_late(phys_addr_t base, phys_addr_t size)
|
||||
phys_addr_t cursor, end;
|
||||
|
||||
end = base + size - 1;
|
||||
memblock_dbg("%s: [%pa-%pa] %pF\n",
|
||||
memblock_dbg("%s: [%pa-%pa] %pS\n",
|
||||
__func__, &base, &end, (void *)_RET_IP_);
|
||||
kmemleak_free_part_phys(base, size);
|
||||
cursor = PFN_UP(base);
|
||||
|
@ -519,7 +519,7 @@ static void print_bad_pte(struct vm_area_struct *vma, unsigned long addr,
|
||||
dump_page(page, "bad pte");
|
||||
pr_alert("addr:%p vm_flags:%08lx anon_vma:%p mapping:%p index:%lx\n",
|
||||
(void *)addr, vma->vm_flags, vma->anon_vma, mapping, index);
|
||||
pr_alert("file:%pD fault:%pf mmap:%pf readpage:%pf\n",
|
||||
pr_alert("file:%pD fault:%ps mmap:%ps readpage:%ps\n",
|
||||
vma->vm_file,
|
||||
vma->vm_ops ? vma->vm_ops->fault : NULL,
|
||||
vma->vm_file ? vma->vm_file->f_op->mmap : NULL,
|
||||
|
@ -493,7 +493,7 @@ static unsigned long do_shrink_slab(struct shrink_control *shrinkctl,
|
||||
|
||||
total_scan += delta;
|
||||
if (total_scan < 0) {
|
||||
pr_err("shrink_slab: %pF negative objects to delete nr=%ld\n",
|
||||
pr_err("shrink_slab: %pS negative objects to delete nr=%ld\n",
|
||||
shrinker->scan_objects, total_scan);
|
||||
total_scan = freeable;
|
||||
next_deferred = nr;
|
||||
|
@ -2398,7 +2398,7 @@ static void finish_request(struct ceph_osd_request *req)
|
||||
|
||||
static void __complete_request(struct ceph_osd_request *req)
|
||||
{
|
||||
dout("%s req %p tid %llu cb %pf result %d\n", __func__, req,
|
||||
dout("%s req %p tid %llu cb %ps result %d\n", __func__, req,
|
||||
req->r_tid, req->r_callback, req->r_result);
|
||||
|
||||
if (req->r_callback)
|
||||
|
@ -258,7 +258,7 @@ static int ptype_seq_show(struct seq_file *seq, void *v)
|
||||
else
|
||||
seq_printf(seq, "%04x", ntohs(pt->type));
|
||||
|
||||
seq_printf(seq, " %-8s %pf\n",
|
||||
seq_printf(seq, " %-8s %ps\n",
|
||||
pt->dev ? pt->dev->name : "", pt->func);
|
||||
}
|
||||
|
||||
|
@ -149,7 +149,7 @@ static void poll_one_napi(struct napi_struct *napi)
|
||||
* indicate that we are clearing the Tx path only.
|
||||
*/
|
||||
work = napi->poll(napi, 0);
|
||||
WARN_ONCE(work, "%pF exceeded budget in poll\n", napi->poll);
|
||||
WARN_ONCE(work, "%pS exceeded budget in poll\n", napi->poll);
|
||||
trace_napi_poll(napi, work, 0);
|
||||
|
||||
clear_bit(NAPI_STATE_NPSVC, &napi->state);
|
||||
@ -346,7 +346,7 @@ void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb,
|
||||
}
|
||||
|
||||
WARN_ONCE(!irqs_disabled(),
|
||||
"netpoll_send_skb_on_dev(): %s enabled interrupts in poll (%pF)\n",
|
||||
"netpoll_send_skb_on_dev(): %s enabled interrupts in poll (%pS)\n",
|
||||
dev->name, dev->netdev_ops->ndo_start_xmit);
|
||||
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user