| /* pcr.c: Generic sparc64 performance counter infrastructure. |
| * |
| * Copyright (C) 2009 David S. Miller (davem@davemloft.net) |
| */ |
| #include <linux/kernel.h> |
| #include <linux/module.h> |
| #include <linux/init.h> |
| #include <linux/irq.h> |
| |
| #include <linux/irq_work.h> |
| #include <linux/ftrace.h> |
| |
| #include <asm/pil.h> |
| #include <asm/pcr.h> |
| #include <asm/nmi.h> |
| |
| /* This code is shared between various users of the performance |
| * counters. Users will be oprofile, pseudo-NMI watchdog, and the |
| * perf_event support layer. |
| */ |
| |
| #define PCR_SUN4U_ENABLE (PCR_PIC_PRIV | PCR_STRACE | PCR_UTRACE) |
| #define PCR_N2_ENABLE (PCR_PIC_PRIV | PCR_STRACE | PCR_UTRACE | \ |
| PCR_N2_TOE_OV1 | \ |
| (2 << PCR_N2_SL1_SHIFT) | \ |
| (0xff << PCR_N2_MASK1_SHIFT)) |
| |
| u64 pcr_enable; |
| unsigned int picl_shift; |
| |
| /* Performance counter interrupts run unmasked at PIL level 15. |
| * Therefore we can't do things like wakeups and other work |
| * that expects IRQ disabling to be adhered to in locking etc. |
| * |
| * Therefore in such situations we defer the work by signalling |
| * a lower level cpu IRQ. |
| */ |
| void __irq_entry deferred_pcr_work_irq(int irq, struct pt_regs *regs) |
| { |
| struct pt_regs *old_regs; |
| |
| clear_softint(1 << PIL_DEFERRED_PCR_WORK); |
| |
| old_regs = set_irq_regs(regs); |
| irq_enter(); |
| #ifdef CONFIG_IRQ_WORK |
| irq_work_run(); |
| #endif |
| irq_exit(); |
| set_irq_regs(old_regs); |
| } |
| |
| void arch_irq_work_raise(void) |
| { |
| set_softint(1 << PIL_DEFERRED_PCR_WORK); |
| } |
| |
| const struct pcr_ops *pcr_ops; |
| EXPORT_SYMBOL_GPL(pcr_ops); |
| |
| static u64 direct_pcr_read(void) |
| { |
| u64 val; |
| |
| read_pcr(val); |
| return val; |
| } |
| |
| static void direct_pcr_write(u64 val) |
| { |
| write_pcr(val); |
| } |
| |
| static const struct pcr_ops direct_pcr_ops = { |
| .read = direct_pcr_read, |
| .write = direct_pcr_write, |
| }; |
| |
| static void n2_pcr_write(u64 val) |
| { |
| unsigned long ret; |
| |
| ret = sun4v_niagara2_setperf(HV_N2_PERF_SPARC_CTL, val); |
| if (val != HV_EOK) |
| write_pcr(val); |
| } |
| |
| static const struct pcr_ops n2_pcr_ops = { |
| .read = direct_pcr_read, |
| .write = n2_pcr_write, |
| }; |
| |
| static unsigned long perf_hsvc_group; |
| static unsigned long perf_hsvc_major; |
| static unsigned long perf_hsvc_minor; |
| |
| static int __init register_perf_hsvc(void) |
| { |
| if (tlb_type == hypervisor) { |
| switch (sun4v_chip_type) { |
| case SUN4V_CHIP_NIAGARA1: |
| perf_hsvc_group = HV_GRP_NIAG_PERF; |
| break; |
| |
| case SUN4V_CHIP_NIAGARA2: |
| perf_hsvc_group = HV_GRP_N2_CPU; |
| break; |
| |
| default: |
| return -ENODEV; |
| } |
| |
| |
| perf_hsvc_major = 1; |
| perf_hsvc_minor = 0; |
| if (sun4v_hvapi_register(perf_hsvc_group, |
| perf_hsvc_major, |
| &perf_hsvc_minor)) { |
| printk("perfmon: Could not register hvapi.\n"); |
| return -ENODEV; |
| } |
| } |
| return 0; |
| } |
| |
| static void __init unregister_perf_hsvc(void) |
| { |
| if (tlb_type != hypervisor) |
| return; |
| sun4v_hvapi_unregister(perf_hsvc_group); |
| } |
| |
| int __init pcr_arch_init(void) |
| { |
| int err = register_perf_hsvc(); |
| |
| if (err) |
| return err; |
| |
| switch (tlb_type) { |
| case hypervisor: |
| pcr_ops = &n2_pcr_ops; |
| pcr_enable = PCR_N2_ENABLE; |
| picl_shift = 2; |
| break; |
| |
| case cheetah: |
| case cheetah_plus: |
| pcr_ops = &direct_pcr_ops; |
| pcr_enable = PCR_SUN4U_ENABLE; |
| break; |
| |
| case spitfire: |
| /* UltraSPARC-I/II and derivatives lack a profile |
| * counter overflow interrupt so we can't make use of |
| * their hardware currently. |
| */ |
| /* fallthrough */ |
| default: |
| err = -ENODEV; |
| goto out_unregister; |
| } |
| |
| return nmi_init(); |
| |
| out_unregister: |
| unregister_perf_hsvc(); |
| return err; |
| } |
| |
| arch_initcall(pcr_arch_init); |