blob: 6a0f2c0eea6c38dde62e9d74cc0a1212ff9c6954 [file] [log] [blame]
/**
* Copyright (C) ARM Limited 2010-2013. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 as
* published by the Free Software Foundation.
*
*/
// This version must match the gator daemon version
static unsigned long gator_protocol_version = 16;
#include <linux/slab.h>
#include <linux/cpu.h>
#include <linux/sched.h>
#include <linux/irq.h>
#include <linux/vmalloc.h>
#include <linux/hardirq.h>
#include <linux/highmem.h>
#include <linux/pagemap.h>
#include <linux/suspend.h>
#include <linux/module.h>
#include <linux/perf_event.h>
#include <linux/utsname.h>
#include <asm/stacktrace.h>
#include <asm/uaccess.h>
#include "gator.h"
#include "gator_events.h"
#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32)
#error kernels prior to 2.6.32 are not supported
#endif
#if !defined(CONFIG_GENERIC_TRACER) && !defined(CONFIG_TRACING)
#error gator requires the kernel to have CONFIG_GENERIC_TRACER or CONFIG_TRACING defined
#endif
#ifndef CONFIG_PROFILING
#error gator requires the kernel to have CONFIG_PROFILING defined
#endif
#ifndef CONFIG_HIGH_RES_TIMERS
#error gator requires the kernel to have CONFIG_HIGH_RES_TIMERS defined to support PC sampling
#endif
#if defined(__arm__) && defined(CONFIG_SMP) && !defined(CONFIG_LOCAL_TIMERS)
//#error gator requires the kernel to have CONFIG_LOCAL_TIMERS defined on SMP systems
#endif
#if (GATOR_PERF_SUPPORT) && (!(GATOR_PERF_PMU_SUPPORT))
#ifndef CONFIG_PERF_EVENTS
#warning gator requires the kernel to have CONFIG_PERF_EVENTS defined to support pmu hardware counters
#elif !defined CONFIG_HW_PERF_EVENTS
#warning gator requires the kernel to have CONFIG_HW_PERF_EVENTS defined to support pmu hardware counters
#endif
#endif
/******************************************************************************
* DEFINES
******************************************************************************/
#define SUMMARY_BUFFER_SIZE (1*1024)
#define BACKTRACE_BUFFER_SIZE (128*1024)
#define NAME_BUFFER_SIZE (64*1024)
#define COUNTER_BUFFER_SIZE (64*1024) // counters have the core as part of the data and the core value in the frame header may be discarded
#define BLOCK_COUNTER_BUFFER_SIZE (128*1024)
#define ANNOTATE_BUFFER_SIZE (128*1024) // annotate counters have the core as part of the data and the core value in the frame header may be discarded
#define SCHED_TRACE_BUFFER_SIZE (128*1024)
#define GPU_TRACE_BUFFER_SIZE (64*1024) // gpu trace counters have the core as part of the data and the core value in the frame header may be discarded
#define IDLE_BUFFER_SIZE (32*1024) // idle counters have the core as part of the data and the core value in the frame header may be discarded
#define NO_COOKIE 0U
#define UNRESOLVED_COOKIE ~0U
#define FRAME_SUMMARY 1
#define FRAME_BACKTRACE 2
#define FRAME_NAME 3
#define FRAME_COUNTER 4
#define FRAME_BLOCK_COUNTER 5
#define FRAME_ANNOTATE 6
#define FRAME_SCHED_TRACE 7
#define FRAME_GPU_TRACE 8
#define FRAME_IDLE 9
#define MESSAGE_END_BACKTRACE 1
#define MESSAGE_COOKIE 1
#define MESSAGE_THREAD_NAME 2
#define HRTIMER_CORE_NAME 3
#define MESSAGE_GPU_START 1
#define MESSAGE_GPU_STOP 2
#define MESSAGE_SCHED_SWITCH 1
#define MESSAGE_SCHED_EXIT 2
#define MESSAGE_SCHED_START 3
#define MESSAGE_IDLE_ENTER 1
#define MESSAGE_IDLE_EXIT 2
#define MAXSIZE_PACK32 5
#define MAXSIZE_PACK64 10
#define FRAME_HEADER_SIZE 3
#if defined(__arm__)
#define PC_REG regs->ARM_pc
#elif defined(__aarch64__)
#define PC_REG regs->pc
#else
#define PC_REG regs->ip
#endif
enum {
SUMMARY_BUF,
BACKTRACE_BUF,
NAME_BUF,
COUNTER_BUF,
BLOCK_COUNTER_BUF,
ANNOTATE_BUF,
SCHED_TRACE_BUF,
GPU_TRACE_BUF,
IDLE_BUF,
NUM_GATOR_BUFS
};
/******************************************************************************
* Globals
******************************************************************************/
static unsigned long gator_cpu_cores;
// Size of the largest buffer. Effectively constant, set in gator_op_create_files
static unsigned long userspace_buffer_size;
static unsigned long gator_backtrace_depth;
// How often to commit the buffers for live in nanoseconds
static u64 gator_live_rate;
static unsigned long gator_started;
static u64 gator_monotonic_started;
static unsigned long gator_buffer_opened;
static unsigned long gator_timer_count;
static unsigned long gator_response_type;
static DEFINE_MUTEX(start_mutex);
static DEFINE_MUTEX(gator_buffer_mutex);
bool event_based_sampling;
static DECLARE_WAIT_QUEUE_HEAD(gator_buffer_wait);
static DECLARE_WAIT_QUEUE_HEAD(gator_annotate_wait);
static struct timer_list gator_buffer_wake_up_timer;
static LIST_HEAD(gator_events);
static DEFINE_PER_CPU(u64, last_timestamp);
static bool printed_monotonic_warning;
static bool sent_core_name[NR_CPUS];
/******************************************************************************
* Prototypes
******************************************************************************/
static void buffer_check(int cpu, int buftype, u64 time);
static void gator_commit_buffer(int cpu, int buftype, u64 time);
static int buffer_bytes_available(int cpu, int buftype);
static bool buffer_check_space(int cpu, int buftype, int bytes);
static int contiguous_space_available(int cpu, int bufytpe);
static void gator_buffer_write_packed_int(int cpu, int buftype, int x);
static void gator_buffer_write_packed_int64(int cpu, int buftype, long long x);
static void gator_buffer_write_bytes(int cpu, int buftype, const char *x, int len);
static void gator_buffer_write_string(int cpu, int buftype, const char *x);
static void gator_add_trace(int cpu, unsigned long address);
static void gator_add_sample(int cpu, struct pt_regs *const regs, u64 time);
static u64 gator_get_time(void);
// Size of the buffer, must be a power of 2. Effectively constant, set in gator_op_setup.
static uint32_t gator_buffer_size[NUM_GATOR_BUFS];
// gator_buffer_size - 1, bitwise and with pos to get offset into the array. Effectively constant, set in gator_op_setup.
static uint32_t gator_buffer_mask[NUM_GATOR_BUFS];
// Read position in the buffer. Initialized to zero in gator_op_setup and incremented after bytes are read by userspace in userspace_buffer_read
static DEFINE_PER_CPU(int[NUM_GATOR_BUFS], gator_buffer_read);
// Write position in the buffer. Initialized to zero in gator_op_setup and incremented after bytes are written to the buffer
static DEFINE_PER_CPU(int[NUM_GATOR_BUFS], gator_buffer_write);
// Commit position in the buffer. Initialized to zero in gator_op_setup and incremented after a frame is ready to be read by userspace
static DEFINE_PER_CPU(int[NUM_GATOR_BUFS], gator_buffer_commit);
// If set to false, decreases the number of bytes returned by buffer_bytes_available. Set in buffer_check_space if no space is remaining. Initialized to true in gator_op_setup
// This means that if we run out of space, continue to report that no space is available until bytes are read by userspace
static DEFINE_PER_CPU(int[NUM_GATOR_BUFS], buffer_space_available);
// The buffer. Allocated in gator_op_setup
static DEFINE_PER_CPU(char *[NUM_GATOR_BUFS], gator_buffer);
// The time after which the buffer should be committed for live display
static DEFINE_PER_CPU(u64, gator_buffer_commit_time);
/******************************************************************************
* Application Includes
******************************************************************************/
#include "gator_marshaling.c"
#include "gator_hrtimer_perf.c"
#include "gator_hrtimer_gator.c"
#include "gator_cookies.c"
#include "gator_annotate.c"
#include "gator_trace_sched.c"
#include "gator_trace_power.c"
#include "gator_trace_gpu.c"
#include "gator_backtrace.c"
#include "gator_fs.c"
#include "gator_pack.c"
/******************************************************************************
* Misc
******************************************************************************/
const struct gator_cpu gator_cpus[] = {
{
.cpuid = ARM1136,
.core_name = "ARM1136",
.pmnc_name = "ARM_ARM11",
.dt_name = "arm,arm1136",
.pmnc_counters = 3,
},
{
.cpuid = ARM1156,
.core_name = "ARM1156",
.pmnc_name = "ARM_ARM11",
.dt_name = "arm,arm1156",
.pmnc_counters = 3,
},
{
.cpuid = ARM1176,
.core_name = "ARM1176",
.pmnc_name = "ARM_ARM11",
.dt_name = "arm,arm1176",
.pmnc_counters = 3,
},
{
.cpuid = ARM11MPCORE,
.core_name = "ARM11MPCore",
.pmnc_name = "ARM_ARM11MPCore",
.dt_name = "arm,arm11mpcore",
.pmnc_counters = 3,
},
{
.cpuid = CORTEX_A5,
.core_name = "Cortex-A5",
.pmu_name = "ARMv7_Cortex_A5",
.pmnc_name = "ARM_Cortex-A5",
.dt_name = "arm,cortex-a5",
.pmnc_counters = 2,
},
{
.cpuid = CORTEX_A7,
.core_name = "Cortex-A7",
.pmu_name = "ARMv7_Cortex_A7",
.pmnc_name = "ARM_Cortex-A7",
.dt_name = "arm,cortex-a7",
.pmnc_counters = 4,
},
{
.cpuid = CORTEX_A8,
.core_name = "Cortex-A8",
.pmu_name = "ARMv7_Cortex_A8",
.pmnc_name = "ARM_Cortex-A8",
.dt_name = "arm,cortex-a8",
.pmnc_counters = 4,
},
{
.cpuid = CORTEX_A9,
.core_name = "Cortex-A9",
.pmu_name = "ARMv7_Cortex_A9",
.pmnc_name = "ARM_Cortex-A9",
.dt_name = "arm,cortex-a9",
.pmnc_counters = 6,
},
{
.cpuid = CORTEX_A12,
.core_name = "Cortex-A12",
.pmu_name = "ARMv7_Cortex_A12",
.pmnc_name = "ARM_Cortex-A12",
.dt_name = "arm,cortex-a12",
.pmnc_counters = 6,
},
{
.cpuid = CORTEX_A15,
.core_name = "Cortex-A15",
.pmu_name = "ARMv7_Cortex_A15",
.pmnc_name = "ARM_Cortex-A15",
.dt_name = "arm,cortex-a15",
.pmnc_counters = 6,
},
{
.cpuid = SCORPION,
.core_name = "Scorpion",
.pmnc_name = "Scorpion",
.pmnc_counters = 4,
},
{
.cpuid = SCORPIONMP,
.core_name = "ScorpionMP",
.pmnc_name = "ScorpionMP",
.pmnc_counters = 4,
},
{
.cpuid = KRAITSIM,
.core_name = "KraitSIM",
.pmnc_name = "Krait",
.pmnc_counters = 4,
},
{
.cpuid = KRAIT,
.core_name = "Krait",
.pmnc_name = "Krait",
.pmnc_counters = 4,
},
{
.cpuid = KRAIT_S4_PRO,
.core_name = "Krait S4 Pro",
.pmnc_name = "Krait",
.pmnc_counters = 4,
},
{
.cpuid = CORTEX_A53,
.core_name = "Cortex-A53",
.pmnc_name = "ARM_Cortex-A53",
.dt_name = "arm,cortex-a53",
.pmnc_counters = 6,
},
{
.cpuid = CORTEX_A57,
.core_name = "Cortex-A57",
.pmnc_name = "ARM_Cortex-A57",
.dt_name = "arm,cortex-a57",
.pmnc_counters = 6,
},
{
.cpuid = AARCH64,
.core_name = "AArch64",
.pmnc_name = "ARM_AArch64",
.pmnc_counters = 6,
},
{
.cpuid = OTHER,
.core_name = "Other",
.pmnc_name = "Other",
.pmnc_counters = 6,
},
{}
};
const struct gator_cpu *gator_find_cpu_by_cpuid(const u32 cpuid)
{
int i;
for (i = 0; gator_cpus[i].cpuid != 0; ++i) {
const struct gator_cpu *const gator_cpu = &gator_cpus[i];
if (gator_cpu->cpuid == cpuid) {
return gator_cpu;
}
}
return NULL;
}
const struct gator_cpu *gator_find_cpu_by_pmu_name(const char *const name)
{
int i;
for (i = 0; gator_cpus[i].cpuid != 0; ++i) {
const struct gator_cpu *const gator_cpu = &gator_cpus[i];
if (gator_cpu->pmu_name != NULL && strcmp(gator_cpu->pmu_name, name) == 0) {
return gator_cpu;
}
}
return NULL;
}
u32 gator_cpuid(void)
{
#if defined(__arm__) || defined(__aarch64__)
u32 val;
#if !defined(__aarch64__)
asm volatile("mrc p15, 0, %0, c0, c0, 0" : "=r" (val));
#else
asm volatile("mrs %0, midr_el1" : "=r" (val));
#endif
return (val >> 4) & 0xfff;
#else
return OTHER;
#endif
}
static void gator_buffer_wake_up(unsigned long data)
{
wake_up(&gator_buffer_wait);
}
/******************************************************************************
* Commit interface
******************************************************************************/
static bool buffer_commit_ready(int *cpu, int *buftype)
{
int cpu_x, x;
for_each_present_cpu(cpu_x) {
for (x = 0; x < NUM_GATOR_BUFS; x++)
if (per_cpu(gator_buffer_commit, cpu_x)[x] != per_cpu(gator_buffer_read, cpu_x)[x]) {
*cpu = cpu_x;
*buftype = x;
return true;
}
}
*cpu = -1;
*buftype = -1;
return false;
}
/******************************************************************************
* Buffer management
******************************************************************************/
static int buffer_bytes_available(int cpu, int buftype)
{
int remaining, filled;
filled = per_cpu(gator_buffer_write, cpu)[buftype] - per_cpu(gator_buffer_read, cpu)[buftype];
if (filled < 0) {
filled += gator_buffer_size[buftype];
}
remaining = gator_buffer_size[buftype] - filled;
if (per_cpu(buffer_space_available, cpu)[buftype]) {
// Give some extra room; also allows space to insert the overflow error packet
remaining -= 200;
} else {
// Hysteresis, prevents multiple overflow messages
remaining -= 2000;
}
return remaining;
}
static int contiguous_space_available(int cpu, int buftype)
{
int remaining = buffer_bytes_available(cpu, buftype);
int contiguous = gator_buffer_size[buftype] - per_cpu(gator_buffer_write, cpu)[buftype];
if (remaining < contiguous)
return remaining;
else
return contiguous;
}
static bool buffer_check_space(int cpu, int buftype, int bytes)
{
int remaining = buffer_bytes_available(cpu, buftype);
if (remaining < bytes) {
per_cpu(buffer_space_available, cpu)[buftype] = false;
} else {
per_cpu(buffer_space_available, cpu)[buftype] = true;
}
return per_cpu(buffer_space_available, cpu)[buftype];
}
static void gator_buffer_write_bytes(int cpu, int buftype, const char *x, int len)
{
int i;
u32 write = per_cpu(gator_buffer_write, cpu)[buftype];
u32 mask = gator_buffer_mask[buftype];
char *buffer = per_cpu(gator_buffer, cpu)[buftype];
for (i = 0; i < len; i++) {
buffer[write] = x[i];
write = (write + 1) & mask;
}
per_cpu(gator_buffer_write, cpu)[buftype] = write;
}
static void gator_buffer_write_string(int cpu, int buftype, const char *x)
{
int len = strlen(x);
gator_buffer_write_packed_int(cpu, buftype, len);
gator_buffer_write_bytes(cpu, buftype, x, len);
}
static void gator_commit_buffer(int cpu, int buftype, u64 time)
{
int type_length, commit, length, byte;
if (!per_cpu(gator_buffer, cpu)[buftype])
return;
// post-populate the length, which does not include the response type length nor the length itself, i.e. only the length of the payload
type_length = gator_response_type ? 1 : 0;
commit = per_cpu(gator_buffer_commit, cpu)[buftype];
length = per_cpu(gator_buffer_write, cpu)[buftype] - commit;
if (length < 0) {
length += gator_buffer_size[buftype];
}
length = length - type_length - sizeof(s32);
if (length <= FRAME_HEADER_SIZE) {
// Nothing to write, only the frame header is present
return;
}
for (byte = 0; byte < sizeof(s32); byte++) {
per_cpu(gator_buffer, cpu)[buftype][(commit + type_length + byte) & gator_buffer_mask[buftype]] = (length >> byte * 8) & 0xFF;
}
per_cpu(gator_buffer_commit, cpu)[buftype] = per_cpu(gator_buffer_write, cpu)[buftype];
if (gator_live_rate > 0) {
while (time > per_cpu(gator_buffer_commit_time, cpu)) {
per_cpu(gator_buffer_commit_time, cpu) += gator_live_rate;
}
}
marshal_frame(cpu, buftype);
// had to delay scheduling work as attempting to schedule work during the context switch is illegal in kernel versions 3.5 and greater
mod_timer(&gator_buffer_wake_up_timer, jiffies + 1);
}
static void buffer_check(int cpu, int buftype, u64 time)
{
int filled = per_cpu(gator_buffer_write, cpu)[buftype] - per_cpu(gator_buffer_commit, cpu)[buftype];
if (filled < 0) {
filled += gator_buffer_size[buftype];
}
if (filled >= ((gator_buffer_size[buftype] * 3) / 4)) {
gator_commit_buffer(cpu, buftype, time);
}
}
static void gator_add_trace(int cpu, unsigned long address)
{
off_t offset = 0;
unsigned long cookie = get_address_cookie(cpu, current, address & ~1, &offset);
if (cookie == NO_COOKIE || cookie == UNRESOLVED_COOKIE) {
offset = address;
}
marshal_backtrace(offset & ~1, cookie);
}
static void gator_add_sample(int cpu, struct pt_regs *const regs, u64 time)
{
bool inKernel;
unsigned long exec_cookie;
if (!regs)
return;
inKernel = !user_mode(regs);
exec_cookie = get_exec_cookie(cpu, current);
if (!marshal_backtrace_header(exec_cookie, current->tgid, current->pid, inKernel, time))
return;
if (inKernel) {
kernel_backtrace(cpu, regs);
} else {
// Cookie+PC
gator_add_trace(cpu, PC_REG);
// Backtrace
if (gator_backtrace_depth)
arm_backtrace_eabi(cpu, regs, gator_backtrace_depth);
}
marshal_backtrace_footer(time);
}
/******************************************************************************
* hrtimer interrupt processing
******************************************************************************/
static void gator_timer_interrupt(void)
{
struct pt_regs *const regs = get_irq_regs();
gator_backtrace_handler(regs);
}
void gator_backtrace_handler(struct pt_regs *const regs)
{
u64 time = gator_get_time();
int cpu = get_physical_cpu();
// Output backtrace
gator_add_sample(cpu, regs, time);
// Collect counters
if (!per_cpu(collecting, cpu)) {
collect_counters(time);
}
}
static int gator_running;
// This function runs in interrupt context and on the appropriate core
static void gator_timer_offline(void *migrate)
{
struct gator_interface *gi;
int i, len, cpu = get_physical_cpu();
int *buffer;
u64 time;
gator_trace_sched_offline();
gator_trace_power_offline();
if (!migrate) {
gator_hrtimer_offline();
}
// Offline any events and output counters
time = gator_get_time();
if (marshal_event_header(time)) {
list_for_each_entry(gi, &gator_events, list) {
if (gi->offline) {
len = gi->offline(&buffer, migrate);
marshal_event(len, buffer);
}
}
// Only check after writing all counters so that time and corresponding counters appear in the same frame
buffer_check(cpu, BLOCK_COUNTER_BUF, time);
}
// Flush all buffers on this core
for (i = 0; i < NUM_GATOR_BUFS; i++)
gator_commit_buffer(cpu, i, time);
}
// This function runs in interrupt context and may be running on a core other than core 'cpu'
static void gator_timer_offline_dispatch(int cpu, bool migrate)
{
struct gator_interface *gi;
list_for_each_entry(gi, &gator_events, list) {
if (gi->offline_dispatch) {
gi->offline_dispatch(cpu, migrate);
}
}
}
static void gator_timer_stop(void)
{
int cpu;
if (gator_running) {
on_each_cpu(gator_timer_offline, NULL, 1);
for_each_online_cpu(cpu) {
gator_timer_offline_dispatch(lcpu_to_pcpu(cpu), false);
}
gator_running = 0;
gator_hrtimer_shutdown();
}
}
#if defined(__arm__) || defined(__aarch64__)
static void gator_send_core_name(int cpu, const u32 cpuid, const struct gator_cpu *const gator_cpu) {
const char *core_name = NULL;
char core_name_buf[32];
if (!sent_core_name[cpu]) {
if (gator_cpu != NULL) {
core_name = gator_cpu->core_name;
} else {
snprintf(core_name_buf, sizeof(core_name_buf), "Unknown (0x%.3x)", cpuid);
core_name = core_name_buf;
}
marshal_core_name(cpu, cpuid, core_name);
sent_core_name[cpu] = true;
}
}
#endif
// This function runs in interrupt context and on the appropriate core
static void gator_timer_online(void *migrate)
{
struct gator_interface *gi;
int len, cpu = get_physical_cpu();
int *buffer;
u64 time;
gator_trace_power_online();
// online any events and output counters
time = gator_get_time();
if (marshal_event_header(time)) {
list_for_each_entry(gi, &gator_events, list) {
if (gi->online) {
len = gi->online(&buffer, migrate);
marshal_event(len, buffer);
}
}
// Only check after writing all counters so that time and corresponding counters appear in the same frame
buffer_check(cpu, BLOCK_COUNTER_BUF, time);
}
if (!migrate) {
gator_hrtimer_online();
}
#if defined(__arm__) || defined(__aarch64__)
if (!sent_core_name[cpu]) {
const u32 cpuid = gator_cpuid();
gator_send_core_name(cpu, cpuid, gator_find_cpu_by_cpuid(cpuid));
}
#endif
}
// This function runs in interrupt context and may be running on a core other than core 'cpu'
static void gator_timer_online_dispatch(int cpu, bool migrate)
{
struct gator_interface *gi;
list_for_each_entry(gi, &gator_events, list) {
if (gi->online_dispatch) {
gi->online_dispatch(cpu, migrate);
}
}
}
#include "gator_iks.c"
int gator_timer_start(unsigned long sample_rate)
{
int cpu;
if (gator_running) {
pr_notice("gator: already running\n");
return 0;
}
gator_running = 1;
// event based sampling trumps hr timer based sampling
if (event_based_sampling) {
sample_rate = 0;
}
if (gator_hrtimer_init(sample_rate, gator_timer_interrupt) == -1)
return -1;
gator_send_iks_core_names();
for_each_online_cpu(cpu) {
gator_timer_online_dispatch(lcpu_to_pcpu(cpu), false);
}
on_each_cpu(gator_timer_online, NULL, 1);
return 0;
}
static u64 gator_get_time(void)
{
struct timespec ts;
u64 timestamp;
u64 prev_timestamp;
u64 delta;
int cpu = smp_processor_id();
// Match clock_gettime(CLOCK_MONOTONIC_RAW, &ts) from userspace
getrawmonotonic(&ts);
timestamp = timespec_to_ns(&ts);
// getrawmonotonic is not monotonic on all systems. Detect and attempt to correct these cases.
// up to 0.5ms delta has been seen on some systems, which can skew Streamline data when viewing at high resolution.
// This doesn't work well with interrupts, but that it's OK - the real concern is to catch big jumps in time
prev_timestamp = per_cpu(last_timestamp, cpu);
if (prev_timestamp <= timestamp) {
per_cpu(last_timestamp, cpu) = timestamp;
} else {
delta = prev_timestamp - timestamp;
// Log the error once
if (!printed_monotonic_warning && delta > 500000) {
printk(KERN_ERR "%s: getrawmonotonic is not monotonic cpu: %i delta: %lli\nSkew in Streamline data may be present at the fine zoom levels\n", __FUNCTION__, cpu, delta);
printed_monotonic_warning = true;
}
timestamp = prev_timestamp;
}
return timestamp - gator_monotonic_started;
}
/******************************************************************************
* cpu hotplug and pm notifiers
******************************************************************************/
static int __cpuinit gator_hotcpu_notify(struct notifier_block *self, unsigned long action, void *hcpu)
{
int cpu = lcpu_to_pcpu((long)hcpu);
switch (action) {
case CPU_DOWN_PREPARE:
case CPU_DOWN_PREPARE_FROZEN:
smp_call_function_single(cpu, gator_timer_offline, NULL, 1);
gator_timer_offline_dispatch(cpu, false);
break;
case CPU_ONLINE:
case CPU_ONLINE_FROZEN:
gator_timer_online_dispatch(cpu, false);
smp_call_function_single(cpu, gator_timer_online, NULL, 1);
break;
}
return NOTIFY_OK;
}
static struct notifier_block __refdata gator_hotcpu_notifier = {
.notifier_call = gator_hotcpu_notify,
};
// n.b. calling "on_each_cpu" only runs on those that are online
// Registered linux events are not disabled, so their counters will continue to collect
static int gator_pm_notify(struct notifier_block *nb, unsigned long event, void *dummy)
{
int cpu;
switch (event) {
case PM_HIBERNATION_PREPARE:
case PM_SUSPEND_PREPARE:
unregister_hotcpu_notifier(&gator_hotcpu_notifier);
unregister_scheduler_tracepoints();
on_each_cpu(gator_timer_offline, NULL, 1);
for_each_online_cpu(cpu) {
gator_timer_offline_dispatch(lcpu_to_pcpu(cpu), false);
}
break;
case PM_POST_HIBERNATION:
case PM_POST_SUSPEND:
for_each_online_cpu(cpu) {
gator_timer_online_dispatch(lcpu_to_pcpu(cpu), false);
}
on_each_cpu(gator_timer_online, NULL, 1);
register_scheduler_tracepoints();
register_hotcpu_notifier(&gator_hotcpu_notifier);
break;
}
return NOTIFY_OK;
}
static struct notifier_block gator_pm_notifier = {
.notifier_call = gator_pm_notify,
};
static int gator_notifier_start(void)
{
int retval;
retval = register_hotcpu_notifier(&gator_hotcpu_notifier);
if (retval == 0)
retval = register_pm_notifier(&gator_pm_notifier);
return retval;
}
static void gator_notifier_stop(void)
{
unregister_pm_notifier(&gator_pm_notifier);
unregister_hotcpu_notifier(&gator_hotcpu_notifier);
}
/******************************************************************************
* Main
******************************************************************************/
static void gator_summary(void)
{
u64 timestamp, uptime;
struct timespec ts;
char uname_buf[512];
void (*m2b)(struct timespec *ts);
unsigned long flags;
snprintf(uname_buf, sizeof(uname_buf), "%s %s %s %s %s GNU/Linux", utsname()->sysname, utsname()->nodename, utsname()->release, utsname()->version, utsname()->machine);
getnstimeofday(&ts);
timestamp = timespec_to_ns(&ts);
do_posix_clock_monotonic_gettime(&ts);
// monotonic_to_bootbased is not defined for some versions of Android
m2b = symbol_get(monotonic_to_bootbased);
if (m2b) {
m2b(&ts);
}
uptime = timespec_to_ns(&ts);
// Disable interrupts as gator_get_time calls smp_processor_id to verify time is monotonic
local_irq_save(flags);
// Set monotonic_started to zero as gator_get_time is uptime minus monotonic_started
gator_monotonic_started = 0;
gator_monotonic_started = gator_get_time();
local_irq_restore(flags);
marshal_summary(timestamp, uptime, gator_monotonic_started, uname_buf);
}
int gator_events_install(struct gator_interface *interface)
{
list_add_tail(&interface->list, &gator_events);
return 0;
}
int gator_events_get_key(void)
{
// key of zero is reserved as a timestamp
static int key = 1;
const int ret = key;
key += 2;
return ret;
}
static int gator_init(void)
{
int i;
calc_first_cluster_size();
// events sources (gator_events.h, generated by gator_events.sh)
for (i = 0; i < ARRAY_SIZE(gator_events_list); i++)
if (gator_events_list[i])
gator_events_list[i]();
gator_trace_sched_init();
gator_trace_power_init();
return 0;
}
static void gator_exit(void)
{
struct gator_interface *gi;
list_for_each_entry(gi, &gator_events, list)
if (gi->shutdown)
gi->shutdown();
}
static int gator_start(void)
{
unsigned long cpu, i;
struct gator_interface *gi;
if (gator_migrate_start())
goto migrate_failure;
// Initialize the buffer with the frame type and core
for_each_present_cpu(cpu) {
for (i = 0; i < NUM_GATOR_BUFS; i++) {
marshal_frame(cpu, i);
}
per_cpu(last_timestamp, cpu) = 0;
}
printed_monotonic_warning = false;
// Capture the start time
gator_summary();
// start all events
list_for_each_entry(gi, &gator_events, list) {
if (gi->start && gi->start() != 0) {
struct list_head *ptr = gi->list.prev;
while (ptr != &gator_events) {
gi = list_entry(ptr, struct gator_interface, list);
if (gi->stop)
gi->stop();
ptr = ptr->prev;
}
goto events_failure;
}
}
// cookies shall be initialized before trace_sched_start() and gator_timer_start()
if (cookies_initialize())
goto cookies_failure;
if (gator_annotate_start())
goto annotate_failure;
if (gator_trace_sched_start())
goto sched_failure;
if (gator_trace_power_start())
goto power_failure;
if (gator_trace_gpu_start())
goto gpu_failure;
if (gator_timer_start(gator_timer_count))
goto timer_failure;
if (gator_notifier_start())
goto notifier_failure;
return 0;
notifier_failure:
gator_timer_stop();
timer_failure:
gator_trace_gpu_stop();
gpu_failure:
gator_trace_power_stop();
power_failure:
gator_trace_sched_stop();
sched_failure:
gator_annotate_stop();
annotate_failure:
cookies_release();
cookies_failure:
// stop all events
list_for_each_entry(gi, &gator_events, list)
if (gi->stop)
gi->stop();
events_failure:
gator_migrate_stop();
migrate_failure:
return -1;
}
static void gator_stop(void)
{
struct gator_interface *gi;
gator_annotate_stop();
gator_trace_sched_stop();
gator_trace_power_stop();
gator_trace_gpu_stop();
// stop all interrupt callback reads before tearing down other interfaces
gator_notifier_stop(); // should be called before gator_timer_stop to avoid re-enabling the hrtimer after it has been offlined
gator_timer_stop();
// stop all events
list_for_each_entry(gi, &gator_events, list)
if (gi->stop)
gi->stop();
gator_migrate_stop();
}
/******************************************************************************
* Filesystem
******************************************************************************/
/* fopen("buffer") */
static int gator_op_setup(void)
{
int err = 0;
int cpu, i;
mutex_lock(&start_mutex);
gator_buffer_size[SUMMARY_BUF] = SUMMARY_BUFFER_SIZE;
gator_buffer_mask[SUMMARY_BUF] = SUMMARY_BUFFER_SIZE - 1;
gator_buffer_size[BACKTRACE_BUF] = BACKTRACE_BUFFER_SIZE;
gator_buffer_mask[BACKTRACE_BUF] = BACKTRACE_BUFFER_SIZE - 1;
gator_buffer_size[NAME_BUF] = NAME_BUFFER_SIZE;
gator_buffer_mask[NAME_BUF] = NAME_BUFFER_SIZE - 1;
gator_buffer_size[COUNTER_BUF] = COUNTER_BUFFER_SIZE;
gator_buffer_mask[COUNTER_BUF] = COUNTER_BUFFER_SIZE - 1;
gator_buffer_size[BLOCK_COUNTER_BUF] = BLOCK_COUNTER_BUFFER_SIZE;
gator_buffer_mask[BLOCK_COUNTER_BUF] = BLOCK_COUNTER_BUFFER_SIZE - 1;
gator_buffer_size[ANNOTATE_BUF] = ANNOTATE_BUFFER_SIZE;
gator_buffer_mask[ANNOTATE_BUF] = ANNOTATE_BUFFER_SIZE - 1;
gator_buffer_size[SCHED_TRACE_BUF] = SCHED_TRACE_BUFFER_SIZE;
gator_buffer_mask[SCHED_TRACE_BUF] = SCHED_TRACE_BUFFER_SIZE - 1;
gator_buffer_size[GPU_TRACE_BUF] = GPU_TRACE_BUFFER_SIZE;
gator_buffer_mask[GPU_TRACE_BUF] = GPU_TRACE_BUFFER_SIZE - 1;
gator_buffer_size[IDLE_BUF] = IDLE_BUFFER_SIZE;
gator_buffer_mask[IDLE_BUF] = IDLE_BUFFER_SIZE - 1;
// Initialize percpu per buffer variables
for (i = 0; i < NUM_GATOR_BUFS; i++) {
// Verify buffers are a power of 2
if (gator_buffer_size[i] & (gator_buffer_size[i] - 1)) {
err = -ENOEXEC;
goto setup_error;
}
for_each_present_cpu(cpu) {
per_cpu(gator_buffer_read, cpu)[i] = 0;
per_cpu(gator_buffer_write, cpu)[i] = 0;
per_cpu(gator_buffer_commit, cpu)[i] = 0;
per_cpu(buffer_space_available, cpu)[i] = true;
per_cpu(gator_buffer_commit_time, cpu) = gator_live_rate;
// Annotation is a special case that only uses a single buffer
if (cpu > 0 && i == ANNOTATE_BUF) {
per_cpu(gator_buffer, cpu)[i] = NULL;
continue;
}
per_cpu(gator_buffer, cpu)[i] = vmalloc(gator_buffer_size[i]);
if (!per_cpu(gator_buffer, cpu)[i]) {
err = -ENOMEM;
goto setup_error;
}
}
}
setup_error:
mutex_unlock(&start_mutex);
return err;
}
/* Actually start profiling (echo 1>/dev/gator/enable) */
static int gator_op_start(void)
{
int err = 0;
mutex_lock(&start_mutex);
if (gator_started || gator_start())
err = -EINVAL;
else
gator_started = 1;
mutex_unlock(&start_mutex);
return err;
}
/* echo 0>/dev/gator/enable */
static void gator_op_stop(void)
{
mutex_lock(&start_mutex);
if (gator_started) {
gator_stop();
mutex_lock(&gator_buffer_mutex);
gator_started = 0;
gator_monotonic_started = 0;
cookies_release();
wake_up(&gator_buffer_wait);
mutex_unlock(&gator_buffer_mutex);
}
mutex_unlock(&start_mutex);
}
static void gator_shutdown(void)
{
int cpu, i;
mutex_lock(&start_mutex);
for_each_present_cpu(cpu) {
mutex_lock(&gator_buffer_mutex);
for (i = 0; i < NUM_GATOR_BUFS; i++) {
vfree(per_cpu(gator_buffer, cpu)[i]);
per_cpu(gator_buffer, cpu)[i] = NULL;
per_cpu(gator_buffer_read, cpu)[i] = 0;
per_cpu(gator_buffer_write, cpu)[i] = 0;
per_cpu(gator_buffer_commit, cpu)[i] = 0;
per_cpu(buffer_space_available, cpu)[i] = true;
per_cpu(gator_buffer_commit_time, cpu) = 0;
}
mutex_unlock(&gator_buffer_mutex);
}
memset(&sent_core_name, 0, sizeof(sent_core_name));
mutex_unlock(&start_mutex);
}
static int gator_set_backtrace(unsigned long val)
{
int err = 0;
mutex_lock(&start_mutex);
if (gator_started)
err = -EBUSY;
else
gator_backtrace_depth = val;
mutex_unlock(&start_mutex);
return err;
}
static ssize_t enable_read(struct file *file, char __user *buf, size_t count, loff_t *offset)
{
return gatorfs_ulong_to_user(gator_started, buf, count, offset);
}
static ssize_t enable_write(struct file *file, char const __user *buf, size_t count, loff_t *offset)
{
unsigned long val;
int retval;
if (*offset)
return -EINVAL;
retval = gatorfs_ulong_from_user(&val, buf, count);
if (retval)
return retval;
if (val)
retval = gator_op_start();
else
gator_op_stop();
if (retval)
return retval;
return count;
}
static const struct file_operations enable_fops = {
.read = enable_read,
.write = enable_write,
};
static int userspace_buffer_open(struct inode *inode, struct file *file)
{
int err = -EPERM;
if (!capable(CAP_SYS_ADMIN))
return -EPERM;
if (test_and_set_bit_lock(0, &gator_buffer_opened))
return -EBUSY;
if ((err = gator_op_setup()))
goto fail;
/* NB: the actual start happens from userspace
* echo 1 >/dev/gator/enable
*/
return 0;
fail:
__clear_bit_unlock(0, &gator_buffer_opened);
return err;
}
static int userspace_buffer_release(struct inode *inode, struct file *file)
{
gator_op_stop();
gator_shutdown();
__clear_bit_unlock(0, &gator_buffer_opened);
return 0;
}
static ssize_t userspace_buffer_read(struct file *file, char __user *buf, size_t count, loff_t *offset)
{
int commit, length1, length2, read;
char *buffer1;
char *buffer2;
int cpu, buftype;
int written = 0;
// ensure there is enough space for a whole frame
if (count < userspace_buffer_size || *offset) {
return -EINVAL;
}
// sleep until the condition is true or a signal is received
// the condition is checked each time gator_buffer_wait is woken up
wait_event_interruptible(gator_buffer_wait, buffer_commit_ready(&cpu, &buftype) || !gator_started);
if (signal_pending(current)) {
return -EINTR;
}
if (buftype == -1 || cpu == -1) {
return 0;
}
mutex_lock(&gator_buffer_mutex);
do {
read = per_cpu(gator_buffer_read, cpu)[buftype];
commit = per_cpu(gator_buffer_commit, cpu)[buftype];
// May happen if the buffer is freed during pending reads.
if (!per_cpu(gator_buffer, cpu)[buftype]) {
break;
}
// determine the size of two halves
length1 = commit - read;
length2 = 0;
buffer1 = &(per_cpu(gator_buffer, cpu)[buftype][read]);
buffer2 = &(per_cpu(gator_buffer, cpu)[buftype][0]);
if (length1 < 0) {
length1 = gator_buffer_size[buftype] - read;
length2 = commit;
}
if (length1 + length2 > count - written) {
break;
}
// start, middle or end
if (length1 > 0 && copy_to_user(&buf[written], buffer1, length1)) {
break;
}
// possible wrap around
if (length2 > 0 && copy_to_user(&buf[written + length1], buffer2, length2)) {
break;
}
per_cpu(gator_buffer_read, cpu)[buftype] = commit;
written += length1 + length2;
// Wake up annotate_write if more space is available
if (buftype == ANNOTATE_BUF) {
wake_up(&gator_annotate_wait);
}
} while (buffer_commit_ready(&cpu, &buftype));
mutex_unlock(&gator_buffer_mutex);
// kick just in case we've lost an SMP event
wake_up(&gator_buffer_wait);
return written > 0 ? written : -EFAULT;
}
const struct file_operations gator_event_buffer_fops = {
.open = userspace_buffer_open,
.release = userspace_buffer_release,
.read = userspace_buffer_read,
};
static ssize_t depth_read(struct file *file, char __user *buf, size_t count, loff_t *offset)
{
return gatorfs_ulong_to_user(gator_backtrace_depth, buf, count, offset);
}
static ssize_t depth_write(struct file *file, char const __user *buf, size_t count, loff_t *offset)
{
unsigned long val;
int retval;
if (*offset)
return -EINVAL;
retval = gatorfs_ulong_from_user(&val, buf, count);
if (retval)
return retval;
retval = gator_set_backtrace(val);
if (retval)
return retval;
return count;
}
static const struct file_operations depth_fops = {
.read = depth_read,
.write = depth_write
};
void gator_op_create_files(struct super_block *sb, struct dentry *root)
{
struct dentry *dir;
struct gator_interface *gi;
int cpu;
/* reinitialize default values */
gator_cpu_cores = 0;
for_each_present_cpu(cpu) {
gator_cpu_cores++;
}
userspace_buffer_size = BACKTRACE_BUFFER_SIZE;
gator_response_type = 1;
gator_live_rate = 0;
gatorfs_create_file(sb, root, "enable", &enable_fops);
gatorfs_create_file(sb, root, "buffer", &gator_event_buffer_fops);
gatorfs_create_file(sb, root, "backtrace_depth", &depth_fops);
gatorfs_create_ro_ulong(sb, root, "cpu_cores", &gator_cpu_cores);
gatorfs_create_ro_ulong(sb, root, "buffer_size", &userspace_buffer_size);
gatorfs_create_ulong(sb, root, "tick", &gator_timer_count);
gatorfs_create_ulong(sb, root, "response_type", &gator_response_type);
gatorfs_create_ro_ulong(sb, root, "version", &gator_protocol_version);
gatorfs_create_ro_u64(sb, root, "started", &gator_monotonic_started);
gatorfs_create_u64(sb, root, "live_rate", &gator_live_rate);
// Annotate interface
gator_annotate_create_files(sb, root);
// Linux Events
dir = gatorfs_mkdir(sb, root, "events");
list_for_each_entry(gi, &gator_events, list)
if (gi->create_files)
gi->create_files(sb, dir);
// Sched Events
sched_trace_create_files(sb, dir);
// Power interface
gator_trace_power_create_files(sb, dir);
}
/******************************************************************************
* Module
******************************************************************************/
static int __init gator_module_init(void)
{
if (gatorfs_register()) {
return -1;
}
if (gator_init()) {
gatorfs_unregister();
return -1;
}
setup_timer(&gator_buffer_wake_up_timer, gator_buffer_wake_up, 0);
return 0;
}
static void __exit gator_module_exit(void)
{
del_timer_sync(&gator_buffer_wake_up_timer);
tracepoint_synchronize_unregister();
gator_exit();
gatorfs_unregister();
}
module_init(gator_module_init);
module_exit(gator_module_exit);
MODULE_LICENSE("GPL");
MODULE_AUTHOR("ARM Ltd");
MODULE_DESCRIPTION("Gator system profiler");