4d4036e0e7
The number of hardware counters is limited. The multiplexing feature enables OProfile to gather more events than counters are provided by the hardware. This is realized by switching between events at an user specified time interval. A new file (/dev/oprofile/time_slice) is added for the user to specify the timer interval in ms. If the number of events to profile is higher than the number of hardware counters available, the patch will schedule a work queue that switches the event counter and re-writes the different sets of values into it. The switching mechanism needs to be implemented for each architecture to support multiplexing. This patch only implements AMD CPU support, but multiplexing can be easily extended for other models and architectures. There are follow-on patches that rework parts of this patch. Signed-off-by: Jason Yeh <jason.yeh@amd.com> Signed-off-by: Robert Richter <robert.richter@amd.com>
89 lines
2.4 KiB
C
89 lines
2.4 KiB
C
/**
|
|
* @file oprofile_stats.c
|
|
*
|
|
* @remark Copyright 2002 OProfile authors
|
|
* @remark Read the file COPYING
|
|
*
|
|
* @author John Levon
|
|
*/
|
|
|
|
#include <linux/oprofile.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/threads.h>
|
|
|
|
#include "oprofile_stats.h"
|
|
#include "cpu_buffer.h"
|
|
|
|
struct oprofile_stat_struct oprofile_stats;
|
|
#ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX
|
|
atomic_t multiplex_counter;
|
|
#endif
|
|
|
|
void oprofile_reset_stats(void)
|
|
{
|
|
struct oprofile_cpu_buffer *cpu_buf;
|
|
int i;
|
|
|
|
for_each_possible_cpu(i) {
|
|
cpu_buf = &per_cpu(cpu_buffer, i);
|
|
cpu_buf->sample_received = 0;
|
|
cpu_buf->sample_lost_overflow = 0;
|
|
cpu_buf->backtrace_aborted = 0;
|
|
cpu_buf->sample_invalid_eip = 0;
|
|
}
|
|
|
|
atomic_set(&oprofile_stats.sample_lost_no_mm, 0);
|
|
atomic_set(&oprofile_stats.sample_lost_no_mapping, 0);
|
|
atomic_set(&oprofile_stats.event_lost_overflow, 0);
|
|
atomic_set(&oprofile_stats.bt_lost_no_mapping, 0);
|
|
#ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX
|
|
atomic_set(&multiplex_counter, 0);
|
|
#endif
|
|
}
|
|
|
|
|
|
void oprofile_create_stats_files(struct super_block *sb, struct dentry *root)
|
|
{
|
|
struct oprofile_cpu_buffer *cpu_buf;
|
|
struct dentry *cpudir;
|
|
struct dentry *dir;
|
|
char buf[10];
|
|
int i;
|
|
|
|
dir = oprofilefs_mkdir(sb, root, "stats");
|
|
if (!dir)
|
|
return;
|
|
|
|
for_each_possible_cpu(i) {
|
|
cpu_buf = &per_cpu(cpu_buffer, i);
|
|
snprintf(buf, 10, "cpu%d", i);
|
|
cpudir = oprofilefs_mkdir(sb, dir, buf);
|
|
|
|
/* Strictly speaking access to these ulongs is racy,
|
|
* but we can't simply lock them, and they are
|
|
* informational only.
|
|
*/
|
|
oprofilefs_create_ro_ulong(sb, cpudir, "sample_received",
|
|
&cpu_buf->sample_received);
|
|
oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_overflow",
|
|
&cpu_buf->sample_lost_overflow);
|
|
oprofilefs_create_ro_ulong(sb, cpudir, "backtrace_aborted",
|
|
&cpu_buf->backtrace_aborted);
|
|
oprofilefs_create_ro_ulong(sb, cpudir, "sample_invalid_eip",
|
|
&cpu_buf->sample_invalid_eip);
|
|
}
|
|
|
|
oprofilefs_create_ro_atomic(sb, dir, "sample_lost_no_mm",
|
|
&oprofile_stats.sample_lost_no_mm);
|
|
oprofilefs_create_ro_atomic(sb, dir, "sample_lost_no_mapping",
|
|
&oprofile_stats.sample_lost_no_mapping);
|
|
oprofilefs_create_ro_atomic(sb, dir, "event_lost_overflow",
|
|
&oprofile_stats.event_lost_overflow);
|
|
oprofilefs_create_ro_atomic(sb, dir, "bt_lost_no_mapping",
|
|
&oprofile_stats.bt_lost_no_mapping);
|
|
#ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX
|
|
oprofilefs_create_ro_atomic(sb, dir, "multiplex_counter",
|
|
&multiplex_counter);
|
|
#endif
|
|
}
|