/**
* @file daemon/opd_sfile.c
* Management of sample files
*
* @remark Copyright 2002, 2005 OProfile authors
* @remark Read the file COPYING
*
* @author John Levon
* @author Philippe Elie
*/
#include "opd_sfile.h"
#include "opd_trans.h"
#include "opd_kernel.h"
#include "opd_mangling.h"
#include "opd_anon.h"
#include "opd_printf.h"
#include "opd_stats.h"
#include "oprofiled.h"
#include "op_libiberty.h"
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#define HASH_SIZE 2048
#define HASH_BITS (HASH_SIZE - 1)
/** All sfiles are hashed into these lists */
static struct list_head hashes[HASH_SIZE];
/** All sfiles are on this list. */
static LIST_HEAD(lru_list);
/* FIXME: can undoubtedly improve this hashing */
/** Hash the transient parameters for lookup. */
static unsigned long
sfile_hash(struct transient const * trans, struct kernel_image * ki)
{
unsigned long val = 0;
if (separate_thread) {
val ^= trans->tid << 2;
val ^= trans->tgid << 2;
}
if (separate_kernel || ((trans->anon || separate_lib) && !ki))
val ^= trans->app_cookie >> (DCOOKIE_SHIFT + 3);
if (separate_cpu)
val ^= trans->cpu;
/* cookie meaningless for kernel, shouldn't hash */
if (trans->in_kernel) {
val ^= ki->start >> 14;
val ^= ki->end >> 7;
return val & HASH_BITS;
}
if (trans->cookie != NO_COOKIE) {
val ^= trans->cookie >> DCOOKIE_SHIFT;
return val & HASH_BITS;
}
if (!separate_thread)
val ^= trans->tgid << 2;
if (trans->anon) {
val ^= trans->anon->start >> VMA_SHIFT;
val ^= trans->anon->end >> (VMA_SHIFT + 1);
}
return val & HASH_BITS;
}
static int
do_match(struct sfile const * sf, cookie_t cookie, cookie_t app_cookie,
struct kernel_image const * ki, struct anon_mapping const * anon,
pid_t tgid, pid_t tid, unsigned int cpu)
{
/* this is a simplified check for "is a kernel image" AND
* "is the right kernel image". Also handles no-vmlinux
* correctly.
*/
if (sf->kernel != ki)
return 0;
if (separate_thread) {
if (sf->tid != tid || sf->tgid != tgid)
return 0;
}
if (separate_cpu) {
if (sf->cpu != cpu)
return 0;
}
if (separate_kernel || ((anon || separate_lib) && !ki)) {
if (sf->app_cookie != app_cookie)
return 0;
}
/* ignore the cached trans->cookie for kernel images,
* it's meaningless and we checked all others already
*/
if (ki)
return 1;
if (sf->anon != anon)
return 0;
return sf->cookie == cookie;
}
static int
trans_match(struct transient const * trans, struct sfile const * sfile,
struct kernel_image const * ki)
{
return do_match(sfile, trans->cookie, trans->app_cookie, ki,
trans->anon, trans->tgid, trans->tid, trans->cpu);
}
static int
sfile_equal(struct sfile const * sf, struct sfile const * sf2)
{
return do_match(sf, sf2->cookie, sf2->app_cookie, sf2->kernel,
sf2->anon, sf2->tgid, sf2->tid, sf2->cpu);
}
static int
is_sf_ignored(struct sfile const * sf)
{
if (sf->kernel) {
if (!is_image_ignored(sf->kernel->name))
return 0;
/* Let a dependent kernel image redeem the sf if we're
* executing on behalf of an application.
*/
return is_cookie_ignored(sf->app_cookie);
}
/* Anon regions are always dependent on the application.
* Otherwise, let a dependent image redeem the sf.
*/
if (sf->anon || is_cookie_ignored(sf->cookie))
return is_cookie_ignored(sf->app_cookie);
return 0;
}
/** create a new sfile matching the current transient parameters */
static struct sfile *
create_sfile(unsigned long hash, struct transient const * trans,
struct kernel_image * ki)
{
size_t i;
struct sfile * sf;
sf = xmalloc(sizeof(struct sfile));
sf->hashval = hash;
/* The logic here: if we're in the kernel, the cached cookie is
* meaningless (though not the app_cookie if separate_kernel)
*/
sf->cookie = trans->in_kernel ? INVALID_COOKIE : trans->cookie;
sf->app_cookie = INVALID_COOKIE;
sf->tid = (pid_t)-1;
sf->tgid = (pid_t)-1;
sf->cpu = 0;
sf->kernel = ki;
sf->anon = trans->anon;
for (i = 0 ; i < op_nr_counters ; ++i)
odb_init(&sf->files[i]);
for (i = 0; i < CG_HASH_SIZE; ++i)
list_init(&sf->cg_hash[i]);
if (separate_thread)
sf->tid = trans->tid;
if (separate_thread || trans->cookie == NO_COOKIE)
sf->tgid = trans->tgid;
if (separate_cpu)
sf->cpu = trans->cpu;
if (separate_kernel || ((trans->anon || separate_lib) && !ki))
sf->app_cookie = trans->app_cookie;
sf->ignored = is_sf_ignored(sf);
sf->embedded_offset = trans->embedded_offset;
/* If embedded_offset is a valid value, it means we're
* processing a Cell BE SPU profile; in which case, we
* want sf->app_cookie to hold trans->app_cookie.
*/
if (trans->embedded_offset != UNUSED_EMBEDDED_OFFSET)
sf->app_cookie = trans->app_cookie;
return sf;
}
struct sfile * sfile_find(struct transient const * trans)
{
struct sfile * sf;
struct list_head * pos;
struct kernel_image * ki = NULL;
unsigned long hash;
if (trans->tracing != TRACING_ON) {
opd_stats[OPD_SAMPLES]++;
opd_stats[trans->in_kernel == 1 ? OPD_KERNEL : OPD_PROCESS]++;
}
/* There is a small race where this *can* happen, see
* caller of cpu_buffer_reset() in the kernel
*/
if (trans->in_kernel == -1) {
verbprintf(vsamples, "Losing sample at 0x%llx of unknown provenance.\n",
trans->pc);
opd_stats[OPD_NO_CTX]++;
return NULL;
}
/* we might need a kernel image start/end to hash on */
if (trans->in_kernel) {
ki = find_kernel_image(trans);
if (!ki) {
verbprintf(vsamples, "Lost kernel sample %llx\n", trans->pc);
opd_stats[OPD_LOST_KERNEL]++;
return NULL;
}
} else if (trans->cookie == NO_COOKIE && !trans->anon) {
if (vsamples) {
char const * app = verbose_cookie(trans->app_cookie);
printf("No anon map for pc %llx, app %s.\n",
trans->pc, app);
}
opd_stats[OPD_LOST_NO_MAPPING]++;
return NULL;
}
hash = sfile_hash(trans, ki);
list_for_each(pos, &hashes[hash]) {
sf = list_entry(pos, struct sfile, hash);
if (trans_match(trans, sf, ki)) {
sfile_get(sf);
goto lru;
}
}
sf = create_sfile(hash, trans, ki);
list_add(&sf->hash, &hashes[hash]);
lru:
sfile_put(sf);
return sf;
}
static void sfile_dup(struct sfile * to, struct sfile * from)
{
size_t i;
memcpy(to, from, sizeof (struct sfile));
for (i = 0 ; i < op_nr_counters ; ++i)
odb_init(&to->files[i]);
for (i = 0; i < CG_HASH_SIZE; ++i)
list_init(&to->cg_hash[i]);
list_init(&to->hash);
list_init(&to->lru);
}
static odb_t * get_file(struct transient const * trans, int is_cg)
{
struct sfile * sf = trans->current;
struct sfile * last = trans->last;
struct cg_entry * cg;
struct list_head * pos;
unsigned long hash;
odb_t * file;
if (trans->event >= op_nr_counters) {
fprintf(stderr, "%s: Invalid counter %lu\n", __FUNCTION__,
trans->event);
abort();
}
file = &sf->files[trans->event];
if (!is_cg)
goto open;
hash = last->hashval & (CG_HASH_SIZE - 1);
/* Need to look for the right 'to'. Since we're looking for
* 'last', we use its hash.
*/
list_for_each(pos, &sf->cg_hash[hash]) {
cg = list_entry(pos, struct cg_entry, hash);
if (sfile_equal(last, &cg->to)) {
file = &cg->to.files[trans->event];
goto open;
}
}
cg = xmalloc(sizeof(struct cg_entry));
sfile_dup(&cg->to, last);
list_add(&cg->hash, &sf->cg_hash[hash]);
file = &cg->to.files[trans->event];
open:
if (!odb_open_count(file))
opd_open_sample_file(file, last, sf, trans->event, is_cg);
/* Error is logged by opd_open_sample_file */
if (!odb_open_count(file))
return NULL;
return file;
}
static void verbose_print_sample(struct sfile * sf, vma_t pc, uint counter)
{
char const * app = verbose_cookie(sf->app_cookie);
printf("0x%llx(%u): ", pc, counter);
if (sf->anon) {
printf("anon (tgid %u, 0x%llx-0x%llx), ",
(unsigned int)sf->anon->tgid,
sf->anon->start, sf->anon->end);
} else if (sf->kernel) {
printf("kern (name %s, 0x%llx-0x%llx), ", sf->kernel->name,
sf->kernel->start, sf->kernel->end);
} else {
printf("%s(%llx), ", verbose_cookie(sf->cookie), sf->cookie);
}
printf("app %s(%llx)", app, sf->app_cookie);
}
static void verbose_sample(struct transient const * trans, vma_t pc)
{
printf("Sample ");
verbose_print_sample(trans->current, pc, trans->event);
printf("\n");
}
static void
verbose_arc(struct transient const * trans, vma_t from, vma_t to)
{
printf("Arc ");
verbose_print_sample(trans->current, from, trans->event);
printf(" -> 0x%llx", to);
printf("\n");
}
static void sfile_log_arc(struct transient const * trans)
{
int err;
vma_t from = trans->pc;
vma_t to = trans->last_pc;
uint64_t key;
odb_t * file;
file = get_file(trans, 1);
/* absolute value -> offset */
if (trans->current->kernel)
from -= trans->current->kernel->start;
if (trans->last->kernel)
to -= trans->last->kernel->start;
if (trans->current->anon)
from -= trans->current->anon->start;
if (trans->last->anon)
to -= trans->last->anon->start;
if (varcs)
verbose_arc(trans, from, to);
if (!file) {
opd_stats[OPD_LOST_SAMPLEFILE]++;
return;
}
/* Possible narrowings to 32-bit value only. */
key = to & (0xffffffff);
key |= ((uint64_t)from) << 32;
err = odb_update_node(file, key);
if (err) {
fprintf(stderr, "%s: %s\n", __FUNCTION__, strerror(err));
abort();
}
}
void sfile_log_sample(struct transient const * trans)
{
int err;
vma_t pc = trans->pc;
odb_t * file;
if (trans->tracing == TRACING_ON) {
/* can happen if kernel sample falls through the cracks,
* see opd_put_sample() */
if (trans->last)
sfile_log_arc(trans);
return;
}
file = get_file(trans, 0);
/* absolute value -> offset */
if (trans->current->kernel)
pc -= trans->current->kernel->start;
if (trans->current->anon)
pc -= trans->current->anon->start;
if (vsamples)
verbose_sample(trans, pc);
if (!file) {
opd_stats[OPD_LOST_SAMPLEFILE]++;
return;
}
err = odb_update_node(file, (uint64_t)pc);
if (err) {
fprintf(stderr, "%s: %s\n", __FUNCTION__, strerror(err));
abort();
}
}
static int close_sfile(struct sfile * sf, void * data __attribute__((unused)))
{
size_t i;
/* it's OK to close a non-open odb file */
for (i = 0; i < op_nr_counters; ++i)
odb_close(&sf->files[i]);
return 0;
}
static void kill_sfile(struct sfile * sf)
{
close_sfile(sf, NULL);
list_del(&sf->hash);
list_del(&sf->lru);
}
static int sync_sfile(struct sfile * sf, void * data __attribute__((unused)))
{
size_t i;
for (i = 0; i < op_nr_counters; ++i)
odb_sync(&sf->files[i]);
return 0;
}
static int is_sfile_kernel(struct sfile * sf, void * data __attribute__((unused)))
{
return !!sf->kernel;
}
static int is_sfile_anon(struct sfile * sf, void * data)
{
return sf->anon == data;
}
typedef int (*sfile_func)(struct sfile *, void *);
static void
for_one_sfile(struct sfile * sf, sfile_func func, void * data)
{
size_t i;
int free_sf = func(sf, data);
for (i = 0; i < CG_HASH_SIZE; ++i) {
struct list_head * pos;
struct list_head * pos2;
list_for_each_safe(pos, pos2, &sf->cg_hash[i]) {
struct cg_entry * cg =
list_entry(pos, struct cg_entry, hash);
if (free_sf || func(&cg->to, data)) {
kill_sfile(&cg->to);
list_del(&cg->hash);
free(cg);
}
}
}
if (free_sf) {
kill_sfile(sf);
free(sf);
}
}
static void for_each_sfile(sfile_func func, void * data)
{
struct list_head * pos;
struct list_head * pos2;
list_for_each_safe(pos, pos2, &lru_list) {
struct sfile * sf = list_entry(pos, struct sfile, lru);
for_one_sfile(sf, func, data);
}
}
void sfile_clear_kernel(void)
{
for_each_sfile(is_sfile_kernel, NULL);
}
void sfile_clear_anon(struct anon_mapping * anon)
{
for_each_sfile(is_sfile_anon, anon);
}
void sfile_sync_files(void)
{
for_each_sfile(sync_sfile, NULL);
}
void sfile_close_files(void)
{
for_each_sfile(close_sfile, NULL);
}
static int always_true(void)
{
return 1;
}
#define LRU_AMOUNT 256
/*
* Clear out older sfiles. Note the current sfiles we're using
* will not be present in this list, due to sfile_get/put() pairs
* around the caller of this.
*/
int sfile_lru_clear(void)
{
struct list_head * pos;
struct list_head * pos2;
int amount = LRU_AMOUNT;
if (list_empty(&lru_list))
return 1;
list_for_each_safe(pos, pos2, &lru_list) {
struct sfile * sf;
if (!--amount)
break;
sf = list_entry(pos, struct sfile, lru);
for_one_sfile(sf, (sfile_func)always_true, NULL);
}
return 0;
}
void sfile_get(struct sfile * sf)
{
if (sf)
list_del(&sf->lru);
}
void sfile_put(struct sfile * sf)
{
if (sf)
list_add_tail(&sf->lru, &lru_list);
}
void sfile_init(void)
{
size_t i = 0;
for (; i < HASH_SIZE; ++i)
list_init(&hashes[i]);
}