/*--------------------------------------------------------------------*/ /*--- Callgrind ---*/ /*--- global.h ---*/ /*--- (C) 2004, 2005 Josef Weidendorfer ---*/ /*--------------------------------------------------------------------*/ #ifndef CLG_GLOBAL #define CLG_GLOBAL #include "pub_tool_basics.h" #include "pub_tool_vki.h" #include "pub_tool_debuginfo.h" #include "pub_tool_libcbase.h" #include "pub_tool_libcassert.h" #include "pub_tool_libcfile.h" #include "pub_tool_libcprint.h" #include "pub_tool_libcproc.h" #include "pub_tool_machine.h" #include "pub_tool_mallocfree.h" #include "pub_tool_options.h" #include "pub_tool_tooliface.h" #include "pub_tool_xarray.h" #include "pub_tool_clientstate.h" #include "pub_tool_machine.h" // VG_(fnptr_to_fnentry) #include "events.h" // defines CLG_ macro #include "costs.h" /*------------------------------------------------------------*/ /*--- Calltree compile options --- */ /*------------------------------------------------------------*/ /* Enable debug output */ #define CLG_ENABLE_DEBUG 1 /* Enable experimental features? */ #define CLG_EXPERIMENTAL 0 /* Syscall Timing in microseconds? * (define to 0 if you get compile errors) */ #define CLG_MICROSYSTIME 0 /* Set to 1 if you want full sanity checks for JCC */ #define JCC_CHECK 0 /*------------------------------------------------------------*/ /*--- Command line options ---*/ /*------------------------------------------------------------*/ #define DEFAULT_OUTFORMAT "callgrind.out.%p" typedef struct _CommandLineOptions CommandLineOptions; struct _CommandLineOptions { /* Dump format options */ Char* out_format; /* Format string for callgrind output file name */ Bool combine_dumps; /* Dump trace parts into same file? */ Bool compress_strings; Bool compress_events; Bool compress_pos; Bool mangle_names; Bool compress_mangled; Bool dump_line; Bool dump_instr; Bool dump_bb; Bool dump_bbs; /* Dump basic block information? */ /* Dump generation options */ ULong dump_every_bb; /* Dump every xxx BBs. */ /* Collection options */ Bool separate_threads; /* Separate threads in dump? */ Int separate_callers; /* Separate dependent on how many callers? */ Int separate_recursions; /* Max level of recursions to separate */ Bool skip_plt; /* Skip functions in PLT section? */ Bool skip_direct_recursion; /* Increment direct recursions the level? */ Bool collect_atstart; /* Start in collecting state ? */ Bool collect_jumps; /* Collect (cond.) jumps in functions ? */ Bool collect_alloc; /* Collect size of allocated memory */ Bool collect_systime; /* Collect time for system calls */ Bool collect_bus; /* Collect global bus events */ /* Instrument options */ Bool instrument_atstart; /* Instrument at start? */ Bool simulate_cache; /* Call into cache simulator ? */ Bool simulate_branch; /* Call into branch prediction simulator ? */ /* Call graph generation */ Bool pop_on_jump; /* Handle a jump between functions as ret+call */ #if CLG_ENABLE_DEBUG Int verbose; ULong verbose_start; #endif }; /*------------------------------------------------------------*/ /*--- Constants ---*/ /*------------------------------------------------------------*/ /* According to IA-32 Intel Architecture Software Developer's Manual: Vol 2 */ #define MAX_x86_INSTR_SIZE 16 /* Minimum cache line size allowed */ #define MIN_LINE_SIZE 16 /* Size of various buffers used for storing strings */ #define FILENAME_LEN VKI_PATH_MAX #define FN_NAME_LEN 4096 /* for C++ code :-) */ #define OBJ_NAME_LEN 256 #define COSTS_LEN 512 /* at least 17x 64bit values */ #define BUF_LEN 512 #define COMMIFY_BUF_LEN 128 #define RESULTS_BUF_LEN 256 #define LINE_BUF_LEN 64 /* Convenience macros */ /* Use this only when size of sprintf args are known to fit into * given buffer; for strings of unknown length, use WRITE_STR below */ #define WRITE_SPRINTF(fd, zz_buf, fmt, args...) \ do { Int len = VG_(sprintf)(zz_buf, fmt, ## args); \ VG_(write)(fd, (void*)zz_buf, len); \ } while (0) #define WRITE_STR(fd, str) \ do { if (str) { Int len = VG_(strlen)(str); \ VG_(write)(fd, (void*)str, len); } \ else VG_(write)(fd, "(null)", 6); \ } while (0) #define WRITE_STR2(fd, str1, str2) \ do { if (str1) { Int len = VG_(strlen)(str1); \ VG_(write)(fd, (void*)str1, len); } \ else VG_(write)(fd, "(null)", 6); \ if (str2) { Int len = VG_(strlen)(str2); \ VG_(write)(fd, (void*)str2, len); } \ else VG_(write)(fd, "(null)", 6); \ } while (0) #define WRITE_STR3(fd, str1, str2, str3) \ do { if (str1) { Int len = VG_(strlen)(str1); \ VG_(write)(fd, (void*)str1, len); } \ else VG_(write)(fd, "(null)", 6); \ if (str2) { Int len = VG_(strlen)(str2); \ VG_(write)(fd, (void*)str2, len); } \ else VG_(write)(fd, "(null)", 6); \ if (str3) { Int len = VG_(strlen)(str3); \ VG_(write)(fd, (void*)str3, len); } \ else VG_(write)(fd, "(null)", 6); \ } while (0) /*------------------------------------------------------------*/ /*--- Statistics ---*/ /*------------------------------------------------------------*/ typedef struct _Statistics Statistics; struct _Statistics { ULong call_counter; ULong jcnd_counter; ULong jump_counter; ULong rec_call_counter; ULong ret_counter; ULong bb_executions; Int context_counter; Int bb_retranslations; Int distinct_objs; Int distinct_files; Int distinct_fns; Int distinct_contexts; Int distinct_bbs; Int distinct_jccs; Int distinct_bbccs; Int distinct_instrs; Int distinct_skips; Int bb_hash_resizes; Int bbcc_hash_resizes; Int jcc_hash_resizes; Int cxt_hash_resizes; Int fn_array_resizes; Int call_stack_resizes; Int fn_stack_resizes; Int full_debug_BBs; Int file_line_debug_BBs; Int fn_name_debug_BBs; Int no_debug_BBs; Int bbcc_lru_misses; Int jcc_lru_misses; Int cxt_lru_misses; Int bbcc_clones; }; /*------------------------------------------------------------*/ /*--- Structure declarations ---*/ /*------------------------------------------------------------*/ typedef struct _Context Context; typedef struct _CC CC; typedef struct _BB BB; typedef struct _Skipped Skipped; typedef struct _BBCC BBCC; typedef struct _jCC jCC; typedef struct _fCC fCC; typedef struct _fn_node fn_node; typedef struct _file_node file_node; typedef struct _obj_node obj_node; typedef struct _fn_config fn_config; typedef struct _call_entry call_entry; typedef struct _thread_info thread_info; /* Costs of event sets. Aliases to arrays of 64-bit values */ typedef ULong* SimCost; /* All events the simulator can produce */ typedef ULong* UserCost; typedef ULong* FullCost; /* Simulator + User */ /* The types of control flow changes that can happen between * execution of two BBs in a thread. */ typedef enum { jk_None = 0, /* no explicit change by a guest instruction */ jk_Jump, /* regular jump */ jk_Call, jk_Return, jk_CondJump /* conditional jump taken (only used as jCC type) */ } ClgJumpKind; /* JmpCall cost center * for subroutine call (from->bb->jmp_addr => to->bb->addr) * * Each BB has at most one CALL instruction. The list of JCC from * this call is a pointer to the list head (stored in BBCC), and * <next_from> in the JCC struct. * * For fast lookup, JCCs are reachable with a hash table, keyed by * the (from_bbcc,to) pair. <next_hash> is used for the JCC chain * of one hash table entry. * * Cost <sum> holds event counts for already returned executions. * <last> are the event counters at last enter of the subroutine. * <sum> is updated on returning from the subroutine by * adding the diff of <last> and current event counters to <sum>. * * After updating, <last> is set to current event counters. Thus, * events are not counted twice for recursive calls (TODO: True?) */ struct _jCC { ClgJumpKind jmpkind; /* jk_Call, jk_Jump, jk_CondJump */ jCC* next_hash; /* for hash entry chain */ jCC* next_from; /* next JCC from a BBCC */ BBCC *from, *to; /* call arc from/to this BBCC */ UInt jmp; /* jump no. in source */ ULong call_counter; /* no wraparound with 64 bit */ FullCost cost; /* simulator + user counters */ }; /* * Info for one instruction of a basic block. */ typedef struct _InstrInfo InstrInfo; struct _InstrInfo { UInt instr_offset; UInt instr_size; UInt cost_offset; EventSet* eventset; }; /* * Info for a side exit in a BB */ typedef struct _CJmpInfo CJmpInfo; struct _CJmpInfo { UInt instr; /* instruction index for BB.instr array */ ClgJumpKind jmpkind; /* jump kind when leaving BB at this side exit */ }; /** * An instrumented basic block (BB). * * BBs are put into a resizable hash to allow for fast detection if a * BB is to be retranslated but cost info is already available. * The key for a BB is a (object, offset) tupel making it independent * from possibly multiple mappings of the same ELF object. * * At the beginning of each instrumented BB, * a call to setup_bbcc(), specifying a pointer to the * according BB structure, is added. * * As cost of a BB has to be distinguished depending on the context, * multiple cost centers for one BB (struct BBCC) exist and the according * BBCC is set by setup_bbcc. */ struct _BB { obj_node* obj; /* ELF object of BB */ PtrdiffT offset; /* offset of BB in ELF object file */ BB* next; /* chaining for a hash entry */ VgSectKind sect_kind; /* section of this BB, e.g. PLT */ UInt instr_count; /* filled by CLG_(get_fn_node) if debug info is available */ fn_node* fn; /* debug info for this BB */ UInt line; Bool is_entry; /* True if this BB is a function entry */ BBCC* bbcc_list; /* BBCCs for same BB (see next_bbcc in BBCC) */ BBCC* last_bbcc; /* Temporary: Cached for faster access (LRU) */ /* filled by CLG_(instrument) if not seen before */ UInt cjmp_count; /* number of side exits */ CJmpInfo* jmp; /* array of info for condition jumps, * allocated directly after this struct */ Bool cjmp_inverted; /* is last side exit actually fall through? */ UInt instr_len; UInt cost_count; InstrInfo instr[0]; /* info on instruction sizes and costs */ }; /** * Function context * * Basic blocks are always executed in the scope of a context. * A function context is a list of function nodes representing * the call chain to the current context: I.e. fn[0] is the * function we are currently in, fn[1] has called fn[0], and so on. * Recursion levels are used for fn[0]. * * To get a unique number for a full execution context, use * rec_index = min(<fn->rec_separation>,<active>) - 1; * unique_no = <number> + rec_index * * For each Context, recursion index and BB, there can be a BBCC. */ struct _Context { UInt size; // number of function dependencies UInt base_number; // for context compression & dump array Context* next; // entry chaining for hash UWord hash; // for faster lookup... fn_node* fn[0]; }; /* * Cost info for a side exits from a BB */ typedef struct _JmpData JmpData; struct _JmpData { ULong ecounter; /* number of times the BB was left at this exit */ jCC* jcc_list; /* JCCs used for this exit */ }; /* * Basic Block Cost Center * * On demand, multiple BBCCs will be created for the same BB * dependend on command line options and: * - current function (it's possible that a BB is executed in the * context of different functions, e.g. in manual assembler/PLT) * - current thread ID * - position where current function is called from * - recursion level of current function * * The cost centres for the instructions of a basic block are * stored in a contiguous array. * They are distinguishable by their tag field. */ struct _BBCC { BB* bb; /* BB for this cost center */ Context* cxt; /* execution context of this BBCC */ ThreadId tid; /* only for assertion check purpose */ UInt rec_index; /* Recursion index in rec->bbcc for this bbcc */ BBCC** rec_array; /* Variable sized array of pointers to * recursion BBCCs. Shared. */ ULong ret_counter; /* how often returned from jccs of this bbcc; * used to check if a dump for this BBCC is needed */ BBCC* next_bbcc; /* Chain of BBCCs for same BB */ BBCC* lru_next_bbcc; /* BBCC executed next the last time */ jCC* lru_from_jcc; /* Temporary: Cached for faster access (LRU) */ jCC* lru_to_jcc; /* Temporary: Cached for faster access (LRU) */ FullCost skipped; /* cost for skipped functions called from * jmp_addr. Allocated lazy */ BBCC* next; /* entry chain in hash */ ULong* cost; /* start of 64bit costs for this BBCC */ ULong ecounter_sum; /* execution counter for first instruction of BB */ JmpData jmp[0]; }; /* the <number> of fn_node, file_node and obj_node are for compressed dumping * and a index into the dump boolean table and fn_info_table */ struct _fn_node { Char* name; UInt number; Context* last_cxt; /* LRU info */ Context* pure_cxt; /* the context with only the function itself */ file_node* file; /* reverse mapping for 2nd hash */ fn_node* next; Bool dump_before :1; Bool dump_after :1; Bool zero_before :1; Bool toggle_collect :1; Bool skip :1; Bool pop_on_jump : 1; Bool is_malloc :1; Bool is_realloc :1; Bool is_free :1; Int group; Int separate_callers; Int separate_recursions; #if CLG_ENABLE_DEBUG Int verbosity; /* Stores old verbosity level while in function */ #endif }; /* Quite arbitrary fixed hash sizes */ #define N_OBJ_ENTRIES 47 #define N_FILE_ENTRIES 53 #define N_FN_ENTRIES 87 #define N_BBCC2_ENTRIES 37 struct _file_node { Char* name; fn_node* fns[N_FN_ENTRIES]; UInt number; obj_node* obj; file_node* next; }; /* If an object is dlopened multiple times, we hope that <name> is unique; * <start> and <offset> can change with each dlopen, and <start> is * zero when object is unmapped (possible at dump time). */ struct _obj_node { Char* name; UInt last_slash_pos; Addr start; /* Start address of text segment mapping */ SizeT size; /* Length of mapping */ PtrdiffT offset; /* Offset between symbol address and file offset */ file_node* files[N_FILE_ENTRIES]; UInt number; obj_node* next; }; /* an entry in the callstack * * <nonskipped> is 0 if the function called is not skipped (usual case). * Otherwise, it is the last non-skipped BBCC. This one gets all * the calls to non-skipped functions and all costs in skipped * instructions. */ struct _call_entry { jCC* jcc; /* jCC for this call */ FullCost enter_cost; /* cost event counters at entering frame */ Addr sp; /* stack pointer directly after call */ Addr ret_addr; /* address to which to return to * is 0 on a simulated call */ BBCC* nonskipped; /* see above */ Context* cxt; /* context before call */ Int fn_sp; /* function stack index before call */ }; /* * Execution state of main thread or a running signal handler in * a thread while interrupted by another signal handler. * As there's no scheduling among running signal handlers of one thread, * we only need a subset of a full thread state: * - event counter * - collect state * - last BB, last jump kind, last nonskipped BB * - callstack pointer for sanity checking and correct unwinding * after exit */ typedef struct _exec_state exec_state; struct _exec_state { /* the signum of the handler, 0 for main thread context */ Int sig; /* the old call stack pointer at entering the signal handler */ Int orig_sp; FullCost cost; Bool collect; Context* cxt; Int jmps_passed; /* number of conditional jumps passed in last BB */ BBCC* bbcc; /* last BB executed */ BBCC* nonskipped; Int call_stack_bottom; /* Index into fn_stack */ }; /* Global state structures */ typedef struct _bb_hash bb_hash; struct _bb_hash { UInt size, entries; BB** table; }; typedef struct _cxt_hash cxt_hash; struct _cxt_hash { UInt size, entries; Context** table; }; /* Thread specific state structures, i.e. parts of a thread state. * There are variables for the current state of each part, * on which a thread state is copied at thread switch. */ typedef struct _bbcc_hash bbcc_hash; struct _bbcc_hash { UInt size, entries; BBCC** table; }; typedef struct _jcc_hash jcc_hash; struct _jcc_hash { UInt size, entries; jCC** table; jCC* spontaneous; }; typedef struct _fn_array fn_array; struct _fn_array { UInt size; UInt* array; }; typedef struct _call_stack call_stack; struct _call_stack { UInt size; Int sp; call_entry* entry; }; typedef struct _fn_stack fn_stack; struct _fn_stack { UInt size; fn_node **bottom, **top; }; /* The maximum number of simultaneous running signal handlers per thread. * This is the number of execution states storable in a thread. */ #define MAX_SIGHANDLERS 10 typedef struct _exec_stack exec_stack; struct _exec_stack { Int sp; /* > 0 if a handler is running */ exec_state* entry[MAX_SIGHANDLERS]; }; /* Thread State * * This structure stores thread specific info while a thread is *not* * running. See function switch_thread() for save/restore on thread switch. * * If --separate-threads=no, BBCCs and JCCs can be shared by all threads, i.e. * only structures of thread 1 are used. * This involves variables fn_info_table, bbcc_table and jcc_table. */ struct _thread_info { /* state */ fn_stack fns; /* function stack */ call_stack calls; /* context call arc stack */ exec_stack states; /* execution states interrupted by signals */ /* dump statistics */ FullCost lastdump_cost; /* Cost at last dump */ FullCost sighandler_cost; /* thread specific data structure containers */ fn_array fn_active; jcc_hash jccs; bbcc_hash bbccs; }; /* Structs used for dumping */ /* Address position inside of a BBCC: * This includes * - the address offset from the BB start address * - file/line from debug info for that address (can change inside a BB) */ typedef struct _AddrPos AddrPos; struct _AddrPos { Addr addr; Addr bb_addr; file_node* file; UInt line; }; /* a simulator cost entity that can be written out in one line */ typedef struct _AddrCost AddrCost; struct _AddrCost { AddrPos p; SimCost cost; }; /* A function in an execution context */ typedef struct _FnPos FnPos; struct _FnPos { file_node* file; fn_node* fn; obj_node* obj; Context* cxt; int rec_index; UInt line; }; /*------------------------------------------------------------*/ /*--- Cache simulator interface ---*/ /*------------------------------------------------------------*/ struct cachesim_if { void (*print_opts)(void); Bool (*parse_opt)(Char* arg); void (*post_clo_init)(void); void (*clear)(void); void (*getdesc)(Char* buf); void (*printstat)(Int,Int,Int); void (*add_icost)(SimCost, BBCC*, InstrInfo*, ULong); void (*finish)(void); void (*log_1I0D)(InstrInfo*) VG_REGPARM(1); void (*log_2I0D)(InstrInfo*, InstrInfo*) VG_REGPARM(2); void (*log_3I0D)(InstrInfo*, InstrInfo*, InstrInfo*) VG_REGPARM(3); void (*log_1I1Dr)(InstrInfo*, Addr, Word) VG_REGPARM(3); void (*log_1I1Dw)(InstrInfo*, Addr, Word) VG_REGPARM(3); void (*log_0I1Dr)(InstrInfo*, Addr, Word) VG_REGPARM(3); void (*log_0I1Dw)(InstrInfo*, Addr, Word) VG_REGPARM(3); // function names of helpers (for debugging generated code) Char *log_1I0D_name, *log_2I0D_name, *log_3I0D_name; Char *log_1I1Dr_name, *log_1I1Dw_name; Char *log_0I1Dr_name, *log_0I1Dw_name; }; // set by setup_bbcc at start of every BB, and needed by log_* helpers extern Addr CLG_(bb_base); extern ULong* CLG_(cost_base); // Event groups #define EG_USE 0 #define EG_IR 1 #define EG_DR 2 #define EG_DW 3 #define EG_BC 4 #define EG_BI 5 #define EG_BUS 6 #define EG_ALLOC 7 #define EG_SYS 8 struct event_sets { EventSet *base, *full; }; extern struct event_sets CLG_(sets); #define fullOffset(group) (CLG_(sets).full->offset[group]) /*------------------------------------------------------------*/ /*--- Functions ---*/ /*------------------------------------------------------------*/ /* from clo.c */ void CLG_(set_clo_defaults)(void); void CLG_(update_fn_config)(fn_node*); Bool CLG_(process_cmd_line_option)(Char*); void CLG_(print_usage)(void); void CLG_(print_debug_usage)(void); /* from sim.c */ extern struct cachesim_if CLG_(cachesim); void CLG_(init_eventsets)(void); /* from main.c */ Bool CLG_(get_debug_info)(Addr, Char filename[FILENAME_LEN], Char fn_name[FN_NAME_LEN], UInt*, DebugInfo**); void CLG_(collectBlockInfo)(IRSB* bbIn, UInt*, UInt*, Bool*); void CLG_(set_instrument_state)(Char*,Bool); void CLG_(dump_profile)(Char* trigger,Bool only_current_thread); void CLG_(zero_all_cost)(Bool only_current_thread); Int CLG_(get_dump_counter)(void); void CLG_(fini)(Int exitcode); /* from bb.c */ void CLG_(init_bb_hash)(void); bb_hash* CLG_(get_bb_hash)(void); BB* CLG_(get_bb)(Addr addr, IRSB* bb_in, Bool *seen_before); void CLG_(delete_bb)(Addr addr); static __inline__ Addr bb_addr(BB* bb) { return bb->offset + bb->obj->offset; } static __inline__ Addr bb_jmpaddr(BB* bb) { UInt off = (bb->instr_count > 0) ? bb->instr[bb->instr_count-1].instr_offset : 0; return off + bb->offset + bb->obj->offset; } /* from fn.c */ void CLG_(init_fn_array)(fn_array*); void CLG_(copy_current_fn_array)(fn_array* dst); fn_array* CLG_(get_current_fn_array)(void); void CLG_(set_current_fn_array)(fn_array*); UInt* CLG_(get_fn_entry)(Int n); void CLG_(init_obj_table)(void); obj_node* CLG_(get_obj_node)(DebugInfo* si); file_node* CLG_(get_file_node)(obj_node*, Char* filename); fn_node* CLG_(get_fn_node)(BB* bb); /* from bbcc.c */ void CLG_(init_bbcc_hash)(bbcc_hash* bbccs); void CLG_(copy_current_bbcc_hash)(bbcc_hash* dst); bbcc_hash* CLG_(get_current_bbcc_hash)(void); void CLG_(set_current_bbcc_hash)(bbcc_hash*); void CLG_(forall_bbccs)(void (*func)(BBCC*)); void CLG_(zero_bbcc)(BBCC* bbcc); BBCC* CLG_(get_bbcc)(BB* bb); BBCC* CLG_(clone_bbcc)(BBCC* orig, Context* cxt, Int rec_index); void CLG_(setup_bbcc)(BB* bb) VG_REGPARM(1); /* from jumps.c */ void CLG_(init_jcc_hash)(jcc_hash*); void CLG_(copy_current_jcc_hash)(jcc_hash* dst); jcc_hash* CLG_(get_current_jcc_hash)(void); void CLG_(set_current_jcc_hash)(jcc_hash*); jCC* CLG_(get_jcc)(BBCC* from, UInt, BBCC* to); /* from callstack.c */ void CLG_(init_call_stack)(call_stack*); void CLG_(copy_current_call_stack)(call_stack* dst); void CLG_(set_current_call_stack)(call_stack*); call_entry* CLG_(get_call_entry)(Int n); void CLG_(push_call_stack)(BBCC* from, UInt jmp, BBCC* to, Addr sp, Bool skip); void CLG_(pop_call_stack)(void); Int CLG_(unwind_call_stack)(Addr sp, Int); /* from context.c */ void CLG_(init_fn_stack)(fn_stack*); void CLG_(copy_current_fn_stack)(fn_stack*); fn_stack* CLG_(get_current_fn_stack)(void); void CLG_(set_current_fn_stack)(fn_stack*); void CLG_(init_cxt_table)(void); cxt_hash* CLG_(get_cxt_hash)(void); Context* CLG_(get_cxt)(fn_node** fn); void CLG_(push_cxt)(fn_node* fn); /* from threads.c */ void CLG_(init_threads)(void); thread_info** CLG_(get_threads)(void); thread_info* CLG_(get_current_thread)(void); void CLG_(switch_thread)(ThreadId tid); void CLG_(forall_threads)(void (*func)(thread_info*)); void CLG_(run_thread)(ThreadId tid); void CLG_(init_exec_state)(exec_state* es); void CLG_(init_exec_stack)(exec_stack*); void CLG_(copy_current_exec_stack)(exec_stack*); void CLG_(set_current_exec_stack)(exec_stack*); void CLG_(pre_signal)(ThreadId tid, Int sigNum, Bool alt_stack); void CLG_(post_signal)(ThreadId tid, Int sigNum); void CLG_(run_post_signal_on_call_stack_bottom)(void); /* from dump.c */ extern FullCost CLG_(total_cost); void CLG_(init_dumps)(void); Char* CLG_(get_out_file)(void); Char* CLG_(get_out_directory)(void); /*------------------------------------------------------------*/ /*--- Exported global variables ---*/ /*------------------------------------------------------------*/ extern CommandLineOptions CLG_(clo); extern Statistics CLG_(stat); extern EventMapping* CLG_(dumpmap); /* Function active counter array, indexed by function number */ extern UInt* CLG_(fn_active_array); extern Bool CLG_(instrument_state); /* min of L1 and LL cache line sizes */ extern Int CLG_(min_line_size); extern call_stack CLG_(current_call_stack); extern fn_stack CLG_(current_fn_stack); extern exec_state CLG_(current_state); extern ThreadId CLG_(current_tid); /*------------------------------------------------------------*/ /*--- Debug output ---*/ /*------------------------------------------------------------*/ #if CLG_ENABLE_DEBUG #define CLG_DEBUGIF(x) \ if ( (CLG_(clo).verbose >x) && \ (CLG_(stat).bb_executions >= CLG_(clo).verbose_start)) #define CLG_DEBUG(x,format,args...) \ CLG_DEBUGIF(x) { \ CLG_(print_bbno)(); \ VG_(printf)(format,##args); \ } #define CLG_ASSERT(cond) \ if (!(cond)) { \ CLG_(print_context)(); \ CLG_(print_bbno)(); \ tl_assert(cond); \ } #else #define CLG_DEBUGIF(x) if (0) #define CLG_DEBUG(x...) {} #define CLG_ASSERT(cond) tl_assert(cond); #endif /* from debug.c */ void CLG_(print_bbno)(void); void CLG_(print_context)(void); void CLG_(print_jcc)(int s, jCC* jcc); void CLG_(print_bbcc)(int s, BBCC* bbcc); void CLG_(print_bbcc_fn)(BBCC* bbcc); void CLG_(print_execstate)(int s, exec_state* es); void CLG_(print_eventset)(int s, EventSet* es); void CLG_(print_cost)(int s, EventSet*, ULong* cost); void CLG_(print_bb)(int s, BB* bb); void CLG_(print_bbcc_cost)(int s, BBCC*); void CLG_(print_cxt)(int s, Context* cxt, int rec_index); void CLG_(print_short_jcc)(jCC* jcc); void CLG_(print_stackentry)(int s, int sp); void CLG_(print_addr)(Addr addr); void CLG_(print_addr_ln)(Addr addr); void* CLG_(malloc)(HChar* cc, UWord s, char* f); void* CLG_(free)(void* p, char* f); #if 0 #define CLG_MALLOC(_cc,x) CLG_(malloc)((_cc),x,__FUNCTION__) #define CLG_FREE(p) CLG_(free)(p,__FUNCTION__) #else #define CLG_MALLOC(_cc,x) VG_(malloc)((_cc),x) #define CLG_FREE(p) VG_(free)(p) #endif #endif /* CLG_GLOBAL */