/* This file is part of drd, a thread error detector. Copyright (C) 2006-2015 Bart Van Assche <bvanassche@acm.org>. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307, USA. The GNU General Public License is contained in the file COPYING. */ #include "drd_error.h" #include "drd_barrier.h" #include "drd_clientobj.h" #include "drd_cond.h" #include "drd_mutex.h" #include "drd_segment.h" #include "drd_semaphore.h" #include "drd_suppression.h" #include "drd_thread.h" #include "pub_tool_vki.h" #include "pub_tool_basics.h" // Addr, SizeT #include "pub_tool_libcassert.h" // tl_assert() #include "pub_tool_libcbase.h" // VG_(strlen)() #include "pub_tool_libcprint.h" // VG_(printf)() #include "pub_tool_machine.h" #include "pub_tool_mallocfree.h" // VG_(malloc)(), VG_(free)() #include "pub_tool_options.h" // VG_(clo_backtrace_size) #include "pub_tool_threadstate.h" // VG_(get_pthread_id)() /* Local functions. */ static void thread_append_segment(const DrdThreadId tid, Segment* const sg); static void thread_discard_segment(const DrdThreadId tid, Segment* const sg); static void thread_compute_conflict_set(struct bitmap** conflict_set, const DrdThreadId tid); static Bool thread_conflict_set_up_to_date(const DrdThreadId tid); /* Local variables. */ static ULong s_context_switch_count; static ULong s_discard_ordered_segments_count; static ULong s_compute_conflict_set_count; static ULong s_update_conflict_set_count; static ULong s_update_conflict_set_new_sg_count; static ULong s_update_conflict_set_sync_count; static ULong s_update_conflict_set_join_count; static ULong s_conflict_set_bitmap_creation_count; static ULong s_conflict_set_bitmap2_creation_count; static ThreadId s_vg_running_tid = VG_INVALID_THREADID; DrdThreadId DRD_(g_drd_running_tid) = DRD_INVALID_THREADID; ThreadInfo* DRD_(g_threadinfo); struct bitmap* DRD_(g_conflict_set); Bool DRD_(verify_conflict_set); static Bool s_trace_context_switches = False; static Bool s_trace_conflict_set = False; static Bool s_trace_conflict_set_bm = False; static Bool s_trace_fork_join = False; static Bool s_segment_merging = True; static Bool s_new_segments_since_last_merge; static int s_segment_merge_interval = 10; static unsigned s_join_list_vol = 10; static unsigned s_deletion_head; static unsigned s_deletion_tail; #if defined(VGO_solaris) Bool DRD_(ignore_thread_creation) = True; #else Bool DRD_(ignore_thread_creation) = False; #endif /* VGO_solaris */ /* Function definitions. */ /** Enables/disables context switch tracing. */ void DRD_(thread_trace_context_switches)(const Bool t) { tl_assert(t == False || t == True); s_trace_context_switches = t; } /** Enables/disables conflict set tracing. */ void DRD_(thread_trace_conflict_set)(const Bool t) { tl_assert(t == False || t == True); s_trace_conflict_set = t; } /** Enables/disables conflict set bitmap tracing. */ void DRD_(thread_trace_conflict_set_bm)(const Bool t) { tl_assert(t == False || t == True); s_trace_conflict_set_bm = t; } /** Report whether fork/join tracing is enabled. */ Bool DRD_(thread_get_trace_fork_join)(void) { return s_trace_fork_join; } /** Enables/disables fork/join tracing. */ void DRD_(thread_set_trace_fork_join)(const Bool t) { tl_assert(t == False || t == True); s_trace_fork_join = t; } /** Enables/disables segment merging. */ void DRD_(thread_set_segment_merging)(const Bool m) { tl_assert(m == False || m == True); s_segment_merging = m; } /** Get the segment merging interval. */ int DRD_(thread_get_segment_merge_interval)(void) { return s_segment_merge_interval; } /** Set the segment merging interval. */ void DRD_(thread_set_segment_merge_interval)(const int i) { s_segment_merge_interval = i; } void DRD_(thread_set_join_list_vol)(const int jlv) { s_join_list_vol = jlv; } void DRD_(thread_init)(void) { DRD_(g_threadinfo) = VG_(malloc)("drd.main.ti.1", DRD_N_THREADS * sizeof DRD_(g_threadinfo)[0]); for (UInt i = 0; i < DRD_N_THREADS; ++i) { static ThreadInfo initval; DRD_(g_threadinfo)[i] = initval; } } /** * Convert Valgrind's ThreadId into a DrdThreadId. * * @return DRD thread ID upon success and DRD_INVALID_THREADID if the passed * Valgrind ThreadId does not yet exist. */ DrdThreadId DRD_(VgThreadIdToDrdThreadId)(const ThreadId tid) { UInt i; if (tid == VG_INVALID_THREADID) return DRD_INVALID_THREADID; for (i = 1; i < DRD_N_THREADS; i++) { if (DRD_(g_threadinfo)[i].vg_thread_exists == True && DRD_(g_threadinfo)[i].vg_threadid == tid) { return i; } } return DRD_INVALID_THREADID; } /** Allocate a new DRD thread ID for the specified Valgrind thread ID. */ static DrdThreadId DRD_(VgThreadIdToNewDrdThreadId)(const ThreadId tid) { UInt i; tl_assert(DRD_(VgThreadIdToDrdThreadId)(tid) == DRD_INVALID_THREADID); for (i = 1; i < DRD_N_THREADS; i++) { if (!DRD_(g_threadinfo)[i].valid) { tl_assert(! DRD_(IsValidDrdThreadId)(i)); DRD_(g_threadinfo)[i].valid = True; DRD_(g_threadinfo)[i].vg_thread_exists = True; DRD_(g_threadinfo)[i].vg_threadid = tid; DRD_(g_threadinfo)[i].pt_threadid = INVALID_POSIX_THREADID; DRD_(g_threadinfo)[i].stack_min = 0; DRD_(g_threadinfo)[i].stack_min_min = 0; DRD_(g_threadinfo)[i].stack_startup = 0; DRD_(g_threadinfo)[i].stack_max = 0; DRD_(thread_set_name)(i, ""); DRD_(g_threadinfo)[i].on_alt_stack = False; DRD_(g_threadinfo)[i].is_recording_loads = True; DRD_(g_threadinfo)[i].is_recording_stores = True; DRD_(g_threadinfo)[i].pthread_create_nesting_level = 0; DRD_(g_threadinfo)[i].synchr_nesting = 0; DRD_(g_threadinfo)[i].deletion_seq = s_deletion_tail - 1; DRD_(g_threadinfo)[i].creator_thread = DRD_INVALID_THREADID; #if defined (VGO_solaris) DRD_(g_threadinfo)[i].bind_guard_flag = 0; #endif /* VGO_solaris */ tl_assert(DRD_(g_threadinfo)[i].sg_first == NULL); tl_assert(DRD_(g_threadinfo)[i].sg_last == NULL); tl_assert(DRD_(IsValidDrdThreadId)(i)); return i; } } VG_(printf)( "\nSorry, but the maximum number of threads supported by DRD has been exceeded." "Aborting.\n"); tl_assert(False); return DRD_INVALID_THREADID; } /** Convert a POSIX thread ID into a DRD thread ID. */ DrdThreadId DRD_(PtThreadIdToDrdThreadId)(const PThreadId tid) { UInt i; if (tid != INVALID_POSIX_THREADID) { for (i = 1; i < DRD_N_THREADS; i++) { if (DRD_(g_threadinfo)[i].posix_thread_exists && DRD_(g_threadinfo)[i].pt_threadid == tid) { return i; } } } return DRD_INVALID_THREADID; } /** Convert a DRD thread ID into a Valgrind thread ID. */ ThreadId DRD_(DrdThreadIdToVgThreadId)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return (DRD_(g_threadinfo)[tid].vg_thread_exists ? DRD_(g_threadinfo)[tid].vg_threadid : VG_INVALID_THREADID); } #ifdef ENABLE_DRD_CONSISTENCY_CHECKS /** * Sanity check of the doubly linked list of segments referenced by a * ThreadInfo struct. * @return True if sane, False if not. */ static Bool DRD_(sane_ThreadInfo)(const ThreadInfo* const ti) { Segment* p; for (p = ti->sg_first; p; p = p->thr_next) { if (p->thr_next && p->thr_next->thr_prev != p) return False; if (p->thr_next == 0 && p != ti->sg_last) return False; } for (p = ti->sg_last; p; p = p->thr_prev) { if (p->thr_prev && p->thr_prev->thr_next != p) return False; if (p->thr_prev == 0 && p != ti->sg_first) return False; } return True; } #endif /** * Create the first segment for a newly started thread. * * This function is called from the handler installed via * VG_(track_pre_thread_ll_create)(). The Valgrind core invokes this handler * from the context of the creator thread, before the new thread has been * created. * * @param[in] creator DRD thread ID of the creator thread. * @param[in] vg_created Valgrind thread ID of the created thread. * * @return DRD thread ID of the created thread. */ DrdThreadId DRD_(thread_pre_create)(const DrdThreadId creator, const ThreadId vg_created) { DrdThreadId created; tl_assert(DRD_(VgThreadIdToDrdThreadId)(vg_created) == DRD_INVALID_THREADID); created = DRD_(VgThreadIdToNewDrdThreadId)(vg_created); tl_assert(0 <= (int)created && created < DRD_N_THREADS && created != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[created].sg_first == NULL); tl_assert(DRD_(g_threadinfo)[created].sg_last == NULL); if (creator != DRD_INVALID_THREADID) { if (DRD_(ignore_thread_creation)) { tl_assert(DRD_(thread_get_synchr_nesting_count)(created) == 0); DRD_(thread_enter_synchr)(created); /* Counterpart in DRD_(thread_set_pthreadid)(). */ } } DRD_(g_threadinfo)[created].creator_thread = creator; /* Create an initial segment for the newly created thread. */ thread_append_segment(created, DRD_(sg_new)(creator, created)); return created; } /** * Initialize DRD_(g_threadinfo)[] for a newly created thread. Must be called * after the thread has been created and before any client instructions are run * on the newly created thread, e.g. from the handler installed via * VG_(track_pre_thread_first_insn)(). * * @param[in] vg_created Valgrind thread ID of the newly created thread. * * @return DRD thread ID for the new thread. */ DrdThreadId DRD_(thread_post_create)(const ThreadId vg_created) { const DrdThreadId created = DRD_(VgThreadIdToDrdThreadId)(vg_created); tl_assert(0 <= (int)created && created < DRD_N_THREADS && created != DRD_INVALID_THREADID); DRD_(g_threadinfo)[created].stack_max = VG_(thread_get_stack_max)(vg_created); DRD_(g_threadinfo)[created].stack_startup = DRD_(g_threadinfo)[created].stack_max; DRD_(g_threadinfo)[created].stack_min = DRD_(g_threadinfo)[created].stack_max; DRD_(g_threadinfo)[created].stack_min_min = DRD_(g_threadinfo)[created].stack_max; DRD_(g_threadinfo)[created].stack_size = VG_(thread_get_stack_size)(vg_created); tl_assert(DRD_(g_threadinfo)[created].stack_max != 0); return created; } static void DRD_(thread_delayed_delete)(const DrdThreadId tid) { UInt j; DRD_(g_threadinfo)[tid].vg_thread_exists = False; DRD_(g_threadinfo)[tid].posix_thread_exists = False; DRD_(g_threadinfo)[tid].deletion_seq = s_deletion_head++; #if 0 VG_(message)(Vg_DebugMsg, "Adding thread %d to the deletion list\n", tid); #endif if (s_deletion_head - s_deletion_tail >= s_join_list_vol) { for (j = 0; j < DRD_N_THREADS; ++j) { if (DRD_(IsValidDrdThreadId)(j) && DRD_(g_threadinfo)[j].deletion_seq == s_deletion_tail) { s_deletion_tail++; #if 0 VG_(message)(Vg_DebugMsg, "Delayed delete of thread %d\n", j); #endif DRD_(thread_delete)(j, False); break; } } } } /** * Process VG_USERREQ__POST_THREAD_JOIN. This client request is invoked just * after thread drd_joiner joined thread drd_joinee. */ void DRD_(thread_post_join)(DrdThreadId drd_joiner, DrdThreadId drd_joinee) { tl_assert(DRD_(IsValidDrdThreadId)(drd_joiner)); tl_assert(DRD_(IsValidDrdThreadId)(drd_joinee)); DRD_(thread_new_segment)(drd_joiner); DRD_(thread_combine_vc_join)(drd_joiner, drd_joinee); DRD_(thread_new_segment)(drd_joinee); if (s_trace_fork_join) { const ThreadId joiner = DRD_(DrdThreadIdToVgThreadId)(drd_joiner); const unsigned msg_size = 256; HChar* msg; msg = VG_(malloc)("drd.main.dptj.1", msg_size); VG_(snprintf)(msg, msg_size, "drd_post_thread_join joiner = %u, joinee = %u", drd_joiner, drd_joinee); if (joiner) { HChar* vc; vc = DRD_(vc_aprint)(DRD_(thread_get_vc)(drd_joiner)); VG_(snprintf)(msg + VG_(strlen)(msg), msg_size - VG_(strlen)(msg), ", new vc: %s", vc); VG_(free)(vc); } DRD_(trace_msg)("%pS", msg); VG_(free)(msg); } if (! DRD_(get_check_stack_accesses)()) { DRD_(finish_suppression)(DRD_(thread_get_stack_max)(drd_joinee) - DRD_(thread_get_stack_size)(drd_joinee), DRD_(thread_get_stack_max)(drd_joinee)); } DRD_(clientobj_delete_thread)(drd_joinee); DRD_(thread_delayed_delete)(drd_joinee); } /** * NPTL hack: NPTL allocates the 'struct pthread' on top of the stack, * and accesses this data structure from multiple threads without locking. * Any conflicting accesses in the range stack_startup..stack_max will be * ignored. */ void DRD_(thread_set_stack_startup)(const DrdThreadId tid, const Addr stack_startup) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].stack_min <= stack_startup); tl_assert(stack_startup <= DRD_(g_threadinfo)[tid].stack_max); DRD_(g_threadinfo)[tid].stack_startup = stack_startup; } /** Return the stack pointer for the specified thread. */ Addr DRD_(thread_get_stack_min)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].stack_min; } /** * Return the lowest value that was ever assigned to the stack pointer * for the specified thread. */ Addr DRD_(thread_get_stack_min_min)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].stack_min_min; } /** Return the top address for the stack of the specified thread. */ Addr DRD_(thread_get_stack_max)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].stack_max; } /** Return the maximum stack size for the specified thread. */ SizeT DRD_(thread_get_stack_size)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].stack_size; } Bool DRD_(thread_get_on_alt_stack)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].on_alt_stack; } void DRD_(thread_set_on_alt_stack)(const DrdThreadId tid, const Bool on_alt_stack) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(on_alt_stack == !!on_alt_stack); DRD_(g_threadinfo)[tid].on_alt_stack = on_alt_stack; } Int DRD_(thread_get_threads_on_alt_stack)(void) { int n = 0; for (UInt i = 1; i < DRD_N_THREADS; i++) n += DRD_(g_threadinfo)[i].on_alt_stack; return n; } /** * Clean up thread-specific data structures. */ void DRD_(thread_delete)(const DrdThreadId tid, const Bool detached) { Segment* sg; Segment* sg_prev; tl_assert(DRD_(IsValidDrdThreadId)(tid)); tl_assert(DRD_(g_threadinfo)[tid].synchr_nesting >= 0); for (sg = DRD_(g_threadinfo)[tid].sg_last; sg; sg = sg_prev) { sg_prev = sg->thr_prev; sg->thr_next = NULL; sg->thr_prev = NULL; DRD_(sg_put)(sg); } DRD_(g_threadinfo)[tid].valid = False; DRD_(g_threadinfo)[tid].vg_thread_exists = False; DRD_(g_threadinfo)[tid].posix_thread_exists = False; if (detached) DRD_(g_threadinfo)[tid].detached_posix_thread = False; else tl_assert(!DRD_(g_threadinfo)[tid].detached_posix_thread); DRD_(g_threadinfo)[tid].sg_first = NULL; DRD_(g_threadinfo)[tid].sg_last = NULL; tl_assert(!DRD_(IsValidDrdThreadId)(tid)); } /** * Called after a thread performed its last memory access and before * thread_delete() is called. Note: thread_delete() is only called for * joinable threads, not for detached threads. */ void DRD_(thread_finished)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); DRD_(g_threadinfo)[tid].vg_thread_exists = False; if (DRD_(g_threadinfo)[tid].detached_posix_thread) { /* * Once a detached thread has finished, its stack is deallocated and * should no longer be taken into account when computing the conflict set. */ DRD_(g_threadinfo)[tid].stack_min = DRD_(g_threadinfo)[tid].stack_max; /* * For a detached thread, calling pthread_exit() invalidates the * POSIX thread ID associated with the detached thread. For joinable * POSIX threads however, the POSIX thread ID remains live after the * pthread_exit() call until pthread_join() is called. */ DRD_(g_threadinfo)[tid].posix_thread_exists = False; } } /** Called just after fork() in the child process. */ void DRD_(drd_thread_atfork_child)(const DrdThreadId tid) { unsigned i; for (i = 1; i < DRD_N_THREADS; i++) { if (i == tid) continue; if (DRD_(IsValidDrdThreadId(i))) DRD_(thread_delete)(i, True); tl_assert(!DRD_(IsValidDrdThreadId(i))); } DRD_(bm_cleanup)(DRD_(g_conflict_set)); DRD_(bm_init)(DRD_(g_conflict_set)); } /** Called just before pthread_cancel(). */ void DRD_(thread_pre_cancel)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID); if (DRD_(thread_get_trace_fork_join)()) DRD_(trace_msg)("[%u] drd_thread_pre_cancel %u", DRD_(g_drd_running_tid), tid); } /** * Store the POSIX thread ID for the specified thread. * * @note This function can be called two times for the same thread -- see also * the comment block preceding the pthread_create() wrapper in * drd_pthread_intercepts.c. */ void DRD_(thread_set_pthreadid)(const DrdThreadId tid, const PThreadId ptid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pt_threadid == INVALID_POSIX_THREADID || DRD_(g_threadinfo)[tid].pt_threadid == ptid); tl_assert(ptid != INVALID_POSIX_THREADID); DRD_(g_threadinfo)[tid].posix_thread_exists = True; DRD_(g_threadinfo)[tid].pt_threadid = ptid; if (DRD_(g_threadinfo)[tid].creator_thread != DRD_INVALID_THREADID) { if (DRD_(ignore_thread_creation)) { DRD_(thread_leave_synchr)(tid); tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0); } } } /** Returns true for joinable threads and false for detached threads. */ Bool DRD_(thread_get_joinable)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return ! DRD_(g_threadinfo)[tid].detached_posix_thread; } /** Store the thread mode: joinable or detached. */ #if defined(VGP_mips32_linux) || defined(VGP_mips64_linux) /* There is a cse related issue in gcc for MIPS. Optimization level has to be lowered, so cse related optimizations are not included.*/ __attribute__((optimize("O1"))) #endif void DRD_(thread_set_joinable)(const DrdThreadId tid, const Bool joinable) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert((!! joinable) == joinable); tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID); DRD_(g_threadinfo)[tid].detached_posix_thread = ! joinable; } /** Tells DRD that the calling thread is about to enter pthread_create(). */ void DRD_(thread_entering_pthread_create)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pthread_create_nesting_level >= 0); DRD_(g_threadinfo)[tid].pthread_create_nesting_level++; if (DRD_(ignore_thread_creation)) { tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0); DRD_(thread_enter_synchr)(tid); } } /** Tells DRD that the calling thread has left pthread_create(). */ void DRD_(thread_left_pthread_create)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID); tl_assert(DRD_(g_threadinfo)[tid].pthread_create_nesting_level > 0); DRD_(g_threadinfo)[tid].pthread_create_nesting_level--; if (DRD_(ignore_thread_creation)) { DRD_(thread_leave_synchr)(tid); tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0); } } #if defined(VGO_solaris) /** Handles the bind_guard() intercept. */ void DRD_(thread_entering_rtld_bind_guard)(const DrdThreadId tid, int flags) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); Int bindflag = (flags & VKI_THR_FLG_RTLD); if ((bindflag & DRD_(g_threadinfo)[tid].bind_guard_flag) == 0) { DRD_(g_threadinfo)[tid].bind_guard_flag |= bindflag; DRD_(thread_enter_synchr)(tid); } } /** * Handles the bind_clear() intercept. * Call to bind_clear(0) is typically used to determine value of bind_flags. */ void DRD_(thread_leaving_rtld_bind_clear)(const DrdThreadId tid, int flags) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); Int bindflag = (flags & VKI_THR_FLG_RTLD); if ((DRD_(g_threadinfo)[tid].bind_guard_flag & bindflag) != 0) { DRD_(g_threadinfo)[tid].bind_guard_flag &= ~bindflag; DRD_(thread_leave_synchr)(tid); } } #endif /* VGO_solaris */ /** Obtain the thread number and the user-assigned thread name. */ const HChar* DRD_(thread_get_name)(const DrdThreadId tid) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); return DRD_(g_threadinfo)[tid].name; } /** Set the name of the specified thread. */ void DRD_(thread_set_name)(const DrdThreadId tid, const HChar* const name) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); if (name == NULL || name[0] == 0) VG_(snprintf)(DRD_(g_threadinfo)[tid].name, sizeof(DRD_(g_threadinfo)[tid].name), "Thread %u", tid); else VG_(snprintf)(DRD_(g_threadinfo)[tid].name, sizeof(DRD_(g_threadinfo)[tid].name), "Thread %u (%s)", tid, name); DRD_(g_threadinfo)[tid].name[sizeof(DRD_(g_threadinfo)[tid].name) - 1] = 0; } /** * Update s_vg_running_tid, DRD_(g_drd_running_tid) and recalculate the * conflict set. */ void DRD_(thread_set_vg_running_tid)(const ThreadId vg_tid) { tl_assert(vg_tid != VG_INVALID_THREADID); if (vg_tid != s_vg_running_tid) { DRD_(thread_set_running_tid)(vg_tid, DRD_(VgThreadIdToDrdThreadId)(vg_tid)); } tl_assert(s_vg_running_tid != VG_INVALID_THREADID); tl_assert(DRD_(g_drd_running_tid) != DRD_INVALID_THREADID); } /** * Update s_vg_running_tid, DRD_(g_drd_running_tid) and recalculate the * conflict set. */ void DRD_(thread_set_running_tid)(const ThreadId vg_tid, const DrdThreadId drd_tid) { tl_assert(vg_tid != VG_INVALID_THREADID); tl_assert(drd_tid != DRD_INVALID_THREADID); if (vg_tid != s_vg_running_tid) { if (s_trace_context_switches && DRD_(g_drd_running_tid) != DRD_INVALID_THREADID) { VG_(message)(Vg_DebugMsg, "Context switch from thread %u to thread %u;" " segments: %llu\n", DRD_(g_drd_running_tid), drd_tid, DRD_(sg_get_segments_alive_count)()); } s_vg_running_tid = vg_tid; DRD_(g_drd_running_tid) = drd_tid; thread_compute_conflict_set(&DRD_(g_conflict_set), drd_tid); s_context_switch_count++; } tl_assert(s_vg_running_tid != VG_INVALID_THREADID); tl_assert(DRD_(g_drd_running_tid) != DRD_INVALID_THREADID); } /** * Increase the synchronization nesting counter. Must be called before the * client calls a synchronization function. */ int DRD_(thread_enter_synchr)(const DrdThreadId tid) { tl_assert(DRD_(IsValidDrdThreadId)(tid)); return DRD_(g_threadinfo)[tid].synchr_nesting++; } /** * Decrease the synchronization nesting counter. Must be called after the * client left a synchronization function. */ int DRD_(thread_leave_synchr)(const DrdThreadId tid) { tl_assert(DRD_(IsValidDrdThreadId)(tid)); tl_assert(DRD_(g_threadinfo)[tid].synchr_nesting >= 1); return --DRD_(g_threadinfo)[tid].synchr_nesting; } /** Returns the synchronization nesting counter. */ int DRD_(thread_get_synchr_nesting_count)(const DrdThreadId tid) { tl_assert(DRD_(IsValidDrdThreadId)(tid)); return DRD_(g_threadinfo)[tid].synchr_nesting; } /** Append a new segment at the end of the segment list. */ static void thread_append_segment(const DrdThreadId tid, Segment* const sg) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid])); #endif // add at tail sg->thr_prev = DRD_(g_threadinfo)[tid].sg_last; sg->thr_next = NULL; if (DRD_(g_threadinfo)[tid].sg_last) DRD_(g_threadinfo)[tid].sg_last->thr_next = sg; DRD_(g_threadinfo)[tid].sg_last = sg; if (DRD_(g_threadinfo)[tid].sg_first == NULL) DRD_(g_threadinfo)[tid].sg_first = sg; #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid])); #endif } /** * Remove a segment from the segment list of thread threadid, and free the * associated memory. */ static void thread_discard_segment(const DrdThreadId tid, Segment* const sg) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid])); #endif if (sg->thr_prev) sg->thr_prev->thr_next = sg->thr_next; if (sg->thr_next) sg->thr_next->thr_prev = sg->thr_prev; if (sg == DRD_(g_threadinfo)[tid].sg_first) DRD_(g_threadinfo)[tid].sg_first = sg->thr_next; if (sg == DRD_(g_threadinfo)[tid].sg_last) DRD_(g_threadinfo)[tid].sg_last = sg->thr_prev; DRD_(sg_put)(sg); #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid])); #endif } /** * Returns a pointer to the vector clock of the most recent segment associated * with thread 'tid'. */ VectorClock* DRD_(thread_get_vc)(const DrdThreadId tid) { Segment* latest_sg; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); latest_sg = DRD_(g_threadinfo)[tid].sg_last; tl_assert(latest_sg); return &latest_sg->vc; } /** * Return the latest segment of thread 'tid' and increment its reference count. */ void DRD_(thread_get_latest_segment)(Segment** sg, const DrdThreadId tid) { Segment* latest_sg; tl_assert(sg); tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); latest_sg = DRD_(g_threadinfo)[tid].sg_last; tl_assert(latest_sg); DRD_(sg_put)(*sg); *sg = DRD_(sg_get)(latest_sg); } /** * Compute the minimum of all latest vector clocks of all threads * (Michiel Ronsse calls this "clock snooping" in his papers about DIOTA). * * @param vc pointer to a vectorclock, holds result upon return. */ static void DRD_(thread_compute_minimum_vc)(VectorClock* vc) { unsigned i; Bool first; Segment* latest_sg; first = True; for (i = 0; i < DRD_N_THREADS; i++) { latest_sg = DRD_(g_threadinfo)[i].sg_last; if (latest_sg) { if (first) DRD_(vc_assign)(vc, &latest_sg->vc); else DRD_(vc_min)(vc, &latest_sg->vc); first = False; } } } /** * Compute the maximum of all latest vector clocks of all threads. * * @param vc pointer to a vectorclock, holds result upon return. */ static void DRD_(thread_compute_maximum_vc)(VectorClock* vc) { unsigned i; Bool first; Segment* latest_sg; first = True; for (i = 0; i < DRD_N_THREADS; i++) { latest_sg = DRD_(g_threadinfo)[i].sg_last; if (latest_sg) { if (first) DRD_(vc_assign)(vc, &latest_sg->vc); else DRD_(vc_combine)(vc, &latest_sg->vc); first = False; } } } /** * Discard all segments that have a defined order against the latest vector * clock of all threads -- these segments can no longer be involved in a * data race. */ static void thread_discard_ordered_segments(void) { unsigned i; VectorClock thread_vc_min; s_discard_ordered_segments_count++; DRD_(vc_init)(&thread_vc_min, 0, 0); DRD_(thread_compute_minimum_vc)(&thread_vc_min); if (DRD_(sg_get_trace)()) { HChar *vc_min, *vc_max; VectorClock thread_vc_max; DRD_(vc_init)(&thread_vc_max, 0, 0); DRD_(thread_compute_maximum_vc)(&thread_vc_max); vc_min = DRD_(vc_aprint)(&thread_vc_min); vc_max = DRD_(vc_aprint)(&thread_vc_max); VG_(message)(Vg_DebugMsg, "Discarding ordered segments -- min vc is %s, max vc is %s\n", vc_min, vc_max); VG_(free)(vc_min); VG_(free)(vc_max); DRD_(vc_cleanup)(&thread_vc_max); } for (i = 0; i < DRD_N_THREADS; i++) { Segment* sg; Segment* sg_next; for (sg = DRD_(g_threadinfo)[i].sg_first; sg && (sg_next = sg->thr_next) && DRD_(vc_lte)(&sg->vc, &thread_vc_min); sg = sg_next) { thread_discard_segment(i, sg); } } DRD_(vc_cleanup)(&thread_vc_min); } /** * An implementation of the property 'equiv(sg1, sg2)' as defined in the paper * by Mark Christiaens e.a. The property equiv(sg1, sg2) holds if and only if * all segments in the set CS are ordered consistently against both sg1 and * sg2. The set CS is defined as the set of segments that can immediately * precede future segments via inter-thread synchronization operations. In * DRD the set CS consists of the latest segment of each thread combined with * all segments for which the reference count is strictly greater than one. * The code below is an optimized version of the following: * * for (i = 0; i < DRD_N_THREADS; i++) * { * Segment* sg; * * for (sg = DRD_(g_threadinfo)[i].first; sg; sg = sg->next) * { * if (sg == DRD_(g_threadinfo)[i].last || DRD_(sg_get_refcnt)(sg) > 1) * { * if ( DRD_(vc_lte)(&sg1->vc, &sg->vc) * != DRD_(vc_lte)(&sg2->vc, &sg->vc) * || DRD_(vc_lte)(&sg->vc, &sg1->vc) * != DRD_(vc_lte)(&sg->vc, &sg2->vc)) * { * return False; * } * } * } * } */ static Bool thread_consistent_segment_ordering(const DrdThreadId tid, Segment* const sg1, Segment* const sg2) { unsigned i; tl_assert(sg1->thr_next); tl_assert(sg2->thr_next); tl_assert(sg1->thr_next == sg2); tl_assert(DRD_(vc_lte)(&sg1->vc, &sg2->vc)); for (i = 0; i < DRD_N_THREADS; i++) { Segment* sg; for (sg = DRD_(g_threadinfo)[i].sg_first; sg; sg = sg->thr_next) { if (!sg->thr_next || DRD_(sg_get_refcnt)(sg) > 1) { if (DRD_(vc_lte)(&sg2->vc, &sg->vc)) break; if (DRD_(vc_lte)(&sg1->vc, &sg->vc)) return False; } } for (sg = DRD_(g_threadinfo)[i].sg_last; sg; sg = sg->thr_prev) { if (!sg->thr_next || DRD_(sg_get_refcnt)(sg) > 1) { if (DRD_(vc_lte)(&sg->vc, &sg1->vc)) break; if (DRD_(vc_lte)(&sg->vc, &sg2->vc)) return False; } } } return True; } /** * Merge all segments that may be merged without triggering false positives * or discarding real data races. For the theoretical background of segment * merging, see also the following paper: Mark Christiaens, Michiel Ronsse * and Koen De Bosschere. Bounding the number of segment histories during * data race detection. Parallel Computing archive, Volume 28, Issue 9, * pp 1221-1238, September 2002. This paper contains a proof that merging * consecutive segments for which the property equiv(s1,s2) holds can be * merged without reducing the accuracy of datarace detection. Furthermore * it is also proven that the total number of all segments will never grow * unbounded if all segments s1, s2 for which equiv(s1, s2) holds are merged * every time a new segment is created. The property equiv(s1, s2) is defined * as follows: equiv(s1, s2) <=> for all segments in the set CS, the vector * clocks of segments s and s1 are ordered in the same way as those of segments * s and s2. The set CS is defined as the set of existing segments s that have * the potential to conflict with not yet created segments, either because the * segment s is the latest segment of a thread or because it can become the * immediate predecessor of a new segment due to a synchronization operation. */ static void thread_merge_segments(void) { unsigned i; s_new_segments_since_last_merge = 0; for (i = 0; i < DRD_N_THREADS; i++) { Segment* sg; #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[i])); #endif for (sg = DRD_(g_threadinfo)[i].sg_first; sg; sg = sg->thr_next) { if (DRD_(sg_get_refcnt)(sg) == 1 && sg->thr_next) { Segment* const sg_next = sg->thr_next; if (DRD_(sg_get_refcnt)(sg_next) == 1 && sg_next->thr_next && thread_consistent_segment_ordering(i, sg, sg_next)) { /* Merge sg and sg_next into sg. */ DRD_(sg_merge)(sg, sg_next); thread_discard_segment(i, sg_next); } } } #ifdef ENABLE_DRD_CONSISTENCY_CHECKS tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[i])); #endif } } /** * Create a new segment for the specified thread, and discard any segments * that cannot cause races anymore. */ void DRD_(thread_new_segment)(const DrdThreadId tid) { Segment* last_sg; Segment* new_sg; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid))); last_sg = DRD_(g_threadinfo)[tid].sg_last; new_sg = DRD_(sg_new)(tid, tid); thread_append_segment(tid, new_sg); if (tid == DRD_(g_drd_running_tid) && last_sg) { DRD_(thread_update_conflict_set)(tid, &last_sg->vc); s_update_conflict_set_new_sg_count++; } tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid))); if (s_segment_merging && ++s_new_segments_since_last_merge >= s_segment_merge_interval) { thread_discard_ordered_segments(); thread_merge_segments(); } } /** Call this function after thread 'joiner' joined thread 'joinee'. */ void DRD_(thread_combine_vc_join)(DrdThreadId joiner, DrdThreadId joinee) { tl_assert(joiner != joinee); tl_assert(0 <= (int)joiner && joiner < DRD_N_THREADS && joiner != DRD_INVALID_THREADID); tl_assert(0 <= (int)joinee && joinee < DRD_N_THREADS && joinee != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[joiner].sg_first); tl_assert(DRD_(g_threadinfo)[joiner].sg_last); tl_assert(DRD_(g_threadinfo)[joinee].sg_first); tl_assert(DRD_(g_threadinfo)[joinee].sg_last); if (DRD_(sg_get_trace)()) { HChar *str1, *str2; str1 = DRD_(vc_aprint)(DRD_(thread_get_vc)(joiner)); str2 = DRD_(vc_aprint)(DRD_(thread_get_vc)(joinee)); VG_(message)(Vg_DebugMsg, "Before join: joiner %s, joinee %s\n", str1, str2); VG_(free)(str1); VG_(free)(str2); } if (joiner == DRD_(g_drd_running_tid)) { VectorClock old_vc; DRD_(vc_copy)(&old_vc, DRD_(thread_get_vc)(joiner)); DRD_(vc_combine)(DRD_(thread_get_vc)(joiner), DRD_(thread_get_vc)(joinee)); DRD_(thread_update_conflict_set)(joiner, &old_vc); s_update_conflict_set_join_count++; DRD_(vc_cleanup)(&old_vc); } else { DRD_(vc_combine)(DRD_(thread_get_vc)(joiner), DRD_(thread_get_vc)(joinee)); } thread_discard_ordered_segments(); if (DRD_(sg_get_trace)()) { HChar* str; str = DRD_(vc_aprint)(DRD_(thread_get_vc)(joiner)); VG_(message)(Vg_DebugMsg, "After join: %s\n", str); VG_(free)(str); } } /** * Update the vector clock of the last segment of thread tid with the * the vector clock of segment sg. */ static void thread_combine_vc_sync(DrdThreadId tid, const Segment* sg) { const VectorClock* const vc = &sg->vc; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(DRD_(g_threadinfo)[tid].sg_first); tl_assert(DRD_(g_threadinfo)[tid].sg_last); tl_assert(sg); tl_assert(vc); if (tid != sg->tid) { VectorClock old_vc; DRD_(vc_copy)(&old_vc, DRD_(thread_get_vc)(tid)); DRD_(vc_combine)(DRD_(thread_get_vc)(tid), vc); if (DRD_(sg_get_trace)()) { HChar *str1, *str2; str1 = DRD_(vc_aprint)(&old_vc); str2 = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid)); VG_(message)(Vg_DebugMsg, "thread %u: vc %s -> %s\n", tid, str1, str2); VG_(free)(str1); VG_(free)(str2); } thread_discard_ordered_segments(); DRD_(thread_update_conflict_set)(tid, &old_vc); s_update_conflict_set_sync_count++; DRD_(vc_cleanup)(&old_vc); } else { tl_assert(DRD_(vc_lte)(vc, DRD_(thread_get_vc)(tid))); } } /** * Create a new segment for thread tid and update the vector clock of the last * segment of this thread with the vector clock of segment sg. Call this * function after thread tid had to wait because of thread synchronization * until the memory accesses in the segment sg finished. */ void DRD_(thread_new_segment_and_combine_vc)(DrdThreadId tid, const Segment* sg) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid))); tl_assert(sg); thread_append_segment(tid, DRD_(sg_new)(tid, tid)); thread_combine_vc_sync(tid, sg); if (s_segment_merging && ++s_new_segments_since_last_merge >= s_segment_merge_interval) { thread_discard_ordered_segments(); thread_merge_segments(); } } /** * Call this function whenever a thread is no longer using the memory * [ a1, a2 [, e.g. because of a call to free() or a stack pointer * increase. */ void DRD_(thread_stop_using_mem)(const Addr a1, const Addr a2) { Segment* p; for (p = DRD_(g_sg_list); p; p = p->g_next) DRD_(bm_clear)(DRD_(sg_bm)(p), a1, a2); DRD_(bm_clear)(DRD_(g_conflict_set), a1, a2); } /** Specify whether memory loads should be recorded. */ void DRD_(thread_set_record_loads)(const DrdThreadId tid, const Bool enabled) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(enabled == !! enabled); DRD_(g_threadinfo)[tid].is_recording_loads = enabled; } /** Specify whether memory stores should be recorded. */ void DRD_(thread_set_record_stores)(const DrdThreadId tid, const Bool enabled) { tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(enabled == !! enabled); DRD_(g_threadinfo)[tid].is_recording_stores = enabled; } /** * Print the segment information for all threads. * * This function is only used for debugging purposes. */ void DRD_(thread_print_all)(void) { UInt i; Segment* p; for (i = 0; i < DRD_N_THREADS; i++) { p = DRD_(g_threadinfo)[i].sg_first; if (p) { VG_(printf)("**************\n" "* thread %3u (%d/%u/%u/%u/0x%lx/%d) *\n" "**************\n", i, DRD_(g_threadinfo)[i].valid, DRD_(g_threadinfo)[i].vg_thread_exists, DRD_(g_threadinfo)[i].vg_threadid, DRD_(g_threadinfo)[i].posix_thread_exists, DRD_(g_threadinfo)[i].pt_threadid, DRD_(g_threadinfo)[i].detached_posix_thread); for ( ; p; p = p->thr_next) DRD_(sg_print)(p); } } } /** Show a call stack involved in a data race. */ static void show_call_stack(const DrdThreadId tid, ExeContext* const callstack) { const ThreadId vg_tid = DRD_(DrdThreadIdToVgThreadId)(tid); if (vg_tid != VG_INVALID_THREADID) { if (callstack) VG_(pp_ExeContext)(callstack); else VG_(get_and_pp_StackTrace)(vg_tid, VG_(clo_backtrace_size)); } else { if (!VG_(clo_xml)) VG_(message)(Vg_UserMsg, " (thread finished, call stack no longer available)\n"); } } /** Print information about the segments involved in a data race. */ static void thread_report_conflicting_segments_segment(const DrdThreadId tid, const Addr addr, const SizeT size, const BmAccessTypeT access_type, const Segment* const p) { unsigned i; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(p); for (i = 0; i < DRD_N_THREADS; i++) { if (i != tid) { Segment* q; for (q = DRD_(g_threadinfo)[i].sg_last; q; q = q->thr_prev) { /* * Since q iterates over the segments of thread i in order of * decreasing vector clocks, if q->vc <= p->vc, then * q->next->vc <= p->vc will also hold. Hence, break out of the * loop once this condition is met. */ if (DRD_(vc_lte)(&q->vc, &p->vc)) break; if (!DRD_(vc_lte)(&p->vc, &q->vc)) { if (DRD_(bm_has_conflict_with)(DRD_(sg_bm)(q), addr, addr + size, access_type)) { Segment* q_next; tl_assert(q->stacktrace); if (VG_(clo_xml)) VG_(printf_xml)(" <other_segment_start>\n"); else VG_(message)(Vg_UserMsg, "Other segment start (thread %u)\n", i); show_call_stack(i, q->stacktrace); if (VG_(clo_xml)) VG_(printf_xml)(" </other_segment_start>\n" " <other_segment_end>\n"); else VG_(message)(Vg_UserMsg, "Other segment end (thread %u)\n", i); q_next = q->thr_next; show_call_stack(i, q_next ? q_next->stacktrace : 0); if (VG_(clo_xml)) VG_(printf_xml)(" </other_segment_end>\n"); } } } } } } /** Print information about all segments involved in a data race. */ void DRD_(thread_report_conflicting_segments)(const DrdThreadId tid, const Addr addr, const SizeT size, const BmAccessTypeT access_type) { Segment* p; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); for (p = DRD_(g_threadinfo)[tid].sg_first; p; p = p->thr_next) { if (DRD_(bm_has)(DRD_(sg_bm)(p), addr, addr + size, access_type)) thread_report_conflicting_segments_segment(tid, addr, size, access_type, p); } } /** * Verify whether the conflict set for thread tid is up to date. Only perform * the check if the environment variable DRD_VERIFY_CONFLICT_SET has been set. */ static Bool thread_conflict_set_up_to_date(const DrdThreadId tid) { Bool result; struct bitmap* computed_conflict_set = 0; if (!DRD_(verify_conflict_set)) return True; thread_compute_conflict_set(&computed_conflict_set, tid); result = DRD_(bm_equal)(DRD_(g_conflict_set), computed_conflict_set); if (! result) { VG_(printf)("actual conflict set:\n"); DRD_(bm_print)(DRD_(g_conflict_set)); VG_(printf)("\n"); VG_(printf)("computed conflict set:\n"); DRD_(bm_print)(computed_conflict_set); VG_(printf)("\n"); } DRD_(bm_delete)(computed_conflict_set); return result; } /** * Compute the conflict set: a bitmap that represents the union of all memory * accesses of all segments that are unordered to the current segment of the * thread tid. */ static void thread_compute_conflict_set(struct bitmap** conflict_set, const DrdThreadId tid) { Segment* p; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(tid == DRD_(g_drd_running_tid)); s_compute_conflict_set_count++; s_conflict_set_bitmap_creation_count -= DRD_(bm_get_bitmap_creation_count)(); s_conflict_set_bitmap2_creation_count -= DRD_(bm_get_bitmap2_creation_count)(); if (*conflict_set) { DRD_(bm_cleanup)(*conflict_set); DRD_(bm_init)(*conflict_set); } else { *conflict_set = DRD_(bm_new)(); } if (s_trace_conflict_set) { HChar* str; str = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid)); VG_(message)(Vg_DebugMsg, "computing conflict set for thread %u with vc %s\n", tid, str); VG_(free)(str); } p = DRD_(g_threadinfo)[tid].sg_last; { unsigned j; if (s_trace_conflict_set) { HChar* vc; vc = DRD_(vc_aprint)(&p->vc); VG_(message)(Vg_DebugMsg, "conflict set: thread [%u] at vc %s\n", tid, vc); VG_(free)(vc); } for (j = 0; j < DRD_N_THREADS; j++) { if (j != tid && DRD_(IsValidDrdThreadId)(j)) { Segment* q; for (q = DRD_(g_threadinfo)[j].sg_last; q; q = q->thr_prev) { if (!DRD_(vc_lte)(&q->vc, &p->vc) && !DRD_(vc_lte)(&p->vc, &q->vc)) { if (s_trace_conflict_set) { HChar* str; str = DRD_(vc_aprint)(&q->vc); VG_(message)(Vg_DebugMsg, "conflict set: [%u] merging segment %s\n", j, str); VG_(free)(str); } DRD_(bm_merge2)(*conflict_set, DRD_(sg_bm)(q)); } else { if (s_trace_conflict_set) { HChar* str; str = DRD_(vc_aprint)(&q->vc); VG_(message)(Vg_DebugMsg, "conflict set: [%u] ignoring segment %s\n", j, str); VG_(free)(str); } } } } } } s_conflict_set_bitmap_creation_count += DRD_(bm_get_bitmap_creation_count)(); s_conflict_set_bitmap2_creation_count += DRD_(bm_get_bitmap2_creation_count)(); if (s_trace_conflict_set_bm) { VG_(message)(Vg_DebugMsg, "[%u] new conflict set:\n", tid); DRD_(bm_print)(*conflict_set); VG_(message)(Vg_DebugMsg, "[%u] end of new conflict set.\n", tid); } } /** * Update the conflict set after the vector clock of thread tid has been * updated from old_vc to its current value, either because a new segment has * been created or because of a synchronization operation. */ void DRD_(thread_update_conflict_set)(const DrdThreadId tid, const VectorClock* const old_vc) { const VectorClock* new_vc; Segment* p; unsigned j; tl_assert(0 <= (int)tid && tid < DRD_N_THREADS && tid != DRD_INVALID_THREADID); tl_assert(old_vc); tl_assert(tid == DRD_(g_drd_running_tid)); tl_assert(DRD_(g_conflict_set)); if (s_trace_conflict_set) { HChar* str; str = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid)); VG_(message)(Vg_DebugMsg, "updating conflict set for thread %u with vc %s\n", tid, str); VG_(free)(str); } new_vc = DRD_(thread_get_vc)(tid); tl_assert(DRD_(vc_lte)(old_vc, new_vc)); DRD_(bm_unmark)(DRD_(g_conflict_set)); for (j = 0; j < DRD_N_THREADS; j++) { Segment* q; if (j == tid || ! DRD_(IsValidDrdThreadId)(j)) continue; for (q = DRD_(g_threadinfo)[j].sg_last; q && !DRD_(vc_lte)(&q->vc, new_vc); q = q->thr_prev) { const Bool included_in_old_conflict_set = !DRD_(vc_lte)(old_vc, &q->vc); const Bool included_in_new_conflict_set = !DRD_(vc_lte)(new_vc, &q->vc); if (UNLIKELY(s_trace_conflict_set)) { HChar* str; str = DRD_(vc_aprint)(&q->vc); VG_(message)(Vg_DebugMsg, "conflict set: [%u] %s segment %s\n", j, included_in_old_conflict_set != included_in_new_conflict_set ? "merging" : "ignoring", str); VG_(free)(str); } if (included_in_old_conflict_set != included_in_new_conflict_set) DRD_(bm_mark)(DRD_(g_conflict_set), DRD_(sg_bm)(q)); } for ( ; q && !DRD_(vc_lte)(&q->vc, old_vc); q = q->thr_prev) { const Bool included_in_old_conflict_set = !DRD_(vc_lte)(old_vc, &q->vc); const Bool included_in_new_conflict_set = !DRD_(vc_lte)(&q->vc, new_vc) && !DRD_(vc_lte)(new_vc, &q->vc); if (UNLIKELY(s_trace_conflict_set)) { HChar* str; str = DRD_(vc_aprint)(&q->vc); VG_(message)(Vg_DebugMsg, "conflict set: [%u] %s segment %s\n", j, included_in_old_conflict_set != included_in_new_conflict_set ? "merging" : "ignoring", str); VG_(free)(str); } if (included_in_old_conflict_set != included_in_new_conflict_set) DRD_(bm_mark)(DRD_(g_conflict_set), DRD_(sg_bm)(q)); } } DRD_(bm_clear_marked)(DRD_(g_conflict_set)); p = DRD_(g_threadinfo)[tid].sg_last; for (j = 0; j < DRD_N_THREADS; j++) { if (j != tid && DRD_(IsValidDrdThreadId)(j)) { Segment* q; for (q = DRD_(g_threadinfo)[j].sg_last; q && !DRD_(vc_lte)(&q->vc, &p->vc); q = q->thr_prev) { if (!DRD_(vc_lte)(&p->vc, &q->vc)) DRD_(bm_merge2_marked)(DRD_(g_conflict_set), DRD_(sg_bm)(q)); } } } DRD_(bm_remove_cleared_marked)(DRD_(g_conflict_set)); s_update_conflict_set_count++; if (s_trace_conflict_set_bm) { VG_(message)(Vg_DebugMsg, "[%u] updated conflict set:\n", tid); DRD_(bm_print)(DRD_(g_conflict_set)); VG_(message)(Vg_DebugMsg, "[%u] end of updated conflict set.\n", tid); } tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid))); } /** Report the number of context switches performed. */ ULong DRD_(thread_get_context_switch_count)(void) { return s_context_switch_count; } /** Report the number of ordered segments that have been discarded. */ ULong DRD_(thread_get_discard_ordered_segments_count)(void) { return s_discard_ordered_segments_count; } /** Return how many times the conflict set has been updated entirely. */ ULong DRD_(thread_get_compute_conflict_set_count)() { return s_compute_conflict_set_count; } /** Return how many times the conflict set has been updated partially. */ ULong DRD_(thread_get_update_conflict_set_count)(void) { return s_update_conflict_set_count; } /** * Return how many times the conflict set has been updated partially * because a new segment has been created. */ ULong DRD_(thread_get_update_conflict_set_new_sg_count)(void) { return s_update_conflict_set_new_sg_count; } /** * Return how many times the conflict set has been updated partially * because of combining vector clocks due to synchronization operations * other than reader/writer lock or barrier operations. */ ULong DRD_(thread_get_update_conflict_set_sync_count)(void) { return s_update_conflict_set_sync_count; } /** * Return how many times the conflict set has been updated partially * because of thread joins. */ ULong DRD_(thread_get_update_conflict_set_join_count)(void) { return s_update_conflict_set_join_count; } /** * Return the number of first-level bitmaps that have been created during * conflict set updates. */ ULong DRD_(thread_get_conflict_set_bitmap_creation_count)(void) { return s_conflict_set_bitmap_creation_count; } /** * Return the number of second-level bitmaps that have been created during * conflict set updates. */ ULong DRD_(thread_get_conflict_set_bitmap2_creation_count)(void) { return s_conflict_set_bitmap2_creation_count; }