static void *memory_top;
static RaceSet * raceset;
+#ifdef COLLECT_STAT
static unsigned int store8_count = 0;
static unsigned int store16_count = 0;
static unsigned int store32_count = 0;
static unsigned int load16_count = 0;
static unsigned int load32_count = 0;
static unsigned int load64_count = 0;
+#endif
static const ModelExecution * get_execution()
{
/** This function looks up the entry in the shadow table corresponding to a
* given address.*/
-static uint64_t * lookupAddressEntry(const void *address)
+static inline uint64_t * lookupAddressEntry(const void *address)
{
struct ShadowTable *currtable = root;
#if BIT48
* @param tid2 The thread ID for the potentially-racing action
* @return true if the current clock allows a race with the event at clock2/tid2
*/
-static bool clock_may_race(ClockVector *clock1, thread_id_t tid1,
+static inline bool clock_may_race(ClockVector *clock1, thread_id_t tid1,
modelclock_t clock2, thread_id_t tid2)
{
return tid1 != tid2 && clock2 != 0 && clock1->getClock(tid2) <= clock2;
return race;
}
-/** This function does race detection on a write. */
-void raceCheckWrite(thread_id_t thread, void *location)
-{
- uint64_t *shadow = lookupAddressEntry(location);
- uint64_t shadowval = *shadow;
- ClockVector *currClock = get_execution()->get_cv(thread);
- if (currClock == NULL)
- return;
-
- struct DataRace * race = NULL;
- /* Do full record */
- if (shadowval != 0 && !ISSHORTRECORD(shadowval)) {
- race = fullRaceCheckWrite(thread, location, shadow, currClock);
- goto Exit;
- }
-
- {
- int threadid = id_to_int(thread);
- modelclock_t ourClock = currClock->getClock(thread);
-
- /* Thread ID is too large or clock is too large. */
- if (threadid > MAXTHREADID || ourClock > MAXWRITEVECTOR) {
- expandRecord(shadow);
- race = fullRaceCheckWrite(thread, location, shadow, currClock);
- goto Exit;
- }
-
- {
- /* Check for datarace against last read. */
- modelclock_t readClock = READVECTOR(shadowval);
- thread_id_t readThread = int_to_id(RDTHREADID(shadowval));
-
- if (clock_may_race(currClock, thread, readClock, readThread)) {
- /* We have a datarace */
- race = reportDataRace(readThread, readClock, false, get_execution()->get_parent_action(thread), true, location);
- goto ShadowExit;
- }
- }
-
- {
- /* Check for datarace against last write. */
- modelclock_t writeClock = WRITEVECTOR(shadowval);
- thread_id_t writeThread = int_to_id(WRTHREADID(shadowval));
-
- if (clock_may_race(currClock, thread, writeClock, writeThread)) {
- /* We have a datarace */
- race = reportDataRace(writeThread, writeClock, true, get_execution()->get_parent_action(thread), true, location);
- goto ShadowExit;
- }
- }
-
-ShadowExit:
- *shadow = ENCODEOP(0, 0, threadid, ourClock);
- }
-
-Exit:
- if (race) {
-#ifdef REPORT_DATA_RACES
- race->numframes=backtrace(race->backtrace, sizeof(race->backtrace)/sizeof(void*));
- if (raceset->add(race))
- assert_race(race);
- else model_free(race);
-#else
- model_free(race);
-#endif
- }
-}
-
/** This function does race detection for a write on an expanded record. */
struct DataRace * atomfullRaceCheckWrite(thread_id_t thread, const void *location, uint64_t *shadow, ClockVector *currClock)
{
/** This function just updates metadata on atomic write. */
void recordCalloc(void *location, size_t size) {
- thread_id_t thread = thread_current()->get_id();
+ thread_id_t thread = thread_current_id();
for(;size != 0;size--) {
uint64_t *shadow = lookupAddressEntry(location);
uint64_t shadowval = *shadow;
if (clock_may_race(currClock, thread, readClock, readThread)) {
/* Still need this read in vector */
if (copytoindex != i) {
- ASSERT(record->thread[i] >= 0);
- record->readClock[copytoindex] = record->readClock[i];
- record->thread[copytoindex] = record->thread[i];
+ ASSERT(readThread >= 0);
+ record->readClock[copytoindex] = readClock;
+ record->thread[copytoindex] = readThread;
}
copytoindex++;
}
int newCapacity = copytoindex * 2;
thread_id_t *newthread = (thread_id_t *)snapshot_malloc(sizeof(thread_id_t) * newCapacity);
modelclock_t *newreadClock = (modelclock_t *)snapshot_malloc(sizeof(modelclock_t) * newCapacity);
- std::memcpy(newthread, record->thread, copytoindex * sizeof(thread_id_t));
- std::memcpy(newreadClock, record->readClock, copytoindex * sizeof(modelclock_t));
+ real_memcpy(newthread, record->thread, copytoindex * sizeof(thread_id_t));
+ real_memcpy(newreadClock, record->readClock, copytoindex * sizeof(modelclock_t));
snapshot_free(record->readClock);
snapshot_free(record->thread);
record->readClock = newreadClock;
return race;
}
-/** This function does race detection on a read. */
-void raceCheckRead(thread_id_t thread, const void *location)
-{
- uint64_t *shadow = lookupAddressEntry(location);
- uint64_t shadowval = *shadow;
- ClockVector *currClock = get_execution()->get_cv(thread);
- if (currClock == NULL)
- return;
-
- struct DataRace * race = NULL;
-
- /* Do full record */
- if (shadowval != 0 && !ISSHORTRECORD(shadowval)) {
- race = fullRaceCheckRead(thread, location, shadow, currClock);
- goto Exit;
- }
-
- {
- int threadid = id_to_int(thread);
- modelclock_t ourClock = currClock->getClock(thread);
-
- /* Thread ID is too large or clock is too large. */
- if (threadid > MAXTHREADID || ourClock > MAXWRITEVECTOR) {
- expandRecord(shadow);
- race = fullRaceCheckRead(thread, location, shadow, currClock);
- goto Exit;
- }
-
- /* Check for datarace against last write. */
-
- modelclock_t writeClock = WRITEVECTOR(shadowval);
- thread_id_t writeThread = int_to_id(WRTHREADID(shadowval));
-
- if (clock_may_race(currClock, thread, writeClock, writeThread)) {
- /* We have a datarace */
- race = reportDataRace(writeThread, writeClock, true, get_execution()->get_parent_action(thread), false, location);
- goto ShadowExit;
- }
-
-ShadowExit:
- {
- modelclock_t readClock = READVECTOR(shadowval);
- thread_id_t readThread = int_to_id(RDTHREADID(shadowval));
-
- if (clock_may_race(currClock, thread, readClock, readThread)) {
- /* We don't subsume this read... Have to expand record. */
- expandRecord(shadow);
- fullRaceCheckRead(thread, location, shadow, currClock);
- goto Exit;
- }
- }
-
- *shadow = ENCODEOP(threadid, ourClock, id_to_int(writeThread), writeClock) | (shadowval & ATOMICMASK);
- }
-Exit:
- if (race) {
-#ifdef REPORT_DATA_RACES
- race->numframes=backtrace(race->backtrace, sizeof(race->backtrace)/sizeof(void*));
- if (raceset->add(race))
- assert_race(race);
- else model_free(race);
-#else
- model_free(race);
-#endif
- }
-}
-
-
/** This function does race detection on a read for an expanded record. */
struct DataRace * atomfullRaceCheckRead(thread_id_t thread, const void *location, uint64_t *shadow, ClockVector *currClock)
{
race = reportDataRace(writeThread, writeClock, true, get_execution()->get_parent_action(thread), false, location);
goto Exit;
}
-
-
}
Exit:
if (race) {
return shadow;
}
-static inline void raceCheckRead_otherIt(thread_id_t thread, const void * location) {
+static inline void raceCheckRead_otherIt(thread_id_t thread, const void * location)
+{
uint64_t *shadow = lookupAddressEntry(location);
-
uint64_t shadowval = *shadow;
-
ClockVector *currClock = get_execution()->get_cv(thread);
if (currClock == NULL)
return;
void raceCheckRead64(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (shadow[7]==old_shadowval)
shadow[7] = new_shadowval;
else goto L7;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
raceCheckRead_otherIt(thread, (const void *)(((uintptr_t)location) + 6));
L7:
raceCheckRead_otherIt(thread, (const void *)(((uintptr_t)location) + 7));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckRead32(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (shadow[3]==old_shadowval)
shadow[3] = new_shadowval;
else goto L3;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
raceCheckRead_otherIt(thread, (const void *)(((uintptr_t)location) + 2));
L3:
raceCheckRead_otherIt(thread, (const void *)(((uintptr_t)location) + 3));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckRead16(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (CHECKBOUNDARY(location, 1)) {
if (shadow[1]==old_shadowval) {
shadow[1] = new_shadowval;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
}
raceCheckRead_otherIt(thread, (const void *)(((uintptr_t)location) + 1));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckRead8(thread_id_t thread, const void *location)
{
- uint64_t old_shadowval, new_shadowval;
- old_shadowval = new_shadowval = INVALIDSHADOWVAL;
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
#ifdef COLLECT_STAT
load8_count++;
#endif
- raceCheckRead_firstIt(thread, location, &old_shadowval, &new_shadowval);
+ raceCheckRead_otherIt(thread, location);
+ RESTORE_MODEL_FLAG(old_flag);
}
static inline uint64_t * raceCheckWrite_firstIt(thread_id_t thread, const void * location, uint64_t *old_val, uint64_t *new_val)
return shadow;
}
-static inline void raceCheckWrite_otherIt(thread_id_t thread, const void * location) {
+static inline void raceCheckWrite_otherIt(thread_id_t thread, const void * location)
+{
uint64_t *shadow = lookupAddressEntry(location);
-
uint64_t shadowval = *shadow;
-
ClockVector *currClock = get_execution()->get_cv(thread);
if (currClock == NULL)
return;
void raceCheckWrite64(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (shadow[7]==old_shadowval)
shadow[7] = new_shadowval;
else goto L7;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
raceCheckWrite_otherIt(thread, (const void *)(((uintptr_t)location) + 6));
L7:
raceCheckWrite_otherIt(thread, (const void *)(((uintptr_t)location) + 7));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckWrite32(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (shadow[3]==old_shadowval)
shadow[3] = new_shadowval;
else goto L3;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
raceCheckWrite_otherIt(thread, (const void *)(((uintptr_t)location) + 2));
L3:
raceCheckWrite_otherIt(thread, (const void *)(((uintptr_t)location) + 3));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckWrite16(thread_id_t thread, const void *location)
{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
uint64_t old_shadowval, new_shadowval;
old_shadowval = new_shadowval = INVALIDSHADOWVAL;
#ifdef COLLECT_STAT
if (CHECKBOUNDARY(location, 1)) {
if (shadow[1]==old_shadowval) {
shadow[1] = new_shadowval;
+ RESTORE_MODEL_FLAG(old_flag);
return;
}
}
raceCheckWrite_otherIt(thread, (const void *)(((uintptr_t)location) + 1));
+ RESTORE_MODEL_FLAG(old_flag);
}
void raceCheckWrite8(thread_id_t thread, const void *location)
{
- uint64_t old_shadowval, new_shadowval;
- old_shadowval = new_shadowval = INVALIDSHADOWVAL;
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
#ifdef COLLECT_STAT
store8_count++;
#endif
- raceCheckWrite_firstIt(thread, location, &old_shadowval, &new_shadowval);
+ raceCheckWrite_otherIt(thread, location);
+ RESTORE_MODEL_FLAG(old_flag);
+}
+
+void raceCheckWriteMemop(thread_id_t thread, const void *location, size_t size)
+{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
+ ClockVector *currClock = get_execution()->get_cv(thread);
+ if (currClock == NULL) {
+ RESTORE_MODEL_FLAG(old_flag);
+ return;
+ }
+
+ bool alreadyHasRace = false;
+ for (uint i = 0; i < size; i++) {
+ uint64_t *shadow = lookupAddressEntry(location);
+ uint64_t shadowval = *shadow;
+
+ struct DataRace * race = NULL;
+ /* Do full record */
+ if (shadowval != 0 && !ISSHORTRECORD(shadowval)) {
+ race = fullRaceCheckWrite(thread, location, shadow, currClock);
+ goto Exit;
+ }
+
+ {
+ int threadid = id_to_int(thread);
+ modelclock_t ourClock = currClock->getClock(thread);
+
+ /* Thread ID is too large or clock is too large. */
+ if (threadid > MAXTHREADID || ourClock > MAXWRITEVECTOR) {
+ expandRecord(shadow);
+ race = fullRaceCheckWrite(thread, location, shadow, currClock);
+ goto Exit;
+ }
+
+ {
+ /* Check for datarace against last read. */
+ modelclock_t readClock = READVECTOR(shadowval);
+ thread_id_t readThread = int_to_id(RDTHREADID(shadowval));
+
+ if (clock_may_race(currClock, thread, readClock, readThread)) {
+ /* We have a datarace */
+ race = reportDataRace(readThread, readClock, false, get_execution()->get_parent_action(thread), true, location);
+ goto ShadowExit;
+ }
+ }
+
+ {
+ /* Check for datarace against last write. */
+ modelclock_t writeClock = WRITEVECTOR(shadowval);
+ thread_id_t writeThread = int_to_id(WRTHREADID(shadowval));
+
+ if (clock_may_race(currClock, thread, writeClock, writeThread)) {
+ /* We have a datarace */
+ race = reportDataRace(writeThread, writeClock, true, get_execution()->get_parent_action(thread), true, location);
+ goto ShadowExit;
+ }
+ }
+
+ShadowExit:
+ *shadow = ENCODEOP(0, 0, threadid, ourClock);
+ }
+
+Exit:
+ if (race) {
+#ifdef REPORT_DATA_RACES
+ if (!alreadyHasRace) {
+ alreadyHasRace = true;
+ race->numframes=backtrace(race->backtrace, sizeof(race->backtrace)/sizeof(void*));
+ if (raceset->add(race))
+ assert_race(race);
+ else model_free(race);
+ } else {
+ model_free(race);
+ }
+#else
+ model_free(race);
+#endif
+ }
+ }
+ RESTORE_MODEL_FLAG(old_flag);
}
+void raceCheckReadMemop(thread_id_t thread, const void * location, size_t size)
+{
+ int old_flag = GET_MODEL_FLAG;
+ ENTER_MODEL_FLAG;
+
+ ClockVector *currClock = get_execution()->get_cv(thread);
+ if (currClock == NULL) {
+ RESTORE_MODEL_FLAG(old_flag);
+ return;
+ }
+
+ bool alreadyHasRace = false;
+ for (uint i = 0; i < size; i++) {
+ uint64_t *shadow = lookupAddressEntry(location);
+ uint64_t shadowval = *shadow;
+ struct DataRace * race = NULL;
+
+ /* Do full record */
+ if (shadowval != 0 && !ISSHORTRECORD(shadowval)) {
+ race = fullRaceCheckRead(thread, location, shadow, currClock);
+ goto Exit;
+ }
+
+ {
+ int threadid = id_to_int(thread);
+ modelclock_t ourClock = currClock->getClock(thread);
+
+ /* Thread ID is too large or clock is too large. */
+ if (threadid > MAXTHREADID || ourClock > MAXWRITEVECTOR) {
+ expandRecord(shadow);
+ race = fullRaceCheckRead(thread, location, shadow, currClock);
+ goto Exit;
+ }
+
+ /* Check for datarace against last write. */
+ modelclock_t writeClock = WRITEVECTOR(shadowval);
+ thread_id_t writeThread = int_to_id(WRTHREADID(shadowval));
+
+ if (clock_may_race(currClock, thread, writeClock, writeThread)) {
+ /* We have a datarace */
+ race = reportDataRace(writeThread, writeClock, true, get_execution()->get_parent_action(thread), false, location);
+ }
+
+ modelclock_t readClock = READVECTOR(shadowval);
+ thread_id_t readThread = int_to_id(RDTHREADID(shadowval));
+
+ if (clock_may_race(currClock, thread, readClock, readThread)) {
+ /* We don't subsume this read... Have to expand record. */
+ expandRecord(shadow);
+ struct RaceRecord *record = (struct RaceRecord *) (*shadow);
+ record->thread[1] = thread;
+ record->readClock[1] = ourClock;
+ record->numReads++;
+
+ goto Exit;
+ }
+
+ *shadow = ENCODEOP(threadid, ourClock, id_to_int(writeThread), writeClock) | (shadowval & ATOMICMASK);
+ }
+Exit:
+ if (race) {
+#ifdef REPORT_DATA_RACES
+ if (!alreadyHasRace) {
+ race->numframes=backtrace(race->backtrace, sizeof(race->backtrace)/sizeof(void*));
+ if (raceset->add(race))
+ assert_race(race);
+ else model_free(race);
+ } else {
+ model_free(race);
+ }
+#else
+ model_free(race);
+#endif
+ }
+ }
+ RESTORE_MODEL_FLAG(old_flag);
+}
+
+#ifdef COLLECT_STAT
void print_normal_accesses()
{
model_print("store 8 count: %u\n", store8_count);
model_print("load 32 count: %u\n", load32_count);
model_print("load 64 count: %u\n", load64_count);
}
+#endif