| /* Copyright (C) 2008-2019 Free Software Foundation, Inc. |
| Contributed by Richard Henderson <rth@redhat.com>. |
| |
| This file is part of the GNU Transactional Memory Library (libitm). |
| |
| Libitm is free software; you can redistribute it and/or modify it |
| under the terms of the GNU General Public License as published by |
| the Free Software Foundation; either version 3 of the License, or |
| (at your option) any later version. |
| |
| Libitm is distributed in the hope that it will be useful, but WITHOUT ANY |
| WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
| FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| more details. |
| |
| Under Section 7 of GPL version 3, you are granted additional |
| permissions described in the GCC Runtime Library Exception, version |
| 3.1, as published by the Free Software Foundation. |
| |
| You should have received a copy of the GNU General Public License and |
| a copy of the GCC Runtime Library Exception along with this program; |
| see the files COPYING3 and COPYING.RUNTIME respectively. If not, see |
| <http://www.gnu.org/licenses/>. */ |
| |
| #include "libitm_i.h" |
| #include <pthread.h> |
| |
| |
| using namespace GTM; |
| |
| #if !defined(HAVE_ARCH_GTM_THREAD) || !defined(HAVE_ARCH_GTM_THREAD_DISP) |
| extern __thread gtm_thread_tls _gtm_thr_tls; |
| #endif |
| |
| // Put this at the start of a cacheline so that serial_lock's writers and |
| // htm_fastpath fields are on the same cacheline, so that HW transactions |
| // only have to pay one cacheline capacity to monitor both. |
| gtm_rwlock GTM::gtm_thread::serial_lock |
| __attribute__((aligned(HW_CACHELINE_SIZE))); |
| gtm_thread *GTM::gtm_thread::list_of_threads = 0; |
| unsigned GTM::gtm_thread::number_of_threads = 0; |
| |
| /* ??? Move elsewhere when we figure out library initialization. */ |
| uint64_t GTM::gtm_spin_count_var = 1000; |
| |
| #ifdef HAVE_64BIT_SYNC_BUILTINS |
| static atomic<_ITM_transactionId_t> global_tid; |
| #else |
| static _ITM_transactionId_t global_tid; |
| static pthread_mutex_t global_tid_lock = PTHREAD_MUTEX_INITIALIZER; |
| #endif |
| |
| |
| // Provides a on-thread-exit callback used to release per-thread data. |
| static pthread_key_t thr_release_key; |
| static pthread_once_t thr_release_once = PTHREAD_ONCE_INIT; |
| |
| /* Allocate a transaction structure. */ |
| void * |
| GTM::gtm_thread::operator new (size_t s) |
| { |
| void *tx; |
| |
| assert(s == sizeof(gtm_thread)); |
| |
| tx = xmalloc (sizeof (gtm_thread), true); |
| memset (tx, 0, sizeof (gtm_thread)); |
| |
| return tx; |
| } |
| |
| /* Free the given transaction. Raises an error if the transaction is still |
| in use. */ |
| void |
| GTM::gtm_thread::operator delete(void *tx) |
| { |
| free(tx); |
| } |
| |
| static void |
| thread_exit_handler(void *) |
| { |
| gtm_thread *thr = gtm_thr(); |
| if (thr) |
| delete thr; |
| set_gtm_thr(0); |
| } |
| |
| static void |
| thread_exit_init() |
| { |
| if (pthread_key_create(&thr_release_key, thread_exit_handler)) |
| GTM_fatal("Creating thread release TLS key failed."); |
| } |
| |
| |
| GTM::gtm_thread::~gtm_thread() |
| { |
| if (nesting > 0) |
| GTM_fatal("Thread exit while a transaction is still active."); |
| |
| // Deregister this transaction. |
| serial_lock.write_lock (); |
| gtm_thread **prev = &list_of_threads; |
| for (; *prev; prev = &(*prev)->next_thread) |
| { |
| if (*prev == this) |
| { |
| *prev = (*prev)->next_thread; |
| break; |
| } |
| } |
| number_of_threads--; |
| number_of_threads_changed(number_of_threads + 1, number_of_threads); |
| serial_lock.write_unlock (); |
| } |
| |
| GTM::gtm_thread::gtm_thread () |
| { |
| // This object's memory has been set to zero by operator new, so no need |
| // to initialize any of the other primitive-type members that do not have |
| // constructors. |
| shared_state.store(-1, memory_order_relaxed); |
| |
| // Register this transaction with the list of all threads' transactions. |
| serial_lock.write_lock (); |
| next_thread = list_of_threads; |
| list_of_threads = this; |
| number_of_threads++; |
| number_of_threads_changed(number_of_threads - 1, number_of_threads); |
| serial_lock.write_unlock (); |
| |
| init_cpp_exceptions (); |
| |
| if (pthread_once(&thr_release_once, thread_exit_init)) |
| GTM_fatal("Initializing thread release TLS key failed."); |
| // Any non-null value is sufficient to trigger destruction of this |
| // transaction when the current thread terminates. |
| if (pthread_setspecific(thr_release_key, this)) |
| GTM_fatal("Setting thread release TLS key failed."); |
| } |
| |
| static inline uint32_t |
| choose_code_path(uint32_t prop, abi_dispatch *disp) |
| { |
| if ((prop & pr_uninstrumentedCode) && disp->can_run_uninstrumented_code()) |
| return a_runUninstrumentedCode; |
| else |
| return a_runInstrumentedCode; |
| } |
| |
| #ifdef TARGET_BEGIN_TRANSACTION_ATTRIBUTE |
| /* This macro can be used to define target specific attributes for this |
| function. For example, S/390 requires floating point to be disabled in |
| begin_transaction. */ |
| TARGET_BEGIN_TRANSACTION_ATTRIBUTE |
| #endif |
| uint32_t |
| GTM::gtm_thread::begin_transaction (uint32_t prop, const gtm_jmpbuf *jb) |
| { |
| static const _ITM_transactionId_t tid_block_size = 1 << 16; |
| |
| gtm_thread *tx; |
| abi_dispatch *disp; |
| uint32_t ret; |
| |
| // ??? pr_undoLogCode is not properly defined in the ABI. Are barriers |
| // omitted because they are not necessary (e.g., a transaction on thread- |
| // local data) or because the compiler thinks that some kind of global |
| // synchronization might perform better? |
| if (unlikely(prop & pr_undoLogCode)) |
| GTM_fatal("pr_undoLogCode not supported"); |
| |
| #ifdef USE_HTM_FASTPATH |
| // HTM fastpath. Only chosen in the absence of transaction_cancel to allow |
| // using an uninstrumented code path. |
| // The fastpath is enabled only by dispatch_htm's method group, which uses |
| // serial-mode methods as fallback. Serial-mode transactions cannot execute |
| // concurrently with HW transactions because the latter monitor the serial |
| // lock's writer flag and thus abort if another thread is or becomes a |
| // serial transaction. Therefore, if the fastpath is enabled, then a |
| // transaction is not executing as a HW transaction iff the serial lock is |
| // write-locked. Also, HW transactions monitor the fastpath control |
| // variable, so that they will only execute if dispatch_htm is still the |
| // current method group. This allows us to use htm_fastpath and the serial |
| // lock's writers flag to reliable determine whether the current thread runs |
| // a HW transaction, and thus we do not need to maintain this information in |
| // per-thread state. |
| // If an uninstrumented code path is not available, we can still run |
| // instrumented code from a HW transaction because the HTM fastpath kicks |
| // in early in both begin and commit, and the transaction is not canceled. |
| // HW transactions might get requests to switch to serial-irrevocable mode, |
| // but these can be ignored because the HTM provides all necessary |
| // correctness guarantees. Transactions cannot detect whether they are |
| // indeed in serial mode, and HW transactions should never need serial mode |
| // for any internal changes (e.g., they never abort visibly to the STM code |
| // and thus do not trigger the standard retry handling). |
| #ifndef HTM_CUSTOM_FASTPATH |
| if (likely(serial_lock.get_htm_fastpath() && (prop & pr_hasNoAbort))) |
| { |
| // Note that the snapshot of htm_fastpath that we take here could be |
| // outdated, and a different method group than dispatch_htm may have |
| // been chosen in the meantime. Therefore, take care not not touch |
| // anything besides the serial lock, which is independent of method |
| // groups. |
| for (uint32_t t = serial_lock.get_htm_fastpath(); t; t--) |
| { |
| uint32_t ret = htm_begin(); |
| if (htm_begin_success(ret)) |
| { |
| // We are executing a transaction now. |
| // Monitor the writer flag in the serial-mode lock, and abort |
| // if there is an active or waiting serial-mode transaction. |
| // Also checks that htm_fastpath is still nonzero and thus |
| // HW transactions are allowed to run. |
| // Note that this can also happen due to an enclosing |
| // serial-mode transaction; we handle this case below. |
| if (unlikely(serial_lock.htm_fastpath_disabled())) |
| htm_abort(); |
| else |
| // We do not need to set a_saveLiveVariables because of HTM. |
| return (prop & pr_uninstrumentedCode) ? |
| a_runUninstrumentedCode : a_runInstrumentedCode; |
| } |
| // The transaction has aborted. Don't retry if it's unlikely that |
| // retrying the transaction will be successful. |
| if (!htm_abort_should_retry(ret)) |
| break; |
| // Check whether the HTM fastpath has been disabled. |
| if (!serial_lock.get_htm_fastpath()) |
| break; |
| // Wait until any concurrent serial-mode transactions have finished. |
| // This is an empty critical section, but won't be elided. |
| if (serial_lock.htm_fastpath_disabled()) |
| { |
| tx = gtm_thr(); |
| if (unlikely(tx == NULL)) |
| { |
| // See below. |
| tx = new gtm_thread(); |
| set_gtm_thr(tx); |
| } |
| // Check whether there is an enclosing serial-mode transaction; |
| // if so, we just continue as a nested transaction and don't |
| // try to use the HTM fastpath. This case can happen when an |
| // outermost relaxed transaction calls unsafe code that starts |
| // a transaction. |
| if (tx->nesting > 0) |
| break; |
| // Another thread is running a serial-mode transaction. Wait. |
| serial_lock.read_lock(tx); |
| serial_lock.read_unlock(tx); |
| // TODO We should probably reset the retry count t here, unless |
| // we have retried so often that we should go serial to avoid |
| // starvation. |
| } |
| } |
| } |
| #else |
| // If we have a custom HTM fastpath in ITM_beginTransaction, we implement |
| // just the retry policy here. We communicate with the custom fastpath |
| // through additional property bits and return codes, and either transfer |
| // control back to the custom fastpath or run the fallback mechanism. The |
| // fastpath synchronization algorithm itself is the same. |
| // pr_HTMRetryableAbort states that a HW transaction started by the custom |
| // HTM fastpath aborted, and that we thus have to decide whether to retry |
| // the fastpath (returning a_tryHTMFastPath) or just proceed with the |
| // fallback method. |
| if (likely(serial_lock.get_htm_fastpath() && (prop & pr_HTMRetryableAbort))) |
| { |
| tx = gtm_thr(); |
| if (unlikely(tx == NULL)) |
| { |
| // See below. |
| tx = new gtm_thread(); |
| set_gtm_thr(tx); |
| } |
| // If this is the first abort, reset the retry count. We abuse |
| // restart_total for the retry count, which is fine because our only |
| // other fallback will use serial transactions, which don't use |
| // restart_total but will reset it when committing. |
| if (!(prop & pr_HTMRetriedAfterAbort)) |
| tx->restart_total = gtm_thread::serial_lock.get_htm_fastpath(); |
| |
| if (--tx->restart_total > 0) |
| { |
| // Wait until any concurrent serial-mode transactions have finished. |
| // Essentially the same code as above. |
| if (!serial_lock.get_htm_fastpath()) |
| goto stop_custom_htm_fastpath; |
| if (serial_lock.htm_fastpath_disabled()) |
| { |
| if (tx->nesting > 0) |
| goto stop_custom_htm_fastpath; |
| serial_lock.read_lock(tx); |
| serial_lock.read_unlock(tx); |
| } |
| // Let ITM_beginTransaction retry the custom HTM fastpath. |
| return a_tryHTMFastPath; |
| } |
| } |
| stop_custom_htm_fastpath: |
| #endif |
| #endif |
| |
| tx = gtm_thr(); |
| if (unlikely(tx == NULL)) |
| { |
| // Create the thread object. The constructor will also set up automatic |
| // deletion on thread termination. |
| tx = new gtm_thread(); |
| set_gtm_thr(tx); |
| } |
| |
| if (tx->nesting > 0) |
| { |
| // This is a nested transaction. |
| // Check prop compatibility: |
| // The ABI requires pr_hasNoFloatUpdate, pr_hasNoVectorUpdate, |
| // pr_hasNoIrrevocable, pr_aWBarriersOmitted, pr_RaRBarriersOmitted, and |
| // pr_hasNoSimpleReads to hold for the full dynamic scope of a |
| // transaction. We could check that these are set for the nested |
| // transaction if they are also set for the parent transaction, but the |
| // ABI does not require these flags to be set if they could be set, |
| // so the check could be too strict. |
| // ??? For pr_readOnly, lexical or dynamic scope is unspecified. |
| |
| if (prop & pr_hasNoAbort) |
| { |
| // We can use flat nesting, so elide this transaction. |
| if (!(prop & pr_instrumentedCode)) |
| { |
| if (!(tx->state & STATE_SERIAL) || |
| !(tx->state & STATE_IRREVOCABLE)) |
| tx->serialirr_mode(); |
| } |
| // Increment nesting level after checking that we have a method that |
| // allows us to continue. |
| tx->nesting++; |
| return choose_code_path(prop, abi_disp()); |
| } |
| |
| // The transaction might abort, so use closed nesting if possible. |
| // pr_hasNoAbort has lexical scope, so the compiler should really have |
| // generated an instrumented code path. |
| assert(prop & pr_instrumentedCode); |
| |
| // Create a checkpoint of the current transaction. |
| gtm_transaction_cp *cp = tx->parent_txns.push(); |
| cp->save(tx); |
| new (&tx->alloc_actions) aa_tree<uintptr_t, gtm_alloc_action>(); |
| |
| // Check whether the current method actually supports closed nesting. |
| // If we can switch to another one, do so. |
| // If not, we assume that actual aborts are infrequent, and rather |
| // restart in _ITM_abortTransaction when we really have to. |
| disp = abi_disp(); |
| if (!disp->closed_nesting()) |
| { |
| // ??? Should we elide the transaction if there is no alternative |
| // method that supports closed nesting? If we do, we need to set |
| // some flag to prevent _ITM_abortTransaction from aborting the |
| // wrong transaction (i.e., some parent transaction). |
| abi_dispatch *cn_disp = disp->closed_nesting_alternative(); |
| if (cn_disp) |
| { |
| disp = cn_disp; |
| set_abi_disp(disp); |
| } |
| } |
| } |
| else |
| { |
| // Outermost transaction |
| disp = tx->decide_begin_dispatch (prop); |
| set_abi_disp (disp); |
| } |
| |
| // Initialization that is common for outermost and nested transactions. |
| tx->prop = prop; |
| tx->nesting++; |
| |
| tx->jb = *jb; |
| |
| // As long as we have not exhausted a previously allocated block of TIDs, |
| // we can avoid an atomic operation on a shared cacheline. |
| if (tx->local_tid & (tid_block_size - 1)) |
| tx->id = tx->local_tid++; |
| else |
| { |
| #ifdef HAVE_64BIT_SYNC_BUILTINS |
| // We don't really care which block of TIDs we get but only that we |
| // acquire one atomically; therefore, relaxed memory order is |
| // sufficient. |
| tx->id = global_tid.fetch_add(tid_block_size, memory_order_relaxed); |
| tx->local_tid = tx->id + 1; |
| #else |
| pthread_mutex_lock (&global_tid_lock); |
| global_tid += tid_block_size; |
| tx->id = global_tid; |
| tx->local_tid = tx->id + 1; |
| pthread_mutex_unlock (&global_tid_lock); |
| #endif |
| } |
| |
| // Log the number of uncaught exceptions if we might have to roll back this |
| // state. |
| if (tx->cxa_uncaught_count_ptr != 0) |
| tx->cxa_uncaught_count = *tx->cxa_uncaught_count_ptr; |
| |
| // Run dispatch-specific restart code. Retry until we succeed. |
| GTM::gtm_restart_reason rr; |
| while ((rr = disp->begin_or_restart()) != NO_RESTART) |
| { |
| tx->decide_retry_strategy(rr); |
| disp = abi_disp(); |
| } |
| |
| // Determine the code path to run. Only irrevocable transactions cannot be |
| // restarted, so all other transactions need to save live variables. |
| ret = choose_code_path(prop, disp); |
| if (!(tx->state & STATE_IRREVOCABLE)) |
| ret |= a_saveLiveVariables; |
| return ret; |
| } |
| |
| |
| void |
| GTM::gtm_transaction_cp::save(gtm_thread* tx) |
| { |
| // Save everything that we might have to restore on restarts or aborts. |
| jb = tx->jb; |
| undolog_size = tx->undolog.size(); |
| |
| /* FIXME! Assignment of an aatree like alloc_actions is unsafe; if either |
| *this or *tx is destroyed, the other ends up pointing to a freed node. */ |
| #pragma GCC diagnostic warning "-Wdeprecated-copy" |
| alloc_actions = tx->alloc_actions; |
| |
| user_actions_size = tx->user_actions.size(); |
| id = tx->id; |
| prop = tx->prop; |
| cxa_catch_count = tx->cxa_catch_count; |
| cxa_uncaught_count = tx->cxa_uncaught_count; |
| disp = abi_disp(); |
| nesting = tx->nesting; |
| } |
| |
| void |
| GTM::gtm_transaction_cp::commit(gtm_thread* tx) |
| { |
| // Restore state that is not persistent across commits. Exception handling, |
| // information, nesting level, and any logs do not need to be restored on |
| // commits of nested transactions. Allocation actions must be committed |
| // before committing the snapshot. |
| tx->jb = jb; |
| tx->alloc_actions = alloc_actions; |
| tx->id = id; |
| tx->prop = prop; |
| } |
| |
| |
| void |
| GTM::gtm_thread::rollback (gtm_transaction_cp *cp, bool aborting) |
| { |
| // The undo log is special in that it used for both thread-local and shared |
| // data. Because of the latter, we have to roll it back before any |
| // dispatch-specific rollback (which handles synchronization with other |
| // transactions). |
| undolog.rollback (this, cp ? cp->undolog_size : 0); |
| |
| // Perform dispatch-specific rollback. |
| abi_disp()->rollback (cp); |
| |
| // Roll back all actions that are supposed to happen around the transaction. |
| rollback_user_actions (cp ? cp->user_actions_size : 0); |
| commit_allocations (true, (cp ? &cp->alloc_actions : 0)); |
| revert_cpp_exceptions (cp); |
| |
| if (cp) |
| { |
| // We do not yet handle restarts of nested transactions. To do that, we |
| // would have to restore some state (jb, id, prop, nesting) not to the |
| // checkpoint but to the transaction that was started from this |
| // checkpoint (e.g., nesting = cp->nesting + 1); |
| assert(aborting); |
| // Roll back the rest of the state to the checkpoint. |
| jb = cp->jb; |
| id = cp->id; |
| prop = cp->prop; |
| if (cp->disp != abi_disp()) |
| set_abi_disp(cp->disp); |
| alloc_actions = cp->alloc_actions; |
| nesting = cp->nesting; |
| } |
| else |
| { |
| // Roll back to the outermost transaction. |
| // Restore the jump buffer and transaction properties, which we will |
| // need for the longjmp used to restart or abort the transaction. |
| if (parent_txns.size() > 0) |
| { |
| jb = parent_txns[0].jb; |
| id = parent_txns[0].id; |
| prop = parent_txns[0].prop; |
| } |
| // Reset the transaction. Do not reset this->state, which is handled by |
| // the callers. Note that if we are not aborting, we reset the |
| // transaction to the point after having executed begin_transaction |
| // (we will return from it), so the nesting level must be one, not zero. |
| nesting = (aborting ? 0 : 1); |
| parent_txns.clear(); |
| } |
| |
| if (this->eh_in_flight) |
| { |
| _Unwind_DeleteException ((_Unwind_Exception *) this->eh_in_flight); |
| this->eh_in_flight = NULL; |
| } |
| } |
| |
| void ITM_REGPARM |
| _ITM_abortTransaction (_ITM_abortReason reason) |
| { |
| gtm_thread *tx = gtm_thr(); |
| |
| assert (reason == userAbort || reason == (userAbort | outerAbort)); |
| assert ((tx->prop & pr_hasNoAbort) == 0); |
| |
| if (tx->state & gtm_thread::STATE_IRREVOCABLE) |
| abort (); |
| |
| // Roll back to innermost transaction. |
| if (tx->parent_txns.size() > 0 && !(reason & outerAbort)) |
| { |
| // If the current method does not support closed nesting but we are |
| // nested and must only roll back the innermost transaction, then |
| // restart with a method that supports closed nesting. |
| abi_dispatch *disp = abi_disp(); |
| if (!disp->closed_nesting()) |
| tx->restart(RESTART_CLOSED_NESTING); |
| |
| // The innermost transaction is a closed nested transaction. |
| gtm_transaction_cp *cp = tx->parent_txns.pop(); |
| uint32_t longjmp_prop = tx->prop; |
| gtm_jmpbuf longjmp_jb = tx->jb; |
| |
| tx->rollback (cp, true); |
| |
| // Jump to nested transaction (use the saved jump buffer). |
| GTM_longjmp (a_abortTransaction | a_restoreLiveVariables, |
| &longjmp_jb, longjmp_prop); |
| } |
| else |
| { |
| // There is no nested transaction or an abort of the outermost |
| // transaction was requested, so roll back to the outermost transaction. |
| tx->rollback (0, true); |
| |
| // Aborting an outermost transaction finishes execution of the whole |
| // transaction. Therefore, reset transaction state. |
| if (tx->state & gtm_thread::STATE_SERIAL) |
| gtm_thread::serial_lock.write_unlock (); |
| else |
| gtm_thread::serial_lock.read_unlock (tx); |
| tx->state = 0; |
| |
| GTM_longjmp (a_abortTransaction | a_restoreLiveVariables, |
| &tx->jb, tx->prop); |
| } |
| } |
| |
| bool |
| GTM::gtm_thread::trycommit () |
| { |
| nesting--; |
| |
| // Skip any real commit for elided transactions. |
| if (nesting > 0 && (parent_txns.size() == 0 || |
| nesting > parent_txns[parent_txns.size() - 1].nesting)) |
| return true; |
| |
| if (nesting > 0) |
| { |
| // Commit of a closed-nested transaction. Remove one checkpoint and add |
| // any effects of this transaction to the parent transaction. |
| gtm_transaction_cp *cp = parent_txns.pop(); |
| commit_allocations(false, &cp->alloc_actions); |
| cp->commit(this); |
| return true; |
| } |
| |
| // Commit of an outermost transaction. |
| gtm_word priv_time = 0; |
| if (abi_disp()->trycommit (priv_time)) |
| { |
| // The transaction is now finished but we will still access some shared |
| // data if we have to ensure privatization safety. |
| bool do_read_unlock = false; |
| if (state & gtm_thread::STATE_SERIAL) |
| { |
| gtm_thread::serial_lock.write_unlock (); |
| // There are no other active transactions, so there's no need to |
| // enforce privatization safety. |
| priv_time = 0; |
| } |
| else |
| { |
| // If we have to ensure privatization safety, we must not yet |
| // release the read lock and become inactive because (1) we still |
| // have to go through the list of all transactions, which can be |
| // modified by serial mode threads, and (2) we interpret each |
| // transactions' shared_state in the context of what we believe to |
| // be the current method group (and serial mode transactions can |
| // change the method group). Therefore, if we have to ensure |
| // privatization safety, delay becoming inactive but set a maximum |
| // snapshot time (we have committed and thus have an empty snapshot, |
| // so it will always be most recent). Use release MO so that this |
| // synchronizes with other threads observing our snapshot time. |
| if (priv_time) |
| { |
| do_read_unlock = true; |
| shared_state.store((~(typeof gtm_thread::shared_state)0) - 1, |
| memory_order_release); |
| } |
| else |
| gtm_thread::serial_lock.read_unlock (this); |
| } |
| state = 0; |
| |
| // We can commit the undo log after dispatch-specific commit and after |
| // making the transaction inactive because we only have to reset |
| // gtm_thread state. |
| undolog.commit (); |
| // Reset further transaction state. |
| cxa_catch_count = 0; |
| restart_total = 0; |
| |
| // Ensure privatization safety, if necessary. |
| if (priv_time) |
| { |
| // There must be a seq_cst fence between the following loads of the |
| // other transactions' shared_state and the dispatch-specific stores |
| // that signal updates by this transaction (e.g., lock |
| // acquisitions). This ensures that if we read prior to other |
| // reader transactions setting their shared_state to 0, then those |
| // readers will observe our updates. We can reuse the seq_cst fence |
| // in serial_lock.read_unlock() if we performed that; if not, we |
| // issue the fence. |
| if (do_read_unlock) |
| atomic_thread_fence (memory_order_seq_cst); |
| // TODO Don't just spin but also block using cond vars / futexes |
| // here. Should probably be integrated with the serial lock code. |
| for (gtm_thread *it = gtm_thread::list_of_threads; it != 0; |
| it = it->next_thread) |
| { |
| if (it == this) continue; |
| // We need to load other threads' shared_state using acquire |
| // semantics (matching the release semantics of the respective |
| // updates). This is necessary to ensure that the other |
| // threads' memory accesses happen before our actions that |
| // assume privatization safety. |
| // TODO Are there any platform-specific optimizations (e.g., |
| // merging barriers)? |
| while (it->shared_state.load(memory_order_acquire) < priv_time) |
| cpu_relax(); |
| } |
| } |
| |
| // After ensuring privatization safety, we are now truly inactive and |
| // thus can release the read lock. We will also execute potentially |
| // privatizing actions (e.g., calling free()). User actions are first. |
| if (do_read_unlock) |
| gtm_thread::serial_lock.read_unlock (this); |
| commit_user_actions (); |
| commit_allocations (false, 0); |
| |
| return true; |
| } |
| return false; |
| } |
| |
| void ITM_NORETURN |
| GTM::gtm_thread::restart (gtm_restart_reason r, bool finish_serial_upgrade) |
| { |
| // Roll back to outermost transaction. Do not reset transaction state because |
| // we will continue executing this transaction. |
| rollback (); |
| |
| // If we have to restart while an upgrade of the serial lock is happening, |
| // we need to finish this here, after rollback (to ensure privatization |
| // safety despite undo writes) and before deciding about the retry strategy |
| // (which could switch to/from serial mode). |
| if (finish_serial_upgrade) |
| gtm_thread::serial_lock.write_upgrade_finish(this); |
| |
| decide_retry_strategy (r); |
| |
| // Run dispatch-specific restart code. Retry until we succeed. |
| abi_dispatch* disp = abi_disp(); |
| GTM::gtm_restart_reason rr; |
| while ((rr = disp->begin_or_restart()) != NO_RESTART) |
| { |
| decide_retry_strategy(rr); |
| disp = abi_disp(); |
| } |
| |
| GTM_longjmp (choose_code_path(prop, disp) | a_restoreLiveVariables, |
| &jb, prop); |
| } |
| |
| void ITM_REGPARM |
| _ITM_commitTransaction(void) |
| { |
| #if defined(USE_HTM_FASTPATH) |
| // HTM fastpath. If we are not executing a HW transaction, then we will be |
| // a serial-mode transaction. If we are, then there will be no other |
| // concurrent serial-mode transaction. |
| // See gtm_thread::begin_transaction. |
| if (likely(!gtm_thread::serial_lock.htm_fastpath_disabled())) |
| { |
| htm_commit(); |
| return; |
| } |
| #endif |
| gtm_thread *tx = gtm_thr(); |
| if (!tx->trycommit ()) |
| tx->restart (RESTART_VALIDATE_COMMIT); |
| } |
| |
| void ITM_REGPARM |
| _ITM_commitTransactionEH(void *exc_ptr) |
| { |
| #if defined(USE_HTM_FASTPATH) |
| // See _ITM_commitTransaction. |
| if (likely(!gtm_thread::serial_lock.htm_fastpath_disabled())) |
| { |
| htm_commit(); |
| return; |
| } |
| #endif |
| gtm_thread *tx = gtm_thr(); |
| if (!tx->trycommit ()) |
| { |
| tx->eh_in_flight = exc_ptr; |
| tx->restart (RESTART_VALIDATE_COMMIT); |
| } |
| } |