Uses different pass count for different parallel queue test cases
[libcds.git] / cds / urcu / details / gpb.h
index 0128b94dda12ab8c253d2a4d78ef12684e4af47c..5ad75d130d202372af6d13f5847a20a658259fa5 100644 (file)
@@ -1,7 +1,7 @@
 /*
     This file is a part of libcds - Concurrent Data Structures library
 
-    (C) Copyright Maxim Khizhinsky (libcds.dev@gmail.com) 2006-2016
+    (C) Copyright Maxim Khizhinsky (libcds.dev@gmail.com) 2006-2017
 
     Source code repo: http://github.com/khizmax/libcds/
     Download: http://sourceforge.net/projects/libcds/files/
@@ -86,7 +86,9 @@ namespace cds { namespace urcu {
         typedef base_class::thread_gc thread_gc ;   ///< Thread-side RCU part
         typedef typename thread_gc::scoped_lock scoped_lock ; ///< Access lock class
 
-        static bool const c_bBuffered = true ; ///< This RCU buffers disposed elements
+        //@cond
+        static bool const c_bBuffered = true ; ///< Bufferized RCU
+        //@endcond
 
     protected:
         //@cond
@@ -105,7 +107,7 @@ namespace cds { namespace urcu {
         /// Returns singleton instance
         static general_buffered * instance()
         {
-            return static_cast<general_buffered *>( base_class::instance() );
+            return static_cast<general_buffered *>( base_class::instance());
         }
         /// Checks if the singleton is created and ready to use
         static bool isUsed()
@@ -140,17 +142,17 @@ namespace cds { namespace urcu {
                     p.free();
                 }
                 else {
-                    push_buffer( std::move(p) );
+                    push_buffer( std::move(p));
                     break;
                 }
             }
         }
 
-        // Return: true - synchronize has been called, false - otherwise
+        // Return: \p true - synchronize has been called, \p false - otherwise
         bool push_buffer( epoch_retired_ptr&& ep )
         {
             bool bPushed = m_Buffer.push( ep );
-            if ( !bPushed || m_Buffer.size() >= capacity() ) {
+            if ( !bPushed || m_Buffer.size() >= capacity()) {
                 synchronize();
                 if ( !bPushed ) {
                     ep.free();
@@ -175,7 +177,7 @@ namespace cds { namespace urcu {
         /// Destroys singleton object
         static void Destruct( bool bDetachAll = false )
         {
-            if ( isUsed() ) {
+            if ( isUsed()) {
                 instance()->clear_buffer( std::numeric_limits< uint64_t >::max());
                 if ( bDetachAll )
                     instance()->m_ThreadList.detach_all();
@@ -191,7 +193,7 @@ namespace cds { namespace urcu {
             When the buffer becomes full \ref synchronize function is called
             to wait for the end of grace period and then to free all pointers from the buffer.
         */
-        virtual void retire_ptr( retired_ptr& p )
+        virtual void retire_ptr( retired_ptr& p ) override
         {
             if ( p.m_p )
                 push_buffer( epoch_retired_ptr( p, m_nCurEpoch.load( atomics::memory_order_relaxed )));
@@ -205,7 +207,7 @@ namespace cds { namespace urcu {
             while ( itFirst != itLast ) {
                 epoch_retired_ptr ep( *itFirst, nEpoch );
                 ++itFirst;
-                push_buffer( std::move(ep) );
+                push_buffer( std::move(ep));
             }
         }
 
@@ -232,17 +234,15 @@ namespace cds { namespace urcu {
         bool synchronize( epoch_retired_ptr& ep )
         {
             uint64_t nEpoch;
-            atomics::atomic_thread_fence( atomics::memory_order_acquire );
             {
                 std::unique_lock<lock_type> sl( m_Lock );
-                if ( ep.m_p && m_Buffer.push( ep ) )
+                if ( ep.m_p && m_Buffer.push( ep ))
                     return false;
                 nEpoch = m_nCurEpoch.fetch_add( 1, atomics::memory_order_relaxed );
                 flip_and_wait();
                 flip_and_wait();
             }
             clear_buffer( nEpoch );
-            atomics::atomic_thread_fence( atomics::memory_order_release );
             return true;
         }
         //@endcond