projects
/
folly.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Remove InlineExecutor.cpp
[folly.git]
/
folly
/
ProducerConsumerQueue.h
diff --git
a/folly/ProducerConsumerQueue.h
b/folly/ProducerConsumerQueue.h
index ff1a4fe13b3dbe35903c4adf5dee3ac0a85f10d4..9ebe65acb46c66dbef1f9ab33e77e07fd968bdeb 100644
(file)
--- a/
folly/ProducerConsumerQueue.h
+++ b/
folly/ProducerConsumerQueue.h
@@
-1,5
+1,5
@@
/*
/*
- * Copyright 201
4
Facebook, Inc.
+ * Copyright 201
7
Facebook, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
@@
-17,17
+17,17
@@
// @author Bo Hu (bhu@fb.com)
// @author Jordan DeLong (delong.j@fb.com)
// @author Bo Hu (bhu@fb.com)
// @author Jordan DeLong (delong.j@fb.com)
-#ifndef PRODUCER_CONSUMER_QUEUE_H_
-#define PRODUCER_CONSUMER_QUEUE_H_
+#pragma once
#include <atomic>
#include <cassert>
#include <cstdlib>
#include <atomic>
#include <cassert>
#include <cstdlib>
+#include <memory>
#include <stdexcept>
#include <type_traits>
#include <utility>
#include <stdexcept>
#include <type_traits>
#include <utility>
-#include <boost/noncopyable.hpp>
-#include <
boost/type_traits.hpp
>
+
+#include <
folly/detail/CacheLocality.h
>
namespace folly {
namespace folly {
@@
-36,9
+36,12
@@
namespace folly {
* without locks.
*/
template<class T>
* without locks.
*/
template<class T>
-struct ProducerConsumerQueue
: private boost::noncopyable
{
+struct ProducerConsumerQueue {
typedef T value_type;
typedef T value_type;
+ ProducerConsumerQueue(const ProducerConsumerQueue&) = delete;
+ ProducerConsumerQueue& operator = (const ProducerConsumerQueue&) = delete;
+
// size must be >= 2.
//
// Also, note that the number of usable slots in the queue at any
// size must be >= 2.
//
// Also, note that the number of usable slots in the queue at any
@@
-60,9
+63,9
@@
struct ProducerConsumerQueue : private boost::noncopyable {
// We need to destruct anything that may still exist in our queue.
// (No real synchronization needed at destructor time: only one
// thread can be doing this.)
// We need to destruct anything that may still exist in our queue.
// (No real synchronization needed at destructor time: only one
// thread can be doing this.)
- if (!
boost::has_trivial_destructor
<T>::value) {
-
in
t read = readIndex_;
-
in
t end = writeIndex_;
+ if (!
std::is_trivially_destructible
<T>::value) {
+
size_
t read = readIndex_;
+
size_
t end = writeIndex_;
while (read != end) {
records_[read].~T();
if (++read == size_) {
while (read != end) {
records_[read].~T();
if (++read == size_) {
@@
-134,16
+137,16
@@
struct ProducerConsumerQueue : private boost::noncopyable {
}
bool isEmpty() const {
}
bool isEmpty() const {
-
return readIndex_.load(std::memory_order_consum
e) ==
-
writeIndex_.load(std::memory_order_consum
e);
+
return readIndex_.load(std::memory_order_acquir
e) ==
+
writeIndex_.load(std::memory_order_acquir
e);
}
bool isFull() const {
}
bool isFull() const {
- auto nextRecord = writeIndex_.load(std::memory_order_
consum
e) + 1;
+ auto nextRecord = writeIndex_.load(std::memory_order_
acquir
e) + 1;
if (nextRecord == size_) {
nextRecord = 0;
}
if (nextRecord == size_) {
nextRecord = 0;
}
- if (nextRecord != readIndex_.load(std::memory_order_
consum
e)) {
+ if (nextRecord != readIndex_.load(std::memory_order_
acquir
e)) {
return false;
}
// queue is full
return false;
}
// queue is full
@@
-156,8
+159,8
@@
struct ProducerConsumerQueue : private boost::noncopyable {
// be removing items concurrently).
// * It is undefined to call this from any other thread.
size_t sizeGuess() const {
// be removing items concurrently).
// * It is undefined to call this from any other thread.
size_t sizeGuess() const {
- int ret = writeIndex_.load(std::memory_order_
consum
e) -
-
readIndex_.load(std::memory_order_consum
e);
+ int ret = writeIndex_.load(std::memory_order_
acquir
e) -
+
readIndex_.load(std::memory_order_acquir
e);
if (ret < 0) {
ret += size_;
}
if (ret < 0) {
ret += size_;
}
@@
-165,13
+168,14
@@
struct ProducerConsumerQueue : private boost::noncopyable {
}
private:
}
private:
+ char pad0_[detail::CacheLocality::kFalseSharingRange];
const uint32_t size_;
T* const records_;
const uint32_t size_;
T* const records_;
- std::atomic<int> readIndex_;
- std::atomic<int> writeIndex_;
+ FOLLY_ALIGN_TO_AVOID_FALSE_SHARING std::atomic<unsigned int> readIndex_;
+ FOLLY_ALIGN_TO_AVOID_FALSE_SHARING std::atomic<unsigned int> writeIndex_;
+
+ char pad1_[detail::CacheLocality::kFalseSharingRange - sizeof(writeIndex_)];
};
}
};
}
-
-#endif