gpt4 book ai didi

C++ 无锁堆栈已损坏

转载 作者:太空宇宙 更新时间:2023-11-04 13:15:11 25 4
gpt4 key购买 nike

我正在尝试实现一个无锁堆栈,以便与来自有界普通 c 数组的外部托管内存一起使用。我知道引用实现(例如 Anthony Williams:Concurrency in Action)以及网络上的其他书籍和博客/文章。

实现遵循这些引用并避免了 ABA 问题,因为外部内存位置是使用唯一索引而不是循环指针寻址的。因此它根本不需要处理内存管理并且很简单。

我编写了一些测试,在高负载和争用(压力测试)和单线程下对该堆栈执行弹出和推送操作。前者因奇怪的问题而失败,我不理解并且对我来说看起来很晦涩。

也许有人有想法?

  1. 问题:将已弹出的节点推回堆栈失败,因为违反了前提条件,即节点没有后继者(下一个)。

    BOOST_ASSERT(!m_aData.m_aNodes[nNode-1].next);

    复制设置:至少 3 个线程和 ~16 的容量。大约500次通过。然后推送操作失败。

  2. 问题:所有线程弹出的元素数量和加入后留在堆栈中的元素数量与容量不匹配(节点在转换中丢失)。

    BOOST_ASSERT(aNodes.size()+nPopped == nCapacity);

    复制设置:2 个线程和 2 个容量。需要很多遍才能发生,对我来说至少 700 个。在那之后堆栈头为 0,但弹出的容器中只存在一个节点。节点 {2,0} 悬空。

我用 vs2005、vs2013 和 vs2015 编译。都有同样的问题(vs2005也是代码看起来像C++03的原因)。

这里是node+stack的基本代码

template <typename sizeT> struct node
{
sizeT cur; //!< construction invariant
atomic<sizeT> next;
atomic<sizeT> data;

explicit node() // invalid node
: cur(0), next(0), data(0)
{}

explicit node(sizeT const& nCur, sizeT const& nNext, sizeT const& nData)
: cur(nCur), next(nNext), data(nData)
{}

node& operator=(node const& rhs)
{
cur = rhs.cur;
next.store(rhs.next.load(memory_order_relaxed));
data.store(rhs.data.load(memory_order_relaxed));
return *this;
}
};

template <typename sizeT> struct stack
{
private:
static memory_order const relaxed = memory_order_relaxed;
atomic<sizeT> m_aHead;

public:
explicit stack(sizeT const& nHead) : m_aHead(nHead) {}

template <typename tagT, typename T, std::size_t N>
typename enable_if<is_same<tagT,Synchronized>,sizeT>::type
pop(T (&aNodes)[N])
{
sizeT nOldHead = m_aHead.load();

for(;;)
{
if(!nOldHead) return 0;

BOOST_ASSERT(nOldHead <= N);
T& aOldHead = aNodes[nOldHead-1];
sizeT const nNewHead = aOldHead.next.load(/*relaxed*/);
BOOST_ASSERT(nNewHead <= N);
sizeT const nExpected = nOldHead;

if(m_aHead.compare_exchange_weak(nOldHead,nNewHead
/*,std::memory_order_acquire,std::memory_order_relaxed*/))
{
BOOST_ASSERT(nExpected == nOldHead);

// <--- from here on aOldHead is thread local ---> //
aOldHead.next.store(0 /*,relaxed*/);

return nOldHead;
}

// TODO: add back-off strategy under contention (use loop var)
}
}

template <typename tagT, typename T, std::size_t N>
typename enable_if<is_same<tagT,Synchronized>,void>::type
push(T (&aNodes)[N], sizeT const& nNewHead)
{
#ifndef NDEBUG
{
BOOST_ASSERT(0 < nNewHead && nNewHead <= N);
sizeT const nNext = aNodes[nNewHead-1].next;
BOOST_ASSERT(!nNext);
}
#endif

sizeT nOldHead = m_aHead.load(/*relaxed*/);

for(;;)
{
aNodes[nNewHead-1].next.store(nOldHead /*,relaxed*/);
sizeT const nExpected = nOldHead;
BOOST_ASSERT(nOldHead <= N);

if(m_aHead.compare_exchange_weak(nOldHead,nNewHead
/*,std::memory_order_release,std::memory_order_relaxed*/))
{
BOOST_ASSERT(nExpected == nOldHead);
return;
}

// TODO: add back-off strategy under contention (use loop var)
}
}
};

还有相当嘈杂的测试课

class StackTest
{
private:

typedef boost::mpl::size_t<64> Capacity;
//typedef boost::uint_t<static_log2_ceil<Capacity::value>::value>::least size_type;
typedef std::size_t size_type;

static size_type const nCapacity = Capacity::value;
static size_type const nNodes = Capacity::value;

typedef node<size_type> Node;
typedef stack<size_type> Stack;

typedef mt19937 Twister;
typedef random::uniform_int_distribution<std::size_t> Distribution;
typedef variate_generator<Twister,Distribution> Die;

struct Data //!< shared along threads
{
Node m_aNodes[nNodes];
Stack m_aStack;

explicit Data() : m_aStack(nNodes)
{
m_aNodes[0] = Node(1,0,0); // tail of stack

for(size_type i=1; i<nNodes; ++i)
{
m_aNodes[i] = Node(static_cast<size_type>(i+1),i,0);
}
}

template <typename syncT>
void Run(
uuids::random_generator& aUUIDGen,
std::size_t const& nPasses,
std::size_t const& nThreads)
{
std::vector<ThreadLocalData> aThreadLocalDatas(nThreads,ThreadLocalData(*this));

{
static std::size_t const N = 100000;
Die aRepetition(Twister(hash_value(aUUIDGen())),Distribution(0,N));
Die aAction(Twister(hash_value(aUUIDGen())),Distribution(0,1));

for(std::size_t i=0; i<nThreads; ++i)
{
std::vector<bool>& aActions = aThreadLocalDatas[i].m_aActions;
std::size_t const nRepetition = aRepetition();
aActions.reserve(nRepetition);

for(std::size_t k=0; k<nRepetition; ++k)
{
aActions.push_back(static_cast<bool>(aAction()));
}
}
}

std::size_t nPopped = 0;

if(nThreads == 1)
{
std::size_t const i = 0;
aThreadLocalDatas[i].Run<syncT>(i);
nPopped += aThreadLocalDatas[i].m_aPopped.size();
}
else
{
std::vector<boost::shared_ptr<thread> > aThreads;
aThreads.reserve(nThreads);

for(std::size_t i=0; i<nThreads; ++i)
{
aThreads.push_back(boost::make_shared<thread>(boost::bind(&ThreadLocalData::Run<syncT>,&aThreadLocalDatas[i],i)));
}

for(std::size_t i=0; i<nThreads; ++i)
{
aThreads[i]->join();
nPopped += aThreadLocalDatas[i].m_aPopped.size();
}
}

std::vector<size_type> aNodes;
aNodes.reserve(nCapacity);

while(size_type const nNode = m_aStack.pop<syncT>(m_aNodes))
{
aNodes.push_back(nNode);
}

std::clog << dump(m_aNodes,4) << std::endl;

BOOST_ASSERT(aNodes.size()+nPopped == nCapacity);
}
};


struct ThreadLocalData //!< local to each thread
{
Data& m_aData; //!< shared along threads
std::vector<bool> m_aActions; //!< either pop or push
std::vector<size_type> m_aPopped; //!< popp'ed nodes

explicit ThreadLocalData(Data& aData)
: m_aData(aData), m_aActions(), m_aPopped()
{
m_aPopped.reserve(nNodes);
}

template <typename syncT>
void Run(std::size_t const& k)
{
BOOST_FOREACH(bool const& aAction, m_aActions)
{
if(aAction)
{
if(size_type const nNode = m_aData.m_aStack.pop<syncT>(m_aData.m_aNodes))
{
BOOST_ASSERT(!m_aData.m_aNodes[nNode-1].next);
m_aPopped.push_back(nNode);
}
}
else
{
if(!m_aPopped.empty())
{
size_type const nNode = m_aPopped.back();
size_type const nNext = m_aData.m_aNodes[nNode-1].next;
ASSERT_IF(!nNext,"nNext=" << nNext << " for " << m_aData.m_aNodes[nNode-1] << "\n\n" << dump(m_aData.m_aNodes));
m_aData.m_aStack.push<syncT>(m_aData.m_aNodes,nNode);
m_aPopped.pop_back();
}
}
}
}
};


template <typename syncT>
static void PushPop(
uuids::random_generator& aUUIDGen,
std::size_t const& nPasses,
std::size_t const& nThreads)
{
BOOST_ASSERT(nThreads > 0);
BOOST_ASSERT(nThreads == 1 || (is_same<syncT,Synchronized>::value));

std::clog << BOOST_CURRENT_FUNCTION << " with threads=" << nThreads << std::endl;

for(std::size_t nPass=0; nPass<nPasses; ++nPass)
{
std::ostringstream s;
s << " " << nPass << "/" << nPasses << ": ...";
std::clog << s.str() << std::endl;

Data().Run<syncT>(aUUIDGen,nPass,nThreads);
}
}

public:

static void Run()
{
typedef StackTest self_t;

uuids::random_generator aUUIDGen;

static std::size_t const nMaxPasses = 1000;
Die aPasses(Twister(hash_value(aUUIDGen())),Distribution(0,nMaxPasses));

{
//std::size_t const nThreads = 2; // thread::hardware_concurrency()+1;
std::size_t const nThreads = thread::hardware_concurrency()+1;
self_t().PushPop<Synchronized>(aUUIDGen,aPasses(),nThreads);
}
}
};

这是一个link下载所有需要的文件。

最佳答案

这两个问题只是 ABA 问题的另一个方面。

堆栈:{2,1},{1,0}

  1. 线程A
    1. 流行
      new_head=1
      ...超出时间片
  2. 线程B
    1. 流行
      堆栈:{1,0},弹出:{2,0}
    2. 流行
      堆栈:{},弹出:{2,0},{1,0}
    3. 推({2,0})
      堆栈:{2,0}
  3. 线程A
    1. 流行继续
      cmp_exch 成功,因为head是2
      stack: {}, head=1 --- 错误,0 是正确的

任何问题都可能出现,因为对节点的访问不再是线程本地的。这包括对弹出节点(问题 1)或丢失节点(问题 2)的下一个意外修改。

head+next 需要在一个cmp_exch 中修改以避免该问题。

关于C++ 无锁堆栈已损坏,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/37746576/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com