gpt4 book ai didi

c++ - 为什么使用 std::multiset 作为优先级队列比使用 std::priority_queue 更快?

转载 作者:可可西里 更新时间:2023-11-01 16:35:15 24 4
gpt4 key购买 nike

我尝试用 std::priority_queue 替换 std::multiset。但我对速度结果感到失望。算法运行时间增加50%...

相应的命令如下:

top() = begin();
pop() = erase(knn.begin());
push() = insert();

我对 priority_queue 的实现速度感到惊讶,我期待不同的结果(对 PQ 更好)...从概念上讲,多重集被用作优先级队列。为什么优先级队列和多重集有如此不同的性能,即使使用 -O2

十个结果的平均值,MSVS 2010,Win XP,32 位,方法 findAllKNN2 ()(请参见下文)

MS
N time [s]
100 000 0.5
1 000 000 8

PQ
N time [s]
100 000 0.8
1 000 000 12

是什么导致了这些结果?没有对源代码进行其他更改...感谢您的帮助...

MS 实现:

template <typename Point>
struct TKDNodePriority
{
KDNode <Point> *node;
typename Point::Type priority;

TKDNodePriority() : node ( NULL ), priority ( 0 ) {}
TKDNodePriority ( KDNode <Point> *node_, typename Point::Type priority_ ) : node ( node_ ), priority ( priority_ ) {}

bool operator < ( const TKDNodePriority <Point> &n1 ) const
{
return priority > n1.priority;
}
};

template <typename Point>
struct TNNeighboursList
{
typedef std::multiset < TKDNodePriority <Point> > Type;
};

方法:

template <typename Point>
template <typename Point2>
void KDTree2D <Point>::findAllKNN2 ( const Point2 * point, typename TNNeighboursList <Point>::Type & knn, unsigned int k, KDNode <Point> *node, const unsigned int depth ) const
{
if ( node == NULL )
{
return;
}

if ( point->getCoordinate ( depth % 2 ) <= node->getData()->getCoordinate ( depth % 2 ) )
{
findAllKNN2 ( point, knn, k, node->getLeft(), depth + 1 );
}

else
{
findAllKNN2 ( point, knn, k, node->getRight(), depth + 1 );
}

typename Point::Type dist_q_node = ( node->getData()->getX() - point->getX() ) * ( node->getData()->getX() - point->getX() ) +
( node->getData()->getY() - point->getY() ) * ( node->getData()->getY() - point->getY() );

if (knn.size() == k)
{
if (dist_q_node < knn.begin()->priority )
{
knn.erase(knn.begin());
knn.insert ( TKDNodePriority <Point> ( node, dist_q_node ) );
}
}

else
{
knn.insert ( TKDNodePriority <Point> ( node, dist_q_node ) );
}

typename Point::Type dist_q_node_straight = ( point->getCoordinate ( node->getDepth() % 2 ) - node->getData()->getCoordinate ( node->getDepth() % 2 ) ) *
( point->getCoordinate ( node->getDepth() % 2 ) - node->getData()->getCoordinate ( node->getDepth() % 2 ) ) ;

typename Point::Type top_priority = knn.begin()->priority;
if ( knn.size() < k || dist_q_node_straight < top_priority )
{
if ( point->getCoordinate ( node->getDepth() % 2 ) < node->getData()->getCoordinate ( node->getDepth() % 2 ) )
{
findAllKNN2 ( point, knn, k, node->getRight(), depth + 1 );
}

else
{
findAllKNN2 ( point, knn, k, node->getLeft(), depth + 1 );
}
}
}

PQ 实现(较慢,为什么?)

template <typename Point>
struct TKDNodePriority
{
KDNode <Point> *node;
typename Point::Type priority;

TKDNodePriority() : node ( NULL ), priority ( 0 ) {}
TKDNodePriority ( KDNode <Point> *node_, typename Point::Type priority_ ) : node ( node_ ), priority ( priority_ ) {}

bool operator < ( const TKDNodePriority <Point> &n1 ) const
{
return priority > n1.priority;
}
};


template <typename Point>
struct TNNeighboursList
{
typedef std::priority_queue< TKDNodePriority <Point> > Type;
};

方法:

template <typename Point>
template <typename Point2>
void KDTree2D <Point>::findAllKNN2 ( const Point2 * point, typename TNNeighboursList <Point>::Type & knn, unsigned int k, KDNode <Point> *node, const unsigned int depth ) const
{

if ( node == NULL )
{
return;
}

if ( point->getCoordinate ( depth % 2 ) <= node->getData()->getCoordinate ( depth % 2 ) )
{
findAllKNN2 ( point, knn, k, node->getLeft(), depth + 1 );
}

else
{
findAllKNN2 ( point, knn, k, node->getRight(), depth + 1 );
}

typename Point::Type dist_q_node = ( node->getData()->getX() - point->getX() ) * ( node->getData()->getX() - point->getX() ) +
( node->getData()->getY() - point->getY() ) * ( node->getData()->getY() - point->getY() );

if (knn.size() == k)
{
if (dist_q_node < knn.top().priority )
{
knn.pop();

knn.push ( TKDNodePriority <Point> ( node, dist_q_node ) );
}
}

else
{
knn.push ( TKDNodePriority <Point> ( node, dist_q_node ) );
}

typename Point::Type dist_q_node_straight = ( point->getCoordinate ( node->getDepth() % 2 ) - node->getData()->getCoordinate ( node->getDepth() % 2 ) ) *
( point->getCoordinate ( node->getDepth() % 2 ) - node->getData()->getCoordinate ( node->getDepth() % 2 ) ) ;

typename Point::Type top_priority = knn.top().priority;
if ( knn.size() < k || dist_q_node_straight < top_priority )
{
if ( point->getCoordinate ( node->getDepth() % 2 ) < node->getData()->getCoordinate ( node->getDepth() % 2 ) )
{
findAllKNN2 ( point, knn, k, node->getRight(), depth + 1 );
}

else
{
findAllKNN2 ( point, knn, k, node->getLeft(), depth + 1 );
}
}
}

最佳答案

首先,作者没有提供导致提到的性能下降的最小代码示例。其次,这个问题是 8 年前提出的,我确信编译器对性能有巨大的提升。

我做了一个基准示例,我在队列中取出第一个元素,然后以另一个优先级推回(模拟插入新元素而不创建一个元素),通过数组 kNodesCount 中的元素计数来做到这一点与 kRunsCount 循环迭代。我正在比较 priority_queuemultisetmultimap .我决定包括 multimap为了更精确的比较。它的简单测试非常接近作者的用例,我也尝试重现他在代码示例中使用的结构。

#include <set>
#include <type_traits>
#include <vector>
#include <chrono>
#include <queue>
#include <map>
#include <iostream>

template<typename T>
struct Point {
static_assert(std::is_integral<T>::value || std::is_floating_point<T>::value, "Incompatible type");
using Type = T;

T x;
T y;
};

template<typename T>
struct Node {
using Type = T;

Node<T> * left;
Node<T> * right;
T data;
};

template <typename T>
struct NodePriority {
using Type = T;
using DataType = typename T::Type;

Node<T> * node = nullptr;
DataType priority = static_cast<DataType>(0);

bool operator < (const NodePriority<T> & n1) const noexcept {
return priority > n1.priority;
}

bool operator > (const NodePriority<T> & n1) const noexcept {
return priority < n1.priority;
}
};

// descending order by default
template <typename T>
using PriorityQueueList = std::priority_queue<T>;

// greater used because of ascending order by default
template <typename T>
using MultisetList = std::multiset<T, std::greater<T>>;

// greater used because of ascending order by default
template <typename T>
using MultimapList = std::multimap<typename T::DataType, T, std::greater<typename T::DataType>>;

struct Inner {
template<template <typename> class C, typename T>
static void Operate(C<T> & list, std::size_t priority);

template<typename T>
static void Operate(PriorityQueueList<T> & list, std::size_t priority) {
if (list.size() % 2 == 0) {
auto el = std::move(list.top());
el.priority = priority;
list.push(std::move(el));
}
else {
list.pop();
}
}

template<typename T>
static void Operate(MultisetList<T> & list, std::size_t priority) {
if (list.size() % 2 == 0) {
auto el = std::move(*list.begin());
el.priority = priority;
list.insert(std::move(el));
}
else {
list.erase(list.begin());
}
}

template<typename T>
static void Operate(MultimapList<T> & list, std::size_t priority) {
if (list.size() % 2 == 0) {
auto el = std::move(*list.begin());
auto & elFirst = const_cast<int&>(el.first);
elFirst = priority;
el.second.priority = priority;
list.insert(std::move(el));
}
else {
list.erase(list.begin());
}
}
};

template<typename T>
void doOperationOnPriorityList(T & list) {
for (std::size_t pos = 0, len = list.size(); pos < len; ++pos) {
// move top element and update priority
auto priority = std::rand() % 10;
Inner::Operate(list, priority);
}
}

template<typename T>
void measureOperationTime(T & list, std::size_t runsCount) {
std::chrono::system_clock::time_point t1, t2;
std::uint64_t totalTime(0);
for (std::size_t i = 0; i < runsCount; ++i) {
t1 = std::chrono::system_clock::now();
doOperationOnPriorityList(list);
t2 = std::chrono::system_clock::now();
auto castedTime = std::chrono::duration_cast<std::chrono::milliseconds>(t2 - t1).count();
std::cout << "Run " << i << " time: " << castedTime << "\n";
totalTime += castedTime;
}

std::cout << "Average time is: " << totalTime / runsCount << " ms" << std::endl;
}

int main() {
// consts
const int kNodesCount = 10'000'000;
const int kRunsCount = 10;

// prepare data
PriorityQueueList<NodePriority<Point<int>>> neighboursList1;
MultisetList<NodePriority<Point<int>>> neighboursList2;
MultimapList<NodePriority<Point<int>>> neighboursList3;
std::vector<Node<Point<int>>> nodes;
nodes.reserve(kNodesCount);
for (auto i = 0; i < kNodesCount; ++i) {
nodes.emplace_back(decltype(nodes)::value_type{ nullptr, nullptr, { 0,0 } });
auto priority = std::rand() % 10;
neighboursList1.emplace(decltype(neighboursList1)::value_type{ &nodes.back(), priority });
neighboursList2.emplace(decltype(neighboursList2)::value_type{ &nodes.back(), priority });
neighboursList3.emplace(decltype(neighboursList3)::value_type{ priority, { &nodes.back(), priority } });
}

// do operation on data
std::cout << "\nPriority queue\n";
measureOperationTime(neighboursList1, kRunsCount);
std::cout << "\nMultiset\n";
measureOperationTime(neighboursList2, kRunsCount);
std::cout << "\nMultimap\n";
measureOperationTime(neighboursList3, kRunsCount);

return 0;
}

我已经使用 VS v15.8.9 使用/Ox 进行了发布构建。查看 10'000'000 个项目在 10 次运行中的结果:

Priority queue
Run 0 time: 764
Run 1 time: 933
Run 2 time: 920
Run 3 time: 813
Run 4 time: 991
Run 5 time: 862
Run 6 time: 902
Run 7 time: 1277
Run 8 time: 774
Run 9 time: 771
Average time is: 900 ms

Multiset
Run 0 time: 2235
Run 1 time: 1811
Run 2 time: 1755
Run 3 time: 1535
Run 4 time: 1475
Run 5 time: 1388
Run 6 time: 1482
Run 7 time: 1431
Run 8 time: 1347
Run 9 time: 1347
Average time is: 1580 ms

Multimap
Run 0 time: 2197
Run 1 time: 1885
Run 2 time: 1725
Run 3 time: 1671
Run 4 time: 1500
Run 5 time: 1403
Run 6 time: 1411
Run 7 time: 1420
Run 8 time: 1409
Run 9 time: 1362
Average time is: 1598 ms

嗯哼,如你所见multisetmultimap 的性能相同和 priority_queue是最快的(大约快 43%)。那为什么会这样呢?

让我们从priority_queue开始, C++ 标准并没有告诉我们如何实现一个或另一个容器或结构,但在大多数情况下它是基于一个 binary heap (寻找 msvc 和 gcc 实现)!如果是priority_queue您无法访问除顶部之外的任何元素,您无法遍历它们,无法通过索引获取,甚至无法获取最后一个元素(它为优化留出了一些空间)。二叉堆的平均插入是 O(1),只有最坏的情况是 O(log n),删除是 O(log n),因为我们从底部取出元素然后搜索下一个高优先级。

multimap呢?和 multiset .它们通常都在 red-black binary tree 上实现(寻找 msvc 和 gcc 实现),其中平均插入是 O(log n) 和删除 O(log n)。

从这个角度来看priority_queue NEVER 可以比 multiset 慢或 multimap .所以,回到你的问题,multiset因为优先级队列priority_queue快本身。可能有很多原因,包括原始 priority_queue在旧编译器上实现或错误使用此结构(问题不包含最小可行示例),除了作者没有提到编译标志或编译器版本之外,有时优化会产生重大变化。

UPDATE 1 根据@noɥʇʎudʎzɐɹƆ 请求

不幸的是,我现在无法访问 linux 环境,但我安装了 mingw-w64,版本信息:g++.exe(x86_64-posix-seh,由 strawberryperl.com 项目构建)8.3.0。使用的处理器与 visual studio 相同:处理器 Intel(R) Core(TM) i7-8550U CPU @ 1.80GHz,2001 Mhz,4 核,8 逻辑处理器。

因此 g++ -O2 的结果是:

Priority queue
Run 0 time: 775
Run 1 time: 995
Run 2 time: 901
Run 3 time: 807
Run 4 time: 930
Run 5 time: 765
Run 6 time: 799
Run 7 time: 1151
Run 8 time: 760
Run 9 time: 780
Average time is: 866 ms

Multiset
Run 0 time: 2280
Run 1 time: 1942
Run 2 time: 1607
Run 3 time: 1344
Run 4 time: 1319
Run 5 time: 1210
Run 6 time: 1129
Run 7 time: 1156
Run 8 time: 1244
Run 9 time: 992
Average time is: 1422 ms

Multimap
Run 0 time: 2530
Run 1 time: 1958
Run 2 time: 1670
Run 3 time: 1390
Run 4 time: 1391
Run 5 time: 1235
Run 6 time: 1088
Run 7 time: 1198
Run 8 time: 1071
Run 9 time: 963
Average time is: 1449 ms

您可能会注意到它与 msvc 的图片几乎相同。

更新 2 感谢@JorgeBellon

A quick-bench.com在线benchmark链接,自己查!

希望看到我的帖子有任何补充,干杯!

关于c++ - 为什么使用 std::multiset 作为优先级队列比使用 std::priority_queue 更快?,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/5895792/

24 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com