gpt4 book ai didi

c++ - 性能损失并行

转载 作者:塔克拉玛干 更新时间:2023-11-03 00:32:43 25 4
gpt4 key购买 nike

我有一个程序或多或少地重复了一些 vector 运算。当我尝试使用 parallel_for 并行执行相同的任务时,我观察到每个任务的时间显着增加。每个任务都读取相同的数据,并且没有进行同步。这是示例代码(它需要任务流库(https://github.com/cpp-taskflow/cpp-taskflow)):

#include <array>
#include <numeric>
#include <x86intrin.h>
#include "taskflow.hpp"

//#define USE_AVX_512 1
constexpr size_t Size = 5000;
struct alignas(64) Vec : public std::array<double, Size> {};

struct SimulationData
{
Vec a_;
Vec b_;
Vec c_;

SimulationData()
{
std::iota(a_.begin(), a_.end(), 10);
std::iota(b_.begin(), b_.end(), 5);
std::iota(c_.begin(), c_.end(), 0);
}
};
struct SimulationTask
{
const SimulationData& data_;
double res_;
double time_;
explicit SimulationTask(const SimulationData& data)
: data_(data), res_(0.0), time_(0.0)
{}

constexpr static int blockSize = 20000;
void sample()
{
auto tbeg = std::chrono::steady_clock::now();
Vec result;
for(auto i=0; i < blockSize; ++i)
{
add(result.data(), data_.a_.data(), data_.b_.data(), Size);
mul(result.data(), result.data(), data_.c_.data(), Size);
res_ += *std::max_element(result.begin(), result.end());
}
auto tend = std::chrono::steady_clock::now();
time_ = std::chrono::duration_cast<std::chrono::milliseconds>(tend-tbeg).count();
}
inline double getResults() const
{
return res_;
}
inline double getTime() const
{
return time_;
}
static void add( double* result, const double* a, const double* b, size_t size)
{
size_t i = 0;
// AVX-512 loop
#ifdef USE_AVX_512
for( ; i < (size & ~0x7); i += 8)
{
const __m512d kA8 = _mm512_load_pd( &a[i] );
const __m512d kB8 = _mm512_load_pd( &b[i] );

const __m512d kRes = _mm512_add_pd( kA8, kB8 );
_mm512_stream_pd( &result[i], kRes );
}
#endif
// AVX loop
for ( ; i < (size & ~0x3); i += 4 )
{
const __m256d kA4 = _mm256_load_pd( &a[i] );
const __m256d kB4 = _mm256_load_pd( &b[i] );

const __m256d kRes = _mm256_add_pd( kA4, kB4 );
_mm256_stream_pd( &result[i], kRes );
}

// SSE2 loop
for ( ; i < (size & ~0x1); i += 2 )
{
const __m128d kA2 = _mm_load_pd( &a[i] );
const __m128d kB2 = _mm_load_pd( &b[i] );

const __m128d kRes = _mm_add_pd( kA2, kB2 );
_mm_stream_pd( &result[i], kRes );
}

// Serial loop
for( ; i < size; i++ )
{
result[i] = a[i] + b[i];
}
}
static void mul( double* result, const double* a, const double* b, size_t size)
{
size_t i = 0;
// AVX-512 loop
#ifdef USE_AVX_512
for( ; i < (size & ~0x7); i += 8)
{
const __m512d kA8 = _mm512_load_pd( &a[i] );
const __m512d kB8 = _mm512_load_pd( &b[i] );

const __m512d kRes = _mm512_mul_pd( kA8, kB8 );
_mm512_stream_pd( &result[i], kRes );
}
#endif
// AVX loop
for ( ; i < (size & ~0x3); i += 4 )
{
const __m256d kA4 = _mm256_load_pd( &a[i] );
const __m256d kB4 = _mm256_load_pd( &b[i] );

const __m256d kRes = _mm256_mul_pd( kA4, kB4 );
_mm256_stream_pd( &result[i], kRes );
}

// SSE2 loop
for ( ; i < (size & ~0x1); i += 2 )
{
const __m128d kA2 = _mm_load_pd( &a[i] );
const __m128d kB2 = _mm_load_pd( &b[i] );

const __m128d kRes = _mm_mul_pd( kA2, kB2 );
_mm_stream_pd( &result[i], kRes );
}

// Serial loop
for( ; i < size; i++ )
{
result[i] = a[i] * b[i];
}
}
};

int main(int argc, const char* argv[])
{
int numOfThreads = 1;
if ( argc > 1 )
numOfThreads = atoi( argv[1] );

try
{
SimulationData data;
std::vector<SimulationTask> tasks;
for (int i = 0; i < numOfThreads; ++i)
tasks.emplace_back(data);

tf::Taskflow tf;
tf.parallel_for(tasks, [](auto &task) { task.sample(); });
tf.wait_for_all();
for (const auto &task : tasks)
{
std::cout << "Result: " << task.getResults() << ", Time: " << task.getTime() << std::endl;
}
}
catch (const std::exception& ex)
{
std::cerr << ex.what() << std::endl;
}
return 0;
}

我在双 E5-2697 v2(每个 CPU 有 12 个物理具有超线程的内核,因此有 48 个硬件线程可用)。当我增加并行任务的数量时,每个任务的时间都会增加很多:

# ./timing 1
Result: 1.0011e+12, Time: 618

使用 12 个任务:

# ./timing 12
Result: 1.0011e+12, Time: 788
Result: 1.0011e+12, Time: 609
Result: 1.0011e+12, Time: 812
Result: 1.0011e+12, Time: 605
Result: 1.0011e+12, Time: 808
Result: 1.0011e+12, Time: 1050
Result: 1.0011e+12, Time: 817
Result: 1.0011e+12, Time: 830
Result: 1.0011e+12, Time: 597
Result: 1.0011e+12, Time: 573
Result: 1.0011e+12, Time: 586
Result: 1.0011e+12, Time: 583

使用 24 个任务:

# ./timing 24
Result: 1.0011e+12, Time: 762
Result: 1.0011e+12, Time: 1033
Result: 1.0011e+12, Time: 735
Result: 1.0011e+12, Time: 1051
Result: 1.0011e+12, Time: 1060
Result: 1.0011e+12, Time: 757
Result: 1.0011e+12, Time: 1075
Result: 1.0011e+12, Time: 758
Result: 1.0011e+12, Time: 745
Result: 1.0011e+12, Time: 1165
Result: 1.0011e+12, Time: 1032
Result: 1.0011e+12, Time: 1160
Result: 1.0011e+12, Time: 757
Result: 1.0011e+12, Time: 743
Result: 1.0011e+12, Time: 736
Result: 1.0011e+12, Time: 1028
Result: 1.0011e+12, Time: 1109
Result: 1.0011e+12, Time: 1018
Result: 1.0011e+12, Time: 1338
Result: 1.0011e+12, Time: 743
Result: 1.0011e+12, Time: 1061
Result: 1.0011e+12, Time: 1046
Result: 1.0011e+12, Time: 1341
Result: 1.0011e+12, Time: 761

使用 48 个任务:

# ./timing 48
Result: 1.0011e+12, Time: 1591
Result: 1.0011e+12, Time: 1776
Result: 1.0011e+12, Time: 1923
Result: 1.0011e+12, Time: 1876
Result: 1.0011e+12, Time: 2002
Result: 1.0011e+12, Time: 1649
Result: 1.0011e+12, Time: 1955
Result: 1.0011e+12, Time: 1728
Result: 1.0011e+12, Time: 1632
Result: 1.0011e+12, Time: 1418
Result: 1.0011e+12, Time: 1904
Result: 1.0011e+12, Time: 1847
Result: 1.0011e+12, Time: 1595
Result: 1.0011e+12, Time: 1910
Result: 1.0011e+12, Time: 1530
Result: 1.0011e+12, Time: 1824
Result: 1.0011e+12, Time: 1588
Result: 1.0011e+12, Time: 1656
Result: 1.0011e+12, Time: 1876
Result: 1.0011e+12, Time: 1683
Result: 1.0011e+12, Time: 1403
Result: 1.0011e+12, Time: 1730
Result: 1.0011e+12, Time: 1476
Result: 1.0011e+12, Time: 1938
Result: 1.0011e+12, Time: 1429
Result: 1.0011e+12, Time: 1888
Result: 1.0011e+12, Time: 1530
Result: 1.0011e+12, Time: 1754
Result: 1.0011e+12, Time: 1794
Result: 1.0011e+12, Time: 1935
Result: 1.0011e+12, Time: 1757
Result: 1.0011e+12, Time: 1572
Result: 1.0011e+12, Time: 1474
Result: 1.0011e+12, Time: 1609
Result: 1.0011e+12, Time: 1394
Result: 1.0011e+12, Time: 1655
Result: 1.0011e+12, Time: 1480
Result: 1.0011e+12, Time: 2061
Result: 1.0011e+12, Time: 2056
Result: 1.0011e+12, Time: 1598
Result: 1.0011e+12, Time: 1630
Result: 1.0011e+12, Time: 1623
Result: 1.0011e+12, Time: 2073
Result: 1.0011e+12, Time: 1395
Result: 1.0011e+12, Time: 1487
Result: 1.0011e+12, Time: 1854
Result: 1.0011e+12, Time: 1569
Result: 1.0011e+12, Time: 1530

这段代码有问题吗?矢量化是 parallel_for 的问题吗?我能否使用 perf 或类似工具获得更好的洞察力?

最佳答案

超线程之所以存在,是因为线程(在现实世界场景中)经常需要等待内存中的数据,从而使物理核心在数据传输过程中基本上处于空闲状态。您的示例(以及 CPU,例如通过预取)正在努力避免这种内存限制,因此通过使线程数量饱和,同一内核上的任何两个超线程都在竞争其 execution ports。 .请注意,在您的 CPU 上,每个核心周期只有 3 个整数 vector ALU 可用 - 调度程序可能会让它们都忙于一个线程的操作。

对于 1 个线程或 12 个线程,您不会真正遇到这种争用。对于 24 个线程,如果每个线程都被调度到其自己的物理内核,您将只能避免此问题,这可能不会发生(因此您开始看到更糟糕的计时)。使用 48 个内核,您肯定会遇到上述问题。

正如 harold 提到的,您可能还受存储限制(超线程对竞争的另一种资源)。

关于c++ - 性能损失并行,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/52165572/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com