简体   繁体   English

Python 中多处理返回值的快速整形

[英]Fast shaping of multiprocessing return values in Python

I have a function with list valued return values that I'm multiprocessing in Python and I need to concatenate them to 1D lists at the end.我有一个 function 和列表值返回值,我在 Python 中进行多重处理,我需要在最后将它们连接到一维列表。 The following is a sample code for demonstration:以下是演示示例代码:

import numpy as np
import multiprocessing as mp
import random as rd

N = 4
L = list(range(0, N))

def F(x):
    a = []
    b = []
    for t in range(0,2):
        a.append('a'+str(t*x))
        b.append('b'+str(t*x))
    return a, b

pool = mp.Pool(mp.cpu_count())
a,b = zip(*pool.map(F, L))
pool.close()

print(a)
print(b)

A = np.concatenate(a)
B = np.concatenate(b)
              
print(A)
print(B)

The output for illustration is:用于说明的 output 是:

(['a0', 'a0'], ['a0', 'a1'], ['a0', 'a2'], ['a0', 'a3'])
(['b0', 'b0'], ['b0', 'b1'], ['b0', 'b2'], ['b0', 'b3'])
['a0' 'a0' 'a0' 'a1' 'a0' 'a2' 'a0' 'a3']
['b0' 'b0' 'b0' 'b1' 'b0' 'b2' 'b0' 'b3']

The problem is that the list L that I'm processing is pretty huge and that the concatenations at the end take a huge amount of time which minimizes the advantage over serial processing considerably.问题是我正在处理的列表L非常大,并且最后的连接需要大量时间,这大大减少了串行处理的优势。

Is there some clever way to avoid the concatenation or alternatively a faster method to perform the concatenation?是否有一些巧妙的方法可以避免串联,或者有更快的方法来执行串联? I've been fiddling with queues but this seems kind of very slow.我一直在摆弄队列,但这似乎有点慢。

Note: This seems to be a similar question as Add result from multiprocessing into array .注意:这似乎是一个与Add result from multiprocessing into array类似的问题。

If the desired output is an input suitable for creating a scipy.sparse.coo_matrix , I would take a very different approach: Don't return anything, just create shared objects that can be modified directly.如果所需的 output 是适合创建scipy.sparse.coo_matrix的输入,我会采取非常不同的方法:不返回任何内容,只需创建可以直接修改的共享对象。

What you need to create a coo_matrix is an array of the data values, an array of the data rows, and an array of the data columns (unless you already have another sparse matrix / dense matrix).创建coo_matrix需要的是数据值数组、数据行数组和数据列数组(除非您已经有另一个稀疏矩阵/密集矩阵)。 I would create 3 shared arrays that each process can dump results directly into using the index of each entry from L .我将创建 3 个共享的 arrays ,每个进程都可以使用L中的每个条目的索引将结果直接转储到其中。 This even allows out of order execution, so you can use imap_unordered instead for better speed:这甚至允许乱序执行,因此您可以使用imap_unordered来提高速度:

from multiprocessing.pool import Pool
from multiprocessing.sharedctypes import RawArray
from random import random, randint # bogus data for testing
import numpy as np

from ctypes import c_int, c_float
from scipy.sparse import coo_matrix

#pool worker globals are only global to that process
worker_globals = {}

def init_worker(data_array, row_array, col_array):
    worker_globals['data'] = np.frombuffer(data_array, dtype=c_float)
    worker_globals['row'] = np.frombuffer(row_array, dtype=c_int)
    worker_globals['col'] = np.frombuffer(col_array, dtype=c_int)

def worker_func(tup):
    i, x = tup #enumerate returns a tuple with the index then the value
    #don't bother with mutexes because we only ever write to array[i] once from a single process
    worker_globals['data'][i] = random() #calculate your data, row, and column, and write back to the shared arrays
    worker_globals['row'][i] = x
    worker_globals['col'][i] = randint(0,1000)

if __name__ == "__main__":
    L = list(range(100, 0, -1)) #some data in L
    
    data_array = RawArray(c_float, len(L))
    row_array = RawArray(c_int, len(L))
    col_array = RawArray(c_int, len(L))
    
    with Pool(initializer=init_worker, initargs=(data_array, row_array, col_array)) as p:
        for _ in p.imap_unordered(worker_func, enumerate(L)):
            pass
    d = np.frombuffer(data_array, dtype=c_float)
    r = np.frombuffer(row_array, dtype=c_int)
    c = np.frombuffer(col_array, dtype=c_int)
    mat = coo_matrix((d, (r, c)), shape=(1000, 1000))

By the way: You should also absolutely always be using if __name__ == "__main__": when using multiprocessing.顺便说一句:在使用多处理时,您也绝对应该始终使用if __name__ == "__main__": It is suggested everywhere, and required on windows.它在任何地方都被建议,并且在 windows 上需要。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM