简体   繁体   English

POST API 请求如何在 python 中使用并行处理? requests.exceptions.ConnectionError:

[英]how do POST API requests use parallel processing in python? requests.exceptions.ConnectionError:

I have code like this:我有这样的代码:

import requests
import multiprocessing as mp
import json
import time

BASE_URL = 'http://127.0.0.1:3000/employees'

with open('data.json', 'r') as f:
    array = json.load(f)

def internet_resource_getter(post_data):
    stuff_got = []
    response = requests.post(BASE_URL, json=post_data)
    stuff_got.append(response.json())
    print(stuff_got)
    time.sleep(1)
    return stuff_got

if __name__ == '__main__':
    # freeze_support() here if program needs to be frozen  
    start=time.time()
    with mp.Pool(mp.cpu_count()) as pool:
        pool.map(internet_resource_getter, array)
    elapsed = (time.time() - start)  
    print("\n","time elapsed is :", elapsed)

in file data.json contains 500 records, for example:在文件 data.json 中包含 500 条记录,例如:

[{"first_name":"John","last_name":"Swen"},{"first_name":"Ricard","last_name":"Candra"}]

in BASE_URL there is data like this:在 BASE_URL 中有这样的数据:

[
  {
    "id": 1,
    "first_name": "Sebastian",
    "last_name": "Eschweiler"
  },
  {
    "id": 2,
    "first_name": "Steve",
    "last_name": "Palmer"
  },
  {
    "id": 3,
    "first_name": "Ann",
    "last_name": "Smith"
  }
]

expected output after POST API: POST API 后预期的 output:

[
  {
    "id": 1,
    "first_name": "Sebastian",
    "last_name": "Eschweiler"
  },
  {
    "id": 2,
    "first_name": "Steve",
    "last_name": "Palmer"
  },
  {
    "id": 3,
    "first_name": "Ann",
    "last_name": "Smith"
  },
 {
    "id": 4,
    "first_name": "John",
    "last_name": "Swen"
  },
{
    "id": 5,
    "first_name": "Ricard",
    "last_name": "Candra"
  },
]

with my code above, the data that enters the url is only 420 records, even though my data.json is 500 records.使用上面的代码,输入 url 的数据只有 420 条记录,即使我的 data.json 是 500 条记录。 how do I solve this so that I post 500 records to url.我该如何解决这个问题,以便将 500 条记录发布到 url。 I don't know why only 400 data are processed.我不知道为什么只处理了 400 个数据。 I have an error like this:我有这样的错误:

multiprocessing.pool.RemoteTraceback:
"""
Traceback (most recent call last):
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 699, in urlopen
    httplib_response = self._make_request(
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 445, in _make_request
    six.raise_from(e, None)
  File "<string>", line 3, in raise_from
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 440, in _make_request
    httplib_response = conn.getresponse()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 1347, in getresponse
    response.begin()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 307, in begin
    version, status, reason = self._read_status()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 276, in _read_status
    raise RemoteDisconnected("Remote end closed connection without"
http.client.RemoteDisconnected: Remote end closed connection without response

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\requests\adapters.py", line 439, in send
    resp = conn.urlopen(
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 755, in urlopen
    retries = retries.increment(
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\util\retry.py", line 531, in increment
    raise six.reraise(type(error), error, _stacktrace)
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\packages\six.py", line 734, in reraise
    raise value.with_traceback(tb)
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 699, in urlopen
    httplib_response = self._make_request(
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 445, in _make_request
    six.raise_from(e, None)
  File "<string>", line 3, in raise_from
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\site-packages\urllib3\connectionpool.py", line 440, in _make_request
    httplib_response = conn.getresponse()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 1347, in getresponse
    response.begin()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 307, in begin
    version, status, reason = self._read_status()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\http\client.py", line 276, in _read_status
    raise RemoteDisconnected("Remote end closed connection without"
urllib3.exceptions.ProtocolError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "paralel_pro.py", line 28, in <module>
    pool.map(internet_resource_getter, array)
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\multiprocessing\pool.py", line 364, in map
    return self._map_async(func, iterable, mapstar, chunksize).get()
  File "C:\Users\admin\AppData\Local\Programs\Python\Python38\lib\multiprocessing\pool.py", line 771, in get
    raise self._value
requests.exceptions.ConnectionError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))

To get your answer, check your server logs.要获得答案,请检查您的服务器日志。 I suspect bad data and/or server not handling unexpected data gracefully.我怀疑坏数据和/或服务器没有优雅地处理意外数据。

Client-Side Check客户端检查

See if the 80 records are always the same 80 records.看看这 80 条记录是否总是相同的 80 条记录。

Server-Side Check服务器端检查

Can you find the logs for your server that is running on http://127.0.0.1:3000/employees ?您能找到在http://127.0.0.1:3000/employees上运行的服务器的日志吗? It looks like it is running on your same machine.看起来它正在您的同一台机器上运行。 The server's application or service logs should have the smoking gun.服务器的应用程序或服务日志应该有确凿的证据。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 requests.exceptions.ConnectionError Python - requests.exceptions.ConnectionError Python Python:requests.exceptions.ConnectionError:HTTPSConnectionPool - Python: requests.exceptions.ConnectionError: HTTPSConnectionPool 如何解决获取请求中的 requests.exceptions.ConnectionError - Python - how to solve requests.exceptions.ConnectionError in get request - Python 意外的请求。异常。连接错误 - Unexpected requests.exceptions.ConnectionError requests.exceptions.ConnectionError:HTTPConnectionPool - requests.exceptions.ConnectionError: HTTPConnectionPool python - 处理连接异常(requests.exceptions.ConnectionError) - python - handle connection exception (requests.exceptions.ConnectionError) 如何避免 requests.exceptions.ConnectionError 打印 url 文本? - How to avoid requests.exceptions.ConnectionError for printing url text? 如何从requests.exceptions.ConnectionError的异常中获取错误号和消息 - How do I get Error Number and Message from exception from requests.exceptions.ConnectionError 如何解决 requests.exceptions.ConnectionError: ('Connection aborted.') in python web 抓取? - How to solve requests.exceptions.ConnectionError: ('Connection aborted.') in python web scraping? requests.exceptions.ConnectionError [Errno 11004] getaddrinfo在Python中使用请求失败。 浏览器工作 - requests.exceptions.ConnectionError [Errno 11004] getaddrinfo failed using Requests in Python. Browser works
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM