0
我有一个函数从维基媒体API获取文章的文本怪异的行为运行的代码通过功能
def get_content(lang, title):
url = "https://"+ lang +".wikipedia.org/w/api.php"
params={'action':'query', 'format':'json', 'titles': title, 'prop':
'linkshere','redirects':'true','prop':'extracts','exlimit':'max','explaintext':'true'}
response = requests.get(url=url,params=params).json()
content = response["query"]["pages"]
content = six.next(six.itervalues(content))['extract']
return content
但是当我调用该函数,出现错误:
Traceback (most recent call last):
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/adapters.py", line 423, in send
timeout=timeout
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/packages/urllib3/connectionpool.py", line 640,
in urlopen
_stacktrace=sys.exc_info()[2])
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/packages/urllib3/util/retry.py", line 287, in i
ncrement
raise MaxRetryError(_pool, url, error or ResponseError(cause))
requests.packages.urllib3.exceptions.MaxRetryError:
HTTPSConnectionPool(host='xn--mesut%20zil-57a.wikipedia.org',
port=443): Max retries exceeded with url: /w/api.php?
exlimit=max&titles=en&action=query&format=json&prop=extract
s&redirects=true&explaintext=true (Caused by
NewConnectionError('<requests.packages.urllib3.connection.VerifiedHT
TPSConnection object at 0x7f7a5e74c0b8>: Failed to establish a new
connection: [Errno -2] Name or service not kno
wn',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "get_data.py", line 165, in <module>
print(get_content('Mesut Özil','en'))
File "get_data.py", line 147, in get_content
response = request_get_data(url=url,params=params).json()
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/home/klux/anaconda3/lib/python3.5/site-
packages/requests/adapters.py", line 487, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='xn--
mesut%20zil-57a.wikipedia.org', port=443): Max
retries exceeded with url: /w/api.php?
exlimit=max&titles=en&action=query&format=json
&prop=extracts&redirects=true&explaintext=true (Caused by
NewConnectionError
('<requests.packages.urllib3.connection.VerifiedHTTPSConnection o
bject at 0x7f7a5e74c0b8>: Failed to establish a new connection: [Errno
-2] Name or service not known',))
首先,我以为我是由维基媒体呼叫太多的请求被阻止,但是当我在“主功能”中运行该代码时,它完全像这样工作。
if __name__ == '__main__':
lang = 'en'
title = 'Mesut Özil'
url = "https://"+ lang +".wikipedia.org/w/api.php"
params={'action':'query', 'format':'json', 'titles': title, 'prop': 'linkshere',
'redirects':'true','prop':'extracts','exlimit':'max','explaintext':'true'}
response = requests.get(url=url, params=params).json()
content = response["query"]["pages"]
content = six.next(six.itervalues(content))['extract']
print(content)
输出:
Mesut Özil (German pronunciation: [ˈmeːzut ˈøːzil], Turkish: [meˈsut
ˈøzil]; born 15 October 1988) is a German fo
otballer who plays for English club Arsenal and the German national
team...
我不知道如何解决这个怪异的行为,我试图通过Visual Studio代码和response
变量get_content
功能调试转向undefined
。那么有人对这种情况有任何解决方法?
不幸的是您的代码段缺乏重现或解决问题所必需的背景。在目前的状态下,写出一个建设性的解决方案是不可能的,因此这个问题对其他读者没有用。我正在投票结束这个问题。如果您选择改进它,问题将自动添加到重新打开队列。请参考这两页进一步的信息:https://stackoverflow.com/help/mcve https://stackoverflow.com/help/reopen-questions – FireAphis