search

Home  >  Q&A  >  body text

Web crawler - python urlopen error timeout: timed out

Today the script always reports timeout errors. What is the exception handling of timeout?
The try except below is useless

        try:
            url_open = urllib.request.urlopen(url)
        except urllib.error.HTTPError:
            print('HTTPError')
            continue
        except urllib.error.URLError:
            print('URLError')
            continue

is to return

  File "F:\Program Files (x86)\Anaconda3\lib\http\client.py", line 612, in _safe_read
    chunk = self.fp.read(min(amt, MAXAMOUNT))
  File "F:\Program Files (x86)\Anaconda3\lib\socket.py", line 586, in readinto
    return self._sock.recv_into(b)
   timeout: timed out

Write except as

except Exception:

It’s no use either

伊谢尔伦伊谢尔伦2751 days ago1088

reply all(1)I'll reply

  • 伊谢尔伦

    伊谢尔伦2017-05-18 11:03:23

    Try requests.get(url)

    reply
    0
  • Cancelreply