Merge pull request #1444 from tobbez/fix_indent_spider
Fix incorrect indentation in spack.util.web._spider
This commit is contained in:
		| @@ -25,8 +25,7 @@ | |||||||
| import re | import re | ||||||
| import os | import os | ||||||
| import sys | import sys | ||||||
| import subprocess | import urllib2 | ||||||
| import urllib2, cookielib |  | ||||||
| import urlparse | import urlparse | ||||||
| from multiprocessing import Pool | from multiprocessing import Pool | ||||||
| from HTMLParser import HTMLParser, HTMLParseError | from HTMLParser import HTMLParser, HTMLParseError | ||||||
| @@ -84,7 +83,7 @@ def _spider(args): | |||||||
|         req.get_method = lambda: "HEAD" |         req.get_method = lambda: "HEAD" | ||||||
|         resp = urllib2.urlopen(req, timeout=TIMEOUT) |         resp = urllib2.urlopen(req, timeout=TIMEOUT) | ||||||
|  |  | ||||||
|         if not "Content-type" in resp.headers: |         if "Content-type" not in resp.headers: | ||||||
|             tty.debug("ignoring page " + url) |             tty.debug("ignoring page " + url) | ||||||
|             return pages, links |             return pages, links | ||||||
|  |  | ||||||
| @@ -142,12 +141,12 @@ def _spider(args): | |||||||
|                 pool.terminate() |                 pool.terminate() | ||||||
|                 pool.join() |                 pool.join() | ||||||
|  |  | ||||||
|     except urllib2.URLError, e: |     except urllib2.URLError as e: | ||||||
|         tty.debug(e) |         tty.debug(e) | ||||||
|         if raise_on_error: |         if raise_on_error: | ||||||
|             raise spack.error.NoNetworkConnectionError(str(e), url) |             raise spack.error.NoNetworkConnectionError(str(e), url) | ||||||
|  |  | ||||||
|     except HTMLParseError, e: |     except HTMLParseError as e: | ||||||
|         # This error indicates that Python's HTML parser sucks. |         # This error indicates that Python's HTML parser sucks. | ||||||
|         msg = "Got an error parsing HTML." |         msg = "Got an error parsing HTML." | ||||||
|  |  | ||||||
| @@ -157,7 +156,7 @@ def _spider(args): | |||||||
|  |  | ||||||
|         tty.warn(msg, url, "HTMLParseError: " + str(e)) |         tty.warn(msg, url, "HTMLParseError: " + str(e)) | ||||||
|  |  | ||||||
|     except Exception, e: |     except Exception as e: | ||||||
|         # Other types of errors are completely ignored, except in debug mode. |         # Other types of errors are completely ignored, except in debug mode. | ||||||
|         tty.debug("Error in _spider: %s" % e) |         tty.debug("Error in _spider: %s" % e) | ||||||
|  |  | ||||||
| @@ -173,7 +172,8 @@ def spider(root_url, **kwargs): | |||||||
|        performance over a sequential fetch. |        performance over a sequential fetch. | ||||||
|     """ |     """ | ||||||
|     max_depth = kwargs.setdefault('depth', 1) |     max_depth = kwargs.setdefault('depth', 1) | ||||||
|     pages, links = _spider((root_url, set(), root_url, None, 1, max_depth, False)) |     pages, links = _spider((root_url, set(), root_url, None, | ||||||
|  |                             1, max_depth, False)) | ||||||
|     return pages, links |     return pages, links | ||||||
|  |  | ||||||
|  |  | ||||||
| @@ -235,7 +235,7 @@ def find_versions_of_archive(*archive_urls, **kwargs): | |||||||
|             try: |             try: | ||||||
|                 ver = spack.url.parse_version(url) |                 ver = spack.url.parse_version(url) | ||||||
|                 versions[ver] = url |                 versions[ver] = url | ||||||
|             except spack.url.UndetectableVersionError as e: |             except spack.url.UndetectableVersionError: | ||||||
|                 continue |                 continue | ||||||
|  |  | ||||||
|     return versions |     return versions | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 becker33
					becker33