Simple Web Spider in Python: Difference between revisions

From XPUB & Lens-Based wiki
No edit summary
 
(3 intermediate revisions by one other user not shown)
Line 1: Line 1:
A web spider that starts at a given URL, and follows links
== Opening an network connection with urllib2 ==
== Opening an network connection with urllib2 ==


Line 20: Line 22:
request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
f=urllib2.urlopen(request)
f=urllib2.urlopen(request)
</source>
== Get the URL of all images on a page and download them in a folder ==
<source lang="python">
import urllib2, urlparse, html5lib, lxml
from lxml.cssselect import CSSSelector
request = urllib2.Request("http://www.volkskrant.nl/")
request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
f=urllib2.urlopen(request)
# f.geturl(), f.info()
parser = html5lib.HTMLParser(tree=html5lib.treebuilders.getTreeBuilder("lxml"), namespaceHTMLElements=False)
page = parser.parse(f)
for elt in CSSSelector('img[src]')(page):
    href = urlparse.urljoin(f.geturl(), elt.attrib['src'])
    print 'downloading ' + href
    localfile = open('dump/'+href.split('/')[-1], "wb")
    localfile.write(remotefile.read())
    localfile.close()
</source>
</source>


Line 117: Line 95:
</source>
</source>


== Same as above and grab all the pictures found on each page ==
== Same as above and grab all the pictures found on each page AND BE SMARTER==
<source lang="python">
<source lang="python">
import random
import random
Line 123: Line 101:
from lxml.cssselect import CSSSelector
from lxml.cssselect import CSSSelector


useragent = "Mozilla/5.001 (windows; U; NT4.0; en-US; rv:1.0) Gecko/25250101"
targets = ['http://www.volkskrant.nl/']
targets = ['http://www.volkskrant.nl/']


Line 131: Line 110:
     # try to open URL ...
     # try to open URL ...
     try:
     try:
         request = urllib2.Request(target)
         request = urllib2.Request(target, None, {'User-Agent': useragent})
        request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
         f=urllib2.urlopen(request)
         f=urllib2.urlopen(request)


Line 144: Line 122:
                 for link in CSSSelector('a[href]')(page):
                 for link in CSSSelector('a[href]')(page):
                     href = urlparse.urljoin(f.geturl(), link.attrib['href'])
                     href = urlparse.urljoin(f.geturl(), link.attrib['href'])
                     if href.split(':')[0] == 'http': # No js links
                     if href.split(':')[0] == 'http' and href != target: # No useless links
                         links.append(href)
                         links.append(href)
             if links: # Anything left?
             if links: # Anything left?
Line 150: Line 128:
                 for elt in CSSSelector('img[src]')(page):
                 for elt in CSSSelector('img[src]')(page):
                     href = urlparse.urljoin(f.geturl(), elt.attrib['src'])
                     href = urlparse.urljoin(f.geturl(), elt.attrib['src'])
                     remotefile = urllib2.urlopen(href)
                    request = urllib2.Request(href, None, {'User-Agent': useragent})
                     remotefile = urllib2.urlopen(request)
                     print 'downloading ' + href
                     print 'downloading ' + href
                     localfile = open('dump/'+href.split('/')[-1], "wb")
                     localfile = open('dump/'+href.split('/')[-1], "wb")
Line 156: Line 135:
                     localfile.close()
                     localfile.close()


         except ValueError, err:
        except IOError:
             print "Value Error:", err, target
            print "Ooops"
 
         except ValueError:
             print "Ooops"
 
        except AssertionError:
            print "Ooops"


     # ... catch HTTP and URL errors
     # ... catch HTTP and URL errors

Latest revision as of 11:51, 2 December 2013

A web spider that starts at a given URL, and follows links

Opening an network connection with urllib2

import urllib2

request = urllib2.Request("http://www.volkskrant.nl/")
f=urllib2.urlopen(request)

print f.geturl()
print f.info()
print f.read()

Some sites require that you set the "User-Agent" header.

import urllib2

request = urllib2.Request("http://www.volkskrant.nl/")
request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
f=urllib2.urlopen(request)

Get the URL of all the links in a page and jump to a random page

import random
import urllib2, urlparse, html5lib, lxml
from lxml.cssselect import CSSSelector

targets = ['http://www.volkskrant.nl/']

while True:
    target = random.choice(targets)
    print '*** '+target+' ***'
    request = urllib2.Request(target)
    request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
    f=urllib2.urlopen(request)

    parser = html5lib.HTMLParser(tree=html5lib.treebuilders.getTreeBuilder("lxml"), namespaceHTMLElements=False)
    page = parser.parse(f)

    targets = []

    for links in CSSSelector('a[href]')(page):
        href = urlparse.urljoin(f.geturl(), links.attrib['href'])
        if href.split(':')[0] == 'http':
            targets.append(href)

Get the URL of all the links in a page and jump to a random page AND BE SMART

import random
import urllib2, urlparse, html5lib, lxml
from lxml.cssselect import CSSSelector

targets = ['http://www.volkskrant.nl/']

while True:
    target = random.choice(targets)
    print '*** '+target+' ***'

    # try to open URL ...
    try:
        request = urllib2.Request(target)
        request.add_header("User-Agent", "Mozilla/5.0 (X11; U; Linux x86_64; fr; rv:1.9.1.5) Gecko/20091109 Ubuntu/9.10 (karmic) Firefox/3.5.5")
        f=urllib2.urlopen(request)

        # Is it really something that I can parse? srsly?
        try:
            parser = html5lib.HTMLParser(tree=html5lib.treebuilders.getTreeBuilder("lxml"), namespaceHTMLElements=False)
            page = parser.parse(f)
        except ValueError, err:
            print "Value Error:", err, target

        links = []
        if CSSSelector('a[href]')(page): # Any links for me?
            for link in CSSSelector('a[href]')(page):
                href = urlparse.urljoin(f.geturl(), link.attrib['href'])
                if href.split(':')[0] == 'http': # No js links
                    links.append(href)
        if links: # Anything left?
            targets = links


    # ... catch HTTP and URL errors
    except urllib2.HTTPError, err:
        print "HTTP Error:",err.code , target
        print "trying other URL"
    except urllib2.URLError, err:
        print "URL Error:",err.reason , target
        print "trying other URL"

Same as above and grab all the pictures found on each page AND BE SMARTER

import random
import urllib2, urlparse, html5lib, lxml
from lxml.cssselect import CSSSelector

useragent = "Mozilla/5.001 (windows; U; NT4.0; en-US; rv:1.0) Gecko/25250101"
targets = ['http://www.volkskrant.nl/']

while True:
    target = random.choice(targets)
    print '*** '+target+' ***'

    # try to open URL ...
    try:
        request = urllib2.Request(target, None, {'User-Agent': useragent})
        f=urllib2.urlopen(request)

        # Is it really something that I can parse? srsly?
        try:
            parser = html5lib.HTMLParser(tree=html5lib.treebuilders.getTreeBuilder("lxml"), namespaceHTMLElements=False)
            page = parser.parse(f)

            links = []
            if CSSSelector('a[href]')(page): # Any links for me?
                for link in CSSSelector('a[href]')(page):
                    href = urlparse.urljoin(f.geturl(), link.attrib['href'])
                    if href.split(':')[0] == 'http' and href != target: # No useless links
                        links.append(href)
            if links: # Anything left?
                targets = links
                for elt in CSSSelector('img[src]')(page):
                    href = urlparse.urljoin(f.geturl(), elt.attrib['src'])
                    request = urllib2.Request(href, None, {'User-Agent': useragent})
                    remotefile = urllib2.urlopen(request)
                    print 'downloading ' + href
                    localfile = open('dump/'+href.split('/')[-1], "wb")
                    localfile.write(remotefile.read())
                    localfile.close()

        except IOError:
            print "Ooops"

        except ValueError:
            print "Ooops"

        except AssertionError:
            print "Ooops"

    # ... catch HTTP and URL errors
    except urllib2.HTTPError, err:
        print "HTTP Error:",err.code , target
        print "trying other URL"
    except urllib2.URLError, err:
        print "URL Error:",err.reason , target
        print "trying other URL"