How To Extract Data From Multiple Url Using Python
Hi i want to scrap data from multiple URL, I am doing like for i in range(493): my_url = 'http://tis.nhai.gov.in/TollInformation?TollPlazaID={}'.format(i) but it not giving m
Solution 1:
Seems like some of the pages are missing your key information, you can use error-catching
for it, like this:
try:
tbody = soup('table', {"class": "tollinfotbl"})[0].find_all('tr')[1:]
except IndexError:
continue # Skip this page if no items were scrapped
You may want to add some logging/print information to keep track of nonexisting tables.
EDIT:
It's showing information from only last page, as you are commiting your transaction outside the for
loop, overwriting your conn
for every i
. Just put conn.commit()
inside for
loop, at the far end.
Post a Comment for "How To Extract Data From Multiple Url Using Python"