Hello There,
I have been working as a freelance developer for the past five years and have very good experience in writing web-scraping scripts, both stand-alone and full-fledged backends.
Following is a list of some projects I have worked on recently.
a) Gigkahuna - Entire backend scraper system using python and lxml and mongodb/mysql
b) purpleforge - Working on top of existing scraper scripts using python and lxml
c) Cyclehire Android App - Entire backend system on Google appengine using python and beautiful soup.
For a detailed list of projects, please check my profile as well as my portfolio here.
I shall be using lxml for parsing the html page. Also I am not sure if mechanize is required or not, as I had used simple cookiejars and other session variables previously for login with great success. Can you please share the URL to be scraped, just to make sure whether mechanize is really required or not..
Also can you tell me the output format for the script? [Basically whether a plain .csv would do or whether it has to be interfaced with a database]
regards,
Sandeep S
Edit:- Just noticed the attached image with the bid and found out that you need to parse a facebook page. Please ignore my first question. Sorry for that.