Crawl Domain Find Expred Domains

Closed Posted 4 years ago Paid on delivery
Closed Paid on delivery

We are looking for a crawler to crawl every page of a website looking for external links pointing to expired domains.

User should definde a list of sites to crawl via text file. Crawler should work logically crawling all pages of a site and not be sitemap dependent. Only unique external domains should be logged to prevent duplicate domain availability lookups.

User should also be able to define a list of urls to ignore checking for availability; eg. [login to view URL] etc. these domains should be user defined in a blacklist text file.

Results should be given in a csv file listing linking domain and available domain.

Python Web Scraping

Project ID: #19760507

About the project

4 proposals Remote project Active 4 years ago

4 freelancers are bidding on average $74 for this job

chirgeo

Hi. I did read the project description and have a few questions. 1. Do you need the script as well or data only? 2. What is the format of the output data? CSV is OK? We can do other formats as well. 3. Which fields do More

$200 USD in 5 days
(155 Reviews)
8.1
smsaurabhv

‌Hi, I have gone through your requirement to scrape lots of websites. I am EXPERT in building scraping tools /scripts. Hence, I can SURELY work on your project. I am having 4 YEARS of EXPERIENCE in developing PHP-PYTHO More

$35 USD in 3 days
(120 Reviews)
6.1
arundevp

Hey, I think i may be able to help you out on this project since i have worked on a similar project at work. Im new to freelancer, though i have years of experience in software development in python.

$20 USD in 1 day
(0 Reviews)
0.0