My website is not indexed by google i want someone who can make some changes like in [login to view URL] and sitemap to make the site indexed fast.
We need a custom dynamic crawl-tool for a car marketplace online portal. We need car data like make, model, car features, price etc. directly from the websites of the car dealers / directly from merchants. Sure, the URLs (about 10000) have different structures /format, with its own semantics (e.g. category /names. The data should be normalized. SO
I'm searching a partner for creating a new website e commerce for automo...English at the minimum. TAKE A LOOK AT [login to view URL] Webresponsive, full design, and associated at a crawler robot for having data from différents automotive websites like [login to view URL] or mobile. De. It have to crawl hundreds thousands cars and to integrate to my website
...programmer myself but I am looking for partner. This cutestat website is crawling all the websites and building pages with lots of information about domains (DNS records, IP, keywords, Alexa data, Whois,... etc). Your job would be to write a script to pull all this data for domain and display it in simple way (no design). You have to find corresponding APIs
I'm looking for someone willing to spend a day walking about their city and gathering a list of small giftshops. These are locally owned giftshops and stores. These aren't chains. The stores must be located in a (BIG or CAPTAL) city in one of the following countries: GERMANY, FRANCE, ITALY or AUSTRIA) I require just a simple list of 20 stores (NAME & ADRESS). To clarify further:...
...links for 90% of the sites and crawl the remaining sites. ( Many input files, the format always remains the same however, the data/names will be different) • All of the data is in a table on the site • All output formats and documentation are written • Basic features such as enabling/disabling sites, custom crawl delay, pause, play, skip, on-scree...
...Admin/users system - Admin needs to be able to add users - users needs to be able to add search requests/queries (form with 5 fields) - these requests dicate the crawl queries - 5 different websites need to get crawled (it currently only crawls 1) - The current code works, but needs refactoring (the current project is very small so this wont take long to
Looking for an experienced big data specialist to use the common crawl data set to get websites that offer tours and travels to [login to view URL] the successful candidate should have experience with the common crawl data set and how to implement the processing of this data with Map Reduce and running it on AWS EMR. You should be able to do this as...
Consolidating these WP sites: [login to view URL], [login to view URL], [login to view URL] ... into this domain [login to view URL], [login to view URL], [login to view URL] 1. Backup 2. Migrate 3. Setup 301 redirects (regex) 4. Site crawl to ensure paths are working with no 404s
...predefined search term on Google (results: 4'100 results). - For every entry: + Go to the site of the result + Crawl 4 information + Create an xls with an entry for every result, including the crawled information (structure of the data will be presented when the job has been assigned). That's it :-) Looking forward to your proposal. Regards, Dario
We are looking for web scraping experts to scrap information from several websites(Chinese) into json output. We plan to run the crawler on a daily/weekly basis. Depending on the websites to crawl some might require downloading files in pdf, doc, or other popular formats. Explicit logging is expected for all scraping tasks. You should be an expert
I am looking for a php expert who can solve issue in php curl Its a simple php curl code to crawl a given url and get title description etc from that url If a url has cloudflare enabled, it returns as "access denied" If you can solve, only then bid
A Powershell script is needed which anonymize the IP, FQDN, username and put the real values in a new table. Should run in a folder and crawl all *.txt and *.log files. search for IP, fqdn, username and put org value into a new file with count, hash, real value and also replace the real value with the hash. eg. source: "https://*.[login to view URL],https://*