Content Grabber with free proxy account integration for business directories scrape

Professional data extraction requires adequate proxying to keep anonymity of scraping robots. When attempting to extract large data sets (over 1M records, ex. business directories) reliable and fast proxy service is needed.

Sequentum has released the Nohodo proxy service integration for Content Grabber. Nohodo provides a free account for Content Grabber users (up to 5000 requests monthly for free). The feature is available for both trial users and regular customers. Here’s how it works…

Writing next generation scraping scripts with Web Robots IDE

http cookie
Most scraping solutions fall into two categories: ­Visual scraping platforms targeted at non-programmers ( Content Grabber, Dexi.io, Import.io, etc.), and scraping code libraries like Scrapy or PhantomJS which require at least some knowledge of how to code.

Web Robots builds scraping IDE that fills the gap in between. Code is not hidden but instead made simple to create, run and debug.

import.io’s New Scraping Process and Features


Web scraping Data platform import.io, announced last week that they have secured $3M in funding from investors that include the founders of Yahoo! and MySQL.

They also released a new beta version of the tool that is essentially a better version of their extraction tool, with some new features and a much cleaner and faster user experience.

Scraping software, services and plugins sum up

scraping-software-services-sum-upSince we have already reviewed classic web harvesting software, we want to sum up some other scraping services and crawlers, scrape plugins and other scrape related tools.

Web scraping is a sphere that can be applied to a vast variety of fields, and in turn it can require other technologies to be involved. SEO needs scrape. Proxying is one of the methods which can help you to stay masked while doing much web data extraction. Crawling is another sub-technology indispensable in scrape for unordered information sources. Data refining follows the scrape, so as to deal with the unavoidable inconsistency of harvested data.
In addition, we will consider fast scrape tools, making our life better, and some services and handy scrapers which enable us to obtain freshly extracted data or images.