I had a project in which I read about 45 million requests (with forms) on an expanded basis. On an ongoing basis, I scraped about 20 simultaneous clients, and my pipe was a bottleneck.
I used Selinium Remote-Control after experimenting with writing my own WebClient, WaTiN / WaTiR and using the Microsoft UI Automation API.
Selenium RC allows you to choose a browser. I used Firefox. Setting up the initial scrambling scripts took about an hour of experimentation and tuning. Selenium was much faster than writing native code and much more robust with little investment. Great tool.
To scale the process, I tried several different approaches, but in the end, what worked best was for each instance of SRC to be located in its own shared virtual machine, and then spawn so many of those supported by the workstation. The equivalent number of SRC instances running on the host, instead of vms, inevitably stops when I get up to +10 instances. This required more overhead and installation time before scrambling, but it will work for several days without interruption.
Another consideration: adjust your Firefox settings so that no page loads, turn off all non-essential ones (spoofing checks, cookies if this is not required for your scratch, images, ad unit and flash block, etc.).
Grant source share