ProxyFinder

Status
Not open for further replies.


I'd say it's the real deal, there were no major changes to the structure of it, just increased flexability, so shouldn't be buggy. Any problems, let me know.
 
Sorry to triple post .. but this program is stupid simple to use.

I had no issues with licenses ... in fact, to transfer to my ubuntu box all I did was copy the files to my ext. hard drive and then open the right file on ubuntu. After about 30 minutes of it running I've got 200 working anon proxies (and growing) out of a 19k list.

Fucking awesome Matt ... would pay 3x your cost for this product. Great work!

Suggestion: Timeout thresholds would be nice ... like if the proxy doesn't work in < x seconds don't include it in the list.

Did you use the default list that was included?? I let this sucker run on my Vista laptop for over 20 minutes without finding any Proxies. This was with the newest version that was posted to this thread.

I ran the previous version for about 15 min and it found about 3 proxies. I can't wait to get this beast going. It looks awesome :). Can I execute this program from my web host? I remember seeing somebody say something about doing that over SSH. I downloaded Putty yesterday and setup SSH over at my shared Host Gator account.
 
allbizniz: it sounds like you're running vista SP1, if you are, either upgrade to SP2 or apply the patch mentioned earlier in the thread. Running this on shared hosting will get you instabanned. Use a VPS for it. I plan on launching a hosted proxyfinder service soon, which you may be interested in.

Thanks
 
The new version seems to return many more usable proxies, although still plenty of dead ones by time you try and use them. Also keep in mind some of these proxies are proxies to specific sites. For example, one proxy returned data that seemed to be from a Chinese Google like site.
 
Rage9, the more often you update the list, generally the better the alive/dead ratio will be. PMing you about dodgy proxies.
 
Rage9, the more often you update the list, generally the better the alive/dead ratio will be. PMing you about dodgy proxies.

Yeah but scraping huge lists of random proxies you're bound to run into these problems. It's actually pretty expected.

I just wanted to put the word out that you need to be more vigilent in how you use them. For example I'm writing a scraper that has proxy support and if the data I expect is not being returned you have to be able to skip those dodgy proxies.

The way I'd solve it is when checking to see if the proxy is any good, query a website or server and check the output to see if it's good. For example you could setup a simple PHP script (or whatever language you choose) on a server and it just returns a small amount of data. If you get that data back from the proxy it would mean it isn't being used in another way. At the same time having a bunch of us constantly bashing that server may be another problem.
 
Yeah but scraping huge lists of random proxies you're bound to run into these problems. It's actually pretty expected.

I just wanted to put the word out that you need to be more vigilent in how you use them. For example I'm writing a scraper that has proxy support and if the data I expect is not being returned you have to be able to skip those dodgy proxies.

The way I'd solve it is when checking to see if the proxy is any good, query a website or server and check the output to see if it's good. For example you could setup a simple PHP script (or whatever language you choose) on a server and it just returns a small amount of data. If you get that data back from the proxy it would mean it isn't being used in another way. At the same time having a bunch of us constantly bashing that server may be another problem.

That's why this is a proxy harvester not a proxy in depth elite tester.
 
That's why this is a proxy harvester not a proxy in depth elite tester.

Well it's my 2 cents on it, he had asked me in a PM about more details on the dodge proxies so that he could figure a way to block them.
 
rage9, i agree you need to write your scripts with the ability to handle screwed up proxies. I tend to have a for loop, which it breaks out of if successful. Obviously I want to get rid of obviously dodgy proxies, but I need examples from people as they find them.
 
*cough* scrapebox with certain footprints, been mentioned on wf by gutterseo *cough*

So you're saying Scrapebox will check the proxies and send them back to some master list of working proxies for all the Scrapebox users to enjoy? I have probably interpreted that all wrong.
 
It was my fault I wasn't getting so many usable proxies. I've updated my thread with updated code to use in your PHP applications.
 
it now does additional checks to strip out planetlab and codeen proxies.

Could you explain why planetlab and codeen proxies are not desired?


I love the product......very nice.

However one problem I'm having is that when I immediately move my list of proxies over to Scrapebox and run the proxy check, it will take a list of say 150 proxies, and grind them down to 10 working proxies......is this typical, or am I missing something? (using the default proxy sources, btw).


Other than that, the product works as advertised.


N~
 
im also confused now with 2 people saying out of the 150-200 working proxies list that only 10-30 are even true anon or working.

I know proxies go up and down, but from 150-200 to 10-30 in a matter of 5 minutes seems a bit odd. shouldnt they all be anon?
 
I just tried the script on my Windows XP 2005 Media Center computer and it only found one proxy. I added about 10 more sites on top of what's provided in the script itself. Do you think I'm going something wrong here? I mean, I don't get any errors. I just let it run for hours.
 
Status
Not open for further replies.