Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue with proxy file #208

Open
lorenzoromani1983 opened this issue Apr 25, 2018 · 5 comments
Open

Issue with proxy file #208

lorenzoromani1983 opened this issue Apr 25, 2018 · 5 comments

Comments

@lorenzoromani1983
Copy link

Hi.
I am not able to run the proxy file.
I have it formatted this way:

Socks4 182.48.90.81:1080
Socks4 36.37.225.50:33012

it is a simple txt file with many rows
i get this error message:

Invalid proxy file. Should have the following format: {}'.format(parse_proxy_file.doc))
Exception: Invalid proxy file. Should have the following format: Parses a proxy file

please, help :)

@fassn
Copy link

fassn commented Apr 25, 2018

This project hasn't been updated for more than a year. Prefer using this project, developed from this one: https://github.com/fassn/SerpScrap (this is just a fork, I didn't started the project)

@lorenzoromani1983
Copy link
Author

thanks.
but that, as far as i understand, needs to be used within code.
this is a "stand-alone" tool...which is easier for me (not a coder, not an expert at all).
did you manage to make the proxies work in http mode?
i need to get a json/csv of 700/800 kw from google.

@fassn
Copy link

fassn commented Apr 27, 2018

Yes, I use proxies on the SerpScrap projet without problems

@lorenzoromani1983
Copy link
Author

lorenzoromani1983 commented Apr 28, 2018

THanks, it looks neat. However, i ran into two main problems:

1) I cant save to CSV:
how to proceed? this is the error i get:

Traceback (most recent call last):
File "C:\Users\Lorenzo\Anaconda\lib\site-packages\serpscrap\csv_writer.py", line 10, in write
with open(file_name, 'w', encoding='utf-8', newline='') as f:
PermissionError: [Errno 13] Permission denied: 'c:/.csv'
None
Traceback (most recent call last):
File "C:\Users\Lorenzo\Anaconda\lib\site-packages\serpscrap\csv_writer.py", line 10, in write
with open(file_name, 'w', encoding='utf-8', newline='') as f:
PermissionError: [Errno 13] Permission denied: 'c:/.csv'

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "C:\Users\Lorenzo\Desktop\scraper.txt", line 17, in
results = scrap.as_csv('c:/')
File "C:\Users\Lorenzo\Anaconda\lib\site-packages\serpscrap\serpscrap.py", line 134, in as_csv
writer.write(file_path + '.csv', self.results)
File "C:\Users\Lorenzo\Anaconda\lib\site-packages\serpscrap\csv_writer.py", line 17, in write
raise Exception
Exception

2) when I try to scrape many kw with proxies, I get this error, and end up banned by Google

2018-04-28 14:00:13,048 - scrapcore.scraper.selenium - WARNING - 'NoneType' object has no attribute 'group'
I am using proxies from free online proxy lists. Maybe they are already blacklisted by google?

@ecoron
Copy link

ecoron commented May 8, 2018

hi,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants