2
2
from project .http .requests .parsers .proxyforeuParser import proxyforeuParser
3
3
from project .http .requests .parsers .rebroweeblyParser import rebroweeblyParser
4
4
from project .http .requests .parsers .samairproxyParser import semairproxyParser
5
-
6
- __author__ = 'pgaref'
7
-
8
5
import requests
9
6
from requests .exceptions import ConnectionError
10
7
import random
11
- import os
12
8
import time
13
- from bs4 import BeautifulSoup
14
9
from requests .exceptions import ReadTimeout
15
10
11
+ __author__ = 'pgaref'
16
12
17
13
class RequestProxy :
18
14
agent_file = '../data/user_agents.txt'
19
15
20
16
def __init__ (self , web_proxy_list = []):
21
17
self .useragents = self .load_user_agents (RequestProxy .agent_file )
18
+
22
19
#####
23
20
# Each of the classes below implements a specific URL Parser
24
- # http://<USERNAME>:<PASSWORD>@<IP-ADDR>:<PORT>
25
21
#####
26
22
parsers = []
27
23
parsers .append (freeproxyParser ('http://free-proxy-list.net' ))
@@ -83,24 +79,24 @@ def generate_proxied_request(self, url, params={}, req_timeout=30):
83
79
request = None
84
80
try :
85
81
rand_proxy = random .choice (self .proxy_list )
86
- print "Using proxy: " + str (rand_proxy )
82
+ print "Using proxy: {0}" . format ( str (rand_proxy ) )
87
83
request = requests .get (test_url , proxies = {"http" : rand_proxy },
88
84
headers = req_headers , timeout = req_timeout )
89
85
except ConnectionError :
90
86
self .proxy_list .remove (rand_proxy )
91
- print "Proxy unreachable - Removed Straggling proxy :" , rand_proxy , " PL Size = " , len (self .proxy_list )
87
+ print "Proxy unreachable - Removed Straggling proxy: {0} PL Size = {1}" . format ( rand_proxy , len (self .proxy_list ) )
92
88
pass
93
89
except ReadTimeout :
94
90
self .proxy_list .remove (rand_proxy )
95
- print "Read timed out - Removed Straggling proxy :" , rand_proxy , " PL Size = " , len (self .proxy_list )
91
+ print "Read timed out - Removed Straggling proxy: {0} PL Size = {1}" . format ( rand_proxy , len (self .proxy_list ) )
96
92
pass
97
93
return request
98
94
99
95
if __name__ == '__main__' :
100
96
101
97
start = time .time ()
102
98
req_proxy = RequestProxy ()
103
- print "Initialization took: " , ( time .time ()- start )
99
+ print "Initialization took: {0} sec" . format (( time .time ()- start ) )
104
100
print "Size : " , len (req_proxy .get_proxy_list ())
105
101
print " ALL = " , req_proxy .get_proxy_list ()
106
102
@@ -109,7 +105,7 @@ def generate_proxied_request(self, url, params={}, req_timeout=30):
109
105
while True :
110
106
start = time .time ()
111
107
request = req_proxy .generate_proxied_request (test_url )
112
- print "Proxied Request Took: " , ( time .time ()- start ), " => Status: " , request .__str__ ()
108
+ print "Proxied Request Took: {0} sec => Status: {1}" . format (( time .time ()- start ), request .__str__ () )
113
109
print "Proxy List Size: " , len (req_proxy .get_proxy_list ())
114
110
115
111
print "-> Going to sleep.."
0 commit comments