@@ -12,19 +12,26 @@ import (
12
12
// Client checks for CDN based IPs which should be excluded
13
13
// during scans since they belong to third party firewalls.
14
14
type Client struct {
15
- Data map [string ]struct {}
16
- ranger cidranger.Ranger
15
+ Options Options
16
+ Data map [string ]struct {}
17
+ ranger cidranger.Ranger
17
18
}
18
19
19
20
var defaultScrapers = map [string ]scraperFunc {
20
- "akamai" : scrapeAkamai ,
21
+ // "akamai": scrapeAkamai,
21
22
"azure" : scrapeAzure ,
22
23
"cloudflare" : scrapeCloudflare ,
23
24
"cloudfront" : scrapeCloudFront ,
24
25
"fastly" : scrapeFastly ,
25
26
"incapsula" : scrapeIncapsula ,
26
- "sucuri" : scrapeSucuri ,
27
- "leaseweb" : scrapeLeaseweb ,
27
+ // "sucuri": scrapeSucuri,
28
+ // "leaseweb": scrapeLeaseweb,
29
+ }
30
+
31
+ var defaultScrapersWithOptions = map [string ]scraperWithOptionsFunc {
32
+ "akamai" : scrapeAkamai ,
33
+ "sucuri" : scrapeSucuri ,
34
+ "leaseweb" : scrapeLeaseweb ,
28
35
}
29
36
30
37
var cachedScrapers = map [string ]scraperFunc {
@@ -33,15 +40,20 @@ var cachedScrapers = map[string]scraperFunc{
33
40
34
41
// New creates a new firewall IP checking client.
35
42
func New () (* Client , error ) {
36
- return new (false )
43
+ return new (& Options {} )
37
44
}
38
45
39
46
// NewWithCache creates a new firewall IP with cached data from project discovery (faster)
40
47
func NewWithCache () (* Client , error ) {
41
- return new (true )
48
+ return new (& Options {Cache : true })
49
+ }
50
+
51
+ // NewWithOptions creates a new instance with options
52
+ func NewWithOptions (Options * Options ) (* Client , error ) {
53
+ return new (Options )
42
54
}
43
55
44
- func new (cache bool ) (* Client , error ) {
56
+ func new (options * Options ) (* Client , error ) {
45
57
httpClient := & http.Client {
46
58
Transport : & http.Transport {
47
59
MaxIdleConns : 100 ,
@@ -55,21 +67,31 @@ func new(cache bool) (*Client, error) {
55
67
}
56
68
client := & Client {}
57
69
58
- var scrapers map [string ]scraperFunc
59
- if cache {
60
- scrapers = cachedScrapers
70
+ if options .Cache {
71
+ for _ , scraper := range cachedScrapers {
72
+ cidrs , err := scraper (httpClient )
73
+ if err != nil {
74
+ return nil , err
75
+ }
76
+ client .parseCidrs (cidrs )
77
+ }
61
78
} else {
62
- scrapers = defaultScrapers
79
+ for _ , scraper := range defaultScrapers {
80
+ cidrs , err := scraper (httpClient )
81
+ if err != nil {
82
+ return nil , err
83
+ }
84
+ client .parseCidrs (cidrs )
85
+ }
63
86
}
64
87
65
- client .Data = make (map [string ]struct {})
66
- for _ , scraper := range scrapers {
67
- cidrs , err := scraper (httpClient )
68
- if err != nil {
69
- return nil , err
70
- }
71
- for _ , cidr := range cidrs {
72
- client .Data [cidr ] = struct {}{}
88
+ if options .HasAuthInfo () {
89
+ for _ , scraper := range defaultScrapersWithOptions {
90
+ cidrs , err := scraper (httpClient , options )
91
+ if err != nil {
92
+ return nil , err
93
+ }
94
+ client .parseCidrs (cidrs )
73
95
}
74
96
}
75
97
@@ -86,6 +108,16 @@ func new(cache bool) (*Client, error) {
86
108
return client , nil
87
109
}
88
110
111
+ // parseCidrs inserts the scraped cidrs to the internal structure
112
+ func (c * Client ) parseCidrs (cidrs []string ) {
113
+ if c .Data == nil {
114
+ c .Data = make (map [string ]struct {})
115
+ }
116
+ for _ , cidr := range cidrs {
117
+ c .Data [cidr ] = struct {}{}
118
+ }
119
+ }
120
+
89
121
// Check checks if an IP is contained in the blacklist
90
122
func (c * Client ) Check (ip net.IP ) (bool , error ) {
91
123
return c .ranger .Contains (ip )
0 commit comments