rem
stringlengths 0
322k
| add
stringlengths 0
2.05M
| context
stringlengths 8
228k
|
---|---|---|
sources = [p('wc/HtmlParser/htmllex.c'), p('wc/HtmlParser/htmlparse.c'), p('wc/HtmlParser/s_util.c'),
|
sources = [normpath('wc/HtmlParser/htmllex.c'), normpath('wc/HtmlParser/htmlparse.c'), normpath('wc/HtmlParser/s_util.c'),
|
def get_exe_bytes (self): if win_cross_compiling: bv = "7.1" # wininst-x.y.exe is in the same directory as bdist_wininst directory = os.path.dirname(distutils.command.__file__) # we must use a wininst-x.y.exe built with the same C compiler # used for python. filename = os.path.join(directory, "wininst-%s.exe" % bv) return open(filename, "rb").read() return super(MyBdistWininst, self).get_exe_bytes()
|
depends = [p("wc/HtmlParser/htmlsax.h"), p('wc/HtmlParser/s_util.h')], include_dirs = include_dirs + [p("wc/HtmlParser")],
|
depends = [normpath("wc/HtmlParser/htmlsax.h"), normpath('wc/HtmlParser/s_util.h')], include_dirs = include_dirs + [normpath("wc/HtmlParser")],
|
def get_exe_bytes (self): if win_cross_compiling: bv = "7.1" # wininst-x.y.exe is in the same directory as bdist_wininst directory = os.path.dirname(distutils.command.__file__) # we must use a wininst-x.y.exe built with the same C compiler # used for python. filename = os.path.join(directory, "wininst-%s.exe" % bv) return open(filename, "rb").read() return super(MyBdistWininst, self).get_exe_bytes()
|
sources = [p('wc/levenshtein.c'),],
|
sources = [normpath('wc/levenshtein.c'),],
|
def get_exe_bytes (self): if win_cross_compiling: bv = "7.1" # wininst-x.y.exe is in the same directory as bdist_wininst directory = os.path.dirname(distutils.command.__file__) # we must use a wininst-x.y.exe built with the same C compiler # used for python. filename = os.path.join(directory, "wininst-%s.exe" % bv) return open(filename, "rb").read() return super(MyBdistWininst, self).get_exe_bytes()
|
def __init__ (self, headers):
|
def __init__ (self):
|
def __init__ (self, headers): """ Initialize internal buffers and flags. """ self.headers = headers self.buf = '' self.bytes_remaining = None self.closed = False
|
self.headers = headers
|
def __init__ (self, headers): """ Initialize internal buffers and flags. """ self.headers = headers self.buf = '' self.bytes_remaining = None self.closed = False
|
|
if i > 0:
|
if i >= 0:
|
def read_footers (self): i = self.buf.find('\r\n\r\n') if i > 0: fp = StringIO.StringIO(self.buf[:i]) self.buf = self.buf[i+4:] msg = wc.http.header.WcMessage(fp) fp.close() for name in msg: for value in msg.getheaders(name): self.headers.addheader(name, value)
|
self.headers.addheader(name, value)
|
pass
|
def read_footers (self): i = self.buf.find('\r\n\r\n') if i > 0: fp = StringIO.StringIO(self.buf[:i]) self.buf = self.buf[i+4:] msg = wc.http.header.WcMessage(fp) fp.close() for name in msg: for value in msg.getheaders(name): self.headers.addheader(name, value)
|
remove_old_warning_headers(headers)
|
remove_warning_headers(headers)
|
def set_proxy_headers (headers): remove_hop_by_hop_headers(headers) set_via_header(headers) set_date_header(headers) remove_old_warning_headers(headers)
|
if self.rule:
|
if self.rule is None: self.folder.end_data(name) else:
|
def end_element (self, name): self.cmode = None if self.rule: self.rule.end_data(name) if name in rulenames: if self.compile_data: self.rule.compile_data() elif name=='folder': if self.compile_data: self.folder.compile_data()
|
self.buffer_append_data([DATA; "<?%s?>"%data])
|
self.buffer_append_data([DATA, "<?%s?>"%data])
|
def pi (self, data): self.buffer_append_data([DATA; "<?%s?>"%data])
|
tracelog = log
|
if tracelog is None: tracelog = log
|
def trace (log): """ Start tracing of the current thread (and the current thread only). """ global tracelog tracelog = log sys.settrace(_traceit)
|
debug(tracelog, "%s:%s: %s", name, lineno, line.rstrip())
|
info(tracelog, "%s:%s: %s", name, lineno, line.rstrip())
|
def _traceit (frame, event, arg): """ Print current executed line. """ if event == "line": lineno = frame.f_lineno filename = frame.f_globals["__file__"] if filename.endswith(".pyc") or filename.endswith(".pyo"): filename = filename[:-1] name = frame.f_globals["__name__"] line = linecache.getline(filename, lineno) debug(tracelog, "%s:%s: %s", name, lineno, line.rstrip()) return _traceit
|
traceback.print_stack(stack, file=s) if PRINT_LOCALVARS: s.write("Locals by frame, innermost last%s" % os.linesep) for frame in stack: s.write(os.linesep) s.write("Frame %s in %s at line %s%s" % (frame.f_code.co_name, frame.f_code.co_filename, frame.f_lineno, os.linesep))
|
s.write('Traceback:') s.write(os.linesep) for frame, fname, lineno, method, lines, i in reversed(stack): s.write(' File %r, line %d, in %s' % (fname, lineno, method)) s.write(os.linesep) s.write(' %s' % lines[0].lstrip()) if PRINT_LOCALVARS:
|
def _stack_format (stack): """ Format a stack trace to a message. @return: formatted stack message @rtype: string """ s = StringIO.StringIO() traceback.print_stack(stack, file=s) if PRINT_LOCALVARS: s.write("Locals by frame, innermost last%s" % os.linesep) for frame in stack: s.write(os.linesep) s.write("Frame %s in %s at line %s%s" % (frame.f_code.co_name, frame.f_code.co_filename, frame.f_lineno, os.linesep)) for key, value in frame.f_locals.items(): s.write("\t%20s = " % key) # be careful not to cause a new error in the error output try: s.write(str(value)) s.write(os.linesep) except: s.write("error in str() call%s" % os.linesep) return s.getvalue()
|
s.write("\t%20s = " % key)
|
s.write(" %s = " % key)
|
def _stack_format (stack): """ Format a stack trace to a message. @return: formatted stack message @rtype: string """ s = StringIO.StringIO() traceback.print_stack(stack, file=s) if PRINT_LOCALVARS: s.write("Locals by frame, innermost last%s" % os.linesep) for frame in stack: s.write(os.linesep) s.write("Frame %s in %s at line %s%s" % (frame.f_code.co_name, frame.f_code.co_filename, frame.f_lineno, os.linesep)) for key, value in frame.f_locals.items(): s.write("\t%20s = " % key) # be careful not to cause a new error in the error output try: s.write(str(value)) s.write(os.linesep) except: s.write("error in str() call%s" % os.linesep) return s.getvalue()
|
s.write(str(value))
|
s.write(repr(value))
|
def _stack_format (stack): """ Format a stack trace to a message. @return: formatted stack message @rtype: string """ s = StringIO.StringIO() traceback.print_stack(stack, file=s) if PRINT_LOCALVARS: s.write("Locals by frame, innermost last%s" % os.linesep) for frame in stack: s.write(os.linesep) s.write("Frame %s in %s at line %s%s" % (frame.f_code.co_name, frame.f_code.co_filename, frame.f_lineno, os.linesep)) for key, value in frame.f_locals.items(): s.write("\t%20s = " % key) # be careful not to cause a new error in the error output try: s.write(str(value)) s.write(os.linesep) except: s.write("error in str() call%s" % os.linesep) return s.getvalue()
|
s.write("error in str() call%s" % os.linesep)
|
s.write("error in repr() call%s" % os.linesep)
|
def _stack_format (stack): """ Format a stack trace to a message. @return: formatted stack message @rtype: string """ s = StringIO.StringIO() traceback.print_stack(stack, file=s) if PRINT_LOCALVARS: s.write("Locals by frame, innermost last%s" % os.linesep) for frame in stack: s.write(os.linesep) s.write("Frame %s in %s at line %s%s" % (frame.f_code.co_name, frame.f_code.co_filename, frame.f_lineno, os.linesep)) for key, value in frame.f_locals.items(): s.write("\t%20s = " % key) # be careful not to cause a new error in the error output try: s.write(str(value)) s.write(os.linesep) except: s.write("error in str() call%s" % os.linesep) return s.getvalue()
|
s = _stack_format(inspect.stack()[2:]) fun(s)
|
fun(_stack_format(inspect.stack()[2:]))
|
def _log (fun, msg, args, tb=False): """ Log a message with given function and an optional traceback. @return: None """ fun(msg, *args) if tb: # note: get rid of last parts of the stack s = _stack_format(inspect.stack()[2:]) fun(s)
|
def debug (log, msg, *args, **kwargs):
|
def debug (logname, msg, *args, **kwargs):
|
def debug (log, msg, *args, **kwargs): """ Log a debug message. return: None """ _log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.DEBUG): _log(log.debug, msg, args, tb=kwargs.get("tb"))
|
def debug (log, msg, *args, **kwargs): """ Log a debug message. return: None """ _log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb"))
|
def info (log, msg, *args, **kwargs):
|
def info (logname, msg, *args, **kwargs):
|
def info (log, msg, *args, **kwargs): """ Log an informational message. return: None """ _log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.INFO): _log(log.info, msg, args, tb=kwargs.get("tb"))
|
def info (log, msg, *args, **kwargs): """ Log an informational message. return: None """ _log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb"))
|
def warn (log, msg, *args, **kwargs):
|
def warn (logname, msg, *args, **kwargs):
|
def warn (log, msg, *args, **kwargs): """ Log a warning. return: None """ _log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.WARN): _log(log.warn, msg, args, tb=kwargs.get("tb"))
|
def warn (log, msg, *args, **kwargs): """ Log a warning. return: None """ _log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb"))
|
def error (log, msg, *args, **kwargs):
|
def error (logname, msg, *args, **kwargs):
|
def error (log, msg, *args, **kwargs): """ Log an error. return: None """ _log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.ERROR): _log(log.error, msg, args, tb=kwargs.get("tb"))
|
def error (log, msg, *args, **kwargs): """ Log an error. return: None """ _log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb"))
|
def critical (log, msg, *args, **kwargs):
|
def critical (logname, msg, *args, **kwargs):
|
def critical (log, msg, *args, **kwargs): """ Log a critical error. return: None """ _log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.CRITICAL): _log(log.critical, msg, args, tb=kwargs.get("tb"))
|
def critical (log, msg, *args, **kwargs): """ Log a critical error. return: None """ _log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb"))
|
def exception (log, msg, *args, **kwargs):
|
def exception (logname, msg, *args, **kwargs):
|
def exception (log, msg, *args, **kwargs): """ Log an exception. return: None """ _log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb"))
|
_log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb"))
|
log = logging.getLogger(logname) if log.isEnabledFor(logging.EXCEPTION): _log(log.exception, msg, args, tb=kwargs.get("tb"))
|
def exception (log, msg, *args, **kwargs): """ Log an exception. return: None """ _log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb"))
|
count, counttype = subkey['DNSServerAddressCount'] values, valuestype = subkey['DNSServerAddresses']
|
values = subkey.get('DNSServerAddresses', "")
|
def init_dns_resolver_nt (): import winreg key = None try: key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\Tcpip\Parameters") except EnvironmentError: try: # for Windows ME key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\VxD\MSTCP") except EnvironmentError: pass if key: for server in winreg.stringdisplay(key.get("NameServer", "")): if server: DnsConfig.nameservers.append(str(server)) for item in winreg.stringdisplay(key.get("SearchList", "")): if item: DnsConfig.search_domains.append(str(item)) if not DnsConfig.nameservers: # XXX the proper way to test this is to search for # the "EnableDhcp" key in the interface adapters... for server in winreg.stringdisplay(key.get("DhcpNameServer", "")): if server: DnsConfig.nameservers.append(str(server)) for item in winreg.stringdisplay(key.get("DhcpDomain", "")): if item: DnsConfig.search_domains.append(str(item)) try: # search adapters key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\Tcpip\Parameters\DNSRegisteredAdapters") for subkey in key.subkeys(): count, counttype = subkey['DNSServerAddressCount'] values, valuestype = subkey['DNSServerAddresses'] for server in winreg.binipdisplay(values): if server: DnsConfig.nameservers.append(server) except EnvironmentError: pass try: # search interfaces key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\Tcpip\Parameters\Interfaces") for subkey in key.subkeys(): for server in winreg.stringdisplay(subkey.get('NameServer', '')): if server: DnsConfig.nameservers.append(server) except EnvironmentError: pass
|
[---------|--------][-------][----------][--... ^-script src tag waitbuf: [--------]
|
[------------------][-------][----------][--... outbuf: [--] buf: [-----] waitbuf: [-------]
|
def _fatalError (self, msg): """signal a fatal filter/parser error""" self._errorfun(msg, "fatalError")
|
data = self.inbuf.getvalue()
|
data = self.inbuf.getvalue() + data
|
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self._debug(ALWAYS, "self.inbuf", `self.inbuf.getvalue()`) return data = self.inbuf.getvalue() self.inbuf.close() self.inbuf = StringIO() if data: # only feed non-empty data #self._debug(NIGHTMARE, "feed", `data`) self.parser.feed(data) else: #self._debug(NIGHTMARE, "feed") pass else: # wait state --> put in input buffer #self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
|
self.js_html = FilterHtmlParser(self.rules, self.url,
|
self.js_html = FilterHtmlParser(self.rules, self.pics, self.url,
|
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" #self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = FilterHtmlParser(self.rules, self.url, comments=self.comments, javascript=self.js_filter, level=self.level+1) # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.detachListener(self) # wait for recursive filter to finish self.jsEndScript(item)
|
p.feed("<img bo\\\nrder=0>")
|
p.debug(1) for c in """</td <td a="b" >""": p.feed(c)
|
def _broken (): p = HtmlPrinter() p.feed("<img bo\\\nrder=0>") p.flush()
|
Htmlparser.__init__(self)
|
HtmlParser.__init__(self)
|
def __init__ (self): Htmlparser.__init__(self) if wc.config['showerrors']: self.error = self._error self.warning = self._warning self.fatalError = self._fatalError self.outbuf = StringIO() self.buf = []
|
if not attrs.has_key('buffer'):
|
if attrs['buffer'].closed:
|
def filter (self, data, **attrs): if not attrs.has_key('buffer'): # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() del attrs['buffer'] return return ''
|
del attrs['buffer'] return
|
return data
|
def filter (self, data, **attrs): if not attrs.has_key('buffer'): # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() del attrs['buffer'] return return ''
|
img = Image.open(buf)
|
img = Image.open(buf, 'r')
|
def check_sizes (self, buf): try: img = Image.open(buf) for size, formats in sizes: if size==img.size: # size matches, look for format restriction if not formats: return False elif img.format.lower() in formats: return False except IOError: # XXX print error pass return True
|
def testUrl (self):
|
def testUrlPathAttack (self):
|
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitech%2Bwingman%2Bextreme%2Bdigital%2B3d%26hl%3Den%26lr%3D%26ie%3DUTF-8%26selm%3D3845B54D.E546F9BD%2540monmouth.com%26rnum%3D2" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEqual(url, nurl)
|
nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip")
|
nurl = "http://server/cskin.zip" self.assertEquals(wc.url.url_quote(wc.url.url_norm(url)), nurl) def testUrlQuoting (self):
|
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitech%2Bwingman%2Bextreme%2Bdigital%2B3d%26hl%3Den%26lr%3D%26ie%3DUTF-8%26selm%3D3845B54D.E546F9BD%2540monmouth.com%26rnum%3D2" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEqual(url, nurl)
|
nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEqual(url, nurl)
|
nurl = url self.assertEqual(wc.url.url_quote(wc.url.url_norm(url)), nurl) def testUrlFixing (self): url = r"http://groups.google.com\test.html" nurl = "http://groups.google.com/test.html" self.assertEqual(wc.url.url_norm(url), nurl) url = r"http://groups.google.com/a\test.html" nurl = "http://groups.google.com/a/test.html" self.assertEqual(wc.url.url_norm(url), nurl) url = r"http://groups.google.com\a\test.html" nurl = "http://groups.google.com/a/test.html" self.assertEqual(wc.url.url_norm(url), nurl) url = r"http://groups.google.com\a/test.html" nurl = "http://groups.google.com/a/test.html" self.assertEqual(wc.url.url_norm(url), nurl)
|
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitech%2Bwingman%2Bextreme%2Bdigital%2B3d%26hl%3Den%26lr%3D%26ie%3DUTF-8%26selm%3D3845B54D.E546F9BD%2540monmouth.com%26rnum%3D2" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEqual(url, nurl)
|
[wc.filter.rules.RewriteRule.STARTTAG, "script",
|
[wc.filter.rules.RewriteRule.STARTTAG, u"script",
|
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0] == 'wait', "non-wait state" wc.log.debug(wc.LOG_JS, "%s jsScriptData %r", self, data) if data is None: if not self.js_script: wc.log.warn(wc.LOG_JS, "empty JavaScript src %s", url) self.js_script = u"// "+\ _("error fetching script from %r") % url self.htmlparser.tagbuf.append( [wc.filter.rules.RewriteRule.STARTTAG, "script", {'type': 'text/javascript'}]) # norm html comments script = wc.js.remove_html_comments(self.js_script) script = u"\n<!--\n%s\n//-->\n" % wc.js.escape_js(script) self.htmlparser.tagbuf.append( [wc.filter.rules.RewriteRule.DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.htmlparser.tagbuf.append( [wc.filter.rules.RewriteRule.ENDTAG, u"script"]) self.js_script = u'' self.htmlparser.state = ('parse',) wc.log.debug(wc.LOG_JS, "%s switching back to parse with", self) self.htmlparser.debugbuf(wc.LOG_JS) else: wc.log.debug(wc.LOG_JS, "JS read %d <= %s", len(data), url) self.js_script += data
|
return JSFilter(self.url, opts)
|
return JSFilter(self.url, self.localhost, opts)
|
def new_instance (self, **opts): """generate new JSFilter instance""" return JSFilter(self.url, opts)
|
except socket.error, err: if err==errno.EAGAIN: return self.handle_error('read error') return
|
def handle_read (self): """read data from SSL connection, put it into recv_buffer and call process_read""" assert self.connected debug(PROXY, '%s Connection.handle_read', self)
|
|
raise ValueError(i18n._("Invalid pics rule tag name `%s',"+\ " check your configuration")%name)
|
raise ValueError( i18n._("Invalid pics rule tag name `%s', check your configuration")%name)
|
def fill_attrs (self, attrs, name): if name=='service': self._service = unxmlify(attrs.get('name')).encode('iso8859-1') self.ratings[self._service] = {} elif name=='category': assert self._service self._category = unxmlify(attrs.get('name')).encode('iso8859-1') elif name=='url': pass elif name=='pics': UrlRule.fill_attrs(self, attrs, name) else: raise ValueError(i18n._("Invalid pics rule tag name `%s',"+\ " check your configuration")%name)
|
return False
|
return
|
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not category.valid_value(realvalue): error['categoryvalue'] = True return False if catname not in currule.ratings: currule.ratings[catname] = value info['rulecategory'] = True elif currule.ratings[catname] != value: currule.ratings[catname] = value info['rulecategory'] = True elif catname in currule.ratings: info['rulecategory'] = True del currule.ratings[catname]
|
return False if catname not in currule.ratings: currule.ratings[catname] = value info['rulecategory'] = True elif currule.ratings[catname] != value: currule.ratings[catname] = value info['rulecategory'] = True elif catname in currule.ratings:
|
return if currule.ratings[catname] != value: currule.ratings[catname] = value currule.compile_values()
|
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not category.valid_value(realvalue): error['categoryvalue'] = True return False if catname not in currule.ratings: currule.ratings[catname] = value info['rulecategory'] = True elif currule.ratings[catname] != value: currule.ratings[catname] = value info['rulecategory'] = True elif catname in currule.ratings: info['rulecategory'] = True del currule.ratings[catname]
|
del currule.ratings[catname]
|
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not category.valid_value(realvalue): error['categoryvalue'] = True return False if catname not in currule.ratings: currule.ratings[catname] = value info['rulecategory'] = True elif currule.ratings[catname] != value: currule.ratings[catname] = value info['rulecategory'] = True elif catname in currule.ratings: info['rulecategory'] = True del currule.ratings[catname]
|
|
assert wc.log.debug(wc.LOG_PROXY, "%s poll handle exception", x)
|
assert wc.log.debug(wc.LOG_PROXY, "poll handle exception %s", x)
|
def proxy_poll (timeout=0.0): """ Look for sockets with pending data and call the appropriate connection handlers. """ handlerCount = 0 if wc.proxy.Dispatcher.socket_map: e = wc.proxy.Dispatcher.socket_map.values() r = [x for x in e if x.readable()] w = [x for x in e if x.writable()] assert wc.log.debug(wc.LOG_PROXY, "select with %f timeout", timeout) try: (r, w, e) = select.select(r, w, e, timeout) except select.error, why: if why.args == (4, 'Interrupted system call'): # this occurs on UNIX systems with a sighup signal return else: raise assert wc.log.debug(wc.LOG_PROXY, "poll result %s", (r, w, e)) # Make sure we only process one type of event at a time, # because if something needs to close the connection we # don't want to call another handle_* on it for x in e: assert wc.log.debug(wc.LOG_PROXY, "%s poll handle exception", x) x.handle_expt_event() handlerCount += 1 for x in w: assert wc.log.debug(wc.LOG_PROXY, "poll handle write %s", x) # note: do not put the following "if" in a list filter if not x.writable(): assert wc.log.debug(wc.LOG_PROXY, "not writable %s", x) continue x.handle_write_event() handlerCount += 1 for x in r: assert wc.log.debug(wc.LOG_PROXY, "poll handle read %s", x) # note: do not put the following "if" in a list filter if not x.readable(): assert wc.log.debug(wc.LOG_PROXY, "not readable %s", x) continue x.handle_read_event() handlerCount += 1 return handlerCount
|
_form_proxypass(base64.encodestring(_getval(form, 'proxypass').strip()), res) else: if config['proxypass']:
|
val = _getval(form, 'proxypass') if val=='__dummy__': val = "" _form_proxypass(base64.encodestring(val).strip(), res) else: config['proxypass'] = '' if config['proxypass'] and config['proxyuser']:
|
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser').strip(), res) else: config['proxyuser'] = '' # proxy pass if form.has_key('proxypass'): _form_proxypass(base64.encodestring(_getval(form, 'proxypass').strip()), res) else: if config['proxypass']: res[0] = 407 config['proxypass'] = '' # parent proxy host if form.has_key('parentproxy'): _form_parentproxy(_getval(form, 'parentproxy').strip()) else: config['parentproxy'] = '' # parent proxy port if form.has_key('parentproxyport'): _form_parentproxyport(_getval(form, 'parentproxyport')) else: config['parentproxyport'] = 3128 # parent proxy user if form.has_key('parentproxyuser'): _form_parentproxyuser(_getval(form, 'parentproxyuser').strip()) else: config['parentproxyuser'] = '' # parent proxy pass if form.has_key('parentproxypass'): val = _getval(form, 'parentproxypass') if val=='__dummy__': # ignore the dummy value val = "" _form_parentproxypass(base64.encodestring(val)) else: config['parentproxypass'] = '' # timeout if form.has_key('timeout'): _form_timeout(_getval(form, 'timeout')) else: config['timeout'] = 0 # filter modules _form_filtermodules(form) # allowed hosts if form.has_key('addallowed') and form.has_key('newallowed'): _form_addallowed(_getval(form, 'newallowed').strip()) elif form.has_key('delallowed') and form.has_key('allowedhosts'): _form_delallowed(form) # no filter hosts if form.has_key('addnofilter') and form.has_key('newnofilter'): _form_addnofilter(_getval(form, 'newnofilter').strip()) elif form.has_key('delnofilter') and form.has_key('nofilterhosts'): _form_delnofilter(form) if info: # write changed config config.write_proxyconf() _daemon.reload() return res[0]
|
config['proxypass'] = ''
|
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser').strip(), res) else: config['proxyuser'] = '' # proxy pass if form.has_key('proxypass'): _form_proxypass(base64.encodestring(_getval(form, 'proxypass').strip()), res) else: if config['proxypass']: res[0] = 407 config['proxypass'] = '' # parent proxy host if form.has_key('parentproxy'): _form_parentproxy(_getval(form, 'parentproxy').strip()) else: config['parentproxy'] = '' # parent proxy port if form.has_key('parentproxyport'): _form_parentproxyport(_getval(form, 'parentproxyport')) else: config['parentproxyport'] = 3128 # parent proxy user if form.has_key('parentproxyuser'): _form_parentproxyuser(_getval(form, 'parentproxyuser').strip()) else: config['parentproxyuser'] = '' # parent proxy pass if form.has_key('parentproxypass'): val = _getval(form, 'parentproxypass') if val=='__dummy__': # ignore the dummy value val = "" _form_parentproxypass(base64.encodestring(val)) else: config['parentproxypass'] = '' # timeout if form.has_key('timeout'): _form_timeout(_getval(form, 'timeout')) else: config['timeout'] = 0 # filter modules _form_filtermodules(form) # allowed hosts if form.has_key('addallowed') and form.has_key('newallowed'): _form_addallowed(_getval(form, 'newallowed').strip()) elif form.has_key('delallowed') and form.has_key('allowedhosts'): _form_delallowed(form) # no filter hosts if form.has_key('addnofilter') and form.has_key('newnofilter'): _form_addnofilter(_getval(form, 'newnofilter').strip()) elif form.has_key('delnofilter') and form.has_key('nofilterhosts'): _form_delnofilter(form) if info: # write changed config config.write_proxyconf() _daemon.reload() return res[0]
|
|
_form_parentproxypass(base64.encodestring(val))
|
_form_parentproxypass(base64.encodestring(val).strip())
|
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser').strip(), res) else: config['proxyuser'] = '' # proxy pass if form.has_key('proxypass'): _form_proxypass(base64.encodestring(_getval(form, 'proxypass').strip()), res) else: if config['proxypass']: res[0] = 407 config['proxypass'] = '' # parent proxy host if form.has_key('parentproxy'): _form_parentproxy(_getval(form, 'parentproxy').strip()) else: config['parentproxy'] = '' # parent proxy port if form.has_key('parentproxyport'): _form_parentproxyport(_getval(form, 'parentproxyport')) else: config['parentproxyport'] = 3128 # parent proxy user if form.has_key('parentproxyuser'): _form_parentproxyuser(_getval(form, 'parentproxyuser').strip()) else: config['parentproxyuser'] = '' # parent proxy pass if form.has_key('parentproxypass'): val = _getval(form, 'parentproxypass') if val=='__dummy__': # ignore the dummy value val = "" _form_parentproxypass(base64.encodestring(val)) else: config['parentproxypass'] = '' # timeout if form.has_key('timeout'): _form_timeout(_getval(form, 'timeout')) else: config['timeout'] = 0 # filter modules _form_filtermodules(form) # allowed hosts if form.has_key('addallowed') and form.has_key('newallowed'): _form_addallowed(_getval(form, 'newallowed').strip()) elif form.has_key('delallowed') and form.has_key('allowedhosts'): _form_delallowed(form) # no filter hosts if form.has_key('addnofilter') and form.has_key('newnofilter'): _form_addnofilter(_getval(form, 'newnofilter').strip()) elif form.has_key('delnofilter') and form.has_key('nofilterhosts'): _form_delnofilter(form) if info: # write changed config config.write_proxyconf() _daemon.reload() return res[0]
|
res[0] = 407
|
def _form_proxyuser (proxyuser, res): if proxyuser != config['proxyuser']: res[0] = 407 config['proxyuser'] = proxyuser info['proxyuser'] = True
|
|
if proxypass != config['proxypass'] and proxypass!='__dummy__': res[0] = 407 config['proxypass'] = base64.encodestring(proxypass)
|
if proxypass != config['proxypass']: config['proxypass'] = proxypass
|
def _form_proxypass (proxypass, res): if proxypass != config['proxypass'] and proxypass!='__dummy__': res[0] = 407 config['proxypass'] = base64.encodestring(proxypass) info['proxypass'] = True
|
return os.path.join(directory, "_%s_configdata.py"%self.get_name())
|
return os.path.join(directory, "_%s2_configdata.py"%self.get_name())
|
def get_conf_filename (self, directory): return os.path.join(directory, "_%s_configdata.py"%self.get_name())
|
debug(PROXY, "expire servers")
|
debug(PROXY, "pool expire servers")
|
def expire_servers (self): """expire server connection that have been unused for too long""" debug(PROXY, "expire servers") expire_time = time.time() - 300 # Unused for five minutes to_expire = [] for addr,set in self.smap.items(): for server,status in set.items(): if status[0] == 'available' and status[1] < expire_time: # It's old .. let's get rid of it to_expire.append((addr,server)) for addr,server in to_expire: debug(PROXY, "expire %s server %s", addr, server) server.close() if addr in self.smap: assert not self.smap[addr].has_key(server), "Not expired: %s"%str(self.smap[addr]) make_timer(60, self.expire_servers)
|
Connect to addr. """
|
Connect to given address. """ debug("Client connect to %s" % str(addr))
|
def connect (self, addr): """ Connect to addr. """ self.socket = socket.socket() self.socket.connect(addr)
|
self.socket.sendall(data)
|
debug("Client send") socket_send(self.socket, data)
|
def send_data (self, data): """ Send complete data to socket. """ self.socket.sendall(data)
|
data = "" while wc.proxy.readable_socket(self.socket): s = self.socket.recv(4096) if not s: break data += s return data
|
debug("Client read") return socket_read(self.socket) class HttpServer (BaseHTTPServer.HTTPServer): """ Custom HTTP server class. """ def handle_request(self): """ Handle one request, possibly blocking. Exceptions are raised again for the test suite to handle. """ try: request, client_address = self.get_request() except socket.error: return if self.verify_request(request, client_address): try: self.process_request(request, client_address) except: self.handle_error(request, client_address) self.close_request(request) raise def handle_error(self, request, client_address): """ Suppress error printing. """ pass
|
def read_data (self): """ Read until no more data is available. """ data = "" while wc.proxy.readable_socket(self.socket): s = self.socket.recv(4096) if not s: break data += s return data
|
content = self.rfile.read(clen)
|
data = self.rfile.read(clen)
|
def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.headers["Content-Length"]) content = self.rfile.read(clen) else: content = "" return HttpRequest(method, uri, version, headers, content=content)
|
content = "" return HttpRequest(method, uri, version, headers, content=content)
|
data = socketfile_read(self.rfile) if "Transfer-Encoding" in self.headers: handler = TrailerHandler(headers) enctype = self.headers["Transfer-Encoding"] data = decode_transfer(enctype, data, handler) handler.handle_trailer() if "Content-Encoding" in self.headers: enctype = self.headers["Content-Encoding"] data = decode_content(enctype, data) return HttpRequest(method, uri, version, headers, content=data)
|
def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.headers["Content-Length"]) content = self.rfile.read(clen) else: content = "" return HttpRequest(method, uri, version, headers, content=content)
|
ServerClass = BaseHTTPServer.HTTPServer
|
ServerClass = HttpServer
|
def start_server (self): """ Start a HTTP server which is ready for use. @return: http server @rtype: BaseHTTPServer.HTTPServer """ port = 8000 server_address = ('', port) HandlerClass = HttpRequestHandler HandlerClass.protocol_version = "HTTP/1.1" ServerClass = BaseHTTPServer.HTTPServer httpd = ServerClass(server_address, HandlerClass) httpd.tester = self return httpd
|
Construct valid HTTP response data string.
|
Construct a HTTP response data string.
|
def construct_response_data (self, response): """ Construct valid HTTP response data string. """ lines = [] version = "HTTP/%d.%d" % response.version lines.append("%s %d %s" % (version, response.status, response.msg)) lines.extend(response.headers) # an empty line ends the headers lines.extend(("", "")) data = "\r\n".join(lines) if response.content: data += response.content return data
|
parts = ['HTTP/1.0', 200, 'Ok']
|
parts = ['HTTP/1.0', "200", 'Ok']
|
def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r from %r", response, url) parts = ['HTTP/1.0', 200, 'Ok'] if not is_http_status(parts[1]): wc.log.error(wc.LOG_PROXY, "Invalid http statuscode %r from %r", parts[1], url) parts[1] = 200 parts[1] = int(parts[1]) return parts
|
parts[1] = 200
|
parts[1] = "200"
|
def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r from %r", response, url) parts = ['HTTP/1.0', 200, 'Ok'] if not is_http_status(parts[1]): wc.log.error(wc.LOG_PROXY, "Invalid http statuscode %r from %r", parts[1], url) parts[1] = 200 parts[1] = int(parts[1]) return parts
|
("\nname: %r\nvalue: %r" % (self.name, self.value))
|
("name %r\nvalue %r\nstage %s" % (self.name, self.value, self.filterstage))
|
def __str__ (self): """return rule data as string""" return super(HeaderRule, self).__str__() + \ ("\nname: %r\nvalue: %r" % (self.name, self.value))
|
wc.config = wc.Configuration(confdir=confdir) wc.config.init_filter_modules()
|
wc.configuration.config = wc.configuration.Configuration(confdir=confdir) wc.configuration.config.init_filter_modules() wc.proxy.dns_lookups.init_resolver()
|
def _main (): """USAGE: test/run.sh test/filterfile.py <config dir> <.html file>""" if len(sys.argv)!=3: print _main.__doc__ sys.exit(1) confdir = sys.argv[1] fname = sys.argv[2] if fname=="-": f = sys.stdin else: f = file(fname) logfile = os.path.join(confdir, "logging.conf") wc.initlog(logfile, wc.Name, filelogs=False) wc.config = wc.Configuration(confdir=confdir) wc.config.init_filter_modules() headers = wc.proxy.Headers.WcMessage() headers['Content-Type'] = "text/html" attrs = wc.filter.get_filterattrs(fname, [wc.filter.FILTER_RESPONSE_MODIFY], headers=headers, serverheaders=headers) filtered = "" data = f.read(2048) while data: print >>sys.stderr, "Test: data", len(data) try: filtered += wc.filter.applyfilter(wc.filter.FILTER_RESPONSE_MODIFY, data, 'filter', attrs) except wc.filter.FilterException, msg: print >>sys.stderr, "Test: exception:", msg pass data = f.read(2048) print >>sys.stderr, "Test: finishing" i = 1 while True: print >>sys.stderr, "Test: finish", i try: filtered += wc.filter.applyfilter(wc.filter.FILTER_RESPONSE_MODIFY, "", 'finish', attrs) break except wc.filter.FilterException, msg: print >>sys.stderr, "Test: finish: exception:", msg wc.proxy.proxy_poll(timeout=max(0, wc.proxy.run_timers())) i += 1 if i==200: # background downloading if javascript is too slow print >>sys.stderr, "Test: oooooops" break print "Filtered:", filtered
|
self._debug(NIGHTMARE, "JS: document.write", `data`)
|
def jsProcessData (self, data): """process data produced by document.write() JavaScript""" self._debug(NIGHTMARE, "JS: document.write", `data`) self.js_output += 1 # parse recursively self.js_html.feed(data)
|
|
self._debug(NIGHTMARE, "JS: popup")
|
def jsProcessPopup (self): """process javascript popup""" self._debug(NIGHTMARE, "JS: popup") self.js_popup += 1
|
|
self._debug(NIGHTMARE, "buf_append_data")
|
def buf_append_data (self, data): """we have to make sure that we have no two following DATA things in the tag buffer. Why? To be 100% sure that an ENCLOSED match really matches enclosed data. """ self._debug(NIGHTMARE, "buf_append_data") if data[0]==DATA and self.buf and self.buf[-1][0]==DATA: self.buf[-1][1] += data[1] else: self.buf.append(data)
|
|
self._debug(NIGHTMARE, "flushbuf")
|
def flushbuf (self): """clear and return the output buffer""" self._debug(NIGHTMARE, "flushbuf") data = self.outbuf.getvalue() self.outbuf.close() self.outbuf = StringIO() return data
|
|
self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`)
|
def _debugbuf (self): """print debugging information about data buffer status""" self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`)
|
|
self._debug(NIGHTMARE, "feed", `data`)
|
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() self.inbuf = StringIO() if data: # only feed non-empty data self._debug(NIGHTMARE, "feed", `data`) self.parser.feed(data) else: self._debug(NIGHTMARE, "feed") pass else: # wait state --> put in input buffer self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
|
|
self._debug(NIGHTMARE, "feed")
|
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() self.inbuf = StringIO() if data: # only feed non-empty data self._debug(NIGHTMARE, "feed", `data`) self.parser.feed(data) else: self._debug(NIGHTMARE, "feed") pass else: # wait state --> put in input buffer self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
|
|
self._debug(NIGHTMARE, "wait")
|
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() self.inbuf = StringIO() if data: # only feed non-empty data self._debug(NIGHTMARE, "feed", `data`) self.parser.feed(data) else: self._debug(NIGHTMARE, "feed") pass else: # wait state --> put in input buffer self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
|
|
self._debug(HURT_ME_PLENTY, "flush")
|
def flush (self): self._debug(HURT_ME_PLENTY, "flush") # flushing in wait state raises a filter exception if self.state=='wait': raise FilterWait("HtmlParser[%d]: waiting for data"%self.level) self.parser.flush()
|
|
self._debug(NIGHTMARE, "replay", waitbuf)
|
def replay (self, waitbuf): """call the handler functions again with buffer data""" self._debug(NIGHTMARE, "replay", waitbuf) for item in waitbuf: if item[0]==DATA: self._data(item[1]) elif item[0]==STARTTAG: self.startElement(item[1], item[2]) elif item[0]==ENDTAG: self.endElement(item[1]) elif item[0]==COMMENT: self.comment(item[1])
|
|
self._debug(NIGHTMARE, "cdata", `data`)
|
def cdata (self, data): """character data""" self._debug(NIGHTMARE, "cdata", `data`) return self._data(data)
|
|
self._debug(NIGHTMARE, "characters", `data`)
|
def characters (self, data): """characters""" self._debug(NIGHTMARE, "characters", `data`) return self._data(data)
|
|
self._debug(NIGHTMARE, "comment", `data`)
|
def comment (self, data): """a comment; accept only non-empty comments""" self._debug(NIGHTMARE, "comment", `data`) item = [COMMENT, data] if self.state=='wait': return self.waitbuf.append(item) if self.comments and data: self.buf.append(item)
|
|
self._debug(NIGHTMARE, "doctype", `data`)
|
def doctype (self, data): self._debug(NIGHTMARE, "doctype", `data`) return self._data("<!DOCTYPE%s>"%data)
|
|
self._debug(NIGHTMARE, "pi", `data`)
|
def pi (self, data): self._debug(NIGHTMARE, "pi", `data`) return self._data("<?%s?>"%data)
|
|
self._debug(NIGHTMARE, "startElement", `tag`)
|
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 if tag=="meta" and \ attrs.get('http-equiv', '').lower() =='pics-label': labels = resolve_html_entities(attrs.get('content', '')) # note: if there are no pics rules, this loop is empty for rule in self.pics: msg = check_pics(rule, labels) if msg: raise FilterPics(msg) # first labels match counts self.pics = [] elif tag=="body": # headers finished if self.pics: # no pics data found self.pics = [] elif tag=="base" and attrs.has_key('href'): self.base_url = resolve_html_entities(attrs['href']) # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item # give'em a chance to replace more than one attribute continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
|
|
self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`))
|
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 if tag=="meta" and \ attrs.get('http-equiv', '').lower() =='pics-label': labels = resolve_html_entities(attrs.get('content', '')) # note: if there are no pics rules, this loop is empty for rule in self.pics: msg = check_pics(rule, labels) if msg: raise FilterPics(msg) # first labels match counts self.pics = [] elif tag=="body": # headers finished if self.pics: # no pics data found self.pics = [] elif tag=="base" and attrs.has_key('href'): self.base_url = resolve_html_entities(attrs['href']) # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item # give'em a chance to replace more than one attribute continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
|
|
self._debug(NIGHTMARE, "put on buffer")
|
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 if tag=="meta" and \ attrs.get('http-equiv', '').lower() =='pics-label': labels = resolve_html_entities(attrs.get('content', '')) # note: if there are no pics rules, this loop is empty for rule in self.pics: msg = check_pics(rule, labels) if msg: raise FilterPics(msg) # first labels match counts self.pics = [] elif tag=="body": # headers finished if self.pics: # no pics data found self.pics = [] elif tag=="base" and attrs.has_key('href'): self.base_url = resolve_html_entities(attrs['href']) # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item # give'em a chance to replace more than one attribute continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
|
|
self._debug(NIGHTMARE, "endElement", `tag`)
|
def endElement (self, tag): """We know the following: if a rule matches, it must be the one on the top of the stack. So we look only at the top rule.
|
|
self._debug(NIGHTMARE, "JS: del", `name`, "from", `tag`)
|
def jsStartElement (self, tag, attrs): """Check popups for onmouseout and onmouseover. Inline extern javascript sources""" changed = 0 self.js_src = None self.js_output = 0 self.js_popup = 0 for name in ('onmouseover', 'onmouseout'): if attrs.has_key(name) and self.jsPopup(attrs, name): self._debug(NIGHTMARE, "JS: del", `name`, "from", `tag`) del attrs[name] changed = 1 if tag=='form': name = attrs.get('name', attrs.get('id')) self.jsForm(name, attrs.get('action', ''), attrs.get('target', '')) elif tag=='script': lang = attrs.get('language', '').lower() url = attrs.get('src', '') scrtype = attrs.get('type', '').lower() is_js = scrtype=='text/javascript' or \ lang.startswith('javascript') or \ not (lang or scrtype) if is_js and url: return self.jsScriptSrc(url, lang) self.buf.append([STARTTAG, tag, attrs])
|
|
self._debug(NIGHTMARE, "JS: jsPopup")
|
def jsPopup (self, attrs, name): """check if attrs[name] javascript opens a popup window""" self._debug(NIGHTMARE, "JS: jsPopup") val = resolve_html_entities(attrs[name]) if not val: return self.js_env.attachListener(self) try: self.js_env.executeScriptAsFunction(val, 0.0) except jslib.error, msg: pass self.js_env.detachListener(self) res = self.js_popup self.js_popup = 0 return res
|
|
self._debug(HURT_ME_PLENTY, "jsForm", `name`, `action`, `target`)
|
def jsForm (self, name, action, target): """when hitting a (named) form, notify the JS engine about that""" if not name: return self._debug(HURT_ME_PLENTY, "jsForm", `name`, `action`, `target`) self.js_env.addForm(name, action, target)
|
|
self._debug(NIGHTMARE, "switching back to parse with")
|
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) script = "\n<!--\n%s\n//-->\n"%escape_js(self.js_script) self.buf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.buf.append([ENDTAG, "script"]) self.js_script = '' self.state = 'parse' self._debug(NIGHTMARE, "switching back to parse with") self._debugbuf() else: self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url) self.js_script += data
|
|
self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url)
|
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) script = "\n<!--\n%s\n//-->\n"%escape_js(self.js_script) self.buf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.buf.append([ENDTAG, "script"]) self.js_script = '' self.state = 'parse' self._debug(NIGHTMARE, "switching back to parse with") self._debugbuf() else: self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url) self.js_script += data
|
|
self._debug(HURT_ME_PLENTY, "JS jsScriptSrc", `url`, `ver`)
|
def jsScriptSrc (self, url, language): """Start a background download for <script src=""> tags""" assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) if self.base_url: url = urlparse.urljoin(self.base_url, url) else: url = urlparse.urljoin(self.url, url) self._debug(HURT_ME_PLENTY, "JS jsScriptSrc", `url`, `ver`) if _has_ws(url): print >> sys.stderr, "HtmlParser[%d]: broken JS url"%self.level,\ `url`, "at", `self.url` return self.state = 'wait' self.waited = 'True' self.js_src = 'True' client = HttpProxyClient(self.jsScriptData, (url, ver)) ClientServerMatchmaker(client, "GET %s HTTP/1.1" % url, #request {}, #headers '', #content {'nofilter': None}, # nofilter 'identity', # compress mime = "application/x-javascript", )
|
|
self._debug(NIGHTMARE, "JS: jsScript", ver, `script`)
|
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = FilterHtmlParser(self.rules, self.pics, self.url, comments=self.comments, javascript=self.js_filter, level=self.level+1) # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.detachListener(self) # wait for recursive filter to finish self.jsEndScript(item)
|
|
self._debug(NIGHTMARE, "JS: endScript")
|
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.js_html.inbuf.getvalue() assert not self.js_html.waitbuf assert len(self.buf) >= 2 self.buf[-2:-2] = [[DATA, self.js_html.outbuf.getvalue()]]+self.js_html.buf self.js_html = None if (self.js_popup + self.js_output) > 0: # delete old script del self.buf[-1] del self.buf[-1] elif not self.filterEndElement(item[1]): self.buf.append(item) self._debug(NIGHTMARE, "JS: switching back to parse with") self._debugbuf() self.state = 'parse'
|
|
self._debug(NIGHTMARE, "JS: switching back to parse with")
|
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.js_html.inbuf.getvalue() assert not self.js_html.waitbuf assert len(self.buf) >= 2 self.buf[-2:-2] = [[DATA, self.js_html.outbuf.getvalue()]]+self.js_html.buf self.js_html = None if (self.js_popup + self.js_output) > 0: # delete old script del self.buf[-1] del self.buf[-1] elif not self.filterEndElement(item[1]): self.buf.append(item) self._debug(NIGHTMARE, "JS: switching back to parse with") self._debugbuf() self.state = 'parse'
|
|
assert None == wc.log.debug(wc.LOG_FILTER, "..not applying")
|
assert None == wc.log.debug(wc.LOG_FILTER, "..not applying (mime %s)" % attrs['mime'])
|
def applyfilter (filterstage, data, fun, attrs): """ Apply all filters which are registered in the given filter stage. For different filter stages we have different data objects. Look at the filter examples. One can prevent all filtering with the 'nofilter' attribute, or deactivate single filter modules with 'nofilter-<name>', for example 'nofilter-blocker'. """ attrs['filterstage'] = filterstage assert None == wc.log.debug(wc.LOG_FILTER, "Filter (%s) %d bytes in %s..", fun, len(data), filterstage) if attrs.get('nofilter') or (fun != 'finish' and not data): assert None == wc.log.debug(wc.LOG_FILTER, "..don't filter") return data for f in wc.configuration.config['filterlist'][filterstage]: assert None == wc.log.debug(wc.LOG_FILTER, "..filter with %s" % f) if f.applies_to_mime(attrs['mime']) and \ not "nofilter-%s" % str(f).lower() in attrs: assert None == wc.log.debug(wc.LOG_FILTER, "..applying") data = getattr(f, fun)(data, attrs) else: assert None == wc.log.debug(wc.LOG_FILTER, "..not applying") assert None == wc.log.debug(wc.LOG_FILTER, "..result %d bytes", len(data)) return data
|
msg, status = _stop(watchfile)
|
msg1, status1 = _stop(watchfile)
|
def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg, status = _stop(watchfile) msg2, status2 = stop(pidfile) return "%s\n%s"%(msg,msg2), status2
|
return "%s\n%s"%(msg,msg2), status2
|
if msg1: if msg2: msg = "%s\n%s"%(msg1,msg2) else: msg = msg1 else: msg = msg2 if status1: status = status1 else: status = status2 return msg, status
|
def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg, status = _stop(watchfile) msg2, status2 = stop(pidfile) return "%s\n%s"%(msg,msg2), status2
|
(wc.ConfigCharset, super(FolderRule, self).toxml(), self.oid)
|
(wc.configuration.ConfigCharset, super(FolderRule, self).toxml(), self.oid)
|
def toxml (self): """Rule data as XML for storing""" s = u"""<?xml version="1.0" encoding="%s"?>
|
if self.headers.has_key('Content-Length'): self.bytes_remaining = int(self.headers['Content-Length'])
|
def process_headers (self): i = self.recv_buffer.find('\r\n\r\n') if i >= 0: # Two newlines ends headers i += 4 # Skip over newline terminator # the first 2 chars are the newline of request data = self.read(i)[2:] self.headers = rfc822.Message(StringIO(data)) #debug(HURT_ME_PLENTY, "Proxy: C/Headers", `self.headers.headers`) # set via header via = self.headers.get('Via', "").strip() if via: via += " " via += "1.1 unknown\r" self.headers['Via'] = via self.headers = applyfilter(FILTER_REQUEST_HEADER, self.headers, fun="finish", attrs=self.nofilter) # remember if client understands gzip self.compress = 'identity' encodings = self.headers.get('Accept-Encoding', '') for accept in encodings.split(','): if ';' in accept: accept, q = accept.split(';', 1) if accept.strip().lower() in ('gzip', 'x-gzip'): self.compress = 'gzip' break # we understand gzip, deflate and identity self.headers['Accept-Encoding'] = \ 'gzip;q=1.0, deflate;q=0.9, identity;q=0.5\r' # add decoders self.decoders = [] # Chunked encoded if self.headers.get('Transfer-Encoding') is not None: debug(BRING_IT_ON, 'Proxy: C/Transfer-encoding:', `self.headers['transfer-encoding']`) self.decoders.append(UnchunkStream()) # remove encoding header to_remove = ["Transfer-Encoding"] if self.headers.get("Content-Length") is not None: print >>sys.stderr, 'Warning: chunked encoding should not have Content-Length' to_remove.append("Content-Length") self.bytes_remaining = None remove_headers(self.headers, to_remove) # add warning self.headers['Warning'] = "214 Transformation applied\r" debug(HURT_ME_PLENTY, "Proxy: C/Headers", `str(self.headers)`) if self.headers.has_key('Content-Length'): self.bytes_remaining = int(self.headers['Content-Length']) if config["proxyuser"] and not self.check_proxy_auth(): return self.error(407, i18n._("Proxy Authentication Required")) if self.method=='OPTIONS': mf = int(self.headers.get('Max-Forwards', -1)) if mf==0: # XXX display options ? self.state = 'done' ServerHandleDirectly(self, 'HTTP/1.0 200 OK\r\n', 'Content-Type: text/plain\r\n\r\n', '') return if mf>0: self.headers['Max-Forwards'] = mf-1 self.state = 'content'
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.