rem
stringlengths
0
322k
add
stringlengths
0
2.05M
context
stringlengths
8
228k
if front(str): start,end = front(str).span() return end-start-1
result = front(str) if result is not None: start, end = result.span() return end-start
def indention(str,front = re.compile("^\s+").match): """ Convert all tabs to the appropriate number of spaces. Find the number of leading spaces. If none, return 0 """ if front(str): start,end = front(str).span() return end-start-1 else: return 0 # no leading spaces
expires=http_date(time.time() + seconds)
expires=rfc1123_date(time.time() + seconds)
def ZCache_set(self, ob, data, view_name, keywords, mtime_func): # Note the blatant ignorance of view_name, keywords, and # mtime_func. Standard HTTP accelerators are not able to make # use of this data. REQUEST = ob.REQUEST RESPONSE = REQUEST.RESPONSE anon = 1 u = REQUEST.get('AUTHENTICATED_USER', None) if u is not None: if u.getUserName() != 'Anonymous User': anon = 0 phys_path = ob.getPhysicalPath() if self.hit_counts.has_key(phys_path): hits = self.hit_counts[phys_path] else: self.hit_counts[phys_path] = hits = [0,0] if anon: hits[0] = hits[0] + 1 else: hits[1] = hits[1] + 1
print "FAILED to write INFO event", event, ":", details
try: print "FAILED to write INFO event", event, ":", details except IOError: pass
def logmsg(self, event): # log a service event using servicemanager.LogMsg try: servicemanager.LogMsg(servicemanager.EVENTLOG_INFORMATION_TYPE, event, (self._svc_name_, " (%s)" % self._svc_display_name_)) except win32api.error, details: # Failed to write a log entry - most likely problem is # that the event log is full. We don't want this to kill us print "FAILED to write INFO event", event, ":", details
print "FAILED to write event log entry:", details print msg
try: print "FAILED to write event log entry:", details print msg except IOError: pass
def _dolog(self, func, msg): try: fullmsg = "%s (%s): %s" % \ (self._svc_name_, self._svc_display_name_, msg) func(fullmsg) except win32api.error, details: # Failed to write a log entry - most likely problem is # that the event log is full. We don't want this to kill us print "FAILED to write event log entry:", details print msg
'GMT-0800','GMT-0900','GMT-1000','GMT-1100','GMT-1200']
'GMT-0800','GMT-0900','GMT-1000','GMT-1100','GMT-1200', 'GMT+1']
def info(self,t=None): idx=self.index(t)[0] zs =self.az[self.tinfo[idx][2]:] return self.tinfo[idx][0],self.tinfo[idx][1],zs[:find(zs,'\000')]
'gmt+1300':'GMT+13','gmt-0100':'GMT-1','gmt-0200':'GMT-2', 'gmt-0300':'GMT-3','gmt-0400':'GMT-4','gmt-0500':'GMT-5', 'gmt-0600':'GMT-6','gmt-0700':'GMT-7','gmt-0800':'GMT-8', 'gmt-0900':'GMT-9','gmt-1000':'GMT-10','gmt-1100':'GMT-11', 'gmt-1200':'GMT-12','greenwich':'Greenwich','hongkong':'Hongkong',
'gmt+1300':'GMT+13', 'gmt-0100':'GMT-1', 'gmt-0200':'GMT-2', 'gmt-0300':'GMT-3', 'gmt-0400':'GMT-4', 'gmt-0500':'GMT-5', 'gmt-0600':'GMT-6', 'gmt-0700':'GMT-7', 'gmt-0800':'GMT-8', 'gmt-0900':'GMT-9', 'gmt-1000':'GMT-10','gmt-1100':'GMT-11','gmt-1200':'GMT-12', 'gmt+1': 'GMT+1', 'gmt+2': 'GMT+2', 'gmt+3': 'GMT+3', 'gmt+4': 'GMT+4', 'gmt+5': 'GMT+5', 'gmt+6': 'GMT+6', 'gmt+7': 'GMT+7', 'gmt+8': 'GMT+8', 'gmt+9': 'GMT+9', 'gmt+10':'GMT+10','gmt+11':'GMT+11','gmt+12':'GMT+12', 'gmt+13':'GMT+13', 'gmt-1': 'GMT-1', 'gmt-2': 'GMT-2', 'gmt-3': 'GMT-3', 'gmt-4': 'GMT-4', 'gmt-5': 'GMT-5', 'gmt-6': 'GMT-6', 'gmt-7': 'GMT-7', 'gmt-8': 'GMT-8', 'gmt-9': 'GMT-9', 'gmt-10':'GMT-10','gmt-11':'GMT-11','gmt-12':'GMT-12', 'greenwich':'Greenwich','hongkong':'Hongkong',
def info(self,t=None): idx=self.index(t)[0] zs =self.az[self.tinfo[idx][2]:] return self.tinfo[idx][0],self.tinfo[idx][1],zs[:find(zs,'\000')]
for k in data.keys(None, delete_end):
for k in list(data.keys(None, delete_end)):
def _housekeep(self, delete_end): """ do garbage collection, bucket replentishing and notification """ data = self._data period = self._period min_spares = self._min_spare_buckets DEBUG and TLOG( '_housekeep: current slice %s' % self._getCurrentTimeslice() ) notify = {} while 1: try: k, v = self.notify_queue.get_nowait() # duplicates will be ignored notify[k] = v except Queue.Empty: break
mod_since=int(DateTime(header).timeTime())
mod_since=long(DateTime(header).timeTime())
def index_html(self, REQUEST, RESPONSE): """ The default view of the contents of a File or Image.
last_mod = int(self._p_mtime)
last_mod = long(self._p_mtime)
def index_html(self, REQUEST, RESPONSE): """ The default view of the contents of a File or Image.
last_mod = 0
last_mod = long(0)
def index_html(self, REQUEST, RESPONSE): """ The default view of the contents of a File or Image.
def _compare(dt1, dt2): '''Compares the internal representation of dt1 with the representation in dt2. Allows sub-millisecond variations. Primarily for testing.''' assert dt1.millis() == dt2.millis(), \ '%s != %s' % (dt1.millis(),dt2.millis()) assert math.floor(dt1._t * 1000.0) == \ math.floor(dt2._t * 1000.0) assert math.floor(dt1._d * 86400000.0) == \ math.floor(dt2._d * 86400000.0) assert math.floor(dt1.time * 86400000.0) == \ math.floor(dt2.time * 86400000.0)
def _compare(dt1, dt2): '''Compares the internal representation of dt1 with the representation in dt2. Allows sub-millisecond variations. Primarily for testing.''' assert dt1.millis() == dt2.millis(), \ '%s != %s' % (dt1.millis(),dt2.millis()) assert math.floor(dt1._t * 1000.0) == \ math.floor(dt2._t * 1000.0) assert math.floor(dt1._d * 86400000.0) == \ math.floor(dt2._d * 86400000.0) assert math.floor(dt1.time * 86400000.0) == \ math.floor(dt2.time * 86400000.0)
_compare(dt, dt1)
self.assertEqual(repr(dt),repr(dt1))
def testConstructor3(self): '''Constructor from date/time string''' dt = DateTime() dt1s = '%d/%d/%d %d:%d:%f %s' % ( dt.year(), dt.month(), dt.day(), dt.hour(), dt.minute(), dt.second(), dt.timezone()) dt1 = DateTime(dt1s) _compare(dt, dt1)
assert dt.debugCompare(dt1), (dt, dt1)
self._compare(dt,dt1)
def testConstructor4(self): '''Constructor from time float''' dt = DateTime() dt1 = DateTime(float(dt)) assert dt.debugCompare(dt1), (dt, dt1)
_compare(dt, dt1)
self._compare(dt, dt1)
def testConstructor6(self): '''Constructor from year and julian date''' # This test must normalize the time zone, or it *will* break when # DST changes! dt1 = DateTime(2000, 5.500000578705) dt = DateTime('2000/1/5 12:00:00.050 pm %s' % dt1.localZone()) _compare(dt, dt1)
assert dt.debugCompare(dt1), (dt, dt1)
self.assertEqual(repr(dt),repr(dt1))
def testConstructor7(self): '''Constructor from parts''' dt = DateTime() dt1 = DateTime( dt.year(), dt.month(), dt.day(), dt.hour(), dt.minute(), dt.second(), dt.timezone()) assert dt.debugCompare(dt1), (dt, dt1)
em=regex.compile("[ \t\n]\*\([^ \t][^\n*]*[^ \t]\)\*\([ \t\n,.:;!?]\)") code=regex.compile("[ \t\n(]'\([^ \t']\([^\n']*[^ \t']\)?\)'\([) \t\n,.:;!?]\)") strong=regex.compile("[ \t\n]\*\*\([^ \t][^\n*]*[^ \t]\)\*\*\([ \t\n,.:;!?]\)") extra_dl=regex.compile("</dl>\n<dl>") extra_ul=regex.compile("</ul>\n<ul>") extra_ol=regex.compile("</ol>\n<ol>")
def structure(list): if not list: return [] i=0 l=len(list) r=[] while i < l: sublen=paragraphs(list,i) i=i+1 r.append((list[i-1][1],structure(list[i:i+sublen]))) i=i+sublen return r
s=regsub.gsub(strong,' <strong>\\1</strong>\\2',s) s=regsub.gsub(code,' <code>\\1</code>\\3',s) s=regsub.gsub(em,' <em>\\1</em>\\2',s)
s=gsub(strong,'\\1<strong>\\2</strong>\\3',s) s=gsub(code, '\\1<code>\\2</code>\\3',s) s=gsub(em, '\\1<em>\\2</em>\\3',s)
def ctag(s): if s is None: s='' s=regsub.gsub(strong,' <strong>\\1</strong>\\2',s) s=regsub.gsub(code,' <code>\\1</code>\\3',s) s=regsub.gsub(em,' <em>\\1</em>\\2',s) return s
def __str__(self):
def __str__(self, extra_dl=regex.compile("</dl>\n<dl>"), extra_ul=regex.compile("</ul>\n<ul>"), extra_ol=regex.compile("</ol>\n<ol>"), ):
def __str__(self):
s=regsub.gsub(extra_dl,'\n',s) s=regsub.gsub(extra_ul,'\n',s) s=regsub.gsub(extra_ol,'\n',s)
s=gsub(extra_dl,'\n',s) s=gsub(extra_ul,'\n',s) s=gsub(extra_ol,'\n',s)
def __str__(self):
elif nl.search(s[0]) < 0 and s[1]:
elif nl.search(s[0]) < 0 and s[1] and s[0][-1:] != ':':
def _str(self,structure,level):
r='<XMP>\n'
r='<PRE>\n'
def pre(self,structure,tagged=0):
r="%s%s\n\n%s" % (r,s[0],self.pre(s[1],1)) if not tagged: r=r+'</XMP>\n'
r="%s%s\n\n%s" % (r,html_quote(s[0]),self.pre(s[1],1)) if not tagged: r=r+'</PRE>\n'
def pre(self,structure,tagged=0):
import regsub text = regsub.gsub(
text = gsub(
def html_with_references(text): import regsub text = regsub.gsub( '[\0\n].. \[\([-_0-9_a-zA-Z]+\)\]', '\n <a name="\\1">[\\1]</a>', text) text = regsub.gsub( '\([\0- ,]\)\[\([0-9_a-zA-Z]+\)\]\([\0- ,.:]\)', '\\1<a href="#\\2">[\\2]</a>\\3', text) text = regsub.gsub( '\([\0- ]\)\([a-z]+://[^\0- ]+\)', '\\1<a href="\\2">\\2</a>', text) return HTML(text,level=1)
text = regsub.gsub(
text = gsub(
def html_with_references(text): import regsub text = regsub.gsub( '[\0\n].. \[\([-_0-9_a-zA-Z]+\)\]', '\n <a name="\\1">[\\1]</a>', text) text = regsub.gsub( '\([\0- ,]\)\[\([0-9_a-zA-Z]+\)\]\([\0- ,.:]\)', '\\1<a href="#\\2">[\\2]</a>\\3', text) text = regsub.gsub( '\([\0- ]\)\([a-z]+://[^\0- ]+\)', '\\1<a href="\\2">\\2</a>', text) return HTML(text,level=1)
print html_with_references(sys.stdin.read())
if '-t' in sys.argv: import regex, string s=sys.stdin.read() r=regex.compile('\([\0-\n]*\n\)') if r.match(s) >= 0: s=s[len(r.group(1)):] s=str(html_with_references(s)) if s[:4]=='<h1>': t=s[4:string.find(s,'</h1>')] s='''<html><head><title>%s</title> </head><body> %s </body></html> ''' % (t,s) print s else: print html_with_references(sys.stdin.read())
def main(): import sys print html_with_references(sys.stdin.read())
def document_src(self, REQUEST, RESPONSE):
def document_src(self, REQUEST=None, RESPONSE=None):
def document_src(self, REQUEST, RESPONSE): """Return unprocessed document source.""" RESPONSE.setHeader('Content-Type', 'text/plain') return self.read()
RESPONSE.setHeader('Content-Type', 'text/plain')
if RESPONSE is not None: RESPONSE.setHeader('Content-Type', 'text/plain')
def document_src(self, REQUEST, RESPONSE): """Return unprocessed document source.""" RESPONSE.setHeader('Content-Type', 'text/plain') return self.read()
if mtime == self._v_last_read: return
if hasattr(self, '_v_program') and mtime == self._v_last_read: return
def _cook_check(self): if self._v_last_read and not DevelopmentMode: return __traceback_info__ = self.filename try: mtime=os.stat(self.filename)[8] except: mtime=0 if mtime == self._v_last_read: return self.pt_edit(open(self.filename), None) self._cook() if self._v_errors: LOG('PageTemplateFile', ERROR, 'Error in template', join(self._v_errors, '\n')) return self._v_last_read = mtime
if hasattr(self, idattr): id=getattr(self, idattr)
if hasattr(item, idattr): id=getattr(item, idattr)
def tpValuesIds(self, branches, args, simple_type={type(''):0, type(1):0, type(1.0):0}.has_key, ): # This should build the ids of subitems which are # expandable (non-empty). Leaves should never be # in the state - it will screw the colspan counting. r=[] idattr=args['id'] try: try: items=getattr(self, branches)() except AttributeError: items=() for item in items: try: if getattr(item, branches)(): if hasattr(self, idattr): id=getattr(self, idattr) if not simple_type(type(id)): id=id() elif hasattr(item, '_p_oid'): id=oid(item) else: id=pyid(item) e=tpValuesIds(item, branches, args) if e: id=[id,e] else: id=[id] r.append(id) except: pass except: pass return r
def set_cdata_mode(self):
cdata_endtag = None def set_cdata_mode(self, endtag=None): self.cdata_endtag = endtag
def set_cdata_mode(self): self.interesting = interesting_cdata
self.set_cdata_mode()
self.set_cdata_mode(tag)
def parse_starttag(self, i): self.__starttag_text = None endpos = self.check_for_whole_start_tag(i) if endpos < 0: return endpos rawdata = self.rawdata self.__starttag_text = rawdata[i:endpos]
tag = match.group(1) self.handle_endtag(tag.lower())
tag = match.group(1).lower() if ( self.cdata_endtag is not None and tag != self.cdata_endtag): self.handle_data(match.group(0)) else: self.handle_endtag(tag)
def parse_endtag(self, i): rawdata = self.rawdata assert rawdata[i:i+2] == "</", "unexpected call to parse_endtag" match = endendtag.search(rawdata, i+1) # > if not match: return -1 j = match.end() match = endtagfind.match(rawdata, i) # </ + tag + > if not match: self.error("bad end tag: %s" % `rawdata[i:j]`) tag = match.group(1) self.handle_endtag(tag.lower()) return j
result = pw_validate(passwrd, password) if not result: LOG('Zope', WARNING, 'invalid password for user %s' % self.getUserName())
result = pw_validate(passwrd, password)
def authenticate(self, password, request): passwrd=self._getPassword()
if domains and result: result = domainSpecMatch(domains, request) if not result: LOG('Zope', WARNING, 'invalid domain for user %s' % self.getUserName()) return result
if domains: return result and domainSpecMatch(domains, request)
def authenticate(self, password, request): passwrd=self._getPassword()
LOG('Zope', WARNING, 'non-basic authentication attempted')
def validate(self,request,auth='',roles=None):
LOG('Zope', WARNING, 'no such user as %s in %s' \ % (name, self.absolute_url()))
def validate(self,request,auth='',roles=None):
LOG('Zope', WARNING, 'user %s is not authorized to access %s' \ % (name, parent.absolute_url()))
def validate(self,request,auth='',roles=None):
LOG('Zope', WARNING, 'no such user as %s in %s' \ % (name, self.absolute_url()))
def validate(self,request,auth='',roles=None): parent=request['PARENTS'][0] e=request.environ if e.has_key('REMOTE_USER'): name=e['REMOTE_USER'] else: for ob in self.getUsers(): domains=ob.getDomains() if domains: if ob.authenticate('', request): if ob.allowed(parent, roles): ob=ob.__of__(self) return ob nobody=self._nobody if self._isTop() and nobody.allowed(parent, roles): ob=nobody.__of__(self) return ob return None
LOG('Zope', WARNING, 'user %s is not authorized to access %s' \ % (name, parent.absolute_url()))
def validate(self,request,auth='',roles=None): parent=request['PARENTS'][0] e=request.environ if e.has_key('REMOTE_USER'): name=e['REMOTE_USER'] else: for ob in self.getUsers(): domains=ob.getDomains() if domains: if ob.authenticate('', request): if ob.allowed(parent, roles): ob=ob.__of__(self) return ob nobody=self._nobody if self._isTop() and nobody.allowed(parent, roles): ob=nobody.__of__(self) return ob return None
<TABLE BORDER="0" WIDTH="100%"> <TR VALIGN="TOP"> <TD WIDTH="10%" ALIGN="CENTER">
<table border="0" width="100%"> <tr valign="top"> <td width="10%" align="center">
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
</TD> <TD WIDTH="90%"> <H2>Site Error</H2> <P>An error was encountered while publishing this resource. </P>""" + \
</td> <td width="90%"> <h2>Site Error</h2> <p>An error was encountered while publishing this resource. </p>""" + \
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
<P><STRONG>%s</STRONG></P>
<p><strong>%s</strong></p>
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
<HR NOSHADE> <P>Troubleshooting Suggestions</P> <UL> <LI>The URL may be incorrect.</LI> <LI>The parameters passed to this resource may be incorrect.</LI> <LI>A resource that this resource relies on may be encountering an error.</LI> </UL> <P>For more detailed information about the error, please
<hr noshade="noshade"/> <p>Troubleshooting Suggestions</p> <ul> <li>The URL may be incorrect.</li> <li>The parameters passed to this resource may be incorrect.</li> <li>A resource that this resource relies on may be encountering an error.</li> </ul> <p>For more detailed information about the error, please
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
</P> <P>If the error persists please contact the site maintainer.
</p> <p>If the error persists please contact the site maintainer.
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
</P> </TD></TR> </TABLE>""")
</p> </td></tr> </table>""")
def _error_html(self,title,body): # XXX could this try to use standard_error_message somehow? return ("""\
def prepare(self, defaulthost=None, dnsresolver=None,
def prepare(self, defaulthost='', dnsresolver=None,
def prepare(self, defaulthost=None, dnsresolver=None, module=None, env=None, portbase=None): if defaulthost: hostname = defaulthost ip = socket.gethostbyname(hostname) elif defaulthost is '': hostname = '' ip = '127.0.0.1' else: hostname, ip = get_default_host_info() if not self.host: self._set_default_host(hostname, ip) else: self.ip = socket.gethostbyname(self.host) self.dnsresolver = dnsresolver self.module = module self.cgienv = env if portbase and self.port is not None: self.port += portbase
if defaulthost: hostname = defaulthost ip = socket.gethostbyname(hostname) elif defaulthost is '': hostname = '' ip = '127.0.0.1' else: hostname, ip = get_default_host_info()
def prepare(self, defaulthost=None, dnsresolver=None, module=None, env=None, portbase=None): if defaulthost: hostname = defaulthost ip = socket.gethostbyname(hostname) elif defaulthost is '': hostname = '' ip = '127.0.0.1' else: hostname, ip = get_default_host_info() if not self.host: self._set_default_host(hostname, ip) else: self.ip = socket.gethostbyname(self.host) self.dnsresolver = dnsresolver self.module = module self.cgienv = env if portbase and self.port is not None: self.port += portbase
self._set_default_host(hostname, ip)
ip = socket.gethostbyname(defaulthost) self._set_default_host(defaulthost, ip)
def prepare(self, defaulthost=None, dnsresolver=None, module=None, env=None, portbase=None): if defaulthost: hostname = defaulthost ip = socket.gethostbyname(hostname) elif defaulthost is '': hostname = '' ip = '127.0.0.1' else: hostname, ip = get_default_host_info() if not self.host: self._set_default_host(hostname, ip) else: self.ip = socket.gethostbyname(self.host) self.dnsresolver = dnsresolver self.module = module self.cgienv = env if portbase and self.port is not None: self.port += portbase
server = HTTPServer.zhttp_server(ip=self.host, port=self.port,
server = HTTPServer.zhttp_server(ip=self.ip, port=self.port,
def create(self): from ZServer import HTTPServer from ZServer.AccessLogger import access_logger handler = self.createHandler() handler._force_connection_close = self.force_connection_close if self.webdav_source_clients: handler.set_webdav_source_clients(self.webdav_source_clients) server = HTTPServer.zhttp_server(ip=self.host, port=self.port, resolver=self.dnsresolver, logger_object=access_logger) server.install_handler(handler) return server
return FTPServer(ip=self.host, hostname=self.host, port=self.port,
return FTPServer(ip=self.ip, hostname=self.host, port=self.port,
def create(self): from ZServer.AccessLogger import access_logger from ZServer.FTPServer import FTPServer return FTPServer(ip=self.host, hostname=self.host, port=self.port, module=self.module, resolver=self.dnsresolver, logger_object=access_logger)
return PCGIServer(ip=self.host, port=self.port,
return PCGIServer(ip=self.ip, port=self.port,
def create(self): from ZServer.AccessLogger import access_logger from ZServer.PCGIServer import PCGIServer return PCGIServer(ip=self.host, port=self.port, module=self.module, resolver=self.dnsresolver, pcgi_file=self.path, logger_object=access_logger)
return FCGIServer(ip=self.host, port=self.port,
return FCGIServer(ip=self.ip, port=self.port,
def create(self): from ZServer.AccessLogger import access_logger from ZServer.FCGIServer import FCGIServer return FCGIServer(ip=self.host, port=self.port, socket_file=self.path, module=self.module, resolver=self.dnsresolver, logger_object=access_logger)
}, self.db.table_names())
}, filter(self.db.database.datadefs.has_key, self.db.table_names()) )
def tables(self,*args,**kw): if self.db is None: self.open() return map(lambda name: { 'TABLE_NAME': name, 'TABLE_TYPE': 'TABLE', }, self.db.table_names())
def translate(self, msgid, domain, mapping=None, context=None, target_language=None, default=None): if context is None: context = self.contexts.get('context')
def translate(self, msgid, domain=None, mapping=None, default=None): context = self.contexts.get('context')
def translate(self, msgid, domain, mapping=None, context=None, target_language=None, default=None): if context is None: context = self.contexts.get('context') return getGlobalTranslationService().translate( domain, msgid, mapping=mapping, context=context, default=default, target_language=target_language)
default=default, target_language=target_language)
default=default)
def translate(self, msgid, domain, mapping=None, context=None, target_language=None, default=None): if context is None: context = self.contexts.get('context') return getGlobalTranslationService().translate( domain, msgid, mapping=mapping, context=context, default=default, target_language=target_language)
del self._unindex[i]
del self._unindex[documentId]
def unindex_object(self, documentId): """ Unindex the object with integer id 'documentId' and don't raise an exception if we fail """
line = f.readline() while line != "------\n":
line = f.readline().strip() while line != "------":
def verifyEntry(self, f, time=None, subsys=None, severity=None, summary=None, detail=None, error=None): # skip to the beginning of next entry line = f.readline() while line != "------\n": if not line: self.fail("can't find entry in log file") line = f.readline() line = f.readline().strip() _time, rest = line.split(" ", 1) if time is not None: self.assertEqual(_time, time) if subsys is not None: self.assert_(rest.find(subsys) != -1, "subsystem mismatch") if severity is not None and severity >= self._severity: s = severity_string[severity] self.assert_(rest.find(s) != -1, "severity mismatch") if summary is not None: self.assert_(rest.find(summary) != -1, "summary mismatch") if detail is not None: line = f.readline() self.assert_(line.find(detail) != -1, "missing detail") if error is not None: line = f.readline() self.assert_(line.startswith('Traceback'), "missing traceback") last = "%s: %s\n" % (error[0], error[1]) if last.startswith("exceptions."): last = last[len("exceptions."):] while 1: line = f.readline() if not line: self.fail("couldn't find end of traceback") if line == "------\n": self.fail("couldn't find end of traceback") if line == last: break
line = f.readline()
line = f.readline().strip()
def verifyEntry(self, f, time=None, subsys=None, severity=None, summary=None, detail=None, error=None): # skip to the beginning of next entry line = f.readline() while line != "------\n": if not line: self.fail("can't find entry in log file") line = f.readline() line = f.readline().strip() _time, rest = line.split(" ", 1) if time is not None: self.assertEqual(_time, time) if subsys is not None: self.assert_(rest.find(subsys) != -1, "subsystem mismatch") if severity is not None and severity >= self._severity: s = severity_string[severity] self.assert_(rest.find(s) != -1, "severity mismatch") if summary is not None: self.assert_(rest.find(summary) != -1, "summary mismatch") if detail is not None: line = f.readline() self.assert_(line.find(detail) != -1, "missing detail") if error is not None: line = f.readline() self.assert_(line.startswith('Traceback'), "missing traceback") last = "%s: %s\n" % (error[0], error[1]) if last.startswith("exceptions."): last = last[len("exceptions."):] while 1: line = f.readline() if not line: self.fail("couldn't find end of traceback") if line == "------\n": self.fail("couldn't find end of traceback") if line == last: break
last = "%s: %s\n" % (error[0], error[1])
last = "%s: %s" % (error[0], error[1])
def verifyEntry(self, f, time=None, subsys=None, severity=None, summary=None, detail=None, error=None): # skip to the beginning of next entry line = f.readline() while line != "------\n": if not line: self.fail("can't find entry in log file") line = f.readline() line = f.readline().strip() _time, rest = line.split(" ", 1) if time is not None: self.assertEqual(_time, time) if subsys is not None: self.assert_(rest.find(subsys) != -1, "subsystem mismatch") if severity is not None and severity >= self._severity: s = severity_string[severity] self.assert_(rest.find(s) != -1, "severity mismatch") if summary is not None: self.assert_(rest.find(summary) != -1, "summary mismatch") if detail is not None: line = f.readline() self.assert_(line.find(detail) != -1, "missing detail") if error is not None: line = f.readline() self.assert_(line.startswith('Traceback'), "missing traceback") last = "%s: %s\n" % (error[0], error[1]) if last.startswith("exceptions."): last = last[len("exceptions."):] while 1: line = f.readline() if not line: self.fail("couldn't find end of traceback") if line == "------\n": self.fail("couldn't find end of traceback") if line == last: break
if line == "------\n":
if line == "------":
def verifyEntry(self, f, time=None, subsys=None, severity=None, summary=None, detail=None, error=None): # skip to the beginning of next entry line = f.readline() while line != "------\n": if not line: self.fail("can't find entry in log file") line = f.readline() line = f.readline().strip() _time, rest = line.split(" ", 1) if time is not None: self.assertEqual(_time, time) if subsys is not None: self.assert_(rest.find(subsys) != -1, "subsystem mismatch") if severity is not None and severity >= self._severity: s = severity_string[severity] self.assert_(rest.find(s) != -1, "severity mismatch") if summary is not None: self.assert_(rest.find(summary) != -1, "summary mismatch") if detail is not None: line = f.readline() self.assert_(line.find(detail) != -1, "missing detail") if error is not None: line = f.readline() self.assert_(line.startswith('Traceback'), "missing traceback") last = "%s: %s\n" % (error[0], error[1]) if last.startswith("exceptions."): last = last[len("exceptions."):] while 1: line = f.readline() if not line: self.fail("couldn't find end of traceback") if line == "------\n": self.fail("couldn't find end of traceback") if line == last: break
roles=parent.__roles__ if roles is None: return 'Anonymous', if 'Shared' in roles: roles=list(roles) roles.remove('Shared') r=r+roles else: try: return r+list(roles) except: return r
def _shared_roles(self, parent): r=[] while 1: if hasattr(parent, 'aq_parent'): while hasattr(parent.aq_self,'aq_self'): parent=parent.aq_self parent=parent.aq_parent else: return r roles=parent.__roles__ if roles is None: return 'Anonymous', if 'Shared' in roles: roles=list(roles) roles.remove('Shared') r=r+roles else: try: return r+list(roles) except: return r
if 'Anonymous' in roles: return 1
if roles is None or 'Anonymous' in roles: return 1 while 'Shared' in roles: roles.remove('Shared') return self.allowed(parent,roles)
def allowed(self,parent,roles=None):
self._storage.tpc_begin(self._transaction) self.assertRaises(POSException.UndoError, self._storage.transactionalUndo, revid, self._transaction) self._storage.tpc_abort(self._transaction)
self._storage.tpc_begin(self._transaction) oids = self._storage.transactionalUndo(revid, self._transaction) self._storage.tpc_vote(self._transaction) self._storage.tpc_finish(self._transaction) assert len(oids) == 1 assert oids[0] == oid self.assertRaises(KeyError, self._storage.load, oid, '') import Full self.assertRaises(Full.ObjectDoesNotExist, self._storage.load, oid, '') try: self._storage.load(oid, '') except Full.ObjectDoesNotExist, e: revid = e.revid self._storage.tpc_begin(self._transaction) oids = self._storage.transactionalUndo(revid, self._transaction) self._storage.tpc_vote(self._transaction) self._storage.tpc_finish(self._transaction) assert len(oids) == 1 assert oids[0] == oid data, revid = self._storage.load(oid, '') assert pickle.loads(data) == 23
def checkSimpleTransactionalUndo(self): oid = self._storage.new_oid() revid = self._dostore(oid, data=23) revid = self._dostore(oid, revid=revid, data=24) revid = self._dostore(oid, revid=revid, data=25) # Now start an undo transaction self._storage.tpc_begin(self._transaction) oids = self._storage.transactionalUndo(revid, self._transaction) self._storage.tpc_vote(self._transaction) self._storage.tpc_finish(self._transaction) assert len(oids) == 1 assert oids[0] == oid data, revid = self._storage.load(oid, '') assert pickle.loads(data) == 24 # Do another one self._storage.tpc_begin(self._transaction) oids = self._storage.transactionalUndo(revid, self._transaction) self._storage.tpc_vote(self._transaction) self._storage.tpc_finish(self._transaction) assert len(oids) == 1 assert oids[0] == oid data, revid = self._storage.load(oid, '') assert pickle.loads(data) == 23 # Can't undo the first record self._storage.tpc_begin(self._transaction) self.assertRaises(POSException.UndoError, self._storage.transactionalUndo, revid, self._transaction) self._storage.tpc_abort(self._transaction)
+ (s,t,v)), tb
% (s,t,v)), tb
def connect(self,s):
manage_options = ({'label':'Overview', 'action':'manage_main'},) + \ SimpleItem.manage_options
manage_options = ({'label':'Overview', 'action':'manage_main'}, {'label':'Query', 'action':'queryLexicon'}, ) + SimpleItem.manage_options
def manage_addLexicon(self, id, title='', elements=[], REQUEST=None): """Add ZCTextIndex Lexicon""" pipeline = [] for el_record in elements: if not hasattr(el_record, 'name'): continue # Skip over records that only specify element group element = element_factory.instantiate(el_record.group, el_record.name) if element is not None: if el_record.group == 'Word Splitter': # I don't like hardcoding this, but its a simple solution # to get the splitter element first in the pipeline pipeline.insert(0, element) else: pipeline.append(element) lexicon = PLexicon(id, title, *pipeline) self._setObject(id, lexicon) if REQUEST is not None: return self.manage_main(self, REQUEST, update_menu=1)
_queryLexicon = DTMLFile('dtml/queryLexicon', globals()) def queryLexicon(self, REQUEST, words=None, page=0, rows=20, cols=4): """Lexicon browser/query user interface """ if words: wids = [] for word in words: wids.extend(self.globToWordIds(word)) words = [self.get_word(wid) for wid in wids] else: words = self.words() word_count = len(words) rows = max(min(rows, 500),1) cols = max(min(cols, 12), 1) page_count = word_count / (rows * cols) + \ (word_count % (rows * cols) > 0) page = max(min(page, page_count - 1), 0) start = rows * cols * page end = min(rows * cols * (page + 1), word_count) if word_count: words = list(words[start:end]) else: words = [] columns = [] i = 0 while i < len(words): columns.append(words[i:i + rows]) i += rows return self._queryLexicon(self, REQUEST, page=page, rows=rows, cols=cols, start_word=start+1, end_word=end, word_count=word_count, page_count=page_count, page_columns=columns)
def getPipelineNames(self): """Return list of names of pipeline element classes""" return [element.__class__.__name__ for element in self._pipeline]
arguments={} for name, arg in q._searchable_arguments().items(): arguments["%s/%s" % (id,name)]=arg
if input_id: for name, arg in q._searchable_arguments().items(): if len(qs) > 1: key="%s/%s" % (id,name) else: key=name arguments[key]=arg keys.append(key)
def add(self, report_id, report_title, report_style, input_id, input_title, queries=[], REQUEST=None): 'add a report' if not queries: raise ValueError, ( 'No <em>searchable objects</em> were selected') if not report_id: raise ValueError, ( 'No <em>report id</em> were specified') if input_title and not input_id: raise ValueError, ( 'No <em>input id</em> were specified') qs=map(lambda q, self=self: _getquery(self, q), queries) for q in qs: id=q.id arguments={} for name, arg in q._searchable_arguments().items(): arguments["%s/%s" % (id,name)]=arg if q._searchable_result_columns() is None: raise 'Unusable Searchable Error',( """The input searchable object, <em>%s</em>, has not been tested. Until it has been tested, it\'s outpuit schema is unknown, and a report cannot be generated. Before creating a report from this query, you must try out the query. To try out the query, <a href="%s">click hear</a>. """ % (q.title_and_id(), q.id)) if input_id: self.manage_addDocument( input_id,input_title, default_input_form(arguments, report_id)) self.manage_addDocument( report_id,report_title, ('<!--#var standard_html_header-->\n%s\n' '<!--#var standard_html_footer-->' % join(map(lambda q, report_style=report_style: custom_default_report(q.id, q, no_table=report_style), qs), '\n<hr>\n'))) if REQUEST: return self.manage_main(self,REQUEST)
""" % (q.title_and_id(), q.id))
""" % (q.title_and_id(), id))
def add(self, report_id, report_title, report_style, input_id, input_title, queries=[], REQUEST=None): 'add a report' if not queries: raise ValueError, ( 'No <em>searchable objects</em> were selected') if not report_id: raise ValueError, ( 'No <em>report id</em> were specified') if input_title and not input_id: raise ValueError, ( 'No <em>input id</em> were specified') qs=map(lambda q, self=self: _getquery(self, q), queries) for q in qs: id=q.id arguments={} for name, arg in q._searchable_arguments().items(): arguments["%s/%s" % (id,name)]=arg if q._searchable_result_columns() is None: raise 'Unusable Searchable Error',( """The input searchable object, <em>%s</em>, has not been tested. Until it has been tested, it\'s outpuit schema is unknown, and a report cannot be generated. Before creating a report from this query, you must try out the query. To try out the query, <a href="%s">click hear</a>. """ % (q.title_and_id(), q.id)) if input_id: self.manage_addDocument( input_id,input_title, default_input_form(arguments, report_id)) self.manage_addDocument( report_id,report_title, ('<!--#var standard_html_header-->\n%s\n' '<!--#var standard_html_footer-->' % join(map(lambda q, report_style=report_style: custom_default_report(q.id, q, no_table=report_style), qs), '\n<hr>\n'))) if REQUEST: return self.manage_main(self,REQUEST)
items.sort()
def default_input_form(arguments,action='query', tabs=''): if arguments: items=arguments.items() items.sort() return ( "%s\n%s%s" % ( '<!--#var standard_html_header-->\n%s\n' '<form action="%s" method="get">\n' '<h2><!--#var document_title--></h2>\n' 'Enter query parameters:<br>' '<table>\n' % (tabs,action), join( map( lambda a: ('<tr>\t<th>%s</th>\n' '\t<td><input name="%s"\n' '\t width=30 value="%s">' '</td></tr>' % (nicify(a[0]), ( a[1].has_key('type') and ("%s:%s" % (a[0],a[1]['type'])) or a[0] ), a[1].has_key('default') and a[1]['default'] or '' )) , items ), '\n'), '\n<tr><td colspan=2 align=center>\n' '<input type="SUBMIT" name="SUBMIT" value="Submit Query">\n' '<!--#if HTTP_REFERER-->\n' ' <input type="SUBMIT" name="SUBMIT" value="Cancel">\n' ' <INPUT NAME="CANCEL_ACTION" TYPE="HIDDEN"\n' ' VALUE="<!--#var HTTP_REFERER-->">\n' '<!--#/if HTTP_REFERER-->\n' '</td></tr>\n</table>\n</form>\n' '<!--#var standard_html_footer-->\n' ) ) else: return ( '<!--#var standard_html_header-->\n%s\n' '<form action="%s" method="get">\n' '<h2><!--#var document_title--></h2>\n' 'This query requires no input.<p>\n' '<input type="SUBMIT" name="SUBMIT" value="Submit Query">\n' '<!--#if HTTP_REFERER-->\n' ' <input type="SUBMIT" name="SUBMIT" value="Cancel">\n' ' <INPUT NAME="CANCEL_ACTION" TYPE="HIDDEN"\n' ' VALUE="<!--#var HTTP_REFERER-->">\n' '<!--#/if HTTP_REFERER-->\n' '</td></tr>\n</table>\n</form>\n' '<!--#var standard_html_footer-->\n' % (tabs, action) )
if not result: LOG('Zope', WARNING, 'invalid password for user %s' % self.getUserName())
def authenticate(self, password, request): passwrd=self._getPassword()
if domains: return result and domainSpecMatch(domains, request)
if domains and result: result = domainSpecMatch(domains, request) if not result: LOG('Zope', WARNING, 'invalid domain for user %s' % self.getUserName()) return result
def authenticate(self, password, request): passwrd=self._getPassword()
self._unindex[i] = tuple(d.keys())
unindex[i] = ()
def index_object(self, i, obj, tupleType=type(()), dictType=type({}), callable=callable): """Recompute index data for data with ids >= start. if 'obj' is passed in, it is indexed instead of _data[i]"""
else: index[word] = i, score
else: index[word] = i, score unindex[i].append(word)
def index_object(self, i, obj, tupleType=type(()), dictType=type({}), callable=callable): """Recompute index data for data with ids >= start. if 'obj' is passed in, it is indexed instead of _data[i]"""
schema(upper(name)]=i
schema[upper(name)]=i
def __init__(self,file,brains=NoBrains, parent=None, zbrains=None):
def aq_base(ob, hasattr=hasattr):
def aq_base(ob, getattr=getattr):
def aq_base(ob, hasattr=hasattr): # Return the aq_base of an object. if hasattr(ob, 'aq_base'): return ob.aq_base return ob
if hasattr(ob, 'aq_base'): return ob.aq_base return ob
return getattr(ob, 'aq_base', ob)
def aq_base(ob, hasattr=hasattr): # Return the aq_base of an object. if hasattr(ob, 'aq_base'): return ob.aq_base return ob
if hasattr(parent, absId): return 0
def is_acquired(ob, hasattr=hasattr, aq_base=aq_base, absattr=absattr): # Return true if this object is not a direct # subobject of its aq_parent object. if not hasattr(ob, 'aq_parent'): return 0 parent = aq_base(ob.aq_parent) absId = absattr(ob.id) if hasattr(parent,'_objects'): if absId+' ' in parent.objectIds(): return 0 if hasattr(parent, absId): return 0 if hasattr(aq_base(ob), 'isTopLevelPrincipiaApplicationObject') and \ ob.isTopLevelPrincipiaApplicationObject: return 0 return 1
i = 3 run = struct[top][1] if level == 0: return struct if level == 1: return run if level == 2: a = numbers[level-2] return run[a-1][1] a = numbers[level-2] run = run[a-1][1] while i <= level: a = numbers[i-2] run = run[a-1][1] i = i + 1 return run def find_level(indent, levels): """ when the indention is less, find which level it is a sublevel of """ if levels.has_key(indent): return levels[indent] for key in levels.keys(): if key > indent: return levels[key]
tmp = [] for x in numbers: if level > x: tmp.append(x) numbers = tmp numbers.append(level) if len(numbers) == 1: return (struct,numbers) run = struct[top][1] if level == numbers[len(numbers)-1]: i = 1 while i < level: run = run[len(run)-1][1] i = i + 1 else: i = 1 while i <= level: run = run[len(run)-1][1] i = i + 1 return run,numbers
def runner(struct,top,level,numbers): """ returns run which is a pointer to struct to be appended to """ i = 3 run = struct[top][1] if level == 0: return struct if level == 1: return run if level == 2: a = numbers[level-2] return run[a-1][1] a = numbers[level-2] run = run[a-1][1] while i <= level: a = numbers[i-2] run = run[a-1][1] i = i + 1 return run
numbers = {0:0}
numbers = [0]
def StructuredText(paragraphs): """ StructuredText accepts paragraphs, which is a list of lines to be parsed. StructuredText creates a structure which mimics the structure of the paragraphs. Structure => [raw_paragraph,parsed_paragraph,[sub-paragraphs]] """ current_level = 0 current_indent = 0 levels = {0:0} # what level ind = [] # structure based on indention levels top = -1 # which header are we under numbers = {0:0} # how many sub-paragraphs already at a level struct = [] # the structure to be returned paragraphs = split_paragraphs(paragraphs) if not paragraphs: result = ["",[]] return result for paragraph in paragraphs: if paragraph == '\n': ind.append([-1, paragraph]) else : ind.append([indention(paragraph), strip(paragraph)+"\n"]) for indent,paragraph in ind : if indent > -1: if indent == 0: """ a new top header, start over, everything underneath is a sub-paragraph """ current_level = 0 current_indent = 0 top = top + 1 levels = {0:0} numbers = {0:0} struct.append([paragraph,[]]) elif indent > current_indent: current_indent = indent current_level = current_level + 1 levels[current_indent] = current_level if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif (indent == current_indent): if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif indent < current_indent : current_level = find_level(indent,levels) current_indent = indent if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 return struct
if indent > -1: if indent == 0: """ a new top header, start over, everything underneath is a sub-paragraph """ current_level = 0 current_indent = 0 top = top + 1 levels = {0:0} numbers = {0:0} struct.append([paragraph,[]]) elif indent > current_indent: current_indent = indent current_level = current_level + 1 levels[current_indent] = current_level if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif (indent == current_indent): if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif indent < current_indent : current_level = find_level(indent,levels)
if indent == 0: struct.append([paragraph,[]]) current_level = 0 current_indent = 0 numbers = [0] levels = {0:0} top = top + 1 elif indent == current_indent: run,numbers = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) elif indent > current_indent: current_level = current_level + 1 current_indent = indent levels[current_level] = indent run,numbers = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) levels[current_level] = indent elif indent < current_indent: l = parent_level(levels,current_level) if indent > 0 and indent < levels[0]: levels[0] = indent
def StructuredText(paragraphs): """ StructuredText accepts paragraphs, which is a list of lines to be parsed. StructuredText creates a structure which mimics the structure of the paragraphs. Structure => [raw_paragraph,parsed_paragraph,[sub-paragraphs]] """ current_level = 0 current_indent = 0 levels = {0:0} # what level ind = [] # structure based on indention levels top = -1 # which header are we under numbers = {0:0} # how many sub-paragraphs already at a level struct = [] # the structure to be returned paragraphs = split_paragraphs(paragraphs) if not paragraphs: result = ["",[]] return result for paragraph in paragraphs: if paragraph == '\n': ind.append([-1, paragraph]) else : ind.append([indention(paragraph), strip(paragraph)+"\n"]) for indent,paragraph in ind : if indent > -1: if indent == 0: """ a new top header, start over, everything underneath is a sub-paragraph """ current_level = 0 current_indent = 0 top = top + 1 levels = {0:0} numbers = {0:0} struct.append([paragraph,[]]) elif indent > current_indent: current_indent = indent current_level = current_level + 1 levels[current_indent] = current_level if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif (indent == current_indent): if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif indent < current_indent : current_level = find_level(indent,levels) current_indent = indent if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 return struct
if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1
run,numbers = runner(struct,top,current_level,numbers) elif find_level(levels,indent) != -1: current_level = find_level(levels,indent) current_indent = indent run,numbers = runner(struct,top,current_level,numbers) elif levels[current_level] > indent and levels[l] < indent: levels[current_level] = indent current_indent = indent run,numbers = runner(struct,top,current_level,numbers) current_level = l else: tmp = {} for i in range(current_level): if indent > levels[i]: tmp[i] = levels[i] elif indent == levels[i]: current_level = i current_indent = indent run,numbers = runner(struct,top,current_level,numbers) levels = tmp run.append([paragraph,[]])
def StructuredText(paragraphs): """ StructuredText accepts paragraphs, which is a list of lines to be parsed. StructuredText creates a structure which mimics the structure of the paragraphs. Structure => [raw_paragraph,parsed_paragraph,[sub-paragraphs]] """ current_level = 0 current_indent = 0 levels = {0:0} # what level ind = [] # structure based on indention levels top = -1 # which header are we under numbers = {0:0} # how many sub-paragraphs already at a level struct = [] # the structure to be returned paragraphs = split_paragraphs(paragraphs) if not paragraphs: result = ["",[]] return result for paragraph in paragraphs: if paragraph == '\n': ind.append([-1, paragraph]) else : ind.append([indention(paragraph), strip(paragraph)+"\n"]) for indent,paragraph in ind : if indent > -1: if indent == 0: """ a new top header, start over, everything underneath is a sub-paragraph """ current_level = 0 current_indent = 0 top = top + 1 levels = {0:0} numbers = {0:0} struct.append([paragraph,[]]) elif indent > current_indent: current_indent = indent current_level = current_level + 1 levels[current_indent] = current_level if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif (indent == current_indent): if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 elif indent < current_indent : current_level = find_level(indent,levels) current_indent = indent if not numbers.has_key(current_level-1): numbers[current_level-1] = 0 run = runner(struct,top,current_level,numbers) run.append([paragraph,[]]) numbers[current_level-1] = numbers[current_level-1] + 1 return struct
self.expr = re.compile('\"[ a-zA-Z0-9.:/;,]+\":[a-zA-Z0-9.:/;,]+(?=(\s+|\.|\!|\?))').search
self.expr = re.compile('\"[ a-zA-Z0-9.:/;,\n]+\":[a-zA-Z0-9.:/;,\n]+(?=(\s+|\.|\!|\?))').search
def __init__(self,str=''): self.expr = re.compile('\"[ a-zA-Z0-9.:/;,]+\":[a-zA-Z0-9.:/;,]+(?=(\s+|\.|\!|\?))').search self.str = [str] self.typ = "href1"
except TypeError: return 0
except (TypeError, AttributeError): return 0
def _supports(self, ob): try: ob[0] except TypeError: return 0 except: pass return 1
self.__wids = OIBTree() self.__words = IOBTree()
self._wids = OIBTree() self._words = IOBTree()
def __init__(self, *pipeline): self.__wids = OIBTree() # word -> wid self.__words = IOBTree() # wid -> word # XXX we're reserving wid 0, but that might be yagni self.__nextwid = 1 self.__pipeline = pipeline
self.__nextwid = 1 self.__pipeline = pipeline
self._nextwid = 1 self._pipeline = pipeline
def __init__(self, *pipeline): self.__wids = OIBTree() # word -> wid self.__words = IOBTree() # wid -> word # XXX we're reserving wid 0, but that might be yagni self.__nextwid = 1 self.__pipeline = pipeline
return self.__nextwid - 1
return self._nextwid - 1
def length(self): """Return the number of unique terms in the lexicon.""" return self.__nextwid - 1
return self.__wids.keys()
return self._wids.keys()
def words(self): return self.__wids.keys()
return self.__words.keys()
return self._words.keys()
def wids(self): return self.__words.keys()
return self.__wids.items()
return self._wids.items()
def items(self): return self.__wids.items()
for element in self.__pipeline:
for element in self._pipeline:
def sourceToWordIds(self, text): last = _text2list(text) for element in self.__pipeline: last = element.process(last) return map(self._getWordIdCreate, last)
for element in self.__pipeline:
for element in self._pipeline:
def termToWordIds(self, text): last = _text2list(text) for element in self.__pipeline: last = element.process(last) wids = [] for word in last: wid = self.__wids.get(word) if wid is not None: wids.append(wid) return wids
wid = self.__wids.get(word)
wid = self._wids.get(word)
def termToWordIds(self, text): last = _text2list(text) for element in self.__pipeline: last = element.process(last) wids = [] for word in last: wid = self.__wids.get(word) if wid is not None: wids.append(wid) return wids
return self.__words[wid]
return self._words[wid]
def get_word(self, wid): """Return the word for the given word id""" return self.__words[wid]
keys = self.__wids.keys(prefix)
keys = self._wids.keys(prefix)
def globToWordIds(self, pattern): if not re.match("^\w+\*$", pattern): return [] pattern = pattern.lower() assert pattern.endswith("*") prefix = pattern[:-1] assert prefix and not prefix.endswith("*") keys = self.__wids.keys(prefix) # Keys starting at prefix wids = [] words = [] for key in keys: if not key.startswith(prefix): break wids.append(self.__wids[key]) words.append(key) return wids
wids.append(self.__wids[key])
wids.append(self._wids[key])
def globToWordIds(self, pattern): if not re.match("^\w+\*$", pattern): return [] pattern = pattern.lower() assert pattern.endswith("*") prefix = pattern[:-1] assert prefix and not prefix.endswith("*") keys = self.__wids.keys(prefix) # Keys starting at prefix wids = [] words = [] for key in keys: if not key.startswith(prefix): break wids.append(self.__wids[key]) words.append(key) return wids
wid = self.__wids.get(word)
wid = self._wids.get(word)
def _getWordIdCreate(self, word): wid = self.__wids.get(word) if wid is None: wid = self.__new_wid() self.__wids[word] = wid self.__words[wid] = word return wid
wid = self.__new_wid() self.__wids[word] = wid self.__words[wid] = word
wid = self._new_wid() self._wids[word] = wid self._words[wid] = word
def _getWordIdCreate(self, word): wid = self.__wids.get(word) if wid is None: wid = self.__new_wid() self.__wids[word] = wid self.__words[wid] = word return wid
def __new_wid(self): wid = self.__nextwid self.__nextwid += 1
def _new_wid(self): wid = self._nextwid self._nextwid += 1
def __new_wid(self): wid = self.__nextwid self.__nextwid += 1 return wid
self._catalog._convertBTrees(threshold *1 )
self._catalog._convertBTrees(threshold)
def manage_convertBTrees(self, threshold=200): """Convert the catalog's data structures to use BTrees package""" tt=time.time() ct=time.clock() self._catalog._convertBTrees(threshold *1 #make sure ints an int) ) tt=time.time()-tt ct=time.clock()-ct return 'Finished conversion in %s seconds (%s cpu)' % (tt, ct)
replace(encodestring('%s:%s' % (self.username,self.password), '\012','')))
replace(encodestring('%s:%s' % (self.username,self.password)), '\012',''))
def __call__(self,*args,**kw): method=self.method if method=='PUT' and len(args)==1 and not kw: query=[args[0]] args=() else: query=[] for i in range(len(args)): try: k=self.args[i] if kw.has_key(k): raise TypeError, 'Keyword arg redefined' kw[k]=args[i] except IndexError: raise TypeError, 'Too many arguments'