rem
stringlengths
0
322k
add
stringlengths
0
2.05M
context
stringlengths
8
228k
assert have_pyrex
assert have_pyrex == cython_needed
def ext_modules(): for pyx_file in glob.glob(os.path.join('djvu', '*.pyx')): module, _ = os.path.splitext(os.path.basename(pyx_file)) yield module
for n in -100, 0, 1 << 32:
for n in -100, 0, (1 + sys.maxint) * 2:
def set_cache_size(n): context.cache_size = n
for n in 1, 100, 1 << 10, 1 << 20, (1 << 32) -1:
n = 1 while n < 3 * sys.maxint:
def set_cache_size(n): context.cache_size = n
>>> page_job.render(RENDER_COLOR, (0, 0, 100000, 100000), (0, 0, 100000, 100000), PixelFormatRgb(), 8)
>>> x = int((sys.maxint//2) ** 0.5) >>> page_job.render(RENDER_COLOR, (0, 0, x, x), (0, 0, x, x), PixelFormatRgb(), 8)
def test_decode(): r''' >>> context = Context() >>> document = context.new_document(FileUri(images + 'test1.djvu')) >>> message = document.get_message() >>> type(message) == DocInfoMessage True >>> page_job = document.pages[0].decode() >>> page_job.is_done True >>> type(page_job) == PageJob True >>> page_job.is_done True >>> page_job.is_error False >>> page_job.status == JobOK True >>> page_job.width 64 >>> page_job.height 48 >>> page_job.size (64, 48) >>> page_job.dpi 300 >>> page_job.gamma 2.2000000000000002 >>> page_job.version 24 >>> page_job.type == PAGE_TYPE_BITONAL True >>> page_job.rotation, page_job.initial_rotation (0, 0) >>> page_job.rotation = 100 Traceback (most recent call last): ... ValueError: rotation must be equal to 0, 90, 180, or 270 >>> page_job.rotation = 180 >>> page_job.rotation, page_job.initial_rotation (180, 0) >>> del page_job.rotation >>> page_job.rotation, page_job.initial_rotation (0, 0)
changelog = file(os.path.join('doc', 'changelog'))
changelog = open(os.path.join('doc', 'changelog'))
def get_version(): changelog = file(os.path.join('doc', 'changelog')) try: return changelog.readline().split()[1].strip('()') finally: changelog.close()
from setuptools.extension import Extension
from setuptools.extension import Extension, have_pyrex assert have_pyrex del have_pyrex
def ext_modules(): for pyx_file in glob.glob(os.path.join('djvu', '*.pyx')): module, _ = os.path.splitext(os.path.basename(pyx_file)) yield module
from distutils.extension import Extension, have_pyrex assert have_pyrex del have_pyrex
from distutils.extension import Extension
def ext_modules(): for pyx_file in glob.glob(os.path.join('djvu', '*.pyx')): module, _ = os.path.splitext(os.path.basename(pyx_file)) yield module
"Returns True to carry on or erase and False to maintain" if path.startswith(self.deposit): debug('Deposit path: ' + path,DEBUG_MEDIUM) return False if os.path.ismount(path):
"Returns True to carry on or erase and False to maintain" if path.startswith(self.deposit): debug('Deposit path: ' + path,DEBUG_MEDIUM) return False if os.path.ismount(path) and path != self.homedir:
def __restore_or_erase(self,path): "Returns True to carry on or erase and False to maintain" #Do nothing with the deposit if it's inside a home directory if path.startswith(self.deposit): debug('Deposit path: ' + path,DEBUG_MEDIUM) return False
print "Debug L: "+ str(text.encode('utf-8'))
print "Debug L: "+ text
def debug(text, level=DEBUG_LOW): "Prints a text in the terminal if the debug level is higher than the requested" try: if level <= debug_level: if level == DEBUG_LOW: print "Debug L: "+ str(text.encode('utf-8')) if level == DEBUG_MEDIUM: print "Debug M: "+ str(text.encode('utf-8')) if level == DEBUG_HIGH: print "Debug H: "+ str(text.encode('utf-8')) except: print_error("Can't print debug message")
print "Debug M: "+ str(text.encode('utf-8'))
print "Debug M: "+ text
def debug(text, level=DEBUG_LOW): "Prints a text in the terminal if the debug level is higher than the requested" try: if level <= debug_level: if level == DEBUG_LOW: print "Debug L: "+ str(text.encode('utf-8')) if level == DEBUG_MEDIUM: print "Debug M: "+ str(text.encode('utf-8')) if level == DEBUG_HIGH: print "Debug H: "+ str(text.encode('utf-8')) except: print_error("Can't print debug message")
print "Debug H: "+ str(text.encode('utf-8'))
print "Debug H: "+ text
def debug(text, level=DEBUG_LOW): "Prints a text in the terminal if the debug level is higher than the requested" try: if level <= debug_level: if level == DEBUG_LOW: print "Debug L: "+ str(text.encode('utf-8')) if level == DEBUG_MEDIUM: print "Debug M: "+ str(text.encode('utf-8')) if level == DEBUG_HIGH: print "Debug H: "+ str(text.encode('utf-8')) except: print_error("Can't print debug message")
print "Warning: "+ str(text.encode('utf-8'))
print "Warning: "+ text
def print_error(text,level=ERROR): "Prints a error or warning message in the terminal" try: if level == WARNING: print "Warning: "+ str(text.encode('utf-8')) return print "Error : "+ str(text.encode('utf-8')) except: print "Error : Can't print error message"
print "Error : "+ str(text.encode('utf-8'))
print "Error : "+ text
def print_error(text,level=ERROR): "Prints a error or warning message in the terminal" try: if level == WARNING: print "Warning: "+ str(text.encode('utf-8')) return print "Error : "+ str(text.encode('utf-8')) except: print "Error : Can't print error message"
raise LexerError(errline, ((line, pos), (line, len(errline))))
raise LexerError(errline, ((line, pos + 1), (line, len(errline))))
def match_specs(specs, str, i, (line, pos)): for spec in specs: m = spec.re.match(str, i) if m is not None: value = m.group() nls = value.count(u'\n') n_line = line + nls if nls == 0: n_pos = pos + len(value) else: n_pos = len(value) - value.rfind(u'\n') - 1 return Token(spec.type, value, ((line, pos), (n_line, n_pos))) else: errline = str.splitlines()[line - 1] raise LexerError(errline, ((line, pos), (line, len(errline))))
CC_FLAGS = "-g -Wall -Wfatal-errors -Werror"
CC_FLAGS = "-g -Wall -Werror"
def ExecProcess(cmd): try: pipe = os.popen4(cmd)[1] except AttributeError: pipe = os.popen(cmd) result = "" for line in pipe.readlines(): result = result + line[:-1] pipe.close() return result
DEFINES += ['IMAGES_USE_LIBPNG']
DEFINES += ['IMAGES_USE_LIBPNG', 'NO_CAIRO']
def ExecProcess(cmd): try: pipe = os.popen4(cmd)[1] except AttributeError: pipe = os.popen(cmd) result = "" for line in pipe.readlines(): result = result + line[:-1] pipe.close() return result
'ole32', 'msvcp60', 'Msimg32', 'png', 'z', 'opengl32', 'glu32', 'cairo'
'ole32', 'msvcp60', 'Msimg32', 'opengl32', 'glu32', 'png', 'z'
def ExecProcess(cmd): try: pipe = os.popen4(cmd)[1] except AttributeError: pipe = os.popen(cmd) result = "" for line in pipe.readlines(): result = result + line[:-1] pipe.close() return result
ccflags += " -fPIC" linkflags += " -shared -fPIC"
if getHost() == "windows": linkflags += " -shared" else: ccflags += " -fPIC" linkflags += " -shared -fPIC"
def Compile(outputfile, files, ccflags="", linkflags="", target="exe", cc=CC): if target == "dll": ccflags += " -fPIC" linkflags += " -shared -fPIC" ofile = DllExt(outputfile) elif target == "exe": ofile = ExeExt(outputfile) else: Error("unknown target " + type) linkCmd = "" for f in files: Exec(Subs("$# -c $# -o $# $#", cc, ccflags, ObjExt(f), f)) linkCmd += " " + ObjExt(f) Exec(Subs("$# $# -o $# $#", cc, linkflags, ofile, linkCmd))
Exec(Subs("$
c = Changed(f, f) if c.check() or force: Exec(Subs("$ c.success()
def Compile(outputfile, files, ccflags="", linkflags="", target="exe", cc=CC): if target == "dll": ccflags += " -fPIC" linkflags += " -shared -fPIC" ofile = DllExt(outputfile) elif target == "exe": ofile = ExeExt(outputfile) else: Error("unknown target " + type) linkCmd = "" for f in files: Exec(Subs("$# -c $# -o $# $#", cc, ccflags, ObjExt(f), f)) linkCmd += " " + ObjExt(f) Exec(Subs("$# $# -o $# $#", cc, linkflags, ofile, linkCmd))
ccflags += ccflags + " -Iclaro"
ccflags += ccflags + " -Iinclude"
def cmd_examples(ccflags): ccflags += ccflags + " -Iclaro" Compile("build/hello", ["examples/helloworld/hello.c"], ccflags) Compile("build/radio", ["examples/radio/radio.c"], ccflags) Compile("build/combo", ["examples/combo/combo.c"], ccflags) Compile("build/layout_test", ["examples/layout/layout_test.c"], ccflags) Compile("build/layout_any", ["examples/layout/layout_any.c"], ccflags) Compile("build/splitter", ["examples/splitter/splitter-test.c"], ccflags) Compile("build/cldlg", ["examples/cldlg/cldlg.c"], ccflags) Compile("build/designer", ["examples/designer/designer.c"], ccflags) Compile("build/progress", ["examples/progress/progress-test.c"], ccflags) Compile("build/image", ["examples/image/image.c"], ccflags) Compile("build/list", ["examples/list/list.c"], ccflags) Compile("build/listview", ["examples/list/listview.c"], ccflags) Compile("build/workspace", ["examples/workspace/workspace.c"], ccflags)
raise self.ConnectionError(*exc.args)
raise ConnectionError(*exc.args)
def connect(self): """ Connects to the STOMP server if not already connected. """ if self._sock: return try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((self.host, self.port)) except socket.error, exc: raise self.ConnectionError(*exc.args) except socket.timeout, exc: raise self.ConnectionTimeoutError(*exc.args) sock.setsockopt(socket.SOL_TCP, socket.TCP_NODELAY, 1) sock.settimeout(self.socket_timeout) self._sock = sock
raise self.ConnectionTimeoutError(*exc.args)
raise ConnectionTimeoutError(*exc.args)
def connect(self): """ Connects to the STOMP server if not already connected. """ if self._sock: return try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((self.host, self.port)) except socket.error, exc: raise self.ConnectionError(*exc.args) except socket.timeout, exc: raise self.ConnectionTimeoutError(*exc.args) sock.setsockopt(socket.SOL_TCP, socket.TCP_NODELAY, 1) sock.settimeout(self.socket_timeout) self._sock = sock
print "Trying to disconnect."
def disconnect(self, conf=None): """ Disconnect from the server, if connected. """ print "Trying to disconnect." if self._sock is None: return try: self._sock.close() except socket.error: pass self._sock = None self._buffer.clear()
print repr(self.connection.connected)
def disconnect(self, conf=None, extra_headers=None): """Disconnect from the server.""" print repr(self.connection.connected) if self.connection.connected: disconnect = frame.DisconnectFrame(extra_headers=extra_headers) result = self.send_frame(disconnect) self.connection.disconnect() return result
subscribe clients, since the message messages are received by a separate thread.
subscribe clients, since a listener thread needs to be able to share the *same* socket with other publisher thread(s).
def get_all_connections(self): "Return a list of all connection objects the manager knows about" return self.connections.values()
This class is notably not thread-safe. You need to use external mechanisms to guard access to connections. This is typically accomplished by using a thread-safe connection pool implementation (e.g. L{stompclient.connection.ThreadLocalConnectionPool}).
This class provides some basic synchronization to avoid threads stepping on eachother. Specifically the following activities are each protected by [their own] C{threading.RLock} instances: - connect() and disconnect() methods (share a lock). - read() - send() It is assumed that send() and recv() should be allowed to happen concurrently, so these do not *share* a lock. If you need more thread-isolation, consider using a thread-safe connection pool implementation (e.g. L{stompclient.connection.ThreadLocalConnectionPool}).
def get_all_connections(self): "Return a list of all connection objects the manager knows about" return self.connections.values()
self._connect_lock = threading.RLock() self._send_lock = threading.RLock() self._read_lock = threading.RLock()
def __init__(self, host, port=61613, socket_timeout=None): self.host = host self.port = port self.socket_timeout = socket_timeout self._sock = None self._buffer = FrameBuffer() self._connected = threading.Event()
if self._sock: return try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((self.host, self.port)) except socket.timeout as exc: raise ConnectionTimeoutError(*exc.args) except socket.error as exc: raise ConnectionError(*exc.args) sock.setsockopt(socket.SOL_TCP, socket.TCP_NODELAY, 1) sock.settimeout(self.socket_timeout) self._sock = sock self._connected.set()
with self._connect_lock: if self._sock: return try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((self.host, self.port)) except socket.timeout as exc: raise ConnectionTimeoutError(*exc.args) except socket.error as exc: raise ConnectionError(*exc.args) sock.setsockopt(socket.SOL_TCP, socket.TCP_NODELAY, 1) sock.settimeout(self.socket_timeout) self._sock = sock self._connected.set()
def connect(self): """ Connects to the STOMP server if not already connected. """ if self._sock: return try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((self.host, self.port)) except socket.timeout as exc: raise ConnectionTimeoutError(*exc.args) except socket.error as exc: raise ConnectionError(*exc.args) sock.setsockopt(socket.SOL_TCP, socket.TCP_NODELAY, 1) sock.settimeout(self.socket_timeout) self._sock = sock self._connected.set()
if self._sock is None: raise NotConnectedError() try: self._sock.close() except socket.error: pass self._sock = None self._buffer.clear() self._connected.clear()
with self._connect_lock: if self._sock is None: raise NotConnectedError() try: self._sock.close() except socket.error: pass self._sock = None self._buffer.clear() self._connected.clear()
def disconnect(self, conf=None): """ Disconnect from the server, if connected. @raise NotConnectedError: If the connection is not currently connected. """ if self._sock is None: raise NotConnectedError() try: self._sock.close() except socket.error: pass self._sock = None self._buffer.clear() self._connected.clear()
self.connect() try: self._sock.sendall(str(frame)) except socket.error, e: if e.args[0] == errno.EPIPE: self.disconnect() raise ConnectionError("Error %s while writing to socket. %s." % e.args)
with self._send_lock: self.connect() try: self._sock.sendall(str(frame)) except socket.error, e: if e.args[0] == errno.EPIPE: self.disconnect() raise ConnectionError("Error %s while writing to socket. %s." % e.args)
def send(self, frame): """ Sends the specified frame to STOMP server. @param frame: The frame to send to server. @type frame: L{stompclient.frame.Frame} """ self.connect() try: self._sock.sendall(str(frame)) except socket.error, e: if e.args[0] == errno.EPIPE: self.disconnect() raise ConnectionError("Error %s while writing to socket. %s." % e.args)
begdata = 900 + (nchannels * 75) enddata = self.info["eventtablepos"] nsamples = ((enddata - begdata)/self.info["nchannels"])/2 self.info["nsamples"] = nsamples
def load_setup(self): """Loads CNT metadata from header.""" self.info = {} self.file.seek(0) self.info["rev"] = self.get('12s') self.info["nextfile"] = self.get('l') self.info["prevfile"] = self.get('L') self.info["type"] = self.get('b') self.info["id"] = self.get('20s') self.info["oper"] = self.get('20s') self.info["doctor"] = self.get('20s') self.info["referral"] = self.get('20s') self.info["hospital"] = self.get('20s') self.info["patient"] = self.get('20s') self.info["age"] = self.get('h') self.info["sex"] = self.get('s') self.info["hand"] = self.get('s') self.info["med"] = self.get('20s') self.info["category"] = self.get('20s') self.info["state"] = self.get('20s') self.info["label"] = self.get('20s') self.info["date"] = self.get('10s') self.info["time"] = self.get('12s') self.info["mean_age"] = self.get('f') self.info["stdev"] = self.get('f') self.info["n"] = self.get('h') self.info["compfile"] = self.get('38s') self.info["spectwincomp"] = self.get('f') self.info["meanaccuracy"] = self.get('f') self.info["meanlatency"] = self.get('f') self.info["sortfile"] = self.get('46s') self.info["numevents"] = self.get('i') self.info["compoper"] = self.get('b') self.info["avgmode"] = self.get('b') self.info["review"] = self.get('b') self.info["nsweeps"] = self.get('H') self.info["compsweeps"] = self.get('H') self.info["acceptcnt"] = self.get('H') self.info["rejectcnt"] = self.get('H') self.info["pnts"] = self.get('H') self.info["nchannels"] = self.get('H') self.info["avgupdate"] = self.get('H') self.info["domain"] = self.get('b') self.info["variance"] = self.get('b') self.info["rate"] = self.get('H') self.info["scale"] = self.get('d') self.info["veogcorrect"] = self.get('b') self.info["heogcorrect"] = self.get('b') self.info["aux1correct"] = self.get('b') self.info["aux2correct"] = self.get('b') self.info["veogtrig"] = self.get('f') self.info["heogtrig"] = self.get('f') self.info["aux1trig"] = self.get('f') self.info["aux2trig"] = self.get('f') self.info["heogchnl"] = self.get('h') self.info["veogchnl"] = self.get('h') self.info["aux1chnl"] = self.get('h') self.info["aux2chnl"] = self.get('h') self.info["veogdir"] = self.get('b') self.info["heogdir"] = self.get('b') self.info["aux1dir"] = self.get('b') self.info["aux2dir"] = self.get('b') self.info["veog_n"] = self.get('h') self.info["heog_n"] = self.get('h') self.info["aux1_n"] = self.get('h') self.info["aux2_n"] = self.get('h') self.info["veogmaxcnt"] = self.get('h') self.info["heogmaxcnt"] = self.get('h') self.info["aux1maxcnt"] = self.get('h') self.info["aux2maxcnt"] = self.get('h') self.info["veogmethod"] = self.get('b') self.info["heogmethod"] = self.get('b') self.info["aux1method"] = self.get('b') self.info["aux2method"] = self.get('b') self.info["ampsensitivity"] = self.get('f') self.info["lowpass"] = self.get('b') self.info["highpass"] = self.get('b') self.info["notch"] = self.get('b') self.info["autoclipadd"] = self.get('b') self.info["baseline"] = self.get('b') self.info["offstart"] = self.get('f') self.info["offstop"] = self.get('f') self.info["reject"] = self.get('b') self.info["rejstart"] = self.get('f') self.info["rejstop"] = self.get('f') self.info["rejmin"] = self.get('f') self.info["rejmax"] = self.get('f') self.info["trigtype"] = self.get('b') self.info["trigval"] = self.get('f') self.info["trigchnl"] = self.get('b') self.info["trigmask"] = self.get('h') self.info["trigisi"] = self.get('f') self.info["trigmin"] = self.get('f') self.info["trigmax"] = self.get('f') self.info["trigdir"] = self.get('b') self.info["autoscale"] = self.get('b') self.info["n2"] = self.get('h') self.info["dir"] = self.get('b') self.info["dispmin"] = self.get('f') self.info["dispmax"] = self.get('f') self.info["xmin"] = self.get('f') self.info["xmax"] = self.get('f') self.info["automin"] = self.get('f') self.info["automax"] = self.get('f') self.info["zmin"] = self.get('f') self.info["zmax"] = self.get('f') self.info["lowcut"] = self.get('f') self.info["highcut"] = self.get('f') self.info["common"] = self.get('b') self.info["savemode"] = self.get('b') self.info["manmode"] = self.get('b') self.info["ref"] = self.get('10s') self.info["rectify"] = self.get('b') self.info["displayxmin"] = self.get('f') self.info["displayxmax"] = self.get('f') self.info["phase"] = self.get('b') self.info["screen"] = self.get('16s') self.info["calmode"] = self.get('h') self.info["calmethod"] = self.get('h') self.info["calupdate"] = self.get('h') self.info["calbaseline"] = self.get('h') self.info["calsweeps"] = self.get('h') self.info["calattenuator"] = self.get('f') self.info["calpulsevolt"] = self.get('f') self.info["calpulsestart"] = self.get('f') self.info["calpulsestop"] = self.get('f') self.info["calfreq"] = self.get('f') self.info["taskfile"] = self.get('34s') self.info["seqfile"] = self.get('34s') self.info["spectmethod"] = self.get('b') self.info["spectscaling"] = self.get('b') self.info["spectwindow"] = self.get('b') self.info["spectwinlength"] = self.get('f') self.info["spectorder"] = self.get('b') self.info["notchfilter"] = self.get('b') self.info["headgain"] = self.get('h') self.info["additionalfiles"] = self.get('i') self.info["unused"] = self.get('5s') self.info["fspstopmethod"] = self.get('h') self.info["fspstopmode"] = self.get('h') self.info["fspfvalue"] = self.get('f') self.info["fsppoint"] = self.get('h') self.info["fspblocksize"] = self.get('h') self.info["fspp1"] = self.get('H') self.info["fspp2"] = self.get('H') self.info["fspalpha"] = self.get('f') self.info["fspnoise"] = self.get('f') self.info["fspv1"] = self.get('h') self.info["montage"] = self.get('40s') self.info["eventfile"] = self.get('40s') self.info["fratio"] = self.get('f') self.info["minor_rev"] = self.get('b') self.info["eegupdate"] = self.get('h') self.info["compressed"] = self.get('b') self.info["xscale"] = self.get('f') self.info["yscale"] = self.get('f') self.info["xsize"] = self.get('f') self.info["ysize"] = self.get('f') self.info["acmode"] = self.get('b') self.info["commonchnl"] = self.get('B') self.info["xtics"] = self.get('b') self.info["xrange"] = self.get('b') self.info["ytics"] = self.get('b') self.info["yrange"] = self.get('b') self.info["xscalevalue"] = self.get('f') self.info["xscaleinterval"] = self.get('f') self.info["yscalevalue"] = self.get('f') self.info["yscaleinterval"] = self.get('f') self.info["scaletoolx1"] = self.get('f') self.info["scaletooly1"] = self.get('f') self.info["scaletoolx2"] = self.get('f') self.info["scaletooly2"] = self.get('f') self.info["port"] = self.get('h') self.info["nsamples"] = self.get('L') self.info["filterflag"] = self.get('b') self.info["lowcutoff"] = self.get('f') self.info["lowpoles"] = self.get('h') self.info["highcutoff"] = self.get('f') self.info["highpoles"] = self.get('h') self.info["filtertype"] = self.get('b') self.info["filterdomain"] = self.get('b') self.info["snrflag"] = self.get('b') self.info["coherenceflag"] = self.get('b') self.info["continuoustype"] = self.get('b') self.info["eventtablepos"] = self.get('L') self.info["continuousseconds"] = self.get('f') self.info["channeloffset"] = self.get('l') self.info["autocorrectflag"] = self.get('b') self.info["dcthreshold"] = self.get('B')
shape = (2, len(self.events))
shape = (len(self.events), 2)
def save_events(self): self.get_event_table() self.load_events() shape = (2, len(self.events)) atom = tables.Int16Atom() filters = tables.Filters(complevel=5, complib='zlib') ca = self.h5.createCArray(self.h5.root, 'triggers', atom, shape, title="Trigger Data", filters=filters) for event in self.events: ca[:, event.event_id] = [event.stimtype, event.offset]
for event in self.events: ca[:, event.event_id] = [event.stimtype, event.offset]
for i, event in enumerate(self.events): ca[i, :] = [event.stimtype, event.offset]
def save_events(self): self.get_event_table() self.load_events() shape = (2, len(self.events)) atom = tables.Int16Atom() filters = tables.Filters(complevel=5, complib='zlib') ca = self.h5.createCArray(self.h5.root, 'triggers', atom, shape, title="Trigger Data", filters=filters) for event in self.events: ca[:, event.event_id] = [event.stimtype, event.offset]
sys.exit() CNTData(cnt_filename)
sys.exit(0) else: print cnt_filename CNTData(cnt_filename)
def convert_bytes_to_points(self, byte): data_offset = 900 + (75 * self.info["nchannels"]) point = (byte - data_offset) / (2 * self.info["nchannels"]) return point
begdata = 900 + (nchannels * 75)
begdata = 900 + (self.info["nchannels"] * 75)
def load_setup(self): """Loads CNT metadata from header.""" self.info = {} self.file.seek(0) self.info["rev"] = self.get('12s') self.info["nextfile"] = self.get('l') self.info["prevfile"] = self.get('L') self.info["type"] = self.get('b') self.info["id"] = self.get('20s') self.info["oper"] = self.get('20s') self.info["doctor"] = self.get('20s') self.info["referral"] = self.get('20s') self.info["hospital"] = self.get('20s') self.info["patient"] = self.get('20s') self.info["age"] = self.get('h') self.info["sex"] = self.get('s') self.info["hand"] = self.get('s') self.info["med"] = self.get('20s') self.info["category"] = self.get('20s') self.info["state"] = self.get('20s') self.info["label"] = self.get('20s') self.info["date"] = self.get('10s') self.info["time"] = self.get('12s') self.info["mean_age"] = self.get('f') self.info["stdev"] = self.get('f') self.info["n"] = self.get('h') self.info["compfile"] = self.get('38s') self.info["spectwincomp"] = self.get('f') self.info["meanaccuracy"] = self.get('f') self.info["meanlatency"] = self.get('f') self.info["sortfile"] = self.get('46s') self.info["numevents"] = self.get('i') self.info["compoper"] = self.get('b') self.info["avgmode"] = self.get('b') self.info["review"] = self.get('b') self.info["nsweeps"] = self.get('H') self.info["compsweeps"] = self.get('H') self.info["acceptcnt"] = self.get('H') self.info["rejectcnt"] = self.get('H') self.info["pnts"] = self.get('H') self.info["nchannels"] = self.get('H') self.info["avgupdate"] = self.get('H') self.info["domain"] = self.get('b') self.info["variance"] = self.get('b') self.info["rate"] = self.get('H') self.info["scale"] = self.get('d') self.info["veogcorrect"] = self.get('b') self.info["heogcorrect"] = self.get('b') self.info["aux1correct"] = self.get('b') self.info["aux2correct"] = self.get('b') self.info["veogtrig"] = self.get('f') self.info["heogtrig"] = self.get('f') self.info["aux1trig"] = self.get('f') self.info["aux2trig"] = self.get('f') self.info["heogchnl"] = self.get('h') self.info["veogchnl"] = self.get('h') self.info["aux1chnl"] = self.get('h') self.info["aux2chnl"] = self.get('h') self.info["veogdir"] = self.get('b') self.info["heogdir"] = self.get('b') self.info["aux1dir"] = self.get('b') self.info["aux2dir"] = self.get('b') self.info["veog_n"] = self.get('h') self.info["heog_n"] = self.get('h') self.info["aux1_n"] = self.get('h') self.info["aux2_n"] = self.get('h') self.info["veogmaxcnt"] = self.get('h') self.info["heogmaxcnt"] = self.get('h') self.info["aux1maxcnt"] = self.get('h') self.info["aux2maxcnt"] = self.get('h') self.info["veogmethod"] = self.get('b') self.info["heogmethod"] = self.get('b') self.info["aux1method"] = self.get('b') self.info["aux2method"] = self.get('b') self.info["ampsensitivity"] = self.get('f') self.info["lowpass"] = self.get('b') self.info["highpass"] = self.get('b') self.info["notch"] = self.get('b') self.info["autoclipadd"] = self.get('b') self.info["baseline"] = self.get('b') self.info["offstart"] = self.get('f') self.info["offstop"] = self.get('f') self.info["reject"] = self.get('b') self.info["rejstart"] = self.get('f') self.info["rejstop"] = self.get('f') self.info["rejmin"] = self.get('f') self.info["rejmax"] = self.get('f') self.info["trigtype"] = self.get('b') self.info["trigval"] = self.get('f') self.info["trigchnl"] = self.get('b') self.info["trigmask"] = self.get('h') self.info["trigisi"] = self.get('f') self.info["trigmin"] = self.get('f') self.info["trigmax"] = self.get('f') self.info["trigdir"] = self.get('b') self.info["autoscale"] = self.get('b') self.info["n2"] = self.get('h') self.info["dir"] = self.get('b') self.info["dispmin"] = self.get('f') self.info["dispmax"] = self.get('f') self.info["xmin"] = self.get('f') self.info["xmax"] = self.get('f') self.info["automin"] = self.get('f') self.info["automax"] = self.get('f') self.info["zmin"] = self.get('f') self.info["zmax"] = self.get('f') self.info["lowcut"] = self.get('f') self.info["highcut"] = self.get('f') self.info["common"] = self.get('b') self.info["savemode"] = self.get('b') self.info["manmode"] = self.get('b') self.info["ref"] = self.get('10s') self.info["rectify"] = self.get('b') self.info["displayxmin"] = self.get('f') self.info["displayxmax"] = self.get('f') self.info["phase"] = self.get('b') self.info["screen"] = self.get('16s') self.info["calmode"] = self.get('h') self.info["calmethod"] = self.get('h') self.info["calupdate"] = self.get('h') self.info["calbaseline"] = self.get('h') self.info["calsweeps"] = self.get('h') self.info["calattenuator"] = self.get('f') self.info["calpulsevolt"] = self.get('f') self.info["calpulsestart"] = self.get('f') self.info["calpulsestop"] = self.get('f') self.info["calfreq"] = self.get('f') self.info["taskfile"] = self.get('34s') self.info["seqfile"] = self.get('34s') self.info["spectmethod"] = self.get('b') self.info["spectscaling"] = self.get('b') self.info["spectwindow"] = self.get('b') self.info["spectwinlength"] = self.get('f') self.info["spectorder"] = self.get('b') self.info["notchfilter"] = self.get('b') self.info["headgain"] = self.get('h') self.info["additionalfiles"] = self.get('i') self.info["unused"] = self.get('5s') self.info["fspstopmethod"] = self.get('h') self.info["fspstopmode"] = self.get('h') self.info["fspfvalue"] = self.get('f') self.info["fsppoint"] = self.get('h') self.info["fspblocksize"] = self.get('h') self.info["fspp1"] = self.get('H') self.info["fspp2"] = self.get('H') self.info["fspalpha"] = self.get('f') self.info["fspnoise"] = self.get('f') self.info["fspv1"] = self.get('h') self.info["montage"] = self.get('40s') self.info["eventfile"] = self.get('40s') self.info["fratio"] = self.get('f') self.info["minor_rev"] = self.get('b') self.info["eegupdate"] = self.get('h') self.info["compressed"] = self.get('b') self.info["xscale"] = self.get('f') self.info["yscale"] = self.get('f') self.info["xsize"] = self.get('f') self.info["ysize"] = self.get('f') self.info["acmode"] = self.get('b') self.info["commonchnl"] = self.get('B') self.info["xtics"] = self.get('b') self.info["xrange"] = self.get('b') self.info["ytics"] = self.get('b') self.info["yrange"] = self.get('b') self.info["xscalevalue"] = self.get('f') self.info["xscaleinterval"] = self.get('f') self.info["yscalevalue"] = self.get('f') self.info["yscaleinterval"] = self.get('f') self.info["scaletoolx1"] = self.get('f') self.info["scaletooly1"] = self.get('f') self.info["scaletoolx2"] = self.get('f') self.info["scaletooly2"] = self.get('f') self.info["port"] = self.get('h') self.info["nsamples"] = self.get('L') self.info["filterflag"] = self.get('b') self.info["lowcutoff"] = self.get('f') self.info["lowpoles"] = self.get('h') self.info["highcutoff"] = self.get('f') self.info["highpoles"] = self.get('h') self.info["filtertype"] = self.get('b') self.info["filterdomain"] = self.get('b') self.info["snrflag"] = self.get('b') self.info["coherenceflag"] = self.get('b') self.info["continuoustype"] = self.get('b') self.info["eventtablepos"] = self.get('L') self.info["continuousseconds"] = self.get('f') self.info["channeloffset"] = self.get('l') self.info["autocorrectflag"] = self.get('b') self.info["dcthreshold"] = self.get('B') # The value stored for nsamples is sometimes wrong, for some reason. # This code corrects it. begdata = 900 + (nchannels * 75) enddata = self.info["eventtablepos"] nsamples = ((enddata - begdata)/self.info["nchannels"])/2 self.info["nsamples"] = nsamples
needs = p._str.split()[1:]
needs = p._str.split()
def check(self, *args, **kwds): _rpmts.check(self, *args, **kwds)
self.ndiffAssertEqual(f(s), '\n'.join(['case when foo = 1 then 2', ' when foo = 3 then 4', ' else 5',
self.ndiffAssertEqual(f(s), '\n'.join(['case', ' when foo = 1 then 2', ' when foo = 3 then 4', ' else 5',
def test_case(self): f = lambda sql: sqlparse.format(sql, reindent=True) s = 'case when foo = 1 then 2 when foo = 3 then 4 else 5 end' self.ndiffAssertEqual(f(s), '\n'.join(['case when foo = 1 then 2', ' when foo = 3 then 4', ' else 5', 'end']))
if x.next()(t):
if next(x)(t):
def _consume_cycle(tl, i): x = itertools.cycle(( lambda y: (y.match(T.Punctuation, '.') or y.ttype is T.Operator), lambda y: (y.ttype in (T.String.Symbol, T.Name, T.Wildcard)))) for t in tl.tokens[i:]: if x.next()(t): yield t else: raise StopIteration
return unicode(self).encode('utf-8')
return self.value or ''
def __str__(self): return unicode(self).encode('utf-8')
def __unicode__(self): return self.value or ''
def __repr__(self): short = self._get_repr_value() return '<%s \'%s\' at 0x%07x>' % (self._get_repr_name(), short, id(self))
def __unicode__(self): return ''.join(unicode(x) for x in self.flatten())
def __unicode__(self): return ''.join(unicode(x) for x in self.flatten())
return unicode(self).encode('utf-8')
return ''.join('%s' % x for x in self.flatten())
def __str__(self): return unicode(self).encode('utf-8')
data['proc_time'] = '%.3f' % proc_time or 0.0
data['proc_time'] = '%.3f' % (proc_time or 0.0)
def index(request): output = None data = {} proc_time = None if request.method == 'POST': logging.debug(request.POST) form = FormOptions(request.POST, request.FILES) if form.is_valid(): start = time.time() output = format_sql(form, format=request.POST.get('format', 'html')) proc_time = time.time()-start else: form = FormOptions() if request.POST.get('format', None) == 'json': logging.warning(form.errors) data['errors'] = str(form.errors) data['output'] = output logging.info('%r', proc_time) data['proc_time'] = '%.3f' % proc_time or 0.0 data = json.dumps(data) return HttpResponse(data, content_type='text/x-json') elif request.POST.get('format', None) == 'text': if not form.is_valid(): data = str(form.errors) # XXX convert to plain text else: data = output return HttpResponse(data, content_type='text/plain') return render_to_response('index.html', {'form': form, 'output': output, 'proc_time': proc_time and '%.3f' % proc_time or None, 'user': users.get_current_user(), 'login_url': users.create_login_url('/'), 'logout_url': users.create_logout_url('/'), 'userimg': _get_user_image(users.get_current_user()), 'examples': _get_examples()})
unittest.TextTestRunner(verbosity=2).run(suite)
return unittest.TextTestRunner(verbosity=2).run(suite)
def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite)
main(args)
result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite)
usage()
__usage()
def build_tree(search_path, outdir): """Build a directory tree under outdir containing symlinks to all LV2 extensions found in search_path, such that the symlink paths correspond to the extension URIs.""" for bundle in __bundles(search_path): # Load manifest into model manifest = RDF.Model() parser = RDF.Parser(name="guess") parser.parse_into_model(manifest, 'file://' + os.path.join(bundle, 'manifest.ttl')) # Query extension URI results = manifest.find_statements(RDF.Statement(None, rdf.type, lv2.Specification)) for r in results: ext_uri = str(r.subject.uri) ext_scheme = ext_uri[0:ext_uri.find(':')] ext_path = os.path.normpath(ext_uri[ext_uri.find(':') + 1:].lstrip('/')) ext_dir = os.path.join(outdir, ext_scheme, ext_path) # Make parent directories __mkdir_p(os.path.dirname(ext_dir)) # Remove existing symlink if necessary if os.access(ext_dir, os.F_OK): mode = os.lstat(ext_dir)[stat.ST_MODE] if stat.S_ISLNK(mode): os.remove(ext_dir) else: raise Exception(ext_dir + " exists and is not a link") # Make symlink to bundle directory os.symlink(bundle, ext_dir)
print_usage()
usage()
def usage(): print """Usage: lv2compatgen.py DATA
path = os.path.join(release_dir, 'lv2-%s-%s.0.tar.gz' % (b, rev)) subprocess.call(['tar', '-czf', path, os.path.join(outdir, '%s.lv2' % b)])
path = os.path.join(os.path.abspath(release_dir), 'lv2-%s-%s.0.tar.gz' % (b, rev)) subprocess.call(['tar', '--exclude-vcs', '-czf', path, bundle[bundle.find('/') + 1:]], cwd=dir)
def gendoc(specgen_dir, bundle_dir, ttl_filename, html_filename): subprocess.call([os.path.join(specgen_dir, 'lv2specgen.py'), os.path.join(bundle_dir, ttl_filename), os.path.join(specgen_dir, 'template.html'), os.path.join(specgen_dir, 'style.css'), os.path.join(out_base, html_filename), os.path.join('..', 'doc'), '-i'])
for name, old_kind_details in cwd:
for name, old_kind_details in cwd.iteritems():
def _gather_deleted_dir(self, path, dirdict): # List what the tree thought it had as deletes. pending = [(path, dirdict)] while pending: dirname, cwd = pending.pop(-1) for name, old_kind_details in cwd: path = dirname and ('%s/%s' % (dirname, name)) or name if type(old_kind_details) is dict: pending.append((path, old_kind_details)) old_kind_details = ('dir',) self.journal.add(path, 'del', old_kind_details)
cwd[name][1])
old_kind_details)
def finished(self): """Return the journal obtained by scanning the disk.""" pending = [''] while pending: dirname = pending.pop(-1) names = self.transport.list_dir(dirname) # NB: quadratic in lookup here due to presence in inner loop: # consider tuning. segments = dirname.split('/') cwd = self.tree for segment in segments: if not segment: continue try: cwd = cwd[segment] except KeyError: # totally new directory - added to journal by the directory # above. cwd = {} tree_names = set(cwd) names = set(names) for name in tree_names - names: # deletes path = dirname and ('%s/%s' % (dirname, name)) or name self.journal.add(path, 'del', cwd[name][1]) new_names = names - tree_names for name in names: path = dirname and ('%s/%s' % (dirname, name)) or name if path.endswith('.lmirror/metadata'): # metadata is transmitted by the act of fetching the # journal. continue statinfo = self.transport.stat(path) # Is it old enough to not check mtime = getattr(statinfo, 'st_mtime', 0) if self.last_timestamp - mtime > 3: continue kind = osutils.file_kind_from_stat_mode(statinfo.st_mode) if kind == 'file': f = self.transport.get(path) try: disk_size, disk_sha1 = osutils.size_sha_file(f) finally: f.close() new_kind_details = (kind, disk_sha1, disk_size) elif kind == 'symlink': new_kind_details = (kind, os.readlink(self.transport.local_abspath(path))) elif kind == 'directory': new_kind_details = ('dir',) pending.append(path) else: raise ValueError('unknown kind %r for %r' % (kind, path))
self.journal.add(path, 'replace', (cwd[name][1], new_kind_details))
old_kind_details = cwd[name] if type(old_kind_details) is dict: old_kind_details = ('dir',) if old_kind_details != new_kind_details: self.journal.add(path, 'replace', (old_kind_details, new_kind_details))
def finished(self): """Return the journal obtained by scanning the disk.""" pending = [''] while pending: dirname = pending.pop(-1) names = self.transport.list_dir(dirname) # NB: quadratic in lookup here due to presence in inner loop: # consider tuning. segments = dirname.split('/') cwd = self.tree for segment in segments: if not segment: continue try: cwd = cwd[segment] except KeyError: # totally new directory - added to journal by the directory # above. cwd = {} tree_names = set(cwd) names = set(names) for name in tree_names - names: # deletes path = dirname and ('%s/%s' % (dirname, name)) or name self.journal.add(path, 'del', cwd[name][1]) new_names = names - tree_names for name in names: path = dirname and ('%s/%s' % (dirname, name)) or name if path.endswith('.lmirror/metadata'): # metadata is transmitted by the act of fetching the # journal. continue statinfo = self.transport.stat(path) # Is it old enough to not check mtime = getattr(statinfo, 'st_mtime', 0) if self.last_timestamp - mtime > 3: continue kind = osutils.file_kind_from_stat_mode(statinfo.st_mode) if kind == 'file': f = self.transport.get(path) try: disk_size, disk_sha1 = osutils.size_sha_file(f) finally: f.close() new_kind_details = (kind, disk_sha1, disk_size) elif kind == 'symlink': new_kind_details = (kind, os.readlink(self.transport.local_abspath(path))) elif kind == 'directory': new_kind_details = ('dir',) pending.append(path) else: raise ValueError('unknown kind %r for %r' % (kind, path))
if content[0] != dir:
if content[0] != 'dir':
def replay(self): """Replay the journal.""" adds = [] deletes = [] replaces = [] for path, (action, content) in self.journal.paths.iteritems(): if action == 'new': adds.append((path, content)) elif action == 'del': deletes.append((path, content)) elif action == 'replace': replaces.append((path, content[0], content[1])) else: raise ValueError('unknown action %r for %r' % (action, path)) # Ordering /can/ be more complex than just adds/replace/deletes: # might have an add blocked on a replace making a dir above it. Likewise # A delete might have to happen before a replace when a dir becomes a # file. When we do smarter things, we'll have to just warn that we may # not honour general goals/policy if the user has given us such a # transform. # For now, simplest thing possible - want to get the concept performance # evaluated. adds.sort() for path, content in adds: self.put_with_check(path, content)() replaces.sort(reverse=True) to_rename = [] try: for path, _, new_content in replaces: # TODO: (again, to do with replacing files with dirs:) # do not delay creating dirs needed for files below them. to_rename.append(self.put_with_check(path, new_content)) for path, old_content, new_content in replaces: # TODO: we may want to warn or perhaps have a strict mode here. self.contentdir.delete(path) finally: for doit in to_rename: doit() # Children go first :) deletes.sort(reverse=True) for path,content in deletes: try: if content[0] != dir: self.contentdir.delete(path) else: self.contentdir.rmdir(path) except errors.NoSuchFile: # Already gone, ignore it. pass
realpath = self.sourcedir.local_abspath(path)
realpath = self.contentdir.local_abspath(path)
def put_with_check(self, path, content): """Put a_file at path checking that as received it matches content.
temppath = self.contentdir.local_abspath(tempname)
def put_with_check(self, path, content): """Put a_file at path checking that as received it matches content.
os.utime(tempname, (content.mtime, content.mtime))
os.utime(temppath, (content.mtime, content.mtime))
def put_with_check(self, path, content): """Put a_file at path checking that as received it matches content.
temppath = self.contentdir.local_abspath(tempname)
def put_with_check(self, path, content): """Put a_file at path checking that as received it matches content.
except OSError, e: if e.errno != errno.ENOENT: raise self.ui.output_log(4, __name__, 'Failed to set mtime for %r' % (temppath,))
def put_with_check(self, path, content): """Put a_file at path checking that as received it matches content.
requires = ['webit', 'gtk', 'gobject', 'keybinder', 'pynotify'],
requires = ['webkit', 'gtk', 'gobject', 'keybinder', 'pynotify'],
def get_data_files(root, data_dir): return [ (root + parent[len(data_dir):], [ os.path.join(parent, fn) for fn in files ]) for parent, dirs, files in os.walk(data_dir) if files ]
print raw_json
def crack_request(params): raw_json = urllib.unquote(params[1]) print raw_json request_info = dict([(k.encode('utf8'), v) for k, v in json.loads(raw_json).items()]) args = ( request_info['uuid'] , request_info['method'] , request_info['url'] , request_info['params'] , request_info['headers']) th = threading.Thread(target = request, args=args) th.start() pass
libwebkit = ctypes.CDLL('libwebkit-1.0.so.2')
try: libwebkit = ctypes.CDLL('libwebkit-1.0.so.2') pass except: libwebkit = ctypes.CDLL('libwebkitgtk-1.0.so.0.2.0') pass
def webkit_set_proxy_uri(uri): if uri and '://' not in uri: uri = 'https://' + uri pass try: if os.name == 'nt': libgobject = ctypes.CDLL('libgobject-2.0-0.dll') libsoup = ctypes.CDLL('libsoup-2.4-1.dll') libwebkit = ctypes.CDLL('libwebkit-1.0-2.dll') pass else: libgobject = ctypes.CDLL('libgobject-2.0.so.0') libsoup = ctypes.CDLL('libsoup-2.4.so.1') libwebkit = ctypes.CDLL('libwebkit-1.0.so.2') pass proxy_uri = libsoup.soup_uri_new(uri) if uri else 0 session = libwebkit.webkit_get_default_session() libgobject.g_object_set(session, "proxy-uri", proxy_uri, None) if proxy_uri: libsoup.soup_uri_free(proxy_uri) pass libgobject.g_object_set(session, "max-conns", 20, None) libgobject.g_object_set(session, "max-conns-per-host", 5, None) return 0 except: print 'error: webkit_set_proxy_uri' return 1 pass
print 'error: webkit_set_proxy_uri'
exctype, value = sys.exc_info()[:2] print 'error: webkit_set_proxy_uri: (%s, %s)' % (exctype,value)
def webkit_set_proxy_uri(uri): if uri and '://' not in uri: uri = 'https://' + uri pass try: if os.name == 'nt': libgobject = ctypes.CDLL('libgobject-2.0-0.dll') libsoup = ctypes.CDLL('libsoup-2.4-1.dll') libwebkit = ctypes.CDLL('libwebkit-1.0-2.dll') pass else: libgobject = ctypes.CDLL('libgobject-2.0.so.0') libsoup = ctypes.CDLL('libsoup-2.4.so.1') libwebkit = ctypes.CDLL('libwebkit-1.0.so.2') pass proxy_uri = libsoup.soup_uri_new(uri) if uri else 0 session = libwebkit.webkit_get_default_session() libgobject.g_object_set(session, "proxy-uri", proxy_uri, None) if proxy_uri: libsoup.soup_uri_free(proxy_uri) pass libgobject.g_object_set(session, "max-conns", 20, None) libgobject.g_object_set(session, "max-conns-per-host", 5, None) return 0 except: print 'error: webkit_set_proxy_uri' return 1 pass
gobject.idle_add(os.system, 'aplay -q "%s"' % utils.get_sound('notify'))
subprocess.Popen(['aplay', '-q', '-N', utils.get_sound('notify')])
def crack_system(params): if params[1] == 'notify' or params[1] == 'notify_with_sound': if not get_prefs('use_native_notify'): return summary = urllib.unquote(params[2]) body = urllib.unquote(params[3]) notify.update(summary, body) notify.show() if params[1] == 'notify_with_sound': gobject.idle_add(os.system, 'aplay -q "%s"' % utils.get_sound('notify')) elif params[1] == 'create_profile': profile = urllib.unquote(params[2]) callback = urllib.unquote(params[3]).replace('\n','') config.create_profile(profile) push_profiles() webv.execute_script(callback) elif params[1] == 'delete_profile': profile = urllib.unquote(params[2]) callback = urllib.unquote(params[3]).replace('\n','') config.delete_profile(profile) webv.execute_script(callback) elif params[1] == 'select_profile': app.active_profile = urllib.unquote(params[2]) app.window.set_title('Hotot | %s' % app.active_profile) config.loads(app.active_profile) apply_config() elif params[1] == 'select_protocol': app.protocol = urllib.unquote(params[2]) select_config(app.protocol) elif params[1] == 'sign_in': app.on_sign_in() elif params[1] == 'sign_out': push_profiles() app.on_sign_out() elif params[1] == 'quit': app.quit() pass
$('
$('.version').text('%s');
def apply_config(): version = 'ver %s (%s)'% (hotot.__version__, hotot.__codename__) exts_enabled = json.dumps(get_prefs('exts_enabled')) webv.execute_script(''' $('#version').text('%s'); ext.exts_enabled = %s; ''' % (version , exts_enabled)) apply_prefs() pass
msg = 'Unknow Errors ... '
msg = 'Unknown Errors ... '
def request(uuid, method, url, params={}, headers={},files=[],additions=''): scripts = '' try: if (method == 'POST'): result = _post(url, params, headers, files, additions) else: result = _get(url, params, headers) pass except urllib2.HTTPError, e: msg = 'Unknow Errors ... ' if http_code_msg_table.has_key(e.getcode()): msg = http_code_msg_table[e.getcode()] pass tech_info = 'HTTP Code: %s\\nURL: %s\\nDetails: %s' % (e.getcode(), e.geturl(), str(e)) content = '<p>%s</p><h3>- Technological Info -</h3><div class="dlg_group"><pre>%s</pre></div>' % (msg, tech_info) scripts = ''' ui.MessageDlg.set_text('%s', '%s'); ui.DialogHelper.open(ui.MessageDlg); lib.twitterapi.error_task_table['%s'](''); ''' % ('Ooops, an Error occurred!', content, uuid); pass except urllib2.URLError, e: content = '<p><label>Error Code:</label>%s<br/><label>Reason:</label> %s, %s<br/></p>' % (e.errno, e.reason, e.strerror) scripts = ''' ui.MessageDlg.set_text('%s', '%s'); ui.DialogHelper.open(ui.MessageDlg); lib.twitterapi.error_task_table['%s'](''); ''' % ('Ooops, an Error occurred!', content, uuid); pass else: if uuid != None: if result[0] != '{' and result[0] != '[': scripts = '''lib.twitterapi.success_task_table['%s']('%s'); ''' % (uuid, result) else: scripts = '''lib.twitterapi.success_task_table['%s'](%s); ''' % (uuid, result) pass scripts += '''delete lib.twitterapi.error_task_table['%s']; delete lib.twitterapi.error_task_table['%s']; ''' % (uuid, uuid); gobject.idle_add(webv.execute_script, scripts) pass
settings.set_property('enable-universal-access-from-file-uris', True) settings.set_property('enable-file-access-from-file-uris', True) settings.set_property('enable-page-cache', True) settings.set_property('tab-key-cycles-through-elements', True) settings.set_property('enable-default-context-menu', False)
try: settings.set_property('enable-universal-access-from-file-uris', True) settings.set_property('enable-default-context-menu', False) settings.set_property('enable-page-cache', True) settings.set_property('tab-key-cycles-through-elements', True) settings.set_property('enable-file-access-from-file-uris', True) except: pass
def __init__(self): WebView.__init__(self) self.load_finish_flag = False self.set_property('can-focus', True) self.set_property('can-default', True) self.set_full_content_zoom(1) self.clipbord = gtk.Clipboard()
mitem_resume.connect('activate', self.on_trayicon_activate);
mitem_resume.connect('activate', self.on_trayicon_activate)
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png')) self.window.set_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png'))
mitem_prefs.connect('activate', self.on_mitem_prefs_activate);
mitem_prefs.connect('activate', self.on_mitem_prefs_activate)
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png')) self.window.set_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png'))
mitem_about.connect('activate', self.on_mitem_about_activate);
mitem_about.connect('activate', self.on_mitem_about_activate)
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png')) self.window.set_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png'))
mitem_quit.connect('activate', self.on_mitem_quit_activate);
mitem_quit.connect('activate', self.on_mitem_quit_activate)
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png')) self.window.set_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png'))
mitem_resume.connect('activate', self.on_mitem_resume_activate);
mitem_resume.connect('activate', self.on_mitem_resume_activate)
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png')) self.window.set_icon_from_file( utils.get_ui_object('imgs/ic64_hotot_classics.png'))
pass
def _on_hotkey_compose(self): if config.get(self.active_profile, 'use_native_input'): if not self.tbox_status.is_focus(): self.inputw.hide() pass self.inputw.present() self.tbox_status.grab_focus() else: if not self.webv.is_focus(): self.window.hide() pass self.window.present() self.webv.grab_focus()
config.loads();
config.loads()
def main(): global HAS_INDICATOR gtk.gdk.threads_init() config.loads(); config.load_sys_conf() if not config.sys_get('use_ubuntu_indicator'): HAS_INDICATOR = False try: import dl libc = dl.open('/lib/libc.so.6') libc.call('prctl', 15, 'hotot', 0, 0, 0) except: pass agent.init_notify() app = Hotot() agent.app = app if HAS_INDICATOR: #TODO the icon is only work when installed to /usr/share/icons/hicolor/ indicator = appindicator.Indicator('hotot', 'hotot', appindicator.CATEGORY_COMMUNICATIONS) indicator.set_status(appindicator.STATUS_ACTIVE) indicator.set_attention_icon(utils.get_ui_object('imgs/ic64_hotot.png')) indicator.set_menu(app.menu_tray) gtk.gdk.threads_enter() gtk.main() gtk.gdk.threads_leave()
settings.set_property('tab-key-cycles-through-elements', False)
settings.set_property('tab-key-cycles-through-elements', True)
def __init__(self): webkit.WebView.__init__(self) self.load_finish_flag = False self.set_property('can-focus', True) self.set_property('can-default', True) self.set_full_content_zoom(1) self.clipbord = gtk.Clipboard()
gtk.window_set_default_icon_name("gtk-dnd") self.window.set_icon_name("gtk-dnd")
gtk.window_set_default_icon_from_file( config.abspath + '/imgs/ic64_hotot.png') self.window.set_icon_from_file( config.abspath + '/imgs/ic64_hotot.png')
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_name("gtk-dnd") self.window.set_icon_name("gtk-dnd") self.window.set_default_size(750, 550) self.window.set_title("Hotot") self.window.set_position(gtk.WIN_POS_CENTER)
self.window.set_title("Hotot")
self.window.set_title('Hotot')
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_name("gtk-dnd") self.window.set_icon_name("gtk-dnd") self.window.set_default_size(750, 550) self.window.set_title("Hotot") self.window.set_position(gtk.WIN_POS_CENTER)
libwebkit = ctypes.CDLL('libwebkitgtk-1.0.so.0.2.0')
libwebkit = ctypes.CDLL('libwebkitgtk-1.0.so.0')
def webkit_set_proxy_uri(uri): if uri and '://' not in uri: uri = 'https://' + uri pass try: if os.name == 'nt': libgobject = ctypes.CDLL('libgobject-2.0-0.dll') libsoup = ctypes.CDLL('libsoup-2.4-1.dll') libwebkit = ctypes.CDLL('libwebkit-1.0-2.dll') pass else: libgobject = ctypes.CDLL('libgobject-2.0.so.0') libsoup = ctypes.CDLL('libsoup-2.4.so.1') try: libwebkit = ctypes.CDLL('libwebkit-1.0.so.2') pass except: libwebkit = ctypes.CDLL('libwebkitgtk-1.0.so.0.2.0') pass pass proxy_uri = libsoup.soup_uri_new(uri) if uri else 0 session = libwebkit.webkit_get_default_session() libgobject.g_object_set(session, "proxy-uri", proxy_uri, None) if proxy_uri: libsoup.soup_uri_free(proxy_uri) pass libgobject.g_object_set(session, "max-conns", 20, None) libgobject.g_object_set(session, "max-conns-per-host", 5, None) return 0 except: exctype, value = sys.exc_info()[:2] print 'error: webkit_set_proxy_uri: (%s, %s)' % (exctype,value) return 1 pass
gobject.idle_add(webv.execute_script, "new Image().src='http://mobile.twitter.com/';")
gobject.idle_add(webv.execute_script, "new Image().src='http://google.com/';")
def apply_proxy_setting(): if get_prefs('use_http_proxy'): proxy_uri = "https://%s:%s" % ( get_prefs('http_proxy_host') , get_prefs('http_proxy_port')) webkit_set_proxy_uri(proxy_uri) pass else: webkit_set_proxy_uri("") pass gobject.idle_add(webv.execute_script, "new Image().src='http://mobile.twitter.com/';") pass
, font_family_list, font_family_used, font_size
, json.dumps(font_family_list), font_family_used, font_size
def push_prefs(): # account settings remember_password = 'true' if config.remember_password else 'false' consumer_key = config.consumer_key consumer_secret = config.consumer_secret # system settings shortcut_summon_hotot = config.shortcut_summon_hotot # display settings font_family_list = [ff.get_name() for ff in gtk.gdk.pango_context_get().list_families()] font_family_list.sort() font_family_used = config.font_family_used font_size = config.font_size use_native_input = 'true' if config.use_native_input else 'false' use_native_notify = 'true' if config.use_native_notify else 'false' # networks settings api_base = config.api_base; webv.execute_script(''' var prefs_obj = { "remember_password": %s , "consumer_key": "%s" , "consumer_secret": "%s" , "shortcut_summon_hotot": "%s" , "font_family_list": %s , "font_family_used": "%s" , "font_size": "%s" , "use_native_input": %s , "use_native_notify": %s , "api_base": "%s" }; ui.PrefsDlg.request_prefs_cb(eval(prefs_obj)); ''' % (remember_password , consumer_key, consumer_secret , shortcut_summon_hotot , font_family_list, font_family_used, font_size , use_native_input, use_native_notify , api_base)); pass
config.set(app.active_profile, 'api_base', 'https://identi.ca/api/')
set_prefs('api_base', 'https://identi.ca/api/')
def select_config(protocol): if protocol == 'identica': config.set(app.active_profile, 'api_base', 'https://identi.ca/api/') execute_script("lib.twitterapi.api_base='https://identi.ca/api/'") pass
return DEFAULT_SCREEN_NAME
return json.dumps(DEFAULT_SCREEN_NAME)
def get_screen_name(webv): if CAN_EVAL_SCRIPT: return webv.ctx().EvaluateScript(''' utility.DB.json(utility.DB.auto_complete_list) ''') else: return DEFAULT_SCREEN_NAME
ui.Header.update_status('%s');
ui.StatusBox.update_status('%s');
def update_status(text): view.execute_script(''' ui.Header.update_status('%s'); ''' % text); pass
if (method == 'POST'): result = _post(url, params, headers)
scripts = '' try: if (method == 'POST'): result = _post(url, params, headers) else: result = _get(url, params, headers) except urllib2.HTTPError, e: content = '<p><label>HTTP Code:</label> %s <br/><label>URL:</label> %s<br/><label>Details:</label> %s<br/></p>' % (e.getcode(), e.geturl(), str(e)) scripts = ''' ui.MessageDlg.set_text('%s', '%s'); ui.MessageDlg.show(); lib.twitterapi.error_task_table['%s'](''); ''' % ('Ooops, an Error occurred!', content, uuid); pass
def request(uuid, method, url, params={}, headers={}): if (method == 'POST'): result = _post(url, params, headers) else: result = _get(url, params, headers) scripts = '' if result[0] != '{' and result[0] != '[': scripts = '''lib.twitterapi.success_task_table['%s']('%s'); ''' % (uuid, result) else: scripts = '''lib.twitterapi.success_task_table['%s'](%s); ''' % (uuid, result) gobject.idle_add(webv.execute_script, scripts) pass
result = _get(url, params, headers) scripts = '' if result[0] != '{' and result[0] != '[': scripts = '''lib.twitterapi.success_task_table['%s']('%s'); ''' % (uuid, result) else: scripts = '''lib.twitterapi.success_task_table['%s'](%s); ''' % (uuid, result)
if result[0] != '{' and result[0] != '[': scripts = '''lib.twitterapi.success_task_table['%s']('%s'); ''' % (uuid, result) else: scripts = '''lib.twitterapi.success_task_table['%s'](%s); ''' % (uuid, result) pass scripts += '''delete lib.twitterapi.error_task_table['%s']; delete lib.twitterapi.error_task_table['%s']; ''' % (uuid, uuid);
def request(uuid, method, url, params={}, headers={}): if (method == 'POST'): result = _post(url, params, headers) else: result = _get(url, params, headers) scripts = '' if result[0] != '{' and result[0] != '[': scripts = '''lib.twitterapi.success_task_table['%s']('%s'); ''' % (uuid, result) else: scripts = '''lib.twitterapi.success_task_table['%s'](%s); ''' % (uuid, result) gobject.idle_add(webv.execute_script, scripts) pass
mitem_resume.connect('activate', self.on_mitem_resume_activate);
mitem_resume.connect('activate', self.on_trayicon_activate);
def build_gui(self): self.window = gtk.Window() gtk.window_set_default_icon_from_file( config.get_ui_object('imgs/ic64_hotot.png')) self.window.set_icon_from_file( config.get_ui_object('imgs/ic64_hotot.png'))
builder.set_field_text('donors', _('Donors'), "Separate multiple entries using comments.")
builder.set_field_text('donors', _('Donors'), "Separate multiple entries using commas.")
def build_package_iati_form(is_admin=False): builder = package.build_package_form() # IATI specifics #Publishing Entity: builder.add_field(AtLeastOneGroupSelectField('groups', allow_empty=False)) #builder.add_field(common.TextExtraField('publisher')) #builder.set_field_text('publisher', _('Publishing entity')) #Publishing Entity Type: (Donor, Recipient, Community Data..) #builder.add_field(SelectExtraField('publisher_type', options=PUBLISHER_TYPES)) #builder.set_field_text('publisher_type', _('Publishing entity type')) #Donor (TODO: Generate from crawler) # Editable List, CSV? builder.add_field(CommaListExtraField('donors')) builder.set_field_text('donors', _('Donors'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_type')) builder.set_field_text('donors_type', _('Donor type'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_country')) builder.set_field_text('donors_country', _('Donor country'), "Separate multiple entries using comments.") # TODO: Enforce validation countries = [(v, k) for k, v in COUNTRIES] builder.add_field(SelectExtraField('country', options=countries)) builder.set_field_text('country', _('Recipient country')) #Verification status: enumeration of statuses (checked, not checked etc) # TODO: Enforce validation, can probably only be set by admins builder.add_field(common.CheckboxExtraField('verified')) builder.set_field_text('verified', _('Verification')) builder.add_field(common.CheckboxExtraField('archive_file')) builder.set_field_text('archive_file', _('Archive')) #Activity period: (Generate from crawler) builder.add_field(common.DateRangeExtraField('activity_period')) builder.set_field_text('activity_period', _('Activitiy Period')) #Resource links: to the actual IATI record #Number of activities: (Generate from crawler) builder.add_field(common.TextExtraField('activity_count')) builder.set_field_text('activity_count', _('Num. Activities')) #Date record updated: builder.add_field(common.TextExtraField('record_updated')) builder.set_field_text('record_updated', _('Record updated')) #Date data updated: builder.add_field(common.TextExtraField('data_updated')) builder.set_field_text('data_updated', _('Data updated')) #License: Need this field even if it may be a standard license builder.add_field(common.TextExtraField('license')) builder.set_field_text('license', _('License')) #Department # TODO: Make this a group property instead? builder.add_field(common.TextExtraField('department')) builder.set_field_text('department', _('Department')) #Contact builder.set_field_text('author', _('Contact')) #Contact e-mail builder.set_field_text('author_email', _('Contact e-mail')) #Licence builder.set_field_text('license_id', _('License')) #Resource format #Resource URL #Resource ID # -- do we have an ID? # Layout field_groups = OrderedDict([ (_('Basic information'), ['name', 'title', 'author', 'author_email', 'department',]), (_('Publishing Entity'), ['groups']), (_('Details'), ['country', 'donors', 'donors_type', 'donors_country', 'record_updated', 'data_updated', 'license_id', 'tags', 'notes']), (_('Resources'), ['resources']), (_('Verification and Analysis'), [ 'activity_period', 'activity_count', 'archive_file', ]), ]) if is_admin: field_groups[_('Verification and Analysis')].append('verified') field_groups[_('Verification and Analysis')].append('state') builder.set_displayed_fields(field_groups) return builder # Strings for i18n: [_('External reference'), _('Date released'), _('Date updated'), _('Update frequency'), _('Geographic granularity'), _('Geographic coverage'), _('Temporal granularity'), _('Temporal coverage'), _('Categories'), _('National Statistic'), _('Precision'), _('Taxonomy URL'), _('Department'), _('Agency'), ]
builder.set_field_text('donors_type', _('Donor type'), "Separate multiple entries using comments.")
builder.set_field_text('donors_type', _('Donor type'), "Separate multiple entries using commas.")
def build_package_iati_form(is_admin=False): builder = package.build_package_form() # IATI specifics #Publishing Entity: builder.add_field(AtLeastOneGroupSelectField('groups', allow_empty=False)) #builder.add_field(common.TextExtraField('publisher')) #builder.set_field_text('publisher', _('Publishing entity')) #Publishing Entity Type: (Donor, Recipient, Community Data..) #builder.add_field(SelectExtraField('publisher_type', options=PUBLISHER_TYPES)) #builder.set_field_text('publisher_type', _('Publishing entity type')) #Donor (TODO: Generate from crawler) # Editable List, CSV? builder.add_field(CommaListExtraField('donors')) builder.set_field_text('donors', _('Donors'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_type')) builder.set_field_text('donors_type', _('Donor type'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_country')) builder.set_field_text('donors_country', _('Donor country'), "Separate multiple entries using comments.") # TODO: Enforce validation countries = [(v, k) for k, v in COUNTRIES] builder.add_field(SelectExtraField('country', options=countries)) builder.set_field_text('country', _('Recipient country')) #Verification status: enumeration of statuses (checked, not checked etc) # TODO: Enforce validation, can probably only be set by admins builder.add_field(common.CheckboxExtraField('verified')) builder.set_field_text('verified', _('Verification')) builder.add_field(common.CheckboxExtraField('archive_file')) builder.set_field_text('archive_file', _('Archive')) #Activity period: (Generate from crawler) builder.add_field(common.DateRangeExtraField('activity_period')) builder.set_field_text('activity_period', _('Activitiy Period')) #Resource links: to the actual IATI record #Number of activities: (Generate from crawler) builder.add_field(common.TextExtraField('activity_count')) builder.set_field_text('activity_count', _('Num. Activities')) #Date record updated: builder.add_field(common.TextExtraField('record_updated')) builder.set_field_text('record_updated', _('Record updated')) #Date data updated: builder.add_field(common.TextExtraField('data_updated')) builder.set_field_text('data_updated', _('Data updated')) #License: Need this field even if it may be a standard license builder.add_field(common.TextExtraField('license')) builder.set_field_text('license', _('License')) #Department # TODO: Make this a group property instead? builder.add_field(common.TextExtraField('department')) builder.set_field_text('department', _('Department')) #Contact builder.set_field_text('author', _('Contact')) #Contact e-mail builder.set_field_text('author_email', _('Contact e-mail')) #Licence builder.set_field_text('license_id', _('License')) #Resource format #Resource URL #Resource ID # -- do we have an ID? # Layout field_groups = OrderedDict([ (_('Basic information'), ['name', 'title', 'author', 'author_email', 'department',]), (_('Publishing Entity'), ['groups']), (_('Details'), ['country', 'donors', 'donors_type', 'donors_country', 'record_updated', 'data_updated', 'license_id', 'tags', 'notes']), (_('Resources'), ['resources']), (_('Verification and Analysis'), [ 'activity_period', 'activity_count', 'archive_file', ]), ]) if is_admin: field_groups[_('Verification and Analysis')].append('verified') field_groups[_('Verification and Analysis')].append('state') builder.set_displayed_fields(field_groups) return builder # Strings for i18n: [_('External reference'), _('Date released'), _('Date updated'), _('Update frequency'), _('Geographic granularity'), _('Geographic coverage'), _('Temporal granularity'), _('Temporal coverage'), _('Categories'), _('National Statistic'), _('Precision'), _('Taxonomy URL'), _('Department'), _('Agency'), ]
builder.set_field_text('donors_country', _('Donor country'), "Separate multiple entries using comments.")
builder.set_field_text('donors_country', _('Donor country'), "Separate multiple entries using commas.")
def build_package_iati_form(is_admin=False): builder = package.build_package_form() # IATI specifics #Publishing Entity: builder.add_field(AtLeastOneGroupSelectField('groups', allow_empty=False)) #builder.add_field(common.TextExtraField('publisher')) #builder.set_field_text('publisher', _('Publishing entity')) #Publishing Entity Type: (Donor, Recipient, Community Data..) #builder.add_field(SelectExtraField('publisher_type', options=PUBLISHER_TYPES)) #builder.set_field_text('publisher_type', _('Publishing entity type')) #Donor (TODO: Generate from crawler) # Editable List, CSV? builder.add_field(CommaListExtraField('donors')) builder.set_field_text('donors', _('Donors'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_type')) builder.set_field_text('donors_type', _('Donor type'), "Separate multiple entries using comments.") builder.add_field(CommaListExtraField('donors_country')) builder.set_field_text('donors_country', _('Donor country'), "Separate multiple entries using comments.") # TODO: Enforce validation countries = [(v, k) for k, v in COUNTRIES] builder.add_field(SelectExtraField('country', options=countries)) builder.set_field_text('country', _('Recipient country')) #Verification status: enumeration of statuses (checked, not checked etc) # TODO: Enforce validation, can probably only be set by admins builder.add_field(common.CheckboxExtraField('verified')) builder.set_field_text('verified', _('Verification')) builder.add_field(common.CheckboxExtraField('archive_file')) builder.set_field_text('archive_file', _('Archive')) #Activity period: (Generate from crawler) builder.add_field(common.DateRangeExtraField('activity_period')) builder.set_field_text('activity_period', _('Activitiy Period')) #Resource links: to the actual IATI record #Number of activities: (Generate from crawler) builder.add_field(common.TextExtraField('activity_count')) builder.set_field_text('activity_count', _('Num. Activities')) #Date record updated: builder.add_field(common.TextExtraField('record_updated')) builder.set_field_text('record_updated', _('Record updated')) #Date data updated: builder.add_field(common.TextExtraField('data_updated')) builder.set_field_text('data_updated', _('Data updated')) #License: Need this field even if it may be a standard license builder.add_field(common.TextExtraField('license')) builder.set_field_text('license', _('License')) #Department # TODO: Make this a group property instead? builder.add_field(common.TextExtraField('department')) builder.set_field_text('department', _('Department')) #Contact builder.set_field_text('author', _('Contact')) #Contact e-mail builder.set_field_text('author_email', _('Contact e-mail')) #Licence builder.set_field_text('license_id', _('License')) #Resource format #Resource URL #Resource ID # -- do we have an ID? # Layout field_groups = OrderedDict([ (_('Basic information'), ['name', 'title', 'author', 'author_email', 'department',]), (_('Publishing Entity'), ['groups']), (_('Details'), ['country', 'donors', 'donors_type', 'donors_country', 'record_updated', 'data_updated', 'license_id', 'tags', 'notes']), (_('Resources'), ['resources']), (_('Verification and Analysis'), [ 'activity_period', 'activity_count', 'archive_file', ]), ]) if is_admin: field_groups[_('Verification and Analysis')].append('verified') field_groups[_('Verification and Analysis')].append('state') builder.set_displayed_fields(field_groups) return builder # Strings for i18n: [_('External reference'), _('Date released'), _('Date updated'), _('Update frequency'), _('Geographic granularity'), _('Geographic coverage'), _('Temporal granularity'), _('Temporal coverage'), _('Categories'), _('National Statistic'), _('Precision'), _('Taxonomy URL'), _('Department'), _('Agency'), ]
label = NomGene+AjouteZero(i)+"-"
label = NomGene+"-"+AjouteZero(i)
def AjouteZero(n): N = str(n) a = [] for i in range(len(N),4): a.append("0") a.append(N) return "".join(a)
print f.filename,
print f.basename,
def AjouteZero(n): N = str(n) a = [] for i in range(len(N),4): a.append("0") a.append(N) return "".join(a)
self.raw_data = None
def __init__(self, label=u'', validators=None, parse_kwargs=None, display_format='%Y-%m-%d %H:%M', **kwargs): super(DateTimeField, self).__init__(label, validators, **kwargs) if parse_kwargs is None: parse_kwargs = {} self.parse_kwargs = parse_kwargs self.display_format = display_format self.raw_data = None
if self.raw_data is not None: return self.raw_data
if self.raw_data: return u' '.join(self.raw_data)
def _value(self): if self.raw_data is not None: return self.raw_data else: return self.data and self.data.strftime(self.display_format) or u''
self.raw_data = str.join(' ', valuelist)
date_str = u' '.join(valuelist)
def process_formdata(self, valuelist): if valuelist: self.raw_data = str.join(' ', valuelist) parse_kwargs = self.parse_kwargs.copy() if 'default' not in parse_kwargs: try: parse_kwargs['default'] = self.default() except TypeError: parse_kwargs['default'] = self.default try: self.data = parser.parse(self.raw_data, **parse_kwargs) except ValueError: self.data = None raise ValidationError(u'Invalid date/time input')
self.data = parser.parse(self.raw_data, **parse_kwargs)
self.data = parser.parse(date_str, **parse_kwargs)
def process_formdata(self, valuelist): if valuelist: self.raw_data = str.join(' ', valuelist) parse_kwargs = self.parse_kwargs.copy() if 'default' not in parse_kwargs: try: parse_kwargs['default'] = self.default() except TypeError: parse_kwargs['default'] = self.default try: self.data = parser.parse(self.raw_data, **parse_kwargs) except ValueError: self.data = None raise ValidationError(u'Invalid date/time input')
if field.raw_data is not None and (not field.raw_data or not field.raw_data[0].strip()):
if field.raw_data is None or not field.raw_data or not field.raw_data[0].strip():
def __call__(self, form, field): if field.raw_data is not None and (not field.raw_data or not field.raw_data[0].strip()): field.errors[:] = [] raise StopValidation()
raise ValueError('Missing reference_class attribute in '
raise TypeError('Missing reference_class attribute in '
def __init__(self, label=u'', validators=None, reference_class=None, label_attr=None, allow_blank=False, blank_text=u'', **kwargs): super(ReferencePropertyField, self).__init__(label, validators, **kwargs) self.label_attr = label_attr self.allow_blank = allow_blank self.blank_text = blank_text self._set_data(None) if reference_class is None: raise ValueError('Missing reference_class attribute in ' 'ReferencePropertyField')
raise ValidationError('Not a valid choice')
raise ValueError(self.gettext(u'Not a valid choice'))
def pre_validate(self, form): if not self.allow_blank or self.data is not None: for obj in self.queryset: if self.data == str(obj.key()): break else: raise ValidationError('Not a valid choice')
self.assert_("does not match format" in form.a.errors[0])
self.assert_("not match format" in form.a.errors[0])
def test(self): d = datetime(2008, 5, 5, 4, 30, 0, 0) form = self.F(DummyPostData(a=['2008-05-05', '04:30:00'], b=['2008-05-05 04:30'])) self.assertEqual(form.a.data, d) self.assertEqual(form.a(), u"""<input id="a" name="a" type="text" value="2008-05-05 04:30:00" />""") self.assertEqual(form.b.data, d) self.assertEqual(form.b(), u"""<input id="b" name="b" type="text" value="2008-05-05 04:30" />""") self.assert_(form.validate()) form = self.F(DummyPostData(a=['2008-05-05'])) self.assert_(not form.validate()) self.assert_("does not match format" in form.a.errors[0])
self.assertEqual(form.name._default, None) self.assertEqual(form.city._default, None) self.assertEqual(form.age._default, None) self.assertEqual(form.is_admin._default, False)
self.assertEqual(form.name.default, None) self.assertEqual(form.city.default, None) self.assertEqual(form.age.default, None) self.assertEqual(form.is_admin.default, False)
def test_default_value(self): form_class = model_form(Author)
for obj in self.queryset:
for obj in self.query:
def pre_validate(self, form): if not self.allow_blank or self.data is not None: for obj in self.queryset: if self.data == str(obj.key()): break else: raise ValueError(self.gettext(u'Not a valid choice'))
return f.DateTimeField(format='%Y-%m-%d', **kwargs)
return f.DateField(format='%Y-%m-%d', **kwargs)
def convert_DateProperty(model, prop, kwargs): """Returns a form field for a ``db.DateProperty``.""" if prop.auto_now or prop.auto_now_add: return None return f.DateTimeField(format='%Y-%m-%d', **kwargs)