rem
stringlengths 0
322k
| add
stringlengths 0
2.05M
| context
stringlengths 8
228k
|
---|---|---|
Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. | Design an Nth order lowpass digital or analog filter and return the filter coefficients in (B,A) (numerator, denominator) or (Z,P,K) form. | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
""" | N -- the order of the filter. Wn -- a scalar or length-2 sequence giving the critical frequencies. rp, rs -- For chebyshev and elliptic filters provides the maximum ripple in the passband and the minimum attenuation in the stop band. btype -- the type of filter (lowpass, highpass, bandpass, or bandstop). analog -- non-zero to return an analog filter, otherwise a digital filter is returned. ftype -- the type of IIR filter (Butterworth, Cauer (Elliptic), Bessel, Chebyshev1, Chebyshev2) output -- 'ba' for (B,A) output, 'zpk' for (Z,P,K) output. """ ftype, btype, output = map(string.lower, (ftype, btype, output)) Wn = asarray(Wn) try: btype = band_dict[btype] except KeyError: raise ValueError, "%s is an invalid bandtype for filter." % btype try: typefunc = filter_dict[ftype][0] except KeyError: raise ValueError, "%s is not a valid basic iir filter." % ftype if output not in ['ba', 'zpk']: raise ValueError, "%s is not a valid output form." % output | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
if btype in ['lowpass', 'highpass']: wo = warped else: bw = warped[1] - warped[0] wo = sqrt(warped[0]*warped[1]) | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
|
if typefunc in [buttap, besselap]: z, p, k = typefunc(N) elif typefunc == cheb1ap: if rp is None: raise ValueError, "passband ripple (rp) must be provided to design a Chebyshev I filter." z, p, k = typefunc(N, rp) elif typefunc == cheb2ap: if rs is None: raise ValueError, "stopband atteunatuion (rs) must be provided to design an Chebyshev II filter." z, p, k = typefunc(N, rs) else: if rs is None or rp is None: raise ValueErrro, "Both rp and rs must be provided to design an elliptic filter." z, p, k = typefunc(N, rp, rs) b, a = zpk2tf(z,p,k) | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
|
pass | b, a = bilinear(b, a, fs=fs) | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
pass def cheby1(): pass def cheby2(): pass def ellip(): pass def besself(): pass | if output == 'zpk': return tf2zpk(b,a) else: return b,a def butter(N, Wn, btype='low', analog=0, output='ba'): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital or analog Butterworth filter and return the filter coefficients in (B,A) or (Z,P,K) form. See also buttord. """ return iirfilter(N, Wn, btype=btype, analog=analog, output=output, ftype='butter') def cheby1(N, rp, Wn, btype='low', analog=0, output='ba'): """Chebyshev type I digital and analog filter design. Description: Design an Nth order lowpass digital or analog Chebyshev type I filter and return the filter coefficients in (B,A) or (Z,P,K) form. See also cheyb1ord. """ return iirfilter(N, Wn, rp=rp, btype=btype, analog=analog, output=output, ftype='cheby1') def cheby2(N, rs, Wn, btype='low', analog=0, output='ba'): """Chebyshev type I digital and analog filter design. Description: Design an Nth order lowpass digital or analog Chebyshev type I filter and return the filter coefficients in (B,A) or (Z,P,K) form. See also cheyb2ord. """ return iirfilter(N, Wn, rs=rs, btype=btype, analog=analog, output=output, ftype='cheby2') def ellip(N, rp, rs, Wn, btype='low', analog=0, output='ba'): """Elliptic (Cauer) digital and analog filter design. Description: Design an Nth order lowpass digital or analog elliptic filter and return the filter coefficients in (B,A) or (Z,P,K) form. See also ellipord. """ return iirfilter(N, Wn, rs=rs, rp=rp, btype=btype, analog=analog, output=output, ftype='elliptic') def bessel(N, Wn, btype='low', analog=0, output='ba'): """Bessel digital and analog filter design. Description: Design an Nth order lowpass digital or analog Bessel filter and return the filter coefficients in (B,A) or (Z,P,K) form. """ return iirfilter(N, Wn, btype=btype, analog=analog, output=output, ftype='bessel') | def butter(N, Wn, bandtype='band', analog=0, output=''): """Butterworth digital and analog filter design. Description: Design an Nth order lowpass digital Butterworth filter and return the filter coefficients in (B,A) form. Inputs: """ #pre-warp frequencies for digital filter design if not analog: warped = 2*fs*tan(pi*Wn/fs) else: warped = Wn # convert to low-pass prototype # Get analog lowpass prototype # transform to lowpass, bandpass, highpass, or bandstop # Find discrete equivalent if necessary if not analog: pass # Transform to proper out type (pole-zero, state-space, numer-denom) pass |
try: return ret[0] except IndexError, TypeError: return ret | return ret | def vratio(u, ineps, mp): [s,c,d,phi] = special.ellpj(u,mp) ret = abs(ineps - s/c) try: return ret[0] except IndexError, TypeError: return ret |
try: return abs(r)[0] except TypeError: return abs(r) | return abs(r) | def kratio(m, k_ratio): if m < 0: m = 0.0 if m > 1: m = 1.0 if abs(m) > EPSILON and (abs(m) + EPSILON) < 1: k = special.ellpk([1-m,m]) r = k[0] / k[1] - k_ratio elif abs(m) > EPSILON: r = -k_ratio else: r = 1e20 try: return abs(r)[0] except TypeError: return abs(r) |
maxfun=250, maxiter=250, printmessg=0) v0 = capk * r / (N*capck1[0]) | maxfun=250, maxiter=250, disp=0) v0 = capk * r / (N*val[0]) | def ellipap(N,rp,rs): """Return (z,p,k) zeros, poles, and gain of an Nth order normalized prototype elliptic analog lowpass filter with rp decibels of ripple in the passband and a stopband rs decibels down. Broken... """ if N == 1: p = -sqrt(1.0/(10**(0.1*rp)-1.0)) k = -p z = [] return z, p, k eps = Num.sqrt(10**(0.1*rp)-1) ck1 = eps / Num.sqrt(10**(0.1*rs)-1) ck1p = Num.sqrt(1-ck1*ck1) if ck1p == 1: raise ValueError, "Cannot design a filter with given rp and rs specifications." wp = 1 val = special.ellpk([1-ck1*ck1,1-ck1p*ck1p]) if abs(1-ck1p*ck1p) < EPSILON: krat = 0 else: krat = N*val[0] / val[1] m = optimize.fmin(kratio, 0.5, args=(krat,), maxfun=250, maxiter=250, disp=0) if m < 0 or m > 1: m = optimize.fminbound(kratio, 0, 1, args=(krat,), maxfun=250, maxiter=250, disp=0) capk = special.ellpk(1-m) ws = wp / sqrt(m) m1 = 1-m j = arange(1-N%2,N,2) jj = len(j) [s,c,d,phi] = special.ellpj(j*capk/N,m*ones(jj)) snew = Num.compress(abs(s) > EPSILON, s) z = 1.0 / (sqrt(m)*snew) z = 1j*z z = Num.concatenate((z,conjugate(z))) r = optimize.fmin(vratio, special.ellpk(1-m), args=(1/eps, ck1p*ck1p), maxfun=250, maxiter=250, printmessg=0) v0 = capk * r / (N*capck1[0]) [sv,cv,dv,phi] = special.ellpj(v0,1-m) p = -(c*d*sv*cv + 1j*s*dv) / (1-(d*sv)**2.0) if N % 2: newp = Num.compress(abs(p.imag) > EPSILON*Num.sqrt(MLab.sum(p*Num.conjugate(p)).real), p) p = Num.concatenate((p,conjugate(newp))) else: p = Num.concatenate((p,conjugate(p))) k = (MLab.prod(-p) / MLab.prod(-z)).real if N % 2 == 0: k = k / Num.sqrt((1+eps*eps)) return z, p, k |
return where(c>0,val,1-val) | cond = c + 0*val return where(cond>0,val,1-val) | def _cdf(self, x, a, c): val = special.gammainc(a,x**c) return where(c>0,val,1-val) |
val2 = special.gammaincinv(a,1-q) | val2 = special.gammaincinv(a,1.0-q) | def _ppf(self, q, a, c): val1 = special.gammaincinv(a,q) val2 = special.gammaincinv(a,1-q) ic = 1.0/c return where(c>0,pow(val1,ic),pow(val2,ic)) |
return where(c>0,pow(val1,ic),pow(val2,ic)) | cond = c+0*val1+0*val2 return where(cond > 0,val1**ic,val2**ic) | def _ppf(self, q, a, c): val1 = special.gammaincinv(a,q) val2 = special.gammaincinv(a,1-q) ic = 1.0/c return where(c>0,pow(val1,ic),pow(val2,ic)) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def diff(x,order=1,period=None, _cache = _cache): """ diff(x, order=1, period=2*pi) -> y Return k-th derivative (or integral) of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = pow(sqrt(-1)*j*2*pi/period, order) * x_j y_0 = 0 if order is not 0. Optional input: order The order of differentiation. Default order is 1. If order is negative, then integration is carried out under the assumption that x_0==0. period The assumed period of the sequence. Default is 2*pi. Notes: If sum(x)=0 then diff(diff(x,k),-k)==x (within numerical accuracy) For odd order and even len(x), the Nyquist mode is taken zero. """ tmp = asarray(x) if order==0: return tmp if tmp.dtype in (scipy.Complex32, scipy.Complex64): return diff(tmp.real,order,period)+1j*diff(tmp.imag,order,period) if period is not None: c = 2*pi/period else: c = 1.0 n = len(x) omega = _cache.get((n,order,c)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,order=order,c=c): if k: return pow(c*k,order) return 0 omega = convolve.init_convolution_kernel(n,kernel,d=order, zero_nyquist=1) _cache[(n,order,c)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=order%2, overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def tilbert(x,h,period=None, _cache = _cache): """ tilbert(x, h, period=2*pi) -> y Return h-Tilbert transform of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = -sqrt(-1)*coth(j*h*2*pi/period) * x_j y_0 = 0 Input: h Defines the parameter of the Tilbert transform. period The assumed period of the sequence. Default period is 2*pi. Notes: If sum(x)==0 and n=len(x) is odd then tilbert(itilbert(x)) == x If 2*pi*h/period is approximately 10 or larger then numerically tilbert == hilbert (theoretically oo-Tilbert == Hilbert). For even len(x), the Nyquist mode of x is taken zero. """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return tilbert(tmp.real,h,period)+\ 1j*tilbert(tmp.imag,h,period) if period is not None: h = h*2*pi/period n = len(x) omega = _cache.get((n,h)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,h=h): if k: return -1/tanh(h*k) return 0 omega = convolve.init_convolution_kernel(n,kernel,d=1) _cache[(n,h)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=1,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def itilbert(x,h,period=None, _cache = _cache): """ itilbert(x, h, period=2*pi) -> y Return inverse h-Tilbert transform of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = sqrt(-1)*tanh(j*h*2*pi/period) * x_j y_0 = 0 Optional input: see tilbert.__doc__ """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return itilbert(tmp.real,h,period)+\ 1j*itilbert(tmp.imag,h,period) if period is not None: h = h*2*pi/period n = len(x) omega = _cache.get((n,h)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,h=h): if k: return tanh(h*k) return 0 omega = convolve.init_convolution_kernel(n,kernel,d=1) _cache[(n,h)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=1,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def hilbert(x, _cache=_cache): """ hilbert(x) -> y Return Hilbert transform of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = -sqrt(-1)*sign(j) * x_j y_0 = 0 Notes: If sum(x)==0 then hilbert(ihilbert(x)) == x For even len(x), the Nyquist mode of x is taken zero. """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return hilbert(tmp.real,tol)+1j*hilbert(tmp.imag) n = len(x) omega = _cache.get(n) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k): if k>0: return -1 elif k<0: return 1 return 0 omega = convolve.init_convolution_kernel(n,kernel,d=1) _cache[n] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=1,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def cs_diff(x, a, b, period=None, _cache = _cache): """ cs_diff(x, a, b, period=2*pi) -> y Return (a,b)-cosh/sinh pseudo-derivative of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = -sqrt(-1)*cosh(j*a*2*pi/period)/sinh(j*b*2*pi/period) * x_j y_0 = 0 Input: a,b Defines the parameters of the cosh/sinh pseudo-differential operator. period The period of the sequence. Default period is 2*pi. Notes: For even len(x), the Nyquist mode of x is taken zero. """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return cs_diff(tmp.real,a,b,period)+\ 1j*cs_diff(tmp.imag,a,b,period) if period is not None: a = a*2*pi/period b = b*2*pi/period n = len(x) omega = _cache.get((n,a,b)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,a=a,b=b): if k: return -cosh(a*k)/sinh(b*k) return 0 omega = convolve.init_convolution_kernel(n,kernel,d=1) _cache[(n,a,b)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=1,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def sc_diff(x, a, b, period=None, _cache = _cache): """ sc_diff(x, a, b, period=2*pi) -> y Return (a,b)-sinh/cosh pseudo-derivative of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = sqrt(-1)*sinh(j*a*2*pi/period)/cosh(j*b*2*pi/period) * x_j y_0 = 0 Input: a,b Defines the parameters of the sinh/cosh pseudo-differential operator. period The period of the sequence x. Default is 2*pi. Notes: sc_diff(cs_diff(x,a,b),b,a) == x For even len(x), the Nyquist mode of x is taken zero. """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return sc_diff(tmp.real,a,b,period)+\ 1j*sc_diff(tmp.imag,a,b,period) if period is not None: a = a*2*pi/period b = b*2*pi/period n = len(x) omega = _cache.get((n,a,b)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,a=a,b=b): if k: return sinh(a*k)/cosh(b*k) return 0 omega = convolve.init_convolution_kernel(n,kernel,d=1) _cache[(n,a,b)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,swap_real_imag=1,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def ss_diff(x, a, b, period=None, _cache = _cache): """ ss_diff(x, a, b, period=2*pi) -> y Return (a,b)-sinh/sinh pseudo-derivative of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = sinh(j*a*2*pi/period)/sinh(j*b*2*pi/period) * x_j y_0 = a/b * x_0 Input: a,b Defines the parameters of the sinh/sinh pseudo-differential operator. period The period of the sequence x. Default is 2*pi. Notes: ss_diff(ss_diff(x,a,b),b,a) == x """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return ss_diff(tmp.real,a,b,period)+\ 1j*ss_diff(tmp.imag,a,b,period) if period is not None: a = a*2*pi/period b = b*2*pi/period n = len(x) omega = _cache.get((n,a,b)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,a=a,b=b): if k: return sinh(a*k)/sinh(b*k) return float(a)/b omega = convolve.init_convolution_kernel(n,kernel) _cache[(n,a,b)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def cc_diff(x, a, b, period=None, _cache = _cache): """ cc_diff(x, a, b, period=2*pi) -> y Return (a,b)-cosh/cosh pseudo-derivative of a periodic sequence x. If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = cosh(j*a*2*pi/period)/cosh(j*b*2*pi/period) * x_j Input: a,b Defines the parameters of the sinh/sinh pseudo-differential operator. Optional input: period The period of the sequence x. Default is 2*pi. Notes: cc_diff(cc_diff(x,a,b),b,a) == x """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return cc_diff(tmp.real,a,b,period)+\ 1j*cc_diff(tmp.imag,a,b,period) if period is not None: a = a*2*pi/period b = b*2*pi/period n = len(x) omega = _cache.get((n,a,b)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel(k,a=a,b=b): return cosh(a*k)/cosh(b*k) omega = convolve.init_convolution_kernel(n,kernel) _cache[(n,a,b)] = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve(tmp,omega,overwrite_x=overwrite_x) |
if tmp.dtype in (scipy.Complex32, scipy.Complex64): | if tmp.dtypechar in (scipy.Complex32, scipy.Complex64): | def shift(x, a, period=None, _cache = _cache): """ shift(x, a, period=2*pi) -> y Shift periodic sequence x by a: y(u) = x(u+a). If x_j and y_j are Fourier coefficients of periodic functions x and y, respectively, then y_j = exp(j*a*2*pi/period*sqrt(-1)) * x_f Optional input: period The period of the sequences x and y. Default period is 2*pi. """ tmp = asarray(x) if tmp.dtype in (scipy.Complex32, scipy.Complex64): return shift(tmp.real,a,period)+1j*shift(tmp.imag,a,period) if period is not None: a = a*2*pi/period n = len(x) omega = _cache.get((n,a)) if omega is None: if len(_cache)>20: while _cache: _cache.popitem() def kernel_real(k,a=a): return cos(a*k) def kernel_imag(k,a=a): return sin(a*k) omega_real = convolve.init_convolution_kernel(n,kernel_real,d=0, zero_nyquist=0) omega_imag = convolve.init_convolution_kernel(n,kernel_imag,d=1, zero_nyquist=0) _cache[(n,a)] = omega_real,omega_imag else: omega_real,omega_imag = omega overwrite_x = tmp is not x and not hasattr(x,'__array__') return convolve.convolve_z(tmp,omega_real,omega_imag, overwrite_x=overwrite_x) |
return abs(norm(loc=c,size=self._size)) | return abs(norm(mu=c,size=self._size)) | def _rvs(self, c): return abs(norm(loc=c,size=self._size)) |
return norm(loc=nc,size=self._size)*sqrt(df) / sqrt(chi2(df,size=self._size)) | return norm(mu=nc,size=self._size)*sqrt(df) / sqrt(chi2(df,size=self._size)) | def _rvs(self, df, nc): return norm(loc=nc,size=self._size)*sqrt(df) / sqrt(chi2(df,size=self._size)) |
if not atype[k] in "".join(Numeric.typecodes.values()): | if not atype[k] in typecodes: | def read_array(fileobject, separator=default, columns=default, comment="#", lines=default, atype=Numeric.Float, linesep='\n', rowsize=10000, missing=0): """Return an array or arrays from ascii_formatted data in |fileobject|. Inputs: fileobject -- An open file object or a string for a valid filename. The string can be prepended by "~/" or "~<name>/" to read a file from the home directory. separator -- a string or a tuple of strings to indicate the column separators. If the length of the string tuple is less than the total number of columns, then the last separator is assumed to be the separator for the rest of the columns. columns -- a tuple of integers and range-tuples which describe the columns to read from the file. A negative entry in the last column specifies the negative skip value to the end. Example: columns=(1, 4, (5, 9), (11, 15, 3), 17, -2) will read [1,4,5,6,7,8,11,14,17,19,21,23,...] If multiple arrays are to be returned, then this argument should be an ordered list of such tuples. There should be one entry in the list for each arraytype in the atype list. lines -- a tuple with the same structure as columns which indicates the lines to read. comment -- the comment character (line will be ignored even if it is specified by the lines tuple) linesep -- separator between rows. missing -- value to insert in array when conversion to number fails. atype -- the typecode of the output array. If multiple outputs are desired, then this should be a list of typecodes. The columns to fill the array represented by the given typecode is determined from the columns argument. If the length of atype does not match the length of the columns list, then, the smallest one is expanded to match the largest by repeatedly copying the last entry. rowsize -- the allocation row size (array grows by this amount as data is read in). Output -- the 1 or 2d array, or a tuple of output arrays of different types, sorted in order of the first column to be placed in the output array. """ # Make separator into a tuple of separators. if type(separator) in [types.StringType, type(default)]: sep = (separator,) else: sep = tuple(separator) # Create ascii_object from |fileobject| argument. ascii_object = ascii_stream(fileobject, lines=lines, comment=comment, linesep=linesep) columns, atype = convert_to_equal_lists(columns, atype) numout = len(atype) # Get the number of columns to read and expand the columns argument colsize, collist = getcolumns(ascii_object, columns, sep) # Intialize the output arrays outrange = range(numout) outarr = [] for k in outrange: if not atype[k] in "".join(Numeric.typecodes.values()): raise ValueError, "One of the array types is invalid, k=%d" % k outarr.append(Numeric.zeros((rowsize, colsize[k]),atype[k])) row = 0 block_row = 0 for line in ascii_object: if line.strip() == '': continue vals = process_line(line, sep, collist, atype, missing) for k in outrange: outarr[k][row] = vals[k] row += 1 block_row += 1 if block_row >= rowsize: for k in outrange: outarr[k].resize((outarr[k].shape[0] + rowsize,colsize)) block_row = 0 for k in outrange: if outarr[k].shape[0] != row: outarr[k].resize((row,colsize[k])) a = outarr[k] if a.shape[0] == 1 or a.shape[1] == 1: outarr[k] = Numeric.ravel(a) if len(outarr) == 1: return outarr[0] else: return tuple(outarr) |
outarr[k].resize((outarr[k].shape[0] + rowsize,colsize)) | outarr[k].resize((outarr[k].shape[0] + rowsize,colsize[k])) | def read_array(fileobject, separator=default, columns=default, comment="#", lines=default, atype=Numeric.Float, linesep='\n', rowsize=10000, missing=0): """Return an array or arrays from ascii_formatted data in |fileobject|. Inputs: fileobject -- An open file object or a string for a valid filename. The string can be prepended by "~/" or "~<name>/" to read a file from the home directory. separator -- a string or a tuple of strings to indicate the column separators. If the length of the string tuple is less than the total number of columns, then the last separator is assumed to be the separator for the rest of the columns. columns -- a tuple of integers and range-tuples which describe the columns to read from the file. A negative entry in the last column specifies the negative skip value to the end. Example: columns=(1, 4, (5, 9), (11, 15, 3), 17, -2) will read [1,4,5,6,7,8,11,14,17,19,21,23,...] If multiple arrays are to be returned, then this argument should be an ordered list of such tuples. There should be one entry in the list for each arraytype in the atype list. lines -- a tuple with the same structure as columns which indicates the lines to read. comment -- the comment character (line will be ignored even if it is specified by the lines tuple) linesep -- separator between rows. missing -- value to insert in array when conversion to number fails. atype -- the typecode of the output array. If multiple outputs are desired, then this should be a list of typecodes. The columns to fill the array represented by the given typecode is determined from the columns argument. If the length of atype does not match the length of the columns list, then, the smallest one is expanded to match the largest by repeatedly copying the last entry. rowsize -- the allocation row size (array grows by this amount as data is read in). Output -- the 1 or 2d array, or a tuple of output arrays of different types, sorted in order of the first column to be placed in the output array. """ # Make separator into a tuple of separators. if type(separator) in [types.StringType, type(default)]: sep = (separator,) else: sep = tuple(separator) # Create ascii_object from |fileobject| argument. ascii_object = ascii_stream(fileobject, lines=lines, comment=comment, linesep=linesep) columns, atype = convert_to_equal_lists(columns, atype) numout = len(atype) # Get the number of columns to read and expand the columns argument colsize, collist = getcolumns(ascii_object, columns, sep) # Intialize the output arrays outrange = range(numout) outarr = [] for k in outrange: if not atype[k] in "".join(Numeric.typecodes.values()): raise ValueError, "One of the array types is invalid, k=%d" % k outarr.append(Numeric.zeros((rowsize, colsize[k]),atype[k])) row = 0 block_row = 0 for line in ascii_object: if line.strip() == '': continue vals = process_line(line, sep, collist, atype, missing) for k in outrange: outarr[k][row] = vals[k] row += 1 block_row += 1 if block_row >= rowsize: for k in outrange: outarr[k].resize((outarr[k].shape[0] + rowsize,colsize)) block_row = 0 for k in outrange: if outarr[k].shape[0] != row: outarr[k].resize((row,colsize[k])) a = outarr[k] if a.shape[0] == 1 or a.shape[1] == 1: outarr[k] = Numeric.ravel(a) if len(outarr) == 1: return outarr[0] else: return tuple(outarr) |
fft_opt_info = get_info('fft_opt') | djbfft_info = {} mkl_info = get_info('mkl') if mkl_info: mkl_info.setdefault('define_macros', []).append(('SCIPY_MKL_H', None)) fft_opt_info = mkl_info else: fft_opt_info = get_info('fftw3') or get_info('fftw2') \ or get_info('dfftw') djbfft_info = get_info('djbfft') | def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration from numpy.distutils.system_info import get_info config = Configuration('fftpack',parent_package, top_path) fft_opt_info = get_info('fft_opt') config.add_data_dir('tests') config.add_library('dfftpack', sources=[join('dfftpack','*.f')]) sources = ['fftpack.pyf','src/zfft.c','src/drfft.c','src/zrfft.c', 'src/zfftnd.c'] config.add_extension('_fftpack', sources=sources, libraries=['dfftpack'], extra_info = fft_opt_info ) config.add_extension('convolve', sources = ['convolve.pyf','src/convolve.c'], libraries = ['dfftpack'], extra_info = fft_opt_info ) return config |
extra_info = fft_opt_info | extra_info = [fft_opt_info, djbfft_info], | def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration from numpy.distutils.system_info import get_info config = Configuration('fftpack',parent_package, top_path) fft_opt_info = get_info('fft_opt') config.add_data_dir('tests') config.add_library('dfftpack', sources=[join('dfftpack','*.f')]) sources = ['fftpack.pyf','src/zfft.c','src/drfft.c','src/zrfft.c', 'src/zfftnd.c'] config.add_extension('_fftpack', sources=sources, libraries=['dfftpack'], extra_info = fft_opt_info ) config.add_extension('convolve', sources = ['convolve.pyf','src/convolve.c'], libraries = ['dfftpack'], extra_info = fft_opt_info ) return config |
raise "Couldn't load level" | raise ("Couldn't load level: ", filename) | def __init__(self, *params): if len(params) == 2: (width, height) = params self.name = "No Name" self.author = "No Author" |
print "Got: ", type, x, y | def find(lst, obj): for i in lst: if i[0] == obj: return i return None |
|
ObjectBrush(make_sprite(config.datadir + object[1]), make_metadata(object[0])).add_to_layer(self.objects, CL_Point(x, y)) | self.objects.add_object(ObjMapSpriteObject(make_sprite(config.datadir + object[1]), CL_Point(x, y), make_metadata(BadGuy(object[0]))).to_object()) | def find(lst, obj): for i in lst: if i[0] == obj: return i return None |
for (obj, data) in []: pos = obj.get_pos() f.write(" (%s (x %d) (y %d))" % (data, pos.x, pos.y)) | for obj in self.objects.get_objects(): badguy = get_python_object(obj.get_metadata()) pos = obj.get_pos() f.write(" (%s (x %d) (y %d))\n" % (badguy.type, pos.x, pos.y)) | def save(self, filename): f = file(filename, 'w') f.write(";; Generated by Flexlay Editor\n" "(supertux-level\n") f.write(" (version 1)\n") f.write(" (name \"%s\")\n" % self.name) f.write(" (author \"%s\")\n" % self.author) f.write(" (width %d)\n" % self.width) f.write(" (height %d)\n" % self.height) |
make_metadata(object[0]))) | make_metadata(BadGuy(object[0])))) | def __init__(self, tileset, gui): self.selector_window = Panel(CL_Rect(CL_Point(800-134, 23+33), CL_Size(128 + 6, 558)), gui.get_component()) self.tileselector = TileSelector(CL_Rect(CL_Point(3, 3), CL_Size(128, 552)), self.selector_window) self.tileselector.set_tileset(tileset) self.tileselector.set_tiles(range(1,100)) self.tileselector.show(False) |
self.foreground = TilemapLayer(tileset, self.width, self.height) | self.foreground = TilemapLayer(supertux_tileset, self.width, self.height) | def __init__(self, filename): print "SuperTuxLevel:__init__" self.me = self |
self.interactive = TilemapLayer(tileset, self.width, self.height) | self.interactive = TilemapLayer(supertux_tileset, self.width, self.height) | def __init__(self, filename): print "SuperTuxLevel:__init__" self.me = self |
self.background = TilemapLayer(tileset, self.width, self.height) | self.background = TilemapLayer(supertux_tileset, self.width, self.height) | def __init__(self, filename): print "SuperTuxLevel:__init__" self.me = self |
workspace.set_current_map(self.editormap) | workspace.set_map(self.editormap) | def activate(self, workspace): #editor_tilemap_set_current(self.interactive.to_layer()) workspace.set_current_map(self.editormap) |
window = CL_Window(CL_Rect(50, 50, 350, 300), "My Window", gui.get_component()) | def __init__(self): self.tileselector_window = CL_Window(CL_Rect(CL_Point(150, 150), CL_Size(210, 210)), "Tile Selector", gui.get_component()) self.tileselector = TileSelector(5, 3, self.tileselector_window.get_client_area()) self.tileselector.set_tileset(tileset) self.tileselector.set_tiles(range(1,100)) |
|
editor = Editor() gui = editor.get_gui_manager() editor_map = EditorMapComponent(CL_Rect(0, 0, 799, 599), gui.get_component()) workspace = Workspace(799, 599) editor_map.set_workspace(workspace) m = EditorMap() workspace.set_current_map(m) | def activate(self, workspace): #editor_tilemap_set_current(self.interactive.to_layer()) workspace.set_current_map(self.editormap) |
|
tilemap = TilemapLayer(tileset, 20, 10) m.add_layer(tilemap.to_layer()) tile = Tile("/home/ingo/cvs/supertux/supertux/data/images/tilesets/bonus1.png", CL_Color(255, 255, 255, 255), CL_Color(255, 0, 0, 128)) tileset.add_tile(0, tile) tileset.add_tile(1, tile) tileset.add_tile(2, tile) | def activate(self, workspace): #editor_tilemap_set_current(self.interactive.to_layer()) workspace.set_current_map(self.editormap) |
|
TilemapLayer_set_current(tilemap) tilemap_paint_tool_set_tilemap(tilemap) editor_set_brush_tile(1) def foo(): print "---My Callback---" gui.quit() g = None def draw_something(): print "Draw something" brush = TileBrush(2, 2) brush.set_opaque() _ = PaintCommand(tilemap, brush) _.add_point(CL_Point(1,1)) _.add_point(CL_Point(2,2)) _.add_point(CL_Point(3,3)) _.add_point(CL_Point(4,4)) _.execute() g = _ print "Draw something done" window = CL_Window(CL_Rect(50, 50, 350, 300), "My Window", gui.get_component()) gui.push_component(window) button1 = CL_Button(CL_Rect(50, 50, 200, 75), "Quit", gui.get_component()) connect(button1.sig_clicked(), foo) button2 = CL_Button(CL_Rect(CL_Point(50, 100), CL_Size(150, 25)), "Draw", gui.get_component()) connect(button2.sig_clicked(), draw_something) def get_data(): print tilemap.get_data() def set_data(): tilemap.set_data((0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0)) button3 = CL_Button(CL_Rect(CL_Point(50, 150), CL_Size(150, 25)), "Get Data", gui.get_component()) connect(button3.sig_clicked(), get_data) button4 = CL_Button(CL_Rect(CL_Point(50, 180), CL_Size(150, 25)), "Get Data", gui.get_component()) connect(button4.sig_clicked(), set_data) gui.pop_component() tileselectorw = CL_Window(CL_Rect(CL_Point(150, 150), CL_Size(210, 210)), "Tile Selector", gui.get_component()) tileselector = TileSelector(5, 3, tileselectorw.get_client_area()) tileselector.set_tileset(tileset) tileselector.set_tiles(range(1,100)) class Menu(CL_Menu): def __init__(self): CL_Menu.__init__(self, gui.get_component()) def add_item(self, name, func): item = self.create_item(name) connect(item.sig_clicked(), func) level = None def menu_file_open(): print "File/Open" level = SuperTuxLevel('/home/ingo/cvs/supertux/supertux/data/levels/world1/level2.stl') print "Loading done" level.activate(workspace) print "Activation done" def menu_file_save(): print "File/Save" def menu_file_save_as(): print "File/Save As" menu = Menu() a = menu.add_item("File/Open...", menu_file_open) a = menu.add_item("File/Save...", menu_file_save) a = menu.add_item("File/Save As...", menu_file_save_as) gui.run() | def do_editor(): editor = Editor() gui = editor.get_gui_manager() editor_map = EditorMapComponent(CL_Rect(0, 0, 799, 599), gui.get_component()) workspace = Workspace(799, 599) editor_map.set_workspace(workspace) m = EditorMap() workspace.set_current_map(m) tilemap = TilemapLayer(tileset, 20, 10) m.add_layer(tilemap.to_layer()) TilemapLayer_set_current(tilemap) tilemap_paint_tool_set_tilemap(tilemap) editor_set_brush_tile(1) def foo(): print "---My Callback---" gui.quit() g = None def draw_something(): print "Draw something" brush = TileBrush(2, 2) brush.set_opaque() _ = PaintCommand(tilemap, brush) _.add_point(CL_Point(1,1)) _.add_point(CL_Point(2,2)) _.add_point(CL_Point(3,3)) _.add_point(CL_Point(4,4)) _.execute() g = _ print "Draw something done" window = CL_Window(CL_Rect(50, 50, 350, 300), "My Window", gui.get_component()) gui.push_component(window) button1 = CL_Button(CL_Rect(50, 50, 200, 75), "Quit", gui.get_component()) connect(button1.sig_clicked(), foo) button2 = CL_Button(CL_Rect(CL_Point(50, 100), CL_Size(150, 25)), "Draw", gui.get_component()) connect(button2.sig_clicked(), draw_something) def get_data(): print tilemap.get_data() def set_data(): tilemap.set_data((0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0)) button3 = CL_Button(CL_Rect(CL_Point(50, 150), CL_Size(150, 25)), "Get Data", gui.get_component()) connect(button3.sig_clicked(), get_data) button4 = CL_Button(CL_Rect(CL_Point(50, 180), CL_Size(150, 25)), "Get Data", gui.get_component()) connect(button4.sig_clicked(), set_data) gui.pop_component() tileselectorw = CL_Window(CL_Rect(CL_Point(150, 150), CL_Size(210, 210)), "Tile Selector", gui.get_component()) tileselector = TileSelector(5, 3, tileselectorw.get_client_area()) tileselector.set_tileset(tileset) tileselector.set_tiles(range(1,100)) class Menu(CL_Menu): def __init__(self): CL_Menu.__init__(self, gui.get_component()) def add_item(self, name, func): item = self.create_item(name) connect(item.sig_clicked(), func) level = None def menu_file_open(): print "File/Open" level = SuperTuxLevel('/home/ingo/cvs/supertux/supertux/data/levels/world1/level2.stl') print "Loading done" level.activate(workspace) print "Activation done" def menu_file_save(): print "File/Save" def menu_file_save_as(): print "File/Save As" menu = Menu() a = menu.add_item("File/Open...", menu_file_open) a = menu.add_item("File/Save...", menu_file_save) a = menu.add_item("File/Save As...", menu_file_save_as) print "- entering main loop" gui.run() del button1 del button2 del button3 del button4 del window print "- exit main loop" do_editor() | def activate(self, workspace): #editor_tilemap_set_current(self.interactive.to_layer()) workspace.set_current_map(self.editormap) |
self.author = get_value_from_tree(["name", "_"], data, "no author") | self.author = get_value_from_tree(["author", "_"], data, "no author") | def __init__(self, *params): if len(params) == 2: (width, height) = params self.name = "No Name" self.author = "No Author" |
f.write(" (%s (x %d) (y %d))\n" % (badguy.type, pos.x, pos.y)) | if (badguy.type != "resetpoint"): f.write(" (%s (x %d) (y %d))\n" % (badguy.type, pos.x, pos.y)) f.write(" )\n\n") f.write(" (reset-points\n") for obj in self.objects.get_objects(): badguy = get_python_object(obj.get_metadata()) pos = obj.get_pos() if (badguy.type == "resetpoint"): f.write(" (point (x %d) (y %d))\n" % (badguy.type, pos.x, pos.y)) | def save(self, filename): f = file(filename, 'w') f.write(";; Generated by Flexlay Editor\n" "(supertux-level\n") f.write(" (version 1)\n") f.write(" (name \"%s\")\n" % self.name) f.write(" (author \"%s\")\n" % self.author) f.write(" (width %d)\n" % self.width) f.write(" (height %d)\n" % self.height) |
PART_REGEXP = re.compile(r'^(?:\*\*\*|-->)\s.*(quit|left)') | PART_REGEXP = re.compile(r'^(?:\*\*\*|<--)\s.*(quit|left)') | def __repr__(self): return self.value |
for filename in sorted(glob.glob(os.path.join(logfile_path, '*.log'))): | files = glob.glob(os.path.join(logfile_path, '*.log')) files.sort() for filename in files: | def search_irc_logs(query, stats=None): if not stats: stats = SearchStats() # will be discarded, but, oh, well query = query.lower() for filename in sorted(glob.glob(os.path.join(logfile_path, '*.log'))): date = date_from_filename(filename) link = link_from_filename(filename) stats.files += 1 for time, event, info in LogParser(file(filename)): if event == LogParser.COMMENT: nick, text = info text = nick + ' ' + text elif event == LogParser.NICKCHANGE: text, oldnick, newnick = info else: text = str(info) stats.lines += 1 if query in text.lower(): stats.matches += 1 yield SearchResult(filename, link, date, time, event, info) |
self.navbar(prev, index, next) | self.navbar(self.prev, self.index, self.next) | def foot(self): print >> self.outfile, self.suffix self.navbar(prev, index, next) print >> self.outfile, """ |
nick_old = m.group(1) nick_new = m.group(2) | oldnick = m.group(1) newnick = m.group(2) | def __iter__(self): for line in self.infile: line = line.rstrip('\r\n') if not line: continue |
def nickchange(self, oldnick, newnick): | def change(self, oldnick, newnick): | def nickchange(self, oldnick, newnick): if oldnick in self.nick_colour: self.nick_colour[newnick] = self.nick_colour.pop(oldnick) |
'<a href="t | '<a href=" | def nicktext(self, time, nick, text, htmlcolour): """Output a comment uttered by someone. |
md.push(treeData) | md._push(treeData) | def tpRender(self, md, section, args): data=[] try: # see if we are being run as a sub-document root=md['tree-root-url'] url=md['tree-item-url'] state=md['tree-state'] or md['state'] substate=md['-tree-substate-'] colspan=md['tree-colspan'] level=md['tree-level'] except KeyError: # Check for collapse all, expand all, and state try: collapse_all=md['collapse_all'] except: collapse_all=None if collapse_all: state=[] else: try: expand_all=md['expand_all'] except: expand_all=None if expand_all: state=tpValuesIds(self) else: try: state=md['tree-state'] or md['state'] or md['-tree-state-'] if state[0] != '[': state=unquote(state) state=list(eval(state,{'__builtins__':{}})) except: state=[] colspan=1+tpStateLevel(state) level = 0 substate=state root=md['URL'] l=rfind(root, '/') if l >= 0: root=root[l+1:] url='' # Save state in a cookie if state: md['RESPONSE'].setCookie('tree-state',quote(str(state)[1:-1]+',')) else: md['RESPONSE'].expireCookie('tree-state') if substate==state: data.append('<TABLE CELLSPACING="0">\n') #level=0 treeData={'tree-root-url': root, 'tree-colspan': colspan, 'tree-state': state } md.push(treeData) try: for item in self.tpValues(): data=tpRenderTABLE(item,root,url,state,substate,data,colspan, section,md,treeData, level, args) if state is substate: data.append('</TABLE>\n') result=join(data,'') finally: md.pop(1) return result |
finally: md.pop(1) | finally: md._pop(1) | def tpRender(self, md, section, args): data=[] try: # see if we are being run as a sub-document root=md['tree-root-url'] url=md['tree-item-url'] state=md['tree-state'] or md['state'] substate=md['-tree-substate-'] colspan=md['tree-colspan'] level=md['tree-level'] except KeyError: # Check for collapse all, expand all, and state try: collapse_all=md['collapse_all'] except: collapse_all=None if collapse_all: state=[] else: try: expand_all=md['expand_all'] except: expand_all=None if expand_all: state=tpValuesIds(self) else: try: state=md['tree-state'] or md['state'] or md['-tree-state-'] if state[0] != '[': state=unquote(state) state=list(eval(state,{'__builtins__':{}})) except: state=[] colspan=1+tpStateLevel(state) level = 0 substate=state root=md['URL'] l=rfind(root, '/') if l >= 0: root=root[l+1:] url='' # Save state in a cookie if state: md['RESPONSE'].setCookie('tree-state',quote(str(state)[1:-1]+',')) else: md['RESPONSE'].expireCookie('tree-state') if substate==state: data.append('<TABLE CELLSPACING="0">\n') #level=0 treeData={'tree-root-url': root, 'tree-colspan': colspan, 'tree-state': state } md.push(treeData) try: for item in self.tpValues(): data=tpRenderTABLE(item,root,url,state,substate,data,colspan, section,md,treeData, level, args) if state is substate: data.append('</TABLE>\n') result=join(data,'') finally: md.pop(1) return result |
md.push(treeData) | md._push(treeData) | def tpRenderTABLE(self, root_url, url, state, substate, data, colspan, section, md, treeData, level=0, args=None): have_arg=args.has_key try: items=self.tpValues() except: items=None if not items and have_arg('leaves'): items=1 tpUrl=self.tpURL() url = (url and ('%s/%s' % (url, tpUrl))) or tpUrl treeData['tree-item-url']=url treeData['tree-level']=level treeData['tree-item-expanded']=0 try: id=self.tpId() except: id=None if id is None: try: id=self._p_oid except: id=None if id is None: id=pyid(self) exp=0 sub=None output=data.append # Add prefix output('<TR>\n') # Add +/- icon if items: if level: if level > 3: output( '<TD COLSPAN="%s"></TD>' % (level-1)) elif level > 1: output('<TD></TD>' * (level-1)) output('<TD WIDTH="16"></TD>\n') output('<TD WIDTH="16" VALIGN="TOP">') for i in range(len(substate)): sub=substate[i] if sub[0]==id: exp=i+1 break if exp: treeData['tree-item-expanded']=1 del substate[exp-1] output('<A HREF="%s?tree-state=%s">%s</A>' % (root_url,quote(str(state)[1:-1]+','), icoMinus)) substate.append(sub) else: substate.append([id]) output('<A HREF="%s?tree-state=%s">%s</A>' % (root_url,quote(str(state)[1:-1]+','), icoPlus)) del substate[-1] output('</TD>\n') else: level=level+1 if level > 3: output('<TD COLSPAN="%s"></TD>' % (level-1)) elif level > 1: output('<TD></TD>' * (level-1)) output('<TD WIDTH="16"></TD>\n') # add item text dataspan=colspan-level output('<TD COLSPAN="%s" VALIGN="TOP">' % dataspan) output(section(self, md)) output('</TD>\n</TR>\n') if exp: level=level+1 if level > 2: h='<TD COLSPAN="%s"></TD>' % (level-1) elif level > 1: h='<TD></TD>' * (level-1) else: h='' if have_arg('header'): if md.has_key(args['header']): output(md.getitem(args['header'],0)( self, md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) if items==1: # leaves treeData['-tree-substate-']=sub treeData['tree-level']=level md.push(treeData) output(md.getitem(args['leaves'],0)( self,md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) md.pop(1) elif have_arg('expand'): treeData['-tree-substate-']=sub treeData['tree-level']=level md.push(treeData) output(md.getitem(args['expand'],0)(self,md)) md.pop(1) else: __traceback_info__=sub, args, state, substate for item in items: if len(sub)==1: sub.append([]) data=tpRenderTABLE(item, root_url,url,state,sub[1],data, colspan, section, md, treeData, level, args) if not sub[1]: del sub[1] if have_arg('footer'): if md.has_key(args['footer']): output(md.getitem(args['footer'],0)( self, md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) return data |
md.pop(1) | md._pop(1) | def tpRenderTABLE(self, root_url, url, state, substate, data, colspan, section, md, treeData, level=0, args=None): have_arg=args.has_key try: items=self.tpValues() except: items=None if not items and have_arg('leaves'): items=1 tpUrl=self.tpURL() url = (url and ('%s/%s' % (url, tpUrl))) or tpUrl treeData['tree-item-url']=url treeData['tree-level']=level treeData['tree-item-expanded']=0 try: id=self.tpId() except: id=None if id is None: try: id=self._p_oid except: id=None if id is None: id=pyid(self) exp=0 sub=None output=data.append # Add prefix output('<TR>\n') # Add +/- icon if items: if level: if level > 3: output( '<TD COLSPAN="%s"></TD>' % (level-1)) elif level > 1: output('<TD></TD>' * (level-1)) output('<TD WIDTH="16"></TD>\n') output('<TD WIDTH="16" VALIGN="TOP">') for i in range(len(substate)): sub=substate[i] if sub[0]==id: exp=i+1 break if exp: treeData['tree-item-expanded']=1 del substate[exp-1] output('<A HREF="%s?tree-state=%s">%s</A>' % (root_url,quote(str(state)[1:-1]+','), icoMinus)) substate.append(sub) else: substate.append([id]) output('<A HREF="%s?tree-state=%s">%s</A>' % (root_url,quote(str(state)[1:-1]+','), icoPlus)) del substate[-1] output('</TD>\n') else: level=level+1 if level > 3: output('<TD COLSPAN="%s"></TD>' % (level-1)) elif level > 1: output('<TD></TD>' * (level-1)) output('<TD WIDTH="16"></TD>\n') # add item text dataspan=colspan-level output('<TD COLSPAN="%s" VALIGN="TOP">' % dataspan) output(section(self, md)) output('</TD>\n</TR>\n') if exp: level=level+1 if level > 2: h='<TD COLSPAN="%s"></TD>' % (level-1) elif level > 1: h='<TD></TD>' * (level-1) else: h='' if have_arg('header'): if md.has_key(args['header']): output(md.getitem(args['header'],0)( self, md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) if items==1: # leaves treeData['-tree-substate-']=sub treeData['tree-level']=level md.push(treeData) output(md.getitem(args['leaves'],0)( self,md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) md.pop(1) elif have_arg('expand'): treeData['-tree-substate-']=sub treeData['tree-level']=level md.push(treeData) output(md.getitem(args['expand'],0)(self,md)) md.pop(1) else: __traceback_info__=sub, args, state, substate for item in items: if len(sub)==1: sub.append([]) data=tpRenderTABLE(item, root_url,url,state,sub[1],data, colspan, section, md, treeData, level, args) if not sub[1]: del sub[1] if have_arg('footer'): if md.has_key(args['footer']): output(md.getitem(args['footer'],0)( self, md, standard_html_header=( '<TR>%s<TD WIDTH="16"></TD>' '<TD COLSPAN="%s" VALIGN="TOP">' % (h, dataspan-1)), standard_html_footer='</TD></TR>', )) return data |
i = rfind(name, ":") + 1 | i = name.rfind(":") + 1 | def attrAction(self, item): name, value, action = item[:3] if action == 1 or (action > 1 and not self.showtal): return 0, name, value macs = self.macroStack if action == 2 and self.metal and macs: if len(macs) > 1 or not macs[-1][2]: # Drop all METAL attributes at a use-depth above one. return 0, name, value # Clear 'entering' flag macs[-1][2] = 0 # Convert or drop depth-one METAL attributes. i = rfind(name, ":") + 1 prefix, suffix = name[:i], name[i:] if suffix == "define-macro": # Convert define-macro as we enter depth one. name = prefix + "use-macro" value = macs[-1][0] # Macro name elif suffix == "define-slot": name = prefix + "fill-slot" elif suffix == "fill-slot": pass else: return 0, name, value |
if self.html and lower(name) in BOOLEAN_HTML_ATTRS: | if self.html and name.lower() in BOOLEAN_HTML_ATTRS: | def attrAction_tal(self, item): name, value, action = item[:3] if action > 1: return self.attrAction(item) ok = 1 if self.html and lower(name) in BOOLEAN_HTML_ATTRS: evalue = self.engine.evaluateBoolean(item[3]) if evalue is self.Default: if action == 1: # Cancelled insert ok = 0 elif evalue: value = None else: ok = 0 else: evalue = self.engine.evaluateText(item[3]) if evalue is self.Default: if action == 1: # Cancelled insert ok = 0 else: if evalue is None: ok = 0 value = evalue if ok: if value is None: value = name value = "%s=%s" % (name, quote(value)) return ok, name, value |
i = rfind(s, '\n') | i = s.rfind('\n') | def do_insertText_tal(self, stuff): text = self.engine.evaluateText(stuff[0]) if text is None: return if text is self.Default: self.interpret(stuff[1]) return s = escape(text) self._stream_write(s) i = rfind(s, '\n') if i < 0: self.col = self.col + len(s) else: self.col = len(s) - (i + 1) |
return sys.modules[__name__].__path__[0] | m=sys.modules[__name__] if hasattr(m,'__path__'): return m.__path__[0] return sys.modules[__name__[:rfind(__name__,'.')]].__path__[0] | def package_home(globals_dict): __name__=globals_dict['__name__'] return sys.modules[__name__].__path__[0] |
self.notifyAdd(item) | def new(self, key): |
|
if self._addCallback: | callback = self._addCallback if type(callback) is type(''): callback = self.aq_parent.unrestrictedTraverse(callback) if callable(callback): | def notifyAdd(self, item): if self._addCallback: try: self._addCallback(item, self) # Use self as context except: pass # Eat all errors |
self._addCallback(item, self) | callback(item, self) | def notifyAdd(self, item): if self._addCallback: try: self._addCallback(item, self) # Use self as context except: pass # Eat all errors |
if self._delCallback: | callback = self._delCallback if type(callback) is type(''): callback = self.aq_parent.unrestrictedTraverse(callback) if callable(callback): | def notifyDestruct(self, item): if self._delCallback: try: self._delCallback(item, self) # Use self as context except: pass # Eat all errors |
self._delCallback(item, self) | callback(item, self) | def notifyDestruct(self, item): if self._delCallback: try: self._delCallback(item, self) # Use self as context except: pass # Eat all errors |
self.notifyDestruct(index[k]) | self.notifyDestruct(index[k][k]) | def _clean(self, b, index): |
cmdline = ( '%s -c "from Zope import configure;' | cmdline = ( '%s -c "from Zope2 import configure;' | def get_startup_cmd(self, python, more): cmdline = ( '%s -c "from Zope import configure;' 'configure(\'%s\');' % (python, self.options.configfile) ) return cmdline + more + '\"' |
$Id: Publish.py,v 1.36 1997/03/20 22:31:46 jim Exp $""" | $Id: Publish.py,v 1.37 1997/03/26 19:05:56 jim Exp $""" | def taste(spam): "a favorable reviewer" return spam,'yum yum, I like ' + spam |
__version__='$Revision: 1.36 $'[11:-2] | __version__='$Revision: 1.37 $'[11:-2] | def taste(spam): "a favorable reviewer" return spam,'yum yum, I like ' + spam |
query['self']=parents[0] | selfarg=parents[0] for i in range(len(parents)): try: p=parents[i].aq_self parents[i]=p except: pass | def function_with_empty_signature(): pass |
if name_index < nrequired: | if argument_name=='self': args.append(selfarg) elif name_index < nrequired: | def function_with_empty_signature(): pass |
startk, endk = r.span(2) starth, endh = r.span(3) | starth, endh = r.span(2) | def doc_img( self, s, expr1=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+)').search, expr2=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+):([a-zA-Z0-9\-.:/;,\n\~]+)').search ): |
return (StructuredTextImage(s[startt:endt], href=s[starth:endh], key=s[startk:endk]), | return (StructuredTextImage(s[startt:endt], href=s[starth:endh]), | def doc_img( self, s, expr1=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+)').search, expr2=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+):([a-zA-Z0-9\-.:/;,\n\~]+)').search ): |
else: r=expr1(s) | return None | def doc_img( self, s, expr1=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+)').search, expr2=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+):([a-zA-Z0-9\-.:/;,\n\~]+)').search ): |
if r: startt, endt = r.span(1) starth, endh = r.span(2) start, end = r.span() return (StructuredTextImage(s[startt:endt], href=s[starth:endh]), start, end) return None | def doc_img( self, s, expr1=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+)').search, expr2=re.compile('\"([ _a-zA-Z0-9*.:/;,\-\n\~]+)\":img:([a-zA-Z0-9\-.:/;,\n\~]+):([a-zA-Z0-9\-.:/;,\n\~]+)').search ): |
|
try: src=self(REQUEST, src__=1) if find(src,'\0'): src=join(split(src,'\0'),'\n'+'-'*60+'\n') result=self(REQUEST) if result._searchable_result_columns(): r=custom_default_report(self.id, result) else: r='This was not a query.' except: r=( '<strong>Error, <em>%s</em>:</strong> %s' % (sys.exc_type, sys.exc_value)) report=DocumentTemplate.HTML( '<html><BODY BGCOLOR=" '<!-- '<hr><strong>SQL used:</strong><br>\n<pre>\n%s\n</pre>\n<hr>\n' '</body></html>' % (r,src)) return apply(report,(self,REQUEST),{self.id:result}) | t=v=tb=None try: try: src=self(REQUEST, src__=1) if find(src,'\0'): src=join(split(src,'\0'),'\n'+'-'*60+'\n') result=self(REQUEST) if result._searchable_result_columns(): r=custom_default_report(self.id, result) else: r='This was not a query.' except: t, v, tb = sys.exc_info() r='<strong>Error, <em>%s</em>:</strong> %s' % (t, v) report=DocumentTemplate.HTML( '<html><BODY BGCOLOR=" '<!-- '<hr><strong>SQL used:</strong><br>\n<pre>\n%s\n</pre>\n<hr>\n' '</body></html>' % (r,src)) report=apply(report,(self,REQUEST),{self.id:result}) if tb is not None: self.raise_standardErrorMessage( None, REQUEST, t, v, tb, None, report) return report finally: tb=None | def manage_test(self, REQUEST): |
self.assertEqual(starter.startup_handler.level, logging.DEBUG) | self.assertEqual(starter.startup_handler.level, 15) | def testSetupStartupHandler(self): import zLOG import sys conf = self.load_config_text(""" instancehome <<INSTANCE_HOME>> debug-mode on <eventlog> level info <logfile> path <<INSTANCE_HOME>>/event.log level info </logfile> </eventlog>""") starter = ZopeStarter(conf) starter.setupStartupHandler() self.assert_(not zLOG._call_initialize) self.assertEqual(starter.startup_handler.formatter, zLOG.EventLogger.formatters['file']) self.assertEqual(starter.startup_handler.level, logging.DEBUG) self.assertEqual(starter.startup_handler, zLOG.EventLogger.EventLogger.logger.handlers[0]) self.assertEqual(len(zLOG.EventLogger.EventLogger.logger.handlers), 1) self.assertEqual(zLOG.EventLogger.EventLogger.logger.level, logging.DEBUG) self.assertEqual(starter.startup_handler.level, logging.DEBUG) self.failUnlessEqual(starter.startup_handler.stream, sys.stderr) conf = self.load_config_text(""" instancehome <<INSTANCE_HOME>> debug-mode off <eventlog> level info <logfile> path <<INSTANCE_HOME>>/event.log level info </logfile> </eventlog>""") starter = ZopeStarter(conf) starter.setupStartupHandler() self.failIfEqual(starter.startup_handler.stream, sys.stderr) |
self.assertEqual(zLOG.EventLogger.EventLogger.logger.level, logging.DEBUG) self.assertEqual(starter.startup_handler.level, logging.DEBUG) | def testSetupStartupHandler(self): import zLOG import sys conf = self.load_config_text(""" instancehome <<INSTANCE_HOME>> debug-mode on <eventlog> level info <logfile> path <<INSTANCE_HOME>>/event.log level info </logfile> </eventlog>""") starter = ZopeStarter(conf) starter.setupStartupHandler() self.assert_(not zLOG._call_initialize) self.assertEqual(starter.startup_handler.formatter, zLOG.EventLogger.formatters['file']) self.assertEqual(starter.startup_handler.level, logging.DEBUG) self.assertEqual(starter.startup_handler, zLOG.EventLogger.EventLogger.logger.handlers[0]) self.assertEqual(len(zLOG.EventLogger.EventLogger.logger.handlers), 1) self.assertEqual(zLOG.EventLogger.EventLogger.logger.level, logging.DEBUG) self.assertEqual(starter.startup_handler.level, logging.DEBUG) self.failUnlessEqual(starter.startup_handler.stream, sys.stderr) conf = self.load_config_text(""" instancehome <<INSTANCE_HOME>> debug-mode off <eventlog> level info <logfile> path <<INSTANCE_HOME>>/event.log level info </logfile> </eventlog>""") starter = ZopeStarter(conf) starter.setupStartupHandler() self.failIfEqual(starter.startup_handler.stream, sys.stderr) |
|
print id(A) print A | def __str__(self): context=getattr(self, 'aq_parent', None) if context is None: return self.__name__ return "(%s: %s of %s)" % (uid(self), self.aq_self, context) |
|
class ExternalMethod(OFS.SimpleItem.Item, Persistent, Explicit, | class ExternalMethod(OFS.SimpleItem.Item, Persistent, Acquisition.Explicit, | def manage_addExternalMethod(self, id, title, module, function, REQUEST=None): """Add an external method to a folder Un addition to the standard Principia object-creation arguments, 'id' and title, the following arguments are defined: function -- The name of the python function. This can be a an ordinary Python function, or a bound method. module -- The name of the file containing the function definition. The module normally resides in the 'Extensions' directory, however, the file name may have a prefix of 'product.', indicating that it should be found in a product directory. For example, if the module is: 'ACMEWidgets.foo', then an attempt will first be made to use the file 'lib/python/Products/ACMEWidgets/Extensions/foo.py'. If this failes, then the file 'Extensions/ACMEWidgets.foo.py' will be used. """ i=ExternalMethod(id,title,module,function) self._setObject(id,i) return self.manage_main(self,REQUEST) |
def isFolderish(self): return self.getClassAttr('isPrincipiaFolderish', 0) | def isFolderish(self): return self.getClassAttr('isPrincipiaFolderish', 0, 1) | def isFolderish(self): return self.getClassAttr('isPrincipiaFolderish', 0) |
index = base(name, self) | index = base(name) | def addIndex(self, name, type,extra=None): # Convert the type by finding an appropriate product which supports # this interface by that name. Bleah |
if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() | self._check_for_update() | def index_html(self, REQUEST=None): "View the Help Topic" if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() |
if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() | self._check_for_update() | def index_html(self, REQUEST=None): """ View the STX Help Topic """ if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() |
if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() | self._check_for_update() | def index_html(self, REQUEST, RESPONSE): "View the Help Topic" if 0 and Globals.DevelopmentMode: try: mtime=os.stat(self.file) except: mtime=0 if mtime != self._v_last_read: self.obj = open(self.file).read() self._v_last_read=mtime self.reindex_object() |
self.__app._manage_add_product_meta_type( | app._manage_add_product_meta_type( | def registerClass(self, instance_class=None, meta_type='', permission=None, constructors=(), icon=None, permissions=None, legacy=(), ): """Register a constructor |
meta_type='', CreateFactory=1, REQUEST=None): | meta_type='', CreateFactory=0, REQUEST=None): | def manage_addZClass(self, id, title='', baseclasses=[], meta_type='', CreateFactory=1, REQUEST=None): """Add a Z Class """ bases=[] for b in baseclasses: if Products.meta_classes.has_key(b): bases.append(Products.meta_classes[b]) else: bases.append(getattr(self, b)) Z=ZClass(id,title,bases) if meta_type: Z._zclass_.meta_type=meta_type self._setObject(id, Z) if CreateFactory and meta_type: self.manage_addDTMLMethod( id+'_addForm', id+' constructor input form', addFormDefault % {'id': id, 'meta_type': meta_type}, ) self.manage_addDTMLMethod( id+'_add', id+' constructor', addDefault % {'id': id}, ) self.manage_addPrincipiaFactory( id+'_factory', id+' factory', meta_type, id+'_addForm', ) if REQUEST is not None: return self.manage_main(self,REQUEST) |
bases.append(getattr(self, b)) | bases.append(self._getOb(b)) | def manage_addZClass(self, id, title='', baseclasses=[], meta_type='', CreateFactory=1, REQUEST=None): """Add a Z Class """ bases=[] for b in baseclasses: if Products.meta_classes.has_key(b): bases.append(Products.meta_classes[b]) else: bases.append(getattr(self, b)) Z=ZClass(id,title,bases) if meta_type: Z._zclass_.meta_type=meta_type self._setObject(id, Z) if CreateFactory and meta_type: self.manage_addDTMLMethod( id+'_addForm', id+' constructor input form', addFormDefault % {'id': id, 'meta_type': meta_type}, ) self.manage_addDTMLMethod( id+'_add', id+' constructor', addDefault % {'id': id}, ) self.manage_addPrincipiaFactory( id+'_factory', id+' factory', meta_type, id+'_addForm', ) if REQUEST is not None: return self.manage_main(self,REQUEST) |
d=parse_cookie(self.environ['HTTP_COOKIE']) for k,item in d.items(): cookies[k]=item if not form.has_key(k): form[k]=item | parse_cookie(self.environ['HTTP_COOKIE'],cookies) for k,item in cookies.items(): if not other.has_key(k): other[k]=item | def __init__(self, |
if hasattr(self, 'getSize'): RESPONSE.setHeader('Content-Length', absattr(self.getSize)) | else: url=urlfix(REQUEST['URL'], 'HEAD') name=filter(None, string.split(url, '/'))[-1] ct, ce=mimetypes.guess_type(name) ct=ct or 'application/octet-stream' ct=string.lower(ct) RESPONSE.setHeader('Content-Type', ct) if hasattr(aq_base(self), 'get_size'): RESPONSE.setHeader('Content-Length', absattr(self.get_size)) if hasattr(self, '_p_mtime'): mtime=rfc1123_date(self._p_mtime) RESPONSE.setHeader('Last-Modified', mtime) | def HEAD(self, REQUEST, RESPONSE): """Retrieve resource information without a response body.""" self.dav__init(REQUEST, RESPONSE) if hasattr(self, 'content_type'): RESPONSE.setHeader('Content-Type', absattr(self.content_type)) if hasattr(self, 'getSize'): RESPONSE.setHeader('Content-Length', absattr(self.getSize)) RESPONSE.setStatus(200) return RESPONSE |
product=self.aq_acquire('_getProducts')()._product(name) dispatcher=FactoryDispatcher(product, self.aq_parent) return dispatcher.__of__(self) | return self.__bobo_traverse__(None, name) | def __getitem__(self, name): product=self.aq_acquire('_getProducts')()._product(name) dispatcher=FactoryDispatcher(product, self.aq_parent) return dispatcher.__of__(self) |
dispatcher=FactoryDispatcher(product, self.aq_parent, REQUEST) | dispatcher_class=getattr( getattr(Products, name, None), '__FactoryDispatcher__', FactoryDispatcher) dispatcher=dispatcher_class(product, self.aq_parent, REQUEST) | def __bobo_traverse__(self, REQUEST, name): product=self.aq_acquire('_getProducts')()._product(name) dispatcher=FactoryDispatcher(product, self.aq_parent, REQUEST) return dispatcher.__of__(self) |
" " | def __bobo_traverse__(self, REQUEST, name): product=self.aq_acquire('_getProducts')()._product(name) dispatcher=FactoryDispatcher(product, self.aq_parent, REQUEST) return dispatcher.__of__(self) |
|
try out the query, <a href="%s">click hear</a>. | try out the query, <a href="%s">click here</a>. | def manage_addZSearch(self, report_id, report_title, report_style, input_id, input_title, queries=[], REQUEST=None): 'add a report' if not queries: raise ValueError, ( 'No <em>searchable objects</em> were selected') if not report_id: raise ValueError, ( 'No <em>report id</em> were specified') if input_title and not input_id: raise ValueError, ( 'No <em>input id</em> were specified') qs=map(lambda q, self=self: _getquery(self, q), queries) arguments={} keys=[] for q in qs: id=q.id if input_id: for name, arg in q._searchable_arguments().items(): if len(qs) > 1: key="%s/%s" % (id,name) else: key=name arguments[key]=arg keys.append(key) if q._searchable_result_columns() is None: raise 'Unusable Searchable Error',( """The input searchable object, <em>%s</em>, has not been tested. Until it has been tested, it\'s output schema is unknown, and a report cannot be generated. Before creating a report from this query, you must try out the query. To try out the query, <a href="%s">click hear</a>. """ % (q.title_and_id(), id)) if input_id: arguments=Args(arguments, keys) self.manage_addDocument( input_id,input_title, default_input_form(arguments, report_id)) self.manage_addDocument( report_id,report_title, ('<!--#var standard_html_header-->\n%s\n' '<!--#var standard_html_footer-->' % join(map(lambda q, report_style=report_style: custom_default_report(q.id, q, no_table=report_style), qs), '\n<hr>\n'))) if REQUEST: return self.manage_main(self,REQUEST) |
self.munge(file.read()) | if type(file) is not type(''): file=file.read() self.munge(file) | def manage_upload(self,file='', REQUEST=None): """Replace the contents of the document with the text in file.""" self._validateProxy(REQUEST) self.munge(file.read()) if REQUEST: return MessageDialog( title ='Success!', message='Your changes have been saved', action ='manage_main') |
ms=string.split(ms, ';')[0] | ms=split(ms, ';')[0] | def _init_headers(self, request, response): # Attempt to handle If-Modified-Since headers. ms=request.get_header('If-Modified-Since', None) if ms is not None: ms=string.split(ms, ';')[0] ms=DateTime(ms).timeTime() if self.lmt > ms: RESPONSE.setStatus(304) return RESPONSE response.setHeader('Content-Type', self.content_type) response.setHeader('Last-Modified', self.lmh) |
RESPONSE.setHeader('Content-type','application/data') RESPONSE.setHeader('Content-Disposition', 'inline;filename=%s.%s' % (id, suffix)) | if RESPONSE is not None: RESPONSE.setHeader('Content-type','application/data') RESPONSE.setHeader('Content-Disposition', 'inline;filename=%s.%s' % (id, suffix)) | def manage_exportObject(self, id='', download=None, toxml=None, RESPONSE=None): """Exports an object to a file and returns that file.""" if not id: # can't use getId() here (breaks on "old" exported objects) id=self.id if hasattr(id, 'im_func'): id=id() ob=self else: ob=self._getOb(id) |
raise 'Bad Request', 'Invalid file name %s' % file | raise BadRequestException, 'Invalid file name %s' % file | def manage_importObject(self, file, REQUEST=None, set_owner=1): """Import an object from a file""" dirname, file=os.path.split(file) if dirname: raise 'Bad Request', 'Invalid file name %s' % file |
raise 'Bad Request', 'File does not exist: %s' % file | raise BadRequestException, 'File does not exist: %s' % file | def manage_importObject(self, file, REQUEST=None, set_owner=1): """Import an object from a file""" dirname, file=os.path.split(file) if dirname: raise 'Bad Request', 'Invalid file name %s' % file |
RESPONSE.redirect(SCRIPT_NAME+'HelpSys/hs_index') | RESPONSE.redirect(SCRIPT_NAME+'/HelpSys/hs_index') | def manage_help(self, RESPONSE, SCRIPT_NAME): "Help!" RESPONSE.redirect(SCRIPT_NAME+'HelpSys/hs_index') return '' |
help_option_=({'label': 'Help', 'action': 'manage_help'},), | help_option_=({'label': 'Help', 'action': 'manage_help', },), | def filtered_manage_options( self, REQUEST=None, help_option_=({'label': 'Help', 'action': 'manage_help'},), ): if REQUEST is None: if hasattr(self, 'REQUEST'): REQUEST=self.REQUEST try: user=REQUEST['AUTHENTICATED_USER'] except: user=None result=[] seen_roles={} |
self._info = self._setupDB('info') | def _setupDBs(self): # Data Type Assumptions: # # - Object ids (oid) are 8-bytes # - Objects have revisions, with each revision being identified by a # unique serial number. We sometimes refer to 16-byte strings of # oid+serial as a revision id. # - Transaction ids (tid) are 8-bytes # - Version ids (vid) are 8-bytes # - Data pickles are of arbitrary length # # The Full storage uses the following tables: # # serials -- {oid -> [serial | serial+tid]} # Maps oids to serial numbers, to make it easy to look up the # serial number for the current revision of the object. The value # combined with the oid provides a revision id (revid) which is # used to point into the other tables. Usually the serial is the # tid of the transaction that modified the object, except in the # case of abortVersion(). Here, the serial number of the object # won't change (by definition), but of course the abortVersion() # happens in a new transaction so the tid pointer must change. To # handle this rare case, the value in the serials table can be a # 16-byte value, in which case it will contain both the serial # number and the tid pointer. # # metadata -- {oid+tid -> vid+nvrevid+lrevid+previd} # Maps object revisions to object metadata. This mapping is used # to find other information about a particular concrete object # revision. Essentially it stitches all the other pieces # together. The object revision is identified by the tid of the # transaction in which the object was modified. Normally this # will be the serial number (IOW, the serial number and tid will # be the same value), except in the case of abortVersion(). See # above for details. # # vid is the id of the version this object revision was modified # in. It will be zero if the object was modified in the # non-version. # # nvrevid is the tid pointing to the most current non-version # object revision. So, if the object is living in a version and # that version is aborted, the nvrevid points to the object # revision that will soon be restored. nvrevid will be zero if # the object was never modified in a version. # # lrevid is the tid pointing to object revision's pickle state (I # think of it as the "live revision id" since it's the state that # gives life to the object described by this metadata record). # # prevrevid is the tid pointing to the previous state of the # object. This is used for undo. # # txnMetadata -- {tid -> userlen+desclen+user+desc+ext} # Maps tids to metadata about a transaction. # # userlen is the length in characters of the `user' field as an # 8-byte unsigned long integer # desclen is the length in characters of the `desc' field as an # 8-byte unsigned long integer # user is the user information passed to tpc_finish() # desc is the description info passed to tpc_finish() # ext is the extra info passed to tpc_finish(). It is a # dictionary that we get already pickled by BaseStorage. # # pickles -- {oid+serial -> pickle} # Maps the object revisions to the revision's pickle data. # # refcounts -- {oid -> count} # Maps the oid to the reference count for the object. This # reference count is updated during the _finish() call. In the # Full storage the refcounts include all the revisions of the # object, so it is never decremented except at pack time. When it # goes to zero, the object is automatically deleted. # # txnoids -- {tid -> [oid]} # Maps transaction ids to the oids of the objects modified by the # transaction. # # pickleRefcounts -- {oid+tid -> count} # Maps an object revision to the reference count of that # revision's pickle. In the face of transactional undo, multiple # revisions can point to a single pickle so that pickle can't be # garbage collected until there are no revisions pointing to it. # # vids -- {version_string -> vid} # Maps version strings (which are arbitrary) to vids. # # versions -- {vid -> version_string} # Maps vids to version strings. # # currentVersions -- {vid -> [oid + tid]} # Maps vids to the revids of the objects modified in that version # for all current versions (except the 0th version, which is the # non-version). # # oids -- [oid] # This is a list of oids of objects that are modified in the # current uncommitted transaction. # # pvids -- [vid] # This is a list of all the version ids that have been created in # the current uncommitted transaction. # # prevrevids -- {oid -> tid} # This is a list of previous revision ids for objects which are # modified by transactionalUndo in the current uncommitted # transaction. It's necessary to properly handle multiple # transactionalUndo()'s in a single ZODB transaction. # # pending -- tid -> 'A' | 'C' # This is an optional flag which says what to do when the database # is recovering from a crash. The flag is normally 'A' which # means any pending data should be aborted. At the start of the # tpc_finish() this flag will be changed to 'C' which means, upon # recovery/restart, all pending data should be committed. Outside # of any transaction (e.g. before the tpc_begin()), there will be # no pending entry. It is a database invariant that if the # pending table is empty, the oids, pvids, and prevrevids tables # must also be empty. # # info -- {key -> value} # This table contains storage metadata information. The keys and # values are simple strings of variable length. Here are the # valid keys: # # packtime - time of the last pack. It is illegal to undo to # before the last pack time. # # version - the version of the database (reserved for ZODB4) # # objrevs -- {newserial+oid -> oldserial} # This table collects object revision information for packing # purposes. Every time a new object revision is committed, we # write an entry to this table. When we run pack, we iterate from # the start of this table until newserial > packtime, deleting old # revisions of objects. Note that when a new revision of an # object is first written to a version, no entry is written here. # We do write an entry when we commit or abort the version. # # delqueue -- [oid] # This is also a Queue, not a BTree. It is used during pack to # list objects for which no more references exist, such that the # objects can be completely packed away. # # packmark -- [oid] # Every object reachable from the root during a classic pack # operation will have its oid present in this table. # # oidqueue -- [oid] # This table is a Queue, not a BTree. It is used during the mark # phase of pack() and contains a list of oids for work to be done. # self._serials = self._setupDB('serials', db.DB_DUP) self._pickles = self._setupDB('pickles') self._refcounts = self._setupDB('refcounts') # Temporary tables which keep information during ZODB transactions self._oids = self._setupDB('oids') self._pvids = self._setupDB('pvids') self._prevrevids = self._setupDB('prevrevids') self._pending = self._setupDB('pending') # Other tables self._vids = self._setupDB('vids') self._versions = self._setupDB('versions') self._currentVersions = self._setupDB('currentVersions', db.DB_DUP) self._metadata = self._setupDB('metadata') self._txnMetadata = self._setupDB('txnMetadata') self._txnoids = self._setupDB('txnoids', db.DB_DUP) self._pickleRefcounts = self._setupDB('pickleRefcounts') # Tables to support packing. self._objrevs = self._setupDB('objrevs', db.DB_DUP) self._packmark = self._setupDB('packmark') self._info = self._setupDB('info') self._oidqueue = self._setupDB('oidqueue', 0, db.DB_QUEUE, 8) self._delqueue = self._setupDB('delqueue', 0, db.DB_QUEUE, 8) # Do recovery and consistency checks self._withlock(self._dorecovery) |
|
self._oids.truncate(txn) | def _docommit(self, txn, tid): self._pending.put(self._serial, COMMIT, txn) # Almost all the data's already written by now so we don't need to do # much more than update reference counts. Even there, our work is # easy because we're not going to decref anything here. deltas = {} co = cs = None try: co = self._oids.cursor(txn=txn) cs = self._serials.cursor(txn=txn) rec = co.first() while rec: oid = rec[0] rec = co.next() # Get the pointer to the live pickle data for this revision metadata = self._metadata[oid + self._serial] lrevid = unpack('>8s', metadata[16:24])[0] # Incref all objects referenced by this pickle, but watch out # for the George Bailey Event, which has no pickle. if lrevid <> DNE: revid = oid + lrevid data = self._pickles[revid] self._update(deltas, data, 1) # Incref this pickle; there's a new revision pointing to it refcount = self._pickleRefcounts.get(revid, ZERO, txn=txn) self._pickleRefcounts.put(revid, incr(refcount, 1), txn=txn) # Now delete all entries from the serials table where the # stored tid is not equal to the committing tid. srec = cs.set(oid) while srec: soid, data = srec if soid <> oid: break if len(data) == 8: stid = data else: # In the face of abortVersion, the first half is the # serial number and the second half is the tid. stid = data[8:] if stid <> tid: cs.delete() srec = cs.next_dup() finally: # There's a small window of opportunity for leaking a cursor here, # if co.close() were to fail. In practice this shouldn't happen. if co: co.close() if cs: cs.close() # Now incref all the object refcounts for oid, delta in deltas.items(): refcount = self._refcounts.get(oid, ZERO, txn=txn) self._refcounts.put(oid, incr(refcount, delta), txn=txn) # Now clean up the temporary log tables self._oids.truncate(txn) self._pvids.truncate(txn) self._prevrevids.truncate(txn) self._pending.truncate(txn) |
|
_stupid_dest(format_exception( | _stupid_dest.write(format_exception( | def stupid_log_write(subsystem, severity, summary, detail, error): if severity < 0: return global _stupid_dest if _stupid_dest is None: import os if os.environ.has_key('STUPID_LOG_FILE'): f=os.environ['STUPID_LOG_FILE'] if f: _stupid_dest=open(f,'a') else: import sys _stupid_dest=sys.stderr else: _stupid_dest=_no_stupid_log if _stupid_dest is _no_stupid_log: return _stupid_dest.write( "------\n" "%s %s %s %s\n%s" % (log_time(), severity_string(severity), subsystem, summary, detail, ) ) _stupid_dest.flush() if error: try: _stupid_dest(format_exception( error[0], error[1], error[2], trailer='\n', limit=100)) except: _stupid_dest("%s: %s\n" % error[:2]) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.