text
stringlengths 226
34.5k
|
---|
Project Gutenberg Python problem?
Question: I am trying to process various texts by regex and NLTK of python -which is at
<http://www.nltk.org/book->. I am trying to create a random text generator and
I am having a hard time with a problem. First, here is my algorithm:
1. Enter a sentence as input -this is called trigger string-
2. Get longest word in trigger string
3. Search all Project Gutenberg database for sentences that contain this word -regardless of uppercase lowercase-
4. Return the longest sentence that has the word I spoke about in step 3
5. Append the sentence in Step 1 and Step4 together
6. Repeat the process. Note that I have to get the longest word in second sentence and continue like that and so on-
So far I have been able to do this for first two sentences but I cannot
perform a case insensitive search. Entire sentence database of Project
Gutenberg is available via `gutenberg.sents()` function but regex - case
insensitive search is practically impossible since the `gutenberg.sents()`
outputs the sentences in books as following -in a list of list format-:
EXAMPLE: all the sentences of shakespeare's macbeth is called by typing
import nltk
from nltk.corpus import gutenberg
gutenberg.sents('shakespeare-macbeth.txt')
into the python shell command line and output is:
[['[', 'The', 'Tragedie', 'of', 'Macbeth', 'by', 'William', 'Shakespeare', '1603', ']'],
['Actus', 'Primus', '.'], .......]
with [The Tragedie of Macbeth by William Shakespare, 1603] and Actus Primus.
being the first two sentences.
How can I find the word I'm looking for regardless of it being
uppercase/lowercase ? I'm desperately in need of help since I have been
tinkering with this for the past two days and it's starting to wear on my
nerves. Thanks a lot.
Answer: Given a list `L` of words, and a target word `t`,
any(t.lower()==w.lower() for w in L)
tells you whether L has word t in a case-insensitive way. It's faster, of
course, to do
lt = t.lower()
any(lt==w.lower() for w in L)
since Python does not "hoist" the constant computation out of the loop and,
unless you hoist it yourself, it will be performed repeatedly.
Given a list of lists `lol`, the longest sub-list including `t` can be found
by
longest = max((L for L in lol if any(lt==w.lower() for w in L)), key=len)
If multiple sub-lists include `t` and are of the same maximal length, this
will give you the first one, as it happens.
|
How do I include a Django app in my PYTHONPATH?
Question: I want to be able to import things from applications in my project without
referring to my project name.
My folder structure is like so; I have a project called djangoproject1, and
inside I have a folder called apps, and then I have my individual apps in that
folder.
djangoproject1, apps, and all my applications have an empty "__init__.py" file
in them.
In my settings.py file I have the following code:
import os
import sys
PROJECT_ROOT = os.path.dirname(__file__)
sys.path.insert(0, os.path.join(PROJECT_ROOT, "apps"))
I've tried adding my apps to INSTALLED_APPS several ways:
'djangoproject1.apps.app1', 'djangoproject1.apps.app2',
or
'apps.app1', 'apps.app2',
or
'app1', 'app2',
but nothing works. Whenever I try to do:
from app1 import *
I got an unresolved import error. I'm using the latest versions of eclipse and
django
Answer: Ok, so I got it to work by adding the apps folder to the PYTHONPATH through
eclipse under Project Properties. Is this eclipse only though? I'm not sure
how this will work when I try and deploy the site. What do you guys think?
|
What is the Python equivalent of Ruby's "inspect"?
Question: I just want to quickly see the properties and values of an object in Python,
how do I do that in the terminal on a mac (very basic stuff, never used
python)?
Specifically, I want to see what `message.attachments` are in [this Google App
Engine MailHandler example](http://pastie.org/680280) (images, videos, docs,
etc.).
Answer: If you want to dump the entire object, you can use the
[`pprint`](http://docs.python.org/library/pprint.html) module to get a pretty-
printed version of it.
from pprint import pprint
pprint(my_object)
# If there are many levels of recursion, and you don't want to see them all
# you can use the depth parameter to limit how many levels it goes down
pprint(my_object, depth=2)
Edit: I may have misread what you meant by 'object' - if you're wanting to
look at class instances, as opposed to basic data structures like dicts, you
may want to look at the
[`inspect`](http://docs.python.org/library/inspect.html) module instead.
|
What is Cloud Computing ? and Why?
Question: > **Possible Duplicates:**
> [What is Cloud computing?](http://stackoverflow.com/questions/108037/what-
> is-cloud-computing)
> [What is cloud computing?](http://stackoverflow.com/questions/1830142/what-
> is-cloud-computing)
Sorry for asking some Simple Question. But I cannot clearly understand the
needs of Could Computing. or better to say I am confused about what actually
Cloud Computing is. I've seen a lot of articles on this topic. But they were
non technical and I cannot understand it properly.
Is Cloud Computing a replacement of dedicated Server ? If yes then why would I
use it instead of dedicated Server ? What is Pay per use ? The Pricing Plans
looks like VPS Hosting Plan Charts !! If I need a Service (e.g. it might be
some kind of XML service) heavily used by multiple Applications. I'll put Only
the Service in a dedicated server and apps might be hosted in Shared Hosting
(doesn't matter). Wait a min.. Does it mean My Application hosted on a Cloud
will get CDNs like Google, AOL have :( Sorry I am Confused.
I see in Google App Engine I am restricted in Python and Java Only (thats the
reason I was not interested in Google App Engine much from when it released).
and in an article I saw in Amazon there is no persistent storage. Hmmmm What
does it mean ?? Will I've to reupload everything ?? I see I need to start VM
Images there .. So do I need to pay by the time frame when my image was Up ??
But I find no reason to my Apps non-24hrs.
Sorry I am confused. Please Help me to figure it out.
Answer: > Is Cloud Computing a replacement of dedicated Server ?
Sometimes
Depending on the service offered you upload either a virtual machine image
containing all the stuff you want to run, or a single java program. It runs,
you (and your customers) use it.
> If yes then why would I use it instead of dedicated Server ?
You need a database server today - instead of waiting for Dell to deliver it,
then for you to check out the HW, install the OS, configure it, get the
network working, cool it, manage it, secure it - you just click a couple of
buttons on Amazon/Google and your server is running.
Then your service takes off and you need 100 servers tomorrow, or you have a
business model that needs 100 servers on the last day of the month but 1
server all the time. Cloud computing lets you create new servers by the
minute.
> What is Pay per use ?
You pay for how many cores/how many IP addresses/how many machines you use,
for only the time you use them. Unlike buying a machine and OS license for
life.
> Amazon there is no persistent storage.
Amazon also sells storage with their S3 product which you can use either from
your cloud server or your desktop - I suspect their cloud offering uses S3 so
you buy storage separately.
> But I find no reason to my Apps non-24hrs.
You don't have to, most cloud users are using it to drive their website based
business - stackoverflow could (but doesn't) run on the cloud. SO 24x7
availability is important - this is one of the advantages of the cloud, if a
machine dies your image is instantly started on another machine possibly in
another data center.
Others use it for heavy computing jobs and just buy the time on 100s of
machines to complete the calculation then quit them.
|
Indentation Error python
Question: I'm using twisted API and was going through this example. I inserted one print
statement print "in getdummydata" with correct indentation. code is as below:
from twisted.internet import reactor, defer
def getDummyData(x):
"""
This function is a dummy which simulates a delayed result and
returns a Deferred which will fire with that result. Don't try too
hard to understand this.
"""
print "in getdummydata"
d = defer.Deferred()
# simulate a delayed result by asking the reactor to fire the
# Deferred in 2 seconds time with the result x * 3
reactor.callLater(2, d.callback, x * 3)
return d
def printData(d):
"""
Data handling function to be added as a callback: handles the
data by printing the result
"""
print d
d = getDummyData(3)
d.addCallback(printData)
# manually set up the end of the process by asking the reactor to
# stop itself in 4 seconds time
reactor.callLater(4, reactor.stop)
# start up the Twisted reactor (event loop handler) manually
reactor.run()
But when I run the code it gives the indentation error below:
File "C:\Python26\programs\twisttest.py", line 9
print "in getdummydata"
^
IndentationError: unexpected indent
Please can anyone explain why?
Answer: It looks like the "def" for all your functions have one blank space in front
of them. By my eye, "def" falls under the "r" in the "from" above rather than
the "f".
Perhaps if you remove those spaces the problem will go away. Whitespace is
important to Python.
|
Python and OpenMP C Extensions
Question: I have a C extension in which I'd like to use OpenMP. When I import my module,
though, I get an import error:
ImportError: /home/.../_entropysplit.so: undefined symbol: GOMP_parallel_end
I've compiled the module with -fopenmp and -lgomp. Is this because my Python
installation wasn't compiled with the -fopenmp flag? Will I have to build
Python from source? Or is there some other possibility? This is the only time
I actually use openmp in my module:
unsigned int feature_index;
#pragma omp parallel for
for (feature_index = 0; feature_index < num_features; feature_index++) {
I'd like to stick with openmp if it's possible, just because it's so easy and
the parallelization in this case suits it well.
EDIT: I bit the bullet and recompiled Python with OpenMP support. My module
works perfectly now, but this isn't really a great solution. I can't really
distribute this if it requires a complete recompile of Python. So does anybody
know some way around this? Would ctypes work, maybe?
SOLVED! It was a simple linking issue. (I rebuilt Python for that?!) OpenMP
wasn't being properly linked during the compilation of the module. So it _IS_
possible to load a C Python extension that uses OpenMP.
Answer: Just to make it clearer, here is what your setup.py should look like:
ext = Extension(
'milk.unsupervised._som',
sources = ['milk/unsupervised/_som.cpp'],
extra_compile_args=['-fopenmp'],
extra_link_args=['-lgomp'])
...
setup(..., ext_modules = [ext])
|
Trying to parse an XML file with Python - what am I doing wrong?
Question: I'm working with XML and Python for the first time. The ultimate goal is to
send a request to a REST service, receive a response in XML, and parse the
values and send emails depending on what was returned. However, the REST
service is not yet in place, so for now I'm experimenting with an XML file
saved on my C drive.
I have a simple bit of code, and I'm confused about why it isn't working.
This is my xml file ("XMLTest.xml"):
<Response>
<exitCode>1</exitCode>
<fileName>C:/Something/</fileName>
<errors>
<error>Error generating report</error>
</errors>
</Response>
This is my code so far:
from xml.dom import minidom
something = open("C:/XMLTest.xml")
something = minidom.parse(something)
nodeList = []
for node in something.getElementsByTagName("Response"):
nodeList.extend(t.nodeValue for t in node.childNodes)
print nodeList
But the results that print out are...
[u'\n\t', None, u'\n\t', None, u'\n\t', None, u'\n']
What am I doing wrong?
I'm trying to get the node values. Is there a better way to do this? Is there
a built-in method in Python to convert an xml file to an object or dictionary?
I'd like to get all the values, preferably with the names attached.
Answer: Does this help?
doc = '''<Response>
<exitCode>1</exitCode>
<fileName>C:/Something/</fileName>
<errors>
<error>Error generating report</error>
</errors>
</Response>'''
from xml.dom import minidom
something = minidom.parseString( doc )
nodeList = [ ]
for node in something.getElementsByTagName( "Response" ):
response = { }
response[ "exit code" ] = node.getElementsByTagName( "exitCode" )[ 0 ].childNodes[ 0 ].nodeValue
response[ "file name" ] = node.getElementsByTagName( "fileName" )[ 0 ].childNodes[ 0 ].nodeValue
errors = node.getElementsByTagName( "errors" )[ 0 ].getElementsByTagName( "error" )
response[ "errors" ] = [ error.childNodes[ 0 ].nodeValue for error in errors ]
nodeList.append( response )
import pprint
pprint.pprint( nodeList )
yields
[{'errors': [u'Error generating report'],
'exit code': u'1',
'file name': u'C:/Something/'}]
|
Python: virtualenv - gtk-2.0
Question: To add gtk-2.0 to my virtualenv I did the following:
$ virtualenv --no-site-packages --python=/usr/bin/python2.6 myvirtualenv
$ cd myvirtualenv
$ source bin/activate
$ cd lib/python2.6/
$ ln -s /usr/lib/pymodules/python2.6/gtk-2.0/
<http://stackoverflow.com/questions/249283/virtualenv-on-ubuntu-with-no-site-
packages>
Now in the Python interpreter when I do import gtk it says: No module named
gtk. When I start the interpreter with sudo it works.
Any reason why I need to use sudo and is there a way to prevent it?
**Update:**
Forgot to mention that cairo and pygtk work but it's not the one I need.
**Update2:**
Here the directory to show that I ain't crazy. 
Answer: `sudo python` imports it just fine because that interpreter isn't using your
virtual environment. So don't do that.
You only linked in one of the necessary items. Do the others mentioned in the
answer to the question you linked as well.
(The pygtk.pth file is of particular importance, since it tells python to
actually put that directory you linked onto the python path)
## Update
Put that stuff in $VIRTUALENV/lib/python2.6/**site-packages/** rather than the
directory above that.
Looks like the .pth files aren't read from that directory - just from site-
packages
|
error while trying to connect with a cisco 2600 router through python in windows
Question:
The code i have given is
import telnetlib
HOST="X"
user ="X"
password="X"
en_password="x"
tn=telnetlib.Telnet(HOST)
Traceback (most recent call last):
File "<pyshell#15>", line 1, in <module>
tn=telnetlib.Telnet(HOST)
File "C:\Python27\lib\telnetlib.py", line 209, in __init__
self.open(host, port, timeout)
File "C:\Python27\lib\telnetlib.py", line 225, in open
self.sock = socket.create_connection((host, port), timeout)
File "C:\Python27\lib\socket.py", line 567, in create_connection
raise error, msg
error: [Errno 10060] A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond
This is the error that i am getting while i am trying to connect to the 2600
router . How to remove this error , and connect to the router through a python
script?
Answer: It's a connection timeout - if you're not having a other networking issues
then it's simply that the router is not accepting connections on the default
telnet port. Are you sure you can connect via port 23? Can you use a telnet
client to connect?
|
wxPython and py2app, CreateActCtx error 0x00000008 (Not enough disk space available)
Question: I've been developing an application that uses wxPython as the GUI librar, and
py2exe so that I can easily distribute it, however I have just now tested
py2exe and the following error appears when the executable is launched.
12:13:08: Debug: src/helpers.cpp(140): 'CreateActCtx' failed with error 0x00000008 (Not enough disk space available.).
Traceback (most recent call last):
File "eYoutubeMacros3.py", line 1, in <module>
File "zipextimporter.pyo", line 82, in load_module
File "application\application.pyo", line 5, in <module>
File "zipextimporter.pyo", line 82, in load_module
File "application\backend\backend.pyo", line 4, in <module>
File "zipextimporter.pyo", line 82, in load_module
File "application\backend\extractor.pyo", line 5, in <module>
File "zipextimporter.pyo", line 82, in load_module
File "twisted\web\client.pyo", line 17, in <module>
File "zipextimporter.pyo", line 82, in load_module
File "twisted\web\error.pyo", line 188, in <module>
ImportError: cannot import name resource
The function causing the error in src/helpers.cpp is
static ULONG_PTR wxPySetActivationContext()
{
OSVERSIONINFO info;
wxZeroMemory(info);
info.dwOSVersionInfoSize = sizeof(OSVERSIONINFO);
GetVersionEx(&info);
if (info.dwMajorVersion < 5)
return 0;
ULONG_PTR cookie = 0;
HANDLE h;
ACTCTX actctx;
TCHAR modulename[MAX_PATH];
GetModuleFileName(wxGetInstance(), modulename, MAX_PATH);
wxZeroMemory(actctx);
actctx.cbSize = sizeof(actctx);
actctx.lpSource = modulename;
actctx.lpResourceName = MAKEINTRESOURCE(2);
actctx.hModule = wxGetInstance();
actctx.dwFlags = ACTCTX_FLAG_HMODULE_VALID | ACTCTX_FLAG_RESOURCE_NAME_VALID;
h = CreateActCtx(&actctx);
if (h == INVALID_HANDLE_VALUE) {
wxLogLastError(wxT("CreateActCtx"));
return 0;
}
if (! ActivateActCtx(h, &cookie))
wxLogLastError(wxT("ActivateActCtx"));
return cookie;
}
And lastly my code for py2exe
setup(
console = [self.target], # Contains some build info, is this is relevant I'll add it
zipfile = 'library.dat',
options = {
'py2exe' : {
'bundle_files' : 1,
'dll_excludes' : ['w9xpopen.exe'],
'optimize' : 2,
'dist_dir' : '../dist/executables/',
'compressed' : True,
#'excludes' : ['doctest', 'pdb', 'unittest', 'difflib', 'inspect'],
}
}
)
Edit: Yes the second error seems to be from twisted but I doubt that causes
the first error. Edit2: Hmm perhaps the first one is just a warning.
Answer: That means common controls stuff does not load. The second error could be a
result of the first error which is non fatal and program continues to run.
try first :
(Don't bundle option) and check if the issue still appears. This should
typically work.
bundle_files = 3
try next:
Since, you are using bundle option 1 , Can you check which MSVC runtime DLL is
located in the dist directory along side the executable. I would suggest that
you also find out all MSVCRXX.dll on your machine and see if there are version
issues
|
Django - make file I/O thread safe
Question: I want to read and write python-source-files from the file system in a thread-
safe way.
open("n2.py","w").write(my_new_py_class)
from myproject import n2
#do something with n2
I assume that this is not thread-safe, since a request2 could modify the file
before request1 is loading and executing it. I would like to achieve something
like that one thread is waiting till the other thread wrote, loaded, executed
and closed the file.
Answer: Why are you making your application modify its own files? This is not only
incredibly evil, metaprogramming is orders of magnitude harder to understand
debug. Plus, python caches modules it imports, so it's not really easy to
reload that. And, last but not least, you don't have to writ the code to a
file to execute it, if you really must execute dynamically generated code.
To answer your question about writing files in a thread safe way, the general
convention is to:
* Write your content to a temporary file on the same filesystem as your target file.
* Rename that temporary file to your target file, overwritting it in the process.
This works, because rename is atomic on POSIX systems, when done on the same
device. So other threads/processes will either still have the old file opened,
which is now detached from the filesystem and will be deleted as soon as those
threads/processes are done with it, or they will open the new file with all of
its contents. You avoid having a file that is only half-written.
In practice, I like to make a temporary directory with python's tempfile
module, and write the file in there, then move it and remove the directory --
this way the file is being created with default umask.
Last but not least, rename is not really atomic on Windows, at least with
default settings, as it won't let you overwrite your old file -- you need to
do two renames, and that introduces a possibility of race condition. I don't
know a good solution for Windows.
|
Any quick Python GUI to display live images from Camera
Question: I am trying to display live images from my 1394 camera. Currently my code is
able to obtain images in a loop from the camera and I was looking for any
quick GUI that will update dynamically (as a separate thread). I can do this
in PyQt maybe using QThreads but is there any recommendation or faster way of
doing this?? Here's my code
#Loop capturing frames from camera
for frame in range(1,500):
print 'frame:',frame
TIME.sleep(1) #capture frame every second
image_binary = pycam.cam.RetrieveBuffer()
#convert to PIL Image
pilimg = PIL.Image.frombuffer("L",(cimg.GetCols(),cimg.GetRows()),image_binary,'raw', "RGBA", 0, 1)
# At this point I want to send my image data to a GUI window and display it
Thank you.
Answer: Here's [wxPython](http://www.wxpython.org/) code that will do it...
import wx
from PIL import Image
SIZE = (640, 480)
def get_image():
# Put your code here to return a PIL image from the camera.
return Image.new('L', SIZE)
def pil_to_wx(image):
width, height = image.size
buffer = image.convert('RGB').tostring()
bitmap = wx.BitmapFromBuffer(width, height, buffer)
return bitmap
class Panel(wx.Panel):
def __init__(self, parent):
super(Panel, self).__init__(parent, -1)
self.SetSize(SIZE)
self.SetBackgroundStyle(wx.BG_STYLE_CUSTOM)
self.Bind(wx.EVT_PAINT, self.on_paint)
self.update()
def update(self):
self.Refresh()
self.Update()
wx.CallLater(15, self.update)
def create_bitmap(self):
image = get_image()
bitmap = pil_to_wx(image)
return bitmap
def on_paint(self, event):
bitmap = self.create_bitmap()
dc = wx.AutoBufferedPaintDC(self)
dc.DrawBitmap(bitmap, 0, 0)
class Frame(wx.Frame):
def __init__(self):
style = wx.DEFAULT_FRAME_STYLE & ~wx.RESIZE_BORDER & ~wx.MAXIMIZE_BOX
super(Frame, self).__init__(None, -1, 'Camera Viewer', style=style)
panel = Panel(self)
self.Fit()
def main():
app = wx.PySimpleApp()
frame = Frame()
frame.Center()
frame.Show()
app.MainLoop()
if __name__ == '__main__':
main()
|
How to make python 3 print() utf8
Question: How to make python 3 (3.1) to print("Some text") to stdout in utf8 ...
or how to output raw bytes..
Test.py
>
> TestText = "Test - āĀēĒčČ..šŠūŪžŽ" # this is UTF-8
> TestText2 = b"Test2 -
> \xc4\x81\xc4\x80\xc4\x93\xc4\x92\xc4\x8d\xc4\x8c..\xc5\xa1\xc5\xa0\xc5\xab\xc5\xaa\xc5\xbe\xc5\xbd"
> # just bytes
> print(sys.getdefaultencoding())
> print(sys.stdout.encoding)
> print(TestText)
> print(TestText.encode("utf8"))
> print(TestText.encode("cp1252","replace"))
> print(TestText2)
>
Output: \\\ in cp1257 and I replaced chars to byte values [xHEX]
> utf-8
> cp1257
> Test - [xE2][xC2][xE7][C7][xE8][xC8]..[xF0][xD0][xFB][xDB][xFE][xDE]
> b'Test -
> \xc4\x81\xc4\x80\xc4\x93\xc4\x92\xc4\x8d\xc4\x8c..\xc5\xa1\xc5\xa0\xc5\xab\xc5\xaa\xc5\xbe\xc5\xbd'
> b'Test - ??????..\x9a\x8a??\x9e\x8e'
> b'Test2 -
> \xc4\x81\xc4\x80\xc4\x93\xc4\x92\xc4\x8d\xc4\x8c..\xc5\xa1\xc5\xa0\xc5\xab\xc5\xaa\xc5\xbe\xc5\xbd'
print() is just too smart... :D
there's no point using encoded text with print (it always show only
representation of bytes not real bytes)
and it's impossible to output bytes at all, because print anyway and always
encodes it in sys.stdout.encoding
for example:
>
> print(chr(255))
>
throws an error
> Traceback (most recent call last):
> File "Test.py", line 1, in
> print(chr(255));
> File "H:\Python31\lib\encodings\cp1257.py", line 19, in encode
> return codecs.charmap_encode(input,self.errors,encoding_table)[0]
> UnicodeEncodeError: 'charmap' codec can't encode character '\xff' in
> position 0: character maps to
by the way print( TestText == TestText2.decode("utf8")); returns False...
although print output is same...
EDIT:
How python 3 gets sys.stdout.encoding and how to change it?
I made printRAW function witch works fine :) (tnx Zack)
(actually it encodes output to UTF-8, so in real it's not raw...)
>
> def printRAW(*Text):
> RAWOut = open(1, 'w', encoding='utf8', closefd=False)
> print(*Text, file=RAWOut)
> RAWOut.flush()
> RAWOut.close()
>
> printRAW("Cool", TestText)
>
output: \\\ now it print in UTF-8
> Cool Test - āĀēĒčČ..šŠūŪžŽ
printRAW(chr(252)) also nicely prints ü (in UTF-8, [xC3][xBC]) and without
errors :)
Now I'm looking for maybe better solution if there's any...
Answer: First, a correction:
TestText = "Test - āĀēĒčČ..šŠūŪžŽ" # this NOT utf-8...it is a Unicode string in Python 3.X.
TestText2 = TestText.encode('utf8') # THIS is "just bytes" in UTF-8.
Now, to send UTF-8 to stdout, regardless of the console's encoding, use the
right tool for the job:
import sys
sys.stdout.buffer.write(TestText2)
"buffer" is a raw interface to stdout.
|
Python: 'Private' module in a package
Question: I have a package `mypack` with modules `mod_a` and `mod_b` in it. I intend the
the package itself and `mod_a` to be imported freely:
import mypack
import mypack.mod_a
However, I'd like to keep `mod_b` for the exclusive use of `mypack`. That's
because it exists merely to organize the latter's internal code.
My first question is, is it an accepted practice in Python programming to have
'private' modules like this?
If yes, my second question is, what is the best way to convey this intention
to the client? Do I prefix the name with an underscore (i.e. `_mod_b`)? Or
would it be a good idea to declare a sub-package `private` and place all such
modules there?
Answer: I prefix private modules with an underscore to communicate the intent to the
user. In your case, this would be `mypack._mod_b`
This is in the same spirit (but not completely analogous to) the PEP8
recommendation to name C-extension modules with a leading underscore when it’s
wrapped by a Python module; i.e., `_socket` and `socket`.
|
Tell pydev to exclude an entire package from analysis?
Question: Today I'm on a mission to remove little red X's from my django project in
pydev. Mostly, this involves fixing import problems with pydev.
I'm using [South](http://south.aeracode.org/) for database migrations. South
(if you don't know) generates python modules, and pydev doesn't like them. I
don't want to edit the south code since it's generated.
Is there a way to instruct pydev to exclude certain packages from analysis?
Something like `#@UndefinedVariable`, except for the entire module? Ideally
I'd like to ignore packages named "migrations".
Answer: In South, I have added a "#@PydevCodeAnalysisIgnore" to the templates in
`south/management/datamigration.py` and `south/management/schemamigration.py`.
It doesn't let me ignore entire packages, but serves my purposes well enough.
|
Creating a simple XML file using python
Question: What are my options if I want to create a simple XML file in python? (library
wise)
The xml I want looks like:
<root>
<doc>
<field1 name="blah">some value1</field1>
<field2 name="asdfasd">some vlaue2</field2>
</doc>
</root>
Answer: These days, the most popular (and very simple) option is the ElementTree API,
which has been included in the standard library since Python 2.5.
The available options for that are:
* ElementTree (Basic, pure-Python implementation of ElementTree. Part of the standard library since 2.5)
* cElementTree (Optimized C implementation of ElementTree. Also offered in the standard library since 2.5)
* LXML (Based on libxml2. Offers a a rich superset of the ElementTree API as well XPath, CSS Selectors, and more)
Here's an example of how to generate your example document using the in-stdlib
cElementTree:
import xml.etree.cElementTree as ET
root = ET.Element("root")
doc = ET.SubElement(root, "doc")
ET.SubElement(doc, "field1", name="blah").text = "some value1"
ET.SubElement(doc, "field2", name="asdfasd").text = "some vlaue2"
tree = ET.ElementTree(root)
tree.write("filename.xml")
I've tested it and it works, but I'm assuming whitespace isn't significant. If
you need "prettyprint" indentation, let me know and I'll look up how to do
that. (It may be an LXML-specific option. I don't use the stdlib
implementation much)
For further reading, here are some useful links:
* [API docs for the implementation in the Python standard library](http://docs.python.org/library/xml.etree.elementtree.html)
* [Introductory Tutorial](http://effbot.org/zone/element-index.htm) (From the original author's site)
* [LXML etree tutorial](http://lxml.de/tutorial.html). (With example code for loading the best available option from all major ElementTree implementations)
As a final note, either cElementTree or LXML should be fast enough for all
your needs (both are optimized C code), but in the event you're in a situation
where you need to squeeze out every last bit of performance, the benchmarks on
the LXML site indicate that:
* LXML clearly wins for serializing (generating) XML
* As a side-effect of implementing proper parent traversal, LXML is a bit slower than cElementTree for parsing.
|
ctypes.windll.user32.GetCursorInfo() - how can I manage this to work? [Python]
Question: I have to get the information about the current mouse cursor from windows but
I'm not managing to work this command...
what should I do? Can someone post one example?
Answer: What information are you trying to get out of the GetCursorInfo() call? It
would be easier to use the [win32
extensions](http://python.net/crew/skippy/win32/Downloads.html) (especially if
you just want cursor position).
>>> import win32gui
>>> win32gui.GetCursorInfo()
(1, 65555, (717, 412))
|
Python matplotlib: memory not being released when specifying figure size
Question: I'm using matplotlib to generate many plots of the results of a numerical
simulation. The plots are used as frames in a video, and so I'm generating
many of them by repeatedly calling a function similar to this one:
from pylab import *
def plot_density(filename,i,t,psi_Na):
figure(figsize=(8,6))
imshow(abs(psi_Na)**2,origin = 'lower')
savefig(filename + '_%04d.png'%i)
clf()
The problem is that the memory usage of the python process grows by a couple
of megabytes with every call to this function. For example if I call it with
this loop:
if __name__ == "__main__":
x = linspace(-6e-6,6e-6,128,endpoint=False)
y = linspace(-6e-6,6e-6,128,endpoint=False)
X,Y = meshgrid(x,y)
k = 1000000
omega = 200
times = linspace(0,100e-3,100,endpoint=False)
for i,t in enumerate(times):
psi_Na = sin(k*X-omega*t)
plot_density('wavefunction',i,t,psi_Na)
print i
then the ram usage grows with time to 600MB. If however I comment out the line
`figure(figsize=(8,6))` in the function definition, then the ram usage stays
steady at 52MB. `(8,6)` is the default figure size and so identical images are
produced in both cases. I'd like to make different sized plots from my
numerical data without running out of ram. How might I force python to free up
this memory?
I've tried `gc.collect()` each loop to force garbage collection, and I've
tried `f = gcf()` to get the current figure and then `del f` to delete it, but
to no avail.
I'm running CPython 2.6.5 on 64 bit Ubuntu 10.04.
Answer: From the docstring for `pylab.figure`:
In [313]: pylab.figure?
> If you are creating many figures, make sure you explicitly call "close" on
> the figures you are not using, because this will enable pylab to properly
> clean up the memory.
So perhaps try:
pylab.close() # closes the current figure
|
Python: How exactly can you take a string, split it, reverse it and join it back together again?
Question: How exactly can you take a string, split it, reverse it and join it back
together again without the brackets, commas, etc. using python?
Answer:
>>> tmp = "a,b,cde"
>>> tmp2 = tmp.split(',')
>>> tmp2.reverse()
>>> "".join(tmp2)
'cdeba'
or simpler:
>>> tmp = "a,b,cde"
>>> ''.join(tmp.split(',')[::-1])
'cdeba'
The important parts here are the [split
function](http://docs.python.org/library/stdtypes.html#str.split) and the
[join function](http://docs.python.org/library/stdtypes.html#str.join). To
reverse the list you can use `reverse()`, which reverses the list in place or
the slicing syntax `[::-1]` which returns a new, reversed list.
|
UDP client and server with Twisted Python
Question: I want to create a server and client that sends and receives UDP packets from
the network using Twisted. I've already written this with sockets in Python,
but want to take advantage of Twisted's callback and threading features.
However, I need help though with the design of Twisted.
I have multiple types of packets I want to receive, but let's pretend there is
just one:
class Packet(object):
def __init__(self, data=None):
self.packet_type = 1
self.payload = ''
self.structure = '!H6s'
if data == None:
return
self.packet_type, self.payload = struct.unpack(self.structure, data)
def pack(self):
return struct.pack(self.structure, self.packet_type, self.payload)
def __str__(self):
return "Type: {0}\nPayload {1}\n\n".format(self.packet_type, self.payload)
I made a protocol class (almost direct copy of the examples), which seems to
work when I send data from another program:
class MyProtocol(DatagramProtocol):
def datagramReceived(self, data, (host, port)):
p = Packet(data)
print p
reactor.listenUDP(3000, MyProtocol())
reactor.run()
What I don't know is how do I create a client which can send arbitrary packets
on the network, which get picked up by the reactor:
# Something like this:
s = Sender()
p = Packet()
p.packet_type = 3
s.send(p.pack())
p.packet_type = 99
s.send(p.pack())
I also need to make sure to set the reuse address flag on the client and
servers so I can run multiple instances of each at the same time on the same
device (e.g. one script is sending heartbeats, another responds to heartbeats,
etc).
Can someone show me how this could be done with Twisted?
**Update** :
This is how I do it with sockets in Python. I can run multiple listeners and
senders at the same time and they all hear each other. How do I get this
result with Twisted? (The listening portion need not be a separate process.)
class Listener(Process):
def __init__(self, ip='127.0.0.1', port=3000):
Process.__init__(self)
self.ip = ip
self.port = port
def run(self):
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
sock.bind((self.ip, self.port))
data, from_ip = sock.recvfrom(4096)
p = Packet(data)
print p
class Sender(object):
def __init__(self, ip='127.255.255.255', port=3000):
self.sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
self.sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
self.ip = (ip, port)
def send(self, data):
self.sock.sendto(data, self.ip)
if __name__ == "__main__":
l = Listener()
l.start()
s = Sender()
p = Packet()
p.packet_type = 4
p.payload = 'jake'
s.send(p.pack())
**Working solution** :
class MySender(DatagramProtocol):
def __init__(self, packet, host='127.255.255.255', port=3000):
self.packet = packet.pack()
self.host = host
self.port = port
def startProtocol(self):
self.transport.write(self.packet, (self.host, self.port))
if __name__ == "__main__":
packet = Packet()
packet.packet_type = 1
packet.payload = 'jake'
s = MySender(packet)
reactor.listenMulticast(3000, MyProtocol(), listenMultiple=True)
reactor.listenMulticast(3000, s, listenMultiple=True)
reactor.callLater(4, reactor.stop)
reactor.run()
Answer: Just like the server example above, there is a client example to. This should
help you get started:
* <http://www.opendocs.net/python/twisted/howto/udp.html>
* <http://twistedmatrix.com/documents/current/core/examples/echoclient_udp.py>
Ok, here is a simple heart beat sender and receiver using datagram protocol.
from twisted.internet.protocol import DatagramProtocol
from twisted.internet import reactor
from twisted.internet.task import LoopingCall
import sys, time
class HeartbeatSender(DatagramProtocol):
def __init__(self, name, host, port):
self.name = name
self.loopObj = None
self.host = host
self.port = port
def startProtocol(self):
# Called when transport is connected
# I am ready to send heart beats
self.loopObj = LoopingCall(self.sendHeartBeat)
self.loopObj.start(2, now=False)
def stopProtocol(self):
"Called after all transport is teared down"
pass
def datagramReceived(self, data, (host, port)):
print "received %r from %s:%d" % (data, host, port)
def sendHeartBeat(self):
self.transport.write(self.name, (self.host, self.port))
class HeartbeatReciever(DatagramProtocol):
def __init__(self):
pass
def startProtocol(self):
"Called when transport is connected"
pass
def stopProtocol(self):
"Called after all transport is teared down"
def datagramReceived(self, data, (host, port)):
now = time.localtime(time.time())
timeStr = str(time.strftime("%y/%m/%d %H:%M:%S",now))
print "received %r from %s:%d at %s" % (data, host, port, timeStr)
heartBeatSenderObj = HeartbeatSender("sender", "127.0.0.1", 8005)
reactor.listenMulticast(8005, HeartbeatReciever(), listenMultiple=True)
reactor.listenMulticast(8005, heartBeatSenderObj, listenMultiple=True)
reactor.run()
The broadcast example simply modifies the above approach:
from twisted.internet.protocol import DatagramProtocol
from twisted.internet import reactor
from twisted.internet.task import LoopingCall
import sys, time
class HeartbeatSender(DatagramProtocol):
def __init__(self, name, host, port):
self.name = name
self.loopObj = None
self.host = host
self.port = port
def startProtocol(self):
# Called when transport is connected
# I am ready to send heart beats
self.transport.joinGroup('224.0.0.1')
self.loopObj = LoopingCall(self.sendHeartBeat)
self.loopObj.start(2, now=False)
def stopProtocol(self):
"Called after all transport is teared down"
pass
def datagramReceived(self, data, (host, port)):
print "received %r from %s:%d" % (data, host, port)
def sendHeartBeat(self):
self.transport.write(self.name, (self.host, self.port))
class HeartbeatReciever(DatagramProtocol):
def __init__(self, name):
self.name = name
def startProtocol(self):
"Called when transport is connected"
self.transport.joinGroup('224.0.0.1')
pass
def stopProtocol(self):
"Called after all transport is teared down"
def datagramReceived(self, data, (host, port)):
now = time.localtime(time.time())
timeStr = str(time.strftime("%y/%m/%d %H:%M:%S",now))
print "%s received %r from %s:%d at %s" % (self.name, data, host, port, timeStr)
heartBeatSenderObj = HeartbeatSender("sender", "224.0.0.1", 8005)
reactor.listenMulticast(8005, HeartbeatReciever("listner1"), listenMultiple=True)
reactor.listenMulticast(8005, HeartbeatReciever("listner2"), listenMultiple=True)
reactor.listenMulticast(8005, heartBeatSenderObj, listenMultiple=True)
reactor.run()
|
Is this a problem with the Django tutorial or a package problem, or is it me?
Question: I'm using Ubuntu 10, python 2.6.5
I'm following this tutorial: <http://www.djangobook.com/en/2.0/chapter02>
I followed all of the steps using cut-and-paste. The following directory
structure was automatically created:
bill@ed-desktop:~/projects$ ls -l mysite
total 36
-rw-r--r-- 1 bill bill 0 2010-09-01 08:18 __init__.py
-rw-r--r-- 1 bill bill 546 2010-09-01 08:18 manage.py
-rw-r--r-- 1 bill bill 20451 2010-09-01 18:50 mysite.wpr
-rw-r--r-- 1 bill bill 3291 2010-09-01 08:18 settings.py
-rw-r--r-- 1 bill bill 127 2010-09-01 11:13 urls.py
-rw-r--r-- 1 bill bill 97 2010-09-01 08:20 views.py
# urls.py
from django.conf.urls.defaults import *
import sys
print sys.path
from mysite.views import hello
urlpatterns = patterns('',
(r'^hello/$', hello),
)
pylint produces this error: Unable to import 'mysite.views'
# views.py
from django.http import HttpResponse
def hello(request):
return HttpResponse("Hello world")
bill@ed-desktop:~/projects/mysite$ python manage.py runserver
Validating models...
0 errors found
Django version 1.2.1, using settings 'mysite.settings'
Development server is running at http://127.0.0.1:8000/
Quit the server with CONTROL-C.
Which resulted in:
Page not found (404)
Request Method: GET
Request URL: http://127.0.0.1:8000/
Using the URLconf defined in mysite.urls, Django tried these URL patterns, in this order:
1. ^hello/$
The current URL, , didn't match any of these.
Why does view.py which is in the main directory contain the following?
from mysite.views import hello
There is no subdirectory 'views'. Although I'm familiar with using packages,
I've never had the need to create my own so I'm a bit confused. I would have
thought that `from views import hello` would be correct.
The step-by-step tutorial looks straight forward and I haven't seen anyone
else come across this problem so I'm a bit perplexed what I've done wrong.
Answer: I'm not sure what your actual question is.
You've requested the root page, `\`, but have only defined a URL for
`\hello\`, so obviously Django can't find what you've requested. If you want
your `hello` view to match against the site root, do this:
urlpatterns = patterns('',
(r'^$', hello),
)
I don't understand the question about the `from mysite.views import hello`.
This will work if the parent of `mysite` is on the Python path.
|
How to write stereo wav files in Python?
Question: The following code writes a simple sine at frequency 400Hz to a mono WAV file.
How should this code be changed in order to produce a **stereo** WAV file. The
second channel should be in a different frequency.
import math
import wave
import struct
freq = 440.0
data_size = 40000
fname = "WaveTest.wav"
frate = 11025.0 # framerate as a float
amp = 64000.0 # multiplier for amplitude
sine_list_x = []
for x in range(data_size):
sine_list_x.append(math.sin(2*math.pi*freq*(x/frate)))
wav_file = wave.open(fname, "w")
nchannels = 1
sampwidth = 2
framerate = int(frate)
nframes = data_size
comptype = "NONE"
compname = "not compressed"
wav_file.setparams((nchannels, sampwidth, framerate, nframes,
comptype, compname))
for s in sine_list_x:
# write the audio frames to file
wav_file.writeframes(struct.pack('h', int(s*amp/2)))
wav_file.close()
Answer: Build a parallel `sine_list_y` list with the other frequency / channel, set
`nchannels=2`, and in the output loop use `for s, t in zip(sine_list_x,
sine_list_y):` as the header clause, and a body with two `writeframes` calls
-- one for `s`, one for `t`. IOW, corresponding frames for the two channels
"alternate" in the file.
See e.g. [this](http://www.sonicspot.com/guide/wavefiles.html) page for a
thorough description of all possible WAV file formats, and I quote:
> Multi-channel digital audio samples are stored as interlaced wave data which
> simply means that the audio samples of a multi-channel (such as stereo and
> surround) wave file are stored by cycling through the audio samples for each
> channel before advancing to the next sample time. This is done so that the
> audio files can be played or streamed before the entire file can be read.
> This is handy when playing a large file from disk (that may not completely
> fit into memory) or streaming a file over the Internet. The values in the
> diagram below would be stored in a Wave file in the order they are listed in
> the Value column (top to bottom).
and the following table clearly shows the channels' samples going left, right,
left, right, ...
|
Alternative to Passing Global Variables Around to Classes and Functions
Question: I'm new to python, and I've been reading that using `global` to pass variables
to other functions is considered noobie, as well as a bad practice. I would
like to move away from using global variables, but I'm not sure what to do
instead.
Right now I have a UI I've created in wxPython as its own separate class, and
I have another class that loads settings from a .ini file. Since the settings
in the UI should match those in the .ini, how do I pass around those values? I
could using something like: `Settings = Settings()` and then define the
variables as something like `self.settings1`, but then I would have to make
`Settings` a global variable to pass it to my UI class (which it wouldn't be
if I assign in it `main()`).
So what is the correct and pythonic way to pass around these variables?
**Edit:** Here is the code that I'm working with, and I'm trying to get it to
work like Alex Martelli's example. The following code is saved in
`Settings.py`:
import ConfigParser
class _Settings():
@property
def enableautodownload(self): return self._enableautodownload
def __init__(self):
self.config = ConfigParser.ConfigParser()
self.config.readfp(open('settings.ini'))
self._enableautodownload=self.config.getboolean('DLSettings', 'enableautodownload')
settings = _Settings()
Whenever I try to refer to `Settings.settings.enableautodownload` from another
file I get: `AttributeError: 'module' object has no attribute 'settings'`.
What am I doing wrong?
**Edit 2:** Never mind about the issue, I retyped the code and it works now,
so it must have been a simple spelling or syntax error.
Answer: The alternatives to `global` variables are many -- mostly:
* explicit arguments to functions, classes called to create one of their instance, etc (this is usually the clearest, since it makes the dependency most explicit, when feasible and not too repetitious);
* instance variables of an object, when the functions that need access to those values are methods on that same object (that's OK too, and a reasonable way to use OOP);
* "accessor functions" that provide the values (or an object which has attributes or properties for the values).
Each of these (esp. the first and third ones) is particularly useful for
values whose names must _not_ be re-bound by all and sundry, but only
accessed. The really big problem with `global` is that it provides a "covert
communication channel" (not in the cryptographic sense, but in the literal
one: apparently separate functions can actually be depending on each other,
influencing each other, via global values that are not "obvious" from the
functions' signatures -- this makes the code hard to test, debug, maintain,
and understand).
For your specific problem, if you never use the `global` statement, but rather
access the settings in a "read-only" way from everywhere (and you can ensure
that more fully by making said object's attributes be read-only properties!),
then having the "read-only" accesses be performed on a single, made-once-then-
not-changed, module-level instance, is not too bad. I.e., in some module
`foo.py`:
class _Settings(object):
@property
def one(self): return self._one
@property
def two(self): return self._two
def __init__(self, one, two):
self._one, self._two = one, two
settings = _Settings(23, 45)
and from everywhere else, `import foo` then just access `foo.settings.one` and
`foo.settings.two` as needed. Note that I've named the class with a single
leading underscore (just like the two instance attributes that underlie the
read-only properties) to suggest that it's not meant to be used from "outside"
the module -- only the `settings` object is supposed to be (there's no
enforcement -- but any user violating such requested privacy is most obviously
the only party responsible for whatever mayhem may ensue;-).
|
ImportError: no module named _jcc
Question: I'm trying to install pylucene on python 2.7 (windows) since four days. It
requires JCC to build and install. After thousands and thousands different
errors I handled myself, at last JCC sucessfully builded and installed. At
least, that was what I thought. After that, I tried to import jcc but I get
this error:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "jcc\__init__.py", line 29, in <module>
from _jcc import initVM
ImportError: No module named _jcc
any ideas? It looks installed perfectly but I can't import it.
Answer: Make sure your current dir is not jcc build directory. I get this error if I
try to import jcc when in build directory, python uses the wrong jcc then.
|
Sql Alchemy > TypeError: 'instancemethod' object does not support item assignment
Question: Here's what I've got:
from sqlalchemy import *
from sqlalchemy.orm import *
from web.models.card import *
connectionString = "postgresql://www:www@localhost/prod"
databaseEngine = create_engine(connectionString)
sessionFactory = sessionmaker(autoflush = True, autocommit = False, bind = databaseEngine)
session = sessionFactory()
CardsCollection = session.query(card).all()
_content = {}
for index in range(0, len(CardsCollection)):
c = CardsCollection[index]
_content[index] = c
print json.dumps(_content)
`
And here's the error:
Traceback (most recent call last):
File "/home/src/py/raspberry/src/dictionaryTest.py", line 15, in
CardsCollection = session.query(card).all()
File "/usr/local/lib/python2.6/dist-packages/sqlalchemy/orm/query.py", line 1453, in all
return list(self)
File "/usr/local/lib/python2.6/dist-packages/sqlalchemy/orm/query.py", line 1676, in instances
rows = [process[0](row, None) for row in fetch]
File "/usr/local/lib/python2.6/dist-packages/sqlalchemy/orm/mapper.py", line 2234, in _instance
populate_state(state, dict_, row, isnew, only_load_props)
File "/usr/local/lib/python2.6/dist-packages/sqlalchemy/orm/mapper.py", line 2113, in populate_state
populator(state, dict_, row)
File "/usr/local/lib/python2.6/dist-packages/sqlalchemy/orm/strategies.py", line 127, in new_execute
dict_[key] = row[col]
TypeError: 'instancemethod' object does not support item assignment
`
Can someone help me out with this? I've tried a few things, and researched
into how dictionaries work... but its just not jumping out at me.
**[edit for strange resolution]** Apparently, overriding the
`self.__dict__(self)` method on the card model is what did it. I'm not
entirely sure why, though.
Answer: `__dict__` is a [special
attribute](http://docs.python.org/reference/datamodel.html#index-862) holding
current state of instance, overwriting it with with method will certainly lead
to troubles.
|
How to create a simple mesh in Blender 2.50 via the Python API
Question: I would like to create a simple mesh in Blender (2.50) via the Python API but
the examples from the API documentation don't work yet.
I tried the following but it's [from API
2.49](http://www.blender.org/documentation/249PythonDoc/Mesh-module.html)
from Blender import *
import bpy
editmode = Window.EditMode() # are we in edit mode? If so ...
if editmode: Window.EditMode(0) # leave edit mode before getting the mesh
# define vertices and faces for a pyramid
coords=[ [-1,-1,-1], [1,-1,-1], [1,1,-1], [-1,1,-1], [0,0,1] ]
faces= [ [3,2,1,0], [0,1,4], [1,2,4], [2,3,4], [3,0,4] ]
me = bpy.data.meshes.new('myMesh') # create a new mesh
me.verts.extend(coords) # add vertices to mesh
me.faces.extend(faces) # add faces to the mesh (also adds edges)
me.vertexColors = 1 # enable vertex colors
me.faces[1].col[0].r = 255 # make each vertex a different color
me.faces[1].col[1].g = 255
me.faces[1].col[2].b = 255
scn = bpy.data.scenes.active # link object to current scene
ob = scn.objects.new(me, 'myObj')
if editmode: Window.EditMode(1) # optional, just being nice
This does not work because the mesh object doesn't have any `faces` or `verts`
members.
Are there any options to do this?
Answer: Try [this](http://www.blender.org/documentation/250PythonDoc/) documentation
for the 2.5x API. I understand that despite the big warnings at the top, the
most used sections are fairly stable now. I've not tried it yet.
EDIT:
I think the relevant bit is [this
section](http://www.blender.org/documentation/blender_python_api_2_72_release/bpy.types.Mesh.html#bpy.types.Mesh.from_pydata)
\- it seems you create a list of vertices faces etc. and pass it to this. This
seems to have changed from the most recent examples I can find. Try looking in
your scripts folder - there might be an example there that you can look at.
EDIT 2: I have updated the link to point to the current live docs. The notes
there suggest that there are probably better ways of doing this now but it is
a long time since I have done any blender scripting so can't help more.
|
Problem decrypting PGP in python with pyme without user interaction
Question: I am trying to decrypt messages using pyme (a python wrapper from gpgme). It
works fine if I type in the password when it prompts but I cannot get the
passphrase callback to work. Here is the code
import pyme.core
def Callback( x, y, z ):
print 'in passphrase callback'
return 'passphrase'
plain = pyme.core.Data()
cipher = pyme.core.Data(sys.stdin.read())
c = pyme.core.Context()
c.set_armor(1)
c.set_passphrase_cb(Callback)
c.op_decrypt( cipher, plain )
plain.seek(0,0)
print plain.read()
When I run this and don't provide the password interactively the program then
tries the Callback printing 'in passphrase callback' but then fails with
error:
pyme.errors.GPGMEError: Invocation of gpgme_op_decrypt: Unspecified source: General error (0,1)
First and foremost, why does the passphrase callback not work? And secondly,
how can I prevent the program from prompting the user for a password before
calling the passphrase callback?
This is running on Ubuntu 10.04
Answer: apparently, you need to interpret the keyword `hook`:
def Callback( x, y, z, hook=None):
...
works perfectly well.
|
how to shift a datetime object by 12 hours in python
Question: Datetime objects hurt my head for some reason. I am writing to figure out how
to shift a date time object by 12 hours. I also need to know how to figure out
if two date time object's differ by say 1 minute or more.
Answer: The [`datetime`](http://docs.python.org/library/datetime.html) library has a
`timedelta` object specifically for this kind of thing:
import datetime
mydatetime = datetime.now() # or whatever value you want
twelvelater = mydatetime + datetime.timedelta(hours=12)
twelveearlier = mydatetime - datetime.timedelta(hours=12)
difference = abs(some_datetime_A - some_datetime_B)
# difference is now a timedelta object
# there are a couple of ways to do this comparision:
if difference > timedelta(minutes=1):
print "Timestamps were more than a minute apart"
# or:
if difference.total_seconds() > 60:
print "Timestamps were more than a minute apart"
|
Python: why does my list change after I've retrieved it from an object
Question:
Simple question, I've scaled down a problem I'm having where a list which I've
retrieve from an object is changing when I append more data to the object. Not
to the list.
Can anyone help my understand the behavior of python?
class a():
def __init__(self):
self.log = []
def clearLog(self):
del self.log[:]
def appendLog(self, info):
self.log.append(str(info))
def getLog(self):
return self.log
if __name__ == '__main__':
obj = a()
obj.appendLog("Hello")
# get an instance as of this moment....
list = obj.getLog()
print list
obj.appendLog("World")
# print list, BUT we want the instance that was obtained
# before the new appendage.
print list
**OutPut:**
['Hello']
['Hello', 'World']
Answer: When you code
`list = obj.getLog()`
(ignoring -- just for a second -- what a **terrible** idea it is to use
identifiers that shadow builtins!!!) you're saying: "make name `list` refer to
exactly the same object that `obj.getLog()` returns" -- which as we know from
the code for `class a` is `obj.log`. So of course since now you have one list
object with two names, when you alter that object through **either** name, all
alterations will be fully visible from both names, of course -- remember,
there _is_ just **one** object, you're just using multiple names for it! You
never asked for a copy, so of course Python made no copies.
When you want a copy, instead of the original, **ask for one**! When you know
the type you require (here, a list), the best way is to call the type, i.e.:
mylist = list(obj.getLog())
This of course becomes impossible if you choose to trample all over the
builtins with your identifiers -- -- which is a good part of why such
identifier choice is a **BAD** idea (I can't stress that enough: it's hard to
think of any worse style choice, to use in your Python coding, than such
naming). So, I've renamed the identifier to `mylist` (and of course you need
to rename it in the two `print` statements).
You **could** use highly unreadable or slower approaches to make up for the
wanton destruction of the normal functionality of built-in identifier `list`,
of course -- e.g.:
import copy
list = copy.copy(obj.getLog()) # somewhat slower
or
list = obj.getLog()[:] # traditional, but ECCH
or
temp = obj.getLog()
list = type(temp)(temp) # abstruse
but **BY FAR** the simplest, cleanest, most recommended approach is to **NOT**
name your identifiers the same as Python built-ins (it's also a nice idea to
avoid naming them just like modules in the standard Python library, for
similar though a bit weaker reasons).
|
How to integrate Sikuli scripts into Selenium?
Question: I'm extensively using [Selenium](http://seleniumhq.org/) for integration
testing. Works great for all normal stuff (HTML/AJAX), but no go when I'm
trying to test third party ActiveX, Java applets and Flash components.
The solution I've found for this is
[Sikuli](http://groups.csail.mit.edu/uid/sikuli/). Works great locally, but
how can I integrate that into Selenium?
btw. if that's relevant, for Selenium I'm using Python API.
Answer: See the Python section of the Selenium RC documentation:
<http://seleniumhq.org/docs/05_selenium_rc.html#python>
You may be able to run Selenium tests from Jython. In that case, you can
simply integrate Selenium scripts into your Sikuli scripts.
Try the following in the Sikuli IDE. You may need to modify the import
statements to point to specific files in the Selenium project.
from selenium import selenium # this will probably need tweaking...
slm = selenium("localhost", 4444, "*firefox", "http://www.google.com/")
slm.start()
# etc...
This guy appears to have had success controlling Selenium from Jython:
<http://adam.goucher.ca/?p=367>
My conclusion: it will probably be easiest to stay in Jython and control
Selenium from there. You could integrate both tools into a single script.
|
How to manipulate the response object in django-piston?
Question: I have some existing python code that uses django-piston which returns a
dictionary as its response. For example:
from piston.handler import BaseHandler
class FooHandler(BaseHandler):
allowed_methods = ('GET',)
@classmethod
def create(self, request):
return { 'foo': 'bar' }
This code works fine, and is serialized into JSON with the appropriate HTTP
header set (I'm assuming this works by some piston magic involving emitters;
for bonus points, feel free to clarify how this behavior works as well, as I'm
still getting to know django-piston).
I need to be able to modify the response in arbitrary ways, e.g. setting
headers, status codes, etc. without using some pre-baked solution designed for
a specific purpose. Is there a convenient way to access the response object in
the context of this code and manipulate it, or has the response object not yet
been created? In order to get access to a response object, will I have to
construct it manually (a la vanilla django), serialize the dictionary, and set
the appropriate headers by hand, and thus lose out on some of the useful magic
of django-piston?
Answer: Every django-piston method returns an HTTPResponse.
When you return that dictionary, django-piston is just serializing it and
sticking it in an HTTPResponse that it has crafted of some variety.
Kind of surprised you didn't pick up on that given that those "return
rc.CREATED" lines in all the django-piston examples in the wiki are just
hyper-simplistic responses with an HTTP code and response body.
Take a look here: <https://bitbucket.org/jespern/django-
piston/src/c4b2d21db51a/piston/utils.py>
at the rc_factory class, which creates a variety of simple HTTPResponse
objects for use with Piston.
At the very least you can observe how they do it, and then craft your own.
But the answer to the essence of your question "can I make a custom
HTTPResponse" is yes, you can. Of course, that's what web servers _do_.
|
How would I go about downloading a file from a submitted link then reuploading to my server for streaming?
Question: I'm working on a project where a user can submit a link to a sound file hosted
on another site through a form. I'd like to download that file to my server
and make it available for streaming. I might have to upload it to Amazon S3.
I'm doing this in Django but I'm new to Python. Can anyone point me in the
right direction for how to do this?
Answer: Here's how I would do it:
1. Create a model like `SoundUpload` like:
class SoundUpload(models.Model):
STATUS_CHOICES = (
(0, 'Unprocessed'),
(1, 'Ready'),
(2, 'Bad File'),
)
uploaded_by = models.ForeignKey(User)
original_url = models.URLField(verify_true=False)
download_url = models.URLField(null=True, blank=True)
status = models.IntegerField(choices=STATUS_CHOICES, default=0)
2. Next create the view w/a `ModelForm` and save the info to the database.
3. Hook up a [post-save signal](http://docs.djangoproject.com/en/dev/ref/signals/#django.db.models.signals.post_save) on the `SoundUpload` model that kicks of a [django-celery](http://pypi.python.org/pypi/django-celery) [Task](http://ask.github.com/celery/userguide/tasks.html). This will ensure that the UI responds while you're processing all the data.
def process_new_sound_upload(sender, **kwargs):
# Bury to prevent circular dependency issues.
from your_project.tasks import ProcessSoundUploadTask
if kwargs.get('created', False):
instance = kwargs.get('instance')
ProcessSoundUploadTask.delay(instance.id)
post_save.connect(process_new_sound_upload, sender=SoundUpload)
4. In the `ProcessSoundUploadTask` task you'll want to:
* Lookup the model object based on the passed in id.
* Using [`pycurl`](http://pycurl.sourceforge.net/) download the file to a temporary folder (w/very limitied permissions).
* Use [`ffmpeg`](http://ffmpeg.org/) (or similar) to ensure it's a real sound file. Do any other virus style checks here (depends on how much you trust your users). If it turn out to be a bad file set the `SoundUpload`.status field to `2` (Bad File), save it, and return to stop processing the task. Perhaps send out an email here.
* Use [`boto`](http://code.google.com/p/boto/) to upload the file to s3. See [this](http://ferrouswheel.me/2009/12/upload-a-file-to-s3-with-boto/) example.
* Update the `SoundUpload`.download_url to be the s3 url, the status to be "processed" and save the object.
* Do any other post-processing (sending notification emails, etc.)
The key to this approach is using `django-celery`. Once the task is kicked off
through the post_save signal the UI can return, thus creating a very "snappy"
experience. This task gets put onto an
[AMQP](http://www.amqp.org/confluence/display/AMQP/Advanced+Message+Queuing+Protocol)
message queue that can be processed by multiple workers (dedicated EC2
instances, etc.), so you'll be able to scale without too much trouble. This
may seem like a bit overkill, but it's really not as much work as it seems.
|
Run custom Django management command over SSH
Question: I have a Django application with a custom management command in one of the
apps. If I log in over SSH I can run it without any problems with
> python manage.py sitedir/mycommand
However, if I try to run the command as a oneliner from my local box like
this:
> ssh myserver python manage.py sitedir/mycommand
I get an ImportError like this:
Traceback (most recent call last):
File "mysite/manage.py", line 11, in <module>
execute_manager(settings)
File "/usr/lib/python2.5/site-packages/django/core/management/__init__.py", line 438, in execute_manager
utility.execute()
File "/usr/lib/python2.5/site-packages/django/core/management/__init__.py", line 379, in execute
self.fetch_command(subcommand).run_from_argv(self.argv)
File "/usr/lib/python2.5/site-packages/django/core/management/__init__.py", line 261, in fetch_command
klass = load_command_class(app_name, subcommand)
File "/usr/lib/python2.5/site-packages/django/core/management/__init__.py", line 67, in load_command_class
module = import_module('%s.management.commands.%s' % (app_name, name))
File "/usr/lib/python2.5/site-packages/django/utils/importlib.py", line 35, in import_module
__import__(name)
ImportError: No module named mysite.myapp.management.commands.mycommand
The real reason is that I want to run this admin command from a Fabric script
but until I can get it to run via the ssh one-liner I guess it will be
impossible. Is there something in the environment that differs when you run it
via the ssh one-liner? The python path seems correct in both cases.
Answer: I think I have a clue what is going on. I don't know how to fix it yet.
To reproduce your scenario I wrote a small script.
#!/usr/bin/python
import sys, django
print django.VERSION
After which I executed it after logging in through SSH as well as remotely
(`ssh yourserver.com "python /home/me/script.py"`) and everything went fine.
Then I changed the script.
#!/usr/bin/python
import os
print os.environ['DJANGO_SETTINGS_MODULE']
This version worked when I logged in to the server but **failed** when I tried
to execute it remotely.
Traceback (most recent call last):
File "/home/me/script.py", line 3, in <module>
print os.environ['DJANGO_SETTINGS_MODULE']
File "/usr/lib/python2.5/UserDict.py", line 22, in __getitem__
raise KeyError(key)
KeyError: 'DJANGO_SETTINGS_MODULE'
Apparently the `DJANGO_SETTINGS_MODULE` environment variable is not set when
you execute the command remotely over SSH. I suspect this _**may**_ be what is
going wrong in your case. You will need to figure out how to make sure that
this variable is properly set before executing the script.
~~Perhaps you can explicitly set it:`os.environ['DJANGO_SETTINGS_MODULE'] =
'foo'`.~~.
Try this:
ssh yourserver.com "python /home/me/script.py" -t
DJANGO_SETTINGS_MODULE=app.settings.custom
|
Why is the Python script unreliable when run from rc.local on first boot?
Question: The script below works great when logged in as root and run from the command
line, but when run at first boot using /etc/rc.local in Ubuntu 10.04, it fails
about 25% of the time- the system root, mysql root and some mysql user
passwords are set correctly, but one will fail with console log reporting
standard mysql login error: "ERROR 1045 (28000): Access denied for user 'root'
@ 'localhost' (using password: YES)"
Is there something about running python scripts from init jobs that I should
account for, such as an environment variable?
#!/usr/bin/env python
# Randomizes and outputs to files the system root and mysql user passwords
files = ['/home/ubuntu/passwords','/opt/data1/alfresco/extensions/
extension/alfresco-global.properties','/opt/data/etc/mysql/
debian.cnf','/home/ubuntu/duncil']
userpasswords = {'root':'ROOTPASSWORD'}
mysqlpasswords =
{'root':'MYSQLPASSWORD','alfresco':'alfrescoPASSWORD','debian-sys-
maint':'debian-sys-maintPASSWORD'}
otherpasswords = ['OTHERPASSWORD']
log = '/var/log/firstrun'
import random, string
import crypt
import re
from subprocess import PIPE, Popen
def getsalt(chars = string.letters + string.digits):
# generate a random 2-character 'salt'
return random.choice(chars) + random.choice(chars)
def getpwd(chars = string.letters + string.digits, len = 12):
retval = "";
for i in range(0, len):
# generate 12 character alphanumeric password
retval += random.choice(chars)
return retval
def replace_pass(filename):
handle = open(filename, 'r')
hbuf = handle.read()
handle.close()
for placeholder, password in pdict.iteritems():
hbuf = re.sub(placeholder, password, hbuf)
try:
# Output file
handle = open(filename, 'w')
handle.write(hbuf)
handle.close()
except:
pass
#logh.write('failed to update ' + filename + "\n")
#logh.write('maybe you don\'t have permision to write to it?\n')
logh = open(log, "a")
logh.write("Starting...\n")
# Generate passwords
pdict = {}
for user, placeholder in userpasswords.iteritems():
syspass = getpwd()
Popen(['usermod', '--password', crypt.crypt(syspass, getsalt()), user])
logh.write(placeholder + ": User " + user + " --> " + syspass + "\n")
pdict[placeholder] = syspass
# Whats the MySQL Root password placeholder?
mplace = mysqlpasswords['root']
for user, placeholder in mysqlpasswords.iteritems():
mpass = getpwd()
if (("root" in mysqlpasswords) and (mysqlpasswords['root'] in pdict)):
mrootpass = pdict[mysqlpasswords['root']]
else:
mrootpass = ""
Popen(['mysql', '-uroot', "--password=" + mrootpass, "-e", "UPDATE user SET Password = PASSWORD('" + mpass + "') WHERE User = '" + user + "';FLUSH PRIVILEGES;","mysql"])
logh.write(placeholder + ": MySQL " + user + " --> " + mpass + "\n")
pdict[placeholder] = mpass
for placeholder in otherpasswords:
opass = getpwd()
logh.write(placeholder + ": " + opass + "\n")
pdict[placeholder] = opass
# Update passwords
for file in files:
logh.write("Replacing placeholders in " + file + "\n")
replace_pass(file)
logh.write("Finished\n")
logh.close
Answer: Doesn't `Popen` execute asynchronously?
It seems that during boot, the load is high and you are getting a race
condition between setting the root password and using it to set the next
password (next command).
Try
p = Popen(['mysql', '-uroot', "--password=" + mrootpass, "-e", "UPDATE user SET Password = PASSWORD('" + mpass + "') WHERE User = '" + user + "';FLUSH PRIVILEGES;","mysql"])
p.wait()
and see if that does it.
|
Python bizarre class problem
Question: I have the following piece of code where I try to override a method:
import Queue
class PriorityQueue(Queue.PriorityQueue):
def put(self, item):
super(PriorityQueue, self).put((item.priority, item))
However, when I run it I get `TypeError` exception:
super() argument 1 must be type, not classobj
What is the problem?
Answer: `Queue.PriorityQueue` is not a new-style class, and `super` [only works with
new-style classes](http://docs.python.org/library/functions.html#super). You
must use
import Queue
class PriorityQueue(Queue.PriorityQueue):
def put(self, item):
Queue.PriorityQueue.put(self,(item.priority, item))
instead.
|
How do I add basic authentication to a Python REST request?
Question: I have the following simple Python code that makes a simple post request to a
REST service -
params= { "param1" : param1,
"param2" : param2,
"param3" : param3 }
xmlResults = urllib.urlopen(MY_APP_PATH, urllib.urlencode(params)).read()
results = MyResponseParser.parse(xmlResults)
The problem is that the url used to call the REST service will now require
basic authentication (username and password). How can I incorporate a username
and password / basic authentication into this code, as simply as possible?
Answer: If basic authentication = HTTP authentication, use this:
import urllib
import urllib2
username = 'foo'
password = 'bar'
passman = urllib2.HTTPPasswordMgrWithDefaultRealm()
passman.add_password(None, MY_APP_PATH, username, password)
authhandler = urllib2.HTTPBasicAuthHandler(passman)
opener = urllib2.build_opener(authhandler)
urllib2.install_opener(opener)
params= { "param1" : param1,
"param2" : param2,
"param3" : param3 }
xmlResults = urllib2.urlopen(MY_APP_PATH, urllib.urlencode(params)).read()
results = MyResponseParser.parse(xmlResults)
If not, use `mechanize` or `cookielib` to make an additional request for
logging in. But if the service you access has an XML API, this API surely
includes auth too.
**2016 edit:** By all means, use the [requests](http://docs.python-
requests.org/en/latest/) library! It provides all of the above in a single
call.
|
IronPython, importing Modules
Question: I follow the example from the best answer here to a T, compiling with Pyc.py.
<http://stackoverflow.com/questions/2139202/build-python-scripts-and-call-
methods-from-c>
I get an exception at `pyScope = pyEngine.ImportModule("MyClass");`
no module named MyClass
I believe this to be a bug as sometimes recompilation with Pyc.py will produce
a dll ImportModule recognizes, but other times it doesn't.
CONCLUSION: As noted below by digEmAll, compiling modules with Pyc.py to be
used in this fashion produces random results. Call clr.CompileModules manually
instead.
Answer: OK, I got it.
The module name is the (case sensitive) name of the original .py module, not
the compiled dll.
I mean, if your original module name was `myClass.py`, then you compiled it in
`MyClass.dll`, you must `ImportModule("myClass")` not
`ImportModule("MyClass")`
* * *
**EDIT:**
the previous code refers to the following compile method:
import clr
clr.CompileModules("CompiledScript.dll", "script.py")
On the contrary, using `pyc.py`, the generated dll contains a module called
`__main__` instead of the `.py` file name.
That's very strange...
IIRC, in python a module call itself `__main__` if it's running standalone
(i.e. not called by another), but I still don't grasp the connection...
|
How to clear cookies in WebKit?
Question: i'm currently working with PyWebKitGtk in python
(http://live.gnome.org/PyWebKitGtk). I would like to clear all cookies in my
own little browser. I found interesting method
webkit.HTTPResponse.clearCookies() but I have no idea how to lay my hands on
instance of HTTPResponse object :/
I wouldn't like to use java script for that task.
Answer: If you look at the current state of the bindings on
[GitHub](https://github.com/jmalonzo/pywebkitgtk/), you'll see PyWebKitGTK
[doesn't yet
provide](https://github.com/jmalonzo/pywebkitgtk/blob/master/webkit/webkit-1.1-types.defs)
quite what you want- there's not mapping for the `HTTPResponse` type it looks
like. Unfortunately, I think Javascript or a proxy are your only options right
now.
EDIT:
...unless, of course, you want it _real_ bad and stay up into the night
learning ctypes. In which case, you can do magic. To clear _all_ the browser's
cookies, try this.
import gtk, webkit, ctypes
libwebkit = ctypes.CDLL('libwebkit-1.0.so')
libgobject = ctypes.CDLL('libgobject-2.0.so')
libsoup = ctypes.CDLL('libsoup-2.4.so')
v = webkit.WebView()
#do whatever it is you do with WebView...
....
#get the cookiejar from the default session
#(assumes one session and one cookiesjar)
generic_cookiejar_type = libgobject.g_type_from_name('SoupCookieJar')
cookiejar = libsoup.soup_session_get_feature(session, generic_cookiejar_type)
#build a callback to delete cookies
DEL_COOKIE_FUNC = ctypes.CFUNCTYPE(None, ctypes.c_void_p)
def del_cookie(cookie):
libsoup.soup_cookie_jar_delete_cookie(cookiejar, cookie)
#run the callback on all the cookies
cookie_list = libsoup.soup_cookie_jar_all_cookies(cookiejar)
libsoup.g_slist_foreach(cookie_list, DEL_COOKIE_FUNC(del_cookie), None)
EDIT:
Just started needing this myself, and while it's the right idea it needed
work. Instead, try this- the function type and cookiejar access are fixed.
#add a new cookie jar
cookiejar = libsoup.soup_cookie_jar_new()
#uncomment the below line for a persistent jar instead
#cookiejar = libsoup.soup_cookie_jar_text_new('/path/to/your/cookies.txt',False)
libsoup.soup_session_add_feature(session, cookiejar)
#build a callback to delete cookies
DEL_COOKIE_FUNC = ctypes.CFUNCTYPE(ctypes.c_int, ctypes.c_void_p, ctypes.c_void_p)
def del_cookie(cookie, userdata):
libsoup.soup_cookie_jar_delete_cookie(cookiejar, cookie)
return 0
#run the callback on all the cookies
cookie_list = libsoup.soup_cookie_jar_all_cookies(cookiejar)
libsoup.g_slist_foreach(cookie_list, DEL_COOKIE_FUNC(del_cookie), None)
Note that you should only do this _before_ using the WebView, or maybe in
WebKit callbacks, or you will have threading issues above and beyond those
usually associated with GTK programming.
|
py2exe problems
Question:
c:\python26\setup.py py2exe
Trying to run py2exe and when I get to command prompt I run the line above.
However as opposed to converting my file it try's to open it. What am I doing
wrong?
Answer: You must create your own `setup.py` and then run it with py2exe:
c:\my_python_scripts>python setup.py py2exe
In your `setup.py` you import `distutils`, `py2exe` and show names of your
scripts to compile. There is template for it. Then I usually create `.bat`
file which compiles my scripts.
Have you read [py2exe tutorial](http://www.py2exe.org/index.cgi/Tutorial)?
|
Propagating Clips Error Messages in PyClips
Question: I'm finding it very difficult to develop with PyClips, because it appears to
replace useful error messages thrown by Clips with a generic "syntax error"
message. This makes debugging very laborious and practically impossible on
large codebases when using PyClips.
Consider the following example. I wrote a very large expression, which
contained the multiplication operator, but I mistakenly forgot to add the
second argument. Instead of simply telling I was missing an argument, PyClips
told me there was a syntax error. What should have taken me 1 second to
correct, took me 5 minutes to correct as I hunted through my large expression,
looking for the mistake.
Here's a condensed version:
In Clips, with a useful error message:
clips
CLIPS> (defrule myrule "" (myfact 123) => (bind ?prob (* (min 1 2))))
[ARGACCES4] Function * expected at least 2 argument(s)
ERROR:
(defrule MAIN::myrule ""
(myfact 123)
=>
(bind ?prob (* (min 1 2))
And in PyClips, with an unuseful error message:
python
>>> import clips
>>> clips.BuildRule('myrule','(myfact 123)','(bind ?prob (* (min 1 2)))','')
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/local/lib/python2.6/dist-packages/clips/_clips_wrap.py", line 2839, in BuildRule
_c.build(construct)
_clips.ClipsError: C08: syntax error, or unable to parse expression
How can I get PyClips to give me the _real_ error thrown by Clips?
Answer: Catch the ClipsError, then read ErrorStream for the details. For example:
engine = clips.Environment()
engine.Reset()
engine.Clear()
try:
engine.Load(os.path.abspath(rule_file))
except clips.ClipsError:
logging.error(clips.ErrorStream.Read())
|
python 2.6.x theading / signals /atexit fail on some versions?
Question: I've seen a lot of questions related to this... but my code _works_ on python
2.6.2 and _fails_ to work on python 2.6.5. Am I wrong in thinking that the
whole atexit "functions registered via this module are not called when the
program is killed by a signal" thing shouldn't count here because I'm catching
the signal and then exiting cleanly? What's going on here? Whats the proper
way to do this?
import atexit, sys, signal, time, threading
terminate = False
threads = []
def test_loop():
while True:
if terminate:
print('stopping thread')
break
else:
print('looping')
time.sleep(1)
@atexit.register
def shutdown():
global terminate
print('shutdown detected')
terminate = True
for thread in threads:
thread.join()
def close_handler(signum, frame):
print('caught signal')
sys.exit(0)
def run():
global threads
thread = threading.Thread(target=test_loop)
thread.start()
threads.append(thread)
while True:
time.sleep(2)
print('main')
signal.signal(signal.SIGINT, close_handler)
if __name__ == "__main__":
run()
python 2.6.2:
$ python halp.py
looping
looping
looping
main
looping
main
looping
looping
looping
main
looping
^Ccaught signal
shutdown detected
stopping thread
python 2.6.5:
$ python halp.py
looping
looping
looping
main
looping
looping
main
looping
looping
main
^Ccaught signal
looping
looping
looping
looping
...
looping
looping
Killed <- kill -9 process at this point
The main thread on 2.6.5 appears to never execute the atexit functions.
Answer: The root difference here is actually unrelated to both signals and atexit, but
rather a change in the behavior of `sys.exit`.
Before around 2.6.5, `sys.exit` (more accurately, SystemExit being caught at
the top level) would cause the interpreter to exit; if threads were still
running, they'd be terminated, just as with POSIX threads.
Around 2.6.5, the behavior changed: the effect of `sys.exit` is now
essentially the same as returning from the main function of the program. When
you do _that_ \--in both versions--the interpreter waits for all threads to be
joined before exiting.
The relevant change is that `Py_Finalize` now calls
`wait_for_thread_shutdown()` near the top, where it didn't before.
This behavioral change seems incorrect, primarily because it no longer
functions as documented, which is simply: "Exit from Python." The practical
effect is no longer to exit from Python, but simply to exit the thread. (As a
side note, `sys.exit` has never exited Python when called from another thread,
but that obscure divergance from documented behavior doesn't justify a much
bigger one.)
I can see the appeal of the new behavior: rather than two ways to exit the
main thread ("exit and wait for threads" and "exit immediately"), there's only
one, as sys.exit is essentially identical to simply returning from the top
function. However, it's a breaking change and diverges from documented
behavior, which far outweighs that.
Because of this change, after `sys.exit` from the signal handler above, the
interpreter sits around waiting for threads to exit and then runs `atexit`
handlers after they do. Since it's the handler itself that tells the threads
to exit, the result is a deadlock.
|
Python set error reporting level like in PHP
Question: How can I set error reporting and warning outputs in Python like in PHP
`error_reporting(E_LEVEL)`?
Answer: A vaguely related option might be the setting of level in the
[`logging`](http://docs.python.org/library/logging.html) module of the Python
standard library, and I quote from Python's docs:
import logging
LOG_FILENAME = 'example.log'
logging.basicConfig(filename=LOG_FILENAME,level=logging.DEBUG)
logging.debug('This message should go to the log file')
That `level=` determines which logging messages are emitted and which ones are
filtered. However this only applies to errors (and other messages) emitted
through `logging` module functions, **not** to (e.g) tracebacks resulting from
exceptions; if you want to control the latter (what kinds of message come out
when the process dies by propagating an exception), you can build something
based on
[sys.excepthook](http://docs.python.org/library/sys.html#sys.excepthook), but
your degrees of freedom will still be somewhat limited (in particular, after
the reporting -- abundant or scarce as it may be -- the process _will_ exit if
an exception has propagated to that point).
|
Efficient way to store dictionary (hash) in file with python?
Question: I'm implementing a Unix userland tool that needs to store a hash on the disk.
The hash will be read every run of the program, pretty frequently. The hash
needs to store "name:path" values only.
I looked at the [bsddb](https://docs.python.org/2/library/bsddb.html) standard
library module for python, but I can see it will be deprecated in Python 3. I
also saw the [pickle](https://docs.python.org/3/library/pickle.html) standard
library module.
I'm not a python guy, so what is the efficient way for hash serialization and
frequent open/read/close operations?
Answer: I would start with the [shelve](http://docs.python.org/library/shelve.html)
module and see if that isn't too slow. It does exactly what you want.
import shelve
d = shelve.open('filename')
d['name'] = 'path'
d.close()
or to read from it
d = shelve.open('filename')
d = hash['name']
It's essentially a wrapper around pickle that provides a dictionary
abstraction.
|
Python ssl problem with multiprocessing
Question: I want to send data from a client to the server in a TLS TCP socket from
multiple client subprocesses so I share the same ssl socket with all
subprocesses. Communication works with one subprocess, but if I use more than
one subprocesses, the TLS server crashes with an `ssl.SSLError`
(SSL3_GET_RECORD:decryption failed or bad record mac).
**More specific:** It does not depend which process first calls the
`SSLSocket.write()` method, but this process is the only one from this time on
which can call it. If another process calls `write()`, the server will result
in the exception described above.
I used this basic code:
**tlsserver.py**
import socket, ssl
def deal_with_client(connstream):
data = connstream.read()
while data:
print data
data = connstream.read()
connstream.close()
bindsocket = socket.socket()
bindsocket.bind(('127.0.0.1', 9998))
bindsocket.listen(5)
while True:
newsocket, fromaddr = bindsocket.accept()
connstream = ssl.wrap_socket(newsocket,
server_side=True,
certfile="srv.crt",
keyfile="srv.key",
ssl_version=ssl.PROTOCOL_TLSv1)
deal_with_client(connstream)
**tlsclient.py**
import socket, ssl
import multiprocessing
class SubProc:
def __init__(self, sock):
self.sock = sock
def do(self):
self.sock.write("Test")
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
ssl_sock = ssl.wrap_socket(s)
ssl_sock.connect(('127.0.0.1', 9998))
print "Connected to", repr(ssl_sock.getpeername())
for x in (1,2):
subproc = SubProc(ssl_sock)
proc = multiprocessing.Process(target=subproc.do)
And this is the backtrace:
Traceback (most recent call last):
File "tlsserver.py", line 21, in <module>
deal_with_client(connstream)
File "tlsserver.py", line 7, in deal_with_client
data = connstream.read()
File "/usr/lib64/python2.6/ssl.py", line 136, in read
return self._sslobj.read(len)
ssl.SSLError: [Errno 1] _ssl.c:1325: error:1408F119:SSL routines:SSL3_GET_RECORD:decryption failed or bad record mac
Answer: The problem is that you're re-using the same connection for both processes.
The way SSL encrypts data makes this fail -- the two processes would have to
communicate with each other about the state of the shared SSL connection. Even
if you do make it work, or if you didn't use SSL, the data would arrive at the
server all jumbled up; you would have no real way of distinguishing which
bytes came from which process.
What you need to do is give each process its own SSL connection, by making the
connection in `subproc.do`. Alternatively, don't have the subprocesses
communicate with the server at all, but rather communicate with the main
process, and have the main process relay it over the SSL connection.
|
Using Python quick insert many columns into Sqlite\Mysql
Question: If Newdata is list of x columns, How would get the number unique columns--
number of members of first tuple. (Len is not important.) Change the number of
"?" to match columns and insert using the statement below.
csr = con.cursor()
csr.execute('Truncate table test.data')
csr.executemany('INSERT INTO test.data VALUES (?,?,?,?)', Newdata)
con.commit()
Answer: By "Newdata is list of x columns", I imagine you mean `x` **tuples** , since
then you continue to speak of "the first tuple". If `Newdata` is a list of
tuples, `y = len(Newdata[0])` is the number of items in the first one of those
tuples.
Assuming that's the number you want (and all tuples had better have the same
number of items, otherwise `executemany` _will_ fail!), the general idea in
@Nathan's answer is right: build the string with the appropriate number of
comma-separated question marks:
holders = ','.join('?' * y)
then insert it in the rest of the SQL statement. @Nathan's way to insert is
right for most Python 2.any versions, but if you have 2.6 or better,
sql = 'INSERT INTO testdata VALUES({0})'.format(holders)
is currently preferred (it also works in Python 3.any).
Finally,
csr.executemany(sql, Newdata)
will do what you desire. Remember to commit the transaction once you're
done!-)
|
Prototype for python?
Question: I just learned Prototype for Javascript. It's super convenient: using the $
shortcut, accessing xml elements is not painful any more!
The question: is there a Prototype-like extension for Python?
Answer: Python has [lxml](http://codespeak.net/lxml/) which has the `xpath` method
wherein you could use xpath expressions to select elements. As I understand
it, $ in prototype searches and returns an element that has a particular id,
in which case could be translated in xpath to `*[@id=<someid>]` like so:
>>> import lxml.etree
>>> tree = lxml.etree.XML("<root><a id='1'/><b id='2'/></root>")
>>> tree.xpath("*[@id=1]")
[<Element a at c3bc30>]
>>> lxml.etree.tostring(tree.xpath("*[@id=1]")[0])
'<a id="1"/>'
I think the Python standard library includes support for a subset of xpath in
ElementTree too so you might be able to implement that there somehow if you do
not wish to install lxml (which isn't included in stdlib)...
|
Python subprocess timeout?
Question: Is there any argument or options to setup a timeout for Python's
subprocess.Popen method?
Something like this:
`subprocess.Popen(['..'], ..., timeout=20)` ?
Answer: I would advise taking a look at the [Timer
class](http://docs.python.org/2/library/threading.html#threading.Timer) in the
threading module. I used it to implement a timeout for a Popen.
First, create a callback:
def timeout( p ):
if p.poll() is None:
print 'Error: process taking too long to complete--terminating'
p.kill()
Then open the process:
proc = Popen( ... )
Then create a timer that will call the callback passing the process to it.
t = threading.Timer( 10.0, timeout, [proc] )
t.start()
t.join()
Somewhere later in the program, you may want to add the line:
t.cancel()
Otherwise, the python program will keep running until the timer has finished
running.
EDIT: I was advised that there is a race condition that the subprocess p may
terminate between the p.poll() and p.kill() calls. I believe the following
code can fix that:
import errno
def timeout( p ):
if p.poll() is None:
try:
p.kill()
print 'Error: process taking too long to complete--terminating'
except OSError as e:
if e.errno != errno.ESRCH:
raise
Though you may want to clean the exception handling to specifically handle
just the particular exception that occurs when the subprocess has already
terminated normally.
|
Is this a good reason to check types in Python?
Question: I know that checking types in Python is bad and you should probably never do
it. But I can't seem to find the disadvantage to this.
class O(object):
def __init__(self, name):
'''Can only be called in derived classes.'''
if type(self) is O:
message = "%(class)s cannot be instantiated, it must be derived."
raise TypeError, message % { "class" : O }
self.name = name
def fn(self):
'''Must be populated in derived classes.'''
raise NotImplementedError
Now if someone tries to instantiate O, a class I never meant to be
instantiated, they know immediately.
Is this still bad form?
Answer: Look at [Abstract Base Classes](http://docs.python.org/library/abc.html) as
they will provide more fine grained control over how the subclasses are
instantiated if this is something that you really want to do.
All in all, this might be a valid use because you are not preventing _me_ from
passing whatever _I_ want to your code but I still wouldn't consider it
pythonic. You are telling _me_ that _I_ can't instantiate your class. What If
I want to?
Using ABC's, it looks like:
import abc
class O(object):
__metaclass__ = abc.ABCMeta
def __init__(self, name):
self.name = name
@abc.abstractmethod
def fn(self):
pass
This has the advantage of not breaking super on the `fn` method. As you have
it with the `raise NotImplementedError`, you are breaking super for multiple
inheritance. If a class derives from two classes that subclass `O` and both
call super (as they should to allow for multiple inheritance) then it will
create the exception that you raise.
So now, you are not just telling me that I can't instantiate your class, you
are telling me that I can't use multiple inheritance when subclassing from
your class.
|
Getting pdb in Emacs to use Python process from current virtualenv
Question: I am debugging some python code in emacs using pdb and getting some import
issues. The dependencies are installed in one of my bespoked virtualenv
environments.
Pdb is stubbornly using /usr/bin/python and not the python process from my
virtualenv.
I use virtualenv.el to support switching of environments within emacs and via
the postactivate hooks described in
<http://jesselegg.com/archives/2010/03/14/emacs-python-
programmers-2-virtualenv-ipython-daemon-mode/>
This works well when running M-x python-shell
>>> import sys
>>> print sys.path
This points to all of my virtualenv libraries indicating that the python-shell
is that of my virtualenv.
This is contradicted however by M-! which python, which gives
_/usr/bin/python_
Does anyone know how I can tell M-x pdb to adopt the python process from the
currently active virtualenv?
Answer: `python-shell` uses variable `python-default-interpreter` to determine which
python interpreter to use. When the value of this variable is `cpython`, the
variables `python-python-command` and `python-python-command-args` are
consulted to determine the interpreter and arguments to use. Those two
variables are manipulated by `virtualenv.el` to set the current virtual
environment.
So when you use `python-shell` command, it uses your virtual environments
without any problem.
But, when you do `M-!` `python`, you're not using the variables `python-
python-command` and `python-python-command-args`. So it uses the python tools
it finds in your path.
When you call `M-x` `pdb` it uses gud-pdb-command-name as the default pdb
tool. To redefine this variable, each time you activate an environment, you
could do something like this :
(defadvice virtualenv-activate (after virtual-pdb)
(custom-set-variables
'(gud-pdb-command-name
(concat virtualenv-active "/bin/pdb" ))))
(ad-activate 'virtualenv-activate)
To have pdb in your virtual environment, do the following :
cp /usr/bin/pdb /path/to/virtual/env/bin
Then edit the first line of /path/to/virtual/env/bin/pdb to have :
#! /usr/bin/env python
Reactivate your env and Pdb should now use your virtualenv python instead of
the system-wide python.
|
How do you convert a stringed dictionary to a Python dictionary?
Question: I have the following string which is a Python dictionary stringified:
some_string = '{123: False, 456: True, 789: False}'
How do I get the Python dictionary out of the above string?
Answer: Use
[**`ast.literal_eval`**](http://docs.python.org/library/ast.html#ast.literal_eval):
> Safely evaluate an expression node or a string containing a Python
> expression. The string or node provided may only consist of the following
> Python literal structures: strings, numbers, tuples, lists, dicts, booleans,
> and None.
>
> This can be used for safely evaluating strings containing Python expressions
> from untrusted sources without the need to parse the values oneself.
Example:
>>> some_string = '{123: False, 456: True, 789: False}'
>>> import ast
>>> ast.literal_eval(some_string)
{456: True, 123: False, 789: False}
|
i got this error: "ImportError: cannot import name python" How do I fix it?
Question:
File "G:\Python25\Lib\site-packages\PyAMF-0.6b2-py2.5-win32.egg\pyamf\util\__init__.py", line 15, in <module>
ImportError: cannot import name python
How do I fix it? If you need any info to know how to fix this problem, I can
explain, just ask.
Thanks
Code:
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.ext import webapp
from TottysGateway import TottysGateway
import logging
def main():
services_root = 'services'
#services = ['users.login']
#gateway = TottysGateway(services, services_root, logger=logging, debug=True)
#app = webapp.WSGIApplication([('/', gateway)], debug=True)
#run_wsgi_app(app)
if __name__ == "__main__":
main()
Code:
from pyamf.remoting.gateway.google import WebAppGateway
import logging
class TottysGateway(WebAppGateway):
def __init__(self, services_available, root_path, not_found_service, logger, debug):
# override the contructor and then call the super
self.services_available = services_available
self.root_path = root_path
self.not_found_service = not_found_service
WebAppGateway.__init__(self, {}, logger=logging, debug=True)
def getServiceRequest(self, request, target):
# override the original getServiceRequest method
try:
# try looking for the service in the services list
return WebAppGateway.getServiceRequest(self, request, target)
except:
pass
try:
# don't know what it does but is an error for now
service_func = self.router(target)
except:
if(target in self.services_available):
# only if is an available service import it's module
# so it doesn't access services that should be hidden
try:
module_path = self.root_path + '.' + target
paths = target.rsplit('.')
func_name = paths[len(paths) - 1]
import_as = '_'.join(paths) + '_' + func_name
import_string = "from "+module_path+" import "+func_name+' as service_func'
exec import_string
except:
service_func = False
if(not service_func):
# if is not found load the default not found service
module_path = self.rootPath + '.' + self.not_found_service
import_string = "from "+module_path+" import "+func_name+' as service_func'
# add the service loaded above
assign_string = "self.addService(service_func, target)"
exec assign_string
return WebAppGateway.getServiceRequest(self, request, target)
Answer: You need to post your full traceback. What you show here isn't all that
useful. I ended up digging up line 15 of pyamf/util/**init**.py. The code you
should have posted is
from pyamf import python
This should not fail unless your local environment is messed up.
Can you 'import pyamf.util' and 'import pyamf.python' in a interactive Python
shell? What about if you start Python while in /tmp (on the assumption that
you might have a file named 'pyamf.py' in the current directory. Which is a
bad thing.)
= (older comment below) =
Fix your question. I can't even tell where line 15 of util/__init__.py is
supposed to be. Since I can't figure that out, I can't answer your question.
Instead, I'll point out ways to improve your question and code.
First, use the markup language correctly, so that all the code is in a code
block. Make sure you've titled the code, so we know it's from util/__init__.py
and not some random file.
In your error message, include the _full_ traceback, and not the last two
lines.
Stop using parens in things like "if(not service_func):" and use a space
instead, so its " if not service_func:". This is discussed in [PEP
8](http://www.python.org/dev/peps/pep-0008/).
Read the Python documentation and learn how to use the language. Something
like "func_name = paths[len(paths) - 1]" should be "func_name = paths[-1]"
Learn about the
[**import**](http://docs.python.org/library/functions.html#__import__)
function and don't use "exec" for this case. Nor do you need the "exec
assign_string" -- just do the "self.addService(service_func, target)"
|
WxPython - Resize WxFrame when adding new content?
Question: Pretty much exactly as it sounds. I have buttons in a `Wx.Frame` that are
created on the fly and I'd like the parent frame to increase in height as I
add new buttons. The height is already being acquire from the total number of
buttons multiplied by an integer equal the each button's height, but I don't
know how to get the frame to change size based on that when new buttons are
added.
As a side question the current method I have for updating the buttons creates
a nasty flicker and I was wondering if anyone had any ideas for fixing that.
import wx
import mmap
import re
class pt:
with open('note.txt', "r+") as note:
buf = mmap.mmap(note.fileno(), 0)
TL = 0
readline = buf.readline
while readline():
TL += 1
readlist = note.readlines()
note.closed
class MainWindow(wx.Frame):
def __init__(self, parent, title):
w, h = wx.GetDisplaySize()
self.x = w * 0
self.y = h - bdepth
self.container = wx.Frame.__init__(self, parent, title = title, pos = (self.x, self.y), size = (224, bdepth), style = wx.STAY_ON_TOP)
self.__DoButtons()
self.Show(True)
def __DoButtons(self):
for i, line in enumerate(pt.readlist):
strip = line.rstrip('\n')
todo = strip.lstrip('!')
self.check = re.match('!', strip)
self.priority = re.search('(\!$)', strip)
if self.check is None and self.priority is None:
bullet = wx.Image('bullet.bmp', wx.BITMAP_TYPE_BMP)
solid = wx.EmptyBitmap(200,64,-1)
dc = wx.MemoryDC()
dc.SelectObject(solid)
solidpen = wx.Pen(wx.Colour(75,75,75),wx.SOLID)
dc.SetPen(solidpen)
dc.DrawRectangle(0, 0, 200, 64)
dc.SetTextForeground(wx.Colour(255, 255, 255))
dc.DrawBitmap(wx.BitmapFromImage(bullet, 32), 10, 28)
dc.DrawText(todo, 30, 24)
dc.SelectObject(wx.NullBitmap)
hover = wx.EmptyBitmap(200,64,-1)
dc = wx.MemoryDC()
dc.SelectObject(hover)
hoverpen = wx.Pen(wx.Colour(100,100,100),wx.SOLID)
dc.SetPen(hoverpen)
dc.DrawRectangle(0, 0, 200, 64)
dc.SetTextForeground(wx.Colour(255, 255, 255))
dc.DrawBitmap(wx.BitmapFromImage(bullet, 32), 10, 28)
dc.DrawText(todo, 30, 24)
dc.SelectObject(wx.NullBitmap)
bmp = solid
elif self.priority is None:
checkmark = wx.Image('check.bmp', wx.BITMAP_TYPE_BMP)
checked = wx.EmptyBitmap(200,64,-1)
dc = wx.MemoryDC()
dc.SelectObject(checked)
checkedpen = wx.Pen(wx.Colour(50,50,50),wx.SOLID)
dc.SetPen(checkedpen)
dc.DrawRectangle(0, 0, 200, 50)
dc.SetTextForeground(wx.Colour(200, 255, 0))
dc.DrawBitmap(wx.BitmapFromImage(checkmark, 32), 6, 24)
dc.DrawText(todo, 30, 24)
dc.SelectObject(wx.NullBitmap)
bmp = checked
else:
exclaim = wx.Image('exclaim.bmp', wx.BITMAP_TYPE_BMP)
important = wx.EmptyBitmap(200,64,-1)
dc = wx.MemoryDC()
dc.SelectObject(important)
importantpen = wx.Pen(wx.Colour(75,75,75),wx.SOLID)
dc.SetPen(importantpen)
dc.DrawRectangle(0, 0, 200, 50)
dc.SetTextForeground(wx.Colour(255, 180, 0))
dc.DrawBitmap(wx.BitmapFromImage(exclaim, 32), 6, 24)
dc.DrawText(todo, 30, 24)
dc.SelectObject(wx.NullBitmap)
importanthover = wx.EmptyBitmap(200,64,-1)
dc = wx.MemoryDC()
dc.SelectObject(importanthover)
importanthoverpen = wx.Pen(wx.Colour(100,100,100),wx.SOLID)
dc.SetPen(importanthoverpen)
dc.DrawRectangle(0, 0, 200, 50)
dc.SetTextForeground(wx.Colour(255, 180, 0))
dc.DrawBitmap(wx.BitmapFromImage(exclaim, 32), 6, 24)
dc.DrawText(todo, 30, 24)
dc.SelectObject(wx.NullBitmap)
bmp = important
b = wx.BitmapButton(self, i + 800, bmp, (10, i * 64), (bmp.GetWidth(), bmp.GetHeight()), style = wx.NO_BORDER)
if self.check is None and self.priority is None:
b.SetBitmapHover(hover)
elif self.priority is None:
b.SetBitmapHover(checked)
else:
b.SetBitmapHover(importanthover)
self.input = wx.TextCtrl(self, -1, "", (16, pt.TL * 64 + 4), (184, 24))
self.Bind(wx.EVT_TEXT_ENTER, self.OnEnter, self.input)
def OnClick(self, event):
button = event.GetEventObject()
button.None
print('cheese')
def OnEnter(self, event):
value = self.input.GetValue()
pt.readlist.append('\n' + value)
self.__DoButtons()
with open('note.txt', "r+") as note:
for item in pt.readlist:
note.write("%s" % item)
note.closed
bdepth = pt.TL * 64 + 32
app = wx.App(False)
frame = MainWindow(None, "Sample editor")
app.SetTopWindow(frame)
app.MainLoop()
Answer: **Don't** double-prefix your methods unless you know what you're doing. This
is not directly related to your question, but it'll result in bugs you won't
understand later.
See this [stackoverflow](http://stackoverflow.com/questions/1301346/the-
meaning-of-a-single-and-a-double-underscore-before-an-object-name-in-python)
question and [the python
documentation](http://docs.python.org/tutorial/classes.html#private-variables)
what/why.
|
Is there some website that has examples of every method in the python standard library?
Question: For example, c++ have
[cplusplus.com/reference](http://cplusplus.com/reference) which contain all of
c++ standard library complete with definitions and more importantly examples,
so I was wondering if there is such a website for python. I know that python
is self documented, like i could use
help(object)
object.__doc__
dir(object)
I know of
[doc.python.org/library](http://doc.python.org/library)
[wiki.python.org](http://wiki.python.org)
But it doesn't have examples of every method. It would be nice if there was
such a website, because when I am learning a new python library I find myself
just testing the methods to see if it does what I want, and it makes my
programming really slow. But this maybe because I have only have 2 years of
programming under my belt. So my question is, is there such a website and is
there a better way to learning a new library in python? Because when learning
a new c++ library, all I need to do is follow by example which makes learning
a new c++ library really easy.
Answer: Try [Python Module of The
Week](http://www.doughellmann.com/PyMOTW/contents.html). It may not be exactly
what your looking for, but you should find it helpfull.
> PyMOTW is a series of blog posts written by Doug Hellmann. It was started as
> a way to build the habit of writing something on a regular basis. The focus
> of the series is building a set of example code for the modules in the
> Python standard library.
|
Python meta-debugging
Question: Heyo,
Just started writing an assembler for the imaginary computer my class is
creating wire-by-wire since the one the TA's provided sucks hard. I chose
python even though I've never really used it that much (but know the basic
syntax) and am loving it.
My favorite ability is how I can take a method I just wrote, paste it into the
shell and then unit test it by hand (I'm using IDLE).
I'm just wondering if there is a way to expose all the symbols in my python
code to the shell automatically, so I can debug without copying and pasting my
code into the shell every time (especially when I make a modification in the
code).
Cheers
Answer: you can import the module that your code is in. This will expose all of the
symbols prefixed with the module name.
The details for the easiest way to do it depend on your operating system but
you can always do:
>>> sys.path.append('/path/to/directory/that/my/module/is/in/')
>>> import mymod #.py
later after you make a change, you can just do
>>>> reload(mymod)
and the symbols will now reference the new values. Note that `from mymod
import foo` will break `reload` in the sense that `foo` will _not_ be updated
after a call to `reload`. So just use `mymod.foo`.
Essentially the trick is to get the directory containing the file on your
`PYTHONPATH` environment variable. You can do this from .bashrc on linux for
example. I don't know how to go about doing it on another operating system. I
use [virualenv](http://pypi.python.org/pypi/virtualenv) with has a [nice
wrapper](http://www.doughellmann.com/projects/virtualenvwrapper/) and `workon`
command so I just have to type `workon foo` and it runs shell scripts (that I
had to write) that add the necessary directories to my python path.
When I was just starting off though, I made one permanent addition to my
`PYTHONPATH` env variable and kept module I wrote in there.
Another alternative is to execute your module with the `-i` option.
$ python -i mymod.py
This will execute the module through to completion and then leave you at the
interpreter. this isn't IDLE though, it's a little rougher but you are now in
your module's namespace (or rather the module's namespace is the global
namespace)
|
Do you think/write differently in vim and emacs?
Question: In many scripts that I write, I often construct programs in a "functional
style". That is to say, I basically define a lot of functions at the
beginning, and then later apply these functions. This results in a sequence of
nested function calls, wherein I type out:
1. function name
2. its arguments
3. next function name
4. its arguments
...and so on.
For cases in which functions are "piped" together, the output of one function
is an argument (usually the first, but not always) to the next function, the
output of which is an argument to the next function, and ad infinitum. In
prefix-notation, the key movements can be very jumpy if you type out this
sequence from left to right. For example, how would you type out the second
line of the following [Python] example (~mul~ is multiply, ~truediv~ is
divide)?
from operator import add, mul, truediv
print(truediv(mul(add(1,1),2),4))
If I had to write the same set of operations linearly (from left to write,
without jumping around), I am more likely to use the notation of function
composition. Building on my previous example in Python, I might write
from functional import foldr, compose, partial, flip
print(foldr(compose,add,(partial(mul,2),partial(flip(truediv),4)))(1,1))
I think this is because I associate each function with its own arguments and
prefer to type them out in succession, rather than filling arguments to
another function before the argument list for the first function is complete
(as would be required to type out the first example from left to right).
I noticed this because I've been an emacs user for a long time and only
recently tried out viper/vimpuse and vim. In emacs, I might do something like
1. [type function name and arguments]
2. C-a
3. [type next function name]
4. C-e
5. [fill in rest of arguments]
6. C-a
7. [type next function name]
8. C-e
9. [fill in rest of arguments]
...and so on, with occasional use of M-b, M-f, M-DEL (backward-word, forward-
word, backward-kill-word) if I mess up or forget something.
I recently found out about C-o in vim, which is a lifesaver - but I find that
the equivalent keys would be
1. [type function name and arguments]
2. C-o 0
3. [type next function name]
4. C-o $
5. [fill in rest of arguments]
6. C-o 0
7. [type next function name]
8. C-o $
9. [fill in rest of arguments]
...and the rest; backward-word, forward-word, and backward-kill-word
equivalents would be C-o b and C-o w, and C-w.
So this got me thinking that to program in vim, I may have to grow a larger
working memory, so that I can pause the construction of one function as I fill
out another, and so on down the stack. Also, in construcing text documents, I
find that I edit (kill, copy, yank) quite frequently even before I finish a
complete thought, which is not so amenable for vim's operational style of
"stay in normal mode, burst of text in insert-mode, and back to normal mode",
which seems to presume that I am capable of producing something worth editing
during my forays into insert-mode. To use vim, I find that I deliberate more
as I type to reduce the frequency of switching between modes. Is this because
I'm naturally spastic, or once I master or commit a suitable range of vim key
commands to muscle memory, I'll stop thinking they're so different?
If you program in both emacs and vim, do you find yourself thinking about and
construcing your programs and blocks of text differently in each editor?
Answer: I used vi from the good old days of 1992 and now I use Emacs since 2001. I
haven't noticed any difference in my thinking when programming functions and
blocks of code. Both editors has their own peculiarities and ways of doing
things, but they are not so strong that they could change your way of thinking
and how you program.
I've always tried to find ways to do what I intend to do. I don't let my
editor force me to do something I don't want. When I do procedural programming
of a new piece of code, I use the technique called "wishful thinking" that is
[mentioned in Structure and Interpretation of Computer
Programs](http://mitpress.mit.edu/sicp/full-text/sicp/book/node28.html) :
You imagine yourself in the perfect world having all the procedures you need
at your disposal. You code your algorithm with all those helpful functions
that you'll need to implement but that you only have prototypes for the
moment. It's similar to a top-down approach.
|
Python programming
Question: My assignment ask to make a function call readFasta that accepts one argument:
the name of a fasta format file (fn) containing one or more sequences. The
function should read the file and return a dictionary where the keys are the
fasta headers and the values are the corresponding sequences from file fn
converted to strings. Make sure that you don’t include any new lines or other
white space characters in the sequences in the dictionary.
For ex, if afile.fa looks like:
>one
atctac
>two
gggaccttgg
>three
gacattac
then the a.readFasta(f) returns:
[‘one’ : ‘atctac’,
‘two’ : ‘gggaccttgg’,
‘three’: ‘gacattac’]
If have tried to write some codes but as I am totally newbie in programming,
it didnt work out very much for me. Can everyone please help me. Thank you so
much. Here are my codes:
import gzip
def readFasta(fn):
if fn.endswith('.gz'):
fh = gzip.gzipfile(fn)
else:
fh = open(fn,'r')
d = {}
while 1:
line = fh.readline()
if not line:
fh.close()
break
vals = line.rstrip().split('\t')
number = vals[0]
sequence = vals[1]
if d.has_key(number):
lst = d[number]
if gene not in lst:
# this test may not be necessary
lst.append(sequence)
else:
d[number] = [sequence]
return d
Here is what I got in my afile.txt
> one atctac
>
> two gggaccttgg
>
> three gacattac
Answer: your post is slightly confusing. I assume that you want it to return a dict.
in that case, you would write it as `{'one': 'actg', 'two': 'aaccttgg' }`. if
you correctly presented the file format, then this function should do the
trick.
import gzip
def read_fasta(filename):
with gzip.open(filename) as f:
return dict(line.split() for line in f)
|
Connect Sphinx autodoc-skip-member to my function
Question: I want to use [sphinx's autodoc-skip-
member](http://sphinx.pocoo.org/ext/autodoc.html#event-autodoc-skip-member)
event to select a portion of the members on a certain python class for
documentation.
But it isn't clear from the sphinx docs, and I can't find any examples that
illustrate: where do I put the code to connect this? I see
[Sphinx.connect](http://sphinx.pocoo.org/ext/appapi.html#sphinx.application.Sphinx.connect)
and I suspect it goes in my conf.py, but when I try variations on this code in
conf.py I can't find the app object that I should connect():
def maybe_skip_member(app, what, name, obj, skip,
options):
print app, what, name, obj, skip, options
return False
# This is not even close to correct:
#from sphinx.application import Sphinx
#Sphinx().connect('autodoc-skip-member', maybe_skip_member)
A pointer to a simple example would be ideal.
Answer: Aha, last ditch effort on a little googling turned up [this
example](http://trac.sagemath.org/sage_trac/attachment/ticket/7813/conf.py),
scroll down to the bottom. Apparently a setup() function in conf.py will get
called with the app. I was able to define the following at the bottom of my
conf.py:
def maybe_skip_member(app, what, name, obj, skip, options):
print app, what, name, obj, skip, options
return True
def setup(app):
app.connect('autodoc-skip-member', maybe_skip_member)
Which is obviously useless (it skips everything), but that's the minimal
example I was looking for and couldn't find...
|
creating xml tree from a textfile with Python
Question: I need to avoid creating double branches in an xml tree when parsing a text
file. Let's say the textfile is as follows (the order of lines is random):
branch1:branch11:message11
branch1:branch12:message12
branch2:branch21:message21
branch2:branch22:message22
So the resulting xml tree should have a root with two branches. Both of those
branches have two subbranches. The Python code I use to parse this textfile is
as follows:
import string
fh = open ('xmlbasic.txt', 'r')
allLines = fh.readlines()
fh.close()
import xml.etree.ElementTree as ET
root = ET.Element('root')
for line in allLines:
tempv = line.split(':')
branch1 = ET.SubElement(root, tempv[0])
branch2 = ET.SubElement(branch1, tempv[1])
branch2.text = tempv[2]
tree = ET.ElementTree(root)
tree.write('xmlbasictree.xml')
The problem with this code is, that a branch in xml tree is created with each
line from the textfile.
Any suggestions how to avoid creating another branch in xml tree if a branch
with this name exists already?
Answer:
with open("xmlbasic.txt") as lines_file:
lines = lines_file.read()
import xml.etree.ElementTree as ET
root = ET.Element('root')
for line in lines:
head, subhead, tail = line.split(":")
head_branch = root.find(head)
if not head_branch:
head_branch = ET.SubElement(root, head)
subhead_branch = head_branch.find(subhead)
if not subhead_branch:
subhead_branch = ET.SubElement(branch1, subhead)
subhead_branch.text = tail
tree = ET.ElementTree(root)
ET.dump(tree)
The logic is simple -- you already stated it in your question! You merely need
to check whether a branch already exists in the tree before creating it.
Note that this is likely inefficient, since you are searching up to the entire
tree for each line. This is because `ElementTree` is not designed for
uniqueness.
* * *
If you require speed (which you may not, especially for smallish trees!), a
more efficient way would be to use a `defaultdict` to store the tree structure
before converting it to an `ElementTree`.
import collections
import xml.etree.ElementTree as ET
with open("xmlbasic.txt") as lines_file:
lines = lines_file.read()
root_dict = collections.defaultdict( dict )
for line in lines:
head, subhead, tail = line.split(":")
root_dict[head][subhead] = tail
root = ET.Element('root')
for head, branch in root_dict.items():
head_element = ET.SubElement(root, head)
for subhead, tail in branch.items():
ET.SubElement(head_element,subhead).text = tail
tree = ET.ElementTree(root)
ET.dump(tree)
|
Is there a way to run a python script that is inside a zip file from bash?
Question: I know there is a way to import modules which are in a zip file with python. I
created kind of custom python package library in a zip file.
I would like to put as well my "task" script in this package, those are using
the library. Then, with bash, I would like to call the desired script in the
zip file without extracting the zip.
The goal is to have only one zip to move in a specified folder when I want to
run my scripts.
Answer: I finally found a way to do this. If I create a zip file, I must create
`__main__.py` at the root of the zip. Thus, it is possible to launch the
script inside the main and call if from bash with the following command :
`python myArchive.zip`
This command will run the `__main__.py` file! :)
Then I can create `.command` file to launch the script with proper parameters.
You can also put some code in the `__main__.py` file to give you more
flexibility if you need to pass arguments for example.
ex: `python __main__.py buildProject`
The reference documentation is here:
<https://docs.python.org/2/library/runpy.html>
|
On Google App Engine (GAE), how do I search on the Key/ID field?
Question: I've got this code (Java, GAE):
// Much earlier:
playerKey = KeyFactory.keyToString(somePlayer.key);
// Then, later...
PersistenceManager pm = assassin.PMF.get().getPersistenceManager();
Key targetKey = KeyFactory.stringToKey(playerKey);
Query query = pm.newQuery(Player.class);
query.setFilter("__key__ == keyParam");
query.declareParameters("com.google.appengine.api.datastore.Key keyParam");
List<Player> players = (List<Player>) query.execute(targetKey); // <-- line 200
which generates this error:
javax.jdo.JDOFatalUserException: Unexpected expression type while parsing query. Are you certain that a field named __key__ exists on your object?
at org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:354)
at org.datanucleus.jdo.JDOQuery.execute(JDOQuery.java:252)
at myapp.Player.validPlayerWithKey(Player.java:200)
// [etc., snip]
But I'm not sure what it wants. I'm trying to search on the JDO id field,
which I I thought I read had the special name `__key__`, [in the
documentation](http://code.google.com/appengine/docs/python/datastore/gqlreference.html).
I've tried it with both
query.setFilter("__key__ == keyParam");
and
query.setFilter("ID == keyParam");
with the same results. So, what am I doing wrong? Or, more importantly, how do
I do it correctly?
Thanks!
**Edit:** For completeness's sake, here is the final, working code (based on
Gordon's answer, which I have accepted as correct):
Player result = null;
if (playerKey == null)
{
log.log(Level.WARNING, "Tried to find player with null key.");
}
else
{
PersistenceManager pm = assassin.PMF.get().getPersistenceManager();
try {
result = (Player) pm.getObjectById(Player.class, playerKey);
} catch (javax.jdo.JDOObjectNotFoundException notFound) {
// Player not found; we will return null.
result = null;
}
pm.close();
}
return result;
Answer: If your objective is to get an object by key, then you should use the
PersistenceManager's getObjectByID() method. More details
[here](http://code.google.com/appengine/docs/java/datastore/creatinggettinganddeletingdata.html#Getting_an_Object_By_Key).
As an aside, trying to construct a query to get something by it's key is
something you shouldn't need to do. Although this is how you would work with
an SQL database, the Google Data Store does things differently, and this is
one of those cases where rather than go through the trouble of constructing a
query, Google App Engine lets you get what you want directly. After all, you
should only have one entity in the database with a particular key, so there's
nothing in the rest of the machinery of a GQL query that you need in this
case, hence it can all be skipped for efficiency.
|
How to make a .exe for Python with good graphics?
Question: I have a Python application and I decided to do a .exe to execute it.
This is the code that I use to do the .exe:
# -*- coding: cp1252 -*-
from distutils.core import setup
import py2exe, sys, os
sys.argv.append('py2exe')
setup(
options = {'py2exe': {'bundle_files': 1}},
windows = [{'script': "SoundLog.py"}],
zipfile = None,
packages=[r"C:\Users\Public\SoundLog\Code\Código Python\SoundLog\Auxiliar", r"C:\Users\Public\SoundLog\Code\Código Python\SoundLog\Plugins"],
)
But when I run my application with the .exe, the graphics are quite different.
In the image bellow you can see the application running thought python at the
left and running thought the .exe at the right. 
How can I make the .exe one be as good as the one that runs thought python?
Answer: I assume you mean the visual style of the toolbar and buttons. You need to add
a manifest file to the EXE file or as a separate file so that Windows applies
the modern style of recent comctl32.dll versions.
Check out [Using Windows XP Visual Styles With Controls on Windows
Forms](http://msdn.microsoft.com/en-us/library/aa289524%28VS.71%29.aspx) on
MSDN. Read the relevant part about creating the ".exe.manifest" file.
A more py2exe-specific tutorial can be found over at the [wxPython
site](http://wiki.wxpython.org/DistributingYourApplication). They explain how
to use setup.py to include the necessary manifest file.
|
Run Python CGI Script on Windows XP
Question: This exact question has been asked before but I am at my wits end! I've spend
4 hours trying to get a SIMPLE Python CGI script to work on Windows XP but I
get errors. Please save my sanity!
Python Script register.py
#!c:/Python30/python.exe -u
print "Content-type: text/html"
print "<P>Hello, World!</p>"
Script is located in:
C:\Program Files\Apache Software Foundation\Apache2.2\cgi-bin\alerter
Apache Error Log:
[Tue Sep 21 19:06:36 2010] [error] [client 127.0.0.1] Premature end of script headers: register.py
[Tue Sep 21 19:06:36 2010] [error] [client 127.0.0.1] File "C:/Program Files/Apache Software Foundation/Apache2.2/cgi-bin/alerter/register.py", line 3\r
[Tue Sep 21 19:06:36 2010] [error] [client 127.0.0.1] print "Content-type: text/html"\r
[Tue Sep 21 19:06:36 2010] [error] [client 127.0.0.1] ^\r
[Tue Sep 21 19:06:36 2010] [error] [client 127.0.0.1] SyntaxError: invalid syntax\r
httpd.conf:
LoadModule cgi_module modules/mod_cgi.so
<Directory "C:/Program Files/Apache Software Foundation/Apache2.2/cgi-bin">
AllowOverride None
Options None
Order allow,deny
Allow from all
Options +ExecCGI
AddHandler cgi-script .py
</Directory>
This should be VERY simple. Yes? I mus be missing that ONE thing that will
make it finally work. I got PHP working a while back with no problems.
Any ideas? Thanks!!!
Answer: Your error is:
> Premature end of script headers
Note that the HTTP protocol specifies that the body of a HTTP response is
separated from it's headers by a blank line (i.e. two times a carriage return
and line feed). I'd go for something like:
import sys
sys.stdout.write("Content-type: text/html\r\n\r\n<p>Body</p>")
|
Python function to solve Ax = b by back substitution
Question: Okay, for my numerical methods class I have the following question:
Write a Python function to solve Ax = b by back substitution, where A is an
upper triangular nonsingular matrix. MATLAB code for this is on page 190 which
you can use as a pseudocode guide if you wish. The function should take as
input A and b and return x. Your function need not check that A is
nonsingular. That is, assume that only nonsingular A will be passed to your
function.
The MATLAB code that it refers to is:
x(n) = c(u)/U(n,n)
for i = n-1 : -1 : 1
x(i) = c(i);
for j = i+1 : n
x(i) = x(i) - U(i,j)*x(j);
end
x(i) = x(i)/U(i,i);
end
My Python code, which I wrote using the MATLAB code snippet, is with an upper
triangular test matrix(not sure if its nonsingular! How do I test for
singularity?):
from scipy import mat
c=[3,2,1]
U=([[6,5,1],[0,1,7],[0,0,2]])
a=0
x=[]
while a<3:
x.append(1)
a=a+1
n=3
i=n-1
x[n-1]=c[n-1]/U[n-1][n-1]
while i>1:
x[i]=c[i]
j=i+1
while j<n-1:
x[i]=x[i]-U[i][j]*x[j];
x[i]=x[i]/U[i][i]
i=i-1
print mat(x)
The answer I am getting is [[1 1 0]] for x. I not sure if I am doing this
correctly. I assume it is wrong and can't figure out what to do next. Any
clues?
Answer:
j=i+1
while j<n-1:
x[i]=x[i]-U[i][j]*x[j];
is infinite ... and never gets executed
your indexing is fubared:
for i in range(n-2,-1,-1):
....
for j in range(i+1,n):
notice, range is half open unlike matlab
|
Python text validation: a-z and comma (",")
Question: I need to check that some text only contains lower-case letters a-z and a
comma (",").
What is the best way to do this in Python?
Answer:
import re
def matches(s):
return re.match("^[a-z,]*$", s) is not None
Which gives you:
>>> matches("tea and cakes")
False
>>> matches("twiddledee,twiddledum")
True
You can optimise a bit with re.compile:
import re
matcher = re.compile("^[a-z,]*$")
def matches(s):
return matcher.match(s) is not None
|
Namespaces in C# vs imports in Java and Python
Question: In the Java and Python world, you look at a source file and know where all the
imports come from (i.e. you know in which file the imported classes are
defined). For example:
In Java:
import javafoo.Bar;
public class MyClass {
private Bar myBar = new Bar();
}
You immediately see that the Bar-class is imported from javafoo. So, Bar is
declared in `/javafoo/Bar.java`
In Python
import pythonbaz
from pythonfoo import Bar
my_bar = Bar()
my_other = pythonbaz.Other()
Here, it is clear that Bar comes from the pythonfoo package and Other is
obviously from pythonbaz.
In C# (correct me if I'm wrong):
using foo
using baz
using anothernamespace
...
public class MyClass
{
private Bar myBar = new Bar();
}
Two questions:
1) How do I know where the Bar-class is declared? Does it come from the
namespace `foo`, or `bar`, or `anothernamespace`? (edit: **without** using
Visual Studio)
2) In Java, the package names correspond to directory names (or, it is a very
strong convention). Thus, when you see which package a class comes from, you
know its directory in the file system.
In C#, there does not seem to be such a convention for namespaces, or am I
missing something? So, how do I know which directory and file to look in
(without Visual Studio)? (after figuring out which namespace the class came
from).
**Edit clarification** : I am aware that Python and/or Java allow wildcard
imports, but the 'culture' in those languages frowns upon them (at least in
Python, in Java I'm not sure). Also, in Java IDEs usually help you create
minimal imports (as Mchl. commented below)
Answer: 1) Well, you can do the same thing in Java too:
import java.util.*;
import java.io.*;
...
InputStream x = ...;
Does `InputStream` come from `java.util` or `java.io`? Of course, you can
choose not to use that feature.
Now, in _theory_ I realise this means when you're looking with a text editor,
you can't tell where the types come from in C#... but in practice, I don't
find that to be a problem. How often are you _actually_ looking at code and
can't use Visual Studio?
2) You can use the same convention in .NET too, of course - and I do, although
I don't have empty directories going up the chain... so if I'm creating a
project with a default namespace of X.Y, then `X.Y.Foo` would be in `Foo.cs`,
and X.Y.Z.Bar would be in `Z\Bar.cs`
That's also what Visual Studio will do by default - if you create a subfolder,
it will create new classes using a namespace based on the project default and
the folder structure.
Of course, you can also declare types in any old file - but _mostly_ people
will follow the normal convention of declaring a type with a corresponding
filename. Before generics made delegate declarations rarer, I used to have a
`Delegates.cs` file containing all the delegate declarations for a particular
namespace (rather than having a bunch of single-declaration files) but these
days that's less of an issue.
|
How to reload modules in django shell?
Question: I am working with Django and use Django shell all the time. The annoying part
is that while the Django server reloads on code changes, the shell does not,
so every time I make a change to a method I am testing, I need to quit the
shell and restart it, re-import all the modules I need, reinitialize all the
variables I need etc. While iPython history saves a lot of typing on this,
this is still a pain. Is there a way to make django shell auto-reload, the
same way django development server does?
I know about reload(), but I import a lot of models and generally use `from
app.models import *` syntax, so reload() is not much help.
Answer: look at the manage.py shell_plus command provided by the [django-
extensions](https://github.com/django-extensions/django-extensions) project.
It will load all your model files on shell startup. and autoreload your any
modify but do not need exit, you can direct call there
|
How can I selectively mask arbitrary data being sent over an insecure link?
Question: I'm using an offsite error logging package for my python web application. When
I send an error I include the contents of (among other things) the POST
variable and some template data. Some of this data must not be sent to the
error logging service (passwords, some other template data).
How can I take a payload that consists of a mix of data -- objects, dicts, et
al -- and mask out (say) every field or entry named `my_private_data`?
What I'd expect is that if an object has a string or integer property (the
private data will always be a number or a string) `my_private_data=SOME
SECRET`, it would be transmitted as `my_private_data=**********`
How do I accomplish this?
Answer: If you have the POST data as a string, you can use the standard modules
"urlparse" and "urllib" to remove certain parameters:
import urlparse
import urllib
postDataAsDict = urlparse.parse_qs("a=5&b=3&c=%26escaped", strict_parsing = True)
print postDataAsDict # prints {'a': ['5'], 'b': ['3'], 'c': ['&escaped']}
del postDataAsDict["a"] # in your case "my_private_data"
print urllib.urlencode(postDataAsDict, True) # prints c=%26escaped&b=3
Note that `parse_qs` correctly supports multiple parameters that have the same
name, so don't worry about that.
|
free implementation of counting user sessions from a web server log?
Question: Web server log analyzers (e.g. Urchin) often display a number of "sessions". A
session is defined as a series of page visits / clicks made by an individual
within a limited, continuous time segment. The attempt is made to identify
these segments using IP addresses, and often supplementary info like user
agent and OS, along with a session timeout threshold such as 15 or 30 minutes.
For certain web sites and applications, a user can be logged in and/or tracked
with a cookie, which means the server can precisely know when a session
begins. I'm not talking about that, but about inferring sessions heuristically
("[session
reconstruction](http://www.dblab.ntua.gr/persdl2007/papers/72.pdf)") when the
web server does not track them.
I could write some code e.g. in Python to try to reconstruct sessions based on
the criteria mentioned above, but I'd rather not reinvent the wheel. I'm
looking at log files of a size around 400K lines, so I'd have to be careful to
use a scalable algorithm.
My goal here is to extract a list of unique IP addresses from a log file, and
for each IP address, to have the number of sessions inferred from that log.
Absolute precision and accuracy are not necessary... pretty-good estimates are
ok.
Based on [this
description](http://www.dblab.ntua.gr/persdl2007/papers/72.pdf):
> a new request is put in an existing session if two conditions are valid:
>
> * the IP address and the user-agent are the same of the requests already
> inserted in the session,
> * the request is done less than fifteen minutes after the last request
> inserted.
>
it would be simple in theory to write a Python program to build up a
dictionary (keyed by IP) of dictionaries (keyed by user-agent) whose value is
a pair: (number of sessions, latest request of latest session).
But I would rather try to use an existing implementation if one's available,
since I might otherwise risk spending a lot of time tuning performance.
FYI lest someone ask for sample input, here is a line of our log file
(sanitized):
#Fields: date time s-ip cs-method cs-uri-stem cs-uri-query s-port cs-username c-ip cs(User-Agent) cs(Referer) sc-status sc-substatus sc-win32-status
2010-09-21 23:59:59 215.51.1.119 GET /graphics/foo.gif - 80 - 128.123.114.141 Mozilla/5.0+(Windows;+U;+Windows+NT+5.1;+en-US;+rv:1.9.2)+Gecko/20100115+Firefox/3.6+(.NET+CLR+3.5.30729) http://www.mysite.org/blarg.htm 200 0 0
Answer: OK, in the absence of any other answer, here's my Python implementation. I'm
not a Python expert. Suggestions for improvement are welcome.
#!/usr/bin/env python
"""Reconstruct sessions: Take a space-delimited web server access log
including IP addresses, timestamps, and User Agent,
and output a list of the IPs, and the number of inferred sessions for each."""
## Input looks like:
# Fields: date time s-ip cs-method cs-uri-stem cs-uri-query s-port cs-username c-ip cs(User-Agent) cs(Referer) sc-status sc-substatus sc-win32-status
# 2010-09-21 23:59:59 172.21.1.119 GET /graphics/foo.gif - 80 - 128.123.114.141 Mozilla/5.0+(Windows;+U;+Windows+NT+5.1;+en-US;+rv:1.9.2)+Gecko/20100115+Firefox/3.6+(.NET+CLR+3.5.30729) http://www.site.org//baz.htm 200 0 0
import datetime
import operator
infileName = "ex100922.log"
outfileName = "visitor-ips.csv"
ipDict = {}
def inputRecords():
infile = open(infileName, "r")
recordsRead = 0
progressThreshold = 100
sessionTimeout = datetime.timedelta(minutes=30)
for line in infile:
if (line[0] == '#'):
continue
else:
recordsRead += 1
fields = line.split()
# print "line of %d records: %s\n" % (len(fields), line)
if (recordsRead >= progressThreshold):
print "Read %d records" % recordsRead
progressThreshold *= 2
# http://www.dblab.ntua.gr/persdl2007/papers/72.pdf
# "a new request is put in an existing session if two conditions are valid:
# * the IP address and the user-agent are the same of the requests already
# inserted in the session,
# * the request is done less than fifteen minutes after the last request inserted."
theDate, theTime = fields[0], fields[1]
newRequestTime = datetime.datetime.strptime(theDate + " " + theTime, "%Y-%m-%d %H:%M:%S")
ipAddr, userAgent = fields[8], fields[9]
if ipAddr not in ipDict:
ipDict[ipAddr] = {userAgent: [1, newRequestTime]}
else:
if userAgent not in ipDict[ipAddr]:
ipDict[ipAddr][userAgent] = [1, newRequestTime]
else:
ipdipaua = ipDict[ipAddr][userAgent]
if newRequestTime - ipdipaua[1] >= sessionTimeout:
ipdipaua[0] += 1
ipdipaua[1] = newRequestTime
infile.close()
return recordsRead
def outputSessions():
outfile = open(outfileName, "w")
outfile.write("#Fields: IPAddr Sessions\n")
recordsWritten = len(ipDict)
# ipDict[ip] is { userAgent1: [numSessions, lastTimeStamp], ... }
for ip, val in ipDict.iteritems():
# TODO: sum over on all keys' values [(v, k) for (k, v) in d.iteritems()].
totalSessions = reduce(operator.add, [v2[0] for v2 in val.itervalues()])
outfile.write("%s\t%d\n" % (ip, totalSessions))
outfile.close()
return recordsWritten
recordsRead = inputRecords()
recordsWritten = outputSessions()
print "Finished session reconstruction: read %d records, wrote %d\n" % (recordsRead, recordsWritten)
Update: This took 39 seconds to input and process 342K records and write 21K
records. That's good enough speed for my purposes. Apparently 3/4 of that time
was spent in `strptime()`!
|
Making all variables in a scope global or importing a module inside another module
Question: I have a package with two modules in it. One is the `__init__` file, and the
other is a separate part of the package. If I try `from mypackage import
separatepart`, the code in the `__init__` module is run, which will run
unneeded code, slowing down the importing by a lot. The code in separate part
won't cause any errors, and so users should be able to directly import it
without importing the `__init__` module.
Since I can't figure out a way to do this, I thought I should include a
function in the `__init__` file that does everything so nothing would be done
directly, but in order to do this, I would need to have any variables set to
be global. Is there any way to tell Python that all variables are global in a
function, or to not run the `__init__` module?
Answer: dthat I know of, there is not way to specify that _all_ variables are global
but you can import the module while you are in the module.~~just make sure
that you do it in a function that isn't called at the top level, you are
playing with infinite recursion here but a simple use should be safe.~~
#module.py
foo = bar = 0 # global
def init()
import module as m
m.foo = 1
m.bar = 2 # access to globals
~~if`init` was called at the top level, then you have infinite recursion but
it sounds like the whole point of this is to avoid this code running at the
top level, so you should be safe.~~ Since you want to do this in the
`__init__.py` file, just import the top level of the package.
It occurred to me on a walk that there's no problem with recursion here
because the top level code will only run once on initial import.
|
How can I replace the class by monkey patching?
Question: How can I replace the ORM class - so it should not cause recursion !!!
_Problem_ :
original class has the super call, when its got replaced - it causes self
inheritance and causes maximum recursion depth exceed exception.
i.e. class orm is calling super(orm, self).... and orm has been replaced by
another class which inherits original orm....
# Package !
addons __init__.py osv run_app.py
./addons:
__init__.py test_app1.py test.py
./osv:
__init__.py orm.py
# contents of orm.py
class orm_template(object):
def __init__(self, *args, **kw):
super(orm_template, self).__init__()
def fields_get(self, fields):
return fields
def browse(self, id):
return id
class orm(orm_template):
def __init__(self, *args, **kw):
super(orm, self).__init__(*args, **kw)
def fields_get(self, fields, context = None):
return super(orm, self).fields_get(fields)
def read(self, fields):
return fields
# contents of addons/**init**.py
import test
def main(app):
print "Running..."
__import__(app, globals(), locals())
# contents of addons/test.py
from osv import orm
import osv
class orm(orm.orm):
def __init__(self, *args, **kw):
super(orm, self).__init__(*args, **kw)
def fields_get(self, *args, **kw):
print "my fields get................."
return super(orm, self).fields_get(*args, **kw)
osv.orm.orm = orm
print "replaced.........................."
# contents of test_app1.py
from osv.orm import orm
class hello(orm):
_name = 'hellos'
def __init__(self, *args, **kw):
super(hello, self).__init__(*args, **kw)
print hello('test').fields_get(['name'])
# contents of run_app.py
import addons
addons.main('test_app1')
### OUTPUT
>>>python run_app.py
replaced..........................
Running...
...
...
super(orm, self).__init__(*args, **kw)
RuntimeError: maximum recursion depth exceeded
I've seen the similar
[question](http://stackoverflow.com/questions/3765222/monkey-patch-python-
class)
Answer: Your `addons/test.py` needs to get and keep a reference to the original
`orm.orm` and use that instead of the replaced version. I.e.:
from osv import orm
import osv
original_orm = osv.orm
class orm(original_orm):
def __init__(self, *args, **kw):
super(orm, self).__init__(*args, **kw)
def fields_get(self, *args, **kw):
print "my fields get................."
return super(orm, self).fields_get(*args, **kw)
osv.orm.orm = orm
print "replaced.........................."
so the monkeypatched-in class inherit from the original rather than from
itself, as you had it in your setup. BTW, if you can avoid monkey-patching by
better design of the `osv` module (e.g. w/a setter function to set what's the
orm) you'll be happier;-).
|
How to submit web forms using Python?
Question: First of all, sorry if this question is a little vague and rambling! I'm ok
with Python, but I've never done anything HTTP related before.
I'm trying to automate submitting a web form, and from reading some of [this
page](http://www.jmarshall.com/easy/http/) I understand that I need to do a
POST request. I also found a code snippet demonstrating the urllib module:
import urllib
params = urllib.urlencode({'spam': 1, 'eggs': 2, 'bacon': 0})
f = urllib.urlopen("http://www.musi-cal.com/cgi-bin/query", params)
print f.read()
But I still don't really understand what I'm doing. I need to trigger "submit"
somehow, and I assume the actual data I'm submitting will go in the params
somewhere?
Answer: The code there should do what you want.
Whatever data you want to use should go into the params as you have in your
example. When the params are included as an argument to urlopen a POST request
will be used (instead of a GET).
By just calling urlopen I believe the POST request will be submitted. If you
want the response however you will need to use f.read().
|
how to read password protected excel in python
Question: I'm new to python programming, and I am trying to read a password protected
file using python, the code is shown below:
import sys
import win32com.client
xlApp = win32com.client.Dispatch("Excel.Application")
print "Excel library version:", xlApp.Version
filename,password = 'C:\myfiles\foo.xls', 'qwerty12'
xlwb = xlApp.Workbooks.Open(filename, Password=password)
But then the xls file is loaded but still prompt me to provide the password, I
can't let python to enter the password for me.
What have I done wrong? Thanks!
Answer: Open takes two types of password, namely:
Password: password required to open a protected workbook.
WriteResPassword : password required to write to a write-reserved workbook
So in your case , is it write protected or protection on open?
Also there is a discussion on SO that says that this does not work with named
parameters, So try providing all parameter values with the defaults
* <http://stackoverflow.com/questions/2887339/how-to-open-write-reserved-excel-file-in-python-with-win32com>
Default values are documented in MSDN
* <http://msdn.microsoft.com/en-us/library/microsoft.office.interop.excel.workbooks.open.aspx>
|
get rhythmbox information from other user
Question: I have Rhythmbox running on my desktop, and I want to be able to control it
from remotely via a web interface. I'm having problems accessing it, however,
because `rhythmbox-client` is complaining that the user (www-data) that is
trying to access it doesn't a) have as X session running, and b) doesn't have
access to my rhythmbox dbus information.
`(rhythmbox-client:13954): Rhythmbox-WARNING **: /bin/dbus-launch terminated
abnormally with the following error: Autolaunch error: X11 initialization
failed.`
I'm running apache, php on ubuntu 10.04. The following methods have also been
tried by calling the php script via php cli from a different user (sshing into
my own box with a different user)
I've tried these approaches:
* calling system() and exec() from the php script with "rhythmbox-client --print-playing --no-start --no-resent"
* using DBus directly in php as well as a python script (calling it from the php script with system()/exec())
* using `setuid()`, `setruid()`, `seteuid()` in a C program and calling it via php.
I'm at a loss now. Is this even possible?
[EDIT] I have used @IvanGoneKrazy's suggestion and taken the environ code from
[this similar question](http://stackoverflow.com/questions/2143785/python-
dbus-session-bus-x11-dependency). Now I have this code with this error
message:
import subprocess, os
p = subprocess.Popen('dbus-launch', shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
for var in p.stdout:
sp = var.split('=', 1)
os.environ[sp[0]] = sp[1][:-1]
os.system('rhythmbox-client --print-playing')
error message:
(rhythmbox-client:15377): Rhythmbox-WARNING **: Launch helper exited with unknown return code 1
Answer: A similar problem and Python solution for is given
[here](http://stackoverflow.com/questions/2143785/python-dbus-session-
bus-x11-dependency). The root issue seems to be that system() and exec() do
not have the correct environment variables set. You can use PHPs
<http://us.php.net/manual/en/function.proc-open.php> to set the environment
vars for your PHP spawned process.
|
How to check if a word is an English word with Python?
Question: I want to check in a Python program if a word is in the English dictionary.
I believe nltk wordnet interface might be the way to go but I have no clue how
to use it for such a simple task.
def is_english_word(word):
pass # how to I implement is_english_word?
is_english_word(token.lower())
In the future, I might want to check if the singular form of a word is in the
dictionary (e.g., properties -> property -> english word). How would I achieve
that?
Answer: For (much) more power and flexibility, use a dedicated spellchecking library
like [`PyEnchant`](http://pythonhosted.org/pyenchant/). There's a
[tutorial](http://pythonhosted.org/pyenchant/tutorial.html), or you could just
dive straight in:
>>> import enchant
>>> d = enchant.Dict("en_US")
>>> d.check("Hello")
True
>>> d.check("Helo")
False
>>> d.suggest("Helo")
['He lo', 'He-lo', 'Hello', 'Helot', 'Help', 'Halo', 'Hell', 'Held', 'Helm', 'Hero', "He'll"]
>>>
`PyEnchant` comes with a few dictionaries (en_GB, en_US, de_DE, fr_FR), but
can use any of the [OpenOffice
ones](http://wiki.services.openoffice.org/wiki/Dictionaries) if you want more
languages.
There appears to be a pluralisation library called
[`inflect`](http://pypi.python.org/pypi/inflect), but I've no idea whether
it's any good.
|
How to open SQL Compact database read only
Question: There is a SQL Compact v3.1 database that I want to quickly read. I'm doing
this in python so I don't have access to managed code.
I've noticed that if I use adodbapi the database file actually gets modified
just by opening it. And sadly when I add 'File mode=Read Only' to the
connection string I get a weird error.
Here is the code I use to connect:
import adodbapi
adodbapi.connect('Provider=Microsoft.SQLSERVER.MOBILE.OLEDB.3.0; Data Source="awesome.sdf"; File mode = Read Only;SSCE:Temp File Directory=c:\temp\\;')
And then I get the error message
OperationalError: (com_error(-2147352567, 'Exception occurred.', (0, u'Microsoft OLE DB Service Components',
u'Multiple-step OLE DB operation generated errors. Check each OLE DB status value, if available. No work was done.',
None, 0, -2147217887), None),
u'Error opening connection: Provider=Microsoft.SQLSERVER.MOBILE.OLEDB.3.0; Data Source="Awesome.sdf";File mode = Read Only;SSCE:Temp File Directory="c:\\\temp\\";')
I added the SSCE because when I wrote a test program in C# it needed it. The
following code works perfectly fine and doesn't modify the file when you do a
simple SELECT query.
conn = new SqlCeConnection("Data Source = awesome.spf; File mode = Read Only;SSCE:Temp File Directory=\"c:\\users\\evelio\\desktop\\\";");
conn.Open();
Thanks for the help,
Evelio
Answer: Look here: <http://social.msdn.microsoft.com/Forums/en-
US/sqlce/thread/bf70c615-b279-4a91-b964-0ff99adc7ab8/#674f6a79-a3b4-4601-a952-860a7e8f3169>
cn.Mode = adModeRead
|
How can you select a random element from a list, and have it be removed?
Question: Let's say I have a list of colours, `colours = ['red', 'blue', 'green',
'purple']`.
I then wish to call this python function that I hope exists, `random_object =
random_choice(colours)`. Now, if random_object holds 'blue', I hope `colours =
['red', 'green', 'purple']`.
Does such a function exist in python?
Answer: Firstly, if you want it removed because you want to do this again and again,
you might want to use `random.shuffle()` in the random module.
`random.choice()` picks one, but does not remove it.
Otherwise, try:
import random
# this will choose one and remove it
def choose_and_remove( items ):
# pick an item index
if items:
index = random.randrange( len(items) )
return items.pop(index)
# nothing left!
return None
|
Mapping Languages to Paradigms
Question: I recently read Eric Steven Raymond's article "How To Become A Hacker" and I
like his suggestion of learning 5 key languages (he suggests Python, C/C++,
Lisp, Java, and Perl) as a way of covering the main programming paradigms in
use today.
His advice is that it's not so important which specific languages a programmer
knows. It's more important to know different approaches to programming, for
two reasons.
The first reason is that it makes it trivial to pick up a new language, once
you know the general approach to the way it solves problems.
The second reason is that there is no one best language - they all have trade-
offs. It would be best to know what type of language to pick given a specific
type of problem. This is what I'm most interested in, but I'm having a problem
really distinguishing between the 5 languages he suggests. There seems to be a
lot of overlap.
So my specific question is, given these 5 languages, what is their intended
programming paradigm, and give one example of the type of problem it would be
best suited for.
An example answer (and I'm not sure this answer is correct):
Perl
- mainly a functional language
- great for quick text substitutions in multiple files from the command line.
I found a few other similar questions posted, but I'd like to know about these
5 languages in particular. I'm just looking for a starting point, nothing too
detailed. Thanks in advance!
Answer: I think you're approaching it wrong. As esr himself says, it's not the
_language_ that matters, it's the _paradigm_. So when you say that
> 1. Perl is a functional language
> 2. It's great for quick text substitutions in multiple files from the
> command line
>
you are missing one of the main points of a functional language which is that
they are great for building large systems using a bottom up approach: solve a
bunch of (well chosen) small problems with well designed functions until we
have a complete system. We cut down on code duplication by identifying what
algorithms that we are using have in common and using higher order functions
to encapsulate their commonality. We minimize (overt) branching behavior by
using higher order functions to cook up just the function that we need for a
given situation.
Likewise, I could say that
> 1. Java is mainly an OOP language
> 2. It's good for writing large, robust systems,
>
but that misses the point that OOP languages are about modeling concepts from
the problem domain in code so that we are left with a clear way to
imperatively solve the problem at hand. We cut down on code duplication by
identifying what the relevant concepts have in common and encapsulating the
code that deals with those commonalities in a class that describes it. We
minimize (overt) branching behavior by providing different subclasses of an
abstraction with appropriately different behavior.
On the whole, the basic point of programming languages and their associated
paradigms is
* to allow you to not think about anything that doesn't affect the quality of the resulting program. If that wasn't a (largely) desirable thing, then we would all be writing machine code.
* This is accomplished by (among other things) providing a set of tools for _building abstractions_.
Shop around and pick one that you like and get good at. Just make sure that
you learn when the other ones allow for a better solution (this will probably
mean getting good at them eventually too ;). I think that you can mainly take
"good solution" to mean, "clear mapping of **code** to **ideas** ". (modulo
concerns about efficiency that would force you (provide an excuse?) to write
in a language like C)
|
python mechanize javascript submit button problem!
Question: im making some script with mechanize.browser module.
one of problem is all other thing is ok, but when submit() form,it not
working,
so i was found some suspicion source part.
in the html source i was found such like following.
im thinking, loginCheck(this) making problem when submit form.
but how to handle this kind of javascript function with mechanize module ,so i
can
successfully submit form and can receive result?
following is websource snippet which related with loginCheck(this) javascript
function.
function init(){
FRMLOGIN.ID.focus();
}
function loginCheck(f){
if(chkNull(f.ID, "아이디를"))
return false;
if(chkNull(f.PWD, "패스워드를"))
return false;
//f.target = "ifrmLoginHidden";
f.action = (f.SECCHK.checked) ? "https://user.buddybuddy.co.kr/Login/Login.asp" : "http://user.buddybuddy.co.kr/Login/Login.asp";
}
i know mechanize not support javascript, so i want to make progammatically
loginCheck()
function with python mechanize code.
anyone would you some help me to make this javascript function to python
mechanize
translated code?
so correctly can login with website?
if so much appreciate!
# -*- coding: cp949-*-
import sys,os
import mechanize, urllib
import cookielib
from BeautifulSoup import BeautifulSoup,BeautifulStoneSoup,Tag
import datetime, time, socket
import re,sys,os,mechanize,urllib,time
br = mechanize.Browser()
cj = cookielib.LWPCookieJar()
br.set_cookiejar(cj)
# Browser options
br.set_handle_equiv(True)
br.set_handle_gzip(True)
br.set_handle_redirect(True)
br.set_handle_referer(True)
br.set_handle_robots(False)
# Follows refresh 0 but not hangs on refresh > 0
br.set_handle_refresh(mechanize._http.HTTPRefreshProcessor(), max_time=1)
# Want debugging messages?
br.set_debug_http(True)
br.set_debug_redirects(True)
br.set_debug_responses(True)
# User-Agent (this is cheating, ok?)
br.addheaders = [('User-agent', 'Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US; rv:1.9.0.6')]
br.open('http://user.buddybuddy.co.kr/Login/LoginForm.asp?URL=')
html = br.response().read()
print html
br.select_form(name='FRMLOGIN')
print br.viewing_html()
br.form['ID']='psh7943'
br.form['PWD']='qkrthgus'
br.submit()
print br.response().read()
if anyone can help me ..much appreciate!!
Answer: You can go through the login process by hand in your browser and check (using
e.g. Firebug in firefox, Developer Tools in Chrome etc.) what requests are
sent to the site when you hit the OK button. Usually this is a POST request
with data taken from the login form. Check what data are sent in this request
and execute your own post request with:
mechanize.urlopen(URL, POST_DATA).
You can extract POST_DATA (and post_url) from mechanize's form object using:
form.click_request_data()
but you may need to do some modifications.
Very simple example:
br.select_form(name='form_name')
br.form['login']='login'
br.form['pass']='pass'
post_url, post_data, headers = br.form.click_request_data()
mechanize.urlopen(post_url, post_data)
|
permissive equality test on string
Question: I'm a python newbie with a problem too hard to tackle.
I have a string defining a path, were all the spaces have been converted to
underscores. How can I find if it corresponds to a real path?
e.g. a string like `/some/path_to/directory_1/and_to/directory_2`
with a real path: `/some/path_to/directory 1/and_to/directory 2`
notice that the real path can contain BOTH spaces and underscores.
How can I feed it to `os.path.exists()` ???
thanks alessandro
Answer: Use [glob](http://docs.python.org/library/glob.html) but replacing every
underscore with a range `[ _]`:
import glob
glob.glob('/some/path_to/directory_1/and_to/directory_2'.replace('_', '[ _]'))
Note that this will fail if your path contains the character `[`. You can fix
this by first replacing `[` with `[[]`.
|
uploading records of list of files in parallel using python to DB
Question: I have a list of files each file have mass of records separting by \n , i need
to proccess those records in parallel and upload them to some sql server could
someone provide an idea what is the best way to do this with python
Answer: The best way might not be to upload in parallell but use SQL Servers bulk
importing mechanisims
e.g.
[BULK INSERT](http://msdn.microsoft.com/en-us/library/ms188365.aspx)
[bcp](http://msdn.microsoft.com/en-us/library/ms162802.aspx)
EDIT:
If you need to process them then a way I have often used is
1) bulk load the data into a staging table
2) Process the data on the database
3) Insert into main tables
Stages 2 and 3 can be combined if the processing is of a reasonable type.
This could be faster as there are less round trips to the server and
processing a set of data rather than row by row is usually quicker.
Also I thing that SQL server will make use of more than one CPU in doing this
processing so you get your processing parallel for free
|
IOError: request data read error
Question: I seem to be getting an IOError: request data read error quite a lot when i'm
doing an Ajax upload. For example out of every 5 file uploads it errors out on
atleast 3.
Other people seem to have had the same issue. Eg.
* <http://stackoverflow.com/questions/2641665/django-upload-failing-on-request-data-read-error>
* <http://stackoverflow.com/questions/411902/django-file-upload-failing-occasionally>
Some other observations:
* It's definitely not my internet connection or a browser issue. Seems to be happening on all browsers chrome/FF/opera.
* I'm running **django 1.1.1 Apache/2.2.14 (Ubuntu) mod_ssl/2.2.14 OpenSSL/0.9.8k mod_wsgi/2.8 Python/2.6.5** on Lucid.
* It is also not the file size. I can sometimes upload 1+ MB files but fail on 180 Kb files.
* * *
**Traceback**
Traceback (most recent call last):
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/core/handlers/base.py", line 98, in get_response
response = middleware_method(request, e)
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/core/handlers/base.py", line 92, in get_response
response = callback(request, *callback_args, **callback_kwargs)
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/contrib/auth/decorators.py", line 78, in __call__
return self.view_func(request, *args, **kwargs)
File "/home/ubuntu/webapps/anonymous_app/app/do_work/views/__init__.py", line 391, in some_form_ajax_upload
f = request.FILES.get('file_upload')
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/core/handlers/wsgi.py", line 187, in _get_files
self._load_post_and_files()
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/core/handlers/wsgi.py", line 137, in _load_post_and_files
self._post, self._files = self.parse_file_upload(self.META, self.environ['wsgi.input'])
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/__init__.py", line 124, in parse_file_upload
return parser.parse()
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 133, in parse
for item_type, meta_data, field_stream in Parser(stream, self._boundary):
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 606, in __iter__
for sub_stream in boundarystream:
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 420, in next
return LazyStream(BoundaryIter(self._stream, self._boundary))
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 446, in __init__
unused_char = self._stream.read(1)
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 299, in read
out = ''.join(parts())
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 292, in parts
chunk = self.next()
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 314, in next
output = self._producer.next()
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 375, in next
data = self.flo.read(self.chunk_size)
File "/home/ubuntu/.virtualenvs/anonymous_app/lib/python2.6/site-packages/django/http/multipartparser.py", line 405, in read
return self._file.read(num_bytes)
IOError: request data read error
<WSGIRequest
GET:<QueryDict: {}>,
POST:<could not parse>,
COOKIES:{'__utma': '168279989.1688771210.1285773436.1285773436.1285773436.1',
'__utmb': '168279989.20.10.1285773436',
'__utmc': '168279989',
'__utmz': '168279989.1285773436.1.1.utmcsr=(direct)|utmccn=(direct)|utmcmd=(none)',
'beta': 'True',
'sessionid': 'b1ecf92f2bba13e1885d07803e10aa03',
'timezone_offset': '-330'},
META:{'CONTENT_LENGTH': '188575',
'CONTENT_TYPE': 'multipart/form-data; boundary=---------------------------57602381214905740261171925981',
'DOCUMENT_ROOT': '/htdocs',
'GATEWAY_INTERFACE': 'CGI/1.1',
'HTTPS': '1',
'HTTP_ACCEPT': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
'HTTP_ACCEPT_CHARSET': 'ISO-8859-1,utf-8;q=0.7,*;q=0.7',
'HTTP_ACCEPT_ENCODING': 'gzip,deflate',
'HTTP_ACCEPT_LANGUAGE': 'en-us,en;q=0.5',
'HTTP_CONNECTION': 'keep-alive',
'HTTP_COOKIE': 'beta=True; __utma=168279989.1688771210.1285773436.1285773436.1285773436.1; __utmb=168279989.20.10.1285773436; __utmc=168279989; __utmz=168279989.1285773436.1.1.utmcsr=(direct)|utmccn=(direct)|utmcmd=(none); sessionid=b1ecf92f2bba13e1885d07803e10aa03; timezone_offset=-330',
'HTTP_HOST': 'xxxxxx.compute-1.amazonaws.com',
'HTTP_KEEP_ALIVE': '115',
'HTTP_REFERER': 'https://ec2-184-72-79-96.compute-1.amazonaws.com/do-my-somees/enter/some-documents/',
'HTTP_USER_AGENT': 'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.10) Gecko/20100915 Ubuntu/10.04 (lucid) Firefox/3.6.10',
'PATH': '/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/X11R6/bin',
'PATH_INFO': u'/do-my-somees/enter/some-documents/ajax-upload/Other-some-Document/',
'PATH_TRANSLATED': '/home/ubuntu/webapps/anonymous_app/settings/apache/qa.wsgi.py/do-my-somees/enter/some-documents/ajax-upload/Other-some-Document/',
'QUERY_STRING': '',
'REMOTE_ADDR': '',
'REMOTE_PORT': '15561',
'REQUEST_METHOD': 'POST',
'REQUEST_URI': '/do-my-somees/enter/some-documents/ajax-upload/Other-some-Document/',
'SCRIPT_FILENAME': '/home/ubuntu/webapps/anonymous_app/settings/apache/qa.wsgi.py',
'SCRIPT_NAME': u'',
'SERVER_ADDR': '10.196.142.182',
'SERVER_ADMIN': 'dev@anonymous_app.com',
'SERVER_NAME': 'ec2-184-72-79-96.compute-1.amazonaws.com',
'SERVER_PORT': '443',
'SERVER_PROTOCOL': 'HTTP/1.1',
'SERVER_SIGNATURE': '<address>Apache/2.2.14 (Ubuntu) Server at ec2-184-72-79-96.compute-1.amazonaws.com Port 443</address>\n',
'SERVER_SOFTWARE': 'Apache/2.2.14 (Ubuntu)',
'SSL_TLS_SNI': 'ec2-184-72-79-96.compute-1.amazonaws.com',
'mod_wsgi.application_group': 'qa.anonymous_app.com|',
'mod_wsgi.callable_object': 'application',
'mod_wsgi.listener_host': '',
'mod_wsgi.listener_port': '443',
'mod_wsgi.process_group': '',
'mod_wsgi.reload_mechanism': '0',
'mod_wsgi.script_reloading': '1',
'mod_wsgi.version': (2, 8),
'wsgi.errors': <mod_wsgi.Log object at 0xb9456860>,
'wsgi.file_wrapper': <built-in method file_wrapper of mod_wsgi.Adapter object at 0xb936a968>,
'wsgi.input': <mod_wsgi.Input object at 0xb9720e30>,
'wsgi.multiprocess': True,
'wsgi.multithread': False,
'wsgi.run_once': False,
'wsgi.url_scheme': 'https',
'wsgi.version': (1, 0)}>
Answer: I get this exception, too. In the Apache error logfile I see this:
[Wed Aug 17 08:30:45 2011] [error] [client 10.114.48.206] (70014)End of file found: mod_wsgi (pid=9722): Unable to get bucket brigade for request., referer: https://egs-work/modwork/beleg/188074/edit/
[Wed Aug 17 08:30:45 2011] [error] [client 10.114.48.206] mod_wsgi (pid=3572): Exception occurred processing WSGI script '/home/modwork_egs_p/modwork_egs/apache/django_wsgi.py'.
[Wed Aug 17 08:30:45 2011] [error] [client 10.114.48.206] IOError: failed to write data
Versions:
apache2-prefork-2.2.15-3.7.x86_64
apache2-mod_wsgi-3.3-1.8.x86_64 WSGIDaemonProcess with threads=1
mod_ssl/2.2.15
Linux egs-work 2.6.34.8-0.2-default #1 SMP 2011-04-06 18:11:26 +0200 x86_64 x86_64 x86_64 GNU/Linux
openSUSE 11.3 (x86_64)
First I was confused, because the last line "failed to **write** data" does
not fit to the django code "load post data". But I guess that django wants to
write an error page to the client. But the client has canceled the tcp
connection. And now http 500 page can't be written to the client.
The client disconnected after sending the request, and before getting the
response:
* The user closed the browser or navigated to an other page.
* The user pressed the reload button.
I have seen this only with POST-Requests (not GET). If POST is used, the
webserver does read at least twice: First to get the headers, the second to
get the data. The second read fails.
It is easy to reproduce:
Insert some code which waits before the first access to request.POST happens
(be sure, that no middleware accesses request.POST before time.sleep()):
def edit(request):
import time
time.sleep(3)
#.....
Now do a big POST (e.g. file upload). I don't know the apache buffer size. But
5 MB should be enough. When the browser shows the hourglass, browse to an
other page. The browser will cancel the request and the exception should be in
the logfile.
This is my Middleware, since I don't want to get the above traceback in our
logfiles:
class HandleExceptionMiddleware:
def process_exception(self, request, exception):
if isinstance(exception, IOError) and 'request data read error' in unicode(exception):
logging.info('%s %s: %s: Request was canceled by the client.' % (
request.build_absolute_uri(), request.user, exception))
return HttpResponseServerError()
|
More pythonic way to write this?
Question: I have this code here:
import re
def get_attr(str, attr):
m = re.search(attr + r'=(\w+)', str)
return None if not m else m.group(1)
str = 'type=greeting hello=world'
print get_attr(str, 'type') # greeting
print get_attr(str, 'hello') # world
print get_attr(str, 'attr') # None
Which works, but I am not particularly fond of this line:
return None if not m else m.group(1)
In my opinion this would look cleaner if we could use a ternary operator:
return (m ? m.group(1) : None)
But that of course isn't there. What do you suggest?
Answer: Python _has_ a ternary operator. You're using it. It's just in the `X if Y
else Z` form.
That said, I'm prone to writing these things out. Fitting things on one line
isn't so great if you sacrifice clarity.
def get_attr(str, attr):
m = re.search(attr + r'=(\w+)', str)
if m:
return m.group(1)
return None
|
Python and C interaction - callback function
Question: I'm trying to make a key logger for Mac OS for one of my research projects. I
have a C code which will grab keystroke and write them to a text file. (The
following code I have taken out some not important stuff)
What I need to do now is just like PyHook, instead of write the data to a text
file, to pass a Python callback function to the C code and make it passes back
the key input to Python, so I can do necessary analysis with Python.
I have look for how to do it, but honestly I have no idea how to approach
this, as I am not used to C programming or Python extensions. Any help would
be greatly appreciated.
#include <Carbon/Carbon.h>
#include <ApplicationServices/ApplicationServices.h>
#include <unistd.h>
#include <stdio.h>
#include <sys/time.h>
#define NUM_RECORDING_EVENT_TYPES 5
#define RECORD 0
#define MOUSEACTION 0
#define KEYSTROKE 1
// maximum expected line length, for fgets
#define LINE_LENGTH 80
#define kShowMouse TRUE
OSStatus RUIRecordingEventOccurred(EventHandlerCallRef nextHandler, EventRef theEvent, void *userData);
void prepareToRecord(); // install the event handler, wait for record signal
// note that keyboard character codes are found in Figure C2 of the document
// Inside Macintosh: Text available from http://developer.apple.com
char * keyStringForKeyCode(int keyCode); // get the representation of the Mac keycode
// Global Variables
int dieNow = 0; // should the program terminate
int ifexit = 0; // Exit state
char *filename = NULL; // Log file name
FILE *fd = NULL; // Log file descriptor
int typecount = 0; // count keystroke to periodically save to a txt file
struct timeval thetime; // for gettimeofday
long currenttime; // the current time in milliseconds
int main()
{
filename = "test.txt";
fd = fopen(filename, "a");
// Get RUI ready to record or play, based off of mode
prepareToRecord();
return EXIT_SUCCESS;
}
// event handler for RUI recorder
OSStatus RUIRecordingEventOccurred(EventHandlerCallRef nextHandler, EventRef theEvent, void *userData)
{
// Determine class and kind of event
int eventClass = GetEventClass(theEvent);
int eventKind = GetEventKind(theEvent);
/* Handle Keyboard Events */
if((eventClass == kEventClassKeyboard) && (eventKind == kEventRawKeyDown)) /* key release implied */ {
int keyCode, modifiers; // what did the user press? any modifier keys down?
// gather keystroke information
GetEventParameter(theEvent, kEventParamKeyCode, typeInteger, NULL, sizeof(keyCode), NULL, &keyCode);
GetEventParameter(theEvent, kEventParamKeyModifiers, typeInteger, NULL, sizeof(modifiers), NULL, &modifiers);
// What time is it?
gettimeofday(&thetime, NULL);
currenttime =(((thetime.tv_sec*1000000) + (thetime.tv_usec)));
fprintf(fd, "%s\n", keyStringForKeyCode(keyCode));
}
return EXIT_SUCCESS;
}
void prepareToRecord()
{
EventRecord event; // holds an event for examination
// Types of events to listen for
EventTypeSpec eventTypes[NUM_RECORDING_EVENT_TYPES] = {{kEventClassKeyboard, kEventRawKeyDown}};
// Install the event handler
InstallEventHandler(GetEventMonitorTarget(), NewEventHandlerUPP(RUIRecordingEventOccurred), NUM_RECORDING_EVENT_TYPES, eventTypes, nil, nil);
// event loop - get events until die command
do {
WaitNextEvent((everyEvent),&event,GetCaretTime(),nil);
} while (dieNow == 0);
}
char * keyStringForKeyCode(int keyCode)
{
// return key char
switch (keyCode) {
case 0: return("a");
default: return("Empty"); // Unknown key, Return "Empty"
}
}
Answer: It's easy - Just Follow the instructions - [Calling Python Functions from
C](http://docs.python.org/release/2.5.2/ext/callingPython.html).
Alternatively if you are trying to call C/C++ functions from Python you can
use [SWIG](http://www.swig.org/) or one of Python's module
[CTypes](http://docs.python.org/library/ctypes.html)
|
Best way to change the value of "settings" from within a Python test case?
Question: I'm writing unit tests in Python for the first time, for a Django app. I've
struck a problem. In order to test a particular piece of functionality, I need
to change the value of one of the app's settings. Here's my first attempt:
def test_in_list(self):
mango.settings.META_LISTS = ('tags',)
tags = Document(filepath).meta['tags']
self.assertEqual(tags, [u'Markdown', u'Django', u'Mango'])
What I'm trying to do is change the value of `META_LISTS` such that the new
value is used when the `Document` object is created. The relevant imports
are...
# tests.py
from mango.models import Document
import mango.settings
# models.py
from mango.settings import *
If I've understood correctly, since `models.py` has already imported the names
from `mango.settings`, changing the value of `META_LISTS` within
`mango.settings` will not alter the value of `META_LISTS` within
`mango.models`.
It's possible – likely even – that I'm going about this in completely the
wrong way. What's the correct way to alter the value of such a "setting" from
within a test case?
**Edit:** I failed to mention that the file `models.py` contains vanilla
Python classes rather than Django models. I certainly need to rename this
file!
Answer: In models.py, use `import mango.settings`. You can then set a variable in your
test code like you would any other:
mango.settings.foo = 'bar'
A module is a singleton. You can change the values in its namespace from
anywhere in your code.
But this won't work if you use `from mango.settings import *`, since that
expression copies the values in the module into the current namespace.
|
dynamic values in kwargs
Question: I have a layer which helps me populating records from the form to tables and
viceversa, it does some input checking, etc.
Now several methods of this layer which are called several times in different
parts of the webform take the same parameters, so I wanted to pack them at the
begining of the codefile.
> kwargs(): return {"tabla":"nombre_tabla","id":[hf_id.Value] ,"container":
> Panel1,"MsgBox1": MsgBox1}
then I call
> IA.search(**kwargs)
but doing that way the values of the dictionary get fixed with the ones they
had in the begining, and one of them is retrieved from a webcontrol so it
needs to be dynamic. So I wrapped them in a function
> def kwargs(): return {"tabla":"nombre_tabla", "id":[hf_id.Value]
> ,"container": Panel1,"MsgBox1": MsgBox1}
and then I call
> IA.search(*_kwargs()) IA.save(_ *kwargs())
etc.
and that way the value of the dictionary which comes from the webform (hf_id)
is dynamic and not fixed. But I was wondering if in this case there is another
way, a pythonic way, to get the values of the dictionary kwargs to be dynamic
and not fixed
Answer: Python objects are pointers (though they are not directly manipulatable by the
user.)
So if you create a list like this:
>>> a = [1, 2, 3]
and then store it in a dictionary:
>>> b = { 'key': a, 'anotherkey': 'spam' }
you will find modifications to the value in the dictionary also modify the
original list:
>>> b['key'].append(4)
>>> print b['key']
[1, 2, 3, 4]
>>> print a
[1, 2, 3, 4]
If you want a copy of an item, so that modifications will not change the
original item, then use the copy module.
>>> from copy import copy
>>> a = [1, 2, 3]
>>> b['key'] = copy(a)
>>> print b['key']
[1, 2, 3]
>>> b['key'].append(4)
>>> print b['key']
[1, 2, 3, 4]
>>> print a
[1, 2, 3]
|
Python, using subprocess.Popen to make linux command line call? I'm getting "[Errno 2] No such file or directory"
Question: I'm trying to follow the info I can find about subprocess.Popen as I want to
make a linux command line call.. I am trying as below but am getting the error
"[Errno 2] No such file or directory". I'm not trying to open a file so I
don't understand this error, and it works fine (although with other issues
relating to waiting for the process to finish when I don't want it to) when I
use a regular os.popen.
I can't seem to figure out how to do this properly, any advice is appreciated.
EDIT: THE COMMAND I AM USING IS COMPLEX AND VARIABLIZED, it would be too out-
of-context to include it here, I think its suffice to say that the code works
when I use `os.popen` and not when I do the new way, so no, the "linux command
line call" is obviously not the call I am using
subprocess.Popen([r"linux command line call"])
>>> [Errno 2] No such file or directory
Answer:
import subprocess
proc=subprocess.Popen(['ls','-l']) # <-- Change the command here
proc.communicate()
`Popen` expects a list of strings. The first string is typically the program
to be run, followed by its arguments. Sometimes when the command is
complicated, it's convenient to use `shlex.split` to compose the list for you:
import shlex
proc=subprocess.Popen(shlex.split('ls -l'))
proc.communicate()
|
Extension Crashing Python on Import?
Question: I have a python extension that is built and installed through distutils (using
mingw on windows). However on import of this module the interpreter crashes.
Is there anyway to debug and figure out why it crashes?
I did look around online and couldn't find anything specific, or any examples.
_EDIT_
Sorry i am trying to compile for python 2.5.4 (we need 2.5.4, since we use
arcgis geoprocessor library):
<http://effbot.org/media/downloads/ftpparse-1.1-20021124.zip>
On windows, i define crash as: "Python.exe has encountered a problem and needs
to close"
I'll try debugging with GDB
_EDIT 2_ For what ever reason, doing a
setup.py clean
For the package and doing:
setup.py install
fixed all the problems. :psyduck:
Answer: I suppose using gdb see <http://oldwiki.mingw.org/index.php/gdb>
|
Python: replacing method in calendar module
Question: I'm trying to replace two methods in calendar module:
import calendar
c = calendar.HTMLCalendar(calendar.MONDAY)
def ext_formatday(self, day, weekday, *notes):
if day == 0:
return '<td class="noday"> </td>'
if len(notes) == 0:
return '<td class="%s">%d<br /></td>' % (self.cssclasses[weekday], day)
else:
return '<td class="%s">%d<br />%s</td>' % (self.cssclasses[weekday], day, notes)
def ext_formatweek(self, theweek, *notes):
if len(notes) == 0:
s = ''.join(self.formatday(d, wd) for (d, wd) in theweek)
else:
s = ''.join(self.formatday(d, wd, notes) for (d, wd) in theweek)
return '<tr>%s</tr>' % s
c.formatday = ext_formatday
c.formatweek = ext_formatweek
print c.formatmonth(2012,1,"foobar")
This won't work - could somebody point me to relevant literature or point out
what I'm doing wrong? I'm trying to implement Alan Hynes suggestion from the
following thread: [thread](http://stackoverflow.com/questions/1101524/python-
calendar-htmlcalendar/1458077#1458077) It way too late for me to think
straight and I've been dancing around that problem for over an hour.
Thanks in advance,
Jakub
Answer: Try replacing the method at the class instead of the instance.
Like this:
import calendar
def ext_formatday(self, day, weekday, *notes):
if day == 0:
return '<td class="noday"> </td>'
if len(notes) == 0:
return '<td class="%s">%d<br /></td>' % (self.cssclasses[weekday], day)
else:
return '<td class="%s">%d<br />%s</td>' % (self.cssclasses[weekday], day, notes)
def ext_formatweek(self, theweek, *notes):
if len(notes) == 0:
s = ''.join(self.formatday(d, wd) for (d, wd) in theweek)
else:
s = ''.join(self.formatday(d, wd, notes) for (d, wd) in theweek)
return '<tr>%s</tr>' % s
calendar.HTMLCalendar.formatday = ext_formatday
calendar.HTMLCalendar.formatweek = ext_formatweek
c = calendar.HTMLCalendar(calendar.MONDAY)
print c.formatmonth(2012,1,"foobar")
|
Best way to generate xml?
Question: I'm creating an web api and need a good way to very quickly generate some well
formatted xml. I cannot find any good way of doing this in python.
Note: Some libraries look promising but either lack documentation or only
output to files.
Answer: [ElementTree](http://docs.python.org/library/xml.etree.elementtree.html) is a
good module for reading xml and writing too e.g.
from xml.etree.ElementTree import Element, SubElement, tostring
root = Element('root')
child = SubElement(root, "child")
child.text = "I am a child"
print tostring(root)
Output:
<root><child>I am a child</child></root>
See this
[tutorial](http://www.doughellmann.com/PyMOTW/xml/etree/ElementTree/create.html)
for more details and how to pretty print.
Alternatively if your XML is simple, do not underestimate the power of string
formatting :)
xmlTemplate = """<root>
<person>
<name>%(name)s</name>
<address>%(address)s</address>
</person>
</root>"""
data = {'name':'anurag', 'address':'Pune, india'}
print xmlTemplate%data
Output:
<root>
<person>
<name>anurag</name>
<address>Pune, india</address>
</person>
</root>
You can use string.Template or some template engine too, for complex
formatting.
|
How to get the duration of a video in Python?
Question: I need to get the video duration in Python. The video formats that I need to
get are [MP4](http://en.wikipedia.org/wiki/MPEG-4_Part_14), Flash video,
[AVI](http://en.wikipedia.org/wiki/Audio_Video_Interleave), and MOV... I have
a shared hosting solution, so I have no
[FFmpeg](http://en.wikipedia.org/wiki/FFmpeg) support.
Answer: You'll probably need to invoke an external program.
[`ffprobe`](http://ffmpeg.org/ffprobe.html) can provide you with that
information:
import subprocess
def getLength(filename):
result = subprocess.Popen(["ffprobe", filename],
stdout = subprocess.PIPE, stderr = subprocess.STDOUT)
return [x for x in result.stdout.readlines() if "Duration" in x]
|
Python Asymmetric Encryption: Using pre-generated prv/pub keys
Question: Ok first off yes I have searched google and stackoverflow and done some
reading (over 4 hours JUST in this sitting) have not found what I need for
these reasons:
* Many of them suggest just launching an exe like gpg.exe (http://stackoverflow.com/questions/1020320)
* Some suggested using PyCrypto or other libraries and looking at them, either a) I can't find how to use any of their API, b) I can't find how to import a pre-existing prv/pub key or c) they use the insecure RandomPool (and me trying to updated it is just asking for trouble)
* Some mention it in passing but I could not find what they where linking to (or their was no link at all.
So I know ask you fellow stackoverflow users how can I do this, take a string
of a public key (or path (I could just write it to a temp file, (I plan to
just have it as a string pub_key = "..." ) ) ) and use it to sign and secure a
string (that will be posted on a forum (JSON for updating my application)?
Also these are RSA keys (Putty Gen 4096 bits SSH-2-RSA) that are generated
using PuttyGen (can be in any format (OpenSSH, ssh.com, ppk)
This is what the public key looks like
\---- BEGIN SSH2 PUBLIC KEY ----
Comment: "rsa-key-20101003"
AAAAB3NzaC1yc2EAAAABJQAAAgEAi+91fFsxZ7k1UuudSe5gZoavwARUyZScCtdf
WQ0ROoJC+XIqW5vVJfgmr+A1jLS5m4wNsrCqeyoX2B22T6iEwqVXrXt3QcbccKMu
WkLKFK1h67q6Coc+3eOTmKrOuZbWc19YQgybdkR/GxF7XAbq4NCGNaCDtMOqX8Q2
L/a9fAYqVdTwg9trpcz3whNmdLk/B0edOABKuVX51UdLV+ZggK503+uAb1JiIIj0
mARwR/HNo4oRLMLf2PjuZsGVYYjJDdVJBU6AN4PUQSRRRPL4+YmsrLJb/TpfJeXA
vj4KZMNJv15YXz7/iMZMKznDtr2RJX5wbSpuTUBNZveA7YiIHxvvvis38b/lX9SJ
SYPfZ9CeQY6MvQgG2zwDTOOvKgOIB4sTGMXfcoxB8AF/QXOcxWFJkZoj36rvMd9n
Po6szLjHXwcEUOUvvQfG4VvdQA0H5gGLHqYL1EehRsgi5qcCoFPaZW2K09ErKcS0
MbrLFjBkQ9KmqAM38bvM8UhCWAMA9VXOGHMxUHBV4Bir9alGS4VX0B8Y0b3dZ+7I
MKkHMCwdEUJf7QVdGxGuSQtVsq8RZbIpk3g7wtv8f6I/iEC58ekdrH35tq5+1ilW
dkk9+rrhUy4qrZ+HFi7AeemybpiumbSnebvnkMaIPAOo23V8C9BQ0iuxx4gIZf10 o+TPSK8=
\---- END SSH2 PUBLIC KEY ----
NOT THIS --> Key Format seems to be PKCS1 so M2Crypto will NOT work (its load
key function expects PEM)
Latest reading I think it is SSH Public Key File Format (RFC:
<http://www.ietf.org/rfc/rfc4716.txt> )
I also think below it wrong, I don't think it handles SSH Public Key File
Format :(
Also looks like Twisted might be where I should look
<http://www.java2s.com/Open-
Source/Python/Network/Twisted/Twisted-1.0.3/Twisted-1.0.3/twisted/conch/ssh/keys.py.htm>
Also why does SO not allow me to post a bounty immediately??
Answer: Ok I found how to load it
from twisted.conch.ssh import keys as Keys
import base64
public_key = """\
---- BEGIN SSH2 PUBLIC KEY ----
Comment: "rsa-key-20101003"
AAAAB3NzaC1yc2EAAAABJQAAAgEAi+91fFsxZ7k1UuudSe5gZoavwARUyZScCtdf
WQ0ROoJC+XIqW5vVJfgmr+A1jLS5m4wNsrCqeyoX2B22T6iEwqVXrXt3QcbccKMu
WkLKFK1h67q6Coc+3eOTmKrOuZbWc19YQgybdkR/GxF7XAbq4NCGNaCDtMOqX8Q2
L/a9fAYqVdTwg9trpcz3whNmdLk/B0edOABKuVX51UdLV+ZggK503+uAb1JiIIj0
mARwR/HNo4oRLMLf2PjuZsGVYYjJDdVJBU6AN4PUQSRRRPL4+YmsrLJb/TpfJeXA
vj4KZMNJv15YXz7/iMZMKznDtr2RJX5wbSpuTUBNZveA7YiIHxvvvis38b/lX9SJ
SYPfZ9CeQY6MvQgG2zwDTOOvKgOIB4sTGMXfcoxB8AF/QXOcxWFJkZoj36rvMd9n
Po6szLjHXwcEUOUvvQfG4VvdQA0H5gGLHqYL1EehRsgi5qcCoFPaZW2K09ErKcS0
MbrLFjBkQ9KmqAM38bvM8UhCWAMA9VXOGHMxUHBV4Bir9alGS4VX0B8Y0b3dZ+7I
MKkHMCwdEUJf7QVdGxGuSQtVsq8RZbIpk3g7wtv8f6I/iEC58ekdrH35tq5+1ilW
dkk9+rrhUy4qrZ+HFi7AeemybpiumbSnebvnkMaIPAOo23V8C9BQ0iuxx4gIZf10
o+TPSK8=
---- END SSH2 PUBLIC KEY ----"""
key_data = ''.join(public_key.splitlines()[2:-1])# remove begin, end tags and comment
blob = base64.decodestring(key_data)
key = Keys.Key._fromString_BLOB(blob)
|
Is there a python equivalent of ruby's "Pathname" module?
Question: Ruby has this really handy module called [Pathname](http://ruby-
doc.org/core/classes/Pathname.html).
Is there a python equivalent to it?
Answer: `pathlib` is the answer to all your python path woos.
Example functionality:
from pathlib import Path
p = Path.cwd()
with (p/'somefile.txt').open() as f:
f.read()
p.is_dir()
|
Python and ADNS, falling in infinite loop somewhere
Question: I have written some code that queries adns. Problem with this code is that it
gets stuck, how? Let me explain it:
* Say my dnslist is ["8.8.4.4", "8.8.8.8", "208.67.220.220", "208.67.222.222", "192.168.50.1"]
* It would pop a dns from the list and query againt it, now that means that DNS will be queried in reverse order
* No matter what i do, It never shows results from the dns it picked up first (in our case 192.168.50.1)
* I was not sure if that dns ever replied so
* First i changed DNS list to contain just that last DNS Server and code executes fine
* Second i used the old list with 5 DNS servers except that the last one was managed by my so i could track if code even queries it or not, and to my surprise the query does take place.
* So query is made, we get result but that result is never inserted into resolved_hosts for some reason, and because that results is not inserted, its length will remain less than the length of dnslist, causing a infinite loop.
What do you think could be causing this problem and how to solve it?
# Code Execution Results
Inside class's init'
Data
host www.yahoo.com
dnslist length 5
intensity 1
Inside resolve()
inside finished_resolving()
Resolved : 0/5
Inside 'while not finished_resolving'
Queue: 0/1
Launching Querying for www.yahoo.com/1 on 192.168.50.1
Queue: 1/1
Launching Querying for www.yahoo.com/1 on 208.67.222.222
inside collect_results()
inside finished_resolving()
Resolved : 0/5
Inside 'while not finished_resolving'
------------------------ CLIPPED ----------------
Inside 'while not finished_resolving'
inside collect_results()
Inside collect_results's for query in self.adns.completed()
DNS used was208.67.222.222
Answered : (0, 'any-fp.wa1.b.yahoo.com', 1286169807, ('69.147.125.65', '67.195.160.76'))
Resolved www.yahoo.com to 69.147.125.65 using 208.67.222.222
Resolved hosts count1
And they are:
{'208.67.222.222': '69.147.125.65'}
inside finished_resolving()
Resolved : 1/5
Inside 'while not finished_resolving'
Queue: 1/1
Launching Querying for www.yahoo.com/1 on 208.67.220.220
inside collect_results()
inside finished_resolving()
Resolved : 1/5
Inside 'while not finished_resolving'
-------------------------- CLIPPED --------------------
inside collect_results()
Inside collect_results's for query in self.adns.completed()
DNS used was208.67.220.220
Answered : (0, 'any-fp.wa1.b.yahoo.com', 1286169790, ('67.195.160.76', '69.147.125.65'))
Resolved www.yahoo.com to 67.195.160.76 using 208.67.220.220
Resolved hosts count2
And they are:
{'208.67.222.222': '69.147.125.65', '208.67.220.220': '67.195.160.76'}
inside finished_resolving()
Resolved : 2/5
Inside 'while not finished_resolving'
Queue: 1/1
Launching Querying for www.yahoo.com/1 on 8.8.8.8
inside collect_results()
inside finished_resolving()
Resolved : 2/5
Inside 'while not finished_resolving'
-------------------------- CLIPPED --------------------
inside collect_results()
Inside collect_results's for query in self.adns.completed()
DNS used was8.8.8.8
Answered : (0, 'eu-fp.wa1.b.yahoo.com', 1286169758, ('87.248.122.122',))
Resolved www.yahoo.com to 87.248.122.122 using 8.8.8.8
Resolved hosts count3
And they are:
{'208.67.222.222': '69.147.125.65', '208.67.220.220': '67.195.160.76', '8.8.8.8': '87.248.122.122'}
inside finished_resolving()
Resolved : 3/5
Inside 'while not finished_resolving'
Queue: 1/1
Launching Querying for www.yahoo.com/1 on 8.8.4.4
inside collect_results()
inside finished_resolving()
Resolved : 3/5
Inside 'while not finished_resolving'
-------------------- CLIPPED -------------------------------------
inside collect_results()
Inside collect_results's for query in self.adns.completed()
DNS used was8.8.4.4
Answered : (0, 'eu-fp.wa1.b.yahoo.com', 1286169757, ('87.248.122.122',))
Resolved www.yahoo.com to 87.248.122.122 using 8.8.4.4
Resolved hosts count4
And they are:
{'208.67.222.222': '69.147.125.65', '208.67.220.220': '67.195.160.76', '8.8.8.8': '87.248.122.122', '8.8.4.4': '87.248.122.122'}
inside finished_resolving()
Resolved : 4/5
Inside 'while not finished_resolving'
inside collect_results()
inside finished_resolving()
Resolved : 4/5
---------------- CLIPPED -------------------------------
(last block keeps repeating until system starts to hang up, load goes upto 24)
# Code
**test.py**
import adns
from time import time
from async_dns import AsyncResolver
dnslist2 = ["8.8.4.4", "8.8.8.8", "208.67.220.220", "208.67.222.222", "192.168.50.1"] #192.168.50.1 is a dns server i manage
host = "www.yahoo.com"
record = adns.rr.A
intensity = len(dnslist2)/5+1
ar = AsyncResolver(dnslist2, host, record, intensity)
start = time()
resolved = ar.resolve()
end = time()
print "\n\n"
for dns, ip in resolved.items():
if ip is None:
print "%s could not resolv %s." % (dns, host)
else:
print "%s resolved it to %s : %s" % (dns, host, ip)
print "\n\n----------------------------------------------------"
print "It took %.2f seconds to query %d dns." % (end-start, len(dnslist))
print "----------------------------------------------------"
**async_dns.py**
#!/usr/bin/python
#
import sys
import adns
from time import time
class AsyncResolver(object):
def __init__(self, dnslist, host, record, intensity=10):
"""
dnslist: a list of dns used to resolve
host : hostname to resolve
record: recordtype to resolve
intensity: how many hosts to resolve at once
"""
print "Inside class's init'"
self.dnslist = dnslist
self.host = host
self.record = record
if intensity >= len(dnslist) :
self.intensity = len(dnslist)/5+1
else:
self.intensity = intensity
print "Data"
print "host " + host
print "dnslist length " + str(len(dnslist))
print "intensity " + str(intensity)
def resolve(self):
""" Resolves hosts and returns a dictionary of { 'dns': 'ip' }. """
print "Inside resolve()"
host = self.host
record = self.record
resolved_hosts = {}
active_queries = {}
dns_queue = self.dnslist[:]
def collect_results():
print "inside collect_results()"
for query in self.adns.completed():
print "Inside collect_results's for query in self.adns.completed()"
answer = query.check()
dns = active_queries[query]
print "DNS used was" + dns
print "Answered : " + str(answer)
del active_queries[query]
if answer[0] == 0:
#print "Inside answer[0] == 0 , ip:" + answer[3][0]
ip = answer[3][0]
resolved_hosts[dns] = ip
print "Resolved %s to %s using %s" % (host, ip, dns)
print "Resolved hosts count" + str(len(resolved_hosts))
print "And they are: "
print str(resolved_hosts)
print "\n"
elif answer[0] == 101 and not record == adns.rr.CNAME: # CNAME if CNAME wasn't required'
print "ooopppps, i got a CNAME, gotta find its A"
print "\n"
query = self.adns.submit(answer[1], adns.rr.A)
active_queries[query] = dns
else:
resolved_hosts[dns] = None
print "THIS COULD NOT BE RESOLVED"
def finished_resolving():
print "inside finished_resolving()"
print "Resolved : " + str(len(resolved_hosts)) + "/" + str(len(self.dnslist))
return len(resolved_hosts) == len(self.dnslist)
while not finished_resolving():
print "Inside 'while not finished_resolving'"
while dns_queue and len(active_queries) <= self.intensity:
print "Queue: " + str(len(active_queries)) + "/" + str(self.intensity)
dns = dns_queue.pop()
self.adns = adns.init(adns.iflags.noautosys,sys.stderr,"nameserver "+dns)
query = self.adns.submit(host, record)
print "Launching Querying for " + host + "/" + str(record) + " on " + dns
active_queries[query] = dns
collect_results()
return resolved_hosts
Answer: I have just attempted to use adns python binding and have experienced a
similar problem - the completed queries array never gets completely filled,
causing the check loop to run forever. In my case, it appeared that domains
that could not be resolved (due to NXDOMAIN etc) were never added to the
completed queries array (with empty/Null value)
I had a look at the C binding but could not see a separate data structure for
holding failed results, so it does appear that these should be added to the
'completed' queries somehow. I suspect there is a bug in the binding or adns
library preventing this from happening.
|
Python doctest example failure
Question: This is probably a silly question.
I am experimenting with python doctest, and I try to run
[this](http://docs.python.org/library/doctest.html#simple-usage-checking-
examples-in-a-text-file) example
ending with
if __name__ == "__main__":
import doctest
doctest.testfile("example.txt")
I have put "example.txt" in the same folder as the source file containing the
example code, but I get the following error:
Traceback (most recent call last):
File "test_av_funktioner.py", line 61, in <module>
doctest.testfile("example.txt")
File "C:\Python26\lib\doctest.py", line 1947, in testfile
text, filename = _load_testfile(filename, package, module_relative)
File "C:\Python26\lib\doctest.py", line 219, in _load_testfile
return open(filename).read(), filename
IOError: [Errno 2] No such file or directory: 'example.txt'
Can I somehow tell/set where the doctest module is searching for the specified
file?
Answer: Doctest searches relative to the calling module's directory by default (but
you can override this).
Quoting the docs for
[`doctest.testfile`](http://docs.python.org/library/doctest.html#doctest.testfile):
> Optional argument `module_relative` specifies how the filename should be
> interpreted:
>
> * If `module_relative` is **True** (the default), then `filename`
> specifies an OS-independent module-relative path. By default, this path is
> relative to the calling module’s directory; but if the `package` argument is
> specified, then it is relative to that package. To ensure OS-independence,
> `filename` should use `/` characters to separate path segments, and may not
> be an absolute path (i.e., it may not begin with `/`).
> * If `module_relative` is **False** , then `filename` specifies an OS-
> specific path. The path may be absolute or relative; relative paths are
> resolved with respect to the current working directory.
>
|
How to distinguish between a sequence and a mapping
Question: I would like to perform an operation on an argument based on the fact that it
might be a map-like object or a sequence-like object. I understand that no
strategy is going to be 100% reliable for type-like checking, but I'm looking
for a robust solution.
Based on this [answer](http://stackoverflow.com/questions/305359/correct-way-
to-detect-sequence-parameter/425567#425567), I know how to determine whether
something is a sequence and I can do this check after checking if the object
is a map.
def ismap(arg):
# How to implement this?
def isseq(arg):
return hasattr(arg,"__iter__")
def operation(arg):
if ismap(arg):
# Do something with a dict-like object
elif isseq(arg):
# Do something with a sequence-like object
else:
# Do something else
Because a sequence can be seen as a map where keys are integers, should I just
try to find a key that is not an integer? Or maybe I could look at the string
representation? or...?
**UPDATE**
I selected SilentGhost's answer because it looks like the most "correct" one,
but for my needs, here is the solution I ended up implementing:
if hasattr(arg, 'keys') and hasattr(arg, '__getitem__'):
# Do something with a map
elif hasattr(arg, '__iter__'):
# Do something with a sequence/iterable
else:
# Do something else
Essentially, I don't want to rely on an ABC because there are many custom
classes that behave like sequences and dictionary but that still do not extend
the python collections ABCs (see @Manoj comment). I thought the keys attribute
(mentioned by someone who removed his/her answer) was a good enough check for
mappings.
Classes extending the Sequence and Mapping ABCs will work with this solution
as well.
Answer:
>>> from collections import Mapping, Sequence
>>> isinstance('ac', Sequence)
True
>>> isinstance('ac', Mapping)
False
>>> isinstance({3:42}, Mapping)
True
>>> isinstance({3:42}, Sequence)
False
[`collections` abstract base classes
(ABCs)](http://docs.python.org/library/collections.html#abcs-abstract-base-
classes)
|
How to write tag deleter script in python
Question: I want to implement a file reader (folders and subfolders) script which
detects some tags and delete those tags from the files.
The files are .cpp, .h .txt and .xml And they are hundreds of files under same
folder.
I have no idea about python, but people told me that I can do it easily.
**EXAMPLE** :
My main folder is A: C:\A
Inside A, I have folders (B,C,D) and some files A.cpp A.h A.txt and A.xml. In
B i have folders B1, B2,B3 and some of them have more subfolders, and files
.cpp, .xml and .h....
* xml files, contains some tags like `<!-- $Mytag: some text$ -->`
* .h and .cpp files contains another kind of tags like `//$TAG some text$`
* .txt has different format tags: `#$This is my tag$`
It always starts and ends with $ symbol but it always have a comment character
(//,
The idea is to run one script and delete all tags from all files so the script
must:
* Read folders and subfolders
* Open files and find tags
* If they are there, delete and save files with changes
**WHAT I HAVE:**
import os
for root, dirs, files in os.walk(os.curdir):
if files.endswith('.cpp'):
%Find //$ and delete until next $
if files.endswith('.h'):
%Find //$ and delete until next $
if files.endswith('.txt'):
%Find #$ and delete until next $
if files.endswith('.xml'):
%Find <!-- $ and delete until next $ and -->
Answer: The general solution would be to:
1. use the `os.walk()` function to traverse the directory tree.
2. Iterate over the filenames and use `fn_name.endswith('.cpp')` with if/elseif to determine which file you're working with
3. Use the `re` module to create a regular expression you can use to determine if a line contains your tag
4. Open the target file and a temporary file (use the `tempfile` module). Iterate over the source file line by line and output the filtered lines to your tempfile.
5. If any lines were replaced, use `os.unlink()` plus `os.rename()` to replace your original file
It's a trivial excercise for a Python adept but for someone new to the
language, it'll probably take a few hours to get working. You probably
couldn't ask for a better task to get introduced to the language though. Good
Luck!
\----- Update -----
The `files` attribute returned by os.walk is a list so you'll need to iterate
over it as well. Also, the `files` attribute will only contain the base name
of the file. You'll need to use the `root` value in conjunction with
`os.path.join()` to convert this to a full path name. Try doing just this:
for root, d, files in os.walk('.'):
for base_filename in files:
full_name = os.path.join(root, base_filename)
if full_name.endswith('.h'):
print full_name, 'is a header!'
elif full_name.endswith('.cpp'):
print full_name, 'is a C++ source file!'
If you're using Python 3, the print statements will need to be function calls
but the general idea remains the same.
|
m2crypto: python 2.7 compatibility and which version of OpenSSL to use?
Question: We've been using M2crypto with Python 2.6 for Windows (32-bit) for some time
with great success. We used one of the user contributed setups to install
M2crypto in our development environments. We would like to move to Python 2.7,
but noticed there are no pre-built Python 2.7 setups for m2crypto.
Questions:
1. Is M2crypto 0.20.2 compatible with Python 2.7 or should we remain at Python 2.6 if we want to continue to use this library?
2. Does anyone have a user built setup for installing M2Crypto 0.20.2 in a Python 2.7 environment? (There are no 2.7 contributions on the M2crypto site and google comes up empty as well)
3. Can we copy our Python 2.6 M2Crypto files (under lib\site-packages\M2Crypto) to the same place in our Python 2.7 setup and avoid a formal installation process?
4. What version of OpenSLL should we use with M2crypto 0.20.2? I checked the OpenSSL website and there appear 2 versions of OpenSSL to choose from: 0.9.8 and 1.0.0a.
Thank you, Malcolm
Answer: 1. Yes, it's compatible with Python 2.7, so you can freely upgrade if you have not already.
2. Yes, here you have `[bdist_wininst](http://myfreefilehosting.com/f/23acda828e_0.33MB)`, `[bdist_egg](http://myfreefilehosting.com/f/e5e68275ff_0.23MB)` and `[bdist](http://myfreefilehosting.com/f/6235a73fed_0.24MB)` for M2Crypto 20.2 built for Python 2.7 with MSVS2008 by me, hope it will fit your needs.
3. No, you will get import error, as .pyd file (which is actually DLL) has `python26.dll` in it's import table, so this will not work. Of course, you can hack it and replace `python26` with `python27`, but that's too dirt and gives you no guarantee it will work all the time.
4. Version I uploaded for you works fine with OpenSSL 0.9.8o, haven't tested it with 1.0.0.
|
programmatically find and replace content dynamically in a string in python
Question: i need to find and replace patterns in a string with a dynamically generated
content.
lets say i want to find all strings within '' in the string and double the
string. a string like:
`my 'cat' is 'white'` should become my `'catcat' is 'whitewhite'`
all matches could also appear twice in the string.
thank you
Answer: Make use of the power of [regular
expressions](http://docs.python.org/library/re.html). In this particular case:
import re
s = "my 'cat' is 'white'"
print re.sub("'([^']+)'", r"'\1\1'", s) # prints my 'catcat' is 'whitewhite'
`\1` refers to the first group in the regex (called `$1` in some other
implementations).
|
Importing assemblies into IronPython from another assembly
Question: I have an IronPython 2.6/2.7 script I am writing which imports a lot of
assemblies.
In other words, at the top of the script it does this...
clr.AddReference( "System.Xml" )
import System.Xml
Except it doesn't do this for 1 assembly, but for 10 assemblies.
Some of the modules are built-in .NET assembllies and some are assemblies I
have made.
I'd like to simplify my script so that it loads one assembly that I will
build. I want to then call a method in that assembly that will do the
"AddReference" and "import" for the 10 assemblies. The primary goal of all
this is to minimize the length/complexity of the script.
So in the end I would see it working like this
clr.AddReferenceToFileAndPath( "d:\\myassembly" )
import MyAssembly
MyAssembly.ImportAllAssembliesIReallyWant()
My core problemis despite reading all the information I could find on
ScriptRuntime, ScriptEngine, scopes, etc. - I still can't figure out how to
write a method in "MyAssembly" that affects what modules are loaded in the
calling script.
Answer: One way to go about this would be to create a built-in module which does this.
You can do this with:
[assembly: PythonModule("mymodule", typeof(MyModuleType)]
public static class MyModuleType {
[SpecialName]
public static void PerformModuleReload(PythonContext context, PythonDictionary dict) {
context.DomainManager.LoadAssembly(typeof(TypeInAssemblyToLoad));
}
}
Just add appropriate LoadAssembly calls for all of the assemblies you care
about. The assembly could also populate members in dict that you want
available.
Another (and possibly simpler) way would be to simply have a .py file which
does all of the clr.AddReference calls you need and have every module import
that one file. The import mechanism will do the appropriate caching so it will
only load once but will ensure all of the assemblies are available for each
module which needs them.
|
How to bulk insert data to mysql with python
Question: Currently i'm using Alchemy as a ORM, and I look for a way to speed up my
insert operation, I have bundle of XML files to import
for name in names:
p=Product()
p.name="xxx"
session.commit()
i use above code to insert my data paser from batch xml file to mysql,it's
very slow also i tried to
for name in names:
p=Product()
p.name="xxx"
session.commit()
but it seems didn't change anything
Answer: You could bypass the ORM for the insertion operation and use the [SQL
Expression
generator](http://www.sqlalchemy.org/docs/core/tutorial.html?highlight=insert#insert-
expressions) instead.
Something like:
conn.execute(Product.insert(), [dict(name=name) for name in names])
That should create a single statement to do your inserting.
That example was taken from [lower down the same
page](http://www.sqlalchemy.org/docs/core/tutorial.html?highlight=insert#executing-
multiple-statements).
(I'd be interested to know what speedup you got from that)
|
Subsets and Splits