text
stringlengths
226
34.5k
How to retrieve nested map values Question: I want to scan AWS DynamoDB table and then pull only a certain value. Here is my code: package main import ( "fmt" "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/session" "github.com/aws/aws-sdk-go/service/dynamodb" ) func main() { svc := dynamodb.New(session.New(), &aws.Config{Region: aws.String("us-west-2")}) params := &dynamodb.ScanInput{ TableName: aws.String("my_Dynamo_table_name"), Limit: aws.Int64(2), } resp, err := svc.Scan(params) if err != nil { fmt.Println(err.Error()) return } fmt.Println(resp) } and The output is: { Count: 2, Items: [{ update_time: { N: "1466495096" }, create_time: { N: "1465655549" } },{ update_time: { N: "1466503947" }, create_time: { N: "1466503947" } }], LastEvaluatedKey: { Prim_key: { S: "1234567890" } }, ScannedCount: 2 } Now, I want to retrieve the `update_time` value for all elements in above output. Here are my attempts: for _, value := range resp.Items { fmt.Println(value["create_time"]["N"]) } and for _, value := range resp.Items { fmt.Println(value.create_time.N) } and for _, value := range resp.Items { fmt.Println(*value.create_time.N) } All above attempts error out with `/var/tmp/dynamo.go:37: invalid operation:` error. I am from perl/python background and recently started learning golang. How to retrieve nested map/array values in this case. Also, any reading references would be of great help. My google search did not reveal anything relevant. Answer: The value of `resp` above is of the type `*ScanOutput`, which has `Items` type as `[]map[string]*AttributeValue`. To access `update_time`, you can try: updateTimes := make([]string, 0) // Items is a slice of map of type map[string]*AttributeValue for _, m := range resp.Items { // m is of type map[string]*AttributeValue timeStrPtr := *m["update_time"].N updateTimes = append(updateTimes, *timeStrPtr) } `updateTimes` should now contains all the `"update_time"` values as strings. More details [here](https://godoc.org/github.com/aws/aws-sdk- go/service/dynamodb#ScanOutput).
Grouping up elements from a list in Python 2.7 Question: Ok, I got a huge text. I extract matches with regex (omitted here because it doesn't matter and I'm bad at this so you don't see how ugly my regex is :) ) and count them. Then, for readability, I split the elements and print them in the fashion I need: import re f = re.findall(r"(...)", PF) a = [[y,f.count(y)] for y in set(f)] (' '.join(map(str, j)) for j in w) for element in w: print element result is something like ['202', 1] ['213', 2] ['210', 2] ['211', 2] ['208', 2] ['304', 1] ['107', 2] ['133', 1] ['132', 1] ['131', 2] What I need is to group up the elements, so that I get an output like A ['133', 1] ['132', 1] ['131', 2] B ['202', 1] ['213', 2] C ['304', 1] ['107', 2] ['210', 2] ['211', 2] ['208', 2] Note that: * in the final result I will need 5 groups (A, B, C, D, E) * the elements can vary, for example tomorrow 131 might not be present but I might have 232 that goes in group A and the number of elements is different every day * it would be perfect, but not mandatory, if the elements in each group would be sorted numerically. * Might sound obvious but I'll make it clear anyway, I know exactly which elements need to go in which group. If it is of any help, group _A can contain (102, 103), B(104,105,106,201,202,203), C(204,205,206,301,302,303,304), D(107,108,109,110,208,209,210,211,213,305,306,307), E(131,132,133,231,232)_. The script needs to take the results that are present that day, compare them to the list above, and sort into the relative groups. Thanks in advance! Answer: You can set up a hash that maps elements to groups. You can then transform each array item from [element,count] to (group,element,count) _(using a tuple to make it more easily sortable and such)_. Sort that array, then use a loop or `reduce` to transform that into your final output. mapElementsToGroups = {'131': 'A', '202': 'B', '304': 'C', …} elementsFoundByGroup = {} for (group, element, count) in sorted( [(mapElementsToGroups[item[0]], item[1], item[2]) for item in a] ): elementsFoundByGroup[group] = elementsFoundByGroup.get(group, []) + [(element, count)] You now have a dictionary mapping each group name found to the list of elements and counts found within that group. The quick print is: print [ group + " " + elements.join("\n " + " "*len(group)) for (group,elements) in sorted(elementsFoundByGroup.items()) ].join("\n")
Python Simon Game: I Got Stuck Finishing the Sequence Question: I'm finally finishing my Simon Game, but I have some doubts about how to complete the sequence. Edit: As you asked I have edited my post. Here I will post my code as it was before this post. So here are my actual problems. 1) I don't know how to add one number to the sequence after each turn. 2) The first turn works fine, but I don't know how to start the next turn, I've tried to call again the function after checking the sequence but it doesn't worked. The necessary code for the two things: from Tkinter import * import Tkinter import random base = Tkinter.Tk() fr = Tkinter.Frame(base, bg='black', width='238', height='250') score = Tkinter.Label(base, bg='black', fg='white', text='Score:') score.place(x = 30, y = 15) s = 0 scoreNum = Tkinter.Label(base, bg='black', fg='white', text = s) scoreNum.place(x = 70, y = 15) clicks = [] color = 0 def yellowClick(): yellow.configure(activebackground='yellow3') yellow.after(500, lambda: yellow.configure(activebackground='yellow')) global clicks global color color = 1 clicks.append(color) yellow = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='yellow', bg='yellow3', command = yellowClick) yellow.place(x = 30, y = 50) def blueClick(): blue.configure(activebackground='medium blue') blue.after(500, lambda: blue.configure(activebackground='blue')) global clicks global color color = 2 clicks.append(color) blue = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='blue', bg='medium blue', command = blueClick) blue.place(x = 125, y = 50) def redClick(): red.configure(activebackground='red3') red.after(500, lambda: red.configure(activebackground='red')) global clicks global color color = 3 clicks.append(color) red = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='red', bg = 'red3', command = redClick) red.place(x = 30, y = 145) def greenClick(): green.configure(activebackground='dark green') green.after(500, lambda: green.configure(activebackground='green4')) global clicks global color color = 4 clicks.append(color) green = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='green4', bg='dark green', command = greenClick) green.place(x = 125, y = 145) def scoreUp(): global s s = s + 1 scoreNum.configure(text = s) sequence = [] def checkSequence(): global clicks global sequence if clicks == sequence: scoreUp() def showSequence(): global sequence global clicks global check r = random.randint(1, 4) if r == 1: yellow.configure(bg='yellow') yellow.after(1000, lambda: yellow.configure(bg='yellow3')) sequence.append(r) base.after(5000, checkSequence) elif r == 2: blue.configure(bg='blue') blue.after(1000, lambda: blue.configure(bg='medium blue')) sequence.append(r) base.after(5000, checkSequence) elif r == 3: red.configure(bg='red') red.after(1000, lambda: red.configure(bg='red3')) sequence.append(r) base.after(5000, checkSequence) elif r == 4: green.configure(bg='green4') green.after(1000, lambda: green.configure(bg='dark green')) sequence.append(r) base.after(5000, checkSequence) base.after(2000, showSequence) fr.pack() base.resizable(False, False) base.mainloop() The checkSequence function is activated by time because I could not find another way to do it. Answer: I've answered this question already but you still had some problems, here is the full and complete code. I've made it more efficient and fixed the problems :) This is working I've tested it. Just comment if there is anything you need to know import Tkinter # you don't need to import Tkinter again, from ... import * moves all the functions from ... into your program import random base = Tkinter.Tk() fr = Tkinter.Frame(base, bg='black', width='238', height='250') fr.pack() score = Tkinter.Label(base, bg='black', fg='white', text='Score:') score.place(x = 30, y = 15) scoreNum = Tkinter.Label(base, bg='black', fg='white', text = 0) scoreNum.place(x = 70, y = 15) global sequence global clicks global s sequence = [] clicks = [] s = 0 yellow = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='yellow', bg='yellow3', command = lambda *args: Click(yellow, "yellow", "yellow3", 1)) blue = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='blue', bg='medium blue', command = lambda *args: Click(blue, "blue", "medium blue", 2)) red = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='red', bg = 'red3', command = lambda *args: Click(red, "red", "red3", 3)) green = Tkinter.Button(base, bd='0', highlightthickness='0', width='7', height='5', activebackground='green4', bg='dark green', command = lambda *args: Click(green, "green", "dark green", 4)) yellow.place(x = 30, y = 50) blue.place(x = 125, y = 50) red.place(x = 30, y = 145) green.place(x = 125, y = 145) def Click(button, colour1, colour2, number): # these arguments change so that you don't have to copy the function 10 times global clicks yellow.configure(activebackground=colour2) yellow.after(500, lambda: yellow.configure(activebackground=colour1)) clicks.append(number) checkSequence() # why not do this straight away? def checkSequence(): global clicks global sequence global s print("clicks: "+str(clicks)) # debug print("sequence: "+str(sequence)) if clicks != sequence[:len(clicks)]: # if all their clicks so far e.g. [1, 2, 4, 3] DONT fit with the sequence e.g. [3, 2, 4, 3, 2, 1] print(" ...Incorrect!") s = 0 scoreNum.configure(text = 0) sequence = [] clicks = [] base.after(1000, showSequence) # now move to the next value in the sequence elif clicks == sequence: # they have completed a sequence print(" ...Match!") s = s + 1 scoreNum.configure(text = s) clicks = [] base.after(1000, showSequence) # now move to the next value in the sequence def showSequence(): global sequence global clicks r = random.randint(1, 4) if r == 1: yellow.configure(bg='yellow') yellow.after(1000, lambda: yellow.configure(bg='yellow3')) elif r == 2: blue.configure(bg='blue') blue.after(1000, lambda: blue.configure(bg='medium blue')) elif r == 3: red.configure(bg='red') red.after(1000, lambda: red.configure(bg='red3')) elif r == 4: green.configure(bg='green') green.after(1000, lambda: green.configure(bg='dark green')) sequence.append(r) # you don't need this in all the if statements, it always happens base.after(100, showSequence) base.resizable(False, False) base.mainloop() NOTE: I'm using python 3.4.1 so I changed all the tkinters to Tkinter
ValueError: script argument must be unicode. - Python3.5, SQLite3 Question: I'm writing following python script: import sqlite3 import sys if len(sys.argv) < 2: print("Error: You must supply at least SQL script.") print("Usage: %s table.db ./sql-dump.sql" % (sys.argv[0])) sys.exit(1) script_path = sys.argv[1] if len(sys.argv) == 3: db = sys.argv[2] else: db = ":memory:" try: con = sqlite3.connect(db) with con: cur = con.cursor() with open(script_path,'rb') as f: cur.executescript(f.read()) except sqlite3.Error as err: print("Error occured: %s" % err) I saved this program as **sqlite_import.py**. I have a database file named **test.db** and a SQL file **world.sql**. Now I tried to run program as: sqlite_import.py test.db world.sql But it shows me error like following: Traceback (most recent call last): File "C:\Users\Jarvis\OneDrive\Documents\Python\Python Data Visualization Cookbook, 2E\2 Knowing Your Data\sqlite_import.py", line 21, in <module> cur.executescript(f.read()) ValueError: script argument must be unicode. Help me to fix this. Answer: You opened the script file as _binary_ : with open(script_path,'rb') as f: This produces a `b'...'` bytes value, not a Unicode `str` object. Remove the `b`, and perhaps add an `encoding` argument to specific what codec to use to decode the data with.
Python: Properly formatting JSON parameters into a proper http request with markit on demand api Question: I am having trouble formatting my request properly in order to use the markitondemand InteractiveChart API. How can I properly do this? reference: <http://dev.markitondemand.com/MODApis/> Here is an example of a proper request: http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json?parameters=%7B%22Normalized%22%3Afalse%2C%22NumberOfDays%22%3A10%2C%22DataPeriod%22%3A%22Day%22%2C%22Elements%22%3A%5B%7B%22Symbol%22%3A%22AAPL%22%2C%22Type%22%3A%22price%22%2C%22Params%22%3A%5B%22c%22%5D%7D%5D%7D Here is my code constructing a request: import requests import json url2 = 'http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json' elements = [ { 'Symbol': 'GOOG', 'Type': 'price', 'Params': {'price': ['c']}, } ] req_obj = { 'Normalized': 'false', 'NumberOfDays': 3, 'DataPeriod': 'Day', 'Elements': elements } resp = requests.get(url2, params={'parameters': json.dumps(req_obj)}) and here is the output when I log the resp.url : http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json?parameters=%7B%22Elements%22%3A+%5B%7B%22Type%22%3A+%22price%22%2C+%22Params%22%3A+%7B%22price%22%3A+%5B%22c%22%5D%7D%2C+%22Symbol%22%3A+%22GOOG%22%7D%5D%2C+%22NumberOfDays%22%3A+3%2C+%22DataPeriod%22%3A+%22Day%22%2C+%22Normalized%22%3A+%22false%22%7D What am I doing wrong here? Answer: Your url end point (by the looks of it) should just be: url2 = 'http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart Then, make your request: resp = requests.get(url2, params=json.dumps({'parameters': req_obj})) Then `requests` will properly handle the `?` and any `&` if they were required.
UnicodeEncodeError Only When Script is Run as a Subprocess Question: I'm running my main script in Python 3.5 using the Spyder IDE, and I want to import functions from a script that happens to only work in Python 3.4. So I was recommended to run this second script as a subprocess like so: import subprocess cmd = [r'c:\python34\pythonw.exe', r'C:\users\John\Desktop\scraper.py'] p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE) stdout, stderr = p.communicate() print(stdout) print(stderr) The script being called is an example from NikolaiT's Web engine scraper: # -*- coding: utf-8 -*- import sys from GoogleScraper import scrape_with_config, GoogleSearchError from GoogleScraper.database import ScraperSearch, SERP, Link def basic_usage(): # See in the config.cfg file for possible values config = { 'SCRAPING': { 'use_own_ip': 'True', 'search_engines': 'baidu', 'num_pages_for_keyword': 3 }, 'keyword': '苹果', 'SELENIUM': { 'sel_browser': 'chrome', }, 'GLOBAL': { 'do_caching': 'False' } } try: sqlalchemy_session = scrape_with_config(config) except GoogleSearchError as e: print(e) # let's inspect what we got link_list = [] for serp in sqlalchemy_session.serps: #print(serp) for link in serp.links: #print(link) link_list.append(link.link) return link_list links = basic_usage() print("test") for link in links: print(link) This script works just fine when running it in Python 3.4's IDLE IDE, but when running it as a subprocess as above, I get the following UnicodeEncodeError printed from my main script: \python34\scripts\googlescraper\GoogleScraper\caching.py", line 413, in parse_all_cached_files store_serp_result(serp, self.config) File "c:\python34\scripts\googlescraper\GoogleScraper\output_converter.py", line 123, in store_serp_result pprint.pprint(data) File "c:\python34\lib\pprint.py", line 52, in pprint printer.pprint(object) File "c:\python34\lib\pprint.py", line 139, in pprint self._format(object, self._stream, 0, 0, {}, 0) File "c:\python34\lib\pprint.py", line 193, in _format allowance + 1, context, level) File "c:\python34\lib\pprint.py", line 268, in _format write(rep) File "c:\python34\lib\encodings\cp1252.py", line 19, in encode return codecs.charmap_encode(input,self.errors,encoding_table)[0] UnicodeEncodeError: \'charmap\' codec can\'t encode characters in position 1-2: character maps to undefined' Why would this only happen when running it indirectly? Thanks for any help, clarification questions, or suggestions for improvement in framing my question. Answer: In short, your issue is as follows: * Python IDE fully supports UTF-8 encoding, and that's why your invoked script runs fine there. * On the other hand, when opening a subprocess, since you are on Windows, you are using Windows1252 charset by default, which doesn't support part of your output. Quick solution: If you invoke the script as `python foobar.py` on Windows command line, you can run `chcp 65001` before invoking the script; if you use Spyder IDE, there should be a setting that allows you to set file / project encoding to UTF-8. (You may also try adding `# -*- coding: utf-8 -*-` to the top of your main Python script.)
File upload Selenium Web driver python in linux machine calling a remote machine Question: Hi I have scenario that needs to upload a file in a webpage. Actually I know that `selenium` will not support file upload scenario. But this can be done in python with external libraries such as `AUTOIT`, `PYWINAUTO`. But the challenge is i have to run my code in a linux server that is going to call a windows remote machine. When i tried installing `pywinauto` in linux server i got an error in importing winreg library. Hence i dont know how to proceed further. Please help me out to solve this scenario. Answer: Both AutoIt and pywinauto are Windows-only libraries (at least for now). If you need to automate file upload on Linux, consider using [AT-SPI accessibility](https://wiki.gnome.org/Accessibility) (say `pyatspi2` package). If it's a server without X and DBus, I think the question is about remote code execution from Linux to Windows. Good option for the SSH remote execution is [Fabric](http://www.fabfile.org/) (very pythonic & nice), but using Cygwin or OpenSSH might be an additional challenge for you. There are many other tools like Ansible etc.
Raspberry Pi getting data from MultiWii Question: I have a Raspberry Pi 3B and a CRIUS All in One Pro (v2.0) MultiWii flight controller. I'm using the MultiWii 2.4 version and the latest NOOBS. I was able to set up both fine, and now I am trying to get the Raspberry Pi to communicate with the MultiWii through a USB/Micro USB cable that connects the two boards. At the moment, the MultiWii isn't returning any data back, and I'm not sure why. From what I can see, I have the protocol correct. I've looked at several working code repos (written in Python of Java for Arduino), and have followed the [MultiWii documentation](http://www.multiwii.com/wiki/index.php?title=Multiwii_Serial_Protocol) and read through the associated [forum post](http://www.multiwii.com/forum/viewtopic.php?f=8&t=1516). Here is the client code that I wrote. package com.jmace.MaceDrone.msp; import com.pi4j.io.serial.Baud; import com.pi4j.io.serial.DataBits; import com.pi4j.io.serial.FlowControl; import com.pi4j.io.serial.Parity; import com.pi4j.io.serial.Serial; import com.pi4j.io.serial.SerialConfig; import com.pi4j.io.serial.SerialDataEvent; import com.pi4j.io.serial.SerialDataEventListener; import com.pi4j.io.serial.SerialFactory; import com.pi4j.io.serial.StopBits; import java.io.IOException; import java.math.BigInteger; public class MultiWiiClient { private final Serial serial; //The preamble is defined by the protocol. //Every message must begin with the characters $M private static final String PREAMBLE = "$M"; //Character that denotes information being passed to the MultiWii private static final char TO_MUTLIWII = '<'; //Character that denotes information being requested from by the MultiWii private static final char FROM_MUTLIWII = '>'; public MultiWiiClient(String usbPort) { SerialConfig config = new SerialConfig(); config.device(usbPort) .baud(Baud._115200) .dataBits(DataBits._8) .parity(Parity.NONE) .stopBits(StopBits._1) .flowControl(FlowControl.NONE); this.serial = SerialFactory.createInstance(); serial.addListener(new SerialDataEventListener() { @Override public void dataReceived(SerialDataEvent event) { try { System.out.println("[HEX DATA] " + event.getHexByteString()); System.out.println("[ASCII DATA] " + event.getAsciiString()); } catch (IOException e) { e.printStackTrace(); } } }); try { this.serial.open(config); } catch (Exception e) { e.printStackTrace(); } } public String sendRequest(MultiWiiRequest request) throws IllegalStateException, IOException { String message = createMessage(request.getId(), false, null); ////////////////////////////////////////////////////////////////////////////////// System.out.println(message); System.out.println(String.format("%040x", new BigInteger(1, message.getBytes()))); ////////////////////////////////////////////////////////////////////////////////// return sendMessage(message); } public String sendCommand(MultiWiiCommand command, String payload) throws IllegalStateException, IOException { String message = createMessage(command.getId(), true, payload); return sendMessage(message); } /** * This method creates the message that will be sent to the MultiWii * * Message format is as follows: * +--------+---------+----+-------+----+---+ * |preamble|direction|size|command|data|crc| * +--------+---------+----+-------+----+---+ * * Preamble (2 bytes): * Marks the start of a new message; always "$M" * * Direction (1 byte): * Either '<' for a command going to the MultiWii or '>' for * information being requested from the MultiWii * * Size (1 byte): * The number of bytes in the payload * * Command (1 byte): * The message ID of the command, as defined in the protocol * 100's for requesting data, and 200's for requesting an action * * Data (variable bytes): * The data to pass along with the command * * CRC (1 byte): * Calculated with an XOR of the size, command, and each byte of data */ private String createMessage(int mutliWiiCommandnumber, boolean isCommand, String payload) { StringBuilder message = new StringBuilder(PREAMBLE); byte checksum=0; //Get the direction of the message if (isCommand) { message.append(TO_MUTLIWII); } else { message.append(FROM_MUTLIWII); } int datalength = (payload != null) ? payload.length() : 0; message.append((char) datalength); checksum ^= datalength; message.append((char) mutliWiiCommandnumber); checksum ^= ((int) mutliWiiCommandnumber); if (payload != null) { for (char c : payload.toCharArray()){ message.append(c); checksum ^= (int) c; } } message.append((char) checksum); return message.toString(); } private String sendMessage(String message) throws IllegalStateException, IOException { serial.write(message.getBytes()); serial.flush(); System.out.println("TESTING ------------------"); return ""; } } I am using "/dev/ttyUSB0" to connect, which I confirmed is the correct location and it seems to be working (no errors when I run it; and if I start running and then disconnect the USB, it throws an exception because it lost the connection). When running, I get the following output (sending command 100, MSP_IDENT): $M>dd 00000000000000000000000000244d3e006464 TESTING ------------------ See [my Git repo](https://github.com/jmace01/MaceDrone/tree/master/src/com/jmace/MaceDrone/msp) for more code context. EDIT: Fixed the checksum code in my post Answer: It looks like it's all as it should be, though I can't see where you're receiving the version byte back from the serial port. in sendMessage, you have `return ""`. Looking at your github repo, I see that sendMessage is a bit different. while (!serial.getCTS()) { try{ Thread.sleep(100); } catch (Exception e) {} } StringBuilder response = new StringBuilder(); while (serial.available() != 0) { response.append(serial.read()); } serial.getCTS() shouldn't be used as you disabled flow control when you initialized the port in the constructor. I'm guessing that CTS is here used as a way to detect when the other device has finished processing the message and has responded. It seems there's no guarantee that after seeing the CTS line go high, that those bytes are actually available to be received; the CTS signal would only mean that the device has finished sending them away to you. You may wish to put another delay in after CTS goes high to ensure all bytes have indeed been transmitted.
Python - Unable to detect face and eye? Question: I am trying to create a face and eye detection using OpenCV library. This is the code I been working with. It's running smoothly with no errors but the only problem is not showing any results no faces and eyes are found with this code import cv2 import sys import numpy as np import os # Get user supplied values imagePath = sys.argv[1] # Create the haar cascade faceCascade = cv2.CascadeClassifier('C:\Users\Karthik\Downloads\Programs\opencv\sources\data\haarcascades\haarcascad_frontalface_default.xml') eyeCascade= cv2.CascadeClassifier('C:\Users\Karthik\Downloads\Programs\opencv\sources\data\haarcascades\haarcascade_eye.xml') # Read the image image = cv2.imread(imagePath) gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) # Detect faces in the image faces = faceCascade.detectMultiScale( gray, scaleFactor=1.2, minNeighbors=5, minSize=(30, 30), flags = cv2.cv.CV_HAAR_SCALE_IMAGE ) print "Found {0} faces!".format(len(faces)) # Draw a rectangle around the faces for (x, y, w, h) in faces: cv2.rectangle(image, (x, y), (x+w, y+h), (255, 0, 0), 2) roi_gray = gray[y:y+h, x:x+w] roi_color = image[y:y+h, x:x+w] eyes = eyeCascade.detectMultiscale(roi_gray) for (ex,ey,ew,eh) in eyes: cv2.rectangle(roi_color,(ex,ey),(ex+ew,ey+eh),(0, 255, 0), 2) cv2.imshow("Faces found", image) print image.shape cv2.waitKey(0) Answer: For me it works in my jupyter notebook on Ubuntu 15.10 using OpenCV 3.1.0-dev with python 3.4 Could it be, that you have a simple typo? `haarcascad_frontalface_default.xml` => `haarcascade_frontalface_default.xml` and here: `eyes = eyeCascade.detectMultiscale(roi_gray)` => `eyeCascade.detectMultiScale(roi_gray)` That's my working code: %matplotlib inline import matplotlib import matplotlib.pyplot as plt import cv2 import sys import numpy as np import os # Create the haar cascade faceCascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml') eyeCascade= cv2.CascadeClassifier('haarcascade_eye.xml') # Read the image image = cv2.imread('lena.png', 0) if image is None: raise ValueError('Image not found') # Detect faces in the image faces = faceCascade.detectMultiScale(image) print('Found {} faces!'.format(len(faces))) # Draw a rectangle around the faces for (x, y, w, h) in faces: cv2.rectangle(image, (x, y), (x+w, y+h), 255, 2) roi = image[y:y+h, x:x+w] eyes = eyeCascade.detectMultiScale(roi) for (ex,ey,ew,eh) in eyes: cv2.rectangle(roi,(ex,ey),(ex+ew,ey+eh), 255, 2) plt.figure() plt.imshow(image, cmap='gray') plt.show() [![enter image description here](http://i.stack.imgur.com/EzAjT.png)](http://i.stack.imgur.com/EzAjT.png)
ImportError: No module named 'theano' Question: I have installed (/library/python/2.7/site-packages) theano on my mac and still get this error. **My code is** import theano theano.test() **and the error** Traceback (most recent call last): File "/Users/mac/Downloads/n.py", line 1, in <module> import theano ImportError: No module named 'theano' I did install theano using pip install theano for python 2.7. Any idea on why its throwing up the error. Also I installed Anaconda and it throws the same error as working on idle. Answer: I had the same problem trying to install pygame, i'm using python 3.4.1 so I'm not 100% sure this will work for you Here is my folder: C:\Python34\Scripts[![enter image description here](http://i.stack.imgur.com/xoYJZ.png)](http://i.stack.imgur.com/xoYJZ.png) **The problem:** This will of course be different for you, but this is what I did anyway: I did `pip install pygame` and it seemed to work but when I tried `import pygame` I got the same error as you **My solution:** instead of installing pygame with pip, I did `pip3 install pygame` and it worked! so for you, look at what files you have in this folder **/library/python/2.7/site-packages** and try all of them, try `pip2 install theano`, try `pip2.7 install theano`, try `pip3 install theano`, try everything! Again I'm not certain this will work for python 2.7, but its worth a try :)
TypeError: 'module' object is not callable in Spacy Python Question: I want to print `Parse Tree` using `Spacy`. But the code below is giving the error > en_nlp = spacy.language('English') TypeError: 'module' object is not > callable The error is on this `en_nlp = spacy.loads('en')` line. I tried to shake off as `en_nlp = spacy.language(English)` by importing `from spacy.en import English` But still it is not working. Can someone help? Code: import spacy from nltk import Tree en_nlp = spacy.loads('en') doc = en_nlp("The quick brown fox jumps over the lazy dog.") def to_nltk_tree(node): if node.n_lefts + node.n_rights > 0: return Tree(node.orth_, [to_nltk_tree(child) for child in node.children]) else: return node.orth_ [to_nltk_tree(sent.root).pretty_print() for sent in doc.sents] Answer: Is it spacy.load('en') or spacy.loads('en') ? The official doc <https://spacy.io/docs/> says : spacy.load('en'). It may be the problem.
Bokeh plots not showing in nbviewer Question: I am working on some visualizations using Bokeh in a Jupyter (ipython) notebook. Though the plots run well within my notebook, it is important for me to make them accessible for users not running the code. I was counting on nbviewer for this, but am having trouble. Using a simple plot for example, I get this great output in my running notebook: [Running notebook in chrome](http://i.stack.imgur.com/dyJwF.png) After pushing to github I view the notebook in nbviewer but my plots are gone: <http://nbviewer.jupyter.org/github/kaplann/bokeh_trouble/blob/master/.ipynb_checkpoints/bokeh%20error- checkpoint.ipynb> I am using chrome Version 52.0.2743.82 (64-bit) on fedora 23, Bokeh - 0.12.1, Python 3.5.1, Anaconda 2.5.0 (64-bit), IPython - 5.1.0 I have a very strong feeling I am missing something, hopefully I am and the fix will be simple. Thanks! Answer: The notebook is an extremely difficult and challenging environment to develop and test Bokeh for. There have been some recent regressions and work to fix them, please refer to GitHub issues [#5014](https://github.com/bokeh/bokeh/issues/5014) and [#5081](https://github.com/bokeh/bokeh/issues/5081) for more information.
is "from flask import request" identical to "import requests"? Question: In other words, is the flask request class identical to the requests library? I consulted: <http://flask.pocoo.org/docs/0.11/api/> <http://docs.python-requests.org/en/master/> but cannot tell for sure. I see code examples where people seem to use them interchangeably. Answer: No these are not only completely different libraries, but completely different purposes. Flask is a web framework which clients make requests to. The Flask `request` object contains the data that the client (eg a browser) has sent to your app - ie the URL parameters, any POST data, etc. The requests library is for your app to make HTTP request to _other_ sites, usually APIs. It makes an outgoing request and returns the response from the external site.
Update a label in tkinter from a button press Question: My question is regarding GUI programming in python by using tkinter. I believe this is Python 3x. My question: While we're executing a program to run the GUI, can a button update a label? More specifically, is there a way to change the labels displayed text after pushing the button? I have consulted stack overflow on this before and adopted the StringVar() method, but it doesn't seem to fix my problem, in fact it omits the text from the GUI completely! Here is the code below from tkinter import * root = Tk() root.title('Copy Text GUI Program') copiedtext = StringVar() copiedtext.set("Text is displayed here") def copytext(): copiedtext.set(textentered.get()) # Write 'Enter Text Here' entertextLabel = Label(root, text="Enter Text Here") entertextLabel.grid(row=0, column=0) # For the user to write text into the gui textentered = Entry(root) textentered.grid(row=0, column=1) # The Copy Text Button copytextButton = Button(root, text="Copy Text") copytextButton.grid(row=1, columnspan=2) # Display the copied text displaytextLabel = Label(root, textvariable=copiedtext) displaytextLabel.grid(row=2,columnspan=2) copytextButton.configure(command=copytext()) root.mainloop() Any help would be appreciated! Answer: What you have to do is to bind a Button event to the copytextButton object like so: copytextButton.bind('<Button-1>', copytext) This means that a callback function - copytext() will be called when you left- click the button. A slight modification is needed in the function itself, since the callback sends an event argument: def copytext(event): copiedtext.set(textentered.get()) Edit: This line is not needed: copytextButton.configure(command=copytext())
How can this tensorflow model converge on a CPU but not on a GPU? Question: We ran into the strange problem that our relatively simple model converges on the CPU, but not on the server with GPU. No modifications to the code are done whatsoever between the two runs. Nor does the code contain any explicit conditional statements to change the workflow on different architectures. What could possibly be the reason? How can this tensorflow model converge on a CPU but not on a GPU? In the likely event that the code is too long for you to read we are still thankful about general speculations and hints. #!/usr/bin/python from __future__ import print_function import tensorflow as tf import os import numpy as np import input_data # copy from tensorflow/examples/tutorials/mnist/input_data.py # wget https://raw.githubusercontent.com/tensorflow/tensorflow/master/tensorflow/examples/tutorials/mnist/input_data.py if needed mnist = input_data.read_data_sets("/tmp/data/", one_hot=True) force_gpu = False debug = True # histogram_summary ... # _cpu='/cpu:0' default_learning_rate=0.001 tensorboard_logs = '/tmp/tensorboard-logs/' # $(sleep 5; open http://0.0.0.0:6006) & tensorboard --debug --logdir=/tmp/tensorboard-logs/ class net(): def __init__(self,model,data,name=0,learning_rate=default_learning_rate,batch_size=64): self.session=sess=session=tf.Session() self.model=model self.data=data # assigned to self.x=net.input via train self.batch_size=batch_size self.layers=[] self.last_width=self.input_width(data) self.learning_rate=learning_rate self.generate_model(model) def generate_model(self,model, name=''): if not model: return self with tf.name_scope('state'): self.keep_prob = tf.placeholder(tf.float32) # 1 for testing! else 1 - dropout self.train_phase = tf.placeholder(tf.bool, name='train_phase') self.global_step = tf.Variable(0) # dont set, feed or increment global_step, tensorflow will do it automatically with tf.name_scope('data'): n_input=28*28 n_classes=10 self.x = x = self.input = tf.placeholder(tf.float32, [None, n_input]) self.last_layer=x self.y = y = self.target = tf.placeholder(tf.float32, [None, n_classes]) if not force_gpu: tf.image_summary("mnist", tf.reshape(self.x, [-1, 28, 28, 1], "mnist_images")) with tf.name_scope('model'): model(self) if(self.last_width!=n_classes): self.classifier() # 10 classes auto def input_width(self,data): return 28*28 def add(self, layer): self.layers.append(layer) self.last_layer = layer self.last_shape = layer.get_shape() def reshape(self,shape): self.last_layer = tf.reshape(self.last_layer,shape) self.last_shape = shape self.last_width = shape[-1] def batchnorm(self): from tensorflow.contrib.layers.python.layers import batch_norm as batch_norm with tf.name_scope('batchnorm') as scope: input = self.last_layer train_op=batch_norm(input, is_training=True, center=False, updates_collections=None, scope=scope) test_op=batch_norm(input, is_training=False, updates_collections=None, center=False,scope=scope, reuse=True) self.add(tf.cond(self.train_phase,lambda:train_op,lambda:test_op)) # Fully connected layer def dense(self, hidden=1024, depth=1, act=tf.nn.tanh, dropout=False, parent=-1): # if parent==-1: parent=self.last_layer shape = self.last_layer.get_shape() if shape and len(shape)>2: self.last_width= int(shape[1]*shape[2]*shape[3]) print("reshapeing ",shape,"to",self.last_width) parent = tf.reshape(parent, [-1, self.last_width]) width = hidden while depth>0: with tf.name_scope('Dense_{:d}'.format(hidden)) as scope: print("Dense ", self.last_width, width) nr = len(self.layers) # if self.last_width == width: # M = closest_unitary(np.random.rand(self.last_width, width) / (self.last_width + width)) # weights = tf.Variable(m, name="weights_dense_" + str(nr)) # else: weights = tf.Variable(tf.random_uniform([self.last_width, width], minval=-1. / width, maxval=1. / width), name="weights_dense") bias = tf.Variable(tf.random_uniform([width],minval=-1./width,maxval=1./width), name="bias_dense") dense1 = tf.matmul(parent, weights, name='dense_'+str(nr))+ bias tf.histogram_summary('dense_'+str(nr),dense1) tf.histogram_summary('weights_'+str(nr),weights) tf.histogram_summary('bias_'+str(nr),bias) tf.histogram_summary('dense_'+str(nr)+'/sparsity', tf.nn.zero_fraction(dense1)) tf.histogram_summary('weights_'+str(nr)+'/sparsity', tf.nn.zero_fraction(weights)) if act: dense1 = act(dense1) # if norm: dense1 = self.norm(dense1,lsize=1) # SHAPE! if dropout: dense1 = tf.nn.dropout(dense1, self.keep_prob) self.layers.append(dense1) self.last_layer = parent = dense1 self.last_width = width depth=depth-1 self.last_shape=[-1,width] # dense # Convolution Layer def conv(self,shape,act=tf.nn.relu,pool=True,dropout=False,norm=True,name=None): # True why dropout bad in tensorflow?? with tf.name_scope('conv'): print("input shape ",self.last_shape) print("conv shape ",shape) width=shape[-1] filters=tf.Variable(tf.random_normal(shape)) # filters = tf.Variable(tf.random_uniform(shape, minval=-1. / width, maxval=1. / width), name="filters") _bias=tf.Variable(tf.random_normal([shape[-1]])) # # conv1 = conv2d('conv', _X, _weights, _bias) conv1=tf.nn.bias_add(tf.nn.conv2d(self.last_layer,filter=filters, strides=[1, 1, 1, 1], padding='SAME'), _bias) if debug: tf.histogram_summary('conv_' + str(len(self.layers)), conv1) if act: conv1=act(conv1) if pool: conv1 = tf.nn.max_pool(conv1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME') if norm: conv1 = tf.nn.lrn(conv1, depth_radius=4, bias=1.0, alpha=0.001 / 9.0, beta=0.75) if debug: tf.histogram_summary('norm_' + str(len(self.layers)), conv1) if dropout: conv1 = tf.nn.dropout(conv1,self.keep_prob) print("output shape ",conv1.get_shape()) self.add(conv1) def classifier(self,classes=10): # Define loss and optimizer with tf.name_scope('prediction'):# prediction if self.last_width!=classes: # print("Automatically adding dense prediction") self.dense(hidden=classes, act= False, dropout = False) # cross_entropy = -tf.reduce_sum(y_*y) with tf.name_scope('classifier'): y_=self.target manual=False # True if classes>100: print("using sampled_softmax_loss") y=prediction=self.last_layer self.cost = tf.reduce_mean(tf.nn.sampled_softmax_loss(y, y_)) # for big vocab elif manual: # prediction = y =self.last_layer=tf.nn.softmax(self.last_layer) # self.cost = cross_entropy = -tf.reduce_sum(y_ * tf.log(y+ 1e-10)) # against NaN! prediction = y = tf.nn.log_softmax(self.last_layer) self.cost = cross_entropy = -tf.reduce_sum(y_ * y) else: y = prediction = self.last_layer self.cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(y, y_)) # prediction, target # if not gpu: tf.scalar_summary('cost', self.cost) # self.cost = tf.Print(self.cost , [self.cost ], "debug cost : ") learning_scheme=self.learning_rate # learning_scheme=tf.train.exponential_decay(self.learning_rate, self.global_step, decay_steps, decay_size) self.optimizer = tf.train.AdamOptimizer(learning_scheme).minimize(self.cost) # Evaluate model correct_pred = tf.equal(tf.argmax(prediction, 1), tf.argmax(self.target, 1)) self.accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32)) if not force_gpu: tf.scalar_summary('accuracy', self.accuracy) # Launch the graph def next_batch(self,batch_size=10): return self.data.train.next_batch(batch_size) def train(self,steps=-1,dropout=None,display_step=10,test_step=200): #epochs=-1, steps = 9999999 if steps==-1 else steps session=self.session # with tf.device(_cpu): # import tensorflow.contrib.layers as layers # t = tf.verify_tensor_all_finite(t, msg) tf.add_check_numerics_ops() self.summaries = tf.merge_all_summaries() self.summary_writer = tf.train.SummaryWriter(tensorboard_logs, session.graph) # if not dropout:dropout=1. # keep all x=self.x y=self.y keep_prob=self.keep_prob session.run([tf.initialize_all_variables()]) step = 0 # show first while step < steps: # print("step %d \r" % step)# end=' ') batch_xs, batch_ys = self.next_batch(self.batch_size) # tf.train.shuffle_batch_join(example_list, batch_size, capacity=min_queue_size + batch_size * 16, min_queue_size) # Fit training using batch data feed_dict = {x: batch_xs, y: batch_ys, keep_prob: dropout, self.train_phase: True} loss,_= session.run([self.cost,self.optimizer], feed_dict=feed_dict) if step % test_step == 0: self.test(step) if step % display_step == 0: # Calculate batch accuracy, loss feed = {x: batch_xs, y: batch_ys, keep_prob: 1., self.train_phase: False} acc , summary = session.run([self.accuracy,self.summaries], feed_dict=feed) # self.summary_writer.add_summary(summary, step) # only test summaries for smoother curve print("\rStep {:d} Loss= {:.6f} Accuracy= {:.3f}".format(step,loss,acc),end=' ') if str(loss)=="nan": return print("\nLoss gradiant explosion, exiting!!!") #restore! step += 1 print("\nOptimization Finished!") self.test(step,number=10000) # final test def inputs(self,data): self.inputs, self.labels = load_data()#...) def test(self,step,number=400):#256 session=sess=self.session run_metadata = tf.RunMetadata() run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE) # Calculate accuracy for 256 mnist test images test_labels = self.data.test.labels[:number] test_images = self.data.test.images[:number] feed_dict = {self.x: test_images, self.y: test_labels, self.keep_prob: 1., self.train_phase:False} accuracy,summary= self.session.run([self.accuracy, self.summaries], feed_dict=feed_dict) # accuracy,summary = session.run([self.accuracy, self.summaries], feed_dict, run_options, run_metadata) print('\t'*3+"Test Accuracy:",accuracy) # self.summary_writer.add_run_metadata(run_metadata, 'step #%03d' % step) self.summary_writer.add_summary(summary,global_step=step) def dense(net): # best with lr ~0.001 # type: (layer.net) -> None # net.batchnorm() # start lower, else no effect # net.dense(400,act=None)# # ~95% we can do better: net.dense(400, act=tf.nn.tanh)# 0.996 YAY only 0.985 on full set, Step 5000 flat return # 0.957% without any model!! def alex(net): # type: (layer.net) -> None print("Building Alex-net") net.reshape(shape=[-1, 28, 28, 1]) # Reshape input pictures # net.batchnorm() net.conv([3, 3, 1, 64]) net.conv([3, 3, 64, 128]) net.conv([3, 3, 128, 256]) net.dense(1024,act=tf.nn.relu) net.dense(1024,act=tf.nn.relu) # net=layer.net(dense,data=mnist, learning_rate=0.01 )#,'mnist' baseline _net=net(alex,data=mnist, learning_rate=0.001)#,'mnist' _net.train(50000,dropout=0.6,display_step=1,test_step=10) Answer: In general floating point computations can be a little nondeterministic when it comes to adding many numbers (and some GPUs are buggy). Did you try retuning hyperparameters (varying learning rates and whatnot) to account for this?
how to process an upload file in Flask? Question: I have a simple Flask app and I would like for it to process an uploaded excel file and display it's data in the webpage. So far I got a page to upload the excel file. main.py from flask import Flask, render_template, send_file, request from flask_uploads import UploadSet, configure_uploads, DOCUMENTS, IMAGES app = Flask(__name__) #the name 'datafiles' must match in app.config to DATAFILES docs = UploadSet('datafiles', DOCUMENTS) app.config['UPLOADED_DATAFILES_DEST'] = 'static/uploads' configure_uploads(app, docs) @app.route("/", methods=['GET']) def index(): # return send_file("templates/index.html") return render_template('index.html') @app.route("/upload", methods = ['GET', 'POST']) def upload(): #user_file is the name value in input element if request.method == 'POST' and 'user_file' in request.files: filename = docs.save(request.files['user_file']) return filename return render_template('upload.html') @app.route("/mergedata", methods=['GET']) def merge_data(): return send_file("templates/mergeDataPage.html") if __name__ == "__main__": app.run(host='0.0.0.0', debug=True) upload.html <!DOCTYPE html> <html> <head> <meta charset="utf-8"> <title>upload</title> </head> <body> <form method="POST" enctype=multipart/form-data action ="{{url_for('upload')}}"> <input type="file" name="user_file"> <input type="submit"> </form> </body> </html> Once the file is uploaded, I would like to process this excel data by using another python script - This script will take the excel file, remove special characters and converted to a csv file/output. I would like to display the csv output in the webpage. The idea is for users to upload the file, press a button to clean the data and have the results printed in an output box. How do I go about in building something like this scenario? I'm pretty new to Flask, but so far has been pretty exciting to learn this cool web framework. Your help will be appreciated, thanks! Answer: You can use [pandas](http://pandas.pydata.org/) module and first convert the excel file. import pandas as pd data_xls = pd.read_excel('your_workbook.xls', 'Sheet1', index_col=None) data_xls.to_csv('your_csv.csv', encoding='utf-8') Pandas is spectacular for dealing with csv files, and the following code would be all you need to read a csv and save an entire column into a variable: import pandas as pd df = pd.read_csv(csv_file) saved_column = df.column_name #you can also use df['column_name'] For instance if you wanted to save all of the info in your column Names into a variable, this is all you need to do: names = df.Names
compare two results (of many) from api data with django/python Question: I'm learning django/python/css/etc... and while doing this, I've decided to build an app for my website that can pull simple movie data from TMDb. What I'm having trouble with is figuring out a way to add a way for the **user to select two different movies, and once selected, see the differences between them (run time, budget, etc).** I've got grabbing the data from the API covered in that doing a search for a movie on my site returns expected results. But now I'm having a really tough time trying to figure out **how to select 1 item from the results to "save" it, search again, select the second movie, and have the comparison show up.** I know it's pretty vague, but any help getting me pointed in the right direction would be greatly appreciated! here's what I'm doing so far with the code: views.py: from django.shortcuts import render from django.conf import settings from .forms import MovieSearch import tmdbsimple as tmdb tmdb.API_KEY = settings.TMDB_API_KEY def search_movie(request): """ Search movie title and return 5 pages of results """ parsed_data = {'results': []} if request.method == 'POST': form = MovieSearch(request.POST) if form.is_valid(): search = tmdb.Search() query = form.cleaned_data['moviename'] response = search.movie(query=query) for movie in response['results']: parsed_data['results'].append( { 'title': movie['title'], 'id': movie['id'], 'poster_path': movie['poster_path'], 'release_date': movie['release_date'][:-6], 'popularity': movie['popularity'], 'overview': movie['overview'] } ) for i in range(2, 5 + 1): response = search.movie(query=query, page=i) for movie in response['results']: parsed_data['results'].append( { 'title': movie['title'], 'id': movie['id'], 'poster_path': movie['poster_path'], 'release_date': movie['release_date'][:-6], 'popularity': movie['popularity'], 'overview': movie['overview'] } ) context = { "form": form, "parsed_data": parsed_data } return render(request, './moviecompare/movies.html', context) else: form = MovieSearch() else: form = MovieSearch() return render(request, './moviecompare/compare.html', {"form": form}) def get_movie(request, movid): """ from search/movie results, get details by movie id (movid) """ movie = tmdb.Movies(movid) response = movie.info() context = { 'response': response } return render(request, './moviecompare/detail.html', context) movies.html: {% extends 'moviecompare/compare.html' %} {% block movies_returned %} <div class="wrap"> <div class="compare-gallery"> {% for key in parsed_data.results|dictsortreversed:'release_date' %} {% if key.poster_path and key.release_date and key.title and key.overview %} <div class="gallery-item"> <img src="http://image.tmdb.org/t/p/w185/{{ key.poster_path }}"> <div class="gallery-text"> <div class="gallery-date"><h5><span><i class="material-icons">date_range</i></span> {{ key.release_date }}</h5></div> <div class="gallery-title"><h3><a href="{% url 'compare:movie_detail' movid=key.id %}">{{ key.title }}</a></h3></div> <div class="gallery-overview">{{ key.overview|truncatechars:80 }}</div> </div> </div> {% endif %} {% endfor %} </div> </div> {% endblock %} and I've got a simple detail.html that doesn't do anything just yet, but it's just for showing detail results for a single movie, so not important to get into yet as it'll just be styling. I'd like for each result in the gallery to have a link to a details page for the movie(done), and also a select box (or similar) to select it as one of the comparison movies. If I can just get some help on how to select two different movies from the search results, and compare those, I think I could work out a way to do the same with two separate movie searches. Thanks for any help! edit: here's what I have so far on [pythonanywhere](http://moniblog.pythonanywhere.com/compare/) \- Answer: Assuming you want to be able to add movies from different searches (e.g. search for rambo, add rambo, search for south park, add south park), you can probably do something like: @require_POST def save_compare(request, pk): """ Async endpoint to add movies to comparison list """ movies = request.session.get('comparing_moviews', []) movies.append(movies) # if you only need the data the API served originally, you could also save that to a data-base when you read it from the API and use it here request.session['comparing_movies'] = movies return JsonResponse("") def show_comparison(request): """ show differing stats like "1 hour 3 mins longer """ mov1, mov2 = request.session.get('comparing_movies') # need to handle wrong number of selected # not sure from the description how to look up the movies by id mov1 = tmdb.Search().search(pk=mov1) mov2 = tmdb.Search().search(pk=mov2) return render_to_response("comparison.html", context={ "mov1": mov1, "mov2": mov2, "length_diff": mov1['length'] - mov2['length']}) to give you the comparison stats you want. You'll also need [session middleware](https://docs.djangoproject.com/en/1.10/topics/http/sessions/) installed. From the search results page, you'll add a button that triggers an async js post to `save_compare` which starts accumulating a comparison list. When they're done, they can click "compare" or something on that list, which will render "comparison.html" with the two movies and the diff however you like it. As an example of some calculation for your "comparison.html", you can grab the hours and minutes from the `length_diff` and render them as "X hours and y minutes longer."
How to download japanese history stock prices automatically from google finance in python Question: I use python to analyze Japanese stock prices. I want to get Japanese historical stock prices to get from google finance. I also refer to googlefinance0.7 ( <https://pypi.python.org/pypi/googlefinance> ) and pandas, but they are not support Japanese stock prices. So how to download japanese history stock prices automatically from google finance? Or are there references to program code in python? Answer: [Quandl](https://www.quandl.com/data/TSE) has all the historical data sets you need. It's designed for easy usage with Python so for [Tokyo Stock Exchange TOPIX Index](https://www.quandl.com/data/TSE/TOPIX-Tokyo-Stock-Exchange-TOPIX-Index) : import quandl mydata = quandl.get("TSE/TOPIX")
numpy ndarray with more that 32 dimensions Question: When I try to create an numpy array with more than 32 dimensions I get an error: import numpy as np np.ndarray([1] * 33) --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-2-78103e601d91> in <module>() ----> 1 np.ndarray([1] * 33) ValueError: sequence too large; cannot be greater than 32 I found this: [Using numpy.array with large number of dimensions](http://stackoverflow.com/questions/35192518/using-numpy-array- with-large-number-of-dimensions) to be related to this question but I want to do this without building my own version. My use case: I am working with Joint Probability Distributions and I am trying to represent each variable on an axis so that computations on it (marginalize, reduce) is a single line operation. For example for a marginalize operation I can simply do a sum over the axis of that variable. For multiplication I can simply do a simple numpy multiplication (after checking if the axes are the same). Is there a possible workaround this? Answer: How about maintaining the data in a 1d, and selectively reshaping to focus on a given dimension To illustrate In [252]: x=np.arange(2*3*4*5) With full reshape In [254]: x.reshape(2,3,4,5).sum(2) Out[254]: array([[[ 30, 34, 38, 42, 46], [110, 114, 118, 122, 126], [190, 194, 198, 202, 206]], [[270, 274, 278, 282, 286], [350, 354, 358, 362, 366], [430, 434, 438, 442, 446]]]) With partial reshape - same numbers, different result shape In [255]: x.reshape(6,4,5).sum(1) Out[255]: array([[ 30, 34, 38, 42, 46], [110, 114, 118, 122, 126], [190, 194, 198, 202, 206], [270, 274, 278, 282, 286], [350, 354, 358, 362, 366], [430, 434, 438, 442, 446]]) I'm not going to test this on anything approaching 32+ dimensions. As was noted in the comment, if many of the dimensions are larger than 1 the overall array size will be excessively large.
python3 global name 'SOL_SOCKET is not defined Question: when I use `gevent`, I can not use `requests`, is my useage wrong? from gevent import monkey; monkey.patch_all() import gevent, requests requests.get('https://haofly.net') it raise the error: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 70, in get return request('get', url, params=params, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 56, in request return session.request(method=method, url=url, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 475, in request resp = self.send(prep, **send_kwargs) File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 596, in send r = adapter.send(request, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/adapters.py", line 423, in send timeout=timeout File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 595, in urlopen chunked=chunked) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 352, in _make_request self._validate_conn(conn) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 831, in _validate_conn conn.connect() File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connection.py", line 289, in connect ssl_version=resolved_ssl_version) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/util/ssl_.py", line 308, in ssl_wrap_socket return context.wrap_socket(sock, server_hostname=server_hostname) File "/usr/local/lib/python3.3/site-packages/gevent/_ssl3.py", line 60, in wrap_socket _context=self) File "/usr/local/lib/python3.3/site-packages/gevent/_ssl3.py", line 143, in __init__ if sock.getsockopt(SOL_SOCKET, SO_TYPE) != SOCK_STREAM: NameError: global name 'SOL_SOCKET' is not defined The version of my Python is 3.3 If I add `ssl=False` to `monkey.patch_all()`, it returns: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 70, in get return request('get', url, params=params, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 56, in request return session.request(method=method, url=url, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 475, in request resp = self.send(prep, **send_kwargs) File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 596, in send r = adapter.send(request, **kwargs) File "/usr/local/lib/python3.3/site-packages/requests/adapters.py", line 423, in send timeout=timeout File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 595, in urlopen chunked=chunked) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 352, in _make_request self._validate_conn(conn) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 831, in _validate_conn conn.connect() File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connection.py", line 289, in connect ssl_version=resolved_ssl_version) File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/util/ssl_.py", line 308, in ssl_wrap_socket return context.wrap_socket(sock, server_hostname=server_hostname) File "/usr/local/lib/python3.3/ssl.py", line 245, in wrap_socket _context=self) File "/usr/local/lib/python3.3/ssl.py", line 335, in __init__ server_hostname) TypeError: _wrap_socket() argument 1 must be _socket.socket, not SSLSocket Answer: This is really a bug in gevent. It is trying to import a [`socket` constant](https://docs.python.org/3/library/socket.html#constants) from the `ssl` library. That specific constant _happens_ to be imported into the `ssl` library because that library makes use of it. But it is only there because a fix for [issue 19422](http://bugs.python.org/issue19422) required it to be imported in to the `ssl` module. The fix for that bug was included in Python 3.3.4, and you have version 3.3.3 and you thus don't have that specific constant available in the `ssl` module as well. You could just fix it by adding `from socket import SOL_SOCKET, SO_TYPE` in `gevent/_ssl3.py` (located in the `/usr/local/lib/python3.3/site-packages` directory on your system). Or you can upgrade to Python 3.3.4 or newer. I've filed this as [issue #856](https://github.com/gevent/gevent/issues/856) with the `gevent` project.
Extract hyper-cubical blocks from a numpy array with unknown number of dimensions Question: I have a bit of python code which currently is hard-wired with two-dimensional arrays as follows: import numpy as np data = np.random.rand(5, 5) width = 3 for y in range(0, data.shape[1] - W + 1): for x in range(0, data.shape[0] - W + 1): block = data[x:x+W, y:y+W] # Do something with this block Now, this is hard coded for a 2-dimensional array and I would like to extend this to 3D and 4D arrays. I could, of course, write more functions for other dimensions but I was wondering if there is a python/numpy trick to generate these sub-blocks without having to replicate this function for multidimensional data. Answer: Here is my wack at this problem. The idea behind the code below is to find the "starting indices" for each slice of data. So for 4x4x4 sub-arrays of a 5x5x5 array, the starting indices would be `(0,0,0), (0,0,1), (0,1,0), (0,1,1), (1,0,0), (1,0,1), (1,1,1)`, and the slices along each dimension would be of length 4. To get the sub-arrays, you just need to iterate over the different tuples of slice objects and pass them to the array. import numpy as np from itertools import product def iterslice(data_shape, width): # check for invalid width assert(all(sh>=width for sh in data_shape), 'all axes lengths must be at least equal to width') # gather all allowed starting indices for the data shape start_indices = [range(sh-width+1) for sh in data_shape] # create tuples of all allowed starting indices start_coords = product(*start_indices) # iterate over tuples of slice objects that have the same dimension # as data_shape, to be passed to the vector for start_coord in start_coords: yield tuple(slice(coord, coord+width) for coord in start_coord) # create 5x5x5 array arr = np.arange(0,5**3).reshape(5,5,5) # create the data slice tuple iterator for 3x3x3 sub-arrays data_slices = iterslice(arr.shape, 3) # the sub-arrays are a list of 3x3x3 arrays, in this case sub_arrays = [arr[ds] for ds in data_slices]
Discovering data type of incoming socket data in python Question: There are couple of devices which are sending socket data over _TCP/IP_ to socket server. Some of the devices are sending data as _Binary encoded Hexadecimal string_ , others are _ASCII string_. Eg.; If device sending data in _ASCII string_ type, script is begin to process immediately without any conversion. If device sending _Binary encoded HEX string_ , script should has to convert _Binary encoded Hex string_ into _Hex string_ first with; data = binascii.hexlify(data) There are two scripts running for different data types for that simple single line. But, I think this could be done in one script if script be aware of the incoming data type. Is there a way to discover type of the incoming socket data in Python? Answer: If you can you should make the sending devices signal what data they are sending eg by using different TCP ports or prepending each message with an `"h"` for hex or an `"a"` for ascii - possibly even use an established protocol like [XML-RPC](https://docs.python.org/3/library/xmlrpc.html) Actually you can only be sure in _some_ cases as all hex-encoded strings are valid ascii and some ascii-strings are valid hex like `"CAFE"`. You can make sure you can decode a string as hex with import string def is_possibly_hex(s): return all(c in string.hexdigits for c in s) or import binascii def is_possibly_hex(s): try: binascii.unhexlify(s) except binascii.Error: return False return True
Running python on server, executing commands from computer Question: I made a login system with python. It works perfectly, but i want to run script on server or web. For example: Steam. Steam wants username and password to log in. So i wanted to do the same for my script. How can i do that? My Code: import os import string import time version = "1.0 Alfa" def login(): print ("----------------------------------------") print (" Login ") print ("----------------------------------------") k_name = input("Enter username: ") if os.path.exists(k_name + ".txt") == False: print ("Username not found.") create() else: k_pass = input("Enter password: ") with open(k_name + ".txt", "r") as f: if k_pass == f.read(): print("Welcome %s!"%k_name) f.close() input() else: print("Password is wrong!") create() def create(): print("You using login system %s" % version) print( "----------------------------------------") print("| Lobby |") print( "----------------------------------------") starting = input("To create user type R, to login type L").upper() if starting == "R": name = input("Enter username: ") password = input("Enter password: ") password2 = input("Enter password again: ") if password == password2: newfile = open(name + ".txt", "w") newfile.write(password) newfile.close() print("User created. Redirecting you to login.") time.sleep(2) login() elif password != password2: print("Passwords doesn't match.") input() create() elif starting == "L": login() else: print("\nWrong button\n") create() create() Answer: Here is the script which you can run : python test.py user pass it will save data if file not found and perform login #!/usr/bin/env python import sys, getopt import os import time version = "1.0 Alfa" def login(username=None, password=None): print ("----------------------------------------") print (" Login ") print ("----------------------------------------") if username: k_name = username else: k_name = input("Enter username: ") if os.path.exists(k_name + ".txt") == False: print ("Username not found.") create(username, password, "R") else: if password: k_pass = password else: k_pass = input("Enter password: ") with open(k_name + ".txt", "r") as f: if k_pass == f.read(): if not username: print("Welcome %s!"%k_name) f.close() input() else: print("Password is wrong!") create() def create(username=None, password=None, mode="L"): print("You using login system %s" % version) print( "----------------------------------------") print("| Lobby |") print( "----------------------------------------") if mode: starting = mode else: starting = input("To create user type R, to login type L").upper() if starting == "R": if username: name = username else: name = input("Enter username: ") if password: password2 = password else: password = input("Enter password: ") password2 = input("Enter password again: ") if password == password2: newfile = open(name + ".txt", "w") newfile.write(password) newfile.close() print("User created. Redirecting you to login.") time.sleep(2) login(username, password) elif password != password2: print("Passwords doesn't match.") input() create() elif starting == "L": login(username, password) else: print("\nWrong button\n") create() def main(argv): print sys.argv if len(sys.argv) < 3: print 'test.py <username> <password>' sys.exit() username = sys.argv[1] password = sys.argv[2] print 'username is ', username print 'password is ', password create(username, password) if __name__ == "__main__": main(sys.argv[1:])
Convert a string to JSON Question: I would like to convert this string to a JSON dict: {u'Processes': [[u'root', u'3606', u'0.0', u'0.2', u'76768', u'16664', u'?', u'Ss', u'20:40', u'0:01', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4088', u'0.0', u'0.2', u'88544', u'20156', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4090', u'0.0', u'0.2', u'88552', u'20140', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4097', u'0.0', u'0.2', u'88552', u'20112', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4110', u'0.0', u'0.2', u'88548', u'20160', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0']], u'Titles': [u'USER', u'PID', u'%CPU', u'%MEM', u'VSZ', u'RSS', u'TTY', u'STAT', u'START', u'TIME', u'COMMAND']} So I used json.dumps and json.loads but the output was not a valid JSON. I understand that JSON needs double quotes instead of simple quotes, but I don't think that the solution is to search and replace is the best way to resolve this problem. Is there a more proper way to do it ? Answer: Use [`ast.literal_eval`](https://docs.python.org/3.5/library/ast.html#ast.literal_eval) to convert string to valid Python object. > Safely evaluate an expression node or a string containing a Python literal > or container display. The string or node provided may only consist of the > following Python literal structures: strings, bytes, numbers, tuples, lists, > dicts, sets, booleans, and None. s = "{u'Processes': [[u'root', u'3606', u'0.0', u'0.2', u'76768', u'16664', u'?', u'Ss', u'20:40', u'0:01', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4088', u'0.0', u'0.2', u'88544', u'20156', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4090', u'0.0', u'0.2', u'88552', u'20140', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4097', u'0.0', u'0.2', u'88552', u'20112', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4110', u'0.0', u'0.2', u'88548', u'20160', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0']], u'Titles': [u'USER', u'PID', u'%CPU', u'%MEM', u'VSZ', u'RSS', u'TTY', u'STAT', u'START', u'TIME', u'COMMAND']}" o = ast.literal_eval(s) assert 'Processes' in o Use [`json.dumps`](https://docs.python.org/3.5/library/json.html#json.dumps) to dump it to JSON string. import json json.dumps(o) # '{"Titles": ["USER", "PID", "%CPU", "%MEM", "VSZ", "RSS", "TTY", "STAT", "START", "TIME", "COMMAND"], "Processes": [["root", "3606", "0.0", "0.2", "76768", "16664", "?", "Ss", "20:40", "0:01", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4088", "0.0", "0.2", "88544", "20156", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4090", "0.0", "0.2", "88552", "20140", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4097", "0.0", "0.2", "88552", "20112", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4110", "0.0", "0.2", "88548", "20160", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"]]}' Or use `json.dump` to dump it to file object, if that's what you want.
imports python3 in another folder Question: When I run launch.py it fails and when I run main.py directly it works. launch.py just imports and runs main.py. Why? ├── dir │   ├── bla.py │   ├── __init__.py │   └── main.py ├── __init__.py └── launch.py launch.py --------- #!/usr/bin/env python # -*- coding: utf-8 -*- from dir import main main.main() main.py ------- #!/usr/bin/env python # -*- coding: utf-8 -*- import bla bla.pront() bla.py ------ def pront(): print('pront') EDITED: [enter image description here](http://i.stack.imgur.com/cgfvU.png) Answer: Using your layout and with the following files, we don't have problems. ### launch.py #!/usr/bin/env python # -*- coding: utf-8 -*- from dir import main if __name__ == "__main__": main. main() ### main.py #!/usr/bin/env python # -*- coding: utf-8 -*- try: from . import bla except: import bla def main(): bla.pront() if __name__ == "__main__": main() The `try ... except` structure is used in case the **_main.py_** was used inside or outside the package. Of course, there is a lot of info about it. You can start with [this](http://stackoverflow.com/questions/16981921/relative-imports-in- python-3).
embedding resources in python scripts Question: I'd like to figure out how to embed binary content in a python script. For instance, I don't want to have any external files around (images, sound, ... ), I want all this content living inside of my python scripts. Little example to clarify, let's say I got this small snippet: from StringIO import StringIO from PIL import Image, ImageFilter embedded_resource = StringIO(open("Lenna.png", "rb").read()) im = Image.open(embedded_resource) im.show() im_sharp = im.filter(ImageFilter.SHARPEN) im_sharp.show() As you can see, the example is reading the external file 'Lenna.png' [![enter image description here](http://i.stack.imgur.com/3p4GL.png)](http://i.stack.imgur.com/3p4GL.png) **Question** How to proceed to embed "Lenna.png" as a resource (variable) into my python script. What's the fastest way to achieve this simple task using python? Answer: The best way to go about this is converting your picture into a python string, and have it in a separate file called something like `resources.py`, then you simply parse it. If you are looking to embed the whole thing inside a single binary, then you're looking at something like [py2exe](http://py2exe.org/). [Here](http://www.py2exe.org/index.cgi/data_files) is an example embedding external files In the first scenario, you could even use `base64` to (de)code the picture, something like this: import base64 file = open('yourImage.png'); encoded = base64.b64encode(file.read()) data = base64.b64decode(encoded) # Don't forget to file.close() !
Shut down Flask SocketIO Server Question: For circumstances outside of my control, I need to use the Flask server to serve basic html files, the Flask SocketIO wrapper to provide a web socket interface between any clients and the server. The `async_mode` has to be `threading` instead of `gevent` or `eventlet`, I understand that it is less efficient to use threading, but I can't use the other two frameworks. In my unit tests, I need to shut down and restart the web socket server. When I attempt to shut down the server, I get the `RunTimeError` 'Cannot stop unknown web server.' This is because the function `werkzeug.server.shutdown` is not found in the Flask Request Environment `flask.request.environ` object. Here is how the server is started. SERVER = flask.Flask(__name__) WEBSOCKET = flask_socketio.SocketIO(SERVER, async_mode='threading') WEBSOCKET.run(SERVER, host='127.0.0.1', port=7777) Here is the short version of how I'm attempting to shut down the server. client = WEBSOCKET.test_client(SERVER) @WEBSOCKET.on('kill') def killed(): WEBSOCKET.stop() try: client.emit('kill') except: pass The stop method must be called from within a flask request context, hence the weird kill event callback. Inside the stop method, the `flask.request.environ` has the value 'CONTENT_LENGTH' (40503696) = {str} '0' 'CONTENT_TYPE' (60436576) = {str} '' 'HTTP_HOST' (61595248) = {str} 'localhost' 'PATH_INFO' (60437104) = {str} '/socket.io' 'QUERY_STRING' (60327808) = {str} '' 'REQUEST_METHOD' (40503648) = {str} 'GET' 'SCRIPT_NAME' (60437296) = {str} '' 'SERVER_NAME' (61595296) = {str} 'localhost' 'SERVER_PORT' (61595392) = {str} '80' 'SERVER_PROTOCOL' (65284592) = {str} 'HTTP/1.1' 'flask.app' (65336784) = {Flask} <Flask 'server'> 'werkzeug.request' (60361056) = {Request} <Request 'http://localhost/socket.io' [GET]> 'wsgi.errors' (65338896) = {file} <open file '<stderr>', mode 'w' at 0x0000000001C92150> 'wsgi.input' (65338848) = {StringO} <cStringIO.StringO object at 0x00000000039902D0> 'wsgi.multiprocess' (65369288) = {bool} False 'wsgi.multithread' (65369232) = {bool} False 'wsgi.run_once' (65338944) = {bool} False 'wsgi.url_scheme' (65338800) = {str} 'http' 'wsgi.version' (65338752) = {tuple} <type 'tuple'>: (1, 0) My question is, how do I set up the Flask server to have the `werkzeug.server.shutdown`method available inside the flask request contexts? Also this is using Python 2.7 Answer: I have good news for you, the testing environment does not use a real server, in that context the client and the server run inside the same process, so the communication between them does not go through the network as it does when you run things for real. Really in this situation there is no server, so there's nothing to stop. It seems you are starting a real server, though. For unit tests, that server is not used, all you need are your unit tests which import the application and then use a test client to issue socket.io events. I think all you need to do is just not start the server, the unit tests should run just fine without it if all you use is the test client as you show above.
assert vs == for testing code in Python? Question: What is the need for importing unittest and running assertTrue (for example) while testing a python function instead of writing a usual python function with == True check for testing? What is the new thing about unittesting, as even the test cases have to be written by user, which can be checked by == instead of assert family of functions in unittest? My question basically is: assert in unittest vs the equality check operation == in Python. Answer: I would say that the most interesting thing is the "easy-to-test" that the unittest package (as well as others, for example pytest) provides. When using unittest you can just run a command and check if your refactoring or your last feature broke something on the rest of the program. I would read something about [test-driven development](https://en.wikipedia.org/wiki/Test- driven_development) to understand better what is the importance of unittesting.
My numpy is latest but tensroflow says it's old one Question: $pip list numpy(1.11.1) My numpy is latest and I am sure it could be used in python environment. >>> import numpy >>> print numpy.__version__ 1.11.1 however I use tensorflow $ tensorboard --logdir=/tmp/basic_rnn/logdir it shows this error RuntimeError: module compiled against API version 0xa but this version of numpy is 0x9 What causes this problem?? Answer: I suspect that the `tensorboard` command is linked against the wrong Python environment. Make sure that `tensorflow` uses the same Python interpreter/virtual environment as you checked your numpy installation with. Also check your `PATH` and `PYTHONPATH` and take a look at `sys.path` and `sys.prefix` to see if there are some paths mixed up.
python import * or a list from other level Question: I'm trying to import a few classes from a module in another level. I can type all the classes, but I' trying to do it dynamically if I do: from ..previous_level.module import * raise: SyntaxError: import * only allowed at module level the same from myapp folder: from myapp.previous_level.module import * raise: SyntaxError: import * only allowed at module level So I thought: my_classes = ['Class_Foo', 'Class_Bar'] for i in my_classes: from ..previous_level.module import i raise: ImportError: cannot import name 'i' and also: my_classes = ['Class_Foo', 'Class_Bar'] for i in my_classes: __import__('myapp').previous_level.module.y raise: AttributeError: module 'myapp.previous_level.module' has no attribute 'y' I've tried `string format` , `getattr()` , `__getattr__` but no success. It's impossible to do import this way, or I'm doing something wrong? Answer: The error `SyntaxError: import * only allowed at module level` happens if you try to import `*` from within a function. This is not visible from the question, but it seems that the original code was simething like: def func(): from ..previous_level.module import * # SyntaxError On the other hand, importing at the beginning of a module would be valid: from ..previous_level.module import * # OK That said, I would still suggest using absolute imports without `*`: from myapp.previous_level.module import ClassA, ClassB, ClassC x = ClassA() # or: import myapp.previous_level.module x = myapp.previous_level.module.ClassA() BTW, this is completely wrong: for i in my_classes: from ..previous_level.module import i
get picture from dynamic content python Question: I'm trying to get the href of the picture from an url without using selenium def(): try: page = urllib2.urlopen('') except httplib.IncompleteRead, e: page = e.partial response = BeautifulSoup(page) print response var = response.find("div", {"id":"il_m"}).find('p') but i got None as a result.What should I do to ge the href ? Answer: You can also get the link from an _anchor_ tag with a _download_ attribute: In [2]: from bs4 import BeautifulSoup In [3]: import urllib2 In [4]: r = urllib2.urlopen('http://icecat.us/index.php/product/image_gallery?num=9010647&id=9409545&lang=us&imgrefurl=philips.com') In [5]: soup = BeautifulSoup(r,"html.parser") In [6]: print(soup.select_one("p a[download]")["href"]) http://images.icecat.biz/img/gallery/9010647-Philips-_FP.jpg You should also take note of the text _Images may be subject to copyright._. on the page.
Russian character decoding in python Question: This question only for python: I have a city name in a string in Russian language and which is in Unicode form like, > `\u041C\u043E\u0441\u043A\u0432\u0430` means > `Москва` How to get original text instead of unicode characters? **Note:** Do not use any import module Answer: >>> a=u"\u041C\u043E\u0441\u043A\u0432\u0430" >>> print a Москва Your string is a unicode string because each character/code point with \u is only usable from a unicode string, you should prefix the string with u. Otherwise is a regular string and each \u counts as a regular ascii character: >>> len(a) 6 >>> b="\u041C\u043E\u0441\u043A\u0432\u0430" >>> len(b) 36
I'm not sure what a certain code line does Question: I was trying to have a program using Python to create a "pyramid" based on a number, n, out of o's and came up with this: ( I would print nn, that would be the lines.) import time n = 0 while True: n += 2 #just another way to show n=n+2 nn = n, "o" #nn would be an amount of o's, based on what #n was time.sleep(1) print (nn, str.center(40, " ")) Not sure how to make nn the o's and not sure what line 6 does either. Does anyone know the answer to either question? ( I'm not in a class just programming for fun.) Answer: import time n = 0 while True: n += 2 #just another way to show n=n+2 nn = "o" * n #nn would be an amount of o's, based on what #n was time.sleep(1) print (nn.center(40," ")) As someone has mentioned in the comments "o" * n will give a a string containing n "o"s. I've fixed the print line to use the correct method a calling the center method.
Selenium Webdriver Python: I can't seem to locate all the text in a label tag Question: <label class="control-label"> Rental Charge: <span class="required" ng-show="vm.rentalInfo.reason">* (Min of $30.00)</span> </label> I used driver.find_element_by_xpath("//label[@ class = 'control-label']/span[@class = 'required']").text and the only `text` I get back is the `asterisk *`, `(Min of $30.00)` is not displayed. Does anyone know why this is? Answer: I am guessing this is a timing issue. [Wait](http://selenium- python.readthedocs.io/waits.html#explicit-waits) for "$" to be present in element explicitly with `text_to_be_present_in_element` expected condition: from selenium.webdriver.common.by import By from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.support import expected_conditions as EC wait = WebDriverWait(driver, 10) elm = wait.until(EC.text_to_be_present_in_element((By.CSS_SELECTOR, 'span[ng-show*="rentalIn‌​fo.reason"]'), "$")) print(elm.text)
Run multiple servers in python at same time (Threading) Question: I have **2 servers** in python, **I want to mix them up in one single .py and run together** : Server.py: import logging, time, os, sys from yowsup.layers import YowLayerEvent, YowParallelLayer from yowsup.layers.auth import AuthError from yowsup.layers.network import YowNetworkLayer from yowsup.stacks.yowstack import YowStackBuilder from layers.notifications.notification_layer import NotificationsLayer from router import RouteLayer class YowsupEchoStack(object): def __init__(self, credentials): "Creates the stacks of the Yowsup Server," self.credentials = credentials stack_builder = YowStackBuilder().pushDefaultLayers(True) stack_builder.push(YowParallelLayer([RouteLayer, NotificationsLayer])) self.stack = stack_builder.build() self.stack.setCredentials(credentials) def start(self): self.stack.broadcastEvent(YowLayerEvent(YowNetworkLayer.EVENT_STATE_CONNECT)) try: logging.info("#" * 50) logging.info("\tServer started. Phone number: %s" % self.credentials[0]) logging.info("#" * 50) self.stack.loop(timeout=0.5, discrete=0.5) except AuthError as e: logging.exception("Authentication Error: %s" % e.message) if "<xml-not-well-formed>" in str(e): os.execl(sys.executable, sys.executable, *sys.argv) except Exception as e: logging.exception("Unexpected Exception: %s" % e.message) if __name__ == "__main__": import sys import config logging.basicConfig(stream=sys.stdout, level=config.logging_level, format=config.log_format) server = YowsupEchoStack(config.auth) while True: # In case of disconnect, keeps connecting... server.start() logging.info("Restarting..") App.py: import web urls = ( '/', 'index' ) app = web.application(urls, globals()) class index: def GET(self): greeting = "Hello World" return greeting if __name__ == "__main__": app.run() I want to **run both together from single .py file together**. If I try to run them from one file, **either of the both starts and other one starts only when first one is done working.** **How can I run 2 servers** in python together? Answer: import thread def run_app1(): #something goes here def run_app2(): #something goes here if __name__=='__main__': thread.start_new_thread(run_app1) thread.start_new_thread(run_app2) if you need to pass args to the functions you can do: thread.start_new_thread(run_app1, (arg1,arg2,....)) if you want more control in your threads you could go: import threading def app1(): #something here def app2(): #something here if __name__=='__main__': t1 = threading.Thread(target=app1) t2 = threading.Thread(target=app2) t1.start() t2.start() if you need to pass args you can go: t1 = threading.Thread(target=app1, args=(arg1,arg2,arg3.....)) What's the differences between thread vs threading? Threading is higher level module than thread and in 3.x thread got renamed to _thread... more info here: <http://docs.python.org/library/threading.html> but that's for another question I guess. So in your case, just make a function that runs the first script, and the second script, and just spawn threads to run them.
Selenium webdriver unable to restart after unexpected exit Question: I haven't been able to start up an instance of python's selenium webdriver after my last use a few days ago. According to the error messages, it unexpectedly quit last time I was using it, and now, after restarting my macbook, uninstalling and reinstalling chromedriver/selenium: `brew rmtree chromedriver && brew install chromedriver` `pip uninstall selenium && pip install selenium` I'm still in the same place. It seems to be selenium itself, because for both Firefox and Chrome, I'm getting error messages. What I'm attempting to run on the python3.5 kernel is: from selenium import webdriver driver = webdriver.Chrome() stacktrace: File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/chrome/webdriver.py", line 62, in __init__self.service.start() File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/common/service.py", line 86, in start self.assert_process_still_running() File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/common/service.py", line 99, in assert_process_still_running % (self.path, return_code) selenium.common.exceptions.WebDriverException: Message: Service chromedriver unexpectedly exited. Status code was: -5 likewise, for Firefox: from selenium import webdriver driver = webdriver.Firefox() and: File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/webdriver.py", line 80, in __init__ self.binary, timeout) File "/Users/myuser/webscraping/env/lib/python3.5/site-packagesyuser/selenium/webdriver/firefox/extension_connection.py", line 52, in __init__ self.binary.launch_browser(self.profile, timeout=timeout) File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/firefox_binary.py", line 68, in launch_browser self._wait_until_connectable(timeout=timeout) File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/firefox_binary.py", line 99, in _wait_until_connectable "The browser appears to have exited " selenium.common.exceptions.WebDriverException: Message: The browser appears to have exited before we could connect. If you specified a log_file in the FirefoxBinary constructor, check it for details. Should I be looking for some rogue process to kill with `ps -e` and `kill -sigint`? Answer: I discovered that homebrew chromedriver was throwing an error related to symlinking the correct dylib. I fixed the issue by following the steps in [this](http://stackoverflow.com/questions/17643509/conflict-between-dynamic- linking-priority-in-osx#answer-35070568) answer to get chromedriver running again, which enabled selenium/chrome webdriver to work as well.
django: link to detail page from list of returned results Question: I'm creating a page that returns a list of movies with basic details after a user search. **After the search, I'd like the user to be able to click on a movie, and get more details about it.** here's a link to the site: (be gentle, I only started learning this stuff 2 months ago! hah) <http://moniblog.pythonanywhere.com/compare/> the data is coming from [TMDB's api](https://www.themoviedb.org/documentation/api) and the initial "generic" search JSON response doesn't have specific details that I'd display on the movie's detail page, but it has an ID that will be used for the specific movie's search. I'm only using views.py to grab/display the search results, and I'm not sure if this is the right way to go, or if I should use a model, but that's probably a different question. forms.py: from django import forms class MovieSearch(forms.Form): moviename = forms.CharField(label="Search", max_length=250) views.py: from django.shortcuts import render, get_object_or_404 from django.conf import settings from .forms import MovieSearch import tmdbsimple as tmdb tmdb.API_KEY = settings.TMDB_API_KEY def search_movie(request): parsed_data = {'results': []} if request.method == 'POST': form = MovieSearch(request.POST) if form.is_valid(): search = tmdb.Search() query = form.cleaned_data['moviename'] response = search.movie(query=query) for movie in response['results']: parsed_data['results'].append( { 'title': movie['title'], 'id': movie['id'], 'poster_path': movie['poster_path'], 'release_date': movie['release_date'][:-6], 'popularity': movie['popularity'], 'overview': movie['overview'] } ) for i in range(2, 5 + 1): response = search.movie(query=query, page=i) for movie in response['results']: parsed_data['results'].append( { 'title': movie['title'], 'id': movie['id'], 'poster_path': movie['poster_path'], 'release_date': movie['release_date'][:-6], 'popularity': movie['popularity'], 'overview': movie['overview'] } ) context = { "form": form, "parsed_data": parsed_data } return render(request, './moviecompare/movies.html', context) else: form = MovieSearch() else: form = MovieSearch() return render(request, './moviecompare/compare.html', {"form": form}) and html: {% extends 'moviecompare/compare.html' %} {% block movies_returned %} <div class="wrap"> <div class="compare-gallery"> {% for key in parsed_data.results|dictsortreversed:'release_date' %} {% if key.poster_path and key.release_date and key.title and key.overview %} <div class="gallery-item"> <img src="http://image.tmdb.org/t/p/w185/{{ key.poster_path }}"> <div class="gallery-text"> <div class="gallery-date"><h5><span><i class="material-icons">date_range</i></span> {{ key.release_date }}</h5></div> <div class="gallery-title"><h3><a href="../detail/{{ key.id }}">{{ key.title }}</a></h3></div> <div class="gallery-overview">{{ key.overview|truncatechars:80 }}</div> </div> </div> {% endif %} {% endfor %} </div> </div> {% endblock %} I've started playing with the urls.py to get something to work, but no luck so far. site's urls.py: urlpatterns = [ url(r'^$', home, name="home"), url(r'^blog/', include('blog.urls', namespace='blog')), url(r'^compare/', include('moviecompare.urls', namespace='compare')), url(r'^movies/', include('moviecompare.urls', namespace='movies')), and the app's urls.py: from django.conf.urls import url from . import views urlpatterns = [ url(r'^', views.search_movie, name='compare'), url(r'^(?P<movid>[0-9])+$', views.get_movie, name='movies') ] edit: adding my 1st (failed) attempt at the movie detail view: def get_movie(request, movid=None): instance = get_object_or_404(request, movid=movid) context = { 'instance': instance } return render(request, './moviecompare/detail.html', context) Answer: I think you should try by fixing the urls.py in this line: url(r'^(?P<movid>[0-9])+$', views.get_movie, name='movies') Move the "+" inside the brackets like that: url(r'^(?P<movid>[0-9]+)$', views.get_movie, name='movies')
I am using beautifulsoup in python 3. But "html.parser" not give me the all code of website Question: My code is import urllib import urllib.request from bs4 import BeautifulSoup fed = "https://www.fedex.com/apps/fedextrack/?action=track&tracknumbers=870915037012296&cntry_code=us&wsch=true" show = urllib.request.urlopen(fed) soup = BeautifulSoup(show,"html.parser") print(soup) I am trying to fetch the delivery address from fedex and status.But this code not fetching all the data of my URL Answer: Why don't use [Fedex SOAP API](http://www.fedex.com/us/developer/) directly. There is a Python wrapper - [`python-fedex`](http://python- fedex.readthedocs.io/en/latest/quick_start.html). Example how to track a package can be found [here](http://python- fedex.readthedocs.io/en/latest/quick_start.html#track-a-package).
Jenkins git triggered build not blocking Question: I am running a build on commit to `origin/master` on my jenkins server that is deploying resources to Amazon AWS. I am using the Execute Shell section to run a python script that handles all unit testing/linting/validation/deployment and everything blocks fine until it gets to the deployment (`deploy.deploy()`) where it returns a success right after kickoff, but doesn't complete deploying. How can I make this block? For reference here is my config: **Execute Shell (Jenkins)** : export DEPLOY_REGION=us-west-2 . build-tools/get_aws_credentials.sh python build-tools/kickoff.py **kickoff.py** if __name__ == "__main__": build_tools_dir="{}".format("/".join(os.path.abspath(__file__).split("/")[0:-1])) sys.path.append(build_tools_dir) base_dir = "/".join(build_tools_dir.split("/")[0:-1]) test_begin = __import__("test_begin") test_all_templates = __import__("test_all_templates") deploy = __import__("deploy") git_plugin = __import__("git_plugin") retval = test_begin.entrypoint("{}/platform/backend".format(base_dir)) if (retval == "SUCCESS"): retval = test_all_templates.entrypoint("{}/platform/backend".format(base_dir)) if (retval == "SUCCESS"): deploy.deploy() **deploy.py** def deploy(): print(". {}/platform/backend/nu.sh --name jenkinsdeploy --stage dev --keyname greig --debug".format("/".join(os.path.abspath(__file__).split("/")[0:-2]))) returnedcode = subprocess.call("sh {}/platform/backend/nu.sh --name jenkinsdeploy --stage dev --keyname colin_greig --debug".format("/".join(os.path.abspath(__file__).split("/")[0:-2])), shell=True) if returnedcode == 0: return "DEPLOY SUCCESS" return "DEPLOY FAILURE" Answer: You can use api calls to aws to retrieve the status and wait until it become 'some status'. The below example is psuedo-code to illustrate the idea: import time import boto3 ec2 = boto3.resource('ec2') while True: response = ec2.describe_instance_status(.....) if response == 'some status': break time.sleep(60) # continue execution
Obtain csv-like parse AND line length byte count? Question: I'm familiar with the `csv` Python module, and believe it's necessary in my case, as I have some fields that contain the delimiter (`|` rather than `,`, but that's irrelevant) within quotes. However, I am also looking for the byte-count length of each original row, _prior_ to splitting into columns. I can't count on the data to always quote a column, and I don't know if/when `csv` will strip off outer quotes, so I don't think (but might be wrong) that simply joining on my delimiter will reproduce the original line string (less CRLF characters). Meaning, I'm not positive the following works: with open(fname) as fh: reader = csv.reader(fh, delimiter="|") for row in reader: original = "|".join(row) ## maybe? I've tried looking at `csv` to see if there was anything in there that I could use/monkey-patch for this purpose, but since `_csv.reader` is a `.so`, I don't know how to mess around with that. In case I'm dealing with an XY problem, my ultimate goal is to read through a CSV file, extracting certain fields and their overall file offsets to create a sort of look-up index. That way, later, when I have a list of candidate values, I can check each one's file-offset and `seek()` there, instead of chugging through the whole file again. As an idea of scale, I might have 100k values to look up across a 10GB file, so re-reading the file 100k times doesn't feel efficient to me. I'm open to other suggestions than the CSV module, but will still need `csv`-like intelligent parsing behavior. EDIT: Not sure how to make it more clear than the title and body already explains - simply `seek()`-ing on a file handle isn't sufficient because I **also** need to parse the lines as a `csv` in order to pull out additional information. Answer: You can't subclass `_csv.reader`, but the _`csvfile`_ argument to the `csv.reader()` [constructor](https://docs.python.org/3/library/csv.html#csv.reader) only has to be a "file-like object". This means you could supply an instance of your own class that does some preprocessing—such as remembering the length of the last line read and file offset. Here's an implementation showing exactly that. Note that the line length does _not_ include the end-of-line character(s). It also shows how the offsets to each line/row could be stored and used after the file is read. import csv class CSVInputFile(object): """ File-like object. """ def __init__(self, file): self.file = file self.offset = None self.linelen = None def __iter__(self): return self def __next__(self): offset = self.file.tell() data = self.file.readline() if not data: raise StopIteration self.offset = offset self.linelen = len(data) return data next = __next__ offsets = [] # remember where each row starts fname = 'unparsed.csv' with open(fname) as fh: csvfile = CSVInputFile(fh) for row in csv.reader(csvfile, delimiter="|"): print('offset: {}, linelen: {}, row: {}'.format( csvfile.offset, csvfile.linelen, row)) # file offset and length of row offsets.append(csvfile.offset) # remember where each row started
How to use yaml.load_all with fileinput.input? Question: Without resorting to `''.join`, is there a Pythonic way to use PyYAML's `yaml.load_all` with `fileinput.input()` for easy streaming of multiple documents from multiple sources? I'm looking for something like the following (non-working example): # example.py import fileinput import yaml for doc in yaml.load_all(fileinput.input()): print(doc) Expected output: $ cat >pre.yaml <<<'--- prefix-doc' $ cat >post.yaml <<<'--- postfix-doc' $ python example.py pre.yaml - post.yaml <<<'--- hello' prefix-doc hello postfix-doc Of course, `yaml.load_all` expects either a string, bytes, or a file-like object and `fileinput.input()` is none of those things, so the above example does not work. Actual output: $ python example.py pre.yaml - post.yaml <<<'--- hello' ... AttributeError: FileInput instance has no attribute 'read' You can make the example work with `''.join`, but that's cheating. I'm looking for a way that does not read the entire stream into memory at once. We might rephrase the question as _Is there some way to emulate a string, bytes, or file-like object that proxies to an underlying iterator of strings?_ However, I doubt that `yaml.load_all` actually needs the entire file-like interface, so that phrasing would ask for more than is strictly necessary. Ideally I'm looking for the minimal adapter that would support something like this: for doc in yaml.load_all(minimal_adapter(fileinput.input())): print(doc) Answer: The problem with `fileinput.input` is that the resulting object doesn't have a `read` method, which is what `yaml.load_all` is looking for. If you're willing to give up `fileinput`, you can just write your own class that will do what you want: import sys import yaml class BunchOFiles (object): def __init__(self, *files): self.files = files self.fditer = self._fditer() self.fd = self.fditer.next() def _fditer(self): for fn in self.files: with sys.stdin if fn == '-' else open(fn, 'r') as fd: yield fd def read(self, size=-1): while True: data = self.fd.read(size) if data: break else: try: self.fd = self.fditer.next() except StopIteration: self.fd = None break return data bunch = BunchOFiles(*sys.argv[1:]) for doc in yaml.load_all(bunch): print doc The `BunchOFiles` class gets you an object with a `read` method that will happily iterate over a list of files until everything is exhausted. Given the above code and your sample input, we get exactly the output you're looking for.
matplotlib multiple values under cursor Question: This question is very similar to those answered here, [matplotlib values under cursor](http://stackoverflow.com/questions/14754931/matplotlib-values-under- cursor) [In a matplotlib figure window (with imshow), how can I remove, hide, or redefine the displayed position of the mouse?](http://stackoverflow.com/questions/14349289/in-a-matplotlib-figure- window-with-imshow-how-can-i-remove-hide-or-redefine) [Interactive pixel information of an image in Python?](http://stackoverflow.com/questions/27704490/interactive-pixel- information-of-an-image-in-python) except that instead of pixel data (x,y,z) I have various measurements associated with (x,y) coordinates that I'd like to portray on a line plot. Specifically, the (x,y) are spatial positions (lat, lon) and at each (lat,lon) point there is a collection of data (speed, RPM, temp, etc.). I just sketched up something quickly to illustrate, a scatter plot with connecting lines, and then when you hover over a data point it displays all of the "z" values associated with that data point. Is there an easy way to do something like this? [![enter image description here](http://i.stack.imgur.com/rJvn9.png)](http://i.stack.imgur.com/rJvn9.png) Answer: You could probably build on something like this example. It doesn't display the information inside the figure (for now only using a `print()` statement), but it demonstrates a simple method of capturing clicks on `scatter` points and showing information for those points: import numpy as np import matplotlib.pylab as pl pl.close('all') n = 10 lat = np.random.random(n) lon = np.random.random(n) speed = np.random.random(n) rpm = np.random.random(n) temp = np.random.random(n) def on_click(event): i = event.ind[0] print('lon={0:.2f}, lat={1:.2f}: V={2:.2f}, RPM={3:.2f}, T={4:.2f}'\ .format(lon[i], lat[i], speed[i], rpm[i], temp[i])) fig=pl.figure() pl.plot(lon, lat, '-') pl.scatter(lon, lat, picker=True) fig.canvas.mpl_connect('pick_event', on_click) Clicking around a bit gives me: lon=0.63, lat=0.58: V=0.51, RPM=0.00, T=0.43 lon=0.41, lat=0.07: V=0.95, RPM=0.59, T=0.98 lon=0.86, lat=0.13: V=0.33, RPM=0.27, T=0.85
python logging module AttributeError: 'str' object has no attribute 'write' Question: I am using tornado,and in its app,I import logging just want to log some info about server. I put this: logging.config.dictConfig(web_LOGGING) right before: tornado.options.parse_command_line() but when I run the server,when I click any link,I get error: Traceback (most recent call last): File "/usr/lib/python2.7/logging/__init__.py", line 874, in emit stream.write(fs % msg) AttributeError: 'str' object has no attribute 'write' Logged from file web.py, line 1946 it just repeats when i click into any link. what is the real problem? I've already changed any file,directory to others to steer clear of namespace conflict... Answer: Psychic debugging says `web_LOGGING` has a key named `stream` with a `str` value (probably a file path); [`stream` is only for already opened files, if you want to pass a file path, it's passed as `filename`](https://docs.python.org/3/library/logging.html#logging.basicConfig).
wxpython phoenix: How to get float values from wx grid cells and perform mathematical operations? Question: I am trying to build a invoice with wx grid, I would like to add the values in quantity column and the values in price column and display it in the row total. import wx import wx.grid as gridlib class MyForm(wx.Frame): def __init__(self): wx.Frame.__init__(self, None, wx.ID_ANY, "Invoice") datag = 0 # Add a panel so it looks the correct on all platforms panel = wx.Panel(self, wx.ID_ANY) grid = gridlib.Grid(panel) grid.CreateGrid(25, 3) note_sizer = wx.BoxSizer() note_sizer.Add(grid, 1, wx.EXPAND) panel.SetSizer(note_sizer) # We can set the sizes of individual rows and columns # in pixels grid.SetColSize(0, 520) # change a couple column labels grid.SetColLabelValue(0, "Product") grid.SetColLabelValue(1, "Quantity") grid.SetColLabelValue(2, "Price") grid.SetColFormatFloat(2) grid.SetCellValue(24, 0, 'Total') if __name__ == "__main__": app = wx.App() frame = MyForm().Show() app.MainLoop() can anyone help? Answer: Once you start wanting to do things like that then it is a good idea to use a custom grid table for containing your data values. In other words, instead of you putting your values into the grid and getting the values back out when needed, the grid simply asks your table for the values to display, on an as- needed basis. This makes it easy to deal with the data in whatever structure and data types makes sense for your application. There are some examples in the wxPython demo that show some ways to implement table classes derived from `GridTableBase` and how to use them.
Class and external method call Question: I am going through a data structures course and I am not understanding how a Class can call a method that's in another Class. The code below has 2 classes: `Printer` and `Task`. Notice that class `Printer` has a method called `startNext`, and this has a variable `self.timeRemaining` that gets assigned the result of `newTask.getPages() * 60/self.pagerate`. How can `newTaks` reference the `getPages()` method from the Task class? The code that passes this object to the Printer class never references the Task class. The code works, since this is what the course gives out but, I just cannot understand how that method is accessed. Code: from pythonds.basic.queue import Queue import random class Printer: def __init__(self, ppm): self.pagerate = ppm self.currentTask = None self.timeRemaining = 0 def tick(self): if self.currentTask != None: self.timeRemaining = self.timeRemaining - 1 if self.timeRemaining <= 0: self.currentTask = None def busy(self): if self.currentTask != None: return True else: return False def startNext(self, newTask): self.currentTask = newTask self.timeRemaining = newTask.getPages() * 60/self.pagerate class Task: def __init__(self, time): self.timeStamp = time self.pages = random.randrange(1, 21) def getStamp(self): return self.timeStamp def getPages(self): return self.pages def waitTime(self, currentTime): return currentTime - self.timeStamp def simulation(numSeconds, pagesPerMinute): labPrinter = Printer(pagesPerMinute) printQueue = Queue() waitingTimes = [] for currentSecond in range(numSeconds): if newPrintTask(): task = Task(currentSecond) printQueue.enqueue(task) if (not labPrinter.busy()) and (not printQueue.isEmpty()): nextTask = printQueue.dequeue() waitingTimes.append(nextTask.waitTime(currentSecond)) labPrinter.startNext(nextTask) labPrinter.tick() averageWait = sum(waitingTimes)/len(waitingTimes) print "Average Wait %6.2f secs %3d tasks remaining." % (averageWait, printQueue.size()) def newPrintTask(): num = random.randrange(1, 181) if num == 180: return True else: return False for i in range(10): simulation(3600, 5) Answer: If I understand clearly your question, it is because you are adding task object to Queue list. Then when you are getting object (list item) back, you are getting again Task object: #creating Task object and adding to Queque list task = Task(currentSecond) printQueue.enqueue(task) class Queue: def __init__(self): #list of task objects self.items = [] def enqueue(self, item): #you are inserting Task object item to list self.items.insert(0,item) def dequeue(self): #returns task object return self.items.pop() So then you can call startNext() method from Printer class, because the dequeue() method is returning Task object. And because of the object in startNext() is type of Task, you can call getPages() method on that object. Is it sufficient answer?
tweepy.error.TweepError: [{u'message': u'Text parameter is missing.', u'code': 38}] Question: I am using `tweepy` twitter api for python, while using it i got some error, I am not able to use `send_direct_message(user/screen_name/user_id, text)` this method Here is my code:- import tweepy consumer_key='XXXXXXXXXXXXXXXXX' consumer_secret='XXXXXXXXXXXXXXXXX' access_token='XXXXXXXXXXXXXXXXX' access_token_secret='XXXXXXXXXXXXXXXXX' auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) API = tweepy.API(auth) user = API.get_user('SSPendse') screen_name="CJ0495" id = 773436067231956992 text="Message Which we have to send must have maximum 140 characters, If it is Greater then the message will be truncated upto 140 characters...." # re = API.send_direct_message(screen_name, text) re = API.send_direct_message(id, text) print re got following error:- Traceback (most recent call last): File "tweetApi.py", line 36, in <module> re = API.send_direct_message(id, text) File "/usr/local/lib/python2.7/dist-packages/tweepy/binder.py", line 245, in _call return method.execute() File "/usr/local/lib/python2.7/dist-packages/tweepy/binder.py", line 229, in execute raise TweepError(error_msg, resp, api_code=api_error_code) tweepy.error.TweepError: [{u'message': u'Text parameter is missing.', u'code': 38}] What will be mistake done by me...??? I also have another problem related to tweepy, How can I moved to page two or got more followers in following code i=1 user = API.get_user('Apple') followers = API.followers(user.id,-1) for follower in followers: print follower,'\t',i i=i+1 if I run the code I got only 5000 followers however if I use `user.followers_count` it gives _362705_ followers (**This no. may be changes while You check it**) How can I see remaining followers Thank You... :) Answer: To solve your first error, replace `re = API.send_direct_message(id, text)` with `re = API.send_direct_message(id, text=text)`. This function only works if you give it the message as a named parameter. The parameter name you need here is "text", so you might want to change your variable name to avoid confusion. Also, I just tried it, since you are sending a direct message, not a tweet, it will **not** be truncated to only the first 140 characters. For your second question, this should do the trick, as explained [here](http://stackoverflow.com/questions/17431807/get-all-follower-ids-in- twitter-by-tweepy): followers = [] for page in tweepy.Cursor(API.followers, screen_name='Apple').pages(): followers.extend(page) time.sleep(60) #this is here to slow down your requests and prevent you from hitting the rate limit and encountering a new error print(followers)
Get "TypeError: cannot create mph' When Using sympy.nsolve() Question: I am attempting to determine the time, angle and speed that something would have to travel to intersect with a moving ellipse. (I actually want these conditions for the minimum time). Right now I was trying to use Sympy to help with this adventure. The following is the code which I am executing: import sympy as sp sp.init_printing() delta_x, delta_y, t = sp.symbols('delta_x delta_y t', real=True, positive=True) V, V_s, x_0, y_0, theta_s, theta_t = sp.symbols('V V_s x_0 y_0 theta_s theta_t', real=True) x, y = sp.symbols('x y', real=True) EQ1 = sp.Eq(((x-(x_0+V*sp.cos(theta_t)*t))/(delta_x+V*sp.cos(theta_t)*t))**2+((y-(y_0+V*sp.sin(theta_t)*t))/(delta_y+V*sp.sin(theta_t)*t))**2-1, 0) sx = sp.Eq(x, V_s*sp.cos(theta_s)*t) sy = sp.Eq(y, V_s*sp.sin(theta_s)*t) mysubs = [(V,5), (x_0, 10), (y_0, 10), (theta_t, 7*(sp.pi/4)), (delta_x, 0), (delta_y, 0)] sp.nsolve((EQ1.subs(mysubs), sx.subs(mysubs), sy.subs(mysubs)), (V_s, theta_s, t), (5, 0.0, 1)) The result of this operation yields: --------------------------------------------------------------------------- TypeError Traceback (most recent call last) /opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/calculus/optimization.py in findroot(ctx, f, x0, solver, tol, verbose, verify, **kwargs) 927 try: --> 928 fx = f(*x0) 929 multidimensional = isinstance(fx, (list, tuple, ctx.matrix)) <string> in <lambda>(_Dummy_75, _Dummy_76, _Dummy_77) /opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/matrices/matrices.py in __init__(self, *args, **kwargs) 300 for j, a in enumerate(row): --> 301 self[i, j] = convert(a) 302 else: /opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/ctx_mp_python.py in convert(ctx, x, strings) 661 return ctx.convert(x._mpmath_(prec, rounding)) --> 662 return ctx._convert_fallback(x, strings) 663 /opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/ctx_mp.py in _convert_fallback(ctx, x, strings) 613 raise ValueError("can only create mpf from zero-width interval") --> 614 raise TypeError("cannot create mpf from " + repr(x)) 615 TypeError: cannot create mpf from 0.08*(x - 13.535533905932737622)**2 + 0.08*(y - 6.464466094067262378)**2 - 1 Is this because the system is not constrained enough? There is a family of angles and velocities which could be used to intercept the moving ellipse. This error does not seem to imply this. (Yes, even when I attempt to constrain V_s in this problem the same error appears). I am using the following versions of things: | Software | Version | |----------|-----------| | python | 3.5.2 | | sympy | 1.0 | | mpmath | 0.19 | Answer: The variables `x` and `y` are still symbolic in the equations. `nsolve` requires all variables to be specified, and at least as many equations as variables. So you either need to include values for them in `mysubs`, or else solve for them (but to do that with `nsolve`, you will need two more equations).
Python PYQT Tabs outside app window [why] Question: Tabs added before app.exec_() is called look and act as any other tabs u met, though if adding another after the app.exec_() call makes the new tab 'detach' from the main app window. Pic below :) Why? How can I make it move inside the window? import threading import time import sys from PyQt5.QtWidgets import QApplication from PyQt5.QtWidgets import QFormLayout from PyQt5.QtWidgets import QLineEdit from PyQt5.QtWidgets import QTabWidget from PyQt5.QtWidgets import QTextEdit from PyQt5.QtWidgets import QWidget class ATry(threading.Thread): def __init__(self): super().__init__() def run(self): time.sleep(1) anotherTextEdit = QTextEdit() anotherLineEdit = QLineEdit() anotherLayout = QFormLayout() anotherLayout.addRow(anotherTextEdit) anotherLayout.addRow(anotherLineEdit) anotherTab = QWidget() anotherTab.setLayout(anotherLayout) md.addTab(anotherTab, "Outside") app = QApplication(sys.argv) md = QTabWidget() aTextEdit = QTextEdit() aLineEdit = QLineEdit() layout = QFormLayout() layout.addRow(aTextEdit) layout.addRow(aLineEdit) thisTab = QWidget() thisTab.setLayout(layout) md.addTab(thisTab, "Inside") a = ATry() a.start() md.show() app.exec_() [Screen describing the problem](http://i.stack.imgur.com/3gXA6.png) Answer: It works with QTimer or signals: import sys import time from PyQt5.QtCore import QObject from PyQt5.QtCore import QThread from PyQt5.QtCore import pyqtSignal from PyQt5.QtWidgets import QApplication from PyQt5.QtWidgets import QFormLayout from PyQt5.QtWidgets import QLineEdit from PyQt5.QtWidgets import QTabWidget from PyQt5.QtWidgets import QTextEdit from PyQt5.QtWidgets import QWidget class ATry(QThread): def __init__(self, pointer): super().__init__() self.pointer = pointer def run(self): time.sleep(2) self.pointer.emit() def addTheTab(): anotherTextEdit = QTextEdit() anotherLineEdit = QLineEdit() anotherLayout = QFormLayout() anotherLayout.addRow(anotherLineEdit) anotherLayout.addRow(anotherTextEdit) anotherTab = QWidget() anotherTab.setLayout(anotherLayout) md.addTab(anotherTab, "Whatever") class MyQObject(QObject): trigger = pyqtSignal() def __init__(self): super().__init__() def connect_and_get_trigger(self): self.trigger.connect(addTheTab) return self.trigger def getGFX(self): app = QApplication(sys.argv) md = QTabWidget() md.show() return app, md obj = MyQObject() app, md = obj.getGFX() addTheTab() a = ATry(obj.connect_and_get_trigger()) a.start() # timer = QTimer() # timer.timeout.connect(proba) # timer.start(3000) app.exec_()
Read a complete data file and round numbers to 2 decimal places and save it with the same format Question: I am trying to learn python and I have the intention to make the a very big data file smaller and later do some statistical Analysis with R. I need to read the data file (see below): SCALAR ND 3 ST 0 TS 10.00 0.0000 0.0000 0.0000 SCALAR ND 3 ST 0 TS 3600.47 255.1744 255.0201 255.2748 SCALAR ND 3 ST 0 TS 7200.42 255.5984 255.4946 255.7014 and find the numbers and round it in two digits after decimal, svae the maximum number with the namber in front of TS. At the end save the data file with the same format like following: SCALAR ND 3 ST 0 TS 10.00 0.00 0.00 0.00 SCALAR ND 3 ST 0 TS 3600.47 255.17 255.02 255.27 SCALAR ND 3 ST 0 TS**MAX** 7200.42 255.60 255.49 255.70 I have written a code like this: import numpy as np import matplotlib.pyplot as plt import pickle # Open file f = open('data.txt', 'r') thefile = open('output.txt', 'wb') # Read and ignore header lines header1 = f.readline() header2 = f.readline() header3 = f.readline() header4 = f.readline() data = [] for line in f: line = line.strip() columns = line.split() source = {} source['WSP'] = columns[0] #source['timestep'] = float(columns[1]) source['timestep'] = columns[1] data.append(source) f.close() but the number in front of TS cannot be read. I wanted to round the numbers but the float which I used does not work. After that I wanted to put it in a loop Any suggestion, Do I write the code in a good way? I will be very thankfull for the help. Answer: Try `"%.2f" % float(columns[1])` to round to two decimal places. Note that it gives you a string, not a float. I don't understand the rest of what you're asking. `>>> "%.2f" % 255.5984` `'255.60'`
Python - Removing vertical bar lines from histogram Question: I'm wanting to remove the vertical bar outlines from my histogram plot, but preserving the "etching" of the histogram, if that makes since. import matplotlib.pyplot as plt import numpy as np bins = 35 fig = plt.figure(figsize=(7,6)) ax = fig.add_subplot(111) ax.hist(subVel_Hydro1, bins=bins, facecolor='none', edgecolor='black', label = 'Pecuiliar Vel') ax.set_xlabel('$v_{_{B|A}} $ [$km\ s^{-1}$]', fontsize = 16) ax.set_ylabel(r'$P\ (r_{_{B|A}} )$', fontsize = 16) ax.legend(frameon=False) Giving [![enter image description here](http://i.stack.imgur.com/DtEHe.png)](http://i.stack.imgur.com/DtEHe.png) Is this doable in matplotlibs histogram functionality? I hope I provided enough clarity. Answer: In `pyplot.hist()` you could set the value of `histtype = 'step'`. Example code: import matplotlib as mpl import matplotlib.pyplot as plt import numpy as np x = np.random.normal(0,1,size=1000) fig = plt.figure() ax = fig.add_subplot(111) ax.hist(x, bins=50, histtype = 'step', fill = None) plt.show() Sample output: [![enter image description here](http://i.stack.imgur.com/621Vo.png)](http://i.stack.imgur.com/621Vo.png)
when i made a 3 handshake with ubuntu in VMware return package R Question: #!/usr/bin/python from scapy.all import * def findWeb(): a = sr1(IP(dst="8.8.8.8")/UDP()/DNS(qd=DNSQR(qname="www.google.com")),verbose=0) return a[DNSRR].rdata def sendPacket(dst,src): ip = IP(dst = dst) SYN = TCP(sport=1500, dport=80, flags='S') SYNACK = sr1(ip/SYN) my_ack = SYNACK.seq + 1 ACK = TCP(sport=1050, dport=80, flags='A', ack=my_ack) send(ip/ACK) payload = "stuff" PUSH = TCP(sport=1050, dport=80, flags='PA', seq=11, ack=my_ack) send(ip/PUSH/payload) http = sr1(ip/TCP()/'GET /index.html HTTP/1.0 \n\n',verbose=0) print http.show() src = '10.0.0.24' dst = findWeb() sendPacket(dst,src) I'm trying to do HTTP packets with SCAPY I am using UBUNTU on VMwaer The problem is that every time I send messages I have RESET How do we fix it? Thanks [sniff package image](http://i.stack.imgur.com/0eJdj.png) Answer: Few things I notice wrong. 1\. You have your sequence number set statically (seq=11) which is wrong. Sequence numbers are always randomly generated and they must be used as per RFC793. So the sequence should be = SYNACK[TCP].ack 2. You set your source port as 1500 during SYN packet, but then you use it as 1050 (typo?) 3. You don't need extra payload/PUSH. Also, have a look at these threads: [How to create HTTP GET request Scapy?](http://stackoverflow.com/questions/37683026/how-to-create-http-get- request-scapy) [Python-Scapy or the like-How can I create an HTTP GET request at the packet level](http://stackoverflow.com/questions/4750793/python-scapy-or-the-like- how-can-i-create-an-http-get-request-at-the-packet-leve)
Complete Algebraic Equations Question: Im trying to build an app that graphs an equation based on user input. The equation would be in slope intercept form: `y = mx + b`, for `m` as slope and `b` as `y intercept`. However, this isn't working for me in python! I tried this: >>> x = 3 >>> 1/2x and was returned this: File "<stdin>", line 1 1/2x ^ SyntaxError: invalid syntax >>> How would I make it so that this returns 1.5? Answer: import re mxb = re.compile('''(?P<slope>[(\d)/.]*)x\s+(?P<sign>[+-])\s+(?P<intercept>[\d./]*)''') s = '1/2x + 5' match = mxb.match(s) if match : print( match.groupdict() ) test the python regex here : [pythex.org](http://pythex.org/?regex=\(%3FP%3Cslope%3E%5B\(%5Cd\)%2F.%5D*\)x%5Cs\(%3FP%3Csign%3E%5B%2B-%5D\)%5Cs\(%3FP%3Cintercept%3E%5B%5Cd.%2F%5D*\)&test_string=x%20%2B%205%0A1%2F2x%20-%202.3%0A.5x%20%2B%201.4&ignorecase=0&multiline=0&dotall=0&verbose=0 "pythex.org")
Python bokeh apply hovertools only on model not on figure Question: I want to have a scatter plot and a (base)line on the same figure. And I want to use `HoverTool` only on the circles of scatter but not on the line. Is it possible? With the code below I get tooltips with `index: 0` and `(x, y): (???, ???)` when I hover on the line (any part of the line). But the `index: 0` data in `source` is totally different (`(x, y): (1, 2)`)... df = pd.DataFrame({'a':[1, 3, 6, 9], 'b':[2, 3, 5, 8]}) from bokeh.models import HoverTool import bokeh.plotting as bplt TOOLS = ['box_zoom', 'box_select', 'wheel_zoom', 'reset', 'pan', 'resize', 'save'] source = bplt.ColumnDataSource(data=df) hover = HoverTool(tooltips=[("index", "$index"), ("(x, y)", "(@a, @b)")]) p = bplt.figure(plot_width=600, plot_height=600, tools=TOOLS+[hover], title="My sample bokeh plot", webgl=True) p.circle('a', 'b', size=10, source=source) p.line([0, 10], [0, 10], color='red') bplt.save(p, 'c:/_teszt.html') Thank you!! Answer: To limit which renderers you want the HoverTool is active on (by default it's active on all) you can either set a `name` attr on your glyphs, then specify which names you want your HoverTool to be active on: p.circle('a', 'b', size=10, name='circle', source=source) hover = HoverTool(names=['circle']) docs: <http://bokeh.pydata.org/en/latest/docs/reference/models/tools.html#bokeh.models.tools.HoverTool.names> or you can add the renderers to the HoverTool. circle = p.circle('a', 'b', size=10, source=source) hover = HoverTool(renderers=['circle']) docs: <http://bokeh.pydata.org/en/latest/docs/reference/models/tools.html#bokeh.models.tools.HoverTool.renderers>
How to see the plot made in python using pandas and matplotlib Question: I am following the tutorial <http://ahmedbesbes.com/how-to-score-08134-in- titanic-kaggle-challenge.html> and following is my code from IPython.core.display import HTML HTML(""" <style> .output_png { display: table-cell; text-align: center; vertical-align: middle; } </style> """) # remove warnings import warnings warnings.filterwarnings('ignore') # --- import pandas as pd pd.options.display.max_columns = 100 import matplotlib matplotlib.use('TkAgg') import matplotlib.pyplot as plt matplotlib.style.use('ggplot') import numpy as np pd.options.display.max_rows = 100 data = pd.read_csv('./data/train.csv') data.head() data['Age'].fillna(data['Age'].median(), inplace=True) survived_sex = data[data['Survived']==1]['Sex'].value_counts() dead_sex = data[data['Survived']==0]['Sex'].value_counts() df = pd.DataFrame([survived_sex,dead_sex]) df.index = ['Survived','Dead'] df.plot(kind='bar',stacked=True, figsize=(15,8)) I dont actually see the plot. How do i see the plot? [![enter image description here](http://i.stack.imgur.com/g24E6.png)](http://i.stack.imgur.com/g24E6.png) Answer: Try using `plt.show()` at the end. EDIT: Also, you may need to add `%matplotlib inline` as explained here: [How to make IPython notebook matplotlib plot inline](http://stackoverflow.com/questions/19410042/how-to-make-ipython- notebook-matplotlib-plot-inline)
AttributeError: type object has no attribute "id" PYTHON Question: So I was trying to make a basic python pong game when this error came up: It seems to say that AttributeError: type object has no attribute "id" which I have no idea what it means. C:\Users\****\AppData\Local\Programs\Python\Python35-32\python.exe C:/Users/****/untitled/src/testing.py Traceback (most recent call last): File "C:/Users/****/untitled/src/testing.py", line 113, in <module> ball.draw() File "C:/Users/****/untitled/src/testing.py", line 36, in draw if self.hit_paddle2(pos) == True: File "C:/Users/****/untitled/src/testing.py", line 53, in hit_paddle2 paddle2_pos = self.canvas.coords(self.paddle2.id) AttributeError: type object 'Paddle2' has no attribute 'id' Process finished with exit code 1 This is the code I used: from tkinter import * import random import time tk = Tk() tk.title("Game") tk.resizable(0, 0) tk.wm_attributes("-topmost", 1) canvas = Canvas(tk, width=500, height=400, bd=0, highlightthickness=0) canvas.pack() tk.update() class Ball: def __init__(self, canvas, paddle, paddle2, color): self.canvas = canvas self.paddle = paddle self.paddle2 = paddle2 self.id = canvas.create_oval(10, 10, 25, 25, fill=color) self.canvas.move(self.id, 245, 100) starts = [-3, -2, -1, 1, 2, 3] random.shuffle(starts) self.x = starts[0] self.y = -3 self.canvas_height = self.canvas.winfo_height() self.canvas_width = self.canvas.winfo_width() self.hit_bottom = False def draw(self): self.canvas.move(self.id, self.x, self.y) pos = self.canvas.coords(self.id) if pos[1] <= 0: self.y = 3 if self.hit_paddle(pos) == True: self.y = -3 if self.hit_paddle2(pos) == True: self.y = 3 if pos[3] >= self.canvas_height: self.hit_bottom = True if pos[0] <= 0: self.x = 3 if pos[2] >= self.canvas_width: self.x = -3 def hit_paddle(self, pos): paddle_pos = self.canvas.coords(self.paddle.id) if pos[2] >= paddle_pos[0] and pos[0] <= paddle_pos[2]: if pos[3] >= paddle_pos[1] and pos[3] <= paddle_pos[3]: return True return False def hit_paddle2(self, pos): paddle2_pos = self.canvas.coords(self.paddle2.id) if pos[2] >= paddle2_pos[0] and pos[0] <= paddle2_pos[2]: if pos[3] >= paddle2_pos[1] and pos[3] <= paddle2_pos[3]: return True return False class Paddle: def __init__(self, canvas, color): self.canvas = canvas self.id = canvas.create_rectangle(0, 0, 100, 10, fill=color) self.canvas.move(self.id, 200, 300) self.x = 0 self.canvas_width = self.canvas.winfo_width() self.canvas.bind_all('<KeyPress-Left>', self.turn_left) self.canvas.bind_all('<KeyPress-Right>', self.turn_right) def turn_left(self, evt): self.x = -2 def turn_right(self, evt): self.x = 2 def draw(self): self.canvas.move(self.id, self.x, 0) pos = self.canvas.coords(self.id) if pos[0] <= 0: self.x = 0 elif pos[2] >= self.canvas_width: self.x = 0 class Paddle2: def __init__(self, canvas, color): self.canvas = canvas self.id = canvas.create_rectangle(0, 10, 100, 10, fill=color) self.canvas.move(self.id, 200, 300) self.x = 0 self.canvas_width = self.canvas.winfo_width() self.canvas.bind_all('<KeyPress-W>', self.turn_left) self.canvas.bind_all('<KeyPress-A>', self.turn_right) def turn_left(self, evt): self.x = -2 def turn_right(self, evt): self.x = 2 def draw(self): self.canvas.move(self.id, self.x, 0) pos = self.canvas.coords(self.id) if pos[0] <= 0: self.x = 0 elif pos[2] >= self.canvas_width: self.x = 0 paddle = Paddle(canvas, 'blue') ball = Ball(canvas, paddle, Paddle2, 'red') paddle2 = Paddle2(canvas, 'blue') while 1: if ball.hit_bottom == False: ball.draw() paddle.draw() paddle2.draw() tk.update_idletasks() tk.update() time.sleep(0.01) Answer: You need to pass instances of `Paddle` and `Paddle2` to `Ball` constructor. paddle = Paddle(canvas, 'blue') paddle2 = Paddle2(canvas, 'blue') ball = Ball(canvas, paddle, paddle2, 'red')
Edited - Python plot persistence windows Question: Each time I launch my program, my plots are erased after each execution. I would like the following situation: 1. Launch program 1 and plot in figure 1 2. Stop the execution of program 1 3. Lauch program 2 and plot in figure 1 4. Retrieve a pdf file where the plot of program 1 and program 2 are superimposed here is my code: z = np.zeros(5) fig,axarr = plt.subplots(nrows=2, ncols=3) fig.subplots_adjust() lig,col = 0,0 axarr[lig,col].plot(tt,z[0,:],'k') #plt.ion() #plt.draw() plt.show(block=True) #plt.show() **EDIT:** Ok maybe I could solve this problem with the following approach: Is it possible to get the ax of a figure that was opened by an other excetion of of program ? Answer: `plt.show(block=True)` should work. Other hack would be using pause function as shown below: from matplotlib import pylab pylab.plot(range(10), range(10)) pylab.pause(2**31-1)
Iterating across multiple columns in Pandas DF and slicing dynamically Question: **TLDR:** How to iterate across all options of multiple columns in a pandas dataframe without specifying the columns or their values explicitly? **Long Version:** I have a pandas dataframe that looks like this, only it has a lot more features or drug dose combinations than are listed here. Instead of just 3 types of features, it could have something like 70...: > dosage_df First Score Last Score A_dose B_dose C_dose 22 28 1 40 130 55 11 2 40 130 15 72 3 40 130 42 67 1 90 130 90 74 2 90 130 87 89 3 90 130 14 43 1 40 700 12 61 2 40 700 41 5 3 40 700 Along with my data frame, I also have a python dictionary with the relevant ranges for each feature. The keys are the feature names, and the different values which it can take are the keys: > dict_of_dose_ranges = {'A_dose': [1, 2, 3], 'B_dose': [40, 90], 'C_dose': [130,700]} For my purposes, I need to generate a particular combination (say A_dose = 1, B_dose = 90, and C_dose = 700), and based on those settings take the relevant slice out of my dataframe, and do relevant calculations from that smaller subset, and save the results somewhere. I need to do this for ALL possible combinations of ALL of my features (far more than the 3 which are here, and which will be variable in the future). In this case, I could easily pop this into SkLearn's Parameter grid, generate the options: > from sklearn.grid_search import ParameterGrid > all_options = list(ParameterGrid(dict_of_dose_ranges)) > all_options and get: [{'A_dose': 1, 'B_dose': 40, 'C_dose': 130}, {'A_dose': 1, 'B_dose': 40, 'C_dose': 700}, {'A_dose': 1, 'B_dose': 90, 'C_dose': 130}, {'A_dose': 1, 'B_dose': 90, 'C_dose': 700}, {'A_dose': 2, 'B_dose': 40, 'C_dose': 130}, {'A_dose': 2, 'B_dose': 40, 'C_dose': 700}, {'A_dose': 2, 'B_dose': 90, 'C_dose': 130}, {'A_dose': 2, 'B_dose': 90, 'C_dose': 700}, {'A_dose': 3, 'B_dose': 40, 'C_dose': 130}, {'A_dose': 3, 'B_dose': 40, 'C_dose': 700}, {'A_dose': 3, 'B_dose': 90, 'C_dose': 130}, {'A_dose': 3, 'B_dose': 90, 'C_dose': 700}] **This is where I run into problems:** **Problem #1)** I can now iterate across `all_options`, but I'm not sure how to now SELECT out of my `dosage_df` from each of the dictionary options (i.e. {'A_dose': 1, 'B_dose': 40, 'C_dose': 130}) WITHOUT doing it explicitly. In the past, I could do something like: dosage_df[(dosage_df.A_dose == 1) & (dosage_df.B_dose == 40) & (dosage_df.C_dose == 130)] First Score Last Score A_dose B_dose C_dose 0 22 28 140 130 But now I'm not sure what to put inside the brackets to slice it dynamically... dosage_df[?????] **Problem #2)** When I actually enter in my full dictionary of features with their respective ranges, I get an error because it deems it as having too many options... from sklearn.grid_search import ParameterGrid all_options = list(ParameterGrid(dictionary_of_features_and_ranges)) all_options --------------------------------------------------------------------------- OverflowError Traceback (most recent call last) <ipython-input-138-7b73d5e248f5> in <module>() 1 from sklearn.grid_search import ParameterGrid ----> 2 all_options = list(ParameterGrid(dictionary_of_features_and_ranges)) 3 all_options OverflowError: long int too large to convert to int I tried a number of alternate approaches including using double while loops, a [tree / recursion method from here](http://stackoverflow.com/questions/23986892/python-recursive-iteration- exceeding-limit-for-tree-implementation), another [recursion method from here](http://stackoverflow.com/questions/13109274/python-recursion- permutations), but it wasn't coming together.... Any help is much appreciated. Answer: You can use [`itertools.product`](https://docs.python.org/3/library/itertools.html#itertools.product) to generate all possible dosage combinations, and [`DataFrame.query`](http://pandas.pydata.org/pandas- docs/stable/generated/pandas.DataFrame.query.html) to do the selection: from itertools import product for dosage_comb in product(*dict_of_dose_ranges.values()): dosage_items = zip(dict_of_dose_ranges.keys(), dosage_comb) query_str = ' & '.join('{} == {}'.format(*x) for x in dosage_items) sub_df = dosage_df.query(query_str) # Do Stuff...
Adding exception to "AttributeError" python Question: So, I have some tweets with some special characters and shapes. I am trying to find a word in those tweets by converting them to lower case. The function throws an "AttributeError" when it encounters those special characters and hence, I want to change my function in a way that it skips those records and processes others. Can I add exception to "AttributeError" in python. I want it to act more like an "iferror resume next"/Error handling statement. I am currently using:- def word_in_text(word, text): try: print text word = word.lower() text = text.lower() match = re.search(word, text) if match: return True else: return False except(AttributeError, Exception) as e: continue error post using @galah92 recommendations :- Traceback (most recent call last): File "<input>", line 1, in <module> File "C:\Python27\lib\site-packages\pandas\core\series.py", line 2220, in apply mapped = lib.map_infer(values, f, convert=convert_dtype) File "pandas\src\inference.pyx", line 1088, in pandas.lib.map_infer (pandas\lib.c:63043) File "<input>", line 1, in <lambda> File "<input>", line 3, in word_in_text File "C:\Python27\lib\re.py", line 146, in search return _compile(pattern, flags).search(string) TypeError: expected string or buffer I am new to Python and self learning it. Any help will be really appreciated. Answer: You can use `re.IGNORECASE` flag when you `search()`. That way you don't need to deal with `lower()` or exceptions. def word_in_text(word, text): print text if re.search(word, text, re.IGNORECASE): return True else: return False * * * As an example, if I run: from __future__ import unicode_literals # see edit notes import re text = "CANCION! You &amp" word = "you" def word_in_text(word, text): print(text) if re.search(word, text, re.IGNORECASE): return True else: return False print(word_in_text(word, text)) The output is: CANCION! You &amp True * * * **EDIT** For Python 2, you should add `from __future__ import unicode_literals` at the top of your script to make sure you encode everything to UTF-8. You can read more about it [here](http://stackoverflow.com/questions/23370025/what-is-unicode-literals- used-for).
<urlopen error (-1, 'SSL exception: Differences between the SSL socket behaviour of cpython vs. jython are explained on the wiki Question: I'm using the following code. import urllib2 #Setting proxy myProxy = {'https':'https://proxy.example.com:8080'} proxy = urllib2.ProxyHandler(myProxy) opener = urllib2.build_opener(proxy) urllib2.install_opener(opener) #Access URL auth_token = "realLy_loNg_AuTheNicaTion_toKen" headers={"Content-Type":"application/json", "charset": "UTF-8", "Authorization": "Bearer %s" %auth_token} url = "https://example.com/something" req = urllib2.Request(url, None, headers) reply = urllib2.urlopen(req) print reply.getcode() I'm running the above as a Jython script in nGrinder. When I run the same script on my system with Jython, it works fine and returns 200(OK status code). When I run it on nGrinder, I get the error (-1, 'SSL exception: Differences between the SSL socket behaviour of cpython vs. jython are explained on the wiki: http://wiki.python.org/jython/NewSocketModule#SSL_Support') Any ideas why this is happening? EDIT: I've been trying and the issue is definitely with the long authentication token. I have a feeling it might be some encoding issue. There was a similar question posted [here](http://stackoverflow.com/questions/33912074/https-get-using-jython) before. I read it but it wasn't described properly. But it could be a good reference to think off of. Answer: It seems issue of certificate trust. <http://bugs.jython.org/issue1688> There are these two options that will solve your problem. > Jython is not like cpython. Cpython does not verify the chain of trust for > server certificates. Jython does verify the chain of trust, and will refuse > to open the connection if it cannot verify the server. > > So you have two options. > > 1. Disable certificate checking on jython > > > <http://jython.xhaus.com/installing-an-all-trusting-security-provider-on- > java-and-jython/> <http://tech.pedersen-live.com/2010/10/trusting-all- > certificates-in-jython/> > > 2. Add your (self-signed?) certificate to your local java trust store, so > that your client will trust your server. > > > Google("java install self-signed certificate") Another reason for this can be due to incompactible version of python and java (<https://github.com/geoscript/geoscript-py/issues/32>) > I just tried with Jython 2.5.3 and JDK 1.7 on Mac and was able to run > ez_setup.py. > > Java version: [Java HotSpot(TM) 64-Bit Server VM (Oracle Corporation)] on > java1.7.0_51 > > Swapping to Oracle JDK 7 did the trick. Please check if this solves the problem, as per implementation of jython and error, issue is due to HTTPS(SSL layer) <https://github.com/int3/jython/blob/master/Lib/socket.py> # These error codes are currently wrong: getting them correct is going to require # some investigation. Cpython 2.6 introduced extensive SSL support. (javax.net.ssl.SSLException, ALL) : lambda x: sslerror(-1, 'SSL exception'+_ssl_message),
Extracting hand writing text out in shape with OpenCV Question: I am very new to OpenCV Python and I really need some help here. So what I am trying to do here is to extract out these words in the image below. [![hand drawn image](http://i.stack.imgur.com/XsSOP.jpg)](http://i.stack.imgur.com/XsSOP.jpg) The words and shapes are all hand drawn, so they are not perfect. I have did some coding below. First of all, I grayscale the image img_final = cv2.imread(file_name) img2gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY) Then I use THRESH_INV to show the content ret, new_img = cv2.threshold(image_final, 100 , 255, cv2.THRESH_BINARY_INV) After which, I dilate the content kernel = cv2.getStructuringElement(cv2.MORPH_CROSS,(3 , 3)) dilated = cv2.dilate(new_img,kernel,iterations = 3) I dilate the image is because I can identify text as one cluster After that, I apply boundingRect around the contour and draw around the rectangle contours, hierarchy = cv2.findContours(dilated,cv2.RETR_EXTERNAL,cv2.CHAIN_APPROX_NONE) # get contours index = 0 for contour in contours: # get rectangle bounding contour [x,y,w,h] = cv2.boundingRect(contour) #Don't plot small false positives that aren't text if w < 10 or h < 10: continue # draw rectangle around contour on original image cv2.rectangle(img,(x,y),(x+w,y+h),(255,0,255),2) This is what I got after that. [![result image](http://i.stack.imgur.com/vxKYi.png)](http://i.stack.imgur.com/vxKYi.png) I am only able to detect one of the text. I have tried many other methods but this is the closet results I have got and it does not fulfill the requirement. The reason for me to identify the text is so that I can get the X and Y coordinate of each of the text in this image by putting a bounding Rectangle "boundingRect()". Please help me out. Thank you so much Answer: You can use the fact that the connected component of the letters are much smaller than the large strokes of the rest of the diagram. I used opencv3 connected components in the code but you can do the same things using findContours. The code: import cv2 import numpy as np # Params maxArea = 150 minArea = 10 # Read image I = cv2.imread('i.jpg') # Convert to gray Igray = cv2.cvtColor(I,cv2.COLOR_RGB2GRAY) # Threshold ret, Ithresh = cv2.threshold(Igray,0,255,cv2.THRESH_BINARY_INV+cv2.THRESH_OTSU) # Keep only small components but not to small comp = cv2.connectedComponentsWithStats(Ithresh) labels = comp[1] labelStats = comp[2] labelAreas = labelStats[:,4] for compLabel in range(1,comp[0],1): if labelAreas[compLabel] > maxArea or labelAreas[compLabel] < minArea: labels[labels==compLabel] = 0 labels[labels>0] = 1 # Do dilation se = cv2.getStructuringElement(cv2.MORPH_ELLIPSE,(25,25)) IdilateText = cv2.morphologyEx(labels.astype(np.uint8),cv2.MORPH_DILATE,se) # Find connected component again comp = cv2.connectedComponentsWithStats(IdilateText) # Draw a rectangle around the text labels = comp[1] labelStats = comp[2] #labelAreas = labelStats[:,4] for compLabel in range(1,comp[0],1): cv2.rectangle(I,(labelStats[compLabel,0],labelStats[compLabel,1]),(labelStats[compLabel,0]+labelStats[compLabel,2],labelStats[compLabel,1]+labelStats[compLabel,3]),(0,0,255),2) [![enter image description here](http://i.stack.imgur.com/k0UdG.jpg)](http://i.stack.imgur.com/k0UdG.jpg)
Reshaping OpenCV Image (numpy) Dimensions Question: I need to convert an image in a numpy array loaded via cv2 into the correct format for the deep learning library mxnet for its convolutional layers. My current images are shaped as follows: (256, 256, 3), or (height, width, channels). From what I've been told, this actually needs to be (3, 256, 256), or (channels, height, width). Unfortunately, my knowledge of numpy/python opencv isn't good enough to know how to manipulate the arrays correctly. I've figured out that I could split the arrays into channels by cv2.split, but I'm uncertain of how to combine them again in the right format (I don't know if using cv2.split is optimal, or if there are better ways in numpy). Thanks for any help. Answer: You can use [`numpy.rollaxis`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.rollaxis.html) as follow: If your `image` as shape `(height, width, channels)` import numpy as np new_shaped_image = np.rollaxis(image, axis=2, start=0) This means that the `2`nd axis of the `new_shaped_image` will be at `0` spot. So `new_shaped_image.shape` will be `(channels, height, width)`
How to fetch JSON data from API, format / encode / write to a file? Question: I need to fetch some data from a weather API, extract certain info and send it to std. output (in my case this is the console/terminal; I am playing around with python API scripting and do not yet have a web site/app do show fetched data). **Example Python code from the API provider (simple to understand and works):** import urllib2 import json API_KEY='mysuperawesomekey' f = urllib2.urlopen('http://api.wunderground.com/api/' + API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json') json_string = f.read() parsed_json = json.loads(json_string) location = parsed_json['location']['city'] temp_f = parsed_json['current_observation']['temp_f'] print "Current temperature in %s is: %s" % (location, temp_f) f.close() Since I am on a free plan, I don't want to use up my API requests but rather fetch the JSON data, save it to a file and use in another script to practice. I found a couple of solutions here on [StackOverflow](http://stackoverflow.com/questions/12309269/how-do-i-write- json-data-to-a-file-in-python) but neither seem to work completely (nice formatting of the file): **Attempt 1. to save the fetched data, add. to the orig. code above:** import io import codecs with open('data.json', 'w') as outfile: json.dump(json_string, outfile, indent=4, sort_keys=True, separators=(",", ':')) **Attempt 2:** with codecs.open('data.json', 'w', 'utf8') as jasonfile: jasonfile.write(json.dumps(parsed_json, sort_keys = True, ensure_ascii=False)) Both of my attempts work ("kind of") as I do get a **.json** file. But, upon inspecting it in my editor (Atom) I am seeing this (first few soft-break lines): **Output:** "\n{\n \"response\": {\n \"version\":\"0.1\",\n\"termsofService\":\"http://www.wunderground.com/weather/api/d/terms.html\",\n \"features\": {\n \"geolookup\": 1\n ,\n \"conditions\": 1\n }\n\t}\n\t\t,\t\"location\": {\n\t\t\"type\":\"CITY\",\n\t\t\"country\":\"US\",\n\t\t\"country_iso3166\":\"US\",\n\t\t\"country_name\":\"USA\",\n\t\t\"state\":\"IA\",\n\t\t\"city\":\"Cedar Rapids\",\n\t\t\"tz_short\":\"CDT\",\n\t\t\"tz_long\":\"America/Chicago\",\n\t\t\"lat\":\"41.97171021\",\n\t\t\"lon\":\"-91.65871429\",\n\t\t\"zip\":\... It is all on one line, with newlines and tabs showing. I have a couple of questions: 1. How can I check what format / kind of data-type API is returning? (FOr example is it just a raw string which needs to be JSON-iffied?) 2. Is my JSON not written in a human readable format because of encoding, delimiters fault? 3. Line `json_string = f.read()` from example code seems to "mimic" working with a "real" internal file object, is this correct? Many thanks! Answer: Seems that it was quite a simple solution. In my original code, I was saving a "non-parsed" variable to a file: import urllib2 import json API_KEY='key' f = urllib2.urlopen('http://api.wunderground.com/api/' + API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json') # Saving the below variable into a .json file gives all the JSON data # on a single line json_string = f.read() with open('data.json', 'w') as outfile: json.dump(json_string, outfile, indent=4, sort_keys=True, separators=(",", ':')) which, of course, resulted in a file with JSON data on a single line. What I should have been doing is saving the **parsed** variable: import urllib2 import json API_KEY='key' f = urllib2.urlopen('http://api.wunderground.com/api/' + API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json') json_string = f.read() # THIS one is correct! parsed_json = json.loads(json_string) with open('data.json', 'w') as outfile: json.dump(parsed_json, outfile, indent=4, sort_keys=True, separators=(",", ':'))
Python won't print expression Question: So, I'm kind of new to programming and have been trying Python. I'm doing a really simple program that converts usd to euroes. This is the text of the problem that I'm trying to solve > You are going to travel to France. You will need to convert dollars to euros > (the currency of the European Union). There are two currency exchange > booths. Each has a display that shows CR: their conversion rate as euros per > dollar and their fee as a percentage. The fee is taken before your money is > converted. Which booth will give you the most euros for your dollars, how > many euros, and how much is the difference. Example 1: Dollars: 200 CR1: > 0.78 Fee: 1 (amount 152.88 euros) CR2: 0.80 Fee: 3 (amount 155.2 euros) > Answer: 2 is the best; difference is 2.32 euros; 155.2 euros And here is my code from __future__ import division usd = int(input("How much in USD? ")) cr1 = int(input("What is the convertion rate of the first one? ")) fee1 = int(input("What is the fee of the first one? ")) cr2 = int(input("What is the convertion rate of the second one? ")) fee2 = int(input("What is the fee of the second one? ")) def convertion (usd, cr, fee): usdwfee = usd - fee convert = usdwfee * cr return convert first = convertion(usd, cr1, fee1) second = convertion(usd, cr2, fee2) fs = first - second sf = second - first def ifstatements (first,second,fs,sf): if first < second: print "1 is the best; difference is ",fs," euroes. 2 converts to ",first," euroes." elif first > second: print "2 is the best; difference is",sf," euroes. 2 converts to", second," euroes." ifstatements(first, second, fs, sf) The problem is that when I run the program it won't print out. It just takes my input and doesn't output anything. Answer: Check your logic more. `cr1 = int(input("What is the convertion rate of the first one? "))` Your conversion rate is in int. As in Integer which means it can't have a floating point (a decimal "CR1: 0.78" from your example). Your cr1 will become 0 if you cast it into an int. Also change your dollar and fees to accept floats too since I'm assuming you want to deal with cents too So change: usd = float(input("How much in USD? ")) cr1 = float(input("What is the convertion rate of the first one? ")) fee1 = float(input("What is the fee of the first one? ")) cr2 = float(input("What is the convertion rate of the second one? ")) fee2 = float(input("What is the fee of the second one? ")) And it should work.
Can't login to a specific ASP.NET website using python requests Question: So I've been trying for the last 6 hours to make this work, but I couldn't and endless searches didn't help, So I guess I'm either doing something very fundamental wrong, or it's just a trivial bug which happens to match my logic so I need extra eyes to help me fix it. The website url is [this](https://www.wes.org/appstatus/). I wrote a piece of messy python code to just login and read the next page, but All I get is a nasty 500 error saying something on the server went wrong processing my request. Here is the request made by a browser which works just fine, no problem. HTTP Response code to this request is 302 (Redirect) POST /appstatus/index.aspx HTTP/1.1 Host: www.wes.org Connection: close Content-Length: 303 Cache-Control: max-age=0 Origin: https://www.wes.org Upgrade-Insecure-Requests: 1 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36 Content-Type: application/x-www-form-urlencoded Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8 Referer: https://www.wes.org/appstatus/index.aspx Accept-Encoding: gzip, deflate, br Accept-Language: en-US,en;q=0.8,fa;q=0.6 Cookie: ASP.NET_SessionId=bu2gemmlh3hvp4f5lqqngrbp; _ga=GA1.2.1842963052.1473348318; _gat=1 __VIEWSTATE=%2FwEPDwUKLTg3MTMwMDc1NA9kFgICAQ9kFgICAQ8PFgIeBFRleHRkZGRk9rP20Uj9SdsjOKNUBlbw55Q01zI%3D&__VIEWSTATEGENERATOR=189D346C&__EVENTVALIDATION=%2FwEWBQK6lf6LBAKf%2B9bUAgK9%2B7qcDgK8w4S2BALowqJjoU1f0Cg%2FEAGU6r2IjpIPG8BO%2BiE%3D&txtUID=Email%40Removed.com&txtPWD=PASSWORDREMOVED&Submit=Log+In&Hidden1= and this one is the request made by my script. POST /appstatus/index.aspx HTTP/1.1 Host: www.wes.org Connection: close Accept-Encoding: gzip, deflate, br User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36 Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8 Upgrade-Insecure-Requests: 1 Content-Type: application/x-www-form-urlencoded Origin: https://www.wes.org Accept-Language: en-US,en;q=0.8,fa;q=0.6 Cache-Control: max-age=0 Referer: https://www.wes.org/appstatus/indexca.aspx Cookie: ASP.NET_SessionId=nxotmb55jjwf5x4511rwiy45 Content-Length: 303 txtPWD=PASSWORDREMOVED&Submit=Log+In&__EVENTVALIDATION=%2FwEWBQK6lf6LBAKf%2B9bUAgK9%2B7qcDgK8w4S2BALowqJjoU1f0Cg%2FEAGU6r2IjpIPG8BO%2BiE%3D&txtUID=Email%40Removed.com&__VIEWSTATE=%2FwEPDwUKLTg3MTMwMDc1NA9kFgICAQ9kFgICAQ8PFgIeBFRleHRkZGRk9rP20Uj9SdsjOKNUBlbw55Q01zI%3D&Hidden1=&__VIEWSTATEGENERATOR=189D346C And this is the script making the request, I'm sorry if it's so messy, just need something quick. import requests import bs4 import urllib.parse def main(): session = requests.Session() headers = {"Origin": "https://www.wes.org", "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8", "Cache-Control": "max-age=0", "Upgrade-Insecure-Requests": "1", "Connection": "close", "User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36", "Referer": "https://www.wes.org/appstatus/indexca.aspx", "Accept-Encoding": "gzip, deflate, br", "Accept-Language": "en-US,en;q=0.8,fa;q=0.6", "Content-Type": "application/x-www-form-urlencoded"} r = session.get('https://www.wes.org/appstatus/index.aspx',headers=headers) cookies = r.cookies soup = bs4.BeautifulSoup(r.content, "html5lib") viewState=urllib.parse.quote(str(soup.select('#__VIEWSTATE')[0]).split('value="')[1].split('"/>')[0]) viewStateGenerator=urllib.parse.quote(str(soup.select('#__VIEWSTATEGENERATOR')[0]).split('value="')[1].split('"/>')[0]) eventValidation=urllib.parse.quote(str(soup.select('#__EVENTVALIDATION')[0]).split('value="')[1].split('"/>')[0]) paramsPost = {} paramsPost.update({'__VIEWSTATE':viewState}) paramsPost.update({'__VIEWSTATEGENERATOR':viewStateGenerator}) paramsPost.update({'__EVENTVALIDATION':eventValidation}) paramsPost.update({"txtUID": "[email protected]"}) paramsPost.update({"txtPWD": "My_So_Called_Password"}) paramsPost.update({"Submit": "Log In"}) paramsPost.update({"Hidden1": ""}) response = session.post("https://www.wes.org/appstatus/index.aspx", data=paramsPost, headers=headers, cookies=cookies) print("Status code:", response.status_code) #Outputs 500. #print("Response body:", response.content) if __name__ == '__main__': main() Any help would be so much appreciated. Answer: You are doing way too much work and in doing so not passing valid data,you extract value attribute directly i.e `.select_one('#__VIEWSTATEGENERATOR')["value"]` and the same for all the rest, the cookies will be set in the Session object after your initial get so the logic boils down to: with requests.Session() as session: headers = {"User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36"} r = session.get('https://www.wes.org/appstatus/index.aspx', headers=headers) soup = bs4.BeautifulSoup(r.content, "html5lib") viewState = soup.select_one('#__VIEWSTATE')["value"] viewStateGenerator = soup.select_one('#__VIEWSTATEGENERATOR')["value"] eventValidation = soup.select_one('#__EVENTVALIDATION')["value"] paramsPost = {'__VIEWSTATE': viewState,'__VIEWSTATEGENERATOR': viewStateGenerator, '__EVENTVALIDATION': eventValidation,"txtUID": "[email protected]", "txtPWD": "My_So_Called_Password", "Submit": "Log In","Hidden1": ""} response = session.post("https://www.wes.org/appstatus/index.aspx", data=paramsPost, headers=headers) print("Status code:", response.status_code) Python by convention uses CamelCase for class names and lowercase with underscores to separate multiple words, you might want to consider applying that to your code.
Cleaner method for finding the shortest distance between points in a python list? Question: I have a list of tuples and an individual point in python e.g. [(1,2) , (2,5), (6,7), (9,3)] and (2,1) , and I want to figure out the fastest path possible created by all combinations of the individual point to the list of points.(Basically I want to find the most efficient way to get to all of the points starting from (2,1)). I have a manhattanDistance function that can take it 2 points and output the distance. However, my algorithm is giving me inconsistent answers (The heuristic is off for some reason) What would be the correct way to accomplish this? Here is my previous algorithm: def bestPath(currentPoint,goalList): sum = 0 bestList = [] while len(goallist) > 0: for point in list: bestList.append((manhattanD(point,currentPoint),point)) bestTup = min(bestList) bestList = [] dist = bestTup[0] newP = bestTup[1] currentPoint = newP sum += dist return sum Answer: Since you don't have so many point, you can easily use a solution that try every possibility. Here is what you can do: First get all combinations: >>> list_of_points = [(1,2) , (2,5), (6,7), (9,3)] >>> list(itertools.permutations(list_of_points)) [((1, 2), (2, 5), (6, 7), (9, 3)), ((1, 2), (2, 5), (9, 3), (6, 7)), ((1, 2), (6, 7), (2, 5), (9, 3)), ((1, 2), (6, 7), (9, 3), (2, 5)), ((1, 2), (9, 3), (2, 5), (6, 7)), ((1, 2), (9, 3), (6, 7), (2, 5)), ((2, 5), (1, 2), (6, 7), (9, 3)), ((2, 5), (1, 2), (9, 3), (6, 7)), ((2, 5), (6, 7), (1, 2), (9, 3)), ((2, 5), (6, 7), (9, 3), (1, 2)), ((2, 5), (9, 3), (1, 2), (6, 7)), ((2, 5), (9, 3), (6, 7), (1, 2)), ((6, 7), (1, 2), (2, 5), (9, 3)), ((6, 7), (1, 2), (9, 3), (2, 5)), ((6, 7), (2, 5), (1, 2), (9, 3)), ((6, 7), (2, 5), (9, 3), (1, 2)), ((6, 7), (9, 3), (1, 2), (2, 5)), ((6, 7), (9, 3), (2, 5), (1, 2)), ((9, 3), (1, 2), (2, 5), (6, 7)), ((9, 3), (1, 2), (6, 7), (2, 5)), ((9, 3), (2, 5), (1, 2), (6, 7)), ((9, 3), (2, 5), (6, 7), (1, 2)), ((9, 3), (6, 7), (1, 2), (2, 5)), ((9, 3), (6, 7), (2, 5), (1, 2))] Then create a function that give you the length of a combination: def combination_length(start_point, combination): lenght = 0 previous = start_point for elem in combination: lenght += manhattanDistance(previous, elem) return length Finally a function that test every possibility: def get_shortest_path(start_point, list_of_point): min = sys.maxint combination_min = None list_of_combinations = list(itertools.permutations(list_of_points)) for combination in list_of_combination: length = combination_length(start_point, combination) if length < min: min = length combination_min = combination return combination_min Then finally you can have: import sys, itertools def combination_length(start_point, combination): lenght = 0 previous = start_point for elem in combination: lenght += manhattanDistance(previous, elem) return length def get_shortest_path(start_point, list_of_point): min = sys.maxint combination_min = None list_of_combinations = list(itertools.permutations(list_of_points)) for combination in list_of_combination: length = combination_length(start_point, combination) if length < min: min = length combination_min = combination return combination_min list_of_points = [(1,2) , (2,5), (6,7), (9,3)] print get_shortest_path((2,1), list_of_points)
Prepare my bigdata with Spark via Python Question: My 100m in size, quantized data: (1424411938', [3885, 7898]) (3333333333', [3885, 7898]) Desired result: (3885, [3333333333, 1424411938]) (7898, [3333333333, 1424411938]) So what I want, is to transform the data so that I group 3885 (for example) with all the `data[0]` that have it). Here is what I did in [python](/questions/tagged/python "show questions tagged 'python'"): def prepare(data): result = [] for point_id, cluster in data: for index, c in enumerate(cluster): found = 0 for res in result: if c == res[0]: found = 1 if(found == 0): result.append((c, [])) for res in result: if c == res[0]: res[1].append(point_id) return result but when I `mapPartitions()`'ed `data` RDD with `prepare()`, it seem to do what I want only in the current partition, thus return a bigger result than the desired. For example, if the 1st record in the start was in the 1st partition and the 2nd in the 2nd, then I would get as a result: (3885, [3333333333]) (7898, [3333333333]) (3885, [1424411938]) (7898, [1424411938]) How to modify my `prepare()` to get the desired effect? Alternatively, how to process the result that `prepare()` produces, so that I can get the desired result? * * * As you may already have noticed from the code, I do not care about speed at all. Here is a way to create the data: data = [] from random import randint for i in xrange(0, 10): data.append((randint(0, 100000000), (randint(0, 16000), randint(0, 16000)))) data = sc.parallelize(data) Answer: You can use a bunch of basic pyspark transformations to achieve this. >>> rdd = sc.parallelize([(1424411938, [3885, 7898]),(3333333333, [3885, 7898])]) >>> r = rdd.flatMap(lambda x: ((a,x[0]) for a in x[1])) We used `flatMap` to have a key, value pair for every item in `x[1]` and we changed the data line format to `(a, x[0])`, the `a` here is every item in `x[1]`. To understand `flatMap` better you can look to the documentation. >>> r2 = r.groupByKey().map(lambda x: (x[0],tuple(x[1]))) We just grouped all key, value pairs by their keys and used tuple function to convert iterable to tuple. >>> r2.collect() [(3885, (1424411938, 3333333333)), (7898, (1424411938, 3333333333))] As you said you can use [:150] to have first 150 elements, I guess this would be proper usage: `r2 = r.groupByKey().map(lambda x: (x[0],tuple(x[1])[:150]))` I tried to be as explanatory as possible. I hope this helps.
Openpyxl. Max_columns giving error Question: I'm using openpyxl-2.4.0-b1 and Python version 34. Following is my code: from openpyxl import load_workbook from openpyxl import Workbook filename= str(input('Please enter the filename name, with the entire path and extension: ')) wb = load_workbook(filename) ws = wb.worksheets[0] row_main = 1 #Main Program starts here. Loop for the entire file print ('Col =', ws.max_column()) while (row_main <(ws.max_row())): #Process rows for each column print ('ROW =', row_main) row_main += 1 It runs into an error: Traceback (most recent call last): print ('Col =', ws.max_column()) TypeError: 'int' object is not callable I can't use _get_ because it has been deprecated. Many thanks in advance. Answer: As `max_column` is `property`, so just use the name to get its value: print('Col =', ws.max_column) Also apply on `max_row`: while (row_main < ws.max_row):
Showing results python command to the web cgi Question: I have a python script and it runs well if executed in a terminal or command line , but after I try even internal server error occurred . how to enhance the script to be run on the web. HTML <html><body> <form enctype="multipart/form-data" action="http://localhost/cgi-bin/coba5.py" method="post"> <p>File: <input type="file" name="file"></p> <p><input type="submit" value="Upload"></p> </form> </body></html> PYTHON #!/usr/bin/python import cgi, os import cgitb; cgitb.enable() import simplejson as json import optparse import sys try: # Windows needs stdio set for binary mode. import msvcrt msvcrt.setmode (0, os.O_BINARY) # stdin = 0 msvcrt.setmode (1, os.O_BINARY) # stdout = 1 except ImportError: pass form = cgi.FieldStorage() # A nested FieldStorage instance holds the file fileitem = form['file'] fn = os.path.basename(fileitem.filename) data = open('/tmp/upload-cgi/' + fn, 'wb').write(fileitem.file.read()) data=fileitem.file.read() location_database = open('/home/bioinformatics2/DW/taxo.json', 'r') database = json.load(location_database) for line in data: for taxonomy in database: if taxonomy["genus"] == line.replace('\n','') : print "Kingdom: %s," % taxonomy['kingdom'], print "phylum: %s," % taxonomy['phylum'], print "class: %s," % taxonomy['class'], print "order: %s," % taxonomy['order'], print "family: %s," % taxonomy['family'], print "genus: %s" % taxonomy['genus'] break else: print ("No found genus on taxanomy") Answer: You're not outputting the HTTP header. At a minimum, you need to output: print "Content-Type: text/plain" print "" You are also reading the file twice: data = open('/tmp/upload-cgi/' + fn, 'wb').write(fileitem.file.read()) data=fileitem.file.read() The second `fileitem.file.read()` will probably not read any content, since the file should already be at the end-of-file, and so `data` will be empty.
Is there a tool to check what names I have used from a "wildly" imported module? Question: I've been using python to do computations for my research. In an effort to clean up my terrible code, I've been reading [Code Like a Pythonista: Idiomatic Python](http://python.net/~goodger/projects/pycon/2007/idiomatic/handout.html) by David Goodger. In this article, Goodger advises against "wild-card" imports of the form from module import * My code uses a lot of these. I'd like to clean my code up, but I'm not sure how. I'm curious if there is a way to check what names from `module` I have used in my code. This way I could either explicitly import these names or replace the names with `module.name`. Is there a tool designed to accomplish such a task? Answer: Use a tool like `pyflakes` (which you should use anyway) to note which names in your code become undefined after you replace `from module import *` with `import module`. Once you've positively identified every instance of a name imported from `module`, you can assess whether to 1. Always use `import module` and `module.x` for `x` imported from `module`. 2. Use `import module as m` and `m.x` if the module name is long. 3. Selectively import some names from `module` into the module namespace with `from module import x, y, z` The above three are not mutually exclusive; as an extreme example, you can use all three in the same module: import really.long.module.name import really.long.module.name as rlmn from really.long.module.name import obvious_name really.long.module.name.foo() # full module name rlmn.bar() # module alias obvious_name() # imported name all in the same code. (I don't _recommend_ using all three in the same module. Stick with either the full module name _or_ the alias throughout a single module, but there is no harm importing common, obvious names directly and using the fully qualified name for more obscure module attributes.)
Python sorting nested dictionary list Question: I am trying to sort a nested dictionary in Python. Right now I have a dictionary of dictionaries. I was able to sort the outer keys using sorted on the list before I started building the dictionary, but I am unable to get the inner keys sorted at this time. I've been trying to mess with the sorted API but am having problems with it. Right now I have: myDict = {'A': {'Key3':4,'Key2':3,'Key4':2,'Key1':1}, 'B': {'Key4':1,'Key3':2,'Key2':3,'Key1':4}, 'C': {'Key1':4,'Key2':2,'Key4':1,'Key3':3}}; But I would like: myDict = {'A': {'Key1':1,'Key2':3,'Key3':4,'Key4':2}, 'B': {'Key1':4,'Key2':3,'Key2':1,'Key4':1}, 'C': {'Key1':4,'Key2':2,'Key3':3,'Key4':1}}; I appreciate the help! Answer: >>> from collections import OrderedDict >>> def sortedDict(items): ... return OrderedDict(sorted(items)) >>> myDict = {'A': {'Key3':4,'Key2':3,'Key4':2,'Key1':1}, ... 'B': {'Key4':1,'Key3':2,'Key2':3,'Key1':4}, ... 'C': {'Key1':4,'Key2':2,'Key4':1,'Key3':3}} >>> sortedDict((key, sortedDict(value.items())) for key, value in myDict.items())
Create gantt chart with hlines? Question: I've tried for several hours to make this work. I tried using 'python-gantt' package, without luck. I also tried plotly (which was beautiful, but I can't host my sensitive data on their site, so that won't work). My starting point is code from here: [How to plot stacked event duration (Gantt Charts) using Python Pandas?](http://stackoverflow.com/questions/31820578/how-to-plot-stacked- event-duration-gantt-charts-using-python-pandas) Three Requirements: * Include the 'Name' on the y axis rather than the numbers. * If someone has multiple events, put all the event periods on one line (this will make pattern identification easier), e.g. Lisa will only have one line on the visual. * Include the 'Event' listed on top of the corresponding line (if possible), e.g. Lisa's first line would say "Hire". The code will need to be dynamic to accommodate many more people and more possible event types... I'm open to suggestions to visualize: I want to show the duration for various staffing events throughout the year, as to help identify patterns. from datetime import datetime import pandas as pd import matplotlib.pyplot as plt import matplotlib.dates as dt df = pd.DataFrame({'Name': ['Joe','Joe','Lisa','Lisa','Lisa','Alice'], 'Event': ['Hire','Term','Hire','Transfer','Term','Term'], 'Start_Date': ["2014-01-01","2014-02-01","2015-01-01","2015-02-01","2015-03-01","2016-01-01"], 'End_Date': ["2014-01-31","2014-03-15","2015-01-31","2015-02-28","2015-05-01","2016-09-01"] }) df = df[['Name','Event','Start_Date','End_Date']] df.Start_Date = pd.to_datetime(df.Start_Date).astype(datetime) df.End_Date = pd.to_datetime(df.End_Date).astype(datetime) fig = plt.figure() ax = fig.add_subplot(111) ax = ax.xaxis_date() ax = plt.hlines(df.index, dt.date2num(df.Start_Date), dt.date2num(df.End_Date)) Answer: I encountered the same problem in the past. You seem to appreciate the aesthetics of [Plotly](https://plot.ly/python/gantt/). Here is a little piece of code which uses [matplotlib.pyplot.broken_barh](http://matplotlib.org/api/pyplot_api.html#matplotlib.pyplot.broken_barh) instead of [matplotlib.pyplot.hlines](http://matplotlib.org/api/pyplot_api.html#matplotlib.pyplot.hlines). [![Gantt Chart with broken_barh](http://i.stack.imgur.com/Bpr3I.png)](http://i.stack.imgur.com/Bpr3I.png) from collections import defaultdict from datetime import datetime from datetime import date import pandas as pd import matplotlib.dates as mdates import matplotlib.patches as mpatches import matplotlib.pyplot as plt df = pd.DataFrame({ 'Name': ['Joe', 'Joe', 'Lisa', 'Lisa', 'Lisa', 'Alice'], 'Event': ['Hire', 'Term', 'Hire', 'Transfer', 'Term', 'Term'], 'Start_Date': ['2014-01-01', '2014-02-01', '2015-01-01', '2015-02-01', '2015-03-01', '2016-01-01'], 'End_Date': ['2014-01-31', '2014-03-15', '2015-01-31', '2015-02-28', '2015-05-01', '2016-09-01'] }) df = df[['Name', 'Event', 'Start_Date', 'End_Date']] df.Start_Date = pd.to_datetime(df.Start_Date).astype(datetime) df.End_Date = pd.to_datetime(df.End_Date).astype(datetime) names = df.Name.unique() nb_names = len(names) fig = plt.figure() ax = fig.add_subplot(111) bar_width = 0.8 default_color = 'blue' colors_dict = defaultdict(lambda: default_color, Hire='green', Term='red', Transfer='orange') # Plot the events for index, name in enumerate(names): mask = df.Name == name start_dates = mdates.date2num(df.loc[mask].Start_Date) end_dates = mdates.date2num(df.loc[mask].End_Date) durations = end_dates - start_dates xranges = zip(start_dates, durations) ymin = index - bar_width / 2.0 ywidth = bar_width yrange = (ymin, ywidth) facecolors = [colors_dict[event] for event in df.loc[mask].Event] ax.broken_barh(xranges, yrange, facecolors=facecolors, alpha=1.0) # you can set alpha to 0.6 to check if there are some overlaps # Shrink the x-axis box = ax.get_position() ax.set_position([box.x0, box.y0, box.width * 0.8, box.height]) # Add the legend patches = [mpatches.Patch(color=color, label=key) for (key, color) in colors_dict.items()] patches = patches + [mpatches.Patch(color=default_color, label='Other')] plt.legend(handles=patches, bbox_to_anchor=(1, 0.5), loc='center left') # Format the x-ticks ax.xaxis.set_major_locator(mdates.YearLocator()) ax.xaxis.set_major_formatter(mdates.DateFormatter('%Y')) ax.xaxis.set_minor_locator(mdates.MonthLocator()) # Format the y-ticks ax.set_yticks(range(nb_names)) ax.set_yticklabels(names) # Set the limits date_min = date(df.Start_Date.min().year, 1, 1) date_max = date(df.End_Date.max().year + 1, 1, 1) ax.set_xlim(date_min, date_max) # Format the coords message box ax.format_xdata = mdates.DateFormatter('%Y-%m-%d') # Set the title ax.set_title('Gantt Chart') plt.show() I hope this will help you.
How to play music using pygame (Python) Question: I am trying to play an .mp3 file using pygame. Here is my code: import pygame pygame.init() pygame.mixer.init() pygame.mixer.music.load('MSM.mp3') pygame.mixer.music.play(0) pygame.event.wait() This however does not actually play the file. The audio file is located within the same folder as the .py file. Answer: The pyGame docs suggest using the pygame.mixer.pre_init() method before the top-level pygame.init() on some platforms. > "Some platforms require the pygame.mixer module for loading and playing > sounds module to be initialized after the display modules have initialized. > The top level pygame.init() takes care of this automatically, but cannot > pass any arguments to the mixer init. To solve this, mixer has a function > pygame.mixer.pre_init() to set the proper defaults before the toplevel init > is used." <https://www.pygame.org/docs/ref/mixer.html#pygame.mixer.init>
scipy.interpolate leads to ImportError Question: my setup is import cartopy.crs as ccrs import matplotlib.pyplot as plt I have scipy `0.17` and cartopy '0.14.2'. All I'm trying to do is plt.axes(projection=ccrs.PlateCarree()) and it leads to this: Traceback (most recent call last): File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/IPython/core/interactiveshell.py", line 2885, in run_code exec(code_obj, self.user_global_ns, self.user_ns) File "<ipython-input-93-636aeb1a7fc6>", line 1, in <module> plt.axes(projection=ccrs.PlateCarree()) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/pyplot.py", line 867, in axes return subplot(111, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/pyplot.py", line 1022, in subplot a = fig.add_subplot(*args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/figure.py", line 987, in add_subplot self, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/projections/__init__.py", line 100, in process_projection_requirements projection_class, extra_kwargs = projection._as_mpl_axes() File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/crs.py", line 150, in _as_mpl_axes import cartopy.mpl.geoaxes as geoaxes File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/mpl/geoaxes.py", line 52, in <module> from cartopy.vector_transform import vector_scalar_to_grid File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/vector_transform.py", line 26, in <module> from scipy.interpolate import griddata File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/interpolate/__init__.py", line 158, in <module> from .interpolate import * File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/interpolate/interpolate.py", line 12, in <module> import scipy.special as spec File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/special/__init__.py", line 629, in <module> from .basic import * File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/special/basic.py", line 14, in <module> from ._ufuncs import (ellipkm1, mathieu_a, mathieu_b, iv, jv, gamma, psi, zeta, ImportError: cannot import name 'zeta' Deep down this appears to be a scipy problem, but I have the newest there - what's going on here? Answer: It turned out to be a problem with the `scipy` installation coming from `conda`. Creating a new environment and freshly installing `scipy` solved the issue.
Matplotlib: Every tick in different color Question: I'm trying to create a scatter-plot with matplotlib (python 3.5) in which every tick on the x-axes has a different color. How is this possible? For example let's say the x-ticks are 'Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su'. Now I want 'Mo' to be green, 'Tu' to be blue, ect... Here is a very simple version of my code: from matplotlib import pyplot as plt plt.figure(figsize=(16, 11)) x = [1, 2, 3, 4, 5, 6, 7] y = [10, 12, 9, 10, 8, 11, 10] plt.scatter(x, y) plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su']) plt.show() I already tried my_colors = ['c', 'b', 'r', 'r', 'g', 'k', 'b'] plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su'], color=my_colors) but that doesn't work. Answer: You can loop over the tick labels (using `plt.gca().get_xticklabels()`) and set their colors after they have been created, using `.set_color()`. For example: from matplotlib import pyplot as plt plt.figure(figsize=(16, 11)) x = [1, 2, 3, 4, 5, 6, 7] y = [10, 12, 9, 10, 8, 11, 10] plt.scatter(x, y) plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su']) my_colors = ['c', 'b', 'r', 'r', 'g', 'k', 'b'] for ticklabel, tickcolor in zip(plt.gca().get_xticklabels(), my_colors): ticklabel.set_color(tickcolor) plt.show() [![enter image description here](http://i.stack.imgur.com/7o24S.png)](http://i.stack.imgur.com/7o24S.png)
i want to scrape data using python script Question: I have written python script to scrape data from <http://www.cricbuzz.com/cricket-stats/icc-rankings/batsmen-rankings> It is a list of 100 players and I successfully scraped this data. The problem is, when i run script instead of scraping data just one time it scraped the same data 3 times. <div class="cb-col cb-col-100 cb-font-14 cb-lst-itm text-center"> <div class="cb-col cb-col-16 cb-rank-tbl cb-font-16">1</div> <div class="cb-col cb-col-50 cb-lst-itm-sm text-left"> <div class="cb-col cb-col-33"> <div class="cb-col cb-col-50"> <span class=" cb-ico" style="position:absolute;"></span>&nbsp;&nbsp;&nbsp;&nbsp;– </div> <div class="cb-col cb-col-50"> <img src="http://i.cricketcb.com/i/stats/fw/50x50/img/faceImages/2250.jpg" class="img-responsive cb-rank-plyr-img"> </div> </div> <div class="cb-col cb-col-67 cb-rank-plyr"> <a class="text-hvr-underline text-bold cb-font-16" href="/profiles/2250/steven-smith" title="Steven Smith's Profile">Steven Smith</a> <div class="cb-font-12 text-gray">AUSTRALIA</div> </div> </div> <div class="cb-col cb-col-17 cb-rank-tbl">906</div> <div class="cb-col cb-col-17 cb-rank-tbl">1</div> </div> And here is python script which i write scrap each player data. import sys,requests,csv,io from bs4 import BeautifulSoup from urllib.parse import urljoin url = "http://www.cricbuzz.com/cricket-stats/icc-rankings/batsmen-rankings" r = requests.get(url) r.content soup = BeautifulSoup(r.content, "html.parser") maindiv = soup.find_all("div", {"class": "text-center"}) for div in maindiv: print(div.text) but instead of scraping the data once, it scrapes the same data 3 times. Where can I make changes to get data just one time? Answer: Select the table and look for the divs in that: maindiv = soup.select("#batsmen-tests div.text-center") for div in maindiv: print(div.text) Your original output and that above gets all the text from the divs as one line which is not really useful, if you just want the player names: anchors = soup.select("#batsmen-tests div.cb-rank-plyr a") for a in anchors: print(a.text) A quick and easy way to get the data in a nice csv format is to just get text from each child: maindiv = soup.select("#batsmen-tests div.text-center") for d in maindiv[1:]: row_data = u",".join(s.strip() for s in filter(None, (t.find(text=True, recursive=False) for t in d.find_all()))) if row_data: print(row_data) Now you get output like: # rank, up/down, name, country, rating, best rank 1,–,Steven Smith,AUSTRALIA,906,1 2,–,Joe Root,ENGLAND,878,1 3,–,Kane Williamson,NEW ZEALAND,876,1 4,–,Hashim Amla,SOUTH AFRICA,847,1 5,–,Younis Khan,PAKISTAN,845,1 6,–,Adam Voges,AUSTRALIA,802,5 7,–,AB de Villiers,SOUTH AFRICA,802,1 8,–,Ajinkya Rahane,INDIA,785,8 9,2,David Warner,AUSTRALIA,772,3 10,–,Alastair Cook,ENGLAND,770,2 11,1,Misbah-ul-Haq,PAKISTAN,764,6 As opposed to: PositionPlayerRatingBest Rank Player 1    –Steven SmithAUSTRALIA9061 2    –Joe RootENGLAND8781 3    –Kane WilliamsonNEW ZEALAND8761 4    –Hashim AmlaSOUTH AFRICA8471 5    –Younis KhanPAKISTAN8451 6    –Adam VogesAUSTRALIA8025
In 'IDA PRO', let 'IDAPython' import default module at startup Question: We know 'IDAPython' loads several modules by default at startup, such as idaapi, idautils.... I wrote a module to let python print all numbers as hex format in the command window, which I wish can be imported each time when python loads those default modules. how to achieve that? Answer: Create a file `%APPDATA%\Hex-Rays\IDA Pro\idapythonrc.py` with the following content: import idaapi idaapi.require('mymodule') With this file in place, you can even keep `mymodule.py` in the same directory. P.S. IDA can tell you the path to this directory, too, which could be handy on other OSes or if the relevant company name changes again ;-). Just enter: get_user_idadir() at the prompt.
Sending gzip compressed data through TCP socket in Python Question: I'm creating an HTTP server in Python without any of the HTTP libraries for learning purposes. Right now it can serve static files fine. The way I serve the file is through this piece of code: with open(self.filename, 'rb') as f: src = f.read() socket.sendall(src) However, I want to optimize its performance a bit by sending compressed data instead of uncompressed. I know that my browser (Chrome) accepts compressed data because it tells me in the header Accept-Encoding: gzip, deflate, sdch So, I changed my code to this with open(self.filename, 'rb') as f: src = zlib.compress(f.read()) socket.sendall(src) But this just outputs garbage. What am I doing wrong? Answer: The zlib library implements the deflate compression algorithm (RFC 1951). There are two encapsulations for the deflate compression: zlib (RFC 1950) and gzip (RFC 1952). These differ only in the kind of header and trailer they provide around the deflate compressed data. `zlib.compress` only provides the raw deflate data, without any header and trailer. To get these you need to use a compression object. For gzip this looks like this: z = zlib.compressobj(-1,zlib.DEFLATED,31) gzip_compressed_data = z.compress(data) + z.flush() The important part here is the `31` as the 3rd argument to `compressobj`. This specifies gzip format which then can be used with `Content-Encoding: gzip`.
Convert strings of bytes to floats in Python Question: I'm using Pandas to handle my data. The first step of my script is to convert a column of strings of bytes to a list of floats. My script is working but its taking too long. Any suggestions on how to speed it up?? def byte_to_hex(byte_str): array = ["%02X" % ord(chr(x)) for x in byte_str] return array for index, row in enumerate(data[column].values): row_new = [] row = byte_to_hex(row) stop = len(row) for i in range(4, stop, 4): sample = "".join(row[i:i + 4]) row_new.append(struct.unpack('!f', bytes.fromhex(sample))[0]) Example: b'\x00\x00\x00\x1cB\x80\x1f\xfcB\x80\x1f\xfcB\x80w\xc8Bz\xa1\x97B\x80\x1f\xfcB}LZB\x80\x1f\xfcBz\xa1\x97Bz\xcaoB\x803\xf5B}\xc5\x84B\x80w\xc8B}\xed\xdbB\x80\x1f\xfcB}\xc5\x84B}LZB\x80\x1f\xfcB}#\xe9B}\xed\xdbB}\xc5\x84B\x803\xf5B\x80\x1f\xfcB}\xc5\x84B\x803\xf5B\x803\xf5Bx\xef\tB\x81\xc4\xdf\x7f\xff\xff\xff' [64.06246948242188, 64.06246948242188, 64.23394775390625, 62.65780258178711, 64.06246948242188, 63.324562072753906, 64.06246948242188, 62.65780258178711, 62.697689056396484, 64.10147857666016, 63.44288635253906, 64.23394775390625, 63.48228073120117, 64.06246948242188, 63.44288635253906, 63.324562072753906, 64.06246948242188, 63.28506851196289, 63.48228073120117, 63.44288635253906, 64.10147857666016, 64.06246948242188, 63.44288635253906, 64.10147857666016, 64.10147857666016] Very thankful for any help :) Answer: I think you are looking for the [Struct](https://docs.python.org/3.4/library/struct.html) package import struct struct.pack('f', 3.14) OUT: b'\xdb\x0' struct.unpack('f', b'\xdb\x0fI@') OUT: (3.1415927410125732,) struct.pack('4f', 1.0, 2.0, 3.0, 4.0) OUT: '\x00\x00\x80?\x00\x00\x00@\x00\x00@@\x00\x00\x80@'
Adding an extra hidden layer using Google's TensorFlow Question: I am trying to create a multi-label classifier using TensorFlow. Though I'm having trouble adding and connecting hidden layers. I was following this tutorial: <http://jrmeyer.github.io/tutorial/2016/02/01/TensorFlow-Tutorial.html> The data I'm using is UCI's Iris data, encoded to one-hot: Training X [105,4] 5,3.2,1.2,0.2 5.5,3.5,1.3,0.2 4.9,3.1,1.5,0.1 4.4,3,1.3,0.2 5.1,3.4,1.5,0.2 . . . Training Y [105,3] 0,0,1 0,0,1 0,0,1 0,0,1 0,0,1 0,0,1 . . . I'm also using testing data X and Y which are [45,4] and [45,3] respectively. Here is my python code: import tensorflow as tf import numpy as np import tarfile import os import matplotlib.pyplot as plt import time ## Import data def csv_to_numpy_array(filePath, delimiter): return np.genfromtxt(filePath, delimiter=delimiter, dtype=None) trainX = csv_to_numpy_array("Iris_training_x.csv", delimiter=",").astype(np.float32) trainY = csv_to_numpy_array("Iris_training_y.csv", delimiter=",").astype(np.float32) testX = csv_to_numpy_array("Iris_testing_x.csv", delimiter=",").astype(np.float32) testY = csv_to_numpy_array("Iris_testing_y.csv", delimiter=",").astype(np.float32) # Data Set Paramaters numFeatures = trainX.shape[1] numLabels = trainY.shape[1] # Training Session Parameters numEpochs = 1000 learningRate = tf.train.exponential_decay(learning_rate=0.008, global_step= 1, decay_steps=trainX.shape[0], decay_rate= 0.95, staircase=True) # Placeholders X=tf.placeholder(tf.float32, [None, numFeatures]) y=tf.placeholder(tf.float32, [None, numLabels]) # Initialize our weights and biases Weights = tf.Variable(tf.random_normal([numFeatures, numLabels], mean=0, stddev=(np.sqrt(6 / numFeatures + numLabels + 1)), name="Weights")) bias = tf.Variable(tf.random_normal([1, numLabels], mean=0, stddev=(np.sqrt(6 / numFeatures + numLabels + 1)), name="bias")) # Prediction algorithm (feedforward) apply_weights_OP = tf.matmul(X, Weights, name="apply_weights") add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias") activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation") numFeatures = activation_OP apply_weights_OP = tf.matmul(X, Weights, name="apply_weights") add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias") activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation") init_OP = tf.initialize_all_variables() # Cost function (Mean Squeared Error) cost_OP = tf.nn.l2_loss(activation_OP-y, name="squared_error_cost") # Optimization Algorithm (Gradient Descent) training_OP = tf.train.GradientDescentOptimizer(learningRate).minimize(cost_OP) # Visualize epoch_values=[] accuracy_values=[] cost_values=[] # Turn on interactive plotting plt.ion() # Create the main, super plot fig = plt.figure() # Create two subplots on their own axes and give titles ax1 = plt.subplot("211") ax1.set_title("TRAINING ACCURACY", fontsize=18) ax2 = plt.subplot("212") ax2.set_title("TRAINING COST", fontsize=18) plt.tight_layout() # Create a tensorflow session sess = tf.Session() # Initialize all tensorflow variables sess.run(init_OP) ## Ops for vizualization # argmax(activation_OP, 1) gives the label our model thought was most likely # argmax(y, 1) is the correct label correct_predictions_OP = tf.equal(tf.argmax(activation_OP,1),tf.argmax(y,1)) # False is 0 and True is 1, what was our average? accuracy_OP = tf.reduce_mean(tf.cast(correct_predictions_OP, "float")) # Summary op for regression output activation_summary_OP = tf.histogram_summary("output", activation_OP) # Summary op for accuracy accuracy_summary_OP = tf.scalar_summary("accuracy", accuracy_OP) # Summary op for cost cost_summary_OP = tf.scalar_summary("cost", cost_OP) # Summary ops to check how variables (W, b) are updating after each iteration weightSummary = tf.histogram_summary("Weights", Weights.eval(session=sess)) biasSummary = tf.histogram_summary("biases", bias.eval(session=sess)) # Merge all summaries all_summary_OPS = tf.merge_all_summaries() # Summary writer writer = tf.train.SummaryWriter("summary_logs", sess.graph_def) # Initialize reporting variables cost = 0 diff = 1 # Training epochs for i in range(numEpochs): if i > 1 and diff < .0001: print("change in cost %g; convergence."%diff) break else: # Run training step step = sess.run(training_OP, feed_dict={X: trainX, y: trainY}) # Report occasional stats if i % 10 == 0: #Add epoch to epoch_values epoch_values.append(i) #Generate accuracy stats on test data summary_results, train_accuracy, newCost = sess.run( [all_summary_OPS, accuracy_OP, cost_OP], feed_dict={X: trainX, y: trainY} ) # Add accuracy to live graphing variable accuracy_values.append(train_accuracy) # Add cost to live graphing variable cost_values.append(newCost) #Write summary stats to writer #writer.add_summary(summary_results, i) # Re-assign values for variables diff = abs(newCost - cost) cost = newCost #generate print statements print("step %d, training accuracy %g"%(i, train_accuracy)) print("step %d, cost %g"%(i, newCost)) print("step %d, change in cost %g"%(i, diff)) # Plot progress to our two subplots accuracyLine, = ax1.plot(epoch_values, accuracy_values) costLine, = ax2.plot(epoch_values, cost_values) fig.canvas.draw() #time.sleep(1) # How well do we perform on held-out test data? print("final accuracy on test set: %s" %str(sess.run(accuracy_OP, feed_dict={X: testX, y: testY}))) # Create Saver saver = tf.train.Saver() # Save variables to .ckpt file # saver.save(sess, "trained_variables.ckpt") # Close tensorflow session sess.close() The problem is here: # Prediction algorithm (feedforward) apply_weights_OP = tf.matmul(X, Weights, name="apply_weights") add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias") activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation") numFeatures = activation_OP apply_weights_OP = tf.matmul(activation_OP, Weights, name="apply_weights") add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias") activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation") My understanding is that the output of one layer should connect to the input of the next one. I just don't know how to modify either the output or input of the layers; it keeps giving me this compatibility error: /usr/bin/python3.5 /home/marco/PycharmProjects/NN_Iris/main Traceback (most recent call last): File "/home/marco/PycharmProjects/NN_Iris/main", line 132, in <module> apply_weights_OP = tf.matmul(activation_OP, Weights, name="apply_weights") File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/ops/math_ops.py", line 1346, in matmul name=name) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/ops/gen_math_ops.py", line 1271, in _mat_mul transpose_b=transpose_b, name=name) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/op_def_library.py", line 703, in apply_op op_def=op_def) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/ops.py", line 2312, in create_op set_shapes_for_outputs(ret) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/ops.py", line 1704, in set_shapes_for_outputs shapes = shape_func(op) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/common_shapes.py", line 94, in matmul_shape inner_a.assert_is_compatible_with(inner_b) File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/tensor_shape.py", line 108, in assert_is_compatible_with % (self, other)) ValueError: Dimensions 3 and 4 are not compatible Process finished with exit code 1 Any suggestions on how to properly connect the two hidden layers? Thanks. Answer: If you want a fully connected network with one hidden layer and an output layer there is how their shapes should look like: # hidden layer weights_hidden = tf.Variable(tf.random_normal([numFeatures, num_nodes]) bias_hidden = tf.Variable(tf.random_normal([num_nodes]) preactivations_hidden = tf.add(tf.matmul(X, weights_hidden), bias_hidden) activations_hidden = tf.nn.sigmoid(preactivations_hidden) # output layer weights_output = tf.Variable(tf.random_normal([num_nodes, numLabels]) bias_output = tf.Variable(tf.random_normal([numLabels]) preactivations_output = tf.add(tf.matmul(activations_hidden, weights_output), bias_output) Where `num_nodes` is number of nodes in the hidden layer that you select yourself. `X` is a `[105, numFeatures]` matrix, `weights_hidden` is `[numFeatures, num_nodes]` matrix, so output of first hidden layer is `[105, num_nodes]`. In the same way, `[105, num_nodes]` multiplied by `[num_nodes, numLabels]` yields `[105, numLabels]` output.
Trouble installing cryptography with pip3 (Ubuntu 16.04 LTS) Question: i'm currently having trouble installing cryptography with pip3. I tried all the other solutions but non of them worked. Here is the output from "sudo -H pip install cryptography": Collecting cryptography Using cached cryptography-1.5.tar.gz Requirement already satisfied (use --upgrade to upgrade): idna>=2.0 in /usr/local/lib/python3.5/dist-packages (from cryptography) Requirement already satisfied (use --upgrade to upgrade): pyasn1>=0.1.8 in /usr/lib/python3/dist-packages (from cryptography) Requirement already satisfied (use --upgrade to upgrade): six>=1.4.1 in /usr/lib/python3/dist-packages (from cryptography) Requirement already satisfied (use --upgrade to upgrade): setuptools>=11.3 in /usr/local/lib/python3.5/dist-packages (from cryptography) Collecting cffi>=1.4.1 (from cryptography) Using cached cffi-1.8.2.tar.gz Requirement already satisfied (use --upgrade to upgrade): pycparser in /usr/local/lib/python3.5/dist-packages (from cffi>=1.4.1->cryptography) Building wheels for collected packages: cryptography, cffi Running setup.py bdist_wheel for cryptography ... error Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cryptography/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" bdist_wheel -d /tmp/tmpme9hnxc5pip-wheel- --python-tag cp35: c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory compilation terminated. Traceback (most recent call last): File "/usr/lib/python3.5/distutils/unixccompiler.py", line 118, in _compile extra_postargs) File "/usr/lib/python3.5/distutils/ccompiler.py", line 909, in spawn spawn(cmd, dry_run=self.dry_run) File "/usr/lib/python3.5/distutils/spawn.py", line 36, in spawn _spawn_posix(cmd, search_path, dry_run=dry_run) File "/usr/lib/python3.5/distutils/spawn.py", line 159, in _spawn_posix % (cmd, exit_status)) distutils.errors.DistutilsExecError: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.5/distutils/core.py", line 148, in setup dist.run_commands() File "/usr/lib/python3.5/distutils/dist.py", line 955, in run_commands self.run_command(cmd) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 161, in run cmd = self.call_command('install_lib', warn_dir=0) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 147, in call_command self.run_command(cmdname) File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command self.distribution.run_command(command) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/install_lib.py", line 11, in run self.build() File "/usr/lib/python3.5/distutils/command/install_lib.py", line 109, in build self.run_command('build_ext') File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command self.distribution.run_command(command) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 77, in run _build_ext.run(self) File "/usr/lib/python3.5/distutils/command/build_ext.py", line 338, in run self.build_extensions() File "/usr/lib/python3.5/distutils/command/build_ext.py", line 447, in build_extensions self._build_extensions_serial() File "/usr/lib/python3.5/distutils/command/build_ext.py", line 472, in _build_extensions_serial self.build_extension(ext) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 198, in build_extension _build_ext.build_extension(self, ext) File "/usr/lib/python3.5/distutils/command/build_ext.py", line 532, in build_extension depends=ext.depends) File "/usr/lib/python3.5/distutils/ccompiler.py", line 574, in compile self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/usr/lib/python3.5/distutils/unixccompiler.py", line 120, in _compile raise CompileError(msg) distutils.errors.CompileError: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules yield saved File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup DirectorySandbox(setup_dir).run(runner) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run return func() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner _execfile(setup_script, ns) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile exec(code, globals, locals) File "/tmp/easy_install-sjss3av8/cffi-1.8.2/setup.py", line 193, in <module> return True File "/usr/lib/python3.5/distutils/core.py", line 163, in setup raise SystemExit("error: " + str(msg)) SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1100, in run_setup run_setup(setup_script, args) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 249, in run_setup raise File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 168, in save_modules saved_exc.resume() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 143, in resume six.reraise(type, exc, self._tb) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/_vendor/six.py", line 685, in reraise raise value.with_traceback(tb) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules yield saved File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup DirectorySandbox(setup_dir).run(runner) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run return func() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner _execfile(setup_script, ns) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile exec(code, globals, locals) File "/tmp/easy_install-sjss3av8/cffi-1.8.2/setup.py", line 193, in <module> return True File "/usr/lib/python3.5/distutils/core.py", line 163, in setup raise SystemExit("error: " + str(msg)) SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmp/pip-build-2saf6qpj/cryptography/setup.py", line 334, in <module> **keywords_with_side_effects(sys.argv) File "/usr/lib/python3.5/distutils/core.py", line 108, in setup _setup_distribution = dist = klass(attrs) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 348, in __init__ self.fetch_build_eggs(attrs['setup_requires']) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 394, in fetch_build_eggs replace_conflicting=True, File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 851, in resolve dist = best[req.key] = env.best_match(req, ws, installer) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1123, in best_match return self.obtain(req, installer) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1135, in obtain return installer(requirement) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 461, in fetch_build_egg return cmd.easy_install(req) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 664, in easy_install return self.install_item(spec, dist.location, tmpdir, deps) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 694, in install_item dists = self.install_eggs(spec, download, tmpdir) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 875, in install_eggs return self.build_and_install(setup_script, setup_base) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1114, in build_and_install self.run_setup(setup_script, setup_base, args) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1102, in run_setup raise DistutilsError("Setup script exited with %s" % (v.args[0],)) distutils.errors.DistutilsError: Setup script exited with error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- Failed building wheel for cryptography Running setup.py clean for cryptography Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cryptography/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" clean --all: c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory compilation terminated. Traceback (most recent call last): File "/usr/lib/python3.5/distutils/unixccompiler.py", line 118, in _compile extra_postargs) File "/usr/lib/python3.5/distutils/ccompiler.py", line 909, in spawn spawn(cmd, dry_run=self.dry_run) File "/usr/lib/python3.5/distutils/spawn.py", line 36, in spawn _spawn_posix(cmd, search_path, dry_run=dry_run) File "/usr/lib/python3.5/distutils/spawn.py", line 159, in _spawn_posix % (cmd, exit_status)) distutils.errors.DistutilsExecError: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.5/distutils/core.py", line 148, in setup dist.run_commands() File "/usr/lib/python3.5/distutils/dist.py", line 955, in run_commands self.run_command(cmd) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 161, in run cmd = self.call_command('install_lib', warn_dir=0) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 147, in call_command self.run_command(cmdname) File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command self.distribution.run_command(command) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/install_lib.py", line 11, in run self.build() File "/usr/lib/python3.5/distutils/command/install_lib.py", line 109, in build self.run_command('build_ext') File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command self.distribution.run_command(command) File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command cmd_obj.run() File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 77, in run _build_ext.run(self) File "/usr/lib/python3.5/distutils/command/build_ext.py", line 338, in run self.build_extensions() File "/usr/lib/python3.5/distutils/command/build_ext.py", line 447, in build_extensions self._build_extensions_serial() File "/usr/lib/python3.5/distutils/command/build_ext.py", line 472, in _build_extensions_serial self.build_extension(ext) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 198, in build_extension _build_ext.build_extension(self, ext) File "/usr/lib/python3.5/distutils/command/build_ext.py", line 532, in build_extension depends=ext.depends) File "/usr/lib/python3.5/distutils/ccompiler.py", line 574, in compile self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/usr/lib/python3.5/distutils/unixccompiler.py", line 120, in _compile raise CompileError(msg) distutils.errors.CompileError: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules yield saved File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup DirectorySandbox(setup_dir).run(runner) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run return func() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner _execfile(setup_script, ns) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile exec(code, globals, locals) File "/tmp/easy_install-rfvxg0ae/cffi-1.8.2/setup.py", line 193, in <module> return True File "/usr/lib/python3.5/distutils/core.py", line 163, in setup raise SystemExit("error: " + str(msg)) SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1100, in run_setup run_setup(setup_script, args) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 249, in run_setup raise File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 168, in save_modules saved_exc.resume() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 143, in resume six.reraise(type, exc, self._tb) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/_vendor/six.py", line 685, in reraise raise value.with_traceback(tb) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules yield saved File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context yield File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup DirectorySandbox(setup_dir).run(runner) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run return func() File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner _execfile(setup_script, ns) File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile exec(code, globals, locals) File "/tmp/easy_install-rfvxg0ae/cffi-1.8.2/setup.py", line 193, in <module> return True File "/usr/lib/python3.5/distutils/core.py", line 163, in setup raise SystemExit("error: " + str(msg)) SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmp/pip-build-2saf6qpj/cryptography/setup.py", line 334, in <module> **keywords_with_side_effects(sys.argv) File "/usr/lib/python3.5/distutils/core.py", line 108, in setup _setup_distribution = dist = klass(attrs) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 348, in __init__ self.fetch_build_eggs(attrs['setup_requires']) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 394, in fetch_build_eggs replace_conflicting=True, File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 851, in resolve dist = best[req.key] = env.best_match(req, ws, installer) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1123, in best_match return self.obtain(req, installer) File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1135, in obtain return installer(requirement) File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 461, in fetch_build_egg return cmd.easy_install(req) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 664, in easy_install return self.install_item(spec, dist.location, tmpdir, deps) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 694, in install_item dists = self.install_eggs(spec, download, tmpdir) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 875, in install_eggs return self.build_and_install(setup_script, setup_base) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1114, in build_and_install self.run_setup(setup_script, setup_base, args) File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1102, in run_setup raise DistutilsError("Setup script exited with %s" % (v.args[0],)) distutils.errors.DistutilsError: Setup script exited with error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- Failed cleaning build dir for cryptography Running setup.py bdist_wheel for cffi ... error Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" bdist_wheel -d /tmp/tmp689jaw0cpip-wheel- --python-tag cp35: running bdist_wheel running build running build_py creating build creating build/lib.linux-x86_64-3.5 creating build/lib.linux-x86_64-3.5/cffi copying cffi/cffi_opcode.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/cparser.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/recompiler.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/vengine_gen.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/commontypes.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/verifier.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/model.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/api.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/setuptools_ext.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/vengine_cpy.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/ffiplatform.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/__init__.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/lock.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/backend_ctypes.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/_cffi_include.h -> build/lib.linux-x86_64-3.5/cffi copying cffi/parse_c_type.h -> build/lib.linux-x86_64-3.5/cffi copying cffi/_embedding.h -> build/lib.linux-x86_64-3.5/cffi running build_ext building '_cffi_backend' extension creating build/temp.linux-x86_64-3.5 creating build/temp.linux-x86_64-3.5/c x86_64-linux-gnu-gcc -pthread -DNDEBUG -g -fwrapv -O2 -Wall -Wstrict-prototypes -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -DUSE__THREAD -I/usr/include/python3.5m -c c/_cffi_backend.c -o build/temp.linux-x86_64-3.5/c/_cffi_backend.o c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory compilation terminated. error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- Failed building wheel for cffi Running setup.py clean for cffi Failed to build cryptography cffi Installing collected packages: cffi, cryptography Running setup.py install for cffi ... error Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" install --record /tmp/pip-7jmqgdpf-record/install-record.txt --single-version-externally-managed --compile: running install running build running build_py creating build creating build/lib.linux-x86_64-3.5 creating build/lib.linux-x86_64-3.5/cffi copying cffi/cffi_opcode.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/cparser.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/recompiler.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/vengine_gen.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/commontypes.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/verifier.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/model.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/api.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/setuptools_ext.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/vengine_cpy.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/ffiplatform.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/__init__.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/lock.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/backend_ctypes.py -> build/lib.linux-x86_64-3.5/cffi copying cffi/_cffi_include.h -> build/lib.linux-x86_64-3.5/cffi copying cffi/parse_c_type.h -> build/lib.linux-x86_64-3.5/cffi copying cffi/_embedding.h -> build/lib.linux-x86_64-3.5/cffi running build_ext building '_cffi_backend' extension creating build/temp.linux-x86_64-3.5 creating build/temp.linux-x86_64-3.5/c x86_64-linux-gnu-gcc -pthread -DNDEBUG -g -fwrapv -O2 -Wall -Wstrict-prototypes -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -DUSE__THREAD -I/usr/include/python3.5m -c c/_cffi_backend.c -o build/temp.linux-x86_64-3.5/c/_cffi_backend.o c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory compilation terminated. error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- Command "/usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" install --record /tmp/pip-7jmqgdpf-record/install-record.txt --single-version-externally-managed --compile" failed with error code 1 in /tmp/pip-build-2saf6qpj/cffi/ I tried understanding the errors but all it says is "failed with error code 1" :/ Checking out other questions, they mentioned about the dependencies for cryptography which is all installed and up-to-date (sudo apt-get install build-essential libssl-dev libffi-dev python-dev): Reading package lists... Done Building dependency tree Reading state information... Done build-essential is already the newest version (12.1ubuntu2). libffi-dev is already the newest version (3.2.1-4). python-dev is already the newest version (2.7.11-1). libssl-dev is already the newest version (1.0.2g-1ubuntu4.2). 0 upgraded, 0 newly installed, 0 to remove and 135 not upgraded. pip3 version: pip 8.1.2 from /usr/local/lib/python3.5/dist-packages (python 3.5) Not even easy_install worked. I really don't know what i'm doing wrong here, please help. Answer: From the `Python.h: No such file or directory` error messages appearing in various places of the log, you need to install `python3-dev`: sudo apt install python3-dev
calculate catalan numbers up to a billion Question: I'm new to python (and programming in general), I was asked in my class to calculate Catalan numbers up to a billion but the program I wrote for it is not working as intended. from numpy import division C=1 n=0 while C<=10000000000: print (C) C=(4*n+2)/(n+2)*C n=n+1 This is what it prints 1, 1, 2, 4, 8, 24, 72, 216, 648, 1944, 5832, 17496, 52488, 157464, 472392, 1417176, 4251528, 12754584, 38263752, 114791256, 344373768, 1033121304, 3099363912, 9298091736, As you can see from my fourth iteration onwards, I get the wrong number and I don't understand why. EDIT: The mathematical definition I used is not wrong! I know the Wiki has another definition but this one is not wrong. Co=1, Cn+1 = (4*n+2)/(n+2)*Cn Answer: C=(4*n+2)/(n+2)*C This applies the calculation in the wrong order. Because you are using integer arithmetic, `(4*n+2)/(n+2)` loses information if you have not already factored in `C`. The correct calculation is: C=C*(4*n+2)/(n+2)
Reassigning a return variable to a function in Python Question: I am trying to figure out why I am getting a "SyntaxError: invalid syntax " the variable title is highlighted red from urllib.request import urlopen from urllib.error import HTTPError from urllib.error import URLError from bs4 import BeautifulSoup def getTitle(url): try: html=urlopen(url) except HTTPError as e: return None try: bsObj=BeautifulSoup(html.read()) title=bsObj.body.h1 except AttributeError as e: return None return title title=getTitle("http://....html") Answer: Python is very sensitive to whitespace, indentation, in order to keep your code looking clean. From what I can see, you need to intent your try/except statements: from urllib.request import urlopen from urllib.error import HTTPError from urllib.error import URLError from bs4 import BeautifulSoup def getTitle(url): try: html=urlopen(url) except HTTPError as e: return None try: bsObj=BeautifulSoup(html.read()) title=bsObj.body.h1 except AttributeError as e: return None return title title=getTitle("http://....html") That should be it, do say if further errors occur.
Extracting text from multiple powerpoint files using python Question: I am trying to find a way to look in a folder and search the contents of all of the powerpoint documents within that folder for specific strings, preferably using Python. When those strings are found, I want to report out the text after that string as well as what document it was found in. I would like to compile the information and report it in a CSV file. So far I've only come across the olefil package, <https://bitbucket.org/decalage/olefileio_pl/wiki/Home>. This provides all of the text contained in a specific document, which is not what I am looking to do. Please help. Answer: `python-pptx` can be used to do what you propose. Just at a high level, you would do something like this (not working code, just and idea of overall approach): from pptx import Presentation for pptx_filename in directory: prs = Presentation(pptx_filename) for slide in prs.slides: for shape in slide.shapes: print shape.text You'd need to add the bits about searching shape text for key strings and adding them to a CSV file or whatever, but this general approach should work just fine. I'll leave it to you to work out the finer points :)
how to parse key value pair request from url using python requests in flask Question: I have spent about a week on this issue and although I have made considerable progress I am stuck at a key point. I am writing a simple client-server program in Python that is supposed to accept key/value pairs from the command line, formulate them into an url, and request the url from the server. The problem appears to be either that the url is not properly formatted or that the server is not parsing it correctly. (That is, the key-value pairs appear to be properly making it from the command line to the function that contains the request.) import sys import requests server_url = "http://0.0.0.0:5006" def test(payload): print('payload in function is ' + payload) r = requests.get('http://0.0.0.0:5006/buy', params=payload) print(r.url) #debug print(r.encoding) print(r.text) print(r.headers) if __name__ == '__main__': payload = sys.argv[2] print('payload from cli is ' + payload) test(payload) Server: import subprocess from flask import Flask from flask import request import request # Configure the app and wallet app = Flask(__name__) @app.route('/buy', methods=['GET', 'POST']) def test(key1, key2): key1 = str(request.args.get('key1')) key2 = str(request.args.get('key2')) print('keys are' + key1 + key2) fortune = subprocess.check_output(['echo', 'key1']) return fortune # Initialize and run the server if __name__ == '__main__': app.run(host='0.0.0.0', port=5006) Client console output: payload from cli is {"key1": "foo", "key2": "bar"} payload in function is {"key1": "foo", "key2": "bar"} http://0.0.0.0:5006/buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D ISO-8859-1 <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN"> <title>500 Internal Server Error</title> <h1>Internal Server Error</h1> <p>The server encountered an internal error and was unable to complete your request. Either the server is overloaded or there is an error in the application.</p> {'Content-Type': 'text/html', 'Content-Length': '291', 'Date': 'Fri, 09 Sep 2016 22:16:44 GMT', 'Server': 'Werkzeug/0.11.10 Python/3.5.2'} Server console output: * Running on http://0.0.0.0:5006/ (Press CTRL+C to quit) [2016-09-09 18:30:33,445] ERROR in app: Exception on /buy [GET] Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1988, in wsgi_app response = self.full_dispatch_request() File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1641, in full_dispatch_request rv = self.handle_user_exception(e) File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1544, in handle_user_exception reraise(exc_type, exc_value, tb) File "/usr/local/lib/python3.5/dist-packages/flask/_compat.py", line 33, in reraise raise value File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1639, in full_dispatch_request rv = self.dispatch_request() File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1625, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) TypeError: test() missing 2 required positional arguments: 'key1' and 'key2' 127.0.0.1 - - [09/Sep/2016 18:30:33] "GET /buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D HTTP/1.1" 500 - I take this as showing that the subprocess call is for some reason not able to decipher key1 and key2 from the URL so is failing when it runs "fortune None None". Answer: The problem seems to be in your payload. The payload needs to be a dictionary. You're giving it a string. `sys.argv[2]` will be a string, even if you format the text to look like a dictionary. So unless there's something missing from your client code snippet, payload isn't actually a dictionary like requests would expect. I can infact confirm this by looking at the URL being generated, which is: http://0.0.0.0:5006/buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D Had the payload been a dictionary and correctly encoded, it would've looked something like this: http://0.0.0.0:5006/buy?key1=foo&key2=bar To see what type payload your really is, do `print(type(payload))` somewhere inside the `test` function. Once you've converted your `payload` into a proper python dictionary (you'll have to parse your `sys.argv[2]`), then requests should work as expected.
python comprehension trubleshooting Question: base=2 digits=set(range(base)) key=range(base**3) dict={ k:[a,b,c] for k in key for a in digits for b in digits for c in digits} print(dict) the output is: {0: [1, 1, 1], 1: [1, 1, 1], 2: [1, 1, 1], 3: [1, 1, 1], 4: [1, 1, 1], 5: [1, 1, 1], 6: [1, 1, 1], 7: [1, 1, 1]} I wonder why the output for [a,b,c] is [1,1,1], not the same as the: [[a,b,c] for a in digits for b in digits for c in digits] which is: [[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1],[1,1,0], [1, 1, 1]] Answer: We can simplify the dictionary comprehension to something like: In [728]: {k:[a,b] for k in [1,2] for a in [1,2] for b in [1,2]} Out[728]: {1: [2, 2], 2: [2, 2]} It is the equivalent to constructing a dictionary with nested loops: In [731]: d={} In [732]: for k in [1,2]: ...: for a in [1,2]: ...: for b in [1,2]: ...: d[k]=[a,b] ...: In [733]: d Out[733]: {1: [2, 2], 2: [2, 2]} But to clarify what is going, let's use a defaultdict, and append the new values: In [730]: from collections import defaultdict In [734]: d=defaultdict(list) In [735]: for k in [1,2]: ...: for a in [1,2]: ...: for b in [1,2]: ...: d[k].append([a,b]) ...: In [736]: d Out[736]: defaultdict(list, {1: [[1, 1], [1, 2], [2, 1], [2, 2]], 2: [[1, 1], [1, 2], [2, 1], [2, 2]]}) It is writing each `[a,b]` pair to the dictionary, but you only end up seeing the last pair for each key. And yes, `@user2357112's` comprehension does the same thing In [737]: {k:[[a,b] for a in [1,2] for b in [1,2]] for k in [1,2]} Out[737]: {1: [[1, 1], [1, 2], [2, 1], [2, 2]], 2: [[1, 1], [1, 2], [2, 1], [2, 2]]}
How to nest or jointly use two template tags in Django templates? Question: I'm trying to use template filters to do run a loop, but I'm unable to combine two python statements within the same statement/template. Whats the correct way to combine two variables in a template? Please see the syntax and explanation below: I'm building a forum with a double index, meaning, I've got a col-md-2 with list of categories. Each categories has forums, and depending on which category is clicked, that category's forums populate the next col-md-2. The remaining col-md-8 gets its content based on which category and which forum is selected. My logic: I've defined a template tag that loads the list of categories, which never change irrespective of which page gets loaded or which category or forum is selected. So that works fine. But based on the selected category, my second column needs to get populated. For this, I'm trying to define a custom filter (below). However, I'm not sure how to use this as it needs to be passed to another template where it runs a loop to render the html. Even if I create the for loop in this template (rather than passing it to another), I still have to do nested template tags, something like: `{% for forum in {{ forum.category|forumindexlistbycategory }} %}` In either cases, I get an error of the type `Invalid filter: 'forumindexlistbycategory'` or `"with" in u'include' tag needs at least one keyword argument`. I've defined the following custom template filter in my pybb_tags.py: from pybb.models import Forum @register.filter def forumindexlistbycat(category): forumlistbycat = Forum.objects.filter(category=category) return forumlistbycat And in my template, I'm trying to load it as follows: {% load i18n pybb_tags %} <div class='category'> {% if category %} <h3>{{ category }}</h3> {% include 'pybb/forumindex_list.html' with forum_list=category.forums_accessed category=category parent_forum='' %} {% else %} <h3>{{ forum.category }}</h3> {% include 'pybb/forumindex_list.html' with forum_list= %}{{ forum.category|forumindexlistbycategory }} {% endif %} </div> Answer: So you must first properly register template tag. from django import template from pybb.models import Forum register = template.Library() @register.filter def forumindexlistbycat(category): forumlistbycat = Forum.objects.filter(category=category) return forumlistbycat Place code from above in file named as your filter, so `forumindexlistbycat.py` and move this file to templatetags folder in your app. If you don't have this folder you must create it. Don't forget to add empty file `__init__.py` in your templatetags folder. And now you could use it in template, so: {% load i18n forumindexlistbycat %} When your templatetag is registered you load it by it's name. And then you use it like: {% include 'pybb/forumindex_list.html' with forum_list=forum.category|forumindexlistbycategory %} Check for more info - [Guide on Custom template tags and filters.](https://docs.djangoproject.com/en/1.10/howto/custom-template-tags/)
Importing functions in R Question: In Python we have chance to import a certain function from a library with a command "import _function_ from _library_ **as smth**. Do we have something similar in R? I know that we can call the function like "_library_ ::_function_()", my question mostly refers to the "as" part. Answer: It is not common and not necessary to do this in R. The assignment operator `<-` can be used to give a new name to an existing function. For example, one could define a function that does exactly the same as lubridate's, `year()` function with: asYear <- lubridate::year One could argue that, by doing so, the `year()` function has been "imported" from the `lubridate` package and that it is now called `asYear()`. In fact, the new function does just the same (which is no surprise, simply because it _is_ the same): asYear(Sys.Date()) #[1] 2016 So it is possible to construct an analogy to "from package import as", but it is not a good idea to do this. Here are a few reasons I can think of: * Debugging a code where library functions have been renamed will be much more difficult. * The documentation is not available for the renamed function. In this example, `?asYear` won't work, in contrast to `?lubridate::year` or `library(lubridate); help(year)`. * The function is not only renamed but it is copied, which clutters the environment and is inefficient in terms of memory usage. * The maintenance of the code becomes unnecessarily difficult. If another programmer (or the original programmer a few years later) looks at a code containing such a redefinition of a function, it will be harder for her or him to understand what this function is doing. There are probably more reasons, but I hope that this is sufficient to discourage the use of such a construction. Different programming languages have different peculiarities, and as a programmer it is necessary to adapt to them. What is common in Python can be awkward in R, and vice versa. A simple and commonly used way to handle such a standard situation in R is to use `library()` to load the entire namespace of the package containing the requested function: library (lubridate) year(Sys.Date()) However, one should be aware of possible namespace clashes, especially if many libraries are loaded simultaneously. Different functions could be defined with the same name in different packages. A well-known example thereof are the contrasting implementations of the `lag()` function in the `dplyr` and `stats` package. In such cases one can use the double colon operator `::` to resolve the namespace that should be addressed. This would be similar to the use of "from" in the case of "import", but such a specification would be needed each time the function is called. lubridate::year(Sys.Date()) #[1] 2016
Using csv training data in tensorflow RNN Question: I am fairly new to tensorflow, and did the obligatory MNIST tutorial successfully. I am trying to train a simple RNN with a set of CSV data. The data is 33 features and a binary output variable at the end (so 34 columns). I have implemented a csv reader that reads in one line at a time. I am trying to read that line and pass it into my tensorflow graph. I feel like the "TensorFlow-way" is starting to become more clear but also that there are some fundamental pieces missing - particularly as it relates to streaming data into your model. I have included an example of what I'm doing below. Most of the code has been stripped for clarity, but the important part remains: import tensorflow as tf import sys import datapipe as datapipe learning_rate = 0.001 n_features = 33 n_hidden = 100 # number of features in the hidden layer - I just made this up n_classes = 2 # 0 or 1 - a binary classification x = tf.placeholder('float', [None, 1, n_features]) y = tf.placeholder('float', [None, n_classes]) transform = tf.transpose(x) with tf.session() as sess: sess.run(tf.initialize_all_variables()) coord = tf.train.Coordinator() threads = tf.train.start_queue_runners(coord=coord) datapipe = datapipe.Datapipe(filename='training.csv', features=33, epochs=100) while not coord.should_stop(): nextline = datapipe.nextline() # I basically want to run "transform" with the nextline of the csv file stuff = sess.run(transform, feed_dict={ x: nextline }) coord.request_stop() coord.join(threads) And `datapipe` is: import tensorflow as tf class Datapipe: def __init__(self, filename=None, features=None, epochs=100): self.filename = filename self.features = features self.epochs = epochs self.defaults = [] for i in range(self.features): self.defaults.append([]) # require all fields to be present def nextline(self): file_queue = tf.train.string_input_producer([self.filename], num_epochs=self.epochs, shuffle=False) reader = tf.TextLineReader() key, csv_str = reader.read(file_queue) return tf.pack(tf.decode_csv(csv_str, record_defaults=self.defaults)) When I run this example, I get the error: `TypeError: The value of a feed cannot be a tf.Tensor object. Acceptable feed values include Python scalars, strings, lists, or numpy ndarrays.` Thank you for all your help! **EDIT** My question is essentially: _How can I feed file data (e.g. csv) into a tensorflow model?_ (the tutorials weren't helpful) **EDIT 12/09/2016** Based on the answer by Sergii, I am doing this now: with open('../data/training2.csv') as f: reader = csv.reader(f) for line in reader: arr = np.array(line) x = arr[0:len(arr)-1:1] y = arr[len(arr)-1:len(arr):1] sess.run(transform, feed_dict={ x: x, y: y }) Answer: `Tensor` objects can not be values of `feed_dict`, it takes actual values, like `numpy` arrays, strings, etc as an input. See for example [this issue.](https://github.com/tensorflow/tensorflow/issues/3389) Try to modify `nextline` method and do not create a `Tensor` there but transform your csv strings into numpy arrays instead.
Python paramiko: redirecting stderr is affected by get_pty = True Question: I am trying to implement an ssh agent that will allow me later, among other things, to execute commands in blocking mode, where output is being read from the channel as soon as it is available. Here's what I have so far: from paramiko import client class SSH_Agent: def __init__(self, server_name, username = getpass.getuser(), password = None, connection_timeout = CONNECTION_TIMEOUT): self.ssh_agent = client.SSHClient() self.ssh_agent.set_missing_host_key_policy(client.AutoAddPolicy()) self.ssh_agent.connect(server_name, username = username, password = password if password is not None else username, timeout = connection_timeout) def execute_command(self, command, out_streams = [sys.stdout], err_streams = [sys.stderr], poll_intervals = POLL_INTERVALS): stdin, stdout, stderr = self.ssh_agent.exec_command(command) channel = stdout.channel stdin.close() channel.shutdown_write() while not channel.closed or channel.recv_ready() or channel.recv_stderr_ready(): got_data = False output_channels = select.select([channel], [], [], poll_intervals)[0] if output_channels: channel = output_channels[0] if channel.recv_ready(): for stream in out_streams: stream.write(channel.recv(len(channel.in_buffer))) stream.flush() got_data = True if channel.recv_stderr_ready(): for stream in err_streams: stream.write(channel.recv_stderr(len(channel.in_stderr_buffer))) stream.flush() got_data = True if not got_data \ and channel.exit_status_ready() \ and not channel.recv_ready() \ and not channel.recv_stderr_ready(): channel.shutdown_read() channel.close() break return channel.recv_exit_status() (this implementation is based on one I found somewhere here in SO) When I test it, it works fine, except that I get this when executing commands: tput: No value for $TERM and no -T specified I read a bit online, and found out that happens because there's no actual terminal behind the `ssh` session. So, I tried to call `paramiko`'s `exec_command()` with `get_pty = True`: stdin, stdout, stderr = self.ssh_agent.exec_command(command, get_pty = True) But then I found out that I'm losing the ability to get data to `stderr` on the channel (everything goes to `stdout` for some reason, i.e. `channel.recv_stderr_ready()` is never `True`). Indeed, I found in the doc the following: > recv_stderr_ready() > > Returns true if data is buffered and ready to be read from this channel’s > stderr stream. **Only channels using exec_command or invoke_shell without a > pty will ever have data on the stderr stream**. > Returns: True if a recv_stderr call on this channel would immediately > return at least one byte; False otherwise. How can I have both? In other words, how can I get rid of this: tput: No value for $TERM and no -T specified while still having the ability to direct `stderr` to wherever I choose? **EDIT** : I Just had an idea... Can I somehow define this `TERM` variable in the remote shell to get rid of that error? Is this a common approach or is it a bad workaround that only hides the problem? Answer: Assuming you're running on a *nix system, you could try setting `TERM` in your command environment instead of creating a pty. For example: def execute_command(self, command, out_streams = [sys.stdout], err_streams = [sys.stderr], poll_intervals = POLL_INTERVALS): # Pre-pend required environment here command = "TERM=xterm " + command stdin, stdout, stderr = self.ssh_agent.exec_command(command) channel = stdout.channel stdin.close() channel.shutdown_write() ... This should still allow you access to the separate streams using your existing code.
Split a 3D numpy array into 3D blocks Question: I would like to split a 3D numpy array into 3D blocks in a 'pythonic' way. I am working with image sequences that are somewhat large arrays (1000X1200X1600), so I need to split them into pieces to do my processing. I have written functions to do this, but I am wondering if there is a native numpy way to accomplish this - numpy.split does not seem to do what I want for 3D arrays (but perhaps I don't understand its functionality) To be clear: the code below accomplishes my task, but I am seeking a faster way to do it. def make_blocks(x,t): #x should be a yXmXn matrix, and t should even divides m,n #returns a list of 3D blocks of size yXtXt down = range(0,x.shape[1],t) across = range(0,x.shape[2],t) reshaped = [] for d in down: for a in across: reshaped.append(x[:,d:d+t,a:a+t]) return reshaped def unmake_blocks(x,d,m,n): #this takes a list of matrix blocks of size dXd that is m*n/d^2 long #returns a 2D array of size mXn rows = [] for i in range(0,int(m/d)): rows.append(np.hstack(x[i*int(n/d):(i+1)*int(n/d)])) return np.vstack(rows) Answer: Here are vectorized versions of those loopy implementations using a combination of permuting dims with [`np.transpose`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.transpose.html) and [`reshaping`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.reshape.html) - def make_blocks_vectorized(x,d): p,m,n = x.shape return x.reshape(-1,m//d,d,n//d,d).transpose(1,3,0,2,4).reshape(-1,p,d,d) def unmake_blocks_vectorized(x,d,m,n): return np.concatenate(x).reshape(m//d,n//d,d,d).transpose(0,2,1,3).reshape(m,n) * * * Sample run for `make_blocks` - In [120]: x = np.random.randint(0,9,(2,4,4)) In [121]: make_blocks(x,2) Out[121]: [array([[[4, 7], [8, 3]], [[0, 5], [3, 2]]]), array([[[5, 7], [4, 0]], [[7, 3], [5, 7]]]), ... and so on. In [122]: make_blocks_vectorized(x,2) Out[122]: array([[[[4, 7], [8, 3]], [[0, 5], [3, 2]]], [[[5, 7], [4, 0]], [[7, 3], [5, 7]]], ... and so on. Sample run for `unmake_blocks` - In [135]: A = [np.random.randint(0,9,(3,3)) for i in range(6)] In [136]: d = 3 In [137]: m,n = 6,9 In [138]: unmake_blocks(A,d,m,n) Out[138]: array([[6, 6, 7, 8, 6, 4, 5, 4, 8], [8, 8, 3, 2, 7, 6, 8, 5, 1], [5, 2, 2, 7, 1, 2, 3, 1, 5], [6, 7, 8, 2, 2, 1, 6, 8, 4], [8, 3, 0, 4, 4, 8, 8, 6, 3], [5, 5, 4, 8, 5, 2, 2, 2, 3]]) In [139]: unmake_blocks_vectorized(A,d,m,n) Out[139]: array([[6, 6, 7, 8, 6, 4, 5, 4, 8], [8, 8, 3, 2, 7, 6, 8, 5, 1], [5, 2, 2, 7, 1, 2, 3, 1, 5], [6, 7, 8, 2, 2, 1, 6, 8, 4], [8, 3, 0, 4, 4, 8, 8, 6, 3], [5, 5, 4, 8, 5, 2, 2, 2, 3]]) * * * Alternative to `make_blocks` with [`view_as_blocks`](http://scikit- image.org/docs/dev/api/skimage.util.html#skimage.util.view_as_blocks) - from skimage.util.shape import view_as_blocks def make_blocks_vectorized_v2(x,d): return view_as_blocks(x,(x.shape[0],d,d)) * * * **Runtime test** 1) `make_blocks` with original and `view_as_blocks` based approaches - In [213]: x = np.random.randint(0,9,(100,160,120)) # scaled down by 10 In [214]: %timeit make_blocks(x,10) 1000 loops, best of 3: 198 µs per loop In [215]: %timeit view_as_blocks(x,(x.shape[0],10,10)) 10000 loops, best of 3: 85.4 µs per loop 2) `unmake_blocks` with original and `transpose+reshape` based approaches - In [237]: A = [np.random.randint(0,9,(10,10)) for i in range(600)] In [238]: d = 10 In [239]: m,n = 10*20,10*30 In [240]: %timeit unmake_blocks(A,d,m,n) 100 loops, best of 3: 2.03 ms per loop In [241]: %timeit unmake_blocks_vectorized(A,d,m,n) 1000 loops, best of 3: 511 µs per loop
Why is my function that I made getting the TypeError: f() takes 0 positional arguments but 1 was given Question: I am working with pandas DataFrames and I am adding new columns for more advanced analysis. My f function is giving me an error TypeError: f() takes 0 positional arguments but 1 was given. I can't figure out why, I have my f function documented in the code if you need to know what it does. from pandas_datareader import data as dreader import pandas as pd from datetime import datetime import dateutil.parser # Sets the max rows that can be displayed # when the program is executed pd.options.display.max_rows = 200 # df is the name of the dataframe, it is # reading the csv file containing date loaded # from yahoo finance(Date,Open,High,Low,Close # volume,adj close,)the name of the ticker # is placed before _data.csv i.e. the ticker aapl # would have a csv file named aapl_data.csv. df = pd.read_csv("cde_data.csv") # the following code will allow for filtering of the datafram # based on the year, day of week (dow), and month. It then gets # applied to the dataframe and then can be used to sort data i.e # print(df[(df.year == 2015) & (df.month == 5) & (df.dow == 4)]) # which will give you all the days in the month of May(df.month == 5), # that fall on a Thursday(df.dow == 4), in the year 2015 # (df.year == 2015) # # Month Day Year # January = 1 Monday = 1 The year will be dispaly in a four # February = 2 Tuesday = 2 digit format i.e. 2015 # March = 3 Wednesday = 3 # April = 4 Thursday = 4 # May = 5 Friday = 5 # June = 6 # July = 7 # August = 8 # September = 9 # October = 10 # November = 11 # December = 12 def year(x): return(x.year) def dow(x): return(x.isoweekday()) def month(x): return(x.month) # f is a function that checks to see if the up_down column # has a value that is greater than, less than, or equal to # zero. The value in the up_down column is derived from # subtracting the opening price of the stock(open column) # from closing price of the stock(close column). If up_down # has a negative value than the stocks price was Down, a positive # value then Up, and no change is Flat def f(): if up_down > 0: x = Up elif up_down < 0: x = Down else: x = Flat return (f) df.reset_index() df.Date = df.Date.apply(dateutil.parser.parse) df['year'] = df.Date.apply(year) df['dow'] = df.Date.apply(dow) df['month'] = df.Date.apply(month) df['up_down'] = df['Close'] - df['Open'] df['up_down_flat'] = df.up_down.apply(f) df2= (df[(df.year > 1984) & (df.month == 5) & (df.dow == 1)]) print (df2) This is the error Traceback (most recent call last): File "dframt1est.py", line 77, in <module> df['up_down_flat'] = df.up_down.apply(f) File "C:\Users\Zac\AppData\Local\Programs\Python\Python35-32\lib\site-packages\pandas\core\series.py", line 2220, in apply mapped = lib.map_infer(values, f, convert=convert_dtype) File "pandas\src\inference.pyx", line 1088, in pandas.lib.map_infer (pandas\lib.c:63043) TypeError: f() takes 0 positional arguments but 1 was given Press any key to continue . . . This a sample csv Date Open High Low Close Volume Adj Close \ 0 1990-04-12 26.875000 26.875000 26.625 26.625 6100 250.576036 1 1990-04-16 26.500000 26.750000 26.375 26.750 500 251.752449 2 1990-04-17 26.750000 26.875000 26.750 26.875 2300 252.928863 3 1990-04-18 26.875000 26.875000 26.500 26.625 3500 250.576036 4 1990-04-19 26.500000 26.750000 26.500 26.750 700 251.752449 5 1990-04-20 26.750000 26.875000 26.750 26.875 2100 252.928863 6 1990-04-23 26.875000 26.875000 26.750 26.875 700 252.928863 7 1990-04-24 27.000000 27.000000 26.000 26.000 2400 244.693970 8 1990-04-25 25.250000 25.250000 24.875 25.125 9300 236.459076 Answer: You're passing `f` as the function in `apply`. That function is called for each row in the dataframe, and needs to take the row as its parameter. Note also that you're returning the function itself as the result; I'm pretty sure you meant to return `x` not `f`. Also, you don't seem to have defined `Up`, `Down` and `Flat` anywhere.
how can i correct this Regex phone number extractor in python Question: The results i'm getting when i run this after copying a set of UK phone numbers to the clipboard are coming out in a very bizarre kind of way. (i have imported both modules before you ask) phoneRegex = re.compile(r'''( (\d{5}|\(\d{5}\))? #area code (\s|-|\.)? #separator (\d{6}) #main 6 digits )''', re.VERBOSE) emailRegex = re.compile(r'''( [a-zA-Z0-9._%+-]+ #username @ #obligatory @ symbol [a-zA-Z0-9.-]+ #domain name (\.[a-zA-Z]{2,5}) #dot-something )''', re.VERBOSE) text = str(pyperclip.paste()) matches = [] for groups in phoneRegex.findall(text): phoneNum = '-'.join([groups[1], groups[3]]) if groups[3] != '': phoneNum += ' ' + groups[3] matches.append(phoneNum) for groups in emailRegex.findall(text): matches.append(groups[0]) if len(matches) > 0: pyperclip.copy('\n'.join(matches)) print('Copied to clipboard:') print('\n'.join(matches)) else: print('No phone numbers or email addresses found') The mistake is somewhere here: for groups in phoneRegex.findall(text): phoneNum = '-'.join([groups[1], groups[3]]) if groups[3] != '': phoneNum += ' ' + groups[3] matches.append(phoneNum) **The numbers copied to clipboard:** 07338 754433 01265768899 (01283)657899 **Expected results:** Copied to clipboard: 07338 754433 01265 768899 01283 657899 **return results:** Copied to clipboard: 07338-754433 754433 -012657 012657 (01283)-657899 657899 Answer: I see three issues: 1. the python code joins the two parts of the phone number together with a `-` and then adds a space and the third part again: phoneNum = '-'.join([groups[1], groups[3]]) if groups[3] != '': phoneNum += ' ' + groups[3] Since `groups[3]` will always not be blank, what you need to do is: if groups[1] != '': phoneNum = ' '.join(groups[1], groups[3]) else: phoneNum = groups[3] 2. Your `phoneRegex` regular expression is not anchored to the beginning and end of the lines. You need to (a) compile it with the `re.MULTILINE` option and (b) anchor the regular expression between `^` and `$`: phoneRegex = re.compile(r'''^( (\d{5}|\(\d{5}\))? #area code (\s|-|\.)? #separator (\d{6}) #main 6 digits )$''', re.VERBOSE + re.MULTILINE) This will prevent a long string of digits with no separator as being just group 3 with a bunch of digits after it. 3. Your match for the area code includes the matched parentheses within the group match. To fix this, you either need to change the regular expression to make sure the parentheses are not part of the group, or you need to change your code to strip the parentheses out if needed. * Avoid parentheses in the regular expression: (?:(\d{5})|\((\d{5})\))? #area code The `(?:...)` is a non-grouping form of parentheses, so it won't be returned by the find. Within that, you have two alternatives: 5 digits in a group - `(\d{5})` \- or literal parentheses that enclose 5 digits in a group - `\((\d{5})\)`. However, this change also affects your phone number recombination logic, because your area code is _either_ in `groups[1]` or `groups[2]`, and your main number is now in `groups[4]`. if groups[1] != '': phoneNum = ' '.join(groups[1], groups[4]) elif groups[2] != '': phoneNum = ' '.join(groups[2], groups[4]) else: phoneNum = groups[4] * This could be made easier by changing the outer set of parentheses and the parentheses around the separator into non-grouping parentheses. You could then do a single join on a filtered result of the groups: phoneRegex = re.compile(r'''(?: (?:(\d{5})|\((\d{5})\))? #area code (?:\s|-|\.)? #separator (\d{6}) #main 6 digits )''', re.VERBOSE) # ... phoneNum = ' '.join([group for group in groups if group != '']) The modified `phoneRegex` ensures that the returned `groups` contain only an optional area code in `groups[0]` or `groups[1]` followed by the main number in `groups[2]`, no extraneous matches returned. The code then filters out any groups that are empty and returns the rest of the groups joined by a space. * Strip parentheses in code: if groups[1] != '': phoneNum = ' '.join(groups[1].lstrip('(').rstrip(')'), groups[3]) else: phoneNum = groups[3]
python os.walk returns nothing Question: I have a problem with using `os.walk` on Mac. If I call it from `python terminal`, it works perfect, but if I call it via a `python script`, it returns empty list. For example: import os path = "/Users/temp/Desktop/test/" for _ ,_ , files in os.walk(path): test = [my_file for my_file in files] print test then, it prints: [] and I am pretty sure that the `path` does exists. Any idea what is the problem? Answer: You most likely need to intantiate the test list outside the for loop, for this to work. import os path = "/Users/temp/Desktop/test/" test = [] for _ ,_ , files in os.walk(path): test.extend([my_file for my_file in files]) print test
The last element (bitstring.BitArray) in list is incorrect after XORing python Question: I have snippet of code: #!/usr/bin/python3 from bitstring import BitArray import itertools # Helper functions def get_bitset_by_letter(letter, encoding): return encoding[letter] if letter in encoding else None def get_letter_by_bitset(bitset, encoding): for letter, encoding_bitset in encoding.items(): if bitset == encoding_bitset: return letter return None class Word(): def __init__(self, word, encoding): self.encoding = encoding self.as_word = '' self.as_bits = [] if isinstance(word, str): self.as_word = word self._encode_to_bits() elif all(isinstance(i, BitArray) for i in word): self.as_bits = word self._decode_to_word() else: raise('class Word: Unexpected type of word arg') def __ixor__(self, other): for index, bitset in enumerate(other.as_bits): print('In XOR: %d-%s-%s' % (index, self.as_bits[index], bitset)) self.as_bits[index] ^= bitset return self def as_bits(self): return self.as_bits def as_word(self): return self.as_word def _encode_to_bits(self): for letter in self.as_word: bitset = get_bitset_by_letter(letter, self.encoding) if bitset is None: raise('Can not find bitset by given letter %s in encoding.' % str(letter)) else: self.as_bits.append(bitset) def _decode_to_word(self): for bitset in self.as_bits: letter = get_letter_by_bitset(bitset, self.encoding) if letter is None: raise('Can not find letter by given bitset %s in encoding.' % str(bitset)) else: self.as_word += letter def main(): encoding = { 'A': BitArray(bin='000'), 'B': BitArray(bin='001'), 'C': BitArray(bin='010'), 'D': BitArray(bin='011'), 'E': BitArray(bin='100'), 'F': BitArray(bin='101'), 'G': BitArray(bin='110'), 'H': BitArray(bin='111'), } print(encoding) print() word_1 = Word('ABCHE', encoding) print('word_1 = %s' % word_1.as_bits) word_2 = Word('FGDEA', encoding) print('word_2 = %s' % word_2.as_bits) word_1 ^= word_2 print(word_1.as_bits) if __name__ == '__main__': main() Details: I use `bitstring.BitArray` class to represent letter as set of three bits and code words this way (some kind of hometask). Also, my code should be able to do XOR for this words. Problem: Every last element for `other.as_bits` list has incorrect value, when it arrives to function. If I comment the line #self.as_bits[index] ^= bitset the last value of the `other.as_bits` is correct. What am I doing wrong? Here is the output (python 3.5.2): {'C': BitArray('0b010'), 'B': BitArray('0b001'), 'G': BitArray('0b110'), 'F': BitArray('0b101'), 'A': BitArray('0b000'), 'H': BitArray('0b111'), 'D': BitArray('0b011'), 'E': BitArray('0b100')} word_1 = [BitArray('0b000'), BitArray('0b001'), BitArray('0b010'), BitArray('0b111'), BitArray('0b100')] word_2 = [BitArray('0b101'), BitArray('0b110'), BitArray('0b011'), BitArray('0b100'), BitArray('0b000')] In XOR: 0-0b000-0b101 In XOR: 1-0b001-0b110 In XOR: 2-0b010-0b011 In XOR: 3-0b111-0b100 In XOR: 4-0b100-0b101 [BitArray('0b101'), BitArray('0b111'), BitArray('0b001'), BitArray('0b011'), BitArray('0b001')] From this you can see, that the last 4th element of `word_2` in `__ixor__` is `0b101`, should be `0b000` Thank you! Answer: I solved this using `__xor__` instead of `__ixor__`.
Transforming string output to JSON Question: I'm getting some data from an external system (Salesforce Marketing Cloud) over API and I'm getting the data back in the format below: Results: [(List){ Client = (ClientID){ ID = 113903 } PartnerKey = None CreatedDate = 2013-07-29 04:43:32.000073 ModifiedDate = 2013-07-29 04:43:32.000073 ID = 1966872 ObjectID = None CustomerKey = "343431CD-031D-43C7-981F-51B778A5A47F" ListName = "PythonSDKList" Category = 578615 Type = "Private" Description = "This list was created with the PythonSDK" ListClassification = "ExactTargetList" }] It's tantalisingly close to JSON, I'd like to format it like a JSON file so that I can import it into our database. Any ideas of how I can do this? Thanks Answer: To convert your array, you could use var jsonString = JSON.stringify(yourArray);
Setting scan coordinates in device options on pyinsane Question: I use Sane's command line utility (`scanimage`) in order to scan films from the transparency unit of my scanner. Here is the command that I have been using with success: scanimage --device-name pixma:04A9190D \ --source 'Transparency Unit' \ --resolution "4800" \ --format "tiff" \ --mode "color" \ -l "80.6" -x "56.2" -t "25.8" -y "219.2" \ > scan.tiff I have decided to move this to Python code, using `pyinsane` in order to enable further integration with my image-processing workflow. This should supposedly give the following in Python code: import pyinsane.abstract as pyinsane device = pyinsane.get_devices()[0] device.options['resolution'].value = 4800 device.options['mode'].value = 'Color' device.options['source'].value = 'Transparency Unit' # Setting coordinates to non-integers fails device.options['tl-y'].value = 25.8 device.options['tl-x'].value = 80.6 device.options['br-y'].value = 219.2 device.options['br-x'].value = 56.2 scan_session = device.scan(multiple=False) try: while True: scan_session.scan.read() except EOFError: pass image = scan_session.images[0] But my first trials have been unsuccessful because I can not figure out how to set the scan coordinates `pyinsane`. As you see, I have found the appropriate options, but I don't know what unit they are in. `scanimage` takes the coordinates in millimetres by default, but `pyinsane` only takes integers. I have tried using pixel coordinates to no avail. I wonder what units the coordinate parameters take, and whether I am using them in the right order. Answer: pyinsane's option descriptions actually say the values are in milimeters: Option: br-x Title: Bottom-right x Desc: Bottom-right x position of scan area. Type: <class 'pyinsane.rawapi.SaneValueType'> : Fixed (2) Unit: <class 'pyinsane.rawapi.SaneUnit'> : Mm (3) Size: 4 Capabilities: <class 'pyinsane.rawapi.SaneCapabilities'> :[ Automatic, Soft_select, Soft_detect,] Constraint type: <class 'pyinsane.rawapi.SaneConstraintType'> : Range (1) Constraint: (0, 14160319, 0) Value: 20 But they are not! I divided the range maximum for `br-x` variable by the width of the scanning area of my scanner, and I got to the number 65536 (which is 2^16). Setting the coordinates to the millimetre value times 65536 works. Maybe these values define the number of steps of the stepper motor? Also not that while scanimage interprets the `-x` and `-y` switches as width and length, and the `-l` and `-t` switches as offset, pyinsane takes bottom- right x (`br-x`), top-left y (`tl-y`), etc.
Encoding issue when writing to CSV file in Python Question: I have some encoding issue while writing an array to CSV. Code: import csv a = [u'eNTfxfwc', 'Pushkar', 'Waghulde', '[email protected]', 'Los Angeles', '', 'UNITED STATES', '2652 Ellendale Pl # 9', '', 'Los Angeles, UNITED STATES', u'90007', u'', u'(213)458-2091', u'None', u'', u'imports/jobvite/eNTfxfwc.txt', '07/17/2012', u'2012-24', u'pWOjJfww', u'Undefined', u'Import', u' ', '', '', u'\u8edf\u9ad4\u5de5\u7a0b\u5e2b (Software Engineer)', '07/17/2012', '', u'Undefined', '', u'None', '', '', u'Undefined', u'Not Considered', '', '07/17/2012', u'import/jobvite/eNTfxfwc_1.txt'] f3 = open('test.csv', 'at') writer = csv.writer(f3,delimiter = ',', lineterminator='\n',quoting=csv.QUOTE_ALL) writer.writerow(a).encode('utf-8') How can I fix this issue and write array into CSV file? Answer: Just encode the elements in a before writing them a = [i.encode('utf-8') for i in a] f3 = open('test.csv', 'at') writer = csv.writer(f3,delimiter = ',',lineterminator='\n',quoting=csv.QUOTE_ALL) writer.writerow(a)
Python Error message when opening .txt file / change in working directory Question: I’m a new python user and have written a python script that prompts for the name of a text file (.txt) to be opened and read by the program. name = raw_input("Enter file:") if len(name) < 1: name = "test.txt" handle = open(name) This was working perfectly fine for me, but then I installed Anaconda and tried to run the same `.py` script and kept getting this error message: $ python /test123.py Enter file:'test.txt' Traceback (most recent call last): File "/test123.py", line 26, in <module> handle = open(name) IOError: [Errno 2] No such file or directory: "'test.txt'" even though I had changed nothing in the script or text file and the text file definitely exists in the directory. So then I uninstalled Anaconda and deleted the Anaconda directory as well as the PATH variable pointing to Anaconda in the `~/.bash_profile` file to see if that would fix the problem. But I’m still getting the same error :( What’s wrong? I do not get any error if I run the file in the Terminal directly from TextWrangler, and my `.py` script contains the shebang line `#!/usr/bin/env python` My info: $ sw_vers ProductName: Mac OS X ProductVersion: 10.11.6 BuildVersion: 15G1004 $ python Python 2.7.12 (v2.7.12:d33e0cf91556, Jun 26 2016, 12:10:39) [GCC 4.2.1 (Apple Inc. build 5666) (dot 3)] on darwin $ which python /Library/Frameworks/Python.framework/Versions/2.7/bin/python $ /usr/bin/python -V Python 2.7.10 Thanks so much for the time and help! **EDIT:** I copied both the `test123.py` document and the `test.txt` file and put them in the directory `/Users/myusername/Documents` and ran the script again, this time inputting `/Users/myusername/Documents/test.txt` (without quotes) at the "Enter file" prompt, and it worked fine! So now I need to modify my question: My understanding--and what had been working fine for me previously--is that when a python script is run, the working directory automatically changes to the directory that the `.py` file is inside. Is this not correct? Or how could I change my settings back to that, since somehow it seemed to have been the case before but now isn't. **EDIT 2:** Now that I've realized the source of my original error was a result of the working directory Python was using, I've been able to solve my problem by adding the following lines to my `test123.py` script: import os abspath = os.path.abspath(__file__) dname = os.path.dirname(abspath) os.chdir(dname) as suggested here: [python: Change the scripts working directory to the script's own directory](http://stackoverflow.com/questions/1432924/python- change-the-scripts-working-directory-to-the-scripts-own-directory) However, I'm also selecting the answer provided by @SomethingSomething because now that I've been able to change the working directory, the input `'test.txt'` (with quotes) is not valid, but `test.txt` (without quotes) works as desired. (The only thing I'm still wondering about is why previously, when I was running python scripts, the working directory automatically changed to the directory that the `.py` file was inside, and then somehow the setting changed, corresponding with my installation of Anaconda. However, now that I've solved my problem this seems unimportant.) Thanks everyone for the help! Answer: It's very simple, your input is wrong: Incorrect: $ python /test123.py Enter file:'test.txt' Correct: $ python /test123.py Enter file:test.txt do not include the `'` characters. It is not code, it's just standard input
Replacing an item in list with items of another list without using dictionaries Question: I am developing a function in python. Here is my content: list = ['cow','orange','mango'] to_replace = 'orange' replace_with = ['banana','cream'] So I want that my list becomes like this after replacement list = ['cow','banana','cream','mango'] I am using this function: def replace_item(list, to_replace, replace_with): for n,i in enumerate(list): if i== to_replace: list[n]=replace_with return list It outputs the list like this: ['cow', ['banana', 'cream'], 'mango'] So how do I modify this function to get the below output? list = ['cow','banana','cream','mango'] NOTE: I found an answer here: [Replacing list item with contents of another list](http://stackoverflow.com/questions/14962931/replacing-list-item-with- contents-of-another-list) but I don't want to involve dictionaries in this. I also want to modify my current function only and keep it simple and straight forward. Answer: This approach is fairly simple and has similar performance to @TadhgMcDonald- Jensen's `iter_replace()` approach (3.6 µs for me): lst = ['cow','orange','mango'] to_replace = 'orange' replace_with = ['banana','cream'] def replace_item(lst, to_replace, replace_with): return sum((replace_with if i==to_replace else [i] for i in lst), []) print replace_item(lst, to_replace, replace_with) # ['cow', 'banana', 'cream', 'mango'] Here's something similar using itertools, but it is slower (5.3 µs): import itertools def replace_item(lst, to_replace, replace_with): return list(itertools.chain.from_iterable( replace_with if i==to_replace else [i] for i in lst )) Or, here's a faster approach using a two-level list comprehension (1.8 µs): def replace_item(lst, to_replace, replace_with): return [j for i in lst for j in (replace_with if i==to_replace else [i])] Or here's a simple, readable version that is fastest of all (1.2 µs): def replace_item(lst, to_replace, replace_with): result = [] for i in lst: if i == to_replace: result.extend(replace_with) else: result.append(i) return result Unlike some answers here, these will do multiple replacements if there are multiple matching items. They are also more efficient than reversing the list or repeatedly inserting values into an existing list (python rewrites the remainder of the list each time you do that, but this only rewrites the list once).
former form fields no longer found by mechanize python script Question: Let me start by apologizing for my utter newbness. I was asked by a friend a couple years ago if I could write a program to automatically grab substitute teaching openings. It wasn't an area I knew anything about, but a couple tutorials allowed me to bang something out despite ignorance about html (and more than a little about Python for that matter). Script worked great since then but this year their site seems to have been redone and broke things, pushing it far beyond my understanding. My previous code that worked: # Create a Browser instance b = mechanize.Browser() # Load the page b.open(loginURL) # Select the form b.select_form(nr=0) # Fill out the form b['id'] = 'XXXXXXXXXX' # Note: I edited out my friend's login info here for privacy b['pin'] = 'XXXX' b.submit(); There is still only one form but the controls are now of type "hidden" and are not the ones I directly need any longer. I can see the old fields in the html when I examine it with developer mode and the names are the same but I can't figure out (tried some things that didn't work) how I would access them now. Here is the html: <form id="loginform" name="loginform" method="post" action="https://www.aesoponline.com/login.asp?x=x&amp;&amp;pswd=&amp;sso="> <input type="hidden" name="location" value=""> <input type="hidden" name="qstring" value=""> <input type="hidden" name="absr_ID" value=""> <input type="hidden" name="foil" value=""> <div style="margin: auto; text-align:center;"> <div id="loginContainer" style="text-align: left;"> <div id="loginContent"> <div id="Div1" style="position:relative; left:65px;" class="hide-me-for-rebranding"> <a href="http://www.frontlinetechnologies.com"> <img src="images/frontlinelogo.png" border="0"> </a> </div> <div id="loginLoginBox" style="position:relative;"> <div id="loginAesopLogo" style="padding-bottom:0px;" class="hide-me-for-rebranding"></div> <!--endloginAesopLogo--> <div id="loginLoginFields" style="margin-top:0px;"> <br> <table> <tbody> <tr height="25px"> <td width="30px"><span class="corrLoginFormText">ID:</span> </td> <td> <input type="text" class="loginFormText" maxlength="80" id="txtLoginID" name="id" value=""> </td> </tr> <tr height="25px"> <td width="30px"><span class="corrLoginFormText">Pin:</span> </td> <td> <input type="password" class="loginFormText" maxlength="20" id="txtPassword" name="pin"> </td> </tr> </tbody> </table> <table> <tbody> <tr height="30px"> <td width="75px" valign="top"> <a class="textButton" id="loginLink" name="loginLink" href="#"><span style="white-space:nowrap;">Login</span></a> <input type="hidden" id="submitLogin" name="submitLogin" value="1"> </td> <td> <div id="loginhelp" style="float:right;"> <img src="images/icon.pinreminder.png" alt="pin" width="10" height="15" align="top"><a href="forgot_pin.asp">Pin Reminder</a> <br> <img src="images/icon.loginproblems.png" alt="login" width="11" height="17" align="top"> <a href="http://help.frontlinek12.com/Employee/Docs/ClientServicesHelpGuide-LoginProblems.pdf">Login Problems</a> </div> </td> </tr> </tbody> </table> </div> <!--endloginLoginFields--> <div id="errorLabel" style="position: absolute; top: 170px; left:5px;margin:0px;"><span class="assistanceText"></span> </div> </div> <!--endloginLoginBox--> <div id="loginContentText"> <span class="loginContentHeader">Welcome To Absence Management</span> <br> <span class="loginContentText"> You are about to enter Frontline Absence Management!<br> Please enter your ID and PIN to login to your account, or click the button below to learn more about Frontline's growing impact on education.</span> <br> <a class="textButton" href="http://www.frontlinek12.com/Products/Aesop.html"><span>Learn More</span></a> </div> <!--endloginContentText--> </div> <!--endLoginContent--> <div id="loginFooterShading" class="hide-me-for-rebranding"> <div id="loginFooterLeft"></div> <div id="loginFooterRight"></div> </div> <!--endloginFooterShading--> <div id="loginFooter" style="text-align:center;width:725px;"> <a href="http://www.frontlinetechnologies.com/Privacy_Policy.html" style="color: rgb(153, 0, 0) ; font-size:9px;" target="_blank">Privacy Policy</a> <br>© Frontline Technologies Group LLC &lt; <parm1>&gt; <br>All rights reserved. Protected under US Patents 6,334,133, 6,675,151, 7,430,519, 7,945,468 and 8,140,366 with additional patents pending. </parm1> </div> <!--endloginReflections--> </div> <!--endLoginContainer--> </div> <!--end margin div --> <!-- MODAL DIALOG --> <div id="basicModalContent" style="display:none"> <span class="assistanceText"></span> </div> </form> Any assistance would be greatly appreciated. Thank you very much. Answer: Try something like this if that html code is exactly what is on that page. When you put b.select_form(nr=0) there is the possibility that for some reason the first form is not what you are selecting. By looking for the form name in the b.select_form() you can ensure you find the correct form. Test it out and see if it works out. import mechanize b = mechanize.Browser() b.open(loginURL) #since the actual form is named loginform just select it b.select_form("loginform") b['id'] name of login input field b['pin'] b.submit()