text
stringlengths 226
34.5k
|
---|
How to retrieve nested map values
Question: I want to scan AWS DynamoDB table and then pull only a certain value. Here is
my code:
package main
import (
"fmt"
"github.com/aws/aws-sdk-go/aws"
"github.com/aws/aws-sdk-go/aws/session"
"github.com/aws/aws-sdk-go/service/dynamodb"
)
func main() {
svc := dynamodb.New(session.New(), &aws.Config{Region: aws.String("us-west-2")})
params := &dynamodb.ScanInput{
TableName: aws.String("my_Dynamo_table_name"),
Limit: aws.Int64(2),
}
resp, err := svc.Scan(params)
if err != nil {
fmt.Println(err.Error())
return
}
fmt.Println(resp)
}
and The output is:
{
Count: 2,
Items: [{
update_time: {
N: "1466495096"
},
create_time: {
N: "1465655549"
}
},{
update_time: {
N: "1466503947"
},
create_time: {
N: "1466503947"
}
}],
LastEvaluatedKey: {
Prim_key: {
S: "1234567890"
}
},
ScannedCount: 2
}
Now, I want to retrieve the `update_time` value for all elements in above
output. Here are my attempts:
for _, value := range resp.Items {
fmt.Println(value["create_time"]["N"])
}
and
for _, value := range resp.Items {
fmt.Println(value.create_time.N)
}
and
for _, value := range resp.Items {
fmt.Println(*value.create_time.N)
}
All above attempts error out with `/var/tmp/dynamo.go:37: invalid operation:`
error. I am from perl/python background and recently started learning golang.
How to retrieve nested map/array values in this case. Also, any reading
references would be of great help. My google search did not reveal anything
relevant.
Answer: The value of `resp` above is of the type `*ScanOutput`, which has `Items` type
as `[]map[string]*AttributeValue`.
To access `update_time`, you can try:
updateTimes := make([]string, 0)
// Items is a slice of map of type map[string]*AttributeValue
for _, m := range resp.Items {
// m is of type map[string]*AttributeValue
timeStrPtr := *m["update_time"].N
updateTimes = append(updateTimes, *timeStrPtr)
}
`updateTimes` should now contains all the `"update_time"` values as strings.
More details [here](https://godoc.org/github.com/aws/aws-sdk-
go/service/dynamodb#ScanOutput).
|
Grouping up elements from a list in Python 2.7
Question: Ok, I got a huge text. I extract matches with regex (omitted here because it
doesn't matter and I'm bad at this so you don't see how ugly my regex is :) )
and count them. Then, for readability, I split the elements and print them in
the fashion I need:
import re
f = re.findall(r"(...)", PF)
a = [[y,f.count(y)] for y in set(f)]
(' '.join(map(str, j)) for j in w)
for element in w:
print element
result is something like
['202', 1]
['213', 2]
['210', 2]
['211', 2]
['208', 2]
['304', 1]
['107', 2]
['133', 1]
['132', 1]
['131', 2]
What I need is to group up the elements, so that I get an output like
A ['133', 1]
['132', 1]
['131', 2]
B ['202', 1]
['213', 2]
C ['304', 1]
['107', 2]
['210', 2]
['211', 2]
['208', 2]
Note that:
* in the final result I will need 5 groups (A, B, C, D, E)
* the elements can vary, for example tomorrow 131 might not be present but I might have 232 that goes in group A and the number of elements is different every day
* it would be perfect, but not mandatory, if the elements in each group would be sorted numerically.
* Might sound obvious but I'll make it clear anyway, I know exactly which elements need to go in which group. If it is of any help, group _A can contain (102, 103), B(104,105,106,201,202,203), C(204,205,206,301,302,303,304), D(107,108,109,110,208,209,210,211,213,305,306,307), E(131,132,133,231,232)_.
The script needs to take the results that are present that day, compare them
to the list above, and sort into the relative groups.
Thanks in advance!
Answer: You can set up a hash that maps elements to groups. You can then transform
each array item from [element,count] to (group,element,count) _(using a tuple
to make it more easily sortable and such)_. Sort that array, then use a loop
or `reduce` to transform that into your final output.
mapElementsToGroups = {'131': 'A', '202': 'B', '304': 'C', …}
elementsFoundByGroup = {}
for (group, element, count) in sorted(
[(mapElementsToGroups[item[0]], item[1], item[2])
for item in a]
):
elementsFoundByGroup[group] = elementsFoundByGroup.get(group, []) + [(element, count)]
You now have a dictionary mapping each group name found to the list of
elements and counts found within that group. The quick print is:
print [
group + " " +
elements.join("\n " + " "*len(group))
for (group,elements) in sorted(elementsFoundByGroup.items())
].join("\n")
|
Python Simon Game: I Got Stuck Finishing the Sequence
Question: I'm finally finishing my Simon Game, but I have some doubts about how to
complete the sequence.
Edit: As you asked I have edited my post. Here I will post my code as it was
before this post. So here are my actual problems.
1) I don't know how to add one number to the sequence after each turn.
2) The first turn works fine, but I don't know how to start the next turn,
I've tried to call again the function after checking the sequence but it
doesn't worked.
The necessary code for the two things:
from Tkinter import *
import Tkinter
import random
base = Tkinter.Tk()
fr = Tkinter.Frame(base, bg='black', width='238', height='250')
score = Tkinter.Label(base, bg='black', fg='white', text='Score:')
score.place(x = 30, y = 15)
s = 0
scoreNum = Tkinter.Label(base, bg='black', fg='white', text = s)
scoreNum.place(x = 70, y = 15)
clicks = []
color = 0
def yellowClick():
yellow.configure(activebackground='yellow3')
yellow.after(500, lambda: yellow.configure(activebackground='yellow'))
global clicks
global color
color = 1
clicks.append(color)
yellow = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='yellow',
bg='yellow3', command = yellowClick)
yellow.place(x = 30, y = 50)
def blueClick():
blue.configure(activebackground='medium blue')
blue.after(500, lambda: blue.configure(activebackground='blue'))
global clicks
global color
color = 2
clicks.append(color)
blue = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='blue',
bg='medium blue', command = blueClick)
blue.place(x = 125, y = 50)
def redClick():
red.configure(activebackground='red3')
red.after(500, lambda: red.configure(activebackground='red'))
global clicks
global color
color = 3
clicks.append(color)
red = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='red',
bg = 'red3', command = redClick)
red.place(x = 30, y = 145)
def greenClick():
green.configure(activebackground='dark green')
green.after(500, lambda: green.configure(activebackground='green4'))
global clicks
global color
color = 4
clicks.append(color)
green = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='green4',
bg='dark green', command = greenClick)
green.place(x = 125, y = 145)
def scoreUp():
global s
s = s + 1
scoreNum.configure(text = s)
sequence = []
def checkSequence():
global clicks
global sequence
if clicks == sequence:
scoreUp()
def showSequence():
global sequence
global clicks
global check
r = random.randint(1, 4)
if r == 1:
yellow.configure(bg='yellow')
yellow.after(1000, lambda: yellow.configure(bg='yellow3'))
sequence.append(r)
base.after(5000, checkSequence)
elif r == 2:
blue.configure(bg='blue')
blue.after(1000, lambda: blue.configure(bg='medium blue'))
sequence.append(r)
base.after(5000, checkSequence)
elif r == 3:
red.configure(bg='red')
red.after(1000, lambda: red.configure(bg='red3'))
sequence.append(r)
base.after(5000, checkSequence)
elif r == 4:
green.configure(bg='green4')
green.after(1000, lambda: green.configure(bg='dark green'))
sequence.append(r)
base.after(5000, checkSequence)
base.after(2000, showSequence)
fr.pack()
base.resizable(False, False)
base.mainloop()
The checkSequence function is activated by time because I could not find
another way to do it.
Answer: I've answered this question already but you still had some problems, here is
the full and complete code. I've made it more efficient and fixed the problems
:)
This is working I've tested it. Just comment if there is anything you need to
know
import Tkinter # you don't need to import Tkinter again, from ... import * moves all the functions from ... into your program
import random
base = Tkinter.Tk()
fr = Tkinter.Frame(base, bg='black', width='238', height='250')
fr.pack()
score = Tkinter.Label(base, bg='black', fg='white', text='Score:')
score.place(x = 30, y = 15)
scoreNum = Tkinter.Label(base, bg='black', fg='white', text = 0)
scoreNum.place(x = 70, y = 15)
global sequence
global clicks
global s
sequence = []
clicks = []
s = 0
yellow = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='yellow',
bg='yellow3', command = lambda *args: Click(yellow, "yellow", "yellow3", 1))
blue = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='blue',
bg='medium blue', command = lambda *args: Click(blue, "blue", "medium blue", 2))
red = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='red',
bg = 'red3', command = lambda *args: Click(red, "red", "red3", 3))
green = Tkinter.Button(base, bd='0', highlightthickness='0',
width='7', height='5', activebackground='green4',
bg='dark green', command = lambda *args: Click(green, "green", "dark green", 4))
yellow.place(x = 30, y = 50)
blue.place(x = 125, y = 50)
red.place(x = 30, y = 145)
green.place(x = 125, y = 145)
def Click(button, colour1, colour2, number): # these arguments change so that you don't have to copy the function 10 times
global clicks
yellow.configure(activebackground=colour2)
yellow.after(500, lambda: yellow.configure(activebackground=colour1))
clicks.append(number)
checkSequence() # why not do this straight away?
def checkSequence():
global clicks
global sequence
global s
print("clicks: "+str(clicks)) # debug
print("sequence: "+str(sequence))
if clicks != sequence[:len(clicks)]: # if all their clicks so far e.g. [1, 2, 4, 3] DONT fit with the sequence e.g. [3, 2, 4, 3, 2, 1]
print(" ...Incorrect!")
s = 0
scoreNum.configure(text = 0)
sequence = []
clicks = []
base.after(1000, showSequence) # now move to the next value in the sequence
elif clicks == sequence: # they have completed a sequence
print(" ...Match!")
s = s + 1
scoreNum.configure(text = s)
clicks = []
base.after(1000, showSequence) # now move to the next value in the sequence
def showSequence():
global sequence
global clicks
r = random.randint(1, 4)
if r == 1:
yellow.configure(bg='yellow')
yellow.after(1000, lambda: yellow.configure(bg='yellow3'))
elif r == 2:
blue.configure(bg='blue')
blue.after(1000, lambda: blue.configure(bg='medium blue'))
elif r == 3:
red.configure(bg='red')
red.after(1000, lambda: red.configure(bg='red3'))
elif r == 4:
green.configure(bg='green')
green.after(1000, lambda: green.configure(bg='dark green'))
sequence.append(r) # you don't need this in all the if statements, it always happens
base.after(100, showSequence)
base.resizable(False, False)
base.mainloop()
NOTE: I'm using python 3.4.1 so I changed all the tkinters to Tkinter
|
ValueError: script argument must be unicode. - Python3.5, SQLite3
Question: I'm writing following python script:
import sqlite3
import sys
if len(sys.argv) < 2:
print("Error: You must supply at least SQL script.")
print("Usage: %s table.db ./sql-dump.sql" % (sys.argv[0]))
sys.exit(1)
script_path = sys.argv[1]
if len(sys.argv) == 3:
db = sys.argv[2]
else:
db = ":memory:"
try:
con = sqlite3.connect(db)
with con:
cur = con.cursor()
with open(script_path,'rb') as f:
cur.executescript(f.read())
except sqlite3.Error as err:
print("Error occured: %s" % err)
I saved this program as **sqlite_import.py**. I have a database file named
**test.db** and a SQL file **world.sql**. Now I tried to run program as:
sqlite_import.py test.db world.sql
But it shows me error like following:
Traceback (most recent call last):
File "C:\Users\Jarvis\OneDrive\Documents\Python\Python Data Visualization Cookbook, 2E\2 Knowing Your Data\sqlite_import.py", line 21, in <module>
cur.executescript(f.read())
ValueError: script argument must be unicode.
Help me to fix this.
Answer: You opened the script file as _binary_ :
with open(script_path,'rb') as f:
This produces a `b'...'` bytes value, not a Unicode `str` object. Remove the
`b`, and perhaps add an `encoding` argument to specific what codec to use to
decode the data with.
|
Python: Properly formatting JSON parameters into a proper http request with markit on demand api
Question: I am having trouble formatting my request properly in order to use the
markitondemand InteractiveChart API. How can I properly do this?
reference: <http://dev.markitondemand.com/MODApis/>
Here is an example of a proper request:
http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json?parameters=%7B%22Normalized%22%3Afalse%2C%22NumberOfDays%22%3A10%2C%22DataPeriod%22%3A%22Day%22%2C%22Elements%22%3A%5B%7B%22Symbol%22%3A%22AAPL%22%2C%22Type%22%3A%22price%22%2C%22Params%22%3A%5B%22c%22%5D%7D%5D%7D
Here is my code constructing a request:
import requests
import json
url2 = 'http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json'
elements = [
{
'Symbol': 'GOOG',
'Type': 'price',
'Params': {'price': ['c']},
}
]
req_obj = {
'Normalized': 'false',
'NumberOfDays': 3,
'DataPeriod': 'Day',
'Elements': elements
}
resp = requests.get(url2, params={'parameters': json.dumps(req_obj)})
and here is the output when I log the resp.url :
http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart/json?parameters=%7B%22Elements%22%3A+%5B%7B%22Type%22%3A+%22price%22%2C+%22Params%22%3A+%7B%22price%22%3A+%5B%22c%22%5D%7D%2C+%22Symbol%22%3A+%22GOOG%22%7D%5D%2C+%22NumberOfDays%22%3A+3%2C+%22DataPeriod%22%3A+%22Day%22%2C+%22Normalized%22%3A+%22false%22%7D
What am I doing wrong here?
Answer: Your url end point (by the looks of it) should just be:
url2 = 'http://dev.markitondemand.com/MODApis/Api/v2/InteractiveChart
Then, make your request:
resp = requests.get(url2, params=json.dumps({'parameters': req_obj}))
Then `requests` will properly handle the `?` and any `&` if they were
required.
|
UnicodeEncodeError Only When Script is Run as a Subprocess
Question: I'm running my main script in Python 3.5 using the Spyder IDE, and I want to
import functions from a script that happens to only work in Python 3.4. So I
was recommended to run this second script as a subprocess like so:
import subprocess
cmd = [r'c:\python34\pythonw.exe', r'C:\users\John\Desktop\scraper.py']
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
stdout, stderr = p.communicate()
print(stdout)
print(stderr)
The script being called is an example from NikolaiT's Web engine scraper:
# -*- coding: utf-8 -*-
import sys
from GoogleScraper import scrape_with_config, GoogleSearchError
from GoogleScraper.database import ScraperSearch, SERP, Link
def basic_usage():
# See in the config.cfg file for possible values
config = {
'SCRAPING': {
'use_own_ip': 'True',
'search_engines': 'baidu',
'num_pages_for_keyword': 3
},
'keyword': '苹果',
'SELENIUM': {
'sel_browser': 'chrome',
},
'GLOBAL': {
'do_caching': 'False'
}
}
try:
sqlalchemy_session = scrape_with_config(config)
except GoogleSearchError as e:
print(e)
# let's inspect what we got
link_list = []
for serp in sqlalchemy_session.serps:
#print(serp)
for link in serp.links:
#print(link)
link_list.append(link.link)
return link_list
links = basic_usage()
print("test")
for link in links:
print(link)
This script works just fine when running it in Python 3.4's IDLE IDE, but when
running it as a subprocess as above, I get the following UnicodeEncodeError
printed from my main script:
\python34\scripts\googlescraper\GoogleScraper\caching.py", line 413, in
parse_all_cached_files store_serp_result(serp, self.config)
File "c:\python34\scripts\googlescraper\GoogleScraper\output_converter.py",
line 123, in store_serp_result pprint.pprint(data)
File "c:\python34\lib\pprint.py", line 52, in pprint printer.pprint(object)
File "c:\python34\lib\pprint.py", line 139, in pprint self._format(object,
self._stream, 0, 0, {}, 0)
File "c:\python34\lib\pprint.py", line 193, in _format allowance + 1, context,
level)
File "c:\python34\lib\pprint.py", line 268, in _format write(rep)
File "c:\python34\lib\encodings\cp1252.py", line 19, in encode return
codecs.charmap_encode(input,self.errors,encoding_table)[0] UnicodeEncodeError:
\'charmap\' codec can\'t encode characters in position 1-2: character maps to
undefined'
Why would this only happen when running it indirectly? Thanks for any help,
clarification questions, or suggestions for improvement in framing my
question.
Answer: In short, your issue is as follows:
* Python IDE fully supports UTF-8 encoding, and that's why your invoked script runs fine there.
* On the other hand, when opening a subprocess, since you are on Windows, you are using Windows1252 charset by default, which doesn't support part of your output.
Quick solution: If you invoke the script as `python foobar.py` on Windows
command line, you can run `chcp 65001` before invoking the script; if you use
Spyder IDE, there should be a setting that allows you to set file / project
encoding to UTF-8.
(You may also try adding `# -*- coding: utf-8 -*-` to the top of your main
Python script.)
|
File upload Selenium Web driver python in linux machine calling a remote machine
Question: Hi I have scenario that needs to upload a file in a webpage. Actually I know
that `selenium` will not support file upload scenario. But this can be done in
python with external libraries such as `AUTOIT`, `PYWINAUTO`. But the
challenge is i have to run my code in a linux server that is going to call a
windows remote machine.
When i tried installing `pywinauto` in linux server i got an error in
importing winreg library. Hence i dont know how to proceed further. Please
help me out to solve this scenario.
Answer: Both AutoIt and pywinauto are Windows-only libraries (at least for now). If
you need to automate file upload on Linux, consider using [AT-SPI
accessibility](https://wiki.gnome.org/Accessibility) (say `pyatspi2` package).
If it's a server without X and DBus, I think the question is about remote code
execution from Linux to Windows. Good option for the SSH remote execution is
[Fabric](http://www.fabfile.org/) (very pythonic & nice), but using Cygwin or
OpenSSH might be an additional challenge for you. There are many other tools
like Ansible etc.
|
Raspberry Pi getting data from MultiWii
Question: I have a Raspberry Pi 3B and a CRIUS All in One Pro (v2.0) MultiWii flight
controller. I'm using the MultiWii 2.4 version and the latest NOOBS. I was
able to set up both fine, and now I am trying to get the Raspberry Pi to
communicate with the MultiWii through a USB/Micro USB cable that connects the
two boards.
At the moment, the MultiWii isn't returning any data back, and I'm not sure
why. From what I can see, I have the protocol correct. I've looked at several
working code repos (written in Python of Java for Arduino), and have followed
the [MultiWii
documentation](http://www.multiwii.com/wiki/index.php?title=Multiwii_Serial_Protocol)
and read through the associated [forum
post](http://www.multiwii.com/forum/viewtopic.php?f=8&t=1516).
Here is the client code that I wrote.
package com.jmace.MaceDrone.msp;
import com.pi4j.io.serial.Baud;
import com.pi4j.io.serial.DataBits;
import com.pi4j.io.serial.FlowControl;
import com.pi4j.io.serial.Parity;
import com.pi4j.io.serial.Serial;
import com.pi4j.io.serial.SerialConfig;
import com.pi4j.io.serial.SerialDataEvent;
import com.pi4j.io.serial.SerialDataEventListener;
import com.pi4j.io.serial.SerialFactory;
import com.pi4j.io.serial.StopBits;
import java.io.IOException;
import java.math.BigInteger;
public class MultiWiiClient {
private final Serial serial;
//The preamble is defined by the protocol.
//Every message must begin with the characters $M
private static final String PREAMBLE = "$M";
//Character that denotes information being passed to the MultiWii
private static final char TO_MUTLIWII = '<';
//Character that denotes information being requested from by the MultiWii
private static final char FROM_MUTLIWII = '>';
public MultiWiiClient(String usbPort) {
SerialConfig config = new SerialConfig();
config.device(usbPort)
.baud(Baud._115200)
.dataBits(DataBits._8)
.parity(Parity.NONE)
.stopBits(StopBits._1)
.flowControl(FlowControl.NONE);
this.serial = SerialFactory.createInstance();
serial.addListener(new SerialDataEventListener() {
@Override
public void dataReceived(SerialDataEvent event) {
try {
System.out.println("[HEX DATA] " + event.getHexByteString());
System.out.println("[ASCII DATA] " + event.getAsciiString());
} catch (IOException e) {
e.printStackTrace();
}
}
});
try {
this.serial.open(config);
} catch (Exception e) {
e.printStackTrace();
}
}
public String sendRequest(MultiWiiRequest request) throws IllegalStateException, IOException {
String message = createMessage(request.getId(), false, null);
//////////////////////////////////////////////////////////////////////////////////
System.out.println(message);
System.out.println(String.format("%040x", new BigInteger(1, message.getBytes())));
//////////////////////////////////////////////////////////////////////////////////
return sendMessage(message);
}
public String sendCommand(MultiWiiCommand command, String payload) throws IllegalStateException, IOException {
String message = createMessage(command.getId(), true, payload);
return sendMessage(message);
}
/**
* This method creates the message that will be sent to the MultiWii
*
* Message format is as follows:
* +--------+---------+----+-------+----+---+
* |preamble|direction|size|command|data|crc|
* +--------+---------+----+-------+----+---+
*
* Preamble (2 bytes):
* Marks the start of a new message; always "$M"
*
* Direction (1 byte):
* Either '<' for a command going to the MultiWii or '>' for
* information being requested from the MultiWii
*
* Size (1 byte):
* The number of bytes in the payload
*
* Command (1 byte):
* The message ID of the command, as defined in the protocol
* 100's for requesting data, and 200's for requesting an action
*
* Data (variable bytes):
* The data to pass along with the command
*
* CRC (1 byte):
* Calculated with an XOR of the size, command, and each byte of data
*/
private String createMessage(int mutliWiiCommandnumber, boolean isCommand, String payload) {
StringBuilder message = new StringBuilder(PREAMBLE);
byte checksum=0;
//Get the direction of the message
if (isCommand) {
message.append(TO_MUTLIWII);
} else {
message.append(FROM_MUTLIWII);
}
int datalength = (payload != null) ? payload.length() : 0;
message.append((char) datalength);
checksum ^= datalength;
message.append((char) mutliWiiCommandnumber);
checksum ^= ((int) mutliWiiCommandnumber);
if (payload != null) {
for (char c : payload.toCharArray()){
message.append(c);
checksum ^= (int) c;
}
}
message.append((char) checksum);
return message.toString();
}
private String sendMessage(String message) throws IllegalStateException, IOException {
serial.write(message.getBytes());
serial.flush();
System.out.println("TESTING ------------------");
return "";
}
}
I am using "/dev/ttyUSB0" to connect, which I confirmed is the correct
location and it seems to be working (no errors when I run it; and if I start
running and then disconnect the USB, it throws an exception because it lost
the connection).
When running, I get the following output (sending command 100, MSP_IDENT):
$M>dd
00000000000000000000000000244d3e006464
TESTING ------------------
See [my Git
repo](https://github.com/jmace01/MaceDrone/tree/master/src/com/jmace/MaceDrone/msp)
for more code context.
EDIT: Fixed the checksum code in my post
Answer: It looks like it's all as it should be, though I can't see where you're
receiving the version byte back from the serial port. in sendMessage, you have
`return ""`.
Looking at your github repo, I see that sendMessage is a bit different.
while (!serial.getCTS()) {
try{
Thread.sleep(100);
} catch (Exception e) {}
}
StringBuilder response = new StringBuilder();
while (serial.available() != 0) {
response.append(serial.read());
}
serial.getCTS() shouldn't be used as you disabled flow control when you
initialized the port in the constructor.
I'm guessing that CTS is here used as a way to detect when the other device
has finished processing the message and has responded.
It seems there's no guarantee that after seeing the CTS line go high, that
those bytes are actually available to be received; the CTS signal would only
mean that the device has finished sending them away to you. You may wish to
put another delay in after CTS goes high to ensure all bytes have indeed been
transmitted.
|
Python - Unable to detect face and eye?
Question: I am trying to create a face and eye detection using OpenCV library. This is
the code I been working with. It's running smoothly with no errors but the
only problem is not showing any results no faces and eyes are found with this
code
import cv2
import sys
import numpy as np
import os
# Get user supplied values
imagePath = sys.argv[1]
# Create the haar cascade
faceCascade = cv2.CascadeClassifier('C:\Users\Karthik\Downloads\Programs\opencv\sources\data\haarcascades\haarcascad_frontalface_default.xml')
eyeCascade= cv2.CascadeClassifier('C:\Users\Karthik\Downloads\Programs\opencv\sources\data\haarcascades\haarcascade_eye.xml')
# Read the image
image = cv2.imread(imagePath)
gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
# Detect faces in the image
faces = faceCascade.detectMultiScale(
gray,
scaleFactor=1.2,
minNeighbors=5,
minSize=(30, 30),
flags = cv2.cv.CV_HAAR_SCALE_IMAGE
)
print "Found {0} faces!".format(len(faces))
# Draw a rectangle around the faces
for (x, y, w, h) in faces:
cv2.rectangle(image, (x, y), (x+w, y+h), (255, 0, 0), 2)
roi_gray = gray[y:y+h, x:x+w]
roi_color = image[y:y+h, x:x+w]
eyes = eyeCascade.detectMultiscale(roi_gray)
for (ex,ey,ew,eh) in eyes:
cv2.rectangle(roi_color,(ex,ey),(ex+ew,ey+eh),(0, 255, 0), 2)
cv2.imshow("Faces found", image)
print image.shape
cv2.waitKey(0)
Answer: For me it works in my jupyter notebook on Ubuntu 15.10 using OpenCV 3.1.0-dev
with python 3.4
Could it be, that you have a simple typo?
`haarcascad_frontalface_default.xml` => `haarcascade_frontalface_default.xml`
and here:
`eyes = eyeCascade.detectMultiscale(roi_gray)` =>
`eyeCascade.detectMultiScale(roi_gray)`
That's my working code:
%matplotlib inline
import matplotlib
import matplotlib.pyplot as plt
import cv2
import sys
import numpy as np
import os
# Create the haar cascade
faceCascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml')
eyeCascade= cv2.CascadeClassifier('haarcascade_eye.xml')
# Read the image
image = cv2.imread('lena.png', 0)
if image is None:
raise ValueError('Image not found')
# Detect faces in the image
faces = faceCascade.detectMultiScale(image)
print('Found {} faces!'.format(len(faces)))
# Draw a rectangle around the faces
for (x, y, w, h) in faces:
cv2.rectangle(image, (x, y), (x+w, y+h), 255, 2)
roi = image[y:y+h, x:x+w]
eyes = eyeCascade.detectMultiScale(roi)
for (ex,ey,ew,eh) in eyes:
cv2.rectangle(roi,(ex,ey),(ex+ew,ey+eh), 255, 2)
plt.figure()
plt.imshow(image, cmap='gray')
plt.show()
[](http://i.stack.imgur.com/EzAjT.png)
|
ImportError: No module named 'theano'
Question: I have installed (/library/python/2.7/site-packages) theano on my mac and
still get this error.
**My code is**
import theano
theano.test()
**and the error**
Traceback (most recent call last):
File "/Users/mac/Downloads/n.py", line 1, in <module>
import theano
ImportError: No module named 'theano'
I did install theano using pip install theano for python 2.7.
Any idea on why its throwing up the error.
Also I installed Anaconda and it throws the same error as working on idle.
Answer: I had the same problem trying to install pygame, i'm using python 3.4.1 so I'm
not 100% sure this will work for you
Here is my folder: C:\Python34\Scripts[](http://i.stack.imgur.com/xoYJZ.png)
**The problem:** This will of course be different for you, but this is what I
did anyway: I did `pip install pygame` and it seemed to work but when I tried
`import pygame` I got the same error as you
**My solution:** instead of installing pygame with pip, I did `pip3 install
pygame` and it worked! so for you, look at what files you have in this folder
**/library/python/2.7/site-packages** and try all of them, try `pip2 install
theano`, try `pip2.7 install theano`, try `pip3 install theano`, try
everything!
Again I'm not certain this will work for python 2.7, but its worth a try :)
|
TypeError: 'module' object is not callable in Spacy Python
Question: I want to print `Parse Tree` using `Spacy`. But the code below is giving the
error
> en_nlp = spacy.language('English') TypeError: 'module' object is not
> callable
The error is on this `en_nlp = spacy.loads('en')` line. I tried to shake off
as `en_nlp = spacy.language(English)` by importing `from spacy.en import
English` But still it is not working. Can someone help?
Code:
import spacy
from nltk import Tree
en_nlp = spacy.loads('en')
doc = en_nlp("The quick brown fox jumps over the lazy dog.")
def to_nltk_tree(node):
if node.n_lefts + node.n_rights > 0:
return Tree(node.orth_, [to_nltk_tree(child) for child in node.children])
else:
return node.orth_
[to_nltk_tree(sent.root).pretty_print() for sent in doc.sents]
Answer:
Is it spacy.load('en') or spacy.loads('en') ?
The official doc <https://spacy.io/docs/> says : spacy.load('en'). It may be
the problem.
|
Bokeh plots not showing in nbviewer
Question: I am working on some visualizations using Bokeh in a Jupyter (ipython)
notebook. Though the plots run well within my notebook, it is important for me
to make them accessible for users not running the code. I was counting on
nbviewer for this, but am having trouble.
Using a simple plot for example, I get this great output in my running
notebook:
[Running notebook in chrome](http://i.stack.imgur.com/dyJwF.png)
After pushing to github I view the notebook in nbviewer but my plots are gone:
<http://nbviewer.jupyter.org/github/kaplann/bokeh_trouble/blob/master/.ipynb_checkpoints/bokeh%20error-
checkpoint.ipynb>
I am using chrome Version 52.0.2743.82 (64-bit) on fedora 23, Bokeh - 0.12.1,
Python 3.5.1, Anaconda 2.5.0 (64-bit), IPython - 5.1.0
I have a very strong feeling I am missing something, hopefully I am and the
fix will be simple. Thanks!
Answer: The notebook is an extremely difficult and challenging environment to develop
and test Bokeh for. There have been some recent regressions and work to fix
them, please refer to GitHub issues
[#5014](https://github.com/bokeh/bokeh/issues/5014) and
[#5081](https://github.com/bokeh/bokeh/issues/5081) for more information.
|
is "from flask import request" identical to "import requests"?
Question: In other words, is the flask request class identical to the requests library?
I consulted:
<http://flask.pocoo.org/docs/0.11/api/>
<http://docs.python-requests.org/en/master/>
but cannot tell for sure. I see code examples where people seem to use them
interchangeably.
Answer: No these are not only completely different libraries, but completely different
purposes.
Flask is a web framework which clients make requests to. The Flask `request`
object contains the data that the client (eg a browser) has sent to your app -
ie the URL parameters, any POST data, etc.
The requests library is for your app to make HTTP request to _other_ sites,
usually APIs. It makes an outgoing request and returns the response from the
external site.
|
Update a label in tkinter from a button press
Question: My question is regarding GUI programming in python by using tkinter. I believe
this is Python 3x.
My question: While we're executing a program to run the GUI, can a button
update a label? More specifically, is there a way to change the labels
displayed text after pushing the button? I have consulted stack overflow on
this before and adopted the StringVar() method, but it doesn't seem to fix my
problem, in fact it omits the text from the GUI completely!
Here is the code below
from tkinter import *
root = Tk()
root.title('Copy Text GUI Program')
copiedtext = StringVar()
copiedtext.set("Text is displayed here")
def copytext():
copiedtext.set(textentered.get())
# Write 'Enter Text Here'
entertextLabel = Label(root, text="Enter Text Here")
entertextLabel.grid(row=0, column=0)
# For the user to write text into the gui
textentered = Entry(root)
textentered.grid(row=0, column=1)
# The Copy Text Button
copytextButton = Button(root, text="Copy Text")
copytextButton.grid(row=1, columnspan=2)
# Display the copied text
displaytextLabel = Label(root, textvariable=copiedtext)
displaytextLabel.grid(row=2,columnspan=2)
copytextButton.configure(command=copytext())
root.mainloop()
Any help would be appreciated!
Answer: What you have to do is to bind a Button event to the copytextButton object
like so:
copytextButton.bind('<Button-1>', copytext)
This means that a callback function - copytext() will be called when you left-
click the button.
A slight modification is needed in the function itself, since the callback
sends an event argument:
def copytext(event):
copiedtext.set(textentered.get())
Edit: This line is not needed:
copytextButton.configure(command=copytext())
|
How can this tensorflow model converge on a CPU but not on a GPU?
Question: We ran into the strange problem that our relatively simple model converges on
the CPU, but not on the server with GPU. No modifications to the code are done
whatsoever between the two runs. Nor does the code contain any explicit
conditional statements to change the workflow on different architectures.
What could possibly be the reason? How can this tensorflow model converge on a
CPU but not on a GPU? In the likely event that the code is too long for you to
read we are still thankful about general speculations and hints.
#!/usr/bin/python
from __future__ import print_function
import tensorflow as tf
import os
import numpy as np
import input_data # copy from tensorflow/examples/tutorials/mnist/input_data.py
# wget https://raw.githubusercontent.com/tensorflow/tensorflow/master/tensorflow/examples/tutorials/mnist/input_data.py if needed
mnist = input_data.read_data_sets("/tmp/data/", one_hot=True)
force_gpu = False
debug = True # histogram_summary ...
# _cpu='/cpu:0'
default_learning_rate=0.001
tensorboard_logs = '/tmp/tensorboard-logs/'
# $(sleep 5; open http://0.0.0.0:6006) & tensorboard --debug --logdir=/tmp/tensorboard-logs/
class net():
def __init__(self,model,data,name=0,learning_rate=default_learning_rate,batch_size=64):
self.session=sess=session=tf.Session()
self.model=model
self.data=data # assigned to self.x=net.input via train
self.batch_size=batch_size
self.layers=[]
self.last_width=self.input_width(data)
self.learning_rate=learning_rate
self.generate_model(model)
def generate_model(self,model, name=''):
if not model: return self
with tf.name_scope('state'):
self.keep_prob = tf.placeholder(tf.float32) # 1 for testing! else 1 - dropout
self.train_phase = tf.placeholder(tf.bool, name='train_phase')
self.global_step = tf.Variable(0) # dont set, feed or increment global_step, tensorflow will do it automatically
with tf.name_scope('data'):
n_input=28*28
n_classes=10
self.x = x = self.input = tf.placeholder(tf.float32, [None, n_input])
self.last_layer=x
self.y = y = self.target = tf.placeholder(tf.float32, [None, n_classes])
if not force_gpu: tf.image_summary("mnist", tf.reshape(self.x, [-1, 28, 28, 1], "mnist_images"))
with tf.name_scope('model'):
model(self)
if(self.last_width!=n_classes): self.classifier() # 10 classes auto
def input_width(self,data):
return 28*28
def add(self, layer):
self.layers.append(layer)
self.last_layer = layer
self.last_shape = layer.get_shape()
def reshape(self,shape):
self.last_layer = tf.reshape(self.last_layer,shape)
self.last_shape = shape
self.last_width = shape[-1]
def batchnorm(self):
from tensorflow.contrib.layers.python.layers import batch_norm as batch_norm
with tf.name_scope('batchnorm') as scope:
input = self.last_layer
train_op=batch_norm(input, is_training=True, center=False, updates_collections=None, scope=scope)
test_op=batch_norm(input, is_training=False, updates_collections=None, center=False,scope=scope, reuse=True)
self.add(tf.cond(self.train_phase,lambda:train_op,lambda:test_op))
# Fully connected layer
def dense(self, hidden=1024, depth=1, act=tf.nn.tanh, dropout=False, parent=-1): #
if parent==-1: parent=self.last_layer
shape = self.last_layer.get_shape()
if shape and len(shape)>2:
self.last_width= int(shape[1]*shape[2]*shape[3])
print("reshapeing ",shape,"to",self.last_width)
parent = tf.reshape(parent, [-1, self.last_width])
width = hidden
while depth>0:
with tf.name_scope('Dense_{:d}'.format(hidden)) as scope:
print("Dense ", self.last_width, width)
nr = len(self.layers)
# if self.last_width == width:
# M = closest_unitary(np.random.rand(self.last_width, width) / (self.last_width + width))
# weights = tf.Variable(m, name="weights_dense_" + str(nr))
# else:
weights = tf.Variable(tf.random_uniform([self.last_width, width], minval=-1. / width, maxval=1. / width), name="weights_dense")
bias = tf.Variable(tf.random_uniform([width],minval=-1./width,maxval=1./width), name="bias_dense")
dense1 = tf.matmul(parent, weights, name='dense_'+str(nr))+ bias
tf.histogram_summary('dense_'+str(nr),dense1)
tf.histogram_summary('weights_'+str(nr),weights)
tf.histogram_summary('bias_'+str(nr),bias)
tf.histogram_summary('dense_'+str(nr)+'/sparsity', tf.nn.zero_fraction(dense1))
tf.histogram_summary('weights_'+str(nr)+'/sparsity', tf.nn.zero_fraction(weights))
if act: dense1 = act(dense1)
# if norm: dense1 = self.norm(dense1,lsize=1) # SHAPE!
if dropout: dense1 = tf.nn.dropout(dense1, self.keep_prob)
self.layers.append(dense1)
self.last_layer = parent = dense1
self.last_width = width
depth=depth-1
self.last_shape=[-1,width] # dense
# Convolution Layer
def conv(self,shape,act=tf.nn.relu,pool=True,dropout=False,norm=True,name=None): # True why dropout bad in tensorflow??
with tf.name_scope('conv'):
print("input shape ",self.last_shape)
print("conv shape ",shape)
width=shape[-1]
filters=tf.Variable(tf.random_normal(shape))
# filters = tf.Variable(tf.random_uniform(shape, minval=-1. / width, maxval=1. / width), name="filters")
_bias=tf.Variable(tf.random_normal([shape[-1]]))
# # conv1 = conv2d('conv', _X, _weights, _bias)
conv1=tf.nn.bias_add(tf.nn.conv2d(self.last_layer,filter=filters, strides=[1, 1, 1, 1], padding='SAME'), _bias)
if debug: tf.histogram_summary('conv_' + str(len(self.layers)), conv1)
if act: conv1=act(conv1)
if pool: conv1 = tf.nn.max_pool(conv1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
if norm: conv1 = tf.nn.lrn(conv1, depth_radius=4, bias=1.0, alpha=0.001 / 9.0, beta=0.75)
if debug: tf.histogram_summary('norm_' + str(len(self.layers)), conv1)
if dropout: conv1 = tf.nn.dropout(conv1,self.keep_prob)
print("output shape ",conv1.get_shape())
self.add(conv1)
def classifier(self,classes=10): # Define loss and optimizer
with tf.name_scope('prediction'):# prediction
if self.last_width!=classes:
# print("Automatically adding dense prediction")
self.dense(hidden=classes, act= False, dropout = False)
# cross_entropy = -tf.reduce_sum(y_*y)
with tf.name_scope('classifier'):
y_=self.target
manual=False # True
if classes>100:
print("using sampled_softmax_loss")
y=prediction=self.last_layer
self.cost = tf.reduce_mean(tf.nn.sampled_softmax_loss(y, y_)) # for big vocab
elif manual:
# prediction = y =self.last_layer=tf.nn.softmax(self.last_layer)
# self.cost = cross_entropy = -tf.reduce_sum(y_ * tf.log(y+ 1e-10)) # against NaN!
prediction = y = tf.nn.log_softmax(self.last_layer)
self.cost = cross_entropy = -tf.reduce_sum(y_ * y)
else:
y = prediction = self.last_layer
self.cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(y, y_)) # prediction, target
# if not gpu:
tf.scalar_summary('cost', self.cost)
# self.cost = tf.Print(self.cost , [self.cost ], "debug cost : ")
learning_scheme=self.learning_rate
# learning_scheme=tf.train.exponential_decay(self.learning_rate, self.global_step, decay_steps, decay_size)
self.optimizer = tf.train.AdamOptimizer(learning_scheme).minimize(self.cost)
# Evaluate model
correct_pred = tf.equal(tf.argmax(prediction, 1), tf.argmax(self.target, 1))
self.accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32))
if not force_gpu: tf.scalar_summary('accuracy', self.accuracy)
# Launch the graph
def next_batch(self,batch_size=10):
return self.data.train.next_batch(batch_size)
def train(self,steps=-1,dropout=None,display_step=10,test_step=200): #epochs=-1,
steps = 9999999 if steps==-1 else steps
session=self.session
# with tf.device(_cpu):
# import tensorflow.contrib.layers as layers
# t = tf.verify_tensor_all_finite(t, msg)
tf.add_check_numerics_ops()
self.summaries = tf.merge_all_summaries()
self.summary_writer = tf.train.SummaryWriter(tensorboard_logs, session.graph) #
if not dropout:dropout=1. # keep all
x=self.x
y=self.y
keep_prob=self.keep_prob
session.run([tf.initialize_all_variables()])
step = 0 # show first
while step < steps:
# print("step %d \r" % step)# end=' ')
batch_xs, batch_ys = self.next_batch(self.batch_size)
# tf.train.shuffle_batch_join(example_list, batch_size, capacity=min_queue_size + batch_size * 16, min_queue_size)
# Fit training using batch data
feed_dict = {x: batch_xs, y: batch_ys, keep_prob: dropout, self.train_phase: True}
loss,_= session.run([self.cost,self.optimizer], feed_dict=feed_dict)
if step % test_step == 0: self.test(step)
if step % display_step == 0:
# Calculate batch accuracy, loss
feed = {x: batch_xs, y: batch_ys, keep_prob: 1., self.train_phase: False}
acc , summary = session.run([self.accuracy,self.summaries], feed_dict=feed)
# self.summary_writer.add_summary(summary, step) # only test summaries for smoother curve
print("\rStep {:d} Loss= {:.6f} Accuracy= {:.3f}".format(step,loss,acc),end=' ')
if str(loss)=="nan": return print("\nLoss gradiant explosion, exiting!!!") #restore!
step += 1
print("\nOptimization Finished!")
self.test(step,number=10000) # final test
def inputs(self,data):
self.inputs, self.labels = load_data()#...)
def test(self,step,number=400):#256
session=sess=self.session
run_metadata = tf.RunMetadata()
run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE)
# Calculate accuracy for 256 mnist test images
test_labels = self.data.test.labels[:number]
test_images = self.data.test.images[:number]
feed_dict = {self.x: test_images, self.y: test_labels, self.keep_prob: 1., self.train_phase:False}
accuracy,summary= self.session.run([self.accuracy, self.summaries], feed_dict=feed_dict)
# accuracy,summary = session.run([self.accuracy, self.summaries], feed_dict, run_options, run_metadata)
print('\t'*3+"Test Accuracy:",accuracy)
# self.summary_writer.add_run_metadata(run_metadata, 'step #%03d' % step)
self.summary_writer.add_summary(summary,global_step=step)
def dense(net): # best with lr ~0.001
# type: (layer.net) -> None
# net.batchnorm() # start lower, else no effect
# net.dense(400,act=None)# # ~95% we can do better:
net.dense(400, act=tf.nn.tanh)# 0.996 YAY only 0.985 on full set, Step 5000 flat
return # 0.957% without any model!!
def alex(net):
# type: (layer.net) -> None
print("Building Alex-net")
net.reshape(shape=[-1, 28, 28, 1]) # Reshape input pictures
# net.batchnorm()
net.conv([3, 3, 1, 64])
net.conv([3, 3, 64, 128])
net.conv([3, 3, 128, 256])
net.dense(1024,act=tf.nn.relu)
net.dense(1024,act=tf.nn.relu)
# net=layer.net(dense,data=mnist, learning_rate=0.01 )#,'mnist' baseline
_net=net(alex,data=mnist, learning_rate=0.001)#,'mnist'
_net.train(50000,dropout=0.6,display_step=1,test_step=10)
Answer: In general floating point computations can be a little nondeterministic when
it comes to adding many numbers (and some GPUs are buggy). Did you try
retuning hyperparameters (varying learning rates and whatnot) to account for
this?
|
how to process an upload file in Flask?
Question: I have a simple Flask app and I would like for it to process an uploaded excel
file and display it's data in the webpage. So far I got a page to upload the
excel file.
main.py
from flask import Flask, render_template, send_file, request
from flask_uploads import UploadSet, configure_uploads, DOCUMENTS, IMAGES
app = Flask(__name__)
#the name 'datafiles' must match in app.config to DATAFILES
docs = UploadSet('datafiles', DOCUMENTS)
app.config['UPLOADED_DATAFILES_DEST'] = 'static/uploads'
configure_uploads(app, docs)
@app.route("/", methods=['GET'])
def index():
# return send_file("templates/index.html")
return render_template('index.html')
@app.route("/upload", methods = ['GET', 'POST'])
def upload():
#user_file is the name value in input element
if request.method == 'POST' and 'user_file' in request.files:
filename = docs.save(request.files['user_file'])
return filename
return render_template('upload.html')
@app.route("/mergedata", methods=['GET'])
def merge_data():
return send_file("templates/mergeDataPage.html")
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=True)
upload.html
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>upload</title>
</head>
<body>
<form method="POST" enctype=multipart/form-data action ="{{url_for('upload')}}">
<input type="file" name="user_file">
<input type="submit">
</form>
</body>
</html>
Once the file is uploaded, I would like to process this excel data by using
another python script - This script will take the excel file, remove special
characters and converted to a csv file/output. I would like to display the csv
output in the webpage.
The idea is for users to upload the file, press a button to clean the data and
have the results printed in an output box. How do I go about in building
something like this scenario?
I'm pretty new to Flask, but so far has been pretty exciting to learn this
cool web framework. Your help will be appreciated, thanks!
Answer: You can use [pandas](http://pandas.pydata.org/) module and first convert the
excel file.
import pandas as pd
data_xls = pd.read_excel('your_workbook.xls', 'Sheet1', index_col=None)
data_xls.to_csv('your_csv.csv', encoding='utf-8')
Pandas is spectacular for dealing with csv files, and the following code would
be all you need to read a csv and save an entire column into a variable:
import pandas as pd
df = pd.read_csv(csv_file)
saved_column = df.column_name #you can also use df['column_name']
For instance if you wanted to save all of the info in your column Names into a
variable, this is all you need to do:
names = df.Names
|
compare two results (of many) from api data with django/python
Question: I'm learning django/python/css/etc... and while doing this, I've decided to
build an app for my website that can pull simple movie data from TMDb. What
I'm having trouble with is figuring out a way to add a way for the **user to
select two different movies, and once selected, see the differences between
them (run time, budget, etc).**
I've got grabbing the data from the API covered in that doing a search for a
movie on my site returns expected results. But now I'm having a really tough
time trying to figure out **how to select 1 item from the results to "save"
it, search again, select the second movie, and have the comparison show up.**
I know it's pretty vague, but any help getting me pointed in the right
direction would be greatly appreciated!
here's what I'm doing so far with the code:
views.py:
from django.shortcuts import render
from django.conf import settings
from .forms import MovieSearch
import tmdbsimple as tmdb
tmdb.API_KEY = settings.TMDB_API_KEY
def search_movie(request):
"""
Search movie title and return 5 pages of results
"""
parsed_data = {'results': []}
if request.method == 'POST':
form = MovieSearch(request.POST)
if form.is_valid():
search = tmdb.Search()
query = form.cleaned_data['moviename']
response = search.movie(query=query)
for movie in response['results']:
parsed_data['results'].append(
{
'title': movie['title'],
'id': movie['id'],
'poster_path': movie['poster_path'],
'release_date': movie['release_date'][:-6],
'popularity': movie['popularity'],
'overview': movie['overview']
}
)
for i in range(2, 5 + 1):
response = search.movie(query=query, page=i)
for movie in response['results']:
parsed_data['results'].append(
{
'title': movie['title'],
'id': movie['id'],
'poster_path': movie['poster_path'],
'release_date': movie['release_date'][:-6],
'popularity': movie['popularity'],
'overview': movie['overview']
}
)
context = {
"form": form,
"parsed_data": parsed_data
}
return render(request, './moviecompare/movies.html', context)
else:
form = MovieSearch()
else:
form = MovieSearch()
return render(request, './moviecompare/compare.html', {"form": form})
def get_movie(request, movid):
"""
from search/movie results, get details by movie id (movid)
"""
movie = tmdb.Movies(movid)
response = movie.info()
context = {
'response': response
}
return render(request, './moviecompare/detail.html', context)
movies.html:
{% extends 'moviecompare/compare.html' %}
{% block movies_returned %}
<div class="wrap">
<div class="compare-gallery">
{% for key in parsed_data.results|dictsortreversed:'release_date' %}
{% if key.poster_path and key.release_date and key.title and key.overview %}
<div class="gallery-item">
<img src="http://image.tmdb.org/t/p/w185/{{ key.poster_path }}">
<div class="gallery-text">
<div class="gallery-date"><h5><span><i class="material-icons">date_range</i></span> {{ key.release_date }}</h5></div>
<div class="gallery-title"><h3><a href="{% url 'compare:movie_detail' movid=key.id %}">{{ key.title }}</a></h3></div>
<div class="gallery-overview">{{ key.overview|truncatechars:80 }}</div>
</div>
</div>
{% endif %}
{% endfor %}
</div>
</div>
{% endblock %}
and I've got a simple detail.html that doesn't do anything just yet, but it's
just for showing detail results for a single movie, so not important to get
into yet as it'll just be styling.
I'd like for each result in the gallery to have a link to a details page for
the movie(done), and also a select box (or similar) to select it as one of the
comparison movies.
If I can just get some help on how to select two different movies from the
search results, and compare those, I think I could work out a way to do the
same with two separate movie searches. Thanks for any help!
edit: here's what I have so far on
[pythonanywhere](http://moniblog.pythonanywhere.com/compare/) \-
Answer: Assuming you want to be able to add movies from different searches (e.g.
search for rambo, add rambo, search for south park, add south park), you can
probably do something like:
@require_POST
def save_compare(request, pk):
""" Async endpoint to add movies to comparison list """
movies = request.session.get('comparing_moviews', [])
movies.append(movies)
# if you only need the data the API served originally, you could also save that to a data-base when you read it from the API and use it here
request.session['comparing_movies'] = movies
return JsonResponse("")
def show_comparison(request):
""" show differing stats like "1 hour 3 mins longer """
mov1, mov2 = request.session.get('comparing_movies') # need to handle wrong number of selected
# not sure from the description how to look up the movies by id
mov1 = tmdb.Search().search(pk=mov1)
mov2 = tmdb.Search().search(pk=mov2)
return render_to_response("comparison.html", context={
"mov1": mov1, "mov2": mov2,
"length_diff": mov1['length'] - mov2['length']})
to give you the comparison stats you want. You'll also need [session
middleware](https://docs.djangoproject.com/en/1.10/topics/http/sessions/)
installed.
From the search results page, you'll add a button that triggers an async js
post to `save_compare` which starts accumulating a comparison list. When
they're done, they can click "compare" or something on that list, which will
render "comparison.html" with the two movies and the diff however you like it.
As an example of some calculation for your "comparison.html", you can grab the
hours and minutes from the `length_diff` and render them as "X hours and y
minutes longer."
|
How to download japanese history stock prices automatically from google finance in python
Question: I use python to analyze Japanese stock prices. I want to get Japanese
historical stock prices to get from google finance. I also refer to
googlefinance0.7 ( <https://pypi.python.org/pypi/googlefinance> ) and pandas,
but they are not support Japanese stock prices. So how to download japanese
history stock prices automatically from google finance? Or are there
references to program code in python?
Answer: [Quandl](https://www.quandl.com/data/TSE) has all the historical data sets you
need.
It's designed for easy usage with Python so for [Tokyo Stock Exchange TOPIX
Index](https://www.quandl.com/data/TSE/TOPIX-Tokyo-Stock-Exchange-TOPIX-Index)
:
import quandl
mydata = quandl.get("TSE/TOPIX")
|
numpy ndarray with more that 32 dimensions
Question: When I try to create an numpy array with more than 32 dimensions I get an
error:
import numpy as np
np.ndarray([1] * 33)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-2-78103e601d91> in <module>()
----> 1 np.ndarray([1] * 33)
ValueError: sequence too large; cannot be greater than 32
I found this: [Using numpy.array with large number of
dimensions](http://stackoverflow.com/questions/35192518/using-numpy-array-
with-large-number-of-dimensions) to be related to this question but I want to
do this without building my own version.
My use case:
I am working with Joint Probability Distributions and I am trying to represent
each variable on an axis so that computations on it (marginalize, reduce) is a
single line operation. For example for a marginalize operation I can simply do
a sum over the axis of that variable. For multiplication I can simply do a
simple numpy multiplication (after checking if the axes are the same).
Is there a possible workaround this?
Answer: How about maintaining the data in a 1d, and selectively reshaping to focus on
a given dimension
To illustrate
In [252]: x=np.arange(2*3*4*5)
With full reshape
In [254]: x.reshape(2,3,4,5).sum(2)
Out[254]:
array([[[ 30, 34, 38, 42, 46],
[110, 114, 118, 122, 126],
[190, 194, 198, 202, 206]],
[[270, 274, 278, 282, 286],
[350, 354, 358, 362, 366],
[430, 434, 438, 442, 446]]])
With partial reshape - same numbers, different result shape
In [255]: x.reshape(6,4,5).sum(1)
Out[255]:
array([[ 30, 34, 38, 42, 46],
[110, 114, 118, 122, 126],
[190, 194, 198, 202, 206],
[270, 274, 278, 282, 286],
[350, 354, 358, 362, 366],
[430, 434, 438, 442, 446]])
I'm not going to test this on anything approaching 32+ dimensions. As was
noted in the comment, if many of the dimensions are larger than 1 the overall
array size will be excessively large.
|
python3 global name 'SOL_SOCKET is not defined
Question: when I use `gevent`, I can not use `requests`, is my useage wrong?
from gevent import monkey; monkey.patch_all()
import gevent, requests
requests.get('https://haofly.net')
it raise the error:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/adapters.py", line 423, in send
timeout=timeout
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 595, in urlopen
chunked=chunked)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 352, in _make_request
self._validate_conn(conn)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 831, in _validate_conn
conn.connect()
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connection.py", line 289, in connect
ssl_version=resolved_ssl_version)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/util/ssl_.py", line 308, in ssl_wrap_socket
return context.wrap_socket(sock, server_hostname=server_hostname)
File "/usr/local/lib/python3.3/site-packages/gevent/_ssl3.py", line 60, in wrap_socket
_context=self)
File "/usr/local/lib/python3.3/site-packages/gevent/_ssl3.py", line 143, in __init__
if sock.getsockopt(SOL_SOCKET, SO_TYPE) != SOCK_STREAM:
NameError: global name 'SOL_SOCKET' is not defined
The version of my Python is 3.3
If I add `ssl=False` to `monkey.patch_all()`, it returns:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 70, in get
return request('get', url, params=params, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/api.py", line 56, in request
return session.request(method=method, url=url, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 475, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/sessions.py", line 596, in send
r = adapter.send(request, **kwargs)
File "/usr/local/lib/python3.3/site-packages/requests/adapters.py", line 423, in send
timeout=timeout
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 595, in urlopen
chunked=chunked)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 352, in _make_request
self._validate_conn(conn)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connectionpool.py", line 831, in _validate_conn
conn.connect()
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/connection.py", line 289, in connect
ssl_version=resolved_ssl_version)
File "/usr/local/lib/python3.3/site-packages/requests/packages/urllib3/util/ssl_.py", line 308, in ssl_wrap_socket
return context.wrap_socket(sock, server_hostname=server_hostname)
File "/usr/local/lib/python3.3/ssl.py", line 245, in wrap_socket
_context=self)
File "/usr/local/lib/python3.3/ssl.py", line 335, in __init__
server_hostname)
TypeError: _wrap_socket() argument 1 must be _socket.socket, not SSLSocket
Answer: This is really a bug in gevent. It is trying to import a [`socket`
constant](https://docs.python.org/3/library/socket.html#constants) from the
`ssl` library.
That specific constant _happens_ to be imported into the `ssl` library because
that library makes use of it. But it is only there because a fix for [issue
19422](http://bugs.python.org/issue19422) required it to be imported in to the
`ssl` module. The fix for that bug was included in Python 3.3.4, and you have
version 3.3.3 and you thus don't have that specific constant available in the
`ssl` module as well.
You could just fix it by adding `from socket import SOL_SOCKET, SO_TYPE` in
`gevent/_ssl3.py` (located in the `/usr/local/lib/python3.3/site-packages`
directory on your system). Or you can upgrade to Python 3.3.4 or newer.
I've filed this as [issue #856](https://github.com/gevent/gevent/issues/856)
with the `gevent` project.
|
Extract hyper-cubical blocks from a numpy array with unknown number of dimensions
Question: I have a bit of python code which currently is hard-wired with two-dimensional
arrays as follows:
import numpy as np
data = np.random.rand(5, 5)
width = 3
for y in range(0, data.shape[1] - W + 1):
for x in range(0, data.shape[0] - W + 1):
block = data[x:x+W, y:y+W]
# Do something with this block
Now, this is hard coded for a 2-dimensional array and I would like to extend
this to 3D and 4D arrays. I could, of course, write more functions for other
dimensions but I was wondering if there is a python/numpy trick to generate
these sub-blocks without having to replicate this function for
multidimensional data.
Answer: Here is my wack at this problem. The idea behind the code below is to find the
"starting indices" for each slice of data. So for 4x4x4 sub-arrays of a 5x5x5
array, the starting indices would be `(0,0,0), (0,0,1), (0,1,0), (0,1,1),
(1,0,0), (1,0,1), (1,1,1)`, and the slices along each dimension would be of
length 4.
To get the sub-arrays, you just need to iterate over the different tuples of
slice objects and pass them to the array.
import numpy as np
from itertools import product
def iterslice(data_shape, width):
# check for invalid width
assert(all(sh>=width for sh in data_shape),
'all axes lengths must be at least equal to width')
# gather all allowed starting indices for the data shape
start_indices = [range(sh-width+1) for sh in data_shape]
# create tuples of all allowed starting indices
start_coords = product(*start_indices)
# iterate over tuples of slice objects that have the same dimension
# as data_shape, to be passed to the vector
for start_coord in start_coords:
yield tuple(slice(coord, coord+width) for coord in start_coord)
# create 5x5x5 array
arr = np.arange(0,5**3).reshape(5,5,5)
# create the data slice tuple iterator for 3x3x3 sub-arrays
data_slices = iterslice(arr.shape, 3)
# the sub-arrays are a list of 3x3x3 arrays, in this case
sub_arrays = [arr[ds] for ds in data_slices]
|
Discovering data type of incoming socket data in python
Question: There are couple of devices which are sending socket data over _TCP/IP_ to
socket server. Some of the devices are sending data as _Binary encoded
Hexadecimal string_ , others are _ASCII string_.
Eg.;
If device sending data in _ASCII string_ type, script is begin to process
immediately without any conversion.
If device sending _Binary encoded HEX string_ , script should has to convert
_Binary encoded Hex string_ into _Hex string_ first with;
data = binascii.hexlify(data)
There are two scripts running for different data types for that simple single
line. But, I think this could be done in one script if script be aware of the
incoming data type. Is there a way to discover type of the incoming socket
data in Python?
Answer: If you can you should make the sending devices signal what data they are
sending eg by using different TCP ports or prepending each message with an
`"h"` for hex or an `"a"` for ascii - possibly even use an established
protocol like [XML-RPC](https://docs.python.org/3/library/xmlrpc.html)
Actually you can only be sure in _some_ cases as all hex-encoded strings are
valid ascii and some ascii-strings are valid hex like `"CAFE"`. You can make
sure you can decode a string as hex with
import string
def is_possibly_hex(s):
return all(c in string.hexdigits for c in s)
or
import binascii
def is_possibly_hex(s):
try:
binascii.unhexlify(s)
except binascii.Error:
return False
return True
|
Running python on server, executing commands from computer
Question: I made a login system with python. It works perfectly, but i want to run
script on server or web. For example: Steam. Steam wants username and password
to log in. So i wanted to do the same for my script. How can i do that?
My Code:
import os
import string
import time
version = "1.0 Alfa"
def login():
print ("----------------------------------------")
print (" Login ")
print ("----------------------------------------")
k_name = input("Enter username: ")
if os.path.exists(k_name + ".txt") == False:
print ("Username not found.")
create()
else:
k_pass = input("Enter password: ")
with open(k_name + ".txt", "r") as f:
if k_pass == f.read():
print("Welcome %s!"%k_name)
f.close()
input()
else:
print("Password is wrong!")
create()
def create():
print("You using login system %s" % version)
print( "----------------------------------------")
print("| Lobby |")
print( "----------------------------------------")
starting = input("To create user type R, to login type L").upper()
if starting == "R":
name = input("Enter username: ")
password = input("Enter password: ")
password2 = input("Enter password again: ")
if password == password2:
newfile = open(name + ".txt", "w")
newfile.write(password)
newfile.close()
print("User created. Redirecting you to login.")
time.sleep(2)
login()
elif password != password2:
print("Passwords doesn't match.")
input()
create()
elif starting == "L":
login()
else:
print("\nWrong button\n")
create()
create()
Answer: Here is the script which you can run : python test.py user pass it will save
data if file not found and perform login
#!/usr/bin/env python
import sys, getopt
import os
import time
version = "1.0 Alfa"
def login(username=None, password=None):
print ("----------------------------------------")
print (" Login ")
print ("----------------------------------------")
if username:
k_name = username
else:
k_name = input("Enter username: ")
if os.path.exists(k_name + ".txt") == False:
print ("Username not found.")
create(username, password, "R")
else:
if password:
k_pass = password
else:
k_pass = input("Enter password: ")
with open(k_name + ".txt", "r") as f:
if k_pass == f.read():
if not username:
print("Welcome %s!"%k_name)
f.close()
input()
else:
print("Password is wrong!")
create()
def create(username=None, password=None, mode="L"):
print("You using login system %s" % version)
print( "----------------------------------------")
print("| Lobby |")
print( "----------------------------------------")
if mode:
starting = mode
else:
starting = input("To create user type R, to login type L").upper()
if starting == "R":
if username:
name = username
else:
name = input("Enter username: ")
if password:
password2 = password
else:
password = input("Enter password: ")
password2 = input("Enter password again: ")
if password == password2:
newfile = open(name + ".txt", "w")
newfile.write(password)
newfile.close()
print("User created. Redirecting you to login.")
time.sleep(2)
login(username, password)
elif password != password2:
print("Passwords doesn't match.")
input()
create()
elif starting == "L":
login(username, password)
else:
print("\nWrong button\n")
create()
def main(argv):
print sys.argv
if len(sys.argv) < 3:
print 'test.py <username> <password>'
sys.exit()
username = sys.argv[1]
password = sys.argv[2]
print 'username is ', username
print 'password is ', password
create(username, password)
if __name__ == "__main__":
main(sys.argv[1:])
|
Convert a string to JSON
Question: I would like to convert this string to a JSON dict:
{u'Processes': [[u'root', u'3606', u'0.0', u'0.2', u'76768', u'16664', u'?', u'Ss', u'20:40', u'0:01', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4088', u'0.0', u'0.2', u'88544', u'20156', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4090', u'0.0', u'0.2', u'88552', u'20140', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4097', u'0.0', u'0.2', u'88552', u'20112', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4110', u'0.0', u'0.2', u'88548', u'20160', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0']], u'Titles': [u'USER', u'PID', u'%CPU', u'%MEM', u'VSZ', u'RSS', u'TTY', u'STAT', u'START', u'TIME', u'COMMAND']}
So I used json.dumps and json.loads but the output was not a valid JSON. I
understand that JSON needs double quotes instead of simple quotes, but I don't
think that the solution is to search and replace is the best way to resolve
this problem. Is there a more proper way to do it ?
Answer: Use
[`ast.literal_eval`](https://docs.python.org/3.5/library/ast.html#ast.literal_eval)
to convert string to valid Python object.
> Safely evaluate an expression node or a string containing a Python literal
> or container display. The string or node provided may only consist of the
> following Python literal structures: strings, bytes, numbers, tuples, lists,
> dicts, sets, booleans, and None.
s = "{u'Processes': [[u'root', u'3606', u'0.0', u'0.2', u'76768', u'16664', u'?', u'Ss', u'20:40', u'0:01', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4088', u'0.0', u'0.2', u'88544', u'20156', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4090', u'0.0', u'0.2', u'88552', u'20140', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4097', u'0.0', u'0.2', u'88552', u'20112', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0'], [u'root', u'4110', u'0.0', u'0.2', u'88548', u'20160', u'?', u'S', u'20:40', u'0:00', u'/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0']], u'Titles': [u'USER', u'PID', u'%CPU', u'%MEM', u'VSZ', u'RSS', u'TTY', u'STAT', u'START', u'TIME', u'COMMAND']}"
o = ast.literal_eval(s)
assert 'Processes' in o
Use [`json.dumps`](https://docs.python.org/3.5/library/json.html#json.dumps)
to dump it to JSON string.
import json
json.dumps(o)
# '{"Titles": ["USER", "PID", "%CPU", "%MEM", "VSZ", "RSS", "TTY", "STAT", "START", "TIME", "COMMAND"], "Processes": [["root", "3606", "0.0", "0.2", "76768", "16664", "?", "Ss", "20:40", "0:01", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4088", "0.0", "0.2", "88544", "20156", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4090", "0.0", "0.2", "88552", "20140", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4097", "0.0", "0.2", "88552", "20112", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"], ["root", "4110", "0.0", "0.2", "88548", "20160", "?", "S", "20:40", "0:00", "/usr/local/bin/python2 /usr/local/bin/gunicorn app:app -b 0.0.0.0:80 --log-file - --access-logfile - --workers 4 --keep-alive 0"]]}'
Or use `json.dump` to dump it to file object, if that's what you want.
|
imports python3 in another folder
Question: When I run launch.py it fails and when I run main.py directly it works.
launch.py just imports and runs main.py. Why?
├── dir
│ ├── bla.py
│ ├── __init__.py
│ └── main.py
├── __init__.py
└── launch.py
launch.py
---------
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dir import main
main.main()
main.py
-------
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import bla
bla.pront()
bla.py
------
def pront():
print('pront')
EDITED:
[enter image description here](http://i.stack.imgur.com/cgfvU.png)
Answer: Using your layout and with the following files, we don't have problems.
### launch.py
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dir import main
if __name__ == "__main__":
main. main()
### main.py
#!/usr/bin/env python
# -*- coding: utf-8 -*-
try:
from . import bla
except:
import bla
def main():
bla.pront()
if __name__ == "__main__":
main()
The `try ... except` structure is used in case the **_main.py_** was used
inside or outside the package.
Of course, there is a lot of info about it. You can start with
[this](http://stackoverflow.com/questions/16981921/relative-imports-in-
python-3).
|
embedding resources in python scripts
Question: I'd like to figure out how to embed binary content in a python script. For
instance, I don't want to have any external files around (images, sound, ...
), I want all this content living inside of my python scripts.
Little example to clarify, let's say I got this small snippet:
from StringIO import StringIO
from PIL import Image, ImageFilter
embedded_resource = StringIO(open("Lenna.png", "rb").read())
im = Image.open(embedded_resource)
im.show()
im_sharp = im.filter(ImageFilter.SHARPEN)
im_sharp.show()
As you can see, the example is reading the external file 'Lenna.png'
[](http://i.stack.imgur.com/3p4GL.png)
**Question**
How to proceed to embed "Lenna.png" as a resource (variable) into my python
script. What's the fastest way to achieve this simple task using python?
Answer: The best way to go about this is converting your picture into a python string,
and have it in a separate file called something like `resources.py`, then you
simply parse it.
If you are looking to embed the whole thing inside a single binary, then
you're looking at something like [py2exe](http://py2exe.org/).
[Here](http://www.py2exe.org/index.cgi/data_files) is an example embedding
external files
In the first scenario, you could even use `base64` to (de)code the picture,
something like this:
import base64
file = open('yourImage.png');
encoded = base64.b64encode(file.read())
data = base64.b64decode(encoded) # Don't forget to file.close() !
|
Shut down Flask SocketIO Server
Question: For circumstances outside of my control, I need to use the Flask server to
serve basic html files, the Flask SocketIO wrapper to provide a web socket
interface between any clients and the server. The `async_mode` has to be
`threading` instead of `gevent` or `eventlet`, I understand that it is less
efficient to use threading, but I can't use the other two frameworks.
In my unit tests, I need to shut down and restart the web socket server. When
I attempt to shut down the server, I get the `RunTimeError` 'Cannot stop
unknown web server.' This is because the function `werkzeug.server.shutdown`
is not found in the Flask Request Environment `flask.request.environ` object.
Here is how the server is started.
SERVER = flask.Flask(__name__)
WEBSOCKET = flask_socketio.SocketIO(SERVER, async_mode='threading')
WEBSOCKET.run(SERVER, host='127.0.0.1', port=7777)
Here is the short version of how I'm attempting to shut down the server.
client = WEBSOCKET.test_client(SERVER)
@WEBSOCKET.on('kill')
def killed():
WEBSOCKET.stop()
try:
client.emit('kill')
except:
pass
The stop method must be called from within a flask request context, hence the
weird kill event callback. Inside the stop method, the `flask.request.environ`
has the value
'CONTENT_LENGTH' (40503696) = {str} '0'
'CONTENT_TYPE' (60436576) = {str} ''
'HTTP_HOST' (61595248) = {str} 'localhost'
'PATH_INFO' (60437104) = {str} '/socket.io'
'QUERY_STRING' (60327808) = {str} ''
'REQUEST_METHOD' (40503648) = {str} 'GET'
'SCRIPT_NAME' (60437296) = {str} ''
'SERVER_NAME' (61595296) = {str} 'localhost'
'SERVER_PORT' (61595392) = {str} '80'
'SERVER_PROTOCOL' (65284592) = {str} 'HTTP/1.1'
'flask.app' (65336784) = {Flask} <Flask 'server'>
'werkzeug.request' (60361056) = {Request} <Request 'http://localhost/socket.io' [GET]>
'wsgi.errors' (65338896) = {file} <open file '<stderr>', mode 'w' at 0x0000000001C92150>
'wsgi.input' (65338848) = {StringO} <cStringIO.StringO object at 0x00000000039902D0>
'wsgi.multiprocess' (65369288) = {bool} False
'wsgi.multithread' (65369232) = {bool} False
'wsgi.run_once' (65338944) = {bool} False
'wsgi.url_scheme' (65338800) = {str} 'http'
'wsgi.version' (65338752) = {tuple} <type 'tuple'>: (1, 0)
My question is, how do I set up the Flask server to have the
`werkzeug.server.shutdown`method available inside the flask request contexts?
Also this is using Python 2.7
Answer: I have good news for you, the testing environment does not use a real server,
in that context the client and the server run inside the same process, so the
communication between them does not go through the network as it does when you
run things for real. Really in this situation there is no server, so there's
nothing to stop.
It seems you are starting a real server, though. For unit tests, that server
is not used, all you need are your unit tests which import the application and
then use a test client to issue socket.io events. I think all you need to do
is just not start the server, the unit tests should run just fine without it
if all you use is the test client as you show above.
|
assert vs == for testing code in Python?
Question: What is the need for importing unittest and running assertTrue (for example)
while testing a python function instead of writing a usual python function
with == True check for testing? What is the new thing about unittesting, as
even the test cases have to be written by user, which can be checked by ==
instead of assert family of functions in unittest? My question basically is:
assert in unittest vs the equality check operation == in Python.
Answer: I would say that the most interesting thing is the "easy-to-test" that the
unittest package (as well as others, for example pytest) provides. When using
unittest you can just run a command and check if your refactoring or your last
feature broke something on the rest of the program. I would read something
about [test-driven development](https://en.wikipedia.org/wiki/Test-
driven_development) to understand better what is the importance of
unittesting.
|
My numpy is latest but tensroflow says it's old one
Question:
$pip list
numpy(1.11.1)
My numpy is latest and I am sure it could be used in python environment.
>>> import numpy
>>> print numpy.__version__
1.11.1
however I use tensorflow
$ tensorboard --logdir=/tmp/basic_rnn/logdir
it shows this error
RuntimeError: module compiled against API version 0xa but this version of numpy is 0x9
What causes this problem??
Answer: I suspect that the `tensorboard` command is linked against the wrong Python
environment. Make sure that `tensorflow` uses the same Python
interpreter/virtual environment as you checked your numpy installation with.
Also check your `PATH` and `PYTHONPATH` and take a look at `sys.path` and
`sys.prefix` to see if there are some paths mixed up.
|
python import * or a list from other level
Question: I'm trying to import a few classes from a module in another level. I can type
all the classes, but I' trying to do it dynamically
if I do:
from ..previous_level.module import *
raise: SyntaxError: import * only allowed at module level
the same from myapp folder:
from myapp.previous_level.module import *
raise: SyntaxError: import * only allowed at module level
So I thought:
my_classes = ['Class_Foo', 'Class_Bar']
for i in my_classes:
from ..previous_level.module import i
raise: ImportError: cannot import name 'i'
and also:
my_classes = ['Class_Foo', 'Class_Bar']
for i in my_classes:
__import__('myapp').previous_level.module.y
raise: AttributeError: module 'myapp.previous_level.module' has no attribute 'y'
I've tried `string format` , `getattr()` , `__getattr__` but no success.
It's impossible to do import this way, or I'm doing something wrong?
Answer: The error `SyntaxError: import * only allowed at module level` happens if you
try to import `*` from within a function. This is not visible from the
question, but it seems that the original code was simething like:
def func():
from ..previous_level.module import * # SyntaxError
On the other hand, importing at the beginning of a module would be valid:
from ..previous_level.module import * # OK
That said, I would still suggest using absolute imports without `*`:
from myapp.previous_level.module import ClassA, ClassB, ClassC
x = ClassA()
# or:
import myapp.previous_level.module
x = myapp.previous_level.module.ClassA()
BTW, this is completely wrong:
for i in my_classes:
from ..previous_level.module import i
|
get picture from dynamic content python
Question: I'm trying to get the href of the picture from an url without using selenium
def():
try:
page = urllib2.urlopen('')
except httplib.IncompleteRead, e:
page = e.partial
response = BeautifulSoup(page)
print response
var = response.find("div", {"id":"il_m"}).find('p')
but i got None as a result.What should I do to ge the href ?
Answer: You can also get the link from an _anchor_ tag with a _download_ attribute:
In [2]: from bs4 import BeautifulSoup
In [3]: import urllib2
In [4]: r = urllib2.urlopen('http://icecat.us/index.php/product/image_gallery?num=9010647&id=9409545&lang=us&imgrefurl=philips.com')
In [5]: soup = BeautifulSoup(r,"html.parser")
In [6]: print(soup.select_one("p a[download]")["href"])
http://images.icecat.biz/img/gallery/9010647-Philips-_FP.jpg
You should also take note of the text _Images may be subject to copyright._.
on the page.
|
Russian character decoding in python
Question: This question only for python:
I have a city name in a string in Russian language and which is in Unicode
form like,
> `\u041C\u043E\u0441\u043A\u0432\u0430`
means
> `Москва`
How to get original text instead of unicode characters?
**Note:** Do not use any import module
Answer:
>>> a=u"\u041C\u043E\u0441\u043A\u0432\u0430"
>>> print a
Москва
Your string is a unicode string because each character/code point with \u is
only usable from a unicode string, you should prefix the string with u.
Otherwise is a regular string and each \u counts as a regular ascii character:
>>> len(a)
6
>>> b="\u041C\u043E\u0441\u043A\u0432\u0430"
>>> len(b)
36
|
I'm not sure what a certain code line does
Question: I was trying to have a program using Python to create a "pyramid" based on a
number, n, out of o's and came up with this: ( I would print nn, that would be
the lines.)
import time
n = 0
while True:
n += 2 #just another way to show n=n+2
nn = n, "o" #nn would be an amount of o's, based on what #n was
time.sleep(1)
print (nn, str.center(40, " "))
Not sure how to make nn the o's and not sure what line 6 does either. Does
anyone know the answer to either question? ( I'm not in a class just
programming for fun.)
Answer:
import time
n = 0
while True:
n += 2 #just another way to show n=n+2
nn = "o" * n #nn would be an amount of o's, based on what #n was
time.sleep(1)
print (nn.center(40," "))
As someone has mentioned in the comments "o" * n will give a a string
containing n "o"s. I've fixed the print line to use the correct method a
calling the center method.
|
Selenium Webdriver Python: I can't seem to locate all the text in a label tag
Question:
<label class="control-label">
Rental Charge:
<span class="required" ng-show="vm.rentalInfo.reason">* (Min of $30.00)</span>
</label>
I used
driver.find_element_by_xpath("//label[@ class = 'control-label']/span[@class = 'required']").text
and the only `text` I get back is the `asterisk *`, `(Min of $30.00)` is not
displayed. Does anyone know why this is?
Answer: I am guessing this is a timing issue. [Wait](http://selenium-
python.readthedocs.io/waits.html#explicit-waits) for "$" to be present in
element explicitly with `text_to_be_present_in_element` expected condition:
from selenium.webdriver.common.by import By
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions as EC
wait = WebDriverWait(driver, 10)
elm = wait.until(EC.text_to_be_present_in_element((By.CSS_SELECTOR, 'span[ng-show*="rentalInfo.reason"]'), "$"))
print(elm.text)
|
Run multiple servers in python at same time (Threading)
Question: I have **2 servers** in python, **I want to mix them up in one single .py and
run together** :
Server.py:
import logging, time, os, sys
from yowsup.layers import YowLayerEvent, YowParallelLayer
from yowsup.layers.auth import AuthError
from yowsup.layers.network import YowNetworkLayer
from yowsup.stacks.yowstack import YowStackBuilder
from layers.notifications.notification_layer import NotificationsLayer
from router import RouteLayer
class YowsupEchoStack(object):
def __init__(self, credentials):
"Creates the stacks of the Yowsup Server,"
self.credentials = credentials
stack_builder = YowStackBuilder().pushDefaultLayers(True)
stack_builder.push(YowParallelLayer([RouteLayer, NotificationsLayer]))
self.stack = stack_builder.build()
self.stack.setCredentials(credentials)
def start(self):
self.stack.broadcastEvent(YowLayerEvent(YowNetworkLayer.EVENT_STATE_CONNECT))
try:
logging.info("#" * 50)
logging.info("\tServer started. Phone number: %s" % self.credentials[0])
logging.info("#" * 50)
self.stack.loop(timeout=0.5, discrete=0.5)
except AuthError as e:
logging.exception("Authentication Error: %s" % e.message)
if "<xml-not-well-formed>" in str(e):
os.execl(sys.executable, sys.executable, *sys.argv)
except Exception as e:
logging.exception("Unexpected Exception: %s" % e.message)
if __name__ == "__main__":
import sys
import config
logging.basicConfig(stream=sys.stdout, level=config.logging_level, format=config.log_format)
server = YowsupEchoStack(config.auth)
while True:
# In case of disconnect, keeps connecting...
server.start()
logging.info("Restarting..")
App.py:
import web
urls = (
'/', 'index'
)
app = web.application(urls, globals())
class index:
def GET(self):
greeting = "Hello World"
return greeting
if __name__ == "__main__":
app.run()
I want to **run both together from single .py file together**. If I try to run
them from one file, **either of the both starts and other one starts only when
first one is done working.**
**How can I run 2 servers** in python together?
Answer:
import thread
def run_app1():
#something goes here
def run_app2():
#something goes here
if __name__=='__main__':
thread.start_new_thread(run_app1)
thread.start_new_thread(run_app2)
if you need to pass args to the functions you can do:
thread.start_new_thread(run_app1, (arg1,arg2,....))
if you want more control in your threads you could go:
import threading
def app1():
#something here
def app2():
#something here
if __name__=='__main__':
t1 = threading.Thread(target=app1)
t2 = threading.Thread(target=app2)
t1.start()
t2.start()
if you need to pass args you can go:
t1 = threading.Thread(target=app1, args=(arg1,arg2,arg3.....))
What's the differences between thread vs threading? Threading is higher level
module than thread and in 3.x thread got renamed to _thread... more info here:
<http://docs.python.org/library/threading.html> but that's for another
question I guess.
So in your case, just make a function that runs the first script, and the
second script, and just spawn threads to run them.
|
Selenium webdriver unable to restart after unexpected exit
Question: I haven't been able to start up an instance of python's selenium webdriver
after my last use a few days ago. According to the error messages, it
unexpectedly quit last time I was using it, and now, after restarting my
macbook, uninstalling and reinstalling chromedriver/selenium:
`brew rmtree chromedriver && brew install chromedriver`
`pip uninstall selenium && pip install selenium`
I'm still in the same place. It seems to be selenium itself, because for both
Firefox and Chrome, I'm getting error messages.
What I'm attempting to run on the python3.5 kernel is:
from selenium import webdriver
driver = webdriver.Chrome()
stacktrace:
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/chrome/webdriver.py",
line 62, in __init__self.service.start()
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/common/service.py",
line 86, in start self.assert_process_still_running()
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/common/service.py",
line 99, in assert_process_still_running % (self.path, return_code)
selenium.common.exceptions.WebDriverException: Message: Service chromedriver unexpectedly exited. Status code was: -5
likewise, for Firefox:
from selenium import webdriver
driver = webdriver.Firefox()
and:
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/webdriver.py",
line 80, in __init__
self.binary, timeout)
File "/Users/myuser/webscraping/env/lib/python3.5/site-packagesyuser/selenium/webdriver/firefox/extension_connection.py",
line 52, in __init__
self.binary.launch_browser(self.profile, timeout=timeout)
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/firefox_binary.py",
line 68, in launch_browser
self._wait_until_connectable(timeout=timeout)
File "/Users/myuser/webscraping/env/lib/python3.5/site-packages/selenium/webdriver/firefox/firefox_binary.py",
line 99, in _wait_until_connectable
"The browser appears to have exited "
selenium.common.exceptions.WebDriverException: Message: The browser appears to have exited before we could connect.
If you specified a log_file in the FirefoxBinary constructor, check it for details.
Should I be looking for some rogue process to kill with `ps -e` and `kill
-sigint`?
Answer: I discovered that homebrew chromedriver was throwing an error related to
symlinking the correct dylib. I fixed the issue by following the steps in
[this](http://stackoverflow.com/questions/17643509/conflict-between-dynamic-
linking-priority-in-osx#answer-35070568) answer to get chromedriver running
again, which enabled selenium/chrome webdriver to work as well.
|
django: link to detail page from list of returned results
Question: I'm creating a page that returns a list of movies with basic details after a
user search.
**After the search, I'd like the user to be able to click on a movie, and get
more details about it.**
here's a link to the site: (be gentle, I only started learning this stuff 2
months ago! hah) <http://moniblog.pythonanywhere.com/compare/>
the data is coming from [TMDB's
api](https://www.themoviedb.org/documentation/api) and the initial "generic"
search JSON response doesn't have specific details that I'd display on the
movie's detail page, but it has an ID that will be used for the specific
movie's search.
I'm only using views.py to grab/display the search results, and I'm not sure
if this is the right way to go, or if I should use a model, but that's
probably a different question.
forms.py:
from django import forms
class MovieSearch(forms.Form):
moviename = forms.CharField(label="Search", max_length=250)
views.py:
from django.shortcuts import render, get_object_or_404
from django.conf import settings
from .forms import MovieSearch
import tmdbsimple as tmdb
tmdb.API_KEY = settings.TMDB_API_KEY
def search_movie(request):
parsed_data = {'results': []}
if request.method == 'POST':
form = MovieSearch(request.POST)
if form.is_valid():
search = tmdb.Search()
query = form.cleaned_data['moviename']
response = search.movie(query=query)
for movie in response['results']:
parsed_data['results'].append(
{
'title': movie['title'],
'id': movie['id'],
'poster_path': movie['poster_path'],
'release_date': movie['release_date'][:-6],
'popularity': movie['popularity'],
'overview': movie['overview']
}
)
for i in range(2, 5 + 1):
response = search.movie(query=query, page=i)
for movie in response['results']:
parsed_data['results'].append(
{
'title': movie['title'],
'id': movie['id'],
'poster_path': movie['poster_path'],
'release_date': movie['release_date'][:-6],
'popularity': movie['popularity'],
'overview': movie['overview']
}
)
context = {
"form": form,
"parsed_data": parsed_data
}
return render(request, './moviecompare/movies.html', context)
else:
form = MovieSearch()
else:
form = MovieSearch()
return render(request, './moviecompare/compare.html', {"form": form})
and html:
{% extends 'moviecompare/compare.html' %}
{% block movies_returned %}
<div class="wrap">
<div class="compare-gallery">
{% for key in parsed_data.results|dictsortreversed:'release_date' %}
{% if key.poster_path and key.release_date and key.title and key.overview %}
<div class="gallery-item">
<img src="http://image.tmdb.org/t/p/w185/{{ key.poster_path }}">
<div class="gallery-text">
<div class="gallery-date"><h5><span><i class="material-icons">date_range</i></span> {{ key.release_date }}</h5></div>
<div class="gallery-title"><h3><a href="../detail/{{ key.id }}">{{ key.title }}</a></h3></div>
<div class="gallery-overview">{{ key.overview|truncatechars:80 }}</div>
</div>
</div>
{% endif %}
{% endfor %}
</div>
</div>
{% endblock %}
I've started playing with the urls.py to get something to work, but no luck so
far.
site's urls.py:
urlpatterns = [
url(r'^$', home, name="home"),
url(r'^blog/', include('blog.urls', namespace='blog')),
url(r'^compare/', include('moviecompare.urls', namespace='compare')),
url(r'^movies/', include('moviecompare.urls', namespace='movies')),
and the app's urls.py:
from django.conf.urls import url
from . import views
urlpatterns = [
url(r'^', views.search_movie, name='compare'),
url(r'^(?P<movid>[0-9])+$', views.get_movie, name='movies')
]
edit: adding my 1st (failed) attempt at the movie detail view:
def get_movie(request, movid=None):
instance = get_object_or_404(request, movid=movid)
context = {
'instance': instance
}
return render(request, './moviecompare/detail.html', context)
Answer: I think you should try by fixing the urls.py in this line:
url(r'^(?P<movid>[0-9])+$', views.get_movie, name='movies')
Move the "+" inside the brackets like that:
url(r'^(?P<movid>[0-9]+)$', views.get_movie, name='movies')
|
I am using beautifulsoup in python 3. But "html.parser" not give me the all code of website
Question: My code is
import urllib
import urllib.request
from bs4 import BeautifulSoup
fed = "https://www.fedex.com/apps/fedextrack/?action=track&tracknumbers=870915037012296&cntry_code=us&wsch=true"
show = urllib.request.urlopen(fed)
soup = BeautifulSoup(show,"html.parser")
print(soup)
I am trying to fetch the delivery address from fedex and status.But this code
not fetching all the data of my URL
Answer: Why don't use [Fedex SOAP API](http://www.fedex.com/us/developer/) directly.
There is a Python wrapper - [`python-fedex`](http://python-
fedex.readthedocs.io/en/latest/quick_start.html).
Example how to track a package can be found [here](http://python-
fedex.readthedocs.io/en/latest/quick_start.html#track-a-package).
|
Jenkins git triggered build not blocking
Question: I am running a build on commit to `origin/master` on my jenkins server that is
deploying resources to Amazon AWS. I am using the Execute Shell section to run
a python script that handles all unit testing/linting/validation/deployment
and everything blocks fine until it gets to the deployment (`deploy.deploy()`)
where it returns a success right after kickoff, but doesn't complete
deploying. How can I make this block?
For reference here is my config:
**Execute Shell (Jenkins)** :
export DEPLOY_REGION=us-west-2
. build-tools/get_aws_credentials.sh
python build-tools/kickoff.py
**kickoff.py**
if __name__ == "__main__":
build_tools_dir="{}".format("/".join(os.path.abspath(__file__).split("/")[0:-1]))
sys.path.append(build_tools_dir)
base_dir = "/".join(build_tools_dir.split("/")[0:-1])
test_begin = __import__("test_begin")
test_all_templates = __import__("test_all_templates")
deploy = __import__("deploy")
git_plugin = __import__("git_plugin")
retval = test_begin.entrypoint("{}/platform/backend".format(base_dir))
if (retval == "SUCCESS"):
retval = test_all_templates.entrypoint("{}/platform/backend".format(base_dir))
if (retval == "SUCCESS"):
deploy.deploy()
**deploy.py**
def deploy():
print(". {}/platform/backend/nu.sh --name jenkinsdeploy --stage dev --keyname greig --debug".format("/".join(os.path.abspath(__file__).split("/")[0:-2])))
returnedcode = subprocess.call("sh {}/platform/backend/nu.sh --name jenkinsdeploy --stage dev --keyname colin_greig --debug".format("/".join(os.path.abspath(__file__).split("/")[0:-2])), shell=True)
if returnedcode == 0:
return "DEPLOY SUCCESS"
return "DEPLOY FAILURE"
Answer: You can use api calls to aws to retrieve the status and wait until it become
'some status'.
The below example is psuedo-code to illustrate the idea:
import time
import boto3
ec2 = boto3.resource('ec2')
while True:
response = ec2.describe_instance_status(.....)
if response == 'some status':
break
time.sleep(60)
# continue execution
|
Obtain csv-like parse AND line length byte count?
Question: I'm familiar with the `csv` Python module, and believe it's necessary in my
case, as I have some fields that contain the delimiter (`|` rather than `,`,
but that's irrelevant) within quotes.
However, I am also looking for the byte-count length of each original row,
_prior_ to splitting into columns. I can't count on the data to always quote a
column, and I don't know if/when `csv` will strip off outer quotes, so I don't
think (but might be wrong) that simply joining on my delimiter will reproduce
the original line string (less CRLF characters). Meaning, I'm not positive the
following works:
with open(fname) as fh:
reader = csv.reader(fh, delimiter="|")
for row in reader:
original = "|".join(row) ## maybe?
I've tried looking at `csv` to see if there was anything in there that I could
use/monkey-patch for this purpose, but since `_csv.reader` is a `.so`, I don't
know how to mess around with that.
In case I'm dealing with an XY problem, my ultimate goal is to read through a
CSV file, extracting certain fields and their overall file offsets to create a
sort of look-up index. That way, later, when I have a list of candidate
values, I can check each one's file-offset and `seek()` there, instead of
chugging through the whole file again. As an idea of scale, I might have 100k
values to look up across a 10GB file, so re-reading the file 100k times
doesn't feel efficient to me. I'm open to other suggestions than the CSV
module, but will still need `csv`-like intelligent parsing behavior.
EDIT: Not sure how to make it more clear than the title and body already
explains - simply `seek()`-ing on a file handle isn't sufficient because I
**also** need to parse the lines as a `csv` in order to pull out additional
information.
Answer: You can't subclass `_csv.reader`, but the _`csvfile`_ argument to the
`csv.reader()`
[constructor](https://docs.python.org/3/library/csv.html#csv.reader) only has
to be a "file-like object". This means you could supply an instance of your
own class that does some preprocessing—such as remembering the length of the
last line read and file offset. Here's an implementation showing exactly that.
Note that the line length does _not_ include the end-of-line character(s). It
also shows how the offsets to each line/row could be stored and used after the
file is read.
import csv
class CSVInputFile(object):
""" File-like object. """
def __init__(self, file):
self.file = file
self.offset = None
self.linelen = None
def __iter__(self):
return self
def __next__(self):
offset = self.file.tell()
data = self.file.readline()
if not data:
raise StopIteration
self.offset = offset
self.linelen = len(data)
return data
next = __next__
offsets = [] # remember where each row starts
fname = 'unparsed.csv'
with open(fname) as fh:
csvfile = CSVInputFile(fh)
for row in csv.reader(csvfile, delimiter="|"):
print('offset: {}, linelen: {}, row: {}'.format(
csvfile.offset, csvfile.linelen, row)) # file offset and length of row
offsets.append(csvfile.offset) # remember where each row started
|
How to use yaml.load_all with fileinput.input?
Question: Without resorting to `''.join`, is there a Pythonic way to use PyYAML's
`yaml.load_all` with `fileinput.input()` for easy streaming of multiple
documents from multiple sources?
I'm looking for something like the following (non-working example):
# example.py
import fileinput
import yaml
for doc in yaml.load_all(fileinput.input()):
print(doc)
Expected output:
$ cat >pre.yaml <<<'--- prefix-doc'
$ cat >post.yaml <<<'--- postfix-doc'
$ python example.py pre.yaml - post.yaml <<<'--- hello'
prefix-doc
hello
postfix-doc
Of course, `yaml.load_all` expects either a string, bytes, or a file-like
object and `fileinput.input()` is none of those things, so the above example
does not work.
Actual output:
$ python example.py pre.yaml - post.yaml <<<'--- hello'
...
AttributeError: FileInput instance has no attribute 'read'
You can make the example work with `''.join`, but that's cheating. I'm looking
for a way that does not read the entire stream into memory at once.
We might rephrase the question as _Is there some way to emulate a string,
bytes, or file-like object that proxies to an underlying iterator of strings?_
However, I doubt that `yaml.load_all` actually needs the entire file-like
interface, so that phrasing would ask for more than is strictly necessary.
Ideally I'm looking for the minimal adapter that would support something like
this:
for doc in yaml.load_all(minimal_adapter(fileinput.input())):
print(doc)
Answer: The problem with `fileinput.input` is that the resulting object doesn't have a
`read` method, which is what `yaml.load_all` is looking for. If you're willing
to give up `fileinput`, you can just write your own class that will do what
you want:
import sys
import yaml
class BunchOFiles (object):
def __init__(self, *files):
self.files = files
self.fditer = self._fditer()
self.fd = self.fditer.next()
def _fditer(self):
for fn in self.files:
with sys.stdin if fn == '-' else open(fn, 'r') as fd:
yield fd
def read(self, size=-1):
while True:
data = self.fd.read(size)
if data:
break
else:
try:
self.fd = self.fditer.next()
except StopIteration:
self.fd = None
break
return data
bunch = BunchOFiles(*sys.argv[1:])
for doc in yaml.load_all(bunch):
print doc
The `BunchOFiles` class gets you an object with a `read` method that will
happily iterate over a list of files until everything is exhausted. Given the
above code and your sample input, we get exactly the output you're looking
for.
|
matplotlib multiple values under cursor
Question: This question is very similar to those answered here,
[matplotlib values under
cursor](http://stackoverflow.com/questions/14754931/matplotlib-values-under-
cursor)
[In a matplotlib figure window (with imshow), how can I remove, hide, or
redefine the displayed position of the
mouse?](http://stackoverflow.com/questions/14349289/in-a-matplotlib-figure-
window-with-imshow-how-can-i-remove-hide-or-redefine)
[Interactive pixel information of an image in
Python?](http://stackoverflow.com/questions/27704490/interactive-pixel-
information-of-an-image-in-python)
except that instead of pixel data (x,y,z) I have various measurements
associated with (x,y) coordinates that I'd like to portray on a line plot.
Specifically, the (x,y) are spatial positions (lat, lon) and at each (lat,lon)
point there is a collection of data (speed, RPM, temp, etc.). I just sketched
up something quickly to illustrate, a scatter plot with connecting lines, and
then when you hover over a data point it displays all of the "z" values
associated with that data point.
Is there an easy way to do something like this?
[](http://i.stack.imgur.com/rJvn9.png)
Answer: You could probably build on something like this example. It doesn't display
the information inside the figure (for now only using a `print()` statement),
but it demonstrates a simple method of capturing clicks on `scatter` points
and showing information for those points:
import numpy as np
import matplotlib.pylab as pl
pl.close('all')
n = 10
lat = np.random.random(n)
lon = np.random.random(n)
speed = np.random.random(n)
rpm = np.random.random(n)
temp = np.random.random(n)
def on_click(event):
i = event.ind[0]
print('lon={0:.2f}, lat={1:.2f}: V={2:.2f}, RPM={3:.2f}, T={4:.2f}'\
.format(lon[i], lat[i], speed[i], rpm[i], temp[i]))
fig=pl.figure()
pl.plot(lon, lat, '-')
pl.scatter(lon, lat, picker=True)
fig.canvas.mpl_connect('pick_event', on_click)
Clicking around a bit gives me:
lon=0.63, lat=0.58: V=0.51, RPM=0.00, T=0.43
lon=0.41, lat=0.07: V=0.95, RPM=0.59, T=0.98
lon=0.86, lat=0.13: V=0.33, RPM=0.27, T=0.85
|
python logging module AttributeError: 'str' object has no attribute 'write'
Question: I am using tornado,and in its app,I import logging just want to log some info
about server. I put this:
logging.config.dictConfig(web_LOGGING)
right before:
tornado.options.parse_command_line()
but when I run the server,when I click any link,I get error:
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 874, in emit
stream.write(fs % msg)
AttributeError: 'str' object has no attribute 'write'
Logged from file web.py, line 1946
it just repeats when i click into any link. what is the real problem?
I've already changed any file,directory to others to steer clear of namespace
conflict...
Answer: Psychic debugging says `web_LOGGING` has a key named `stream` with a `str`
value (probably a file path); [`stream` is only for already opened files, if
you want to pass a file path, it's passed as
`filename`](https://docs.python.org/3/library/logging.html#logging.basicConfig).
|
wxpython phoenix: How to get float values from wx grid cells and perform mathematical operations?
Question: I am trying to build a invoice with wx grid, I would like to add the values in
quantity column and the values in price column and display it in the row
total.
import wx
import wx.grid as gridlib
class MyForm(wx.Frame):
def __init__(self):
wx.Frame.__init__(self, None, wx.ID_ANY, "Invoice")
datag = 0
# Add a panel so it looks the correct on all platforms
panel = wx.Panel(self, wx.ID_ANY)
grid = gridlib.Grid(panel)
grid.CreateGrid(25, 3)
note_sizer = wx.BoxSizer()
note_sizer.Add(grid, 1, wx.EXPAND)
panel.SetSizer(note_sizer)
# We can set the sizes of individual rows and columns
# in pixels
grid.SetColSize(0, 520)
# change a couple column labels
grid.SetColLabelValue(0, "Product")
grid.SetColLabelValue(1, "Quantity")
grid.SetColLabelValue(2, "Price")
grid.SetColFormatFloat(2)
grid.SetCellValue(24, 0, 'Total')
if __name__ == "__main__":
app = wx.App()
frame = MyForm().Show()
app.MainLoop()
can anyone help?
Answer: Once you start wanting to do things like that then it is a good idea to use a
custom grid table for containing your data values. In other words, instead of
you putting your values into the grid and getting the values back out when
needed, the grid simply asks your table for the values to display, on an as-
needed basis. This makes it easy to deal with the data in whatever structure
and data types makes sense for your application. There are some examples in
the wxPython demo that show some ways to implement table classes derived from
`GridTableBase` and how to use them.
|
Class and external method call
Question: I am going through a data structures course and I am not understanding how a
Class can call a method that's in another Class.
The code below has 2 classes: `Printer` and `Task`.
Notice that class `Printer` has a method called `startNext`, and this has a
variable `self.timeRemaining` that gets assigned the result of
`newTask.getPages() * 60/self.pagerate`.
How can `newTaks` reference the `getPages()` method from the Task class?
The code that passes this object to the Printer class never references the
Task class.
The code works, since this is what the course gives out but, I just cannot
understand how that method is accessed.
Code:
from pythonds.basic.queue import Queue
import random
class Printer:
def __init__(self, ppm):
self.pagerate = ppm
self.currentTask = None
self.timeRemaining = 0
def tick(self):
if self.currentTask != None:
self.timeRemaining = self.timeRemaining - 1
if self.timeRemaining <= 0:
self.currentTask = None
def busy(self):
if self.currentTask != None:
return True
else:
return False
def startNext(self, newTask):
self.currentTask = newTask
self.timeRemaining = newTask.getPages() * 60/self.pagerate
class Task:
def __init__(self, time):
self.timeStamp = time
self.pages = random.randrange(1, 21)
def getStamp(self):
return self.timeStamp
def getPages(self):
return self.pages
def waitTime(self, currentTime):
return currentTime - self.timeStamp
def simulation(numSeconds, pagesPerMinute):
labPrinter = Printer(pagesPerMinute)
printQueue = Queue()
waitingTimes = []
for currentSecond in range(numSeconds):
if newPrintTask():
task = Task(currentSecond)
printQueue.enqueue(task)
if (not labPrinter.busy()) and (not printQueue.isEmpty()):
nextTask = printQueue.dequeue()
waitingTimes.append(nextTask.waitTime(currentSecond))
labPrinter.startNext(nextTask)
labPrinter.tick()
averageWait = sum(waitingTimes)/len(waitingTimes)
print "Average Wait %6.2f secs %3d tasks remaining." % (averageWait, printQueue.size())
def newPrintTask():
num = random.randrange(1, 181)
if num == 180:
return True
else:
return False
for i in range(10):
simulation(3600, 5)
Answer: If I understand clearly your question, it is because you are adding task
object to Queue list. Then when you are getting object (list item) back, you
are getting again Task object:
#creating Task object and adding to Queque list
task = Task(currentSecond)
printQueue.enqueue(task)
class Queue:
def __init__(self):
#list of task objects
self.items = []
def enqueue(self, item):
#you are inserting Task object item to list
self.items.insert(0,item)
def dequeue(self):
#returns task object
return self.items.pop()
So then you can call startNext() method from Printer class, because the
dequeue() method is returning Task object.
And because of the object in startNext() is type of Task, you can call
getPages() method on that object.
Is it sufficient answer?
|
tweepy.error.TweepError: [{u'message': u'Text parameter is missing.', u'code': 38}]
Question: I am using `tweepy` twitter api for python, while using it i got some error, I
am not able to use `send_direct_message(user/screen_name/user_id, text)` this
method
Here is my code:-
import tweepy
consumer_key='XXXXXXXXXXXXXXXXX'
consumer_secret='XXXXXXXXXXXXXXXXX'
access_token='XXXXXXXXXXXXXXXXX'
access_token_secret='XXXXXXXXXXXXXXXXX'
auth = tweepy.OAuthHandler(consumer_key, consumer_secret)
auth.set_access_token(access_token, access_token_secret)
API = tweepy.API(auth)
user = API.get_user('SSPendse')
screen_name="CJ0495"
id = 773436067231956992
text="Message Which we have to send must have maximum 140 characters, If it is Greater then the message will be truncated upto 140 characters...."
# re = API.send_direct_message(screen_name, text)
re = API.send_direct_message(id, text)
print re
got following error:-
Traceback (most recent call last):
File "tweetApi.py", line 36, in <module>
re = API.send_direct_message(id, text)
File "/usr/local/lib/python2.7/dist-packages/tweepy/binder.py", line 245, in _call
return method.execute()
File "/usr/local/lib/python2.7/dist-packages/tweepy/binder.py", line 229, in execute
raise TweepError(error_msg, resp, api_code=api_error_code)
tweepy.error.TweepError: [{u'message': u'Text parameter is missing.', u'code': 38}]
What will be mistake done by me...???
I also have another problem related to tweepy, How can I moved to page two or
got more followers in following code
i=1
user = API.get_user('Apple')
followers = API.followers(user.id,-1)
for follower in followers:
print follower,'\t',i
i=i+1
if I run the code I got only 5000 followers however if I use
`user.followers_count` it gives _362705_ followers (**This no. may be changes
while You check it**) How can I see remaining followers
Thank You... :)
Answer: To solve your first error, replace `re = API.send_direct_message(id, text)`
with `re = API.send_direct_message(id, text=text)`. This function only works
if you give it the message as a named parameter. The parameter name you need
here is "text", so you might want to change your variable name to avoid
confusion. Also, I just tried it, since you are sending a direct message, not
a tweet, it will **not** be truncated to only the first 140 characters.
For your second question, this should do the trick, as explained
[here](http://stackoverflow.com/questions/17431807/get-all-follower-ids-in-
twitter-by-tweepy):
followers = []
for page in tweepy.Cursor(API.followers, screen_name='Apple').pages():
followers.extend(page)
time.sleep(60) #this is here to slow down your requests and prevent you from hitting the rate limit and encountering a new error
print(followers)
|
Get "TypeError: cannot create mph' When Using sympy.nsolve()
Question: I am attempting to determine the time, angle and speed that something would
have to travel to intersect with a moving ellipse. (I actually want these
conditions for the minimum time). Right now I was trying to use Sympy to help
with this adventure. The following is the code which I am executing:
import sympy as sp
sp.init_printing()
delta_x, delta_y, t = sp.symbols('delta_x delta_y t', real=True, positive=True)
V, V_s, x_0, y_0, theta_s, theta_t = sp.symbols('V V_s x_0 y_0 theta_s theta_t', real=True)
x, y = sp.symbols('x y', real=True)
EQ1 = sp.Eq(((x-(x_0+V*sp.cos(theta_t)*t))/(delta_x+V*sp.cos(theta_t)*t))**2+((y-(y_0+V*sp.sin(theta_t)*t))/(delta_y+V*sp.sin(theta_t)*t))**2-1, 0)
sx = sp.Eq(x, V_s*sp.cos(theta_s)*t)
sy = sp.Eq(y, V_s*sp.sin(theta_s)*t)
mysubs = [(V,5), (x_0, 10), (y_0, 10), (theta_t, 7*(sp.pi/4)), (delta_x, 0), (delta_y, 0)]
sp.nsolve((EQ1.subs(mysubs), sx.subs(mysubs), sy.subs(mysubs)), (V_s, theta_s, t), (5, 0.0, 1))
The result of this operation yields:
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
/opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/calculus/optimization.py in findroot(ctx, f, x0, solver, tol, verbose, verify, **kwargs)
927 try:
--> 928 fx = f(*x0)
929 multidimensional = isinstance(fx, (list, tuple, ctx.matrix))
<string> in <lambda>(_Dummy_75, _Dummy_76, _Dummy_77)
/opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/matrices/matrices.py in __init__(self, *args, **kwargs)
300 for j, a in enumerate(row):
--> 301 self[i, j] = convert(a)
302 else:
/opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/ctx_mp_python.py in convert(ctx, x, strings)
661 return ctx.convert(x._mpmath_(prec, rounding))
--> 662 return ctx._convert_fallback(x, strings)
663
/opt/local/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/sympy/mpmath/ctx_mp.py in _convert_fallback(ctx, x, strings)
613 raise ValueError("can only create mpf from zero-width interval")
--> 614 raise TypeError("cannot create mpf from " + repr(x))
615
TypeError: cannot create mpf from 0.08*(x - 13.535533905932737622)**2 + 0.08*(y - 6.464466094067262378)**2 - 1
Is this because the system is not constrained enough? There is a family of
angles and velocities which could be used to intercept the moving ellipse.
This error does not seem to imply this. (Yes, even when I attempt to constrain
V_s in this problem the same error appears).
I am using the following versions of things:
| Software | Version |
|----------|-----------|
| python | 3.5.2 |
| sympy | 1.0 |
| mpmath | 0.19 |
Answer: The variables `x` and `y` are still symbolic in the equations. `nsolve`
requires all variables to be specified, and at least as many equations as
variables. So you either need to include values for them in `mysubs`, or else
solve for them (but to do that with `nsolve`, you will need two more
equations).
|
Python PYQT Tabs outside app window [why]
Question: Tabs added before app.exec_() is called look and act as any other tabs u met,
though if adding another after the app.exec_() call makes the new tab 'detach'
from the main app window. Pic below :)
Why? How can I make it move inside the window?
import threading
import time
import sys
from PyQt5.QtWidgets import QApplication
from PyQt5.QtWidgets import QFormLayout
from PyQt5.QtWidgets import QLineEdit
from PyQt5.QtWidgets import QTabWidget
from PyQt5.QtWidgets import QTextEdit
from PyQt5.QtWidgets import QWidget
class ATry(threading.Thread):
def __init__(self):
super().__init__()
def run(self):
time.sleep(1)
anotherTextEdit = QTextEdit()
anotherLineEdit = QLineEdit()
anotherLayout = QFormLayout()
anotherLayout.addRow(anotherTextEdit)
anotherLayout.addRow(anotherLineEdit)
anotherTab = QWidget()
anotherTab.setLayout(anotherLayout)
md.addTab(anotherTab, "Outside")
app = QApplication(sys.argv)
md = QTabWidget()
aTextEdit = QTextEdit()
aLineEdit = QLineEdit()
layout = QFormLayout()
layout.addRow(aTextEdit)
layout.addRow(aLineEdit)
thisTab = QWidget()
thisTab.setLayout(layout)
md.addTab(thisTab, "Inside")
a = ATry()
a.start()
md.show()
app.exec_()
[Screen describing the problem](http://i.stack.imgur.com/3gXA6.png)
Answer: It works with QTimer or signals:
import sys
import time
from PyQt5.QtCore import QObject
from PyQt5.QtCore import QThread
from PyQt5.QtCore import pyqtSignal
from PyQt5.QtWidgets import QApplication
from PyQt5.QtWidgets import QFormLayout
from PyQt5.QtWidgets import QLineEdit
from PyQt5.QtWidgets import QTabWidget
from PyQt5.QtWidgets import QTextEdit
from PyQt5.QtWidgets import QWidget
class ATry(QThread):
def __init__(self, pointer):
super().__init__()
self.pointer = pointer
def run(self):
time.sleep(2)
self.pointer.emit()
def addTheTab():
anotherTextEdit = QTextEdit()
anotherLineEdit = QLineEdit()
anotherLayout = QFormLayout()
anotherLayout.addRow(anotherLineEdit)
anotherLayout.addRow(anotherTextEdit)
anotherTab = QWidget()
anotherTab.setLayout(anotherLayout)
md.addTab(anotherTab, "Whatever")
class MyQObject(QObject):
trigger = pyqtSignal()
def __init__(self):
super().__init__()
def connect_and_get_trigger(self):
self.trigger.connect(addTheTab)
return self.trigger
def getGFX(self):
app = QApplication(sys.argv)
md = QTabWidget()
md.show()
return app, md
obj = MyQObject()
app, md = obj.getGFX()
addTheTab()
a = ATry(obj.connect_and_get_trigger())
a.start()
# timer = QTimer()
# timer.timeout.connect(proba)
# timer.start(3000)
app.exec_()
|
Read a complete data file and round numbers to 2 decimal places and save it with the same format
Question: I am trying to learn python and I have the intention to make the a very big
data file smaller and later do some statistical Analysis with R. I need to
read the data file (see below):
SCALAR
ND 3
ST 0
TS 10.00
0.0000
0.0000
0.0000
SCALAR
ND 3
ST 0
TS 3600.47
255.1744
255.0201
255.2748
SCALAR
ND 3
ST 0
TS 7200.42
255.5984
255.4946
255.7014
and find the numbers and round it in two digits after decimal, svae the
maximum number with the namber in front of TS. At the end save the data file
with the same format like following:
SCALAR
ND 3
ST 0
TS 10.00
0.00
0.00
0.00
SCALAR
ND 3
ST 0
TS 3600.47
255.17
255.02
255.27
SCALAR
ND 3
ST 0
TS**MAX** 7200.42
255.60
255.49
255.70
I have written a code like this:
import numpy as np
import matplotlib.pyplot as plt
import pickle
# Open file
f = open('data.txt', 'r')
thefile = open('output.txt', 'wb')
# Read and ignore header lines
header1 = f.readline()
header2 = f.readline()
header3 = f.readline()
header4 = f.readline()
data = []
for line in f:
line = line.strip()
columns = line.split()
source = {}
source['WSP'] = columns[0]
#source['timestep'] = float(columns[1])
source['timestep'] = columns[1]
data.append(source)
f.close()
but the number in front of TS cannot be read. I wanted to round the numbers
but the float which I used does not work. After that I wanted to put it in a
loop Any suggestion, Do I write the code in a good way? I will be very
thankfull for the help.
Answer: Try `"%.2f" % float(columns[1])` to round to two decimal places. Note that it
gives you a string, not a float. I don't understand the rest of what you're
asking.
`>>> "%.2f" % 255.5984`
`'255.60'`
|
Python - Removing vertical bar lines from histogram
Question: I'm wanting to remove the vertical bar outlines from my histogram plot, but
preserving the "etching" of the histogram, if that makes since.
import matplotlib.pyplot as plt
import numpy as np
bins = 35
fig = plt.figure(figsize=(7,6))
ax = fig.add_subplot(111)
ax.hist(subVel_Hydro1, bins=bins, facecolor='none',
edgecolor='black', label = 'Pecuiliar Vel')
ax.set_xlabel('$v_{_{B|A}} $ [$km\ s^{-1}$]', fontsize = 16)
ax.set_ylabel(r'$P\ (r_{_{B|A}} )$', fontsize = 16)
ax.legend(frameon=False)
Giving
[](http://i.stack.imgur.com/DtEHe.png)
Is this doable in matplotlibs histogram functionality? I hope I provided
enough clarity.
Answer: In `pyplot.hist()` you could set the value of `histtype = 'step'`. Example
code:
import matplotlib as mpl
import matplotlib.pyplot as plt
import numpy as np
x = np.random.normal(0,1,size=1000)
fig = plt.figure()
ax = fig.add_subplot(111)
ax.hist(x, bins=50, histtype = 'step', fill = None)
plt.show()
Sample output:
[](http://i.stack.imgur.com/621Vo.png)
|
when i made a 3 handshake with ubuntu in VMware return package R
Question:
#!/usr/bin/python
from scapy.all import *
def findWeb():
a = sr1(IP(dst="8.8.8.8")/UDP()/DNS(qd=DNSQR(qname="www.google.com")),verbose=0)
return a[DNSRR].rdata
def sendPacket(dst,src):
ip = IP(dst = dst)
SYN = TCP(sport=1500, dport=80, flags='S')
SYNACK = sr1(ip/SYN)
my_ack = SYNACK.seq + 1
ACK = TCP(sport=1050, dport=80, flags='A', ack=my_ack)
send(ip/ACK)
payload = "stuff"
PUSH = TCP(sport=1050, dport=80, flags='PA', seq=11, ack=my_ack)
send(ip/PUSH/payload)
http = sr1(ip/TCP()/'GET /index.html HTTP/1.0 \n\n',verbose=0)
print http.show()
src = '10.0.0.24'
dst = findWeb()
sendPacket(dst,src)
I'm trying to do HTTP packets with SCAPY I am using UBUNTU on VMwaer
The problem is that every time I send messages I have RESET How do we fix it?
Thanks
[sniff package image](http://i.stack.imgur.com/0eJdj.png)
Answer: Few things I notice wrong. 1\. You have your sequence number set statically
(seq=11) which is wrong. Sequence numbers are always randomly generated and
they must be used as per RFC793. So the sequence should be = SYNACK[TCP].ack
2. You set your source port as 1500 during SYN packet, but then you use it as 1050 (typo?)
3. You don't need extra payload/PUSH.
Also, have a look at these threads:
[How to create HTTP GET request
Scapy?](http://stackoverflow.com/questions/37683026/how-to-create-http-get-
request-scapy)
[Python-Scapy or the like-How can I create an HTTP GET request at the packet
level](http://stackoverflow.com/questions/4750793/python-scapy-or-the-like-
how-can-i-create-an-http-get-request-at-the-packet-leve)
|
Complete Algebraic Equations
Question: Im trying to build an app that graphs an equation based on user input. The
equation would be in slope intercept form: `y = mx + b`, for `m` as slope and
`b` as `y intercept`. However, this isn't working for me in python!
I tried this:
>>> x = 3
>>> 1/2x
and was returned this:
File "<stdin>", line 1
1/2x
^
SyntaxError: invalid syntax
>>>
How would I make it so that this returns 1.5?
Answer:
import re
mxb = re.compile('''(?P<slope>[(\d)/.]*)x\s+(?P<sign>[+-])\s+(?P<intercept>[\d./]*)''')
s = '1/2x + 5'
match = mxb.match(s)
if match :
print( match.groupdict() )
test the python regex here :
[pythex.org](http://pythex.org/?regex=\(%3FP%3Cslope%3E%5B\(%5Cd\)%2F.%5D*\)x%5Cs\(%3FP%3Csign%3E%5B%2B-%5D\)%5Cs\(%3FP%3Cintercept%3E%5B%5Cd.%2F%5D*\)&test_string=x%20%2B%205%0A1%2F2x%20-%202.3%0A.5x%20%2B%201.4&ignorecase=0&multiline=0&dotall=0&verbose=0
"pythex.org")
|
Python bokeh apply hovertools only on model not on figure
Question: I want to have a scatter plot and a (base)line on the same figure. And I want
to use `HoverTool` only on the circles of scatter but not on the line. Is it
possible?
With the code below I get tooltips with `index: 0` and `(x, y): (???, ???)`
when I hover on the line (any part of the line). But the `index: 0` data in
`source` is totally different (`(x, y): (1, 2)`)...
df = pd.DataFrame({'a':[1, 3, 6, 9], 'b':[2, 3, 5, 8]})
from bokeh.models import HoverTool
import bokeh.plotting as bplt
TOOLS = ['box_zoom', 'box_select', 'wheel_zoom', 'reset', 'pan', 'resize', 'save']
source = bplt.ColumnDataSource(data=df)
hover = HoverTool(tooltips=[("index", "$index"), ("(x, y)", "(@a, @b)")])
p = bplt.figure(plot_width=600, plot_height=600, tools=TOOLS+[hover],
title="My sample bokeh plot", webgl=True)
p.circle('a', 'b', size=10, source=source)
p.line([0, 10], [0, 10], color='red')
bplt.save(p, 'c:/_teszt.html')
Thank you!!
Answer: To limit which renderers you want the HoverTool is active on (by default it's
active on all) you can either set a `name` attr on your glyphs, then specify
which names you want your HoverTool to be active on:
p.circle('a', 'b', size=10, name='circle', source=source)
hover = HoverTool(names=['circle'])
docs:
<http://bokeh.pydata.org/en/latest/docs/reference/models/tools.html#bokeh.models.tools.HoverTool.names>
or you can add the renderers to the HoverTool.
circle = p.circle('a', 'b', size=10, source=source)
hover = HoverTool(renderers=['circle'])
docs:
<http://bokeh.pydata.org/en/latest/docs/reference/models/tools.html#bokeh.models.tools.HoverTool.renderers>
|
How to see the plot made in python using pandas and matplotlib
Question: I am following the tutorial <http://ahmedbesbes.com/how-to-score-08134-in-
titanic-kaggle-challenge.html> and following is my code
from IPython.core.display import HTML
HTML("""
<style>
.output_png {
display: table-cell;
text-align: center;
vertical-align: middle;
}
</style>
""")
# remove warnings
import warnings
warnings.filterwarnings('ignore')
# ---
import pandas as pd
pd.options.display.max_columns = 100
import matplotlib
matplotlib.use('TkAgg')
import matplotlib.pyplot as plt
matplotlib.style.use('ggplot')
import numpy as np
pd.options.display.max_rows = 100
data = pd.read_csv('./data/train.csv')
data.head()
data['Age'].fillna(data['Age'].median(), inplace=True)
survived_sex = data[data['Survived']==1]['Sex'].value_counts()
dead_sex = data[data['Survived']==0]['Sex'].value_counts()
df = pd.DataFrame([survived_sex,dead_sex])
df.index = ['Survived','Dead']
df.plot(kind='bar',stacked=True, figsize=(15,8))
I dont actually see the plot. How do i see the plot?
[](http://i.stack.imgur.com/g24E6.png)
Answer: Try using `plt.show()` at the end.
EDIT:
Also, you may need to add `%matplotlib inline` as explained here: [How to make
IPython notebook matplotlib plot
inline](http://stackoverflow.com/questions/19410042/how-to-make-ipython-
notebook-matplotlib-plot-inline)
|
AttributeError: type object has no attribute "id" PYTHON
Question: So I was trying to make a basic python pong game when this error came up: It
seems to say that AttributeError: type object has no attribute "id" which I
have no idea what it means.
C:\Users\****\AppData\Local\Programs\Python\Python35-32\python.exe C:/Users/****/untitled/src/testing.py
Traceback (most recent call last):
File "C:/Users/****/untitled/src/testing.py", line 113, in <module>
ball.draw()
File "C:/Users/****/untitled/src/testing.py", line 36, in draw
if self.hit_paddle2(pos) == True:
File "C:/Users/****/untitled/src/testing.py", line 53, in hit_paddle2
paddle2_pos = self.canvas.coords(self.paddle2.id)
AttributeError: type object 'Paddle2' has no attribute 'id'
Process finished with exit code 1
This is the code I used:
from tkinter import *
import random
import time
tk = Tk()
tk.title("Game")
tk.resizable(0, 0)
tk.wm_attributes("-topmost", 1)
canvas = Canvas(tk, width=500, height=400, bd=0, highlightthickness=0)
canvas.pack()
tk.update()
class Ball:
def __init__(self, canvas, paddle, paddle2, color):
self.canvas = canvas
self.paddle = paddle
self.paddle2 = paddle2
self.id = canvas.create_oval(10, 10, 25, 25, fill=color)
self.canvas.move(self.id, 245, 100)
starts = [-3, -2, -1, 1, 2, 3]
random.shuffle(starts)
self.x = starts[0]
self.y = -3
self.canvas_height = self.canvas.winfo_height()
self.canvas_width = self.canvas.winfo_width()
self.hit_bottom = False
def draw(self):
self.canvas.move(self.id, self.x, self.y)
pos = self.canvas.coords(self.id)
if pos[1] <= 0:
self.y = 3
if self.hit_paddle(pos) == True:
self.y = -3
if self.hit_paddle2(pos) == True:
self.y = 3
if pos[3] >= self.canvas_height:
self.hit_bottom = True
if pos[0] <= 0:
self.x = 3
if pos[2] >= self.canvas_width:
self.x = -3
def hit_paddle(self, pos):
paddle_pos = self.canvas.coords(self.paddle.id)
if pos[2] >= paddle_pos[0] and pos[0] <= paddle_pos[2]:
if pos[3] >= paddle_pos[1] and pos[3] <= paddle_pos[3]:
return True
return False
def hit_paddle2(self, pos):
paddle2_pos = self.canvas.coords(self.paddle2.id)
if pos[2] >= paddle2_pos[0] and pos[0] <= paddle2_pos[2]:
if pos[3] >= paddle2_pos[1] and pos[3] <= paddle2_pos[3]:
return True
return False
class Paddle:
def __init__(self, canvas, color):
self.canvas = canvas
self.id = canvas.create_rectangle(0, 0, 100, 10, fill=color)
self.canvas.move(self.id, 200, 300)
self.x = 0
self.canvas_width = self.canvas.winfo_width()
self.canvas.bind_all('<KeyPress-Left>', self.turn_left)
self.canvas.bind_all('<KeyPress-Right>', self.turn_right)
def turn_left(self, evt):
self.x = -2
def turn_right(self, evt):
self.x = 2
def draw(self):
self.canvas.move(self.id, self.x, 0)
pos = self.canvas.coords(self.id)
if pos[0] <= 0:
self.x = 0
elif pos[2] >= self.canvas_width:
self.x = 0
class Paddle2:
def __init__(self, canvas, color):
self.canvas = canvas
self.id = canvas.create_rectangle(0, 10, 100, 10, fill=color)
self.canvas.move(self.id, 200, 300)
self.x = 0
self.canvas_width = self.canvas.winfo_width()
self.canvas.bind_all('<KeyPress-W>', self.turn_left)
self.canvas.bind_all('<KeyPress-A>', self.turn_right)
def turn_left(self, evt):
self.x = -2
def turn_right(self, evt):
self.x = 2
def draw(self):
self.canvas.move(self.id, self.x, 0)
pos = self.canvas.coords(self.id)
if pos[0] <= 0:
self.x = 0
elif pos[2] >= self.canvas_width:
self.x = 0
paddle = Paddle(canvas, 'blue')
ball = Ball(canvas, paddle, Paddle2, 'red')
paddle2 = Paddle2(canvas, 'blue')
while 1:
if ball.hit_bottom == False:
ball.draw()
paddle.draw()
paddle2.draw()
tk.update_idletasks()
tk.update()
time.sleep(0.01)
Answer: You need to pass instances of `Paddle` and `Paddle2` to `Ball` constructor.
paddle = Paddle(canvas, 'blue')
paddle2 = Paddle2(canvas, 'blue')
ball = Ball(canvas, paddle, paddle2, 'red')
|
Edited - Python plot persistence windows
Question: Each time I launch my program, my plots are erased after each execution.
I would like the following situation:
1. Launch program 1 and plot in figure 1
2. Stop the execution of program 1
3. Lauch program 2 and plot in figure 1
4. Retrieve a pdf file where the plot of program 1 and program 2 are superimposed
here is my code:
z = np.zeros(5)
fig,axarr = plt.subplots(nrows=2, ncols=3)
fig.subplots_adjust()
lig,col = 0,0
axarr[lig,col].plot(tt,z[0,:],'k')
#plt.ion()
#plt.draw()
plt.show(block=True)
#plt.show()
**EDIT:** Ok maybe I could solve this problem with the following approach: Is
it possible to get the ax of a figure that was opened by an other excetion of
of program ?
Answer: `plt.show(block=True)` should work.
Other hack would be using pause function as shown below:
from matplotlib import pylab
pylab.plot(range(10), range(10))
pylab.pause(2**31-1)
|
Iterating across multiple columns in Pandas DF and slicing dynamically
Question: **TLDR:** How to iterate across all options of multiple columns in a pandas
dataframe without specifying the columns or their values explicitly?
**Long Version:** I have a pandas dataframe that looks like this, only it has
a lot more features or drug dose combinations than are listed here. Instead of
just 3 types of features, it could have something like 70...:
> dosage_df
First Score Last Score A_dose B_dose C_dose
22 28 1 40 130
55 11 2 40 130
15 72 3 40 130
42 67 1 90 130
90 74 2 90 130
87 89 3 90 130
14 43 1 40 700
12 61 2 40 700
41 5 3 40 700
Along with my data frame, I also have a python dictionary with the relevant
ranges for each feature. The keys are the feature names, and the different
values which it can take are the keys:
> dict_of_dose_ranges = {'A_dose': [1, 2, 3], 'B_dose': [40, 90], 'C_dose': [130,700]}
For my purposes, I need to generate a particular combination (say A_dose = 1,
B_dose = 90, and C_dose = 700), and based on those settings take the relevant
slice out of my dataframe, and do relevant calculations from that smaller
subset, and save the results somewhere.
I need to do this for ALL possible combinations of ALL of my features (far
more than the 3 which are here, and which will be variable in the future).
In this case, I could easily pop this into SkLearn's Parameter grid, generate
the options:
> from sklearn.grid_search import ParameterGrid
> all_options = list(ParameterGrid(dict_of_dose_ranges))
> all_options
and get:
[{'A_dose': 1, 'B_dose': 40, 'C_dose': 130},
{'A_dose': 1, 'B_dose': 40, 'C_dose': 700},
{'A_dose': 1, 'B_dose': 90, 'C_dose': 130},
{'A_dose': 1, 'B_dose': 90, 'C_dose': 700},
{'A_dose': 2, 'B_dose': 40, 'C_dose': 130},
{'A_dose': 2, 'B_dose': 40, 'C_dose': 700},
{'A_dose': 2, 'B_dose': 90, 'C_dose': 130},
{'A_dose': 2, 'B_dose': 90, 'C_dose': 700},
{'A_dose': 3, 'B_dose': 40, 'C_dose': 130},
{'A_dose': 3, 'B_dose': 40, 'C_dose': 700},
{'A_dose': 3, 'B_dose': 90, 'C_dose': 130},
{'A_dose': 3, 'B_dose': 90, 'C_dose': 700}]
**This is where I run into problems:**
**Problem #1)** I can now iterate across `all_options`, but I'm not sure how
to now SELECT out of my `dosage_df` from each of the dictionary options (i.e.
{'A_dose': 1, 'B_dose': 40, 'C_dose': 130}) WITHOUT doing it explicitly.
In the past, I could do something like:
dosage_df[(dosage_df.A_dose == 1) & (dosage_df.B_dose == 40) & (dosage_df.C_dose == 130)]
First Score Last Score A_dose B_dose C_dose
0 22 28 140 130
But now I'm not sure what to put inside the brackets to slice it
dynamically...
dosage_df[?????]
**Problem #2)** When I actually enter in my full dictionary of features with
their respective ranges, I get an error because it deems it as having too many
options...
from sklearn.grid_search import ParameterGrid
all_options = list(ParameterGrid(dictionary_of_features_and_ranges))
all_options
---------------------------------------------------------------------------
OverflowError Traceback (most recent call last)
<ipython-input-138-7b73d5e248f5> in <module>()
1 from sklearn.grid_search import ParameterGrid
----> 2 all_options = list(ParameterGrid(dictionary_of_features_and_ranges))
3 all_options
OverflowError: long int too large to convert to int
I tried a number of alternate approaches including using double while loops, a
[tree / recursion method from
here](http://stackoverflow.com/questions/23986892/python-recursive-iteration-
exceeding-limit-for-tree-implementation), another [recursion method from
here](http://stackoverflow.com/questions/13109274/python-recursion-
permutations), but it wasn't coming together.... Any help is much appreciated.
Answer: You can use
[`itertools.product`](https://docs.python.org/3/library/itertools.html#itertools.product)
to generate all possible dosage combinations, and
[`DataFrame.query`](http://pandas.pydata.org/pandas-
docs/stable/generated/pandas.DataFrame.query.html) to do the selection:
from itertools import product
for dosage_comb in product(*dict_of_dose_ranges.values()):
dosage_items = zip(dict_of_dose_ranges.keys(), dosage_comb)
query_str = ' & '.join('{} == {}'.format(*x) for x in dosage_items)
sub_df = dosage_df.query(query_str)
# Do Stuff...
|
Adding exception to "AttributeError" python
Question: So, I have some tweets with some special characters and shapes. I am trying to
find a word in those tweets by converting them to lower case. The function
throws an "AttributeError" when it encounters those special characters and
hence, I want to change my function in a way that it skips those records and
processes others.
Can I add exception to "AttributeError" in python. I want it to act more like
an "iferror resume next"/Error handling statement.
I am currently using:-
def word_in_text(word, text):
try:
print text
word = word.lower()
text = text.lower()
match = re.search(word, text)
if match:
return True
else:
return False
except(AttributeError, Exception) as e:
continue
error post using @galah92 recommendations :-
Traceback (most recent call last):
File "<input>", line 1, in <module>
File "C:\Python27\lib\site-packages\pandas\core\series.py", line 2220, in apply
mapped = lib.map_infer(values, f, convert=convert_dtype)
File "pandas\src\inference.pyx", line 1088, in pandas.lib.map_infer (pandas\lib.c:63043)
File "<input>", line 1, in <lambda>
File "<input>", line 3, in word_in_text
File "C:\Python27\lib\re.py", line 146, in search
return _compile(pattern, flags).search(string)
TypeError: expected string or buffer
I am new to Python and self learning it. Any help will be really appreciated.
Answer: You can use `re.IGNORECASE` flag when you `search()`.
That way you don't need to deal with `lower()` or exceptions.
def word_in_text(word, text):
print text
if re.search(word, text, re.IGNORECASE):
return True
else:
return False
* * *
As an example, if I run:
from __future__ import unicode_literals # see edit notes
import re
text = "CANCION! You &"
word = "you"
def word_in_text(word, text):
print(text)
if re.search(word, text, re.IGNORECASE):
return True
else:
return False
print(word_in_text(word, text))
The output is:
CANCION! You &
True
* * *
**EDIT**
For Python 2, you should add `from __future__ import unicode_literals` at the
top of your script to make sure you encode everything to UTF-8.
You can read more about it
[here](http://stackoverflow.com/questions/23370025/what-is-unicode-literals-
used-for).
|
<urlopen error (-1, 'SSL exception: Differences between the SSL socket behaviour of cpython vs. jython are explained on the wiki
Question: I'm using the following code.
import urllib2
#Setting proxy
myProxy = {'https':'https://proxy.example.com:8080'}
proxy = urllib2.ProxyHandler(myProxy)
opener = urllib2.build_opener(proxy)
urllib2.install_opener(opener)
#Access URL
auth_token = "realLy_loNg_AuTheNicaTion_toKen"
headers={"Content-Type":"application/json",
"charset": "UTF-8",
"Authorization": "Bearer %s" %auth_token}
url = "https://example.com/something"
req = urllib2.Request(url, None, headers)
reply = urllib2.urlopen(req)
print reply.getcode()
I'm running the above as a Jython script in nGrinder. When I run the same
script on my system with Jython, it works fine and returns 200(OK status
code). When I run it on nGrinder, I get the error
(-1, 'SSL exception: Differences between the SSL socket behaviour of cpython vs. jython are explained on the wiki: http://wiki.python.org/jython/NewSocketModule#SSL_Support')
Any ideas why this is happening?
EDIT: I've been trying and the issue is definitely with the long
authentication token. I have a feeling it might be some encoding issue. There
was a similar question posted
[here](http://stackoverflow.com/questions/33912074/https-get-using-jython)
before. I read it but it wasn't described properly. But it could be a good
reference to think off of.
Answer: It seems issue of certificate trust. <http://bugs.jython.org/issue1688>
There are these two options that will solve your problem.
> Jython is not like cpython. Cpython does not verify the chain of trust for
> server certificates. Jython does verify the chain of trust, and will refuse
> to open the connection if it cannot verify the server.
>
> So you have two options.
>
> 1. Disable certificate checking on jython
>
>
> <http://jython.xhaus.com/installing-an-all-trusting-security-provider-on-
> java-and-jython/> <http://tech.pedersen-live.com/2010/10/trusting-all-
> certificates-in-jython/>
>
> 2. Add your (self-signed?) certificate to your local java trust store, so
> that your client will trust your server.
>
>
> Google("java install self-signed certificate")
Another reason for this can be due to incompactible version of python and java
(<https://github.com/geoscript/geoscript-py/issues/32>)
> I just tried with Jython 2.5.3 and JDK 1.7 on Mac and was able to run
> ez_setup.py.
>
> Java version: [Java HotSpot(TM) 64-Bit Server VM (Oracle Corporation)] on
> java1.7.0_51
>
> Swapping to Oracle JDK 7 did the trick.
Please check if this solves the problem, as per implementation of jython and
error, issue is due to HTTPS(SSL layer)
<https://github.com/int3/jython/blob/master/Lib/socket.py>
# These error codes are currently wrong: getting them correct is going to require
# some investigation. Cpython 2.6 introduced extensive SSL support.
(javax.net.ssl.SSLException, ALL) : lambda x: sslerror(-1, 'SSL exception'+_ssl_message),
|
Extracting hand writing text out in shape with OpenCV
Question: I am very new to OpenCV Python and I really need some help here.
So what I am trying to do here is to extract out these words in the image
below.
[](http://i.stack.imgur.com/XsSOP.jpg)
The words and shapes are all hand drawn, so they are not perfect. I have did
some coding below.
First of all, I grayscale the image
img_final = cv2.imread(file_name)
img2gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)
Then I use THRESH_INV to show the content
ret, new_img = cv2.threshold(image_final, 100 , 255, cv2.THRESH_BINARY_INV)
After which, I dilate the content
kernel = cv2.getStructuringElement(cv2.MORPH_CROSS,(3 , 3))
dilated = cv2.dilate(new_img,kernel,iterations = 3)
I dilate the image is because I can identify text as one cluster
After that, I apply boundingRect around the contour and draw around the
rectangle
contours, hierarchy = cv2.findContours(dilated,cv2.RETR_EXTERNAL,cv2.CHAIN_APPROX_NONE) # get contours
index = 0
for contour in contours:
# get rectangle bounding contour
[x,y,w,h] = cv2.boundingRect(contour)
#Don't plot small false positives that aren't text
if w < 10 or h < 10:
continue
# draw rectangle around contour on original image
cv2.rectangle(img,(x,y),(x+w,y+h),(255,0,255),2)
This is what I got after that.
[](http://i.stack.imgur.com/vxKYi.png)
I am only able to detect one of the text. I have tried many other methods but
this is the closet results I have got and it does not fulfill the requirement.
The reason for me to identify the text is so that I can get the X and Y
coordinate of each of the text in this image by putting a bounding Rectangle
"boundingRect()".
Please help me out. Thank you so much
Answer: You can use the fact that the connected component of the letters are much
smaller than the large strokes of the rest of the diagram.
I used opencv3 connected components in the code but you can do the same things
using findContours.
The code:
import cv2
import numpy as np
# Params
maxArea = 150
minArea = 10
# Read image
I = cv2.imread('i.jpg')
# Convert to gray
Igray = cv2.cvtColor(I,cv2.COLOR_RGB2GRAY)
# Threshold
ret, Ithresh = cv2.threshold(Igray,0,255,cv2.THRESH_BINARY_INV+cv2.THRESH_OTSU)
# Keep only small components but not to small
comp = cv2.connectedComponentsWithStats(Ithresh)
labels = comp[1]
labelStats = comp[2]
labelAreas = labelStats[:,4]
for compLabel in range(1,comp[0],1):
if labelAreas[compLabel] > maxArea or labelAreas[compLabel] < minArea:
labels[labels==compLabel] = 0
labels[labels>0] = 1
# Do dilation
se = cv2.getStructuringElement(cv2.MORPH_ELLIPSE,(25,25))
IdilateText = cv2.morphologyEx(labels.astype(np.uint8),cv2.MORPH_DILATE,se)
# Find connected component again
comp = cv2.connectedComponentsWithStats(IdilateText)
# Draw a rectangle around the text
labels = comp[1]
labelStats = comp[2]
#labelAreas = labelStats[:,4]
for compLabel in range(1,comp[0],1):
cv2.rectangle(I,(labelStats[compLabel,0],labelStats[compLabel,1]),(labelStats[compLabel,0]+labelStats[compLabel,2],labelStats[compLabel,1]+labelStats[compLabel,3]),(0,0,255),2)
[](http://i.stack.imgur.com/k0UdG.jpg)
|
Reshaping OpenCV Image (numpy) Dimensions
Question: I need to convert an image in a numpy array loaded via cv2 into the correct
format for the deep learning library mxnet for its convolutional layers.
My current images are shaped as follows: (256, 256, 3), or (height, width,
channels).
From what I've been told, this actually needs to be (3, 256, 256), or
(channels, height, width).
Unfortunately, my knowledge of numpy/python opencv isn't good enough to know
how to manipulate the arrays correctly.
I've figured out that I could split the arrays into channels by cv2.split, but
I'm uncertain of how to combine them again in the right format (I don't know
if using cv2.split is optimal, or if there are better ways in numpy).
Thanks for any help.
Answer: You can use
[`numpy.rollaxis`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.rollaxis.html)
as follow: If your `image` as shape `(height, width, channels)`
import numpy as np
new_shaped_image = np.rollaxis(image, axis=2, start=0)
This means that the `2`nd axis of the `new_shaped_image` will be at `0` spot.
So `new_shaped_image.shape` will be `(channels, height, width)`
|
How to fetch JSON data from API, format / encode / write to a file?
Question: I need to fetch some data from a weather API, extract certain info and send it
to std. output (in my case this is the console/terminal; I am playing around
with python API scripting and do not yet have a web site/app do show fetched
data).
**Example Python code from the API provider (simple to understand and
works):**
import urllib2
import json
API_KEY='mysuperawesomekey'
f = urllib2.urlopen('http://api.wunderground.com/api/' + API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json')
json_string = f.read()
parsed_json = json.loads(json_string)
location = parsed_json['location']['city']
temp_f = parsed_json['current_observation']['temp_f']
print "Current temperature in %s is: %s" % (location, temp_f)
f.close()
Since I am on a free plan, I don't want to use up my API requests but rather
fetch the JSON data, save it to a file and use in another script to practice.
I found a couple of solutions here on
[StackOverflow](http://stackoverflow.com/questions/12309269/how-do-i-write-
json-data-to-a-file-in-python) but neither seem to work completely (nice
formatting of the file):
**Attempt 1. to save the fetched data, add. to the orig. code above:**
import io
import codecs
with open('data.json', 'w') as outfile:
json.dump(json_string, outfile, indent=4, sort_keys=True, separators=(",", ':'))
**Attempt 2:**
with codecs.open('data.json', 'w', 'utf8') as jasonfile:
jasonfile.write(json.dumps(parsed_json, sort_keys = True, ensure_ascii=False))
Both of my attempts work ("kind of") as I do get a **.json** file. But, upon
inspecting it in my editor (Atom) I am seeing this (first few soft-break
lines):
**Output:**
"\n{\n \"response\": {\n \"version\":\"0.1\",\n\"termsofService\":\"http://www.wunderground.com/weather/api/d/terms.html\",\n \"features\": {\n \"geolookup\": 1\n ,\n \"conditions\": 1\n }\n\t}\n\t\t,\t\"location\": {\n\t\t\"type\":\"CITY\",\n\t\t\"country\":\"US\",\n\t\t\"country_iso3166\":\"US\",\n\t\t\"country_name\":\"USA\",\n\t\t\"state\":\"IA\",\n\t\t\"city\":\"Cedar Rapids\",\n\t\t\"tz_short\":\"CDT\",\n\t\t\"tz_long\":\"America/Chicago\",\n\t\t\"lat\":\"41.97171021\",\n\t\t\"lon\":\"-91.65871429\",\n\t\t\"zip\":\...
It is all on one line, with newlines and tabs showing. I have a couple of
questions:
1. How can I check what format / kind of data-type API is returning? (FOr example is it just a raw string which needs to be JSON-iffied?)
2. Is my JSON not written in a human readable format because of encoding, delimiters fault?
3. Line `json_string = f.read()` from example code seems to "mimic" working with a "real" internal file object, is this correct?
Many thanks!
Answer: Seems that it was quite a simple solution. In my original code, I was saving a
"non-parsed" variable to a file:
import urllib2
import json
API_KEY='key'
f = urllib2.urlopen('http://api.wunderground.com/api/'
+ API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json')
# Saving the below variable into a .json file gives all the JSON data
# on a single line
json_string = f.read()
with open('data.json', 'w') as outfile:
json.dump(json_string, outfile, indent=4, sort_keys=True, separators=(",", ':'))
which, of course, resulted in a file with JSON data on a single line. What I
should have been doing is saving the **parsed** variable:
import urllib2
import json
API_KEY='key'
f = urllib2.urlopen('http://api.wunderground.com/api/' + API_KEY + '/geolookup/conditions/q/IA/Cedar_Rapids.json')
json_string = f.read()
# THIS one is correct!
parsed_json = json.loads(json_string)
with open('data.json', 'w') as outfile:
json.dump(parsed_json, outfile, indent=4, sort_keys=True, separators=(",", ':'))
|
Python won't print expression
Question: So, I'm kind of new to programming and have been trying Python. I'm doing a
really simple program that converts usd to euroes.
This is the text of the problem that I'm trying to solve
> You are going to travel to France. You will need to convert dollars to euros
> (the currency of the European Union). There are two currency exchange
> booths. Each has a display that shows CR: their conversion rate as euros per
> dollar and their fee as a percentage. The fee is taken before your money is
> converted. Which booth will give you the most euros for your dollars, how
> many euros, and how much is the difference. Example 1: Dollars: 200 CR1:
> 0.78 Fee: 1 (amount 152.88 euros) CR2: 0.80 Fee: 3 (amount 155.2 euros)
> Answer: 2 is the best; difference is 2.32 euros; 155.2 euros
And here is my code
from __future__ import division
usd = int(input("How much in USD? "))
cr1 = int(input("What is the convertion rate of the first one? "))
fee1 = int(input("What is the fee of the first one? "))
cr2 = int(input("What is the convertion rate of the second one? "))
fee2 = int(input("What is the fee of the second one? "))
def convertion (usd, cr, fee):
usdwfee = usd - fee
convert = usdwfee * cr
return convert
first = convertion(usd, cr1, fee1)
second = convertion(usd, cr2, fee2)
fs = first - second
sf = second - first
def ifstatements (first,second,fs,sf):
if first < second:
print "1 is the best; difference is ",fs," euroes. 2 converts to ",first," euroes."
elif first > second:
print "2 is the best; difference is",sf," euroes. 2 converts to", second," euroes."
ifstatements(first, second, fs, sf)
The problem is that when I run the program it won't print out. It just takes
my input and doesn't output anything.
Answer: Check your logic more.
`cr1 = int(input("What is the convertion rate of the first one? "))`
Your conversion rate is in int. As in Integer which means it can't have a
floating point (a decimal "CR1: 0.78" from your example). Your cr1 will become
0 if you cast it into an int. Also change your dollar and fees to accept
floats too since I'm assuming you want to deal with cents too
So change:
usd = float(input("How much in USD? "))
cr1 = float(input("What is the convertion rate of the first one? "))
fee1 = float(input("What is the fee of the first one? "))
cr2 = float(input("What is the convertion rate of the second one? "))
fee2 = float(input("What is the fee of the second one? "))
And it should work.
|
Can't login to a specific ASP.NET website using python requests
Question: So I've been trying for the last 6 hours to make this work, but I couldn't and
endless searches didn't help, So I guess I'm either doing something very
fundamental wrong, or it's just a trivial bug which happens to match my logic
so I need extra eyes to help me fix it.
The website url is [this](https://www.wes.org/appstatus/).
I wrote a piece of messy python code to just login and read the next page, but
All I get is a nasty 500 error saying something on the server went wrong
processing my request.
Here is the request made by a browser which works just fine, no problem.
HTTP Response code to this request is 302 (Redirect)
POST /appstatus/index.aspx HTTP/1.1
Host: www.wes.org
Connection: close
Content-Length: 303
Cache-Control: max-age=0
Origin: https://www.wes.org
Upgrade-Insecure-Requests: 1
User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36
Content-Type: application/x-www-form-urlencoded
Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8
Referer: https://www.wes.org/appstatus/index.aspx
Accept-Encoding: gzip, deflate, br
Accept-Language: en-US,en;q=0.8,fa;q=0.6
Cookie: ASP.NET_SessionId=bu2gemmlh3hvp4f5lqqngrbp; _ga=GA1.2.1842963052.1473348318; _gat=1
__VIEWSTATE=%2FwEPDwUKLTg3MTMwMDc1NA9kFgICAQ9kFgICAQ8PFgIeBFRleHRkZGRk9rP20Uj9SdsjOKNUBlbw55Q01zI%3D&__VIEWSTATEGENERATOR=189D346C&__EVENTVALIDATION=%2FwEWBQK6lf6LBAKf%2B9bUAgK9%2B7qcDgK8w4S2BALowqJjoU1f0Cg%2FEAGU6r2IjpIPG8BO%2BiE%3D&txtUID=Email%40Removed.com&txtPWD=PASSWORDREMOVED&Submit=Log+In&Hidden1=
and this one is the request made by my script.
POST /appstatus/index.aspx HTTP/1.1
Host: www.wes.org
Connection: close
Accept-Encoding: gzip, deflate, br
User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36
Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8
Upgrade-Insecure-Requests: 1
Content-Type: application/x-www-form-urlencoded
Origin: https://www.wes.org
Accept-Language: en-US,en;q=0.8,fa;q=0.6
Cache-Control: max-age=0
Referer: https://www.wes.org/appstatus/indexca.aspx
Cookie: ASP.NET_SessionId=nxotmb55jjwf5x4511rwiy45
Content-Length: 303
txtPWD=PASSWORDREMOVED&Submit=Log+In&__EVENTVALIDATION=%2FwEWBQK6lf6LBAKf%2B9bUAgK9%2B7qcDgK8w4S2BALowqJjoU1f0Cg%2FEAGU6r2IjpIPG8BO%2BiE%3D&txtUID=Email%40Removed.com&__VIEWSTATE=%2FwEPDwUKLTg3MTMwMDc1NA9kFgICAQ9kFgICAQ8PFgIeBFRleHRkZGRk9rP20Uj9SdsjOKNUBlbw55Q01zI%3D&Hidden1=&__VIEWSTATEGENERATOR=189D346C
And this is the script making the request, I'm sorry if it's so messy, just
need something quick.
import requests
import bs4
import urllib.parse
def main():
session = requests.Session()
headers = {"Origin": "https://www.wes.org",
"Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8",
"Cache-Control": "max-age=0", "Upgrade-Insecure-Requests": "1", "Connection": "close",
"User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36",
"Referer": "https://www.wes.org/appstatus/indexca.aspx", "Accept-Encoding": "gzip, deflate, br",
"Accept-Language": "en-US,en;q=0.8,fa;q=0.6", "Content-Type": "application/x-www-form-urlencoded"}
r = session.get('https://www.wes.org/appstatus/index.aspx',headers=headers)
cookies = r.cookies
soup = bs4.BeautifulSoup(r.content, "html5lib")
viewState=urllib.parse.quote(str(soup.select('#__VIEWSTATE')[0]).split('value="')[1].split('"/>')[0])
viewStateGenerator=urllib.parse.quote(str(soup.select('#__VIEWSTATEGENERATOR')[0]).split('value="')[1].split('"/>')[0])
eventValidation=urllib.parse.quote(str(soup.select('#__EVENTVALIDATION')[0]).split('value="')[1].split('"/>')[0])
paramsPost = {}
paramsPost.update({'__VIEWSTATE':viewState})
paramsPost.update({'__VIEWSTATEGENERATOR':viewStateGenerator})
paramsPost.update({'__EVENTVALIDATION':eventValidation})
paramsPost.update({"txtUID": "[email protected]"})
paramsPost.update({"txtPWD": "My_So_Called_Password"})
paramsPost.update({"Submit": "Log In"})
paramsPost.update({"Hidden1": ""})
response = session.post("https://www.wes.org/appstatus/index.aspx", data=paramsPost, headers=headers,
cookies=cookies)
print("Status code:", response.status_code) #Outputs 500.
#print("Response body:", response.content)
if __name__ == '__main__':
main()
Any help would be so much appreciated.
Answer: You are doing way too much work and in doing so not passing valid data,you
extract value attribute directly i.e
`.select_one('#__VIEWSTATEGENERATOR')["value"]` and the same for all the rest,
the cookies will be set in the Session object after your initial get so the
logic boils down to:
with requests.Session() as session:
headers = {"User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36"}
r = session.get('https://www.wes.org/appstatus/index.aspx', headers=headers)
soup = bs4.BeautifulSoup(r.content, "html5lib")
viewState = soup.select_one('#__VIEWSTATE')["value"]
viewStateGenerator = soup.select_one('#__VIEWSTATEGENERATOR')["value"]
eventValidation = soup.select_one('#__EVENTVALIDATION')["value"]
paramsPost = {'__VIEWSTATE': viewState,'__VIEWSTATEGENERATOR': viewStateGenerator,
'__EVENTVALIDATION': eventValidation,"txtUID": "[email protected]",
"txtPWD": "My_So_Called_Password",
"Submit": "Log In","Hidden1": ""}
response = session.post("https://www.wes.org/appstatus/index.aspx", data=paramsPost, headers=headers)
print("Status code:", response.status_code)
Python by convention uses CamelCase for class names and lowercase with
underscores to separate multiple words, you might want to consider applying
that to your code.
|
Cleaner method for finding the shortest distance between points in a python list?
Question: I have a list of tuples and an individual point in python e.g. [(1,2) , (2,5),
(6,7), (9,3)] and (2,1) , and I want to figure out the fastest path possible
created by all combinations of the individual point to the list of
points.(Basically I want to find the most efficient way to get to all of the
points starting from (2,1)). I have a manhattanDistance function that can take
it 2 points and output the distance. However, my algorithm is giving me
inconsistent answers (The heuristic is off for some reason)
What would be the correct way to accomplish this?
Here is my previous algorithm:
def bestPath(currentPoint,goalList):
sum = 0
bestList = []
while len(goallist) > 0:
for point in list:
bestList.append((manhattanD(point,currentPoint),point))
bestTup = min(bestList)
bestList = []
dist = bestTup[0]
newP = bestTup[1]
currentPoint = newP
sum += dist
return sum
Answer: Since you don't have so many point, you can easily use a solution that try
every possibility.
Here is what you can do:
First get all combinations:
>>> list_of_points = [(1,2) , (2,5), (6,7), (9,3)]
>>> list(itertools.permutations(list_of_points))
[((1, 2), (2, 5), (6, 7), (9, 3)),
((1, 2), (2, 5), (9, 3), (6, 7)),
((1, 2), (6, 7), (2, 5), (9, 3)),
((1, 2), (6, 7), (9, 3), (2, 5)),
((1, 2), (9, 3), (2, 5), (6, 7)),
((1, 2), (9, 3), (6, 7), (2, 5)),
((2, 5), (1, 2), (6, 7), (9, 3)),
((2, 5), (1, 2), (9, 3), (6, 7)),
((2, 5), (6, 7), (1, 2), (9, 3)),
((2, 5), (6, 7), (9, 3), (1, 2)),
((2, 5), (9, 3), (1, 2), (6, 7)),
((2, 5), (9, 3), (6, 7), (1, 2)),
((6, 7), (1, 2), (2, 5), (9, 3)),
((6, 7), (1, 2), (9, 3), (2, 5)),
((6, 7), (2, 5), (1, 2), (9, 3)),
((6, 7), (2, 5), (9, 3), (1, 2)),
((6, 7), (9, 3), (1, 2), (2, 5)),
((6, 7), (9, 3), (2, 5), (1, 2)),
((9, 3), (1, 2), (2, 5), (6, 7)),
((9, 3), (1, 2), (6, 7), (2, 5)),
((9, 3), (2, 5), (1, 2), (6, 7)),
((9, 3), (2, 5), (6, 7), (1, 2)),
((9, 3), (6, 7), (1, 2), (2, 5)),
((9, 3), (6, 7), (2, 5), (1, 2))]
Then create a function that give you the length of a combination:
def combination_length(start_point, combination):
lenght = 0
previous = start_point
for elem in combination:
lenght += manhattanDistance(previous, elem)
return length
Finally a function that test every possibility:
def get_shortest_path(start_point, list_of_point):
min = sys.maxint
combination_min = None
list_of_combinations = list(itertools.permutations(list_of_points))
for combination in list_of_combination:
length = combination_length(start_point, combination)
if length < min:
min = length
combination_min = combination
return combination_min
Then finally you can have:
import sys, itertools
def combination_length(start_point, combination):
lenght = 0
previous = start_point
for elem in combination:
lenght += manhattanDistance(previous, elem)
return length
def get_shortest_path(start_point, list_of_point):
min = sys.maxint
combination_min = None
list_of_combinations = list(itertools.permutations(list_of_points))
for combination in list_of_combination:
length = combination_length(start_point, combination)
if length < min:
min = length
combination_min = combination
return combination_min
list_of_points = [(1,2) , (2,5), (6,7), (9,3)]
print get_shortest_path((2,1), list_of_points)
|
Prepare my bigdata with Spark via Python
Question: My 100m in size, quantized data:
(1424411938', [3885, 7898])
(3333333333', [3885, 7898])
Desired result:
(3885, [3333333333, 1424411938])
(7898, [3333333333, 1424411938])
So what I want, is to transform the data so that I group 3885 (for example)
with all the `data[0]` that have it). Here is what I did in
[python](/questions/tagged/python "show questions tagged 'python'"):
def prepare(data):
result = []
for point_id, cluster in data:
for index, c in enumerate(cluster):
found = 0
for res in result:
if c == res[0]:
found = 1
if(found == 0):
result.append((c, []))
for res in result:
if c == res[0]:
res[1].append(point_id)
return result
but when I `mapPartitions()`'ed `data` RDD with `prepare()`, it seem to do
what I want only in the current partition, thus return a bigger result than
the desired.
For example, if the 1st record in the start was in the 1st partition and the
2nd in the 2nd, then I would get as a result:
(3885, [3333333333])
(7898, [3333333333])
(3885, [1424411938])
(7898, [1424411938])
How to modify my `prepare()` to get the desired effect? Alternatively, how to
process the result that `prepare()` produces, so that I can get the desired
result?
* * *
As you may already have noticed from the code, I do not care about speed at
all.
Here is a way to create the data:
data = []
from random import randint
for i in xrange(0, 10):
data.append((randint(0, 100000000), (randint(0, 16000), randint(0, 16000))))
data = sc.parallelize(data)
Answer: You can use a bunch of basic pyspark transformations to achieve this.
>>> rdd = sc.parallelize([(1424411938, [3885, 7898]),(3333333333, [3885, 7898])])
>>> r = rdd.flatMap(lambda x: ((a,x[0]) for a in x[1]))
We used `flatMap` to have a key, value pair for every item in `x[1]` and we
changed the data line format to `(a, x[0])`, the `a` here is every item in
`x[1]`. To understand `flatMap` better you can look to the documentation.
>>> r2 = r.groupByKey().map(lambda x: (x[0],tuple(x[1])))
We just grouped all key, value pairs by their keys and used tuple function to
convert iterable to tuple.
>>> r2.collect()
[(3885, (1424411938, 3333333333)), (7898, (1424411938, 3333333333))]
As you said you can use [:150] to have first 150 elements, I guess this would
be proper usage:
`r2 = r.groupByKey().map(lambda x: (x[0],tuple(x[1])[:150]))`
I tried to be as explanatory as possible. I hope this helps.
|
Openpyxl. Max_columns giving error
Question: I'm using openpyxl-2.4.0-b1 and Python version 34. Following is my code:
from openpyxl import load_workbook from openpyxl import Workbook
filename= str(input('Please enter the filename name, with the entire path and extension: '))
wb = load_workbook(filename)
ws = wb.worksheets[0]
row_main = 1
#Main Program starts here. Loop for the entire file
print ('Col =', ws.max_column())
while (row_main <(ws.max_row())): #Process rows for each column
print ('ROW =', row_main)
row_main += 1
It runs into an error: Traceback (most recent call last): print ('Col =',
ws.max_column()) TypeError: 'int' object is not callable
I can't use _get_ because it has been deprecated. Many thanks in advance.
Answer: As `max_column` is `property`, so just use the name to get its value:
print('Col =', ws.max_column)
Also apply on `max_row`:
while (row_main < ws.max_row):
|
Showing results python command to the web cgi
Question: I have a python script and it runs well if executed in a terminal or command
line , but after I try even internal server error occurred . how to enhance
the script to be run on the web.
HTML
<html><body>
<form enctype="multipart/form-data" action="http://localhost/cgi-bin/coba5.py" method="post">
<p>File: <input type="file" name="file"></p>
<p><input type="submit" value="Upload"></p>
</form>
</body></html>
PYTHON
#!/usr/bin/python
import cgi, os
import cgitb; cgitb.enable()
import simplejson as json
import optparse
import sys
try: # Windows needs stdio set for binary mode.
import msvcrt
msvcrt.setmode (0, os.O_BINARY) # stdin = 0
msvcrt.setmode (1, os.O_BINARY) # stdout = 1
except ImportError:
pass
form = cgi.FieldStorage()
# A nested FieldStorage instance holds the file
fileitem = form['file']
fn = os.path.basename(fileitem.filename)
data = open('/tmp/upload-cgi/' + fn, 'wb').write(fileitem.file.read())
data=fileitem.file.read()
location_database = open('/home/bioinformatics2/DW/taxo.json', 'r')
database = json.load(location_database)
for line in data:
for taxonomy in database:
if taxonomy["genus"] == line.replace('\n','') :
print "Kingdom: %s," % taxonomy['kingdom'],
print "phylum: %s," % taxonomy['phylum'],
print "class: %s," % taxonomy['class'],
print "order: %s," % taxonomy['order'],
print "family: %s," % taxonomy['family'],
print "genus: %s" % taxonomy['genus']
break
else:
print ("No found genus on taxanomy")
Answer: You're not outputting the HTTP header. At a minimum, you need to output:
print "Content-Type: text/plain"
print ""
You are also reading the file twice:
data = open('/tmp/upload-cgi/' + fn, 'wb').write(fileitem.file.read())
data=fileitem.file.read()
The second `fileitem.file.read()` will probably not read any content, since
the file should already be at the end-of-file, and so `data` will be empty.
|
Is there a tool to check what names I have used from a "wildly" imported module?
Question: I've been using python to do computations for my research. In an effort to
clean up my terrible code, I've been reading [Code Like a Pythonista:
Idiomatic
Python](http://python.net/~goodger/projects/pycon/2007/idiomatic/handout.html)
by David Goodger.
In this article, Goodger advises against "wild-card" imports of the form
from module import *
My code uses a lot of these. I'd like to clean my code up, but I'm not sure
how. I'm curious if there is a way to check what names from `module` I have
used in my code. This way I could either explicitly import these names or
replace the names with `module.name`. Is there a tool designed to accomplish
such a task?
Answer: Use a tool like `pyflakes` (which you should use anyway) to note which names
in your code become undefined after you replace `from module import *` with
`import module`. Once you've positively identified every instance of a name
imported from `module`, you can assess whether to
1. Always use `import module` and `module.x` for `x` imported from `module`.
2. Use `import module as m` and `m.x` if the module name is long.
3. Selectively import some names from `module` into the module namespace with `from module import x, y, z`
The above three are not mutually exclusive; as an extreme example, you can use
all three in the same module:
import really.long.module.name
import really.long.module.name as rlmn
from really.long.module.name import obvious_name
really.long.module.name.foo() # full module name
rlmn.bar() # module alias
obvious_name() # imported name
all in the same code. (I don't _recommend_ using all three in the same module.
Stick with either the full module name _or_ the alias throughout a single
module, but there is no harm importing common, obvious names directly and
using the fully qualified name for more obscure module attributes.)
|
Python sorting nested dictionary list
Question: I am trying to sort a nested dictionary in Python. Right now I have a
dictionary of dictionaries. I was able to sort the outer keys using sorted on
the list before I started building the dictionary, but I am unable to get the
inner keys sorted at this time. I've been trying to mess with the sorted API
but am having problems with it.
Right now I have:
myDict = {'A': {'Key3':4,'Key2':3,'Key4':2,'Key1':1},
'B': {'Key4':1,'Key3':2,'Key2':3,'Key1':4},
'C': {'Key1':4,'Key2':2,'Key4':1,'Key3':3}};
But I would like:
myDict = {'A': {'Key1':1,'Key2':3,'Key3':4,'Key4':2},
'B': {'Key1':4,'Key2':3,'Key2':1,'Key4':1},
'C': {'Key1':4,'Key2':2,'Key3':3,'Key4':1}};
I appreciate the help!
Answer:
>>> from collections import OrderedDict
>>> def sortedDict(items):
... return OrderedDict(sorted(items))
>>> myDict = {'A': {'Key3':4,'Key2':3,'Key4':2,'Key1':1},
... 'B': {'Key4':1,'Key3':2,'Key2':3,'Key1':4},
... 'C': {'Key1':4,'Key2':2,'Key4':1,'Key3':3}}
>>> sortedDict((key, sortedDict(value.items())) for key, value in myDict.items())
|
Create gantt chart with hlines?
Question: I've tried for several hours to make this work. I tried using 'python-gantt'
package, without luck. I also tried plotly (which was beautiful, but I can't
host my sensitive data on their site, so that won't work).
My starting point is code from here: [How to plot stacked event duration
(Gantt Charts) using Python
Pandas?](http://stackoverflow.com/questions/31820578/how-to-plot-stacked-
event-duration-gantt-charts-using-python-pandas)
Three Requirements:
* Include the 'Name' on the y axis rather than the numbers.
* If someone has multiple events, put all the event periods on one line (this will make pattern identification easier), e.g. Lisa will only have one line on the visual.
* Include the 'Event' listed on top of the corresponding line (if possible), e.g. Lisa's first line would say "Hire".
The code will need to be dynamic to accommodate many more people and more
possible event types...
I'm open to suggestions to visualize: I want to show the duration for various
staffing events throughout the year, as to help identify patterns.
from datetime import datetime
import pandas as pd
import matplotlib.pyplot as plt
import matplotlib.dates as dt
df = pd.DataFrame({'Name': ['Joe','Joe','Lisa','Lisa','Lisa','Alice'],
'Event': ['Hire','Term','Hire','Transfer','Term','Term'],
'Start_Date': ["2014-01-01","2014-02-01","2015-01-01","2015-02-01","2015-03-01","2016-01-01"],
'End_Date': ["2014-01-31","2014-03-15","2015-01-31","2015-02-28","2015-05-01","2016-09-01"]
})
df = df[['Name','Event','Start_Date','End_Date']]
df.Start_Date = pd.to_datetime(df.Start_Date).astype(datetime)
df.End_Date = pd.to_datetime(df.End_Date).astype(datetime)
fig = plt.figure()
ax = fig.add_subplot(111)
ax = ax.xaxis_date()
ax = plt.hlines(df.index, dt.date2num(df.Start_Date), dt.date2num(df.End_Date))
Answer: I encountered the same problem in the past. You seem to appreciate the
aesthetics of [Plotly](https://plot.ly/python/gantt/). Here is a little piece
of code which uses
[matplotlib.pyplot.broken_barh](http://matplotlib.org/api/pyplot_api.html#matplotlib.pyplot.broken_barh)
instead of
[matplotlib.pyplot.hlines](http://matplotlib.org/api/pyplot_api.html#matplotlib.pyplot.hlines).
[](http://i.stack.imgur.com/Bpr3I.png)
from collections import defaultdict
from datetime import datetime
from datetime import date
import pandas as pd
import matplotlib.dates as mdates
import matplotlib.patches as mpatches
import matplotlib.pyplot as plt
df = pd.DataFrame({
'Name': ['Joe', 'Joe', 'Lisa', 'Lisa', 'Lisa', 'Alice'],
'Event': ['Hire', 'Term', 'Hire', 'Transfer', 'Term', 'Term'],
'Start_Date': ['2014-01-01', '2014-02-01', '2015-01-01', '2015-02-01', '2015-03-01', '2016-01-01'],
'End_Date': ['2014-01-31', '2014-03-15', '2015-01-31', '2015-02-28', '2015-05-01', '2016-09-01']
})
df = df[['Name', 'Event', 'Start_Date', 'End_Date']]
df.Start_Date = pd.to_datetime(df.Start_Date).astype(datetime)
df.End_Date = pd.to_datetime(df.End_Date).astype(datetime)
names = df.Name.unique()
nb_names = len(names)
fig = plt.figure()
ax = fig.add_subplot(111)
bar_width = 0.8
default_color = 'blue'
colors_dict = defaultdict(lambda: default_color, Hire='green', Term='red', Transfer='orange')
# Plot the events
for index, name in enumerate(names):
mask = df.Name == name
start_dates = mdates.date2num(df.loc[mask].Start_Date)
end_dates = mdates.date2num(df.loc[mask].End_Date)
durations = end_dates - start_dates
xranges = zip(start_dates, durations)
ymin = index - bar_width / 2.0
ywidth = bar_width
yrange = (ymin, ywidth)
facecolors = [colors_dict[event] for event in df.loc[mask].Event]
ax.broken_barh(xranges, yrange, facecolors=facecolors, alpha=1.0)
# you can set alpha to 0.6 to check if there are some overlaps
# Shrink the x-axis
box = ax.get_position()
ax.set_position([box.x0, box.y0, box.width * 0.8, box.height])
# Add the legend
patches = [mpatches.Patch(color=color, label=key) for (key, color) in colors_dict.items()]
patches = patches + [mpatches.Patch(color=default_color, label='Other')]
plt.legend(handles=patches, bbox_to_anchor=(1, 0.5), loc='center left')
# Format the x-ticks
ax.xaxis.set_major_locator(mdates.YearLocator())
ax.xaxis.set_major_formatter(mdates.DateFormatter('%Y'))
ax.xaxis.set_minor_locator(mdates.MonthLocator())
# Format the y-ticks
ax.set_yticks(range(nb_names))
ax.set_yticklabels(names)
# Set the limits
date_min = date(df.Start_Date.min().year, 1, 1)
date_max = date(df.End_Date.max().year + 1, 1, 1)
ax.set_xlim(date_min, date_max)
# Format the coords message box
ax.format_xdata = mdates.DateFormatter('%Y-%m-%d')
# Set the title
ax.set_title('Gantt Chart')
plt.show()
I hope this will help you.
|
How to play music using pygame (Python)
Question: I am trying to play an .mp3 file using pygame. Here is my code:
import pygame
pygame.init()
pygame.mixer.init()
pygame.mixer.music.load('MSM.mp3')
pygame.mixer.music.play(0)
pygame.event.wait()
This however does not actually play the file. The audio file is located within
the same folder as the .py file.
Answer: The pyGame docs suggest using the pygame.mixer.pre_init() method before the
top-level pygame.init() on some platforms.
> "Some platforms require the pygame.mixer module for loading and playing
> sounds module to be initialized after the display modules have initialized.
> The top level pygame.init() takes care of this automatically, but cannot
> pass any arguments to the mixer init. To solve this, mixer has a function
> pygame.mixer.pre_init() to set the proper defaults before the toplevel init
> is used." <https://www.pygame.org/docs/ref/mixer.html#pygame.mixer.init>
|
scipy.interpolate leads to ImportError
Question: my setup is
import cartopy.crs as ccrs
import matplotlib.pyplot as plt
I have scipy `0.17` and cartopy '0.14.2'.
All I'm trying to do is
plt.axes(projection=ccrs.PlateCarree())
and it leads to this:
Traceback (most recent call last):
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/IPython/core/interactiveshell.py", line 2885, in run_code
exec(code_obj, self.user_global_ns, self.user_ns)
File "<ipython-input-93-636aeb1a7fc6>", line 1, in <module>
plt.axes(projection=ccrs.PlateCarree())
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/pyplot.py", line 867, in axes
return subplot(111, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/pyplot.py", line 1022, in subplot
a = fig.add_subplot(*args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/figure.py", line 987, in add_subplot
self, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/matplotlib/projections/__init__.py", line 100, in process_projection_requirements
projection_class, extra_kwargs = projection._as_mpl_axes()
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/crs.py", line 150, in _as_mpl_axes
import cartopy.mpl.geoaxes as geoaxes
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/mpl/geoaxes.py", line 52, in <module>
from cartopy.vector_transform import vector_scalar_to_grid
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/cartopy/vector_transform.py", line 26, in <module>
from scipy.interpolate import griddata
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/interpolate/__init__.py", line 158, in <module>
from .interpolate import *
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/interpolate/interpolate.py", line 12, in <module>
import scipy.special as spec
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/special/__init__.py", line 629, in <module>
from .basic import *
File "/opt/pycharm-2016.2/helpers/pydev/_pydev_bundle/pydev_import_hook.py", line 21, in do_import
module = self._system_import(name, *args, **kwargs)
File "/usr/local/anaconda2/envs/myenv3/lib/python3.5/site-packages/scipy/special/basic.py", line 14, in <module>
from ._ufuncs import (ellipkm1, mathieu_a, mathieu_b, iv, jv, gamma, psi, zeta,
ImportError: cannot import name 'zeta'
Deep down this appears to be a scipy problem, but I have the newest there -
what's going on here?
Answer: It turned out to be a problem with the `scipy` installation coming from
`conda`. Creating a new environment and freshly installing `scipy` solved the
issue.
|
Matplotlib: Every tick in different color
Question: I'm trying to create a scatter-plot with matplotlib (python 3.5) in which
every tick on the x-axes has a different color. How is this possible?
For example let's say the x-ticks are 'Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa',
'Su'. Now I want 'Mo' to be green, 'Tu' to be blue, ect...
Here is a very simple version of my code:
from matplotlib import pyplot as plt
plt.figure(figsize=(16, 11))
x = [1, 2, 3, 4, 5, 6, 7]
y = [10, 12, 9, 10, 8, 11, 10]
plt.scatter(x, y)
plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su'])
plt.show()
I already tried
my_colors = ['c', 'b', 'r', 'r', 'g', 'k', 'b']
plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su'], color=my_colors)
but that doesn't work.
Answer: You can loop over the tick labels (using `plt.gca().get_xticklabels()`) and
set their colors after they have been created, using `.set_color()`. For
example:
from matplotlib import pyplot as plt
plt.figure(figsize=(16, 11))
x = [1, 2, 3, 4, 5, 6, 7]
y = [10, 12, 9, 10, 8, 11, 10]
plt.scatter(x, y)
plt.xticks(x, ['Mo', 'Tu', 'We', 'Th', 'Fr', 'Sa', 'Su'])
my_colors = ['c', 'b', 'r', 'r', 'g', 'k', 'b']
for ticklabel, tickcolor in zip(plt.gca().get_xticklabels(), my_colors):
ticklabel.set_color(tickcolor)
plt.show()
[](http://i.stack.imgur.com/7o24S.png)
|
i want to scrape data using python script
Question: I have written python script to scrape data from
<http://www.cricbuzz.com/cricket-stats/icc-rankings/batsmen-rankings> It is a
list of 100 players and I successfully scraped this data. The problem is, when
i run script instead of scraping data just one time it scraped the same data 3
times.
<div class="cb-col cb-col-100 cb-font-14 cb-lst-itm text-center">
<div class="cb-col cb-col-16 cb-rank-tbl cb-font-16">1</div>
<div class="cb-col cb-col-50 cb-lst-itm-sm text-left">
<div class="cb-col cb-col-33">
<div class="cb-col cb-col-50">
<span class=" cb-ico" style="position:absolute;"></span> –
</div>
<div class="cb-col cb-col-50">
<img src="http://i.cricketcb.com/i/stats/fw/50x50/img/faceImages/2250.jpg" class="img-responsive cb-rank-plyr-img">
</div>
</div>
<div class="cb-col cb-col-67 cb-rank-plyr">
<a class="text-hvr-underline text-bold cb-font-16" href="/profiles/2250/steven-smith" title="Steven Smith's Profile">Steven Smith</a>
<div class="cb-font-12 text-gray">AUSTRALIA</div>
</div>
</div>
<div class="cb-col cb-col-17 cb-rank-tbl">906</div>
<div class="cb-col cb-col-17 cb-rank-tbl">1</div>
</div>
And here is python script which i write scrap each player data.
import sys,requests,csv,io
from bs4 import BeautifulSoup
from urllib.parse import urljoin
url = "http://www.cricbuzz.com/cricket-stats/icc-rankings/batsmen-rankings"
r = requests.get(url)
r.content
soup = BeautifulSoup(r.content, "html.parser")
maindiv = soup.find_all("div", {"class": "text-center"})
for div in maindiv:
print(div.text)
but instead of scraping the data once, it scrapes the same data 3 times.
Where can I make changes to get data just one time?
Answer: Select the table and look for the divs in that:
maindiv = soup.select("#batsmen-tests div.text-center")
for div in maindiv:
print(div.text)
Your original output and that above gets all the text from the divs as one
line which is not really useful, if you just want the player names:
anchors = soup.select("#batsmen-tests div.cb-rank-plyr a")
for a in anchors:
print(a.text)
A quick and easy way to get the data in a nice csv format is to just get text
from each child:
maindiv = soup.select("#batsmen-tests div.text-center")
for d in maindiv[1:]:
row_data = u",".join(s.strip() for s in filter(None, (t.find(text=True, recursive=False) for t in d.find_all())))
if row_data:
print(row_data)
Now you get output like:
# rank, up/down, name, country, rating, best rank
1,–,Steven Smith,AUSTRALIA,906,1
2,–,Joe Root,ENGLAND,878,1
3,–,Kane Williamson,NEW ZEALAND,876,1
4,–,Hashim Amla,SOUTH AFRICA,847,1
5,–,Younis Khan,PAKISTAN,845,1
6,–,Adam Voges,AUSTRALIA,802,5
7,–,AB de Villiers,SOUTH AFRICA,802,1
8,–,Ajinkya Rahane,INDIA,785,8
9,2,David Warner,AUSTRALIA,772,3
10,–,Alastair Cook,ENGLAND,770,2
11,1,Misbah-ul-Haq,PAKISTAN,764,6
As opposed to:
PositionPlayerRatingBest Rank
Player
1 –Steven SmithAUSTRALIA9061
2 –Joe RootENGLAND8781
3 –Kane WilliamsonNEW ZEALAND8761
4 –Hashim AmlaSOUTH AFRICA8471
5 –Younis KhanPAKISTAN8451
6 –Adam VogesAUSTRALIA8025
|
In 'IDA PRO', let 'IDAPython' import default module at startup
Question: We know 'IDAPython' loads several modules by default at startup, such as
idaapi, idautils.... I wrote a module to let python print all numbers as hex
format in the command window, which I wish can be imported each time when
python loads those default modules. how to achieve that?
Answer: Create a file `%APPDATA%\Hex-Rays\IDA Pro\idapythonrc.py` with the following
content:
import idaapi
idaapi.require('mymodule')
With this file in place, you can even keep `mymodule.py` in the same
directory.
P.S. IDA can tell you the path to this directory, too, which could be handy on
other OSes or if the relevant company name changes again ;-). Just enter:
get_user_idadir()
at the prompt.
|
Sending gzip compressed data through TCP socket in Python
Question: I'm creating an HTTP server in Python without any of the HTTP libraries for
learning purposes. Right now it can serve static files fine.
The way I serve the file is through this piece of code:
with open(self.filename, 'rb') as f:
src = f.read()
socket.sendall(src)
However, I want to optimize its performance a bit by sending compressed data
instead of uncompressed. I know that my browser (Chrome) accepts compressed
data because it tells me in the header
Accept-Encoding: gzip, deflate, sdch
So, I changed my code to this
with open(self.filename, 'rb') as f:
src = zlib.compress(f.read())
socket.sendall(src)
But this just outputs garbage. What am I doing wrong?
Answer: The zlib library implements the deflate compression algorithm (RFC 1951).
There are two encapsulations for the deflate compression: zlib (RFC 1950) and
gzip (RFC 1952). These differ only in the kind of header and trailer they
provide around the deflate compressed data.
`zlib.compress` only provides the raw deflate data, without any header and
trailer. To get these you need to use a compression object. For gzip this
looks like this:
z = zlib.compressobj(-1,zlib.DEFLATED,31)
gzip_compressed_data = z.compress(data) + z.flush()
The important part here is the `31` as the 3rd argument to `compressobj`. This
specifies gzip format which then can be used with `Content-Encoding: gzip`.
|
Convert strings of bytes to floats in Python
Question: I'm using Pandas to handle my data. The first step of my script is to convert
a column of strings of bytes to a list of floats. My script is working but its
taking too long. Any suggestions on how to speed it up??
def byte_to_hex(byte_str):
array = ["%02X" % ord(chr(x)) for x in byte_str]
return array
for index, row in enumerate(data[column].values):
row_new = []
row = byte_to_hex(row)
stop = len(row)
for i in range(4, stop, 4):
sample = "".join(row[i:i + 4])
row_new.append(struct.unpack('!f', bytes.fromhex(sample))[0])
Example:
b'\x00\x00\x00\x1cB\x80\x1f\xfcB\x80\x1f\xfcB\x80w\xc8Bz\xa1\x97B\x80\x1f\xfcB}LZB\x80\x1f\xfcBz\xa1\x97Bz\xcaoB\x803\xf5B}\xc5\x84B\x80w\xc8B}\xed\xdbB\x80\x1f\xfcB}\xc5\x84B}LZB\x80\x1f\xfcB}#\xe9B}\xed\xdbB}\xc5\x84B\x803\xf5B\x80\x1f\xfcB}\xc5\x84B\x803\xf5B\x803\xf5Bx\xef\tB\x81\xc4\xdf\x7f\xff\xff\xff'
[64.06246948242188, 64.06246948242188, 64.23394775390625, 62.65780258178711, 64.06246948242188, 63.324562072753906, 64.06246948242188, 62.65780258178711, 62.697689056396484, 64.10147857666016, 63.44288635253906, 64.23394775390625, 63.48228073120117, 64.06246948242188, 63.44288635253906, 63.324562072753906, 64.06246948242188, 63.28506851196289, 63.48228073120117, 63.44288635253906, 64.10147857666016, 64.06246948242188, 63.44288635253906, 64.10147857666016, 64.10147857666016]
Very thankful for any help :)
Answer: I think you are looking for the
[Struct](https://docs.python.org/3.4/library/struct.html) package
import struct
struct.pack('f', 3.14)
OUT: b'\xdb\x0'
struct.unpack('f', b'\xdb\x0fI@')
OUT: (3.1415927410125732,)
struct.pack('4f', 1.0, 2.0, 3.0, 4.0)
OUT: '\x00\x00\x80?\x00\x00\x00@\x00\x00@@\x00\x00\x80@'
|
Adding an extra hidden layer using Google's TensorFlow
Question: I am trying to create a multi-label classifier using TensorFlow. Though I'm
having trouble adding and connecting hidden layers.
I was following this tutorial:
<http://jrmeyer.github.io/tutorial/2016/02/01/TensorFlow-Tutorial.html>
The data I'm using is UCI's Iris data, encoded to one-hot:
Training X [105,4]
5,3.2,1.2,0.2
5.5,3.5,1.3,0.2
4.9,3.1,1.5,0.1
4.4,3,1.3,0.2
5.1,3.4,1.5,0.2
.
.
.
Training Y [105,3]
0,0,1
0,0,1
0,0,1
0,0,1
0,0,1
0,0,1
.
.
.
I'm also using testing data X and Y which are [45,4] and [45,3] respectively.
Here is my python code:
import tensorflow as tf
import numpy as np
import tarfile
import os
import matplotlib.pyplot as plt
import time
## Import data
def csv_to_numpy_array(filePath, delimiter):
return np.genfromtxt(filePath, delimiter=delimiter, dtype=None)
trainX = csv_to_numpy_array("Iris_training_x.csv", delimiter=",").astype(np.float32)
trainY = csv_to_numpy_array("Iris_training_y.csv", delimiter=",").astype(np.float32)
testX = csv_to_numpy_array("Iris_testing_x.csv", delimiter=",").astype(np.float32)
testY = csv_to_numpy_array("Iris_testing_y.csv", delimiter=",").astype(np.float32)
# Data Set Paramaters
numFeatures = trainX.shape[1]
numLabels = trainY.shape[1]
# Training Session Parameters
numEpochs = 1000
learningRate = tf.train.exponential_decay(learning_rate=0.008,
global_step= 1,
decay_steps=trainX.shape[0],
decay_rate= 0.95,
staircase=True)
# Placeholders
X=tf.placeholder(tf.float32, [None, numFeatures])
y=tf.placeholder(tf.float32, [None, numLabels])
# Initialize our weights and biases
Weights = tf.Variable(tf.random_normal([numFeatures, numLabels],
mean=0,
stddev=(np.sqrt(6 / numFeatures + numLabels + 1)),
name="Weights"))
bias = tf.Variable(tf.random_normal([1, numLabels],
mean=0,
stddev=(np.sqrt(6 / numFeatures + numLabels + 1)),
name="bias"))
# Prediction algorithm (feedforward)
apply_weights_OP = tf.matmul(X, Weights, name="apply_weights")
add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias")
activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation")
numFeatures = activation_OP
apply_weights_OP = tf.matmul(X, Weights, name="apply_weights")
add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias")
activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation")
init_OP = tf.initialize_all_variables()
# Cost function (Mean Squeared Error)
cost_OP = tf.nn.l2_loss(activation_OP-y, name="squared_error_cost")
# Optimization Algorithm (Gradient Descent)
training_OP = tf.train.GradientDescentOptimizer(learningRate).minimize(cost_OP)
# Visualize
epoch_values=[]
accuracy_values=[]
cost_values=[]
# Turn on interactive plotting
plt.ion()
# Create the main, super plot
fig = plt.figure()
# Create two subplots on their own axes and give titles
ax1 = plt.subplot("211")
ax1.set_title("TRAINING ACCURACY", fontsize=18)
ax2 = plt.subplot("212")
ax2.set_title("TRAINING COST", fontsize=18)
plt.tight_layout()
# Create a tensorflow session
sess = tf.Session()
# Initialize all tensorflow variables
sess.run(init_OP)
## Ops for vizualization
# argmax(activation_OP, 1) gives the label our model thought was most likely
# argmax(y, 1) is the correct label
correct_predictions_OP = tf.equal(tf.argmax(activation_OP,1),tf.argmax(y,1))
# False is 0 and True is 1, what was our average?
accuracy_OP = tf.reduce_mean(tf.cast(correct_predictions_OP, "float"))
# Summary op for regression output
activation_summary_OP = tf.histogram_summary("output", activation_OP)
# Summary op for accuracy
accuracy_summary_OP = tf.scalar_summary("accuracy", accuracy_OP)
# Summary op for cost
cost_summary_OP = tf.scalar_summary("cost", cost_OP)
# Summary ops to check how variables (W, b) are updating after each iteration
weightSummary = tf.histogram_summary("Weights", Weights.eval(session=sess))
biasSummary = tf.histogram_summary("biases", bias.eval(session=sess))
# Merge all summaries
all_summary_OPS = tf.merge_all_summaries()
# Summary writer
writer = tf.train.SummaryWriter("summary_logs", sess.graph_def)
# Initialize reporting variables
cost = 0
diff = 1
# Training epochs
for i in range(numEpochs):
if i > 1 and diff < .0001:
print("change in cost %g; convergence."%diff)
break
else:
# Run training step
step = sess.run(training_OP, feed_dict={X: trainX, y: trainY})
# Report occasional stats
if i % 10 == 0:
#Add epoch to epoch_values
epoch_values.append(i)
#Generate accuracy stats on test data
summary_results, train_accuracy, newCost = sess.run(
[all_summary_OPS, accuracy_OP, cost_OP],
feed_dict={X: trainX, y: trainY}
)
# Add accuracy to live graphing variable
accuracy_values.append(train_accuracy)
# Add cost to live graphing variable
cost_values.append(newCost)
#Write summary stats to writer
#writer.add_summary(summary_results, i)
# Re-assign values for variables
diff = abs(newCost - cost)
cost = newCost
#generate print statements
print("step %d, training accuracy %g"%(i, train_accuracy))
print("step %d, cost %g"%(i, newCost))
print("step %d, change in cost %g"%(i, diff))
# Plot progress to our two subplots
accuracyLine, = ax1.plot(epoch_values, accuracy_values)
costLine, = ax2.plot(epoch_values, cost_values)
fig.canvas.draw()
#time.sleep(1)
# How well do we perform on held-out test data?
print("final accuracy on test set: %s" %str(sess.run(accuracy_OP, feed_dict={X: testX, y: testY})))
# Create Saver
saver = tf.train.Saver()
# Save variables to .ckpt file
# saver.save(sess, "trained_variables.ckpt")
# Close tensorflow session
sess.close()
The problem is here:
# Prediction algorithm (feedforward)
apply_weights_OP = tf.matmul(X, Weights, name="apply_weights")
add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias")
activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation")
numFeatures = activation_OP
apply_weights_OP = tf.matmul(activation_OP, Weights, name="apply_weights")
add_bias_OP = tf.add(apply_weights_OP, bias, name="add_bias")
activation_OP = tf.nn.sigmoid(add_bias_OP, name="activation")
My understanding is that the output of one layer should connect to the input
of the next one. I just don't know how to modify either the output or input of
the layers; it keeps giving me this compatibility error:
/usr/bin/python3.5 /home/marco/PycharmProjects/NN_Iris/main
Traceback (most recent call last):
File "/home/marco/PycharmProjects/NN_Iris/main", line 132, in <module>
apply_weights_OP = tf.matmul(activation_OP, Weights, name="apply_weights")
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/ops/math_ops.py", line 1346, in matmul
name=name)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/ops/gen_math_ops.py", line 1271, in _mat_mul
transpose_b=transpose_b, name=name)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/op_def_library.py", line 703, in apply_op
op_def=op_def)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/ops.py", line 2312, in create_op
set_shapes_for_outputs(ret)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/ops.py", line 1704, in set_shapes_for_outputs
shapes = shape_func(op)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/common_shapes.py", line 94, in matmul_shape
inner_a.assert_is_compatible_with(inner_b)
File "/usr/local/lib/python3.5/dist-packages/tensorflow/python/framework/tensor_shape.py", line 108, in assert_is_compatible_with
% (self, other))
ValueError: Dimensions 3 and 4 are not compatible
Process finished with exit code 1
Any suggestions on how to properly connect the two hidden layers? Thanks.
Answer: If you want a fully connected network with one hidden layer and an output
layer there is how their shapes should look like:
# hidden layer
weights_hidden = tf.Variable(tf.random_normal([numFeatures, num_nodes])
bias_hidden = tf.Variable(tf.random_normal([num_nodes])
preactivations_hidden = tf.add(tf.matmul(X, weights_hidden), bias_hidden)
activations_hidden = tf.nn.sigmoid(preactivations_hidden)
# output layer
weights_output = tf.Variable(tf.random_normal([num_nodes, numLabels])
bias_output = tf.Variable(tf.random_normal([numLabels])
preactivations_output = tf.add(tf.matmul(activations_hidden, weights_output), bias_output)
Where `num_nodes` is number of nodes in the hidden layer that you select
yourself. `X` is a `[105, numFeatures]` matrix, `weights_hidden` is
`[numFeatures, num_nodes]` matrix, so output of first hidden layer is `[105,
num_nodes]`. In the same way, `[105, num_nodes]` multiplied by `[num_nodes,
numLabels]` yields `[105, numLabels]` output.
|
Trouble installing cryptography with pip3 (Ubuntu 16.04 LTS)
Question: i'm currently having trouble installing cryptography with pip3. I tried all
the other solutions but non of them worked.
Here is the output from "sudo -H pip install cryptography":
Collecting cryptography
Using cached cryptography-1.5.tar.gz
Requirement already satisfied (use --upgrade to upgrade): idna>=2.0 in /usr/local/lib/python3.5/dist-packages (from cryptography)
Requirement already satisfied (use --upgrade to upgrade): pyasn1>=0.1.8 in /usr/lib/python3/dist-packages (from cryptography)
Requirement already satisfied (use --upgrade to upgrade): six>=1.4.1 in /usr/lib/python3/dist-packages (from cryptography)
Requirement already satisfied (use --upgrade to upgrade): setuptools>=11.3 in /usr/local/lib/python3.5/dist-packages (from cryptography)
Collecting cffi>=1.4.1 (from cryptography)
Using cached cffi-1.8.2.tar.gz
Requirement already satisfied (use --upgrade to upgrade): pycparser in /usr/local/lib/python3.5/dist-packages (from cffi>=1.4.1->cryptography)
Building wheels for collected packages: cryptography, cffi
Running setup.py bdist_wheel for cryptography ... error
Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cryptography/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" bdist_wheel -d /tmp/tmpme9hnxc5pip-wheel- --python-tag cp35:
c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory
compilation terminated.
Traceback (most recent call last):
File "/usr/lib/python3.5/distutils/unixccompiler.py", line 118, in _compile
extra_postargs)
File "/usr/lib/python3.5/distutils/ccompiler.py", line 909, in spawn
spawn(cmd, dry_run=self.dry_run)
File "/usr/lib/python3.5/distutils/spawn.py", line 36, in spawn
_spawn_posix(cmd, search_path, dry_run=dry_run)
File "/usr/lib/python3.5/distutils/spawn.py", line 159, in _spawn_posix
% (cmd, exit_status))
distutils.errors.DistutilsExecError: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.5/distutils/core.py", line 148, in setup
dist.run_commands()
File "/usr/lib/python3.5/distutils/dist.py", line 955, in run_commands
self.run_command(cmd)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 161, in run
cmd = self.call_command('install_lib', warn_dir=0)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 147, in call_command
self.run_command(cmdname)
File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/install_lib.py", line 11, in run
self.build()
File "/usr/lib/python3.5/distutils/command/install_lib.py", line 109, in build
self.run_command('build_ext')
File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 77, in run
_build_ext.run(self)
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 338, in run
self.build_extensions()
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 447, in build_extensions
self._build_extensions_serial()
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 472, in _build_extensions_serial
self.build_extension(ext)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 198, in build_extension
_build_ext.build_extension(self, ext)
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 532, in build_extension
depends=ext.depends)
File "/usr/lib/python3.5/distutils/ccompiler.py", line 574, in compile
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/usr/lib/python3.5/distutils/unixccompiler.py", line 120, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules
yield saved
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup
DirectorySandbox(setup_dir).run(runner)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run
return func()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner
_execfile(setup_script, ns)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile
exec(code, globals, locals)
File "/tmp/easy_install-sjss3av8/cffi-1.8.2/setup.py", line 193, in <module>
return True
File "/usr/lib/python3.5/distutils/core.py", line 163, in setup
raise SystemExit("error: " + str(msg))
SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1100, in run_setup
run_setup(setup_script, args)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 249, in run_setup
raise
File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 168, in save_modules
saved_exc.resume()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 143, in resume
six.reraise(type, exc, self._tb)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/_vendor/six.py", line 685, in reraise
raise value.with_traceback(tb)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules
yield saved
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup
DirectorySandbox(setup_dir).run(runner)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run
return func()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner
_execfile(setup_script, ns)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile
exec(code, globals, locals)
File "/tmp/easy_install-sjss3av8/cffi-1.8.2/setup.py", line 193, in <module>
return True
File "/usr/lib/python3.5/distutils/core.py", line 163, in setup
raise SystemExit("error: " + str(msg))
SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-build-2saf6qpj/cryptography/setup.py", line 334, in <module>
**keywords_with_side_effects(sys.argv)
File "/usr/lib/python3.5/distutils/core.py", line 108, in setup
_setup_distribution = dist = klass(attrs)
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 348, in __init__
self.fetch_build_eggs(attrs['setup_requires'])
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 394, in fetch_build_eggs
replace_conflicting=True,
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 851, in resolve
dist = best[req.key] = env.best_match(req, ws, installer)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1123, in best_match
return self.obtain(req, installer)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1135, in obtain
return installer(requirement)
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 461, in fetch_build_egg
return cmd.easy_install(req)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 664, in easy_install
return self.install_item(spec, dist.location, tmpdir, deps)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 694, in install_item
dists = self.install_eggs(spec, download, tmpdir)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 875, in install_eggs
return self.build_and_install(setup_script, setup_base)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1114, in build_and_install
self.run_setup(setup_script, setup_base, args)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1102, in run_setup
raise DistutilsError("Setup script exited with %s" % (v.args[0],))
distutils.errors.DistutilsError: Setup script exited with error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
----------------------------------------
Failed building wheel for cryptography
Running setup.py clean for cryptography
Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cryptography/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" clean --all:
c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory
compilation terminated.
Traceback (most recent call last):
File "/usr/lib/python3.5/distutils/unixccompiler.py", line 118, in _compile
extra_postargs)
File "/usr/lib/python3.5/distutils/ccompiler.py", line 909, in spawn
spawn(cmd, dry_run=self.dry_run)
File "/usr/lib/python3.5/distutils/spawn.py", line 36, in spawn
_spawn_posix(cmd, search_path, dry_run=dry_run)
File "/usr/lib/python3.5/distutils/spawn.py", line 159, in _spawn_posix
% (cmd, exit_status))
distutils.errors.DistutilsExecError: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.5/distutils/core.py", line 148, in setup
dist.run_commands()
File "/usr/lib/python3.5/distutils/dist.py", line 955, in run_commands
self.run_command(cmd)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 161, in run
cmd = self.call_command('install_lib', warn_dir=0)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/bdist_egg.py", line 147, in call_command
self.run_command(cmdname)
File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/install_lib.py", line 11, in run
self.build()
File "/usr/lib/python3.5/distutils/command/install_lib.py", line 109, in build
self.run_command('build_ext')
File "/usr/lib/python3.5/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 77, in run
_build_ext.run(self)
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 338, in run
self.build_extensions()
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 447, in build_extensions
self._build_extensions_serial()
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 472, in _build_extensions_serial
self.build_extension(ext)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/build_ext.py", line 198, in build_extension
_build_ext.build_extension(self, ext)
File "/usr/lib/python3.5/distutils/command/build_ext.py", line 532, in build_extension
depends=ext.depends)
File "/usr/lib/python3.5/distutils/ccompiler.py", line 574, in compile
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/usr/lib/python3.5/distutils/unixccompiler.py", line 120, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules
yield saved
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup
DirectorySandbox(setup_dir).run(runner)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run
return func()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner
_execfile(setup_script, ns)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile
exec(code, globals, locals)
File "/tmp/easy_install-rfvxg0ae/cffi-1.8.2/setup.py", line 193, in <module>
return True
File "/usr/lib/python3.5/distutils/core.py", line 163, in setup
raise SystemExit("error: " + str(msg))
SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1100, in run_setup
run_setup(setup_script, args)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 249, in run_setup
raise
File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/lib/python3.5/contextlib.py", line 77, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 168, in save_modules
saved_exc.resume()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 143, in resume
six.reraise(type, exc, self._tb)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/_vendor/six.py", line 685, in reraise
raise value.with_traceback(tb)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 156, in save_modules
yield saved
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 197, in setup_context
yield
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 246, in run_setup
DirectorySandbox(setup_dir).run(runner)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 276, in run
return func()
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 245, in runner
_execfile(setup_script, ns)
File "/usr/local/lib/python3.5/dist-packages/setuptools/sandbox.py", line 47, in _execfile
exec(code, globals, locals)
File "/tmp/easy_install-rfvxg0ae/cffi-1.8.2/setup.py", line 193, in <module>
return True
File "/usr/lib/python3.5/distutils/core.py", line 163, in setup
raise SystemExit("error: " + str(msg))
SystemExit: error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-build-2saf6qpj/cryptography/setup.py", line 334, in <module>
**keywords_with_side_effects(sys.argv)
File "/usr/lib/python3.5/distutils/core.py", line 108, in setup
_setup_distribution = dist = klass(attrs)
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 348, in __init__
self.fetch_build_eggs(attrs['setup_requires'])
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 394, in fetch_build_eggs
replace_conflicting=True,
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 851, in resolve
dist = best[req.key] = env.best_match(req, ws, installer)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1123, in best_match
return self.obtain(req, installer)
File "/usr/local/lib/python3.5/dist-packages/pkg_resources/__init__.py", line 1135, in obtain
return installer(requirement)
File "/usr/local/lib/python3.5/dist-packages/setuptools/dist.py", line 461, in fetch_build_egg
return cmd.easy_install(req)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 664, in easy_install
return self.install_item(spec, dist.location, tmpdir, deps)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 694, in install_item
dists = self.install_eggs(spec, download, tmpdir)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 875, in install_eggs
return self.build_and_install(setup_script, setup_base)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1114, in build_and_install
self.run_setup(setup_script, setup_base, args)
File "/usr/local/lib/python3.5/dist-packages/setuptools/command/easy_install.py", line 1102, in run_setup
raise DistutilsError("Setup script exited with %s" % (v.args[0],))
distutils.errors.DistutilsError: Setup script exited with error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
----------------------------------------
Failed cleaning build dir for cryptography
Running setup.py bdist_wheel for cffi ... error
Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" bdist_wheel -d /tmp/tmp689jaw0cpip-wheel- --python-tag cp35:
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-x86_64-3.5
creating build/lib.linux-x86_64-3.5/cffi
copying cffi/cffi_opcode.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/cparser.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/recompiler.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/vengine_gen.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/commontypes.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/verifier.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/model.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/api.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/setuptools_ext.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/vengine_cpy.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/ffiplatform.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/__init__.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/lock.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/backend_ctypes.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/_cffi_include.h -> build/lib.linux-x86_64-3.5/cffi
copying cffi/parse_c_type.h -> build/lib.linux-x86_64-3.5/cffi
copying cffi/_embedding.h -> build/lib.linux-x86_64-3.5/cffi
running build_ext
building '_cffi_backend' extension
creating build/temp.linux-x86_64-3.5
creating build/temp.linux-x86_64-3.5/c
x86_64-linux-gnu-gcc -pthread -DNDEBUG -g -fwrapv -O2 -Wall -Wstrict-prototypes -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -DUSE__THREAD -I/usr/include/python3.5m -c c/_cffi_backend.c -o build/temp.linux-x86_64-3.5/c/_cffi_backend.o
c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory
compilation terminated.
error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
----------------------------------------
Failed building wheel for cffi
Running setup.py clean for cffi
Failed to build cryptography cffi
Installing collected packages: cffi, cryptography
Running setup.py install for cffi ... error
Complete output from command /usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" install --record /tmp/pip-7jmqgdpf-record/install-record.txt --single-version-externally-managed --compile:
running install
running build
running build_py
creating build
creating build/lib.linux-x86_64-3.5
creating build/lib.linux-x86_64-3.5/cffi
copying cffi/cffi_opcode.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/cparser.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/recompiler.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/vengine_gen.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/commontypes.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/verifier.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/model.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/api.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/setuptools_ext.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/vengine_cpy.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/ffiplatform.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/__init__.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/lock.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/backend_ctypes.py -> build/lib.linux-x86_64-3.5/cffi
copying cffi/_cffi_include.h -> build/lib.linux-x86_64-3.5/cffi
copying cffi/parse_c_type.h -> build/lib.linux-x86_64-3.5/cffi
copying cffi/_embedding.h -> build/lib.linux-x86_64-3.5/cffi
running build_ext
building '_cffi_backend' extension
creating build/temp.linux-x86_64-3.5
creating build/temp.linux-x86_64-3.5/c
x86_64-linux-gnu-gcc -pthread -DNDEBUG -g -fwrapv -O2 -Wall -Wstrict-prototypes -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -DUSE__THREAD -I/usr/include/python3.5m -c c/_cffi_backend.c -o build/temp.linux-x86_64-3.5/c/_cffi_backend.o
c/_cffi_backend.c:2:20: fatal error: Python.h: No such file or directory
compilation terminated.
error: command 'x86_64-linux-gnu-gcc' failed with exit status 1
----------------------------------------
Command "/usr/bin/python3 -u -c "import setuptools, tokenize;__file__='/tmp/pip-build-2saf6qpj/cffi/setup.py';exec(compile(getattr(tokenize, 'open', open)(__file__).read().replace('\r\n', '\n'), __file__, 'exec'))" install --record /tmp/pip-7jmqgdpf-record/install-record.txt --single-version-externally-managed --compile" failed with error code 1 in /tmp/pip-build-2saf6qpj/cffi/
I tried understanding the errors but all it says is "failed with error code 1"
:/
Checking out other questions, they mentioned about the dependencies for
cryptography which is all installed and up-to-date (sudo apt-get install
build-essential libssl-dev libffi-dev python-dev):
Reading package lists... Done
Building dependency tree
Reading state information... Done
build-essential is already the newest version (12.1ubuntu2).
libffi-dev is already the newest version (3.2.1-4).
python-dev is already the newest version (2.7.11-1).
libssl-dev is already the newest version (1.0.2g-1ubuntu4.2).
0 upgraded, 0 newly installed, 0 to remove and 135 not upgraded.
pip3 version:
pip 8.1.2 from /usr/local/lib/python3.5/dist-packages (python 3.5)
Not even easy_install worked. I really don't know what i'm doing wrong here,
please help.
Answer: From the `Python.h: No such file or directory` error messages appearing in
various places of the log, you need to install `python3-dev`:
sudo apt install python3-dev
|
calculate catalan numbers up to a billion
Question: I'm new to python (and programming in general), I was asked in my class to
calculate Catalan numbers up to a billion but the program I wrote for it is
not working as intended.
from numpy import division
C=1
n=0
while C<=10000000000:
print (C)
C=(4*n+2)/(n+2)*C
n=n+1
This is what it prints
1, 1, 2, 4, 8, 24, 72, 216, 648, 1944, 5832, 17496, 52488, 157464, 472392,
1417176, 4251528, 12754584, 38263752, 114791256, 344373768, 1033121304,
3099363912, 9298091736,
As you can see from my fourth iteration onwards, I get the wrong number and I
don't understand why.
EDIT: The mathematical definition I used is not wrong! I know the Wiki has
another definition but this one is not wrong. Co=1, Cn+1 = (4*n+2)/(n+2)*Cn
Answer:
C=(4*n+2)/(n+2)*C
This applies the calculation in the wrong order. Because you are using integer
arithmetic, `(4*n+2)/(n+2)` loses information if you have not already factored
in `C`. The correct calculation is:
C=C*(4*n+2)/(n+2)
|
Reassigning a return variable to a function in Python
Question: I am trying to figure out why I am getting a "SyntaxError: invalid syntax "
the variable title is highlighted red
from urllib.request import urlopen
from urllib.error import HTTPError
from urllib.error import URLError
from bs4 import BeautifulSoup
def getTitle(url):
try:
html=urlopen(url)
except HTTPError as e:
return None
try:
bsObj=BeautifulSoup(html.read())
title=bsObj.body.h1
except AttributeError as e:
return None
return title
title=getTitle("http://....html")
Answer: Python is very sensitive to whitespace, indentation, in order to keep your
code looking clean. From what I can see, you need to intent your try/except
statements:
from urllib.request import urlopen
from urllib.error import HTTPError
from urllib.error import URLError
from bs4 import BeautifulSoup
def getTitle(url):
try:
html=urlopen(url)
except HTTPError as e:
return None
try:
bsObj=BeautifulSoup(html.read())
title=bsObj.body.h1
except AttributeError as e:
return None
return title
title=getTitle("http://....html")
That should be it, do say if further errors occur.
|
Extracting text from multiple powerpoint files using python
Question: I am trying to find a way to look in a folder and search the contents of all
of the powerpoint documents within that folder for specific strings,
preferably using Python. When those strings are found, I want to report out
the text after that string as well as what document it was found in. I would
like to compile the information and report it in a CSV file.
So far I've only come across the olefil package,
<https://bitbucket.org/decalage/olefileio_pl/wiki/Home>. This provides all of
the text contained in a specific document, which is not what I am looking to
do. Please help.
Answer: `python-pptx` can be used to do what you propose. Just at a high level, you
would do something like this (not working code, just and idea of overall
approach):
from pptx import Presentation
for pptx_filename in directory:
prs = Presentation(pptx_filename)
for slide in prs.slides:
for shape in slide.shapes:
print shape.text
You'd need to add the bits about searching shape text for key strings and
adding them to a CSV file or whatever, but this general approach should work
just fine. I'll leave it to you to work out the finer points :)
|
how to parse key value pair request from url using python requests in flask
Question: I have spent about a week on this issue and although I have made considerable
progress I am stuck at a key point.
I am writing a simple client-server program in Python that is supposed to
accept key/value pairs from the command line, formulate them into an url, and
request the url from the server. The problem appears to be either that the url
is not properly formatted or that the server is not parsing it correctly.
(That is, the key-value pairs appear to be properly making it from the command
line to the function that contains the request.)
import sys
import requests
server_url = "http://0.0.0.0:5006"
def test(payload):
print('payload in function is ' + payload)
r = requests.get('http://0.0.0.0:5006/buy', params=payload)
print(r.url) #debug
print(r.encoding)
print(r.text)
print(r.headers)
if __name__ == '__main__':
payload = sys.argv[2]
print('payload from cli is ' + payload)
test(payload)
Server:
import subprocess
from flask import Flask
from flask import request
import request
# Configure the app and wallet
app = Flask(__name__)
@app.route('/buy', methods=['GET', 'POST'])
def test(key1, key2):
key1 = str(request.args.get('key1'))
key2 = str(request.args.get('key2'))
print('keys are' + key1 + key2)
fortune = subprocess.check_output(['echo', 'key1'])
return fortune
# Initialize and run the server
if __name__ == '__main__':
app.run(host='0.0.0.0', port=5006)
Client console output:
payload from cli is {"key1": "foo", "key2": "bar"}
payload in function is {"key1": "foo", "key2": "bar"}
http://0.0.0.0:5006/buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D
ISO-8859-1
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>500 Internal Server Error</title>
<h1>Internal Server Error</h1>
<p>The server encountered an internal error and was unable to complete your request. Either the server is overloaded or there is an error in the application.</p>
{'Content-Type': 'text/html', 'Content-Length': '291', 'Date': 'Fri, 09 Sep 2016 22:16:44 GMT', 'Server': 'Werkzeug/0.11.10 Python/3.5.2'}
Server console output:
* Running on http://0.0.0.0:5006/ (Press CTRL+C to quit)
[2016-09-09 18:30:33,445] ERROR in app: Exception on /buy [GET]
Traceback (most recent call last):
File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1988, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1641, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1544, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/usr/local/lib/python3.5/dist-packages/flask/_compat.py", line 33, in reraise
raise value
File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1639, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1625, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
TypeError: test() missing 2 required positional arguments: 'key1' and 'key2'
127.0.0.1 - - [09/Sep/2016 18:30:33] "GET /buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D HTTP/1.1" 500 -
I take this as showing that the subprocess call is for some reason not able to
decipher key1 and key2 from the URL so is failing when it runs "fortune None
None".
Answer: The problem seems to be in your payload.
The payload needs to be a dictionary. You're giving it a string.
`sys.argv[2]` will be a string, even if you format the text to look like a
dictionary. So unless there's something missing from your client code snippet,
payload isn't actually a dictionary like requests would expect.
I can infact confirm this by looking at the URL being generated, which is:
http://0.0.0.0:5006/buy?%7B%22key1%22:%20%22foo%22,%20%22key2%22:%20%22bar%22%7D
Had the payload been a dictionary and correctly encoded, it would've looked
something like this:
http://0.0.0.0:5006/buy?key1=foo&key2=bar
To see what type payload your really is, do `print(type(payload))` somewhere
inside the `test` function.
Once you've converted your `payload` into a proper python dictionary (you'll
have to parse your `sys.argv[2]`), then requests should work as expected.
|
python comprehension trubleshooting
Question:
base=2
digits=set(range(base))
key=range(base**3)
dict={ k:[a,b,c] for k in key for a in digits for b in digits for c in digits}
print(dict)
the output is:
{0: [1, 1, 1], 1: [1, 1, 1], 2: [1, 1, 1], 3: [1, 1, 1], 4: [1, 1, 1], 5: [1, 1, 1], 6: [1, 1, 1], 7: [1, 1, 1]}
I wonder why the output for [a,b,c] is [1,1,1], not the same as the:
[[a,b,c] for a in digits for b in digits for c in digits]
which is:
[[0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1],[1,1,0], [1, 1, 1]]
Answer: We can simplify the dictionary comprehension to something like:
In [728]: {k:[a,b] for k in [1,2] for a in [1,2] for b in [1,2]}
Out[728]: {1: [2, 2], 2: [2, 2]}
It is the equivalent to constructing a dictionary with nested loops:
In [731]: d={}
In [732]: for k in [1,2]:
...: for a in [1,2]:
...: for b in [1,2]:
...: d[k]=[a,b]
...:
In [733]: d
Out[733]: {1: [2, 2], 2: [2, 2]}
But to clarify what is going, let's use a defaultdict, and append the new
values:
In [730]: from collections import defaultdict
In [734]: d=defaultdict(list)
In [735]: for k in [1,2]:
...: for a in [1,2]:
...: for b in [1,2]:
...: d[k].append([a,b])
...:
In [736]: d
Out[736]:
defaultdict(list,
{1: [[1, 1], [1, 2], [2, 1], [2, 2]],
2: [[1, 1], [1, 2], [2, 1], [2, 2]]})
It is writing each `[a,b]` pair to the dictionary, but you only end up seeing
the last pair for each key.
And yes, `@user2357112's` comprehension does the same thing
In [737]: {k:[[a,b] for a in [1,2] for b in [1,2]] for k in [1,2]}
Out[737]: {1: [[1, 1], [1, 2], [2, 1], [2, 2]], 2: [[1, 1], [1, 2], [2, 1], [2, 2]]}
|
How to nest or jointly use two template tags in Django templates?
Question: I'm trying to use template filters to do run a loop, but I'm unable to combine
two python statements within the same statement/template. Whats the correct
way to combine two variables in a template? Please see the syntax and
explanation below:
I'm building a forum with a double index, meaning, I've got a col-md-2 with
list of categories. Each categories has forums, and depending on which
category is clicked, that category's forums populate the next col-md-2. The
remaining col-md-8 gets its content based on which category and which forum is
selected.
My logic:
I've defined a template tag that loads the list of categories, which never
change irrespective of which page gets loaded or which category or forum is
selected. So that works fine. But based on the selected category, my second
column needs to get populated. For this, I'm trying to define a custom filter
(below). However, I'm not sure how to use this as it needs to be passed to
another template where it runs a loop to render the html. Even if I create the
for loop in this template (rather than passing it to another), I still have to
do nested template tags, something like: `{% for forum in {{
forum.category|forumindexlistbycategory }} %}` In either cases, I get an error
of the type `Invalid filter: 'forumindexlistbycategory'` or `"with" in
u'include' tag needs at least one keyword argument`.
I've defined the following custom template filter in my pybb_tags.py:
from pybb.models import Forum
@register.filter
def forumindexlistbycat(category):
forumlistbycat = Forum.objects.filter(category=category)
return forumlistbycat
And in my template, I'm trying to load it as follows:
{% load i18n pybb_tags %}
<div class='category'>
{% if category %}
<h3>{{ category }}</h3>
{% include 'pybb/forumindex_list.html' with forum_list=category.forums_accessed category=category parent_forum='' %}
{% else %}
<h3>{{ forum.category }}</h3>
{% include 'pybb/forumindex_list.html' with forum_list= %}{{ forum.category|forumindexlistbycategory }}
{% endif %}
</div>
Answer: So you must first properly register template tag.
from django import template
from pybb.models import Forum
register = template.Library()
@register.filter
def forumindexlistbycat(category):
forumlistbycat = Forum.objects.filter(category=category)
return forumlistbycat
Place code from above in file named as your filter, so
`forumindexlistbycat.py` and move this file to templatetags folder in your
app. If you don't have this folder you must create it. Don't forget to add
empty file `__init__.py` in your templatetags folder. And now you could use it
in template, so:
{% load i18n forumindexlistbycat %}
When your templatetag is registered you load it by it's name. And then you use
it like:
{% include 'pybb/forumindex_list.html' with forum_list=forum.category|forumindexlistbycategory %}
Check for more info - [Guide on Custom template tags and
filters.](https://docs.djangoproject.com/en/1.10/howto/custom-template-tags/)
|
Importing functions in R
Question: In Python we have chance to import a certain function from a library with a
command "import _function_ from _library_ **as smth**. Do we have something
similar in R? I know that we can call the function like "_library_
::_function_()", my question mostly refers to the "as" part.
Answer: It is not common and not necessary to do this in R. The assignment operator
`<-` can be used to give a new name to an existing function. For example, one
could define a function that does exactly the same as lubridate's, `year()`
function with:
asYear <- lubridate::year
One could argue that, by doing so, the `year()` function has been "imported"
from the `lubridate` package and that it is now called `asYear()`. In fact,
the new function does just the same (which is no surprise, simply because it
_is_ the same):
asYear(Sys.Date())
#[1] 2016
So it is possible to construct an analogy to "from package import as", but it
is not a good idea to do this. Here are a few reasons I can think of:
* Debugging a code where library functions have been renamed will be much more difficult.
* The documentation is not available for the renamed function. In this example, `?asYear` won't work, in contrast to `?lubridate::year` or `library(lubridate); help(year)`.
* The function is not only renamed but it is copied, which clutters the environment and is inefficient in terms of memory usage.
* The maintenance of the code becomes unnecessarily difficult. If another programmer (or the original programmer a few years later) looks at a code containing such a redefinition of a function, it will be harder for her or him to understand what this function is doing.
There are probably more reasons, but I hope that this is sufficient to
discourage the use of such a construction. Different programming languages
have different peculiarities, and as a programmer it is necessary to adapt to
them. What is common in Python can be awkward in R, and vice versa.
A simple and commonly used way to handle such a standard situation in R is to
use `library()` to load the entire namespace of the package containing the
requested function:
library (lubridate)
year(Sys.Date())
However, one should be aware of possible namespace clashes, especially if many
libraries are loaded simultaneously. Different functions could be defined with
the same name in different packages. A well-known example thereof are the
contrasting implementations of the `lag()` function in the `dplyr` and `stats`
package. In such cases one can use the double colon operator `::` to resolve
the namespace that should be addressed. This would be similar to the use of
"from" in the case of "import", but such a specification would be needed each
time the function is called.
lubridate::year(Sys.Date())
#[1] 2016
|
Using csv training data in tensorflow RNN
Question: I am fairly new to tensorflow, and did the obligatory MNIST tutorial
successfully.
I am trying to train a simple RNN with a set of CSV data. The data is 33
features and a binary output variable at the end (so 34 columns).
I have implemented a csv reader that reads in one line at a time. I am trying
to read that line and pass it into my tensorflow graph. I feel like the
"TensorFlow-way" is starting to become more clear but also that there are some
fundamental pieces missing - particularly as it relates to streaming data into
your model.
I have included an example of what I'm doing below. Most of the code has been
stripped for clarity, but the important part remains:
import tensorflow as tf
import sys
import datapipe as datapipe
learning_rate = 0.001
n_features = 33
n_hidden = 100 # number of features in the hidden layer - I just made this up
n_classes = 2 # 0 or 1 - a binary classification
x = tf.placeholder('float', [None, 1, n_features])
y = tf.placeholder('float', [None, n_classes])
transform = tf.transpose(x)
with tf.session() as sess:
sess.run(tf.initialize_all_variables())
coord = tf.train.Coordinator()
threads = tf.train.start_queue_runners(coord=coord)
datapipe = datapipe.Datapipe(filename='training.csv', features=33, epochs=100)
while not coord.should_stop():
nextline = datapipe.nextline()
# I basically want to run "transform" with the nextline of the csv file
stuff = sess.run(transform, feed_dict={ x: nextline })
coord.request_stop()
coord.join(threads)
And `datapipe` is:
import tensorflow as tf
class Datapipe:
def __init__(self, filename=None, features=None, epochs=100):
self.filename = filename
self.features = features
self.epochs = epochs
self.defaults = []
for i in range(self.features):
self.defaults.append([]) # require all fields to be present
def nextline(self):
file_queue = tf.train.string_input_producer([self.filename], num_epochs=self.epochs, shuffle=False)
reader = tf.TextLineReader()
key, csv_str = reader.read(file_queue)
return tf.pack(tf.decode_csv(csv_str, record_defaults=self.defaults))
When I run this example, I get the error:
`TypeError: The value of a feed cannot be a tf.Tensor object. Acceptable feed
values include Python scalars, strings, lists, or numpy ndarrays.`
Thank you for all your help!
**EDIT**
My question is essentially: _How can I feed file data (e.g. csv) into a
tensorflow model?_ (the tutorials weren't helpful)
**EDIT 12/09/2016**
Based on the answer by Sergii, I am doing this now:
with open('../data/training2.csv') as f:
reader = csv.reader(f)
for line in reader:
arr = np.array(line)
x = arr[0:len(arr)-1:1]
y = arr[len(arr)-1:len(arr):1]
sess.run(transform, feed_dict={ x: x, y: y })
Answer: `Tensor` objects can not be values of `feed_dict`, it takes actual values,
like `numpy` arrays, strings, etc as an input. See for example [this
issue.](https://github.com/tensorflow/tensorflow/issues/3389)
Try to modify `nextline` method and do not create a `Tensor` there but
transform your csv strings into numpy arrays instead.
|
Python paramiko: redirecting stderr is affected by get_pty = True
Question: I am trying to implement an ssh agent that will allow me later, among other
things, to execute commands in blocking mode, where output is being read from
the channel as soon as it is available.
Here's what I have so far:
from paramiko import client
class SSH_Agent:
def __init__(self, server_name, username = getpass.getuser(), password = None, connection_timeout = CONNECTION_TIMEOUT):
self.ssh_agent = client.SSHClient()
self.ssh_agent.set_missing_host_key_policy(client.AutoAddPolicy())
self.ssh_agent.connect(server_name, username = username, password = password if password is not None else username, timeout = connection_timeout)
def execute_command(self, command, out_streams = [sys.stdout], err_streams = [sys.stderr], poll_intervals = POLL_INTERVALS):
stdin, stdout, stderr = self.ssh_agent.exec_command(command)
channel = stdout.channel
stdin.close()
channel.shutdown_write()
while not channel.closed or channel.recv_ready() or channel.recv_stderr_ready():
got_data = False
output_channels = select.select([channel], [], [], poll_intervals)[0]
if output_channels:
channel = output_channels[0]
if channel.recv_ready():
for stream in out_streams:
stream.write(channel.recv(len(channel.in_buffer)))
stream.flush()
got_data = True
if channel.recv_stderr_ready():
for stream in err_streams:
stream.write(channel.recv_stderr(len(channel.in_stderr_buffer)))
stream.flush()
got_data = True
if not got_data \
and channel.exit_status_ready() \
and not channel.recv_ready() \
and not channel.recv_stderr_ready():
channel.shutdown_read()
channel.close()
break
return channel.recv_exit_status()
(this implementation is based on one I found somewhere here in SO)
When I test it, it works fine, except that I get this when executing commands:
tput: No value for $TERM and no -T specified
I read a bit online, and found out that happens because there's no actual
terminal behind the `ssh` session.
So, I tried to call `paramiko`'s `exec_command()` with `get_pty = True`:
stdin, stdout, stderr = self.ssh_agent.exec_command(command, get_pty = True)
But then I found out that I'm losing the ability to get data to `stderr` on
the channel (everything goes to `stdout` for some reason, i.e.
`channel.recv_stderr_ready()` is never `True`). Indeed, I found in the doc the
following:
> recv_stderr_ready()
>
> Returns true if data is buffered and ready to be read from this channel’s
> stderr stream. **Only channels using exec_command or invoke_shell without a
> pty will ever have data on the stderr stream**.
> Returns: True if a recv_stderr call on this channel would immediately
> return at least one byte; False otherwise.
How can I have both?
In other words, how can I get rid of this:
tput: No value for $TERM and no -T specified
while still having the ability to direct `stderr` to wherever I choose?
**EDIT** :
I Just had an idea...
Can I somehow define this `TERM` variable in the remote shell to get rid of
that error? Is this a common approach or is it a bad workaround that only
hides the problem?
Answer: Assuming you're running on a *nix system, you could try setting `TERM` in your
command environment instead of creating a pty.
For example:
def execute_command(self, command, out_streams = [sys.stdout], err_streams = [sys.stderr], poll_intervals = POLL_INTERVALS):
# Pre-pend required environment here
command = "TERM=xterm " + command
stdin, stdout, stderr = self.ssh_agent.exec_command(command)
channel = stdout.channel
stdin.close()
channel.shutdown_write()
...
This should still allow you access to the separate streams using your existing
code.
|
Split a 3D numpy array into 3D blocks
Question: I would like to split a 3D numpy array into 3D blocks in a 'pythonic' way. I
am working with image sequences that are somewhat large arrays
(1000X1200X1600), so I need to split them into pieces to do my processing.
I have written functions to do this, but I am wondering if there is a native
numpy way to accomplish this - numpy.split does not seem to do what I want for
3D arrays (but perhaps I don't understand its functionality)
To be clear: the code below accomplishes my task, but I am seeking a faster
way to do it.
def make_blocks(x,t):
#x should be a yXmXn matrix, and t should even divides m,n
#returns a list of 3D blocks of size yXtXt
down = range(0,x.shape[1],t)
across = range(0,x.shape[2],t)
reshaped = []
for d in down:
for a in across:
reshaped.append(x[:,d:d+t,a:a+t])
return reshaped
def unmake_blocks(x,d,m,n):
#this takes a list of matrix blocks of size dXd that is m*n/d^2 long
#returns a 2D array of size mXn
rows = []
for i in range(0,int(m/d)):
rows.append(np.hstack(x[i*int(n/d):(i+1)*int(n/d)]))
return np.vstack(rows)
Answer: Here are vectorized versions of those loopy implementations using a
combination of permuting dims with
[`np.transpose`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.transpose.html)
and
[`reshaping`](http://docs.scipy.org/doc/numpy/reference/generated/numpy.reshape.html)
-
def make_blocks_vectorized(x,d):
p,m,n = x.shape
return x.reshape(-1,m//d,d,n//d,d).transpose(1,3,0,2,4).reshape(-1,p,d,d)
def unmake_blocks_vectorized(x,d,m,n):
return np.concatenate(x).reshape(m//d,n//d,d,d).transpose(0,2,1,3).reshape(m,n)
* * *
Sample run for `make_blocks` -
In [120]: x = np.random.randint(0,9,(2,4,4))
In [121]: make_blocks(x,2)
Out[121]:
[array([[[4, 7],
[8, 3]],
[[0, 5],
[3, 2]]]), array([[[5, 7],
[4, 0]],
[[7, 3],
[5, 7]]]), ... and so on.
In [122]: make_blocks_vectorized(x,2)
Out[122]:
array([[[[4, 7],
[8, 3]],
[[0, 5],
[3, 2]]],
[[[5, 7],
[4, 0]],
[[7, 3],
[5, 7]]], ... and so on.
Sample run for `unmake_blocks` -
In [135]: A = [np.random.randint(0,9,(3,3)) for i in range(6)]
In [136]: d = 3
In [137]: m,n = 6,9
In [138]: unmake_blocks(A,d,m,n)
Out[138]:
array([[6, 6, 7, 8, 6, 4, 5, 4, 8],
[8, 8, 3, 2, 7, 6, 8, 5, 1],
[5, 2, 2, 7, 1, 2, 3, 1, 5],
[6, 7, 8, 2, 2, 1, 6, 8, 4],
[8, 3, 0, 4, 4, 8, 8, 6, 3],
[5, 5, 4, 8, 5, 2, 2, 2, 3]])
In [139]: unmake_blocks_vectorized(A,d,m,n)
Out[139]:
array([[6, 6, 7, 8, 6, 4, 5, 4, 8],
[8, 8, 3, 2, 7, 6, 8, 5, 1],
[5, 2, 2, 7, 1, 2, 3, 1, 5],
[6, 7, 8, 2, 2, 1, 6, 8, 4],
[8, 3, 0, 4, 4, 8, 8, 6, 3],
[5, 5, 4, 8, 5, 2, 2, 2, 3]])
* * *
Alternative to `make_blocks` with [`view_as_blocks`](http://scikit-
image.org/docs/dev/api/skimage.util.html#skimage.util.view_as_blocks) -
from skimage.util.shape import view_as_blocks
def make_blocks_vectorized_v2(x,d):
return view_as_blocks(x,(x.shape[0],d,d))
* * *
**Runtime test**
1) `make_blocks` with original and `view_as_blocks` based approaches -
In [213]: x = np.random.randint(0,9,(100,160,120)) # scaled down by 10
In [214]: %timeit make_blocks(x,10)
1000 loops, best of 3: 198 µs per loop
In [215]: %timeit view_as_blocks(x,(x.shape[0],10,10))
10000 loops, best of 3: 85.4 µs per loop
2) `unmake_blocks` with original and `transpose+reshape` based approaches -
In [237]: A = [np.random.randint(0,9,(10,10)) for i in range(600)]
In [238]: d = 10
In [239]: m,n = 10*20,10*30
In [240]: %timeit unmake_blocks(A,d,m,n)
100 loops, best of 3: 2.03 ms per loop
In [241]: %timeit unmake_blocks_vectorized(A,d,m,n)
1000 loops, best of 3: 511 µs per loop
|
Why is my function that I made getting the TypeError: f() takes 0 positional arguments but 1 was given
Question: I am working with pandas DataFrames and I am adding new columns for more
advanced analysis. My f function is giving me an error TypeError: f() takes 0
positional arguments but 1 was given. I can't figure out why, I have my f
function documented in the code if you need to know what it does.
from pandas_datareader import data as dreader
import pandas as pd
from datetime import datetime
import dateutil.parser
# Sets the max rows that can be displayed
# when the program is executed
pd.options.display.max_rows = 200
# df is the name of the dataframe, it is
# reading the csv file containing date loaded
# from yahoo finance(Date,Open,High,Low,Close
# volume,adj close,)the name of the ticker
# is placed before _data.csv i.e. the ticker aapl
# would have a csv file named aapl_data.csv.
df = pd.read_csv("cde_data.csv")
# the following code will allow for filtering of the datafram
# based on the year, day of week (dow), and month. It then gets
# applied to the dataframe and then can be used to sort data i.e
# print(df[(df.year == 2015) & (df.month == 5) & (df.dow == 4)])
# which will give you all the days in the month of May(df.month == 5),
# that fall on a Thursday(df.dow == 4), in the year 2015
# (df.year == 2015)
#
# Month Day Year
# January = 1 Monday = 1 The year will be dispaly in a four
# February = 2 Tuesday = 2 digit format i.e. 2015
# March = 3 Wednesday = 3
# April = 4 Thursday = 4
# May = 5 Friday = 5
# June = 6
# July = 7
# August = 8
# September = 9
# October = 10
# November = 11
# December = 12
def year(x):
return(x.year)
def dow(x):
return(x.isoweekday())
def month(x):
return(x.month)
# f is a function that checks to see if the up_down column
# has a value that is greater than, less than, or equal to
# zero. The value in the up_down column is derived from
# subtracting the opening price of the stock(open column)
# from closing price of the stock(close column). If up_down
# has a negative value than the stocks price was Down, a positive
# value then Up, and no change is Flat
def f():
if up_down > 0:
x = Up
elif up_down < 0:
x = Down
else:
x = Flat
return (f)
df.reset_index()
df.Date = df.Date.apply(dateutil.parser.parse)
df['year'] = df.Date.apply(year)
df['dow'] = df.Date.apply(dow)
df['month'] = df.Date.apply(month)
df['up_down'] = df['Close'] - df['Open']
df['up_down_flat'] = df.up_down.apply(f)
df2= (df[(df.year > 1984) & (df.month == 5) & (df.dow == 1)])
print (df2)
This is the error
Traceback (most recent call last):
File "dframt1est.py", line 77, in <module>
df['up_down_flat'] = df.up_down.apply(f)
File "C:\Users\Zac\AppData\Local\Programs\Python\Python35-32\lib\site-packages\pandas\core\series.py", line 2220, in apply
mapped = lib.map_infer(values, f, convert=convert_dtype)
File "pandas\src\inference.pyx", line 1088, in pandas.lib.map_infer (pandas\lib.c:63043)
TypeError: f() takes 0 positional arguments but 1 was given
Press any key to continue . . .
This a sample csv
Date Open High Low Close Volume Adj Close \
0 1990-04-12 26.875000 26.875000 26.625 26.625 6100 250.576036
1 1990-04-16 26.500000 26.750000 26.375 26.750 500 251.752449
2 1990-04-17 26.750000 26.875000 26.750 26.875 2300 252.928863
3 1990-04-18 26.875000 26.875000 26.500 26.625 3500 250.576036
4 1990-04-19 26.500000 26.750000 26.500 26.750 700 251.752449
5 1990-04-20 26.750000 26.875000 26.750 26.875 2100 252.928863
6 1990-04-23 26.875000 26.875000 26.750 26.875 700 252.928863
7 1990-04-24 27.000000 27.000000 26.000 26.000 2400 244.693970
8 1990-04-25 25.250000 25.250000 24.875 25.125 9300 236.459076
Answer: You're passing `f` as the function in `apply`. That function is called for
each row in the dataframe, and needs to take the row as its parameter.
Note also that you're returning the function itself as the result; I'm pretty
sure you meant to return `x` not `f`. Also, you don't seem to have defined
`Up`, `Down` and `Flat` anywhere.
|
how can i correct this Regex phone number extractor in python
Question: The results i'm getting when i run this after copying a set of UK phone
numbers to the clipboard are coming out in a very bizarre kind of way. (i have
imported both modules before you ask)
phoneRegex = re.compile(r'''(
(\d{5}|\(\d{5}\))? #area code
(\s|-|\.)? #separator
(\d{6}) #main 6 digits
)''', re.VERBOSE)
emailRegex = re.compile(r'''(
[a-zA-Z0-9._%+-]+ #username
@ #obligatory @ symbol
[a-zA-Z0-9.-]+ #domain name
(\.[a-zA-Z]{2,5}) #dot-something
)''', re.VERBOSE)
text = str(pyperclip.paste())
matches = []
for groups in phoneRegex.findall(text):
phoneNum = '-'.join([groups[1], groups[3]])
if groups[3] != '':
phoneNum += ' ' + groups[3]
matches.append(phoneNum)
for groups in emailRegex.findall(text):
matches.append(groups[0])
if len(matches) > 0:
pyperclip.copy('\n'.join(matches))
print('Copied to clipboard:')
print('\n'.join(matches))
else:
print('No phone numbers or email addresses found')
The mistake is somewhere here:
for groups in phoneRegex.findall(text):
phoneNum = '-'.join([groups[1], groups[3]])
if groups[3] != '':
phoneNum += ' ' + groups[3]
matches.append(phoneNum)
**The numbers copied to clipboard:**
07338 754433
01265768899
(01283)657899
**Expected results:**
Copied to clipboard:
07338 754433
01265 768899
01283 657899
**return results:**
Copied to clipboard:
07338-754433 754433
-012657 012657
(01283)-657899 657899
Answer: I see three issues:
1. the python code joins the two parts of the phone number together with a `-` and then adds a space and the third part again:
phoneNum = '-'.join([groups[1], groups[3]])
if groups[3] != '':
phoneNum += ' ' + groups[3]
Since `groups[3]` will always not be blank, what you need to do is:
if groups[1] != '':
phoneNum = ' '.join(groups[1], groups[3])
else:
phoneNum = groups[3]
2. Your `phoneRegex` regular expression is not anchored to the beginning and end of the lines. You need to (a) compile it with the `re.MULTILINE` option and (b) anchor the regular expression between `^` and `$`:
phoneRegex = re.compile(r'''^(
(\d{5}|\(\d{5}\))? #area code
(\s|-|\.)? #separator
(\d{6}) #main 6 digits
)$''', re.VERBOSE + re.MULTILINE)
This will prevent a long string of digits with no separator as being just
group 3 with a bunch of digits after it.
3. Your match for the area code includes the matched parentheses within the group match. To fix this, you either need to change the regular expression to make sure the parentheses are not part of the group, or you need to change your code to strip the parentheses out if needed.
* Avoid parentheses in the regular expression:
(?:(\d{5})|\((\d{5})\))? #area code
The `(?:...)` is a non-grouping form of parentheses, so it won't be returned
by the find. Within that, you have two alternatives: 5 digits in a group -
`(\d{5})` \- or literal parentheses that enclose 5 digits in a group -
`\((\d{5})\)`.
However, this change also affects your phone number recombination logic,
because your area code is _either_ in `groups[1]` or `groups[2]`, and your
main number is now in `groups[4]`.
if groups[1] != '':
phoneNum = ' '.join(groups[1], groups[4])
elif groups[2] != '':
phoneNum = ' '.join(groups[2], groups[4])
else:
phoneNum = groups[4]
* This could be made easier by changing the outer set of parentheses and the parentheses around the separator into non-grouping parentheses. You could then do a single join on a filtered result of the groups:
phoneRegex = re.compile(r'''(?:
(?:(\d{5})|\((\d{5})\))? #area code
(?:\s|-|\.)? #separator
(\d{6}) #main 6 digits
)''', re.VERBOSE)
# ...
phoneNum = ' '.join([group for group in groups if group != ''])
The modified `phoneRegex` ensures that the returned `groups` contain only an
optional area code in `groups[0]` or `groups[1]` followed by the main number
in `groups[2]`, no extraneous matches returned. The code then filters out any
groups that are empty and returns the rest of the groups joined by a space.
* Strip parentheses in code:
if groups[1] != '':
phoneNum = ' '.join(groups[1].lstrip('(').rstrip(')'), groups[3])
else:
phoneNum = groups[3]
|
python os.walk returns nothing
Question: I have a problem with using `os.walk` on Mac. If I call it from `python
terminal`, it works perfect, but if I call it via a `python script`, it
returns empty list. For example:
import os
path = "/Users/temp/Desktop/test/"
for _ ,_ , files in os.walk(path):
test = [my_file for my_file in files]
print test
then, it prints:
[]
and I am pretty sure that the `path` does exists. Any idea what is the
problem?
Answer: You most likely need to intantiate the test list outside the for loop, for
this to work.
import os
path = "/Users/temp/Desktop/test/"
test = []
for _ ,_ , files in os.walk(path):
test.extend([my_file for my_file in files])
print test
|
The last element (bitstring.BitArray) in list is incorrect after XORing python
Question: I have snippet of code:
#!/usr/bin/python3
from bitstring import BitArray
import itertools
# Helper functions
def get_bitset_by_letter(letter, encoding):
return encoding[letter] if letter in encoding else None
def get_letter_by_bitset(bitset, encoding):
for letter, encoding_bitset in encoding.items():
if bitset == encoding_bitset:
return letter
return None
class Word():
def __init__(self, word, encoding):
self.encoding = encoding
self.as_word = ''
self.as_bits = []
if isinstance(word, str):
self.as_word = word
self._encode_to_bits()
elif all(isinstance(i, BitArray) for i in word):
self.as_bits = word
self._decode_to_word()
else:
raise('class Word: Unexpected type of word arg')
def __ixor__(self, other):
for index, bitset in enumerate(other.as_bits):
print('In XOR: %d-%s-%s' % (index, self.as_bits[index], bitset))
self.as_bits[index] ^= bitset
return self
def as_bits(self):
return self.as_bits
def as_word(self):
return self.as_word
def _encode_to_bits(self):
for letter in self.as_word:
bitset = get_bitset_by_letter(letter, self.encoding)
if bitset is None:
raise('Can not find bitset by given letter %s in encoding.' % str(letter))
else:
self.as_bits.append(bitset)
def _decode_to_word(self):
for bitset in self.as_bits:
letter = get_letter_by_bitset(bitset, self.encoding)
if letter is None:
raise('Can not find letter by given bitset %s in encoding.' % str(bitset))
else:
self.as_word += letter
def main():
encoding = {
'A': BitArray(bin='000'),
'B': BitArray(bin='001'),
'C': BitArray(bin='010'),
'D': BitArray(bin='011'),
'E': BitArray(bin='100'),
'F': BitArray(bin='101'),
'G': BitArray(bin='110'),
'H': BitArray(bin='111'),
}
print(encoding)
print()
word_1 = Word('ABCHE', encoding)
print('word_1 = %s' % word_1.as_bits)
word_2 = Word('FGDEA', encoding)
print('word_2 = %s' % word_2.as_bits)
word_1 ^= word_2
print(word_1.as_bits)
if __name__ == '__main__':
main()
Details: I use `bitstring.BitArray` class to represent letter as set of three
bits and code words this way (some kind of hometask). Also, my code should be
able to do XOR for this words. Problem: Every last element for `other.as_bits`
list has incorrect value, when it arrives to function. If I comment the line
#self.as_bits[index] ^= bitset
the last value of the `other.as_bits` is correct. What am I doing wrong?
Here is the output (python 3.5.2):
{'C': BitArray('0b010'), 'B': BitArray('0b001'), 'G': BitArray('0b110'),
'F': BitArray('0b101'), 'A': BitArray('0b000'), 'H': BitArray('0b111'),
'D': BitArray('0b011'), 'E': BitArray('0b100')}
word_1 = [BitArray('0b000'), BitArray('0b001'), BitArray('0b010'), BitArray('0b111'), BitArray('0b100')]
word_2 = [BitArray('0b101'), BitArray('0b110'), BitArray('0b011'), BitArray('0b100'), BitArray('0b000')]
In XOR: 0-0b000-0b101
In XOR: 1-0b001-0b110
In XOR: 2-0b010-0b011
In XOR: 3-0b111-0b100
In XOR: 4-0b100-0b101
[BitArray('0b101'), BitArray('0b111'), BitArray('0b001'), BitArray('0b011'), BitArray('0b001')]
From this you can see, that the last 4th element of `word_2` in `__ixor__` is
`0b101`, should be `0b000`
Thank you!
Answer: I solved this using `__xor__` instead of `__ixor__`.
|
Transforming string output to JSON
Question: I'm getting some data from an external system (Salesforce Marketing Cloud)
over API and I'm getting the data back in the format below:
Results: [(List){
Client =
(ClientID){
ID = 113903
}
PartnerKey = None
CreatedDate = 2013-07-29 04:43:32.000073
ModifiedDate = 2013-07-29 04:43:32.000073
ID = 1966872
ObjectID = None
CustomerKey = "343431CD-031D-43C7-981F-51B778A5A47F"
ListName = "PythonSDKList"
Category = 578615
Type = "Private"
Description = "This list was created with the PythonSDK"
ListClassification = "ExactTargetList"
}]
It's tantalisingly close to JSON, I'd like to format it like a JSON file so
that I can import it into our database. Any ideas of how I can do this?
Thanks
Answer: To convert your array, you could use var jsonString =
JSON.stringify(yourArray);
|
Setting scan coordinates in device options on pyinsane
Question: I use Sane's command line utility (`scanimage`) in order to scan films from
the transparency unit of my scanner. Here is the command that I have been
using with success:
scanimage --device-name pixma:04A9190D \
--source 'Transparency Unit' \
--resolution "4800" \
--format "tiff" \
--mode "color" \
-l "80.6" -x "56.2" -t "25.8" -y "219.2" \
> scan.tiff
I have decided to move this to Python code, using `pyinsane` in order to
enable further integration with my image-processing workflow. This should
supposedly give the following in Python code:
import pyinsane.abstract as pyinsane
device = pyinsane.get_devices()[0]
device.options['resolution'].value = 4800
device.options['mode'].value = 'Color'
device.options['source'].value = 'Transparency Unit'
# Setting coordinates to non-integers fails
device.options['tl-y'].value = 25.8
device.options['tl-x'].value = 80.6
device.options['br-y'].value = 219.2
device.options['br-x'].value = 56.2
scan_session = device.scan(multiple=False)
try:
while True:
scan_session.scan.read()
except EOFError:
pass
image = scan_session.images[0]
But my first trials have been unsuccessful because I can not figure out how to
set the scan coordinates `pyinsane`. As you see, I have found the appropriate
options, but I don't know what unit they are in. `scanimage` takes the
coordinates in millimetres by default, but `pyinsane` only takes integers. I
have tried using pixel coordinates to no avail. I wonder what units the
coordinate parameters take, and whether I am using them in the right order.
Answer: pyinsane's option descriptions actually say the values are in milimeters:
Option: br-x
Title: Bottom-right x
Desc: Bottom-right x position of scan area.
Type: <class 'pyinsane.rawapi.SaneValueType'> : Fixed (2)
Unit: <class 'pyinsane.rawapi.SaneUnit'> : Mm (3)
Size: 4
Capabilities: <class 'pyinsane.rawapi.SaneCapabilities'> :[ Automatic, Soft_select, Soft_detect,]
Constraint type: <class 'pyinsane.rawapi.SaneConstraintType'> : Range (1)
Constraint: (0, 14160319, 0)
Value: 20
But they are not! I divided the range maximum for `br-x` variable by the width
of the scanning area of my scanner, and I got to the number 65536 (which is
2^16). Setting the coordinates to the millimetre value times 65536 works.
Maybe these values define the number of steps of the stepper motor?
Also not that while scanimage interprets the `-x` and `-y` switches as width
and length, and the `-l` and `-t` switches as offset, pyinsane takes bottom-
right x (`br-x`), top-left y (`tl-y`), etc.
|
Encoding issue when writing to CSV file in Python
Question: I have some encoding issue while writing an array to CSV.
Code:
import csv
a = [u'eNTfxfwc', 'Pushkar', 'Waghulde', '[email protected]', 'Los Angeles', '', 'UNITED STATES', '2652 Ellendale Pl # 9', '', 'Los Angeles, UNITED STATES', u'90007', u'', u'(213)458-2091', u'None', u'', u'imports/jobvite/eNTfxfwc.txt', '07/17/2012', u'2012-24', u'pWOjJfww', u'Undefined', u'Import', u' ', '', '', u'\u8edf\u9ad4\u5de5\u7a0b\u5e2b (Software Engineer)', '07/17/2012', '', u'Undefined', '', u'None', '', '', u'Undefined', u'Not Considered', '', '07/17/2012', u'import/jobvite/eNTfxfwc_1.txt']
f3 = open('test.csv', 'at')
writer = csv.writer(f3,delimiter = ',', lineterminator='\n',quoting=csv.QUOTE_ALL)
writer.writerow(a).encode('utf-8')
How can I fix this issue and write array into CSV file?
Answer: Just encode the elements in a before writing them
a = [i.encode('utf-8') for i in a]
f3 = open('test.csv', 'at')
writer = csv.writer(f3,delimiter = ',',lineterminator='\n',quoting=csv.QUOTE_ALL)
writer.writerow(a)
|
Python Error message when opening .txt file / change in working directory
Question: I’m a new python user and have written a python script that prompts for the
name of a text file (.txt) to be opened and read by the program.
name = raw_input("Enter file:")
if len(name) < 1:
name = "test.txt"
handle = open(name)
This was working perfectly fine for me, but then I installed Anaconda and
tried to run the same `.py` script and kept getting this error message:
$ python /test123.py
Enter file:'test.txt'
Traceback (most recent call last):
File "/test123.py", line 26, in <module>
handle = open(name)
IOError: [Errno 2] No such file or directory: "'test.txt'"
even though I had changed nothing in the script or text file and the text file
definitely exists in the directory. So then I uninstalled Anaconda and deleted
the Anaconda directory as well as the PATH variable pointing to Anaconda in
the `~/.bash_profile` file to see if that would fix the problem. But I’m still
getting the same error :( What’s wrong? I do not get any error if I run the
file in the Terminal directly from TextWrangler, and my `.py` script contains
the shebang line `#!/usr/bin/env python`
My info:
$ sw_vers
ProductName: Mac OS X
ProductVersion: 10.11.6
BuildVersion: 15G1004
$ python
Python 2.7.12 (v2.7.12:d33e0cf91556, Jun 26 2016, 12:10:39)
[GCC 4.2.1 (Apple Inc. build 5666) (dot 3)] on darwin
$ which python
/Library/Frameworks/Python.framework/Versions/2.7/bin/python
$ /usr/bin/python -V
Python 2.7.10
Thanks so much for the time and help!
**EDIT:** I copied both the `test123.py` document and the `test.txt` file and
put them in the directory `/Users/myusername/Documents` and ran the script
again, this time inputting `/Users/myusername/Documents/test.txt` (without
quotes) at the "Enter file" prompt, and it worked fine! So now I need to
modify my question:
My understanding--and what had been working fine for me previously--is that
when a python script is run, the working directory automatically changes to
the directory that the `.py` file is inside. Is this not correct? Or how could
I change my settings back to that, since somehow it seemed to have been the
case before but now isn't.
**EDIT 2:** Now that I've realized the source of my original error was a
result of the working directory Python was using, I've been able to solve my
problem by adding the following lines to my `test123.py` script:
import os
abspath = os.path.abspath(__file__)
dname = os.path.dirname(abspath)
os.chdir(dname)
as suggested here: [python: Change the scripts working directory to the
script's own directory](http://stackoverflow.com/questions/1432924/python-
change-the-scripts-working-directory-to-the-scripts-own-directory)
However, I'm also selecting the answer provided by @SomethingSomething because
now that I've been able to change the working directory, the input
`'test.txt'` (with quotes) is not valid, but `test.txt` (without quotes) works
as desired.
(The only thing I'm still wondering about is why previously, when I was
running python scripts, the working directory automatically changed to the
directory that the `.py` file was inside, and then somehow the setting
changed, corresponding with my installation of Anaconda. However, now that
I've solved my problem this seems unimportant.)
Thanks everyone for the help!
Answer: It's very simple, your input is wrong:
Incorrect:
$ python /test123.py
Enter file:'test.txt'
Correct:
$ python /test123.py
Enter file:test.txt
do not include the `'` characters. It is not code, it's just standard input
|
Replacing an item in list with items of another list without using dictionaries
Question: I am developing a function in python. Here is my content:
list = ['cow','orange','mango']
to_replace = 'orange'
replace_with = ['banana','cream']
So I want that my list becomes like this after replacement
list = ['cow','banana','cream','mango']
I am using this function:
def replace_item(list, to_replace, replace_with):
for n,i in enumerate(list):
if i== to_replace:
list[n]=replace_with
return list
It outputs the list like this:
['cow', ['banana', 'cream'], 'mango']
So how do I modify this function to get the below output?
list = ['cow','banana','cream','mango']
NOTE: I found an answer here: [Replacing list item with contents of another
list](http://stackoverflow.com/questions/14962931/replacing-list-item-with-
contents-of-another-list) but I don't want to involve dictionaries in this. I
also want to modify my current function only and keep it simple and straight
forward.
Answer: This approach is fairly simple and has similar performance to @TadhgMcDonald-
Jensen's `iter_replace()` approach (3.6 µs for me):
lst = ['cow','orange','mango']
to_replace = 'orange'
replace_with = ['banana','cream']
def replace_item(lst, to_replace, replace_with):
return sum((replace_with if i==to_replace else [i] for i in lst), [])
print replace_item(lst, to_replace, replace_with)
# ['cow', 'banana', 'cream', 'mango']
Here's something similar using itertools, but it is slower (5.3 µs):
import itertools
def replace_item(lst, to_replace, replace_with):
return list(itertools.chain.from_iterable(
replace_with if i==to_replace else [i] for i in lst
))
Or, here's a faster approach using a two-level list comprehension (1.8 µs):
def replace_item(lst, to_replace, replace_with):
return [j for i in lst for j in (replace_with if i==to_replace else [i])]
Or here's a simple, readable version that is fastest of all (1.2 µs):
def replace_item(lst, to_replace, replace_with):
result = []
for i in lst:
if i == to_replace:
result.extend(replace_with)
else:
result.append(i)
return result
Unlike some answers here, these will do multiple replacements if there are
multiple matching items. They are also more efficient than reversing the list
or repeatedly inserting values into an existing list (python rewrites the
remainder of the list each time you do that, but this only rewrites the list
once).
|
former form fields no longer found by mechanize python script
Question: Let me start by apologizing for my utter newbness. I was asked by a friend a
couple years ago if I could write a program to automatically grab substitute
teaching openings. It wasn't an area I knew anything about, but a couple
tutorials allowed me to bang something out despite ignorance about html (and
more than a little about Python for that matter). Script worked great since
then but this year their site seems to have been redone and broke things,
pushing it far beyond my understanding.
My previous code that worked:
# Create a Browser instance
b = mechanize.Browser()
# Load the page
b.open(loginURL)
# Select the form
b.select_form(nr=0)
# Fill out the form
b['id'] = 'XXXXXXXXXX' # Note: I edited out my friend's login info here for privacy
b['pin'] = 'XXXX'
b.submit();
There is still only one form but the controls are now of type "hidden" and are
not the ones I directly need any longer. I can see the old fields in the html
when I examine it with developer mode and the names are the same but I can't
figure out (tried some things that didn't work) how I would access them now.
Here is the html:
<form id="loginform" name="loginform" method="post" action="https://www.aesoponline.com/login.asp?x=x&&pswd=&sso=">
<input type="hidden" name="location" value="">
<input type="hidden" name="qstring" value="">
<input type="hidden" name="absr_ID" value="">
<input type="hidden" name="foil" value="">
<div style="margin: auto; text-align:center;">
<div id="loginContainer" style="text-align: left;">
<div id="loginContent">
<div id="Div1" style="position:relative; left:65px;" class="hide-me-for-rebranding">
<a href="http://www.frontlinetechnologies.com">
<img src="images/frontlinelogo.png" border="0">
</a>
</div>
<div id="loginLoginBox" style="position:relative;">
<div id="loginAesopLogo" style="padding-bottom:0px;" class="hide-me-for-rebranding"></div>
<!--endloginAesopLogo-->
<div id="loginLoginFields" style="margin-top:0px;">
<br>
<table>
<tbody>
<tr height="25px">
<td width="30px"><span class="corrLoginFormText">ID:</span>
</td>
<td>
<input type="text" class="loginFormText" maxlength="80" id="txtLoginID" name="id" value="">
</td>
</tr>
<tr height="25px">
<td width="30px"><span class="corrLoginFormText">Pin:</span>
</td>
<td>
<input type="password" class="loginFormText" maxlength="20" id="txtPassword" name="pin">
</td>
</tr>
</tbody>
</table>
<table>
<tbody>
<tr height="30px">
<td width="75px" valign="top">
<a class="textButton" id="loginLink" name="loginLink" href="#"><span style="white-space:nowrap;">Login</span></a>
<input type="hidden" id="submitLogin" name="submitLogin" value="1">
</td>
<td>
<div id="loginhelp" style="float:right;">
<img src="images/icon.pinreminder.png" alt="pin" width="10" height="15" align="top"><a href="forgot_pin.asp">Pin Reminder</a>
<br>
<img src="images/icon.loginproblems.png" alt="login" width="11" height="17" align="top"> <a href="http://help.frontlinek12.com/Employee/Docs/ClientServicesHelpGuide-LoginProblems.pdf">Login Problems</a>
</div>
</td>
</tr>
</tbody>
</table>
</div>
<!--endloginLoginFields-->
<div id="errorLabel" style="position: absolute; top: 170px; left:5px;margin:0px;"><span class="assistanceText"></span>
</div>
</div>
<!--endloginLoginBox-->
<div id="loginContentText">
<span class="loginContentHeader">Welcome To Absence Management</span>
<br>
<span class="loginContentText">
You are about to enter Frontline Absence Management!<br> Please enter your ID and PIN to login to your account, or click the button below to learn more about Frontline's growing impact on education.</span>
<br>
<a class="textButton" href="http://www.frontlinek12.com/Products/Aesop.html"><span>Learn More</span></a>
</div>
<!--endloginContentText-->
</div>
<!--endLoginContent-->
<div id="loginFooterShading" class="hide-me-for-rebranding">
<div id="loginFooterLeft"></div>
<div id="loginFooterRight"></div>
</div>
<!--endloginFooterShading-->
<div id="loginFooter" style="text-align:center;width:725px;">
<a href="http://www.frontlinetechnologies.com/Privacy_Policy.html" style="color: rgb(153, 0, 0) ; font-size:9px;" target="_blank">Privacy Policy</a>
<br>© Frontline Technologies Group LLC <
<parm1>>
<br>All rights reserved. Protected under US Patents 6,334,133, 6,675,151, 7,430,519, 7,945,468 and 8,140,366 with additional patents pending.
</parm1>
</div>
<!--endloginReflections-->
</div>
<!--endLoginContainer-->
</div>
<!--end margin div -->
<!-- MODAL DIALOG -->
<div id="basicModalContent" style="display:none">
<span class="assistanceText"></span>
</div>
</form>
Any assistance would be greatly appreciated. Thank you very much.
Answer: Try something like this if that html code is exactly what is on that page.
When you put b.select_form(nr=0) there is the possibility that for some reason
the first form is not what you are selecting. By looking for the form name in
the b.select_form() you can ensure you find the correct form. Test it out and
see if it works out.
import mechanize
b = mechanize.Browser()
b.open(loginURL)
#since the actual form is named loginform just select it
b.select_form("loginform")
b['id'] name of login input field
b['pin']
b.submit()
|
Subsets and Splits